首页> 外文会议>AAAI Conference on Artificial Intelligence >Faster Gradient-Free Proximal Stochastic Methods for Nonconvex Nonsmooth Optimization
【24h】

Faster Gradient-Free Proximal Stochastic Methods for Nonconvex Nonsmooth Optimization

机译:更快的渐变近端随机方法,用于非凸起的非耦合NonsMooth优化

获取原文

摘要

Proximal gradient method has been playing an important role to solve many machine learning tasks, especially for the non-smooth problems. However, in some machine learning problems such as the bandit model and the black-box learning problem, proximal gradient method could fail because the explicit gradients of these problems are difficult or infeasible to obtain. The gradient-free (zeroth-order) method can address these problems because only the objective function values are required in the optimization. Recently, the first zeroth-order proximal stochastic algorithm was proposed to solve the non-convex nonsmooth problems. However, its convergence rate is O(1/√T) for the nonconvex problems, which is significantly slower than the best convergence rate O(1/T) of the zeroth-order stochastic algorithm, where T is the iteration number. To fill this gap, in the paper, we propose a class of faster zeroth-order proximal stochastic methods with the variance reduction techniques of SVRG and SAGA, which are denoted as ZO-ProxSVRG and ZO-ProxSAGA, respectively. In theoretical analysis, we address the main challenge that an unbiased estimate of the true gradient does not hold in the zeroth-order case, which was required in previous theoretical analysis of both SVRG and SAGA. Moreover, we prove that both ZO-ProxSVRG and ZO-ProxSAGA algorithms have O(1/T) convergence rates. Finally, the experimental results verify that our algorithms have a faster convergence rate than the existing zeroth-order proximal stochastic algorithm.
机译:近端渐变方法一直在发挥重要作用来解决许多机器学习任务,特别是对于非平滑问题。然而,在一些机器学习问题,如强盗模型和黑匣子学习问题,近端梯度方法可能会失败,因为这些问题的显式梯度是难以或不可行的。梯度(Zeroth Order)方法可以解决这些问题,因为在优化中只需要客观函数值。最近,提出了第一种零阶近端随机算法来解决非凸起的非凸形问题。然而,其收敛速率是O(1 /√T),用于非耦合问题,这显着慢于零级随机算法的最佳收敛速率O(1 / T),其中T是迭代号。为了填补这一差距,我们提出了一类更快的Zeroth近端随机方法,具有SVRG和SAGA的方差减少技术,分别表示为ZO-Proxsvrg和ZO-Proxsaga。在理论分析中,我们解决了对真正梯度的无偏见估计没有在零顺序案件中保持的主要挑战,这是在先前的SVRG和SAGA的理论分析中所必需的。此外,我们证明了Zo-ProxSVRG和ZO-ProxSaga算法都具有O(1 / T)收敛速率。最后,实验结果验证了我们的算法具有比现有的Zeroth近端随机算法更快的收敛速度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号