Optimal Rates for Zero-Order Convex Optimization: The Power of Two Function Evaluations

Information Theory, IEEE Transactions  (2015)

引用 461|浏览81
暂无评分
摘要
We consider derivative-free algorithms for stochastic and nonstochastic convex optimization problems that use only function values rather than gradients. Focusing on nonasymptotic bounds on convergence rates, we show that if pairs of function values are available, algorithms for -dimensional optimization that use gradient estimates based on random perturbations suffer a factor of at most in convergence rate over traditional stochastic gradient methods. We establish such results for both smooth and nonsmooth cases, sharpening previous analyses that suggested a worse dimension dependence, and extend our results to the case of multiple ( ) evaluations. We complement our algorithmic development with information-theoretic lower bounds on the minimax convergence rate of such problems, establishing the sharpness of our achievable results up to constant (sometimes logarithmic) factors.
更多
查看译文
关键词
optimization,lower bounds,on-line learning,statistical learning,zero-order optimization,convex programming,stochastic programming,vectors,convex functions,estimation theory,convergence
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要