谷歌浏览器插件
订阅小程序
在清言上使用

Convergence Analysis of AdaBound with Relaxed Bound Functions for Non-Convex Optimization

Neural Networks(2022)

引用 10|浏览23
暂无评分
摘要
Clipping on learning rates in Adam leads to an effective stochastic algorithm-AdaBound. In spite of its effectiveness in practice, convergence analysis of AdaBound has not been fully explored, especially for non-convex optimization. To this end, we address the convergence of the last individual output of AdaBound for non-convex stochastic optimization problems, which is called individual convergence. We prove that, with the iteration of the AdaBound, the cost function converges to a finite value and the corresponding gradient converges to zero. The novelty of this proof is that the convergence conditions on the bound functions and momentum factors are much more relaxed than the existing results, especially when we remove the monotonicity and convergence of the bound functions, and only keep their boundedness. The momentum factors can be fixed to be constant, without the restriction of monotonically decreasing. This provides a new perspective on understanding the bound functions and momentum factors of AdaBound. At last, numerical experiments are provided to corroborate our theory and show that the convergence of AdaBound extends to more general bound functions.
更多
查看译文
关键词
Non-convex optimization,Individual convergence,AdaBound,Bound functions,Deep learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要