Improved Convergence for $\ell_\infty$ and $\ell_1$ Regression via Iteratively Reweighted Least Squares.

International Conference on Machine Learning(2019)

引用 23|浏览49
暂无评分
摘要
The iteratively reweighted least squares method (IRLS) is a popular technique used in practice for solving regression problems. Various versions of this method have been proposed, but their theoretical analyses failed to capture the good practical performance. In this paper we propose a simple and natural version of IRLS for solving $ell_infty$ and $ell_1$ regression, which provably converges to a $(1+epsilon)$-approximate solution in $O(m^{1/3}log(1/epsilon)/epsilon + log(m/epsilon)/epsilon^2)$ iterations, where $m$ is the number of rows of the input matrix. Interestingly, this running time is independent of the conditioning of the input, and the dominant term of the running time depends only linearly in $epsilon^{-1}$, despite the fact that the problem it is solving is non-smooth, and the algorithm is not using any regularization. This improves upon the more complex algorithms of Chin et al. (ITCS u002712), and Christiano et al. (STOC u002711) by a factor of at least $1/epsilon^{5/3}$, and yields a truly efficient natural algorithm for the slime mold dynamics (Straszak-Vishnoi, SODA u002716, ITCS u002716, ITCS u002717).
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要