谷歌浏览器插件
订阅小程序
在清言上使用

Transition to Linearity of Wide Neural Networks is an Emerging Property of Assembling Weak Models

International Conference on Learning Representations (ICLR)(2022)

引用 3|浏览29
暂无评分
摘要
Wide neural networks with linear output layer have been shown to be near-linear, and to have near-constant neural tangent kernel (NTK), in a region containing the optimization path of gradient descent. These findings seem counter-intuitive since in general neural networks are highly complex models. Why does a linear structure emerge when the networks become wide? In this work, we provide a new perspective on this "transition to linearity" by considering a neural network as an assembly model recursively built from a set of sub-models corresponding to individual neurons. In this view, we show that the linearity of wide neural networks is, in fact, an emerging property of assembling a large number of diverse "weak" sub-models, none of which dominate the assembly.
更多
查看译文
关键词
Assembling,linearity,Transition to linearity,wide neural networks
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要