Optimal Rate of Convergence for Deep Neural Network Classifiers under the Teacher-Student Setting

CoRR(2020)

引用 0|浏览1
暂无评分
摘要
Classifiers built with neural networks handle large-scale high-dimensional data, such as facial images from computer vision, extremely well while traditional statistical methods often fail miserably. In this paper, we attempt to understand this empirical success in high dimensional classification by deriving the convergence rates of excess risk. In particular, a teacher-student framework is proposed that assumes the Bayes classifier to be expressed as ReLU neural networks. In this setup, we obtain a dimension-independent and un-improvable rate of convergence, i.e., $O(n^{-2/3})$, for classifiers trained based on either 0-1 loss or hinge loss. This rate can be further improved to $O(n^{-1})$ when data is separable. Here, $n$ represents the sample size.
更多
查看译文
关键词
deep neural network classifiers,deep neural network,neural network,optimal rate,teacher-student
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要