Analysis of deep neural networks with the extended data Jacobian matrix
ICML'16 Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48(2016)
摘要
Deep neural networks have achieved great success on a variety of machine learning tasks. There are many fundamental and open questions yet to be answered, however. We introduce the Extended Data Jacobian Matrix (EDJM) as an architecture-independent tool to analyze neural networks at the manifold of interest. The spectrum of the EDJM is found to be highly correlated with the complexity of the learned functions. After studying the effect of dropout, ensembles, and model distillation using EDJM, we propose a novel spectral regularization method, which improves network performance.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络