Training deep neural networks with gradual deconvexification

2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN)(2016)

引用 2|浏览102
暂无评分
摘要
A new method of training deep neural networks including the convolutional network is proposed. The method deconvexifies the normalized risk-averting error (NRAE) gradually and switches to the risk-averting error (RAE) whenever RAE is computationally manageable. The method creates tunnels between the depressed regions around saddle points, tilts the plateaus, and eliminates nonglobal local minima. Numerical experiments show the effectiveness of gradual deconvexification as compared with unsupervised pretraining. After the minimization process, a statistical pruning method is used to enhance the generalization capability of the neural network under training. Numerical results show further reduction of the testing criterion.
更多
查看译文
关键词
deep neural network training,gradual deconvexification,convolutional network,normalized risk-averting error,NRAE,depressed regions,saddle points,nonglobal local minima elimination,unsupervised pretraining,minimization process,statistical pruning method,neural network generalization capability
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要