谷歌浏览器插件
订阅小程序
在清言上使用

RepAn: Enhanced Annealing Through Re-parameterization

CVPR 2024(2024)

引用 0|浏览16
暂无评分
摘要
The simulated annealing algorithm aims to improve model convergence through multiple restarts of training. However, existing annealing algorithms overlook the cor-relation between different cycles, neglecting the potential for incremental learning. We contend that a fixed network structure prevents the model from recognizing distinct features at different training stages. To this end, we propose RepAn, redesigning the irreversible re-parameterization (Rep) method and integrating it with annealing to enhance training. Specifically, the network goes through Rep, ex-pansion, restoration, and backpropagation operations during training, and iterating through these processes in each annealing round. Such a method exhibits good generalization and is easy to apply, and we provide theoretical expla-nations for its effectiveness. Experiments demonstrate that our method improves baseline performance by 6.38% on the CIFAR-100 dataset and 2.80% on ImageNet, achieving state-of-the-art performance in the Rep field. The code is available at https://github.com/xfey/RepAn.
更多
查看译文
关键词
Theoretical Explanation,Incremental Learning,CIFAR-100 Dataset,Expansion Operation,Performance Improvement,Convolutional Neural Network,Convolutional Layers,Single Layer,Batch Normalization,Fusion Method,Input Channels,Batch Normalization Layer,Residual Connection,Bias Term,Forward Propagation,Single Branch,Memory Footprint,Training Paradigm,Parallel Branches,Lossless Compression,Early Stage Of Training,Additional Branch,Single Convolutional Layer,Output Channels,Cycles Of Expansion
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要