Stct: Sequentially Training Convolutional Networks For Visual Tracking

2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)(2016)

引用 325|浏览62
暂无评分
摘要
Due to the limited amount of training samples, finetuning pre-trained deep models online is prone to overfitting. In this paper, we propose a sequential training method for convolutional neural networks (CNNs) to effectively transfer pre-trained deep features for online applications. We regard a CNN as an ensemble with each channel of the output feature map as an individual base learner. Each base learner is trained using different loss criterions to reduce correlation and avoid over-training. To achieve the best ensemble online, all the base learners are sequentially sampled into the ensemble via important sampling. To further improve the robustness of each base learner, we propose to train the convolutional layers with random binary masks, which serves as a regularization to enforce each base learner to focus on different input features.The proposed online training method is applied to visual tracking problem by transferring deep features trained on massive annotated visual data and is shown to significantly improve tracking performance. Extensive experiments are conducted on two challenging benchmark data set and demonstrate that our tracking algorithm can outperform state-of-the-art methods with a considerable margin.
更多
查看译文
关键词
STCT,pre-trained deep models,sequential training method,convolutional neural networks,CNN,online applications,output feature map,ensemble online,base learners,important sampling,convolutional layers,random binary masks,online training method,visual tracking problem,massive annotated visual data
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要