Temporally Distributed Networks for Fast Video Semantic Segmentation

CVPR(2020)

引用 184|浏览280
暂无评分
摘要
We present TDNet, a temporally distributed network designed for fast and accurate video semantic segmentation. We observe that features extracted from a certain high-level layer of a deep CNN can be approximated by composing features extracted from several shallower sub-networks. Leveraging the inherent temporal continuity in videos, we distribute these sub-networks over sequential frames. Therefore, at each time step, we only need to perform a lightweight computation to extract a sub-features group from a single sub-network. The full features used for segmentation are then recomposed by application of a novel attention propagation module that compensates for geometry deformation between frames. A grouped knowledge distillation loss is also introduced to further improve the representation power at both full and sub-feature levels. Experiments on Cityscapes, CamVid, and NYUD-v2 demonstrate that our method achieves state-of-the-art accuracy with significantly faster speed and lower latency.
更多
查看译文
关键词
TDNet,temporally distributed network,video semantic segmentation,deep CNN,sequential frames,attention propagation module,grouped knowledge distillation loss,subfeature levels,subfeatures group,feature extraction,geometry deformation,Cityscapes,CamVid,NYUD-v2
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要