Normalized Cut Loss for Weakly-supervised CNN Segmentation

2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition(2018)

引用 325|浏览144
暂无评分
摘要
Most recent semantic segmentation methods train deep convolutional neural networks with fully annotated masks requiring pixel-accuracy for good quality training. Common weakly-supervised approaches generate full masks from partial input (e.g. scribbles or seeds) using standard interactive segmentation methods as preprocessing. But, errors in such masks result in poorer training since standard loss functions (e.g. cross-entropy) do not distinguish seeds from potentially mislabeled other pixels. Inspired by the general ideas in semi-supervised learning, we address these problems via a new principled loss function evaluating network output with criteria standard in "shallow" segmentation, e.g. normalized cut. Unlike prior work, the cross entropy part of our loss evaluates only seeds where labels are known while normalized cut softly evaluates consistency of all pixels. We focus on normalized cut loss where dense Gaussian kernel is efficiently implemented in linear time by fast Bilateral filtering. Our normalized cut loss approach to segmentation brings the quality of weakly-supervised training significantly closer to fully supervised methods.
更多
查看译文
关键词
weakly-supervised approaches,standard interactive segmentation methods,semantic segmentation methods,dense Gaussian kernel,bilateral filtering,fully annotated masks,deep convolutional neural networks,weakly-supervised CNN segmentation,fully supervised methods,weakly-supervised training,normalized cut loss approach,principled loss function,semisupervised learning,cross-entropy,standard loss functions
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要