Good View Hunting: Learning Photo Composition from Dense View Pairs

2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR)(2018)

引用 86|浏览147
暂无评分
摘要
Finding views with good photo composition is a challenging task for machine learning methods. A key difficulty is the lack of well annotated large scale datasets. Most existing datasets only provide a limited number of annotations for good views, while ignoring the comparative nature of view selection. In this work, we present the first large scale Comparative Photo Composition dataset, which contains over one million comparative view pairs annotated using a cost-effective crowdsourcing workflow. We show that these comparative view annotations are essential for training a robust neural network model for composition. In addition, we propose a novel knowledge transfer framework to train a fast view proposal network, which runs at 75+ FPS and achieves state-of-the-art performance in image cropping and thumbnail generation tasks on three benchmark datasets. The superiority of our method is also demonstrated in a user study on a challenging experiment, where our method significantly outperforms the baseline methods in producing diversified well-composed views.
更多
查看译文
关键词
view selection,million comparative view pairs,cost-effective crowdsourcing workflow,comparative view annotations,robust neural network model,fast view proposal network,image cropping,thumbnail generation tasks,benchmark datasets,baseline methods,well-composed views,dense view pairs,machine learning methods,comparative photo composition dataset,view hunting,photo composition,annotated large scale datasets
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要