RayNet: Learning Volumetric 3D Reconstruction with Ray Potentials

2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR)(2019)

引用 96|浏览63
暂无评分
摘要
In this paper, we consider the problem of reconstructing a dense 3D model using images captured from different views. Recent methods based on convolutional neural networks (CNN) allow learning the entire task from data. However, they do not incorporate the physics of image formation such as perspective geometry and occlusion. Instead, classical approaches based on Markov Random Fields (MRF) with ray-potentials explicitly model these physical processes, but they cannot cope with large surface appearance variations across different viewpoints. In this paper, we propose RayNet, which combines the strengths of both frameworks. RayNet integrates a CNN that learns view-invariant feature representations with an MRF that explicitly encodes the physics of perspective projection and occlusion. We train RayNet end-to-end using empirical risk minimization. We thoroughly evaluate our approach on challenging real-world datasets and demonstrate its benefits over a piece-wise trained baseline, hand-crafted models as well as other learning-based approaches.
更多
查看译文
关键词
Markov random fields,convolutional neural networks,dense 3D model,ray potentials,volumetric 3D reconstruction,learning-based approaches,hand-crafted models,RayNet end-to-end,perspective projection,view-invariant feature representations,surface appearance variations,physical processes,ray-potentials,classical approaches,perspective geometry,image formation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要