A simple yet effective baseline for 3d human pose estimation

2017 IEEE International Conference on Computer Vision (ICCV)(2017)

引用 1545|浏览279
暂无评分
摘要
Following the success of deep convolutional networks, state-of-the-art methods for 3d human pose estimation have focused on deep end-to-end systems that predict 3d joint locations given raw image pixels. Despite their excellent performance, it is often not easy to understand whether their remaining error stems from a limited 2d pose (visual) understanding, or from a failure to map 2d poses into 3-dimensional positions. With the goal of understanding these sources of error, we set out to build a system that given 2d joint locations predicts 3d positions. Much to our surprise, we have found that, with current technology, "lifting" ground truth 2d joint locations to 3d space is a task that can be solved with a remarkably low error rate: a relatively simple deep feed-forward network outperforms the best reported result by about 30% on Human3.6M, the largest publicly available 3d pose estimation benchmark. Furthermore, training our system on the output of an off-the-shelf state-of-the-art 2d detector (, using images as input) yields state of the art results – this includes an array of systems that have been trained end-to-end specifically for this task. Our results indicate that a large portion of the error of modern deep 3d pose estimation systems stems from their visual analysis, and suggests directions to further advance the state of the art in 3d human pose estimation.
更多
查看译文
关键词
simple yet effective baseline,deep convolutional networks,deep end-to-end systems,3d joint locations,raw image pixels,3dimensional positions,relatively simple deep feedforward network,deep 3d human pose estimation systems,visual analysis,Human3.6M,ground truth 2d joint locations,2d pose understanding,publicly available 3d pose estimation benchmark
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要