Joint Estimation Of 3d Hand Position And Gestures From Monocular Video For Mobile Interaction

CHI '15: CHI Conference on Human Factors in Computing Systems Seoul Republic of Korea April, 2015(2015)

引用 28|浏览79
暂无评分
摘要
We present a machine learning technique to recognize gestures and estimate metric depth of hands for 3D interaction, relying only on monocular RGB video input. We aim to enable spatial interaction with small, body-worn devices where rich 3D input is desired but the usage of conventional depth sensors is prohibitive due to their power consumption and size. We propose a hybrid classification-regression approach to learn and predict a mapping of RGB colors to absolute, metric depth in real time. We also classify distinct hand gestures, allowing for a variety of 3D interactions. We demonstrate our technique with three mobile interaction scenarios and evaluate the method quantitatively and qualitatively.
更多
查看译文
关键词
mobile interaction,gesture recognition,machine learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要