WiFi and Vision Multimodal Learning for Accurate and Robust Device-Free Human Activity Recognition.

CVPR Workshops(2019)

引用 78|浏览755
暂无评分
摘要
Human activity recognition plays an indispensable role in a myriad of emerging applications in context-aware services. Accurate activity recognition systems usually require the user to carry mobile or wearable devices, which is inconvenient for long term usage. In this paper, we design WiVi, a novel human activity recognition scheme that is able to identify common human activities in an accurate and device-free manner via multimodal machine learning using only commercial WiFi-enabled IoT devices and camera. For sensing using WiFi, a new platform is developed to extract fine-grained WiFi channel information and transform them into WiFi frames. A tailored convolutional neural network model is designed to extract high-level representative features among the WiFi frames in order to provide human activity estimation. We utilized a variant of C3D model for activity sensing using vision. Following this, WiVi performs multimodal fusion at the decision level to combine the strength of WiFi and vision by constructing an ensembled DNN model. Extensive experiments are conducted in an indoor environment, demonstrating that WiVi achieves 97.5% activity recognition accuracy and is robust under unfavorable situations, as each modality provides the complementary sensing when the other faces its limiting conditions.
更多
查看译文
关键词
WiVi,human activity recognition scheme,common human activities,multimodal machine,camera,fine-grained WiFi channel information,WiFi frames,tailored convolutional neural network model,high-level representative features,human activity estimation,activity sensing,multimodal fusion,ensembled DNN model,vision multimodal learning,context-aware services,accurate activity recognition systems,mobile devices,wearable devices,long term usage,C3D model
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要