End-to-end Learning of Action Detection from Frame Glimpses in Videos

2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)(2015)

引用 708|浏览105
暂无评分
摘要
In this work we introduce a fully end-to-end approach for action detection in videos that learns to directly predict the temporal bounds of actions. Our intuition is that the process of detecting actions is naturally one of observation and refinement: observing moments in video, and refining hypotheses about when an action is occurring. Based on this insight, we formulate our model as a recurrent neural network-based agent that interacts with a video over time. The agent observes video frames and decides both where to look next and when to emit a prediction. Since backpropagation is not adequate in this non-differentiable setting, we use REINFORCE to learn the agent's decision policy. Our model achieves state-of-the-art results on the THUMOS'14 and ActivityNet datasets while observing only a fraction (2% or less) of the video frames.
更多
查看译文
关键词
end-to-end learning,action detection,video frame glimpses,action temporal bound prediction,recurrent neural network-based agent,backpropagation,REINFORCE,agent decision policy learning,THUMOS'14 dataset,ActivityNet dataset
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要