Unsupervised Extraction of Video Highlights via Robust Recurrent Auto-Encoders

ICCV(2015)

引用 195|浏览122
暂无评分
摘要
With the growing popularity of short-form video sharing platforms such as Instagram and Vine, there has been an increasing need for techniques that automatically extract highlights from video. Whereas prior works have approached this problem with heuristic rules or supervised learning, we present an unsupervised learning approach that takes advantage of the abundance of user-edited videos on social media websites such as YouTube. Based on the idea that the most significant sub-events within a video class are commonly present among edited videos while less interesting ones appear less frequently, we identify the significant sub-events via a robust recurrent auto-encoder trained on a collection of user-edited videos queried for each particular class of interest. The auto-encoder is trained using a proposed shrinking exponential loss function that makes it robust to noise in the web-crawled training data, and is configured with bidirectional long short term memory (LSTM) [5] cells to better model the temporal structure of highlight segments. Different from supervised techniques, our method can infer highlights using only a set of downloaded edited videos, without also needing their pre-edited counterparts which are rarely available online. Extensive experiments indicate the promise of our proposed solution in this challenging unsupervised setting.
更多
查看译文
关键词
unsupervised extraction,video highlight,robust recurrent auto-encoder,short-form video sharing platform,Instagram,Vine,heuristic rule,unsupervised learning approach,user-edited video,social media Web site,YouTube,video class,shrinking exponential loss function,Web-crawled training data,bidirectional long short term memory cell,LSTM cell,unsupervised setting
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要