Event-LSTM: An Unsupervised and Asynchronous Learning-Based Representation for Event-Based Data

IEEE ROBOTICS AND AUTOMATION LETTERS(2022)

引用 4|浏览7
暂无评分
摘要
Event cameras are activity-driven bio-inspired vision sensors that respond asynchronously to intensity changes resulting in sparse data known as events. It has potential advantages over conventional cameras, such as high temporal resolution, low latency, and low power consumption. Given the sparse and asynchronous spatio-temporal nature of the data, event processing is predominantly solved by transforming events into a 2D spatial grid representation and applying standard vision pipelines. In this work, we propose an auto-encoder architecture named as Event-LSTM to generate 2D spatial grid representation. Ours has the following main advantages 1) Unsupervised, task-agnostic learning of 2D spatial grid. Ours is ideally suited for the event domain, where task-specific labeled data is scarce, 2) Asynchronous sampling of event 2D spatial grid. This leads to speed invariant and energy-efficient representation. Evaluations on appearance-based and motion-based tasks demonstrate that our approach yields improvement over state-of-the-art techniques while providing the flexibility to learn spatial grid representation from unlabelled data.
更多
查看译文
关键词
Deep learning methods, deep learning for visual perception, representation learning, event camera, LSTM
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要