Abstract:
Event cameras are activity-driven bio-inspired vision sensors that respond asynchronously to intensity changes resulting in sparse data known as events. It has potential ...Show MoreMetadata
Abstract:
Event cameras are activity-driven bio-inspired vision sensors that respond asynchronously to intensity changes resulting in sparse data known as events. It has potential advantages over conventional cameras, such as high temporal resolution, low latency, and low power consumption. Given the sparse and asynchronous spatio-temporal nature of the data, event processing is predominantly solved by transforming events into a 2D spatial grid representation and applying standard vision pipelines. In this work, we propose an auto-encoder architecture named as Event-LSTM to generate 2D spatial grid representation. Ours has the following main advantages 1) Unsupervised, task-agnostic learning of 2D spatial grid. Ours is ideally suited for the event domain, where task-specific labeled data is scarce, 2) Asynchronous sampling of event 2D spatial grid. This leads to speed invariant and energy-efficient representation. Evaluations on appearance-based and motion-based tasks demonstrate that our approach yields improvement over state-of-the-art techniques while providing the flexibility to learn spatial grid representation from unlabelled data.
Published in: IEEE Robotics and Automation Letters ( Volume: 7, Issue: 2, April 2022)