Event-LSTM: An Unsupervised and Asynchronous Learning-Based Representation for Event-Based Data

Event cameras are activity-driven bio-inspired vision sensors that respond asynchronously to intensity changes resulting in sparse data known as events. It has potential advantages over conventional cameras, such as high temporal resolution, low latency, and low power consumption. Given the sparse a...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE robotics and automation letters 2022-04, Vol.7 (2), p.4678-4685
Hauptverfasser: Annamalai, Lakshmi, Ramanathan, Vignesh, Thakur, Chetan Singh
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Event cameras are activity-driven bio-inspired vision sensors that respond asynchronously to intensity changes resulting in sparse data known as events. It has potential advantages over conventional cameras, such as high temporal resolution, low latency, and low power consumption. Given the sparse and asynchronous spatio-temporal nature of the data, event processing is predominantly solved by transforming events into a 2D spatial grid representation and applying standard vision pipelines. In this work, we propose an auto-encoder architecture named as Event-LSTM to generate 2D spatial grid representation. Ours has the following main advantages 1) Unsupervised, task-agnostic learning of 2D spatial grid. Ours is ideally suited for the event domain, where task-specific labeled data is scarce, 2) Asynchronous sampling of event 2D spatial grid. This leads to speed invariant and energy-efficient representation. Evaluations on appearance-based and motion-based tasks demonstrate that our approach yields improvement over state-of-the-art techniques while providing the flexibility to learn spatial grid representation from unlabelled data.
ISSN:2377-3766
2377-3766
DOI:10.1109/LRA.2022.3151426