MIPD: A Multi-sensory Interactive Perception Dataset for Embodied Intelligent Driving
During the process of driving, humans usually rely on multiple senses to gather information and make decisions. Analogously, in order to achieve embodied intelligence in autonomous driving, it is essential to integrate multidimensional sensory information in order to facilitate interaction with the...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | During the process of driving, humans usually rely on multiple senses to
gather information and make decisions. Analogously, in order to achieve
embodied intelligence in autonomous driving, it is essential to integrate
multidimensional sensory information in order to facilitate interaction with
the environment. However, the current multi-modal fusion sensing schemes often
neglect these additional sensory inputs, hindering the realization of fully
autonomous driving. This paper considers multi-sensory information and proposes
a multi-modal interactive perception dataset named MIPD, enabling expanding the
current autonomous driving algorithm framework, for supporting the research on
embodied intelligent driving. In addition to the conventional camera, lidar,
and 4D radar data, our dataset incorporates multiple sensor inputs including
sound, light intensity, vibration intensity and vehicle speed to enrich the
dataset comprehensiveness. Comprising 126 consecutive sequences, many exceeding
twenty seconds, MIPD features over 8,500 meticulously synchronized and
annotated frames. Moreover, it encompasses many challenging scenarios, covering
various road and lighting conditions. The dataset has undergone thorough
experimental validation, producing valuable insights for the exploration of
next-generation autonomous driving frameworks. |
---|---|
DOI: | 10.48550/arxiv.2411.05881 |