EVT: Efficient View Transformation for Multi-Modal 3D Object Detection
Multi-modal sensor fusion in bird's-eye-view (BEV) representation has become the leading approach in 3D object detection. However, existing methods often rely on depth estimators or transformer encoders for view transformation, incurring substantial computational overhead. Furthermore, the lack...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Multi-modal sensor fusion in bird's-eye-view (BEV) representation has become
the leading approach in 3D object detection. However, existing methods often
rely on depth estimators or transformer encoders for view transformation,
incurring substantial computational overhead. Furthermore, the lack of precise
geometric correspondence between 2D and 3D spaces leads to spatial and
ray-directional misalignments, restricting the effectiveness of BEV
representations. To address these challenges, we propose a novel 3D object
detector via efficient view transformation (EVT), which leverages a
well-structured BEV representation to enhance accuracy and efficiency. EVT
focuses on two main areas. First, it employs Adaptive Sampling and Adaptive
Projection (ASAP), using LiDAR guidance to generate 3D sampling points and
adaptive kernels. The generated points and kernels are then used to facilitate
the transformation of image features into BEV space and refine the BEV
features. Second, EVT includes an improved transformer-based detection
framework, which contains a group-wise query initialization method and an
enhanced query update framework. It is designed to effectively utilize the
obtained multi-modal BEV features within the transformer decoder. By leveraging
the geometric properties of object queries, this framework significantly
enhances detection performance, especially in a multi-layer transformer decoder
structure. EVT achieves state-of-the-art performance on the nuScenes test set
with real-time inference speed. |
---|---|
DOI: | 10.48550/arxiv.2411.10715 |