Sparse4D: Multi-view 3D Object Detection with Sparse Spatial-Temporal Fusion

19 Nov 2022  ·  Xuewu Lin, Tianwei Lin, Zixiang Pei, Lichao Huang, Zhizhong Su ·

Bird-eye-view (BEV) based methods have made great progress recently in multi-view 3D detection task. Comparing with BEV based methods, sparse based methods lag behind in performance, but still have lots of non-negligible merits. To push sparse 3D detection further, in this work, we introduce a novel method, named Sparse4D, which does the iterative refinement of anchor boxes via sparsely sampling and fusing spatial-temporal features. (1) Sparse 4D Sampling: for each 3D anchor, we assign multiple 4D keypoints, which are then projected to multi-view/scale/timestamp image features to sample corresponding features; (2) Hierarchy Feature Fusion: we hierarchically fuse sampled features of different view/scale, different timestamp and different keypoints to generate high-quality instance feature. In this way, Sparse4D can efficiently and effectively achieve 3D detection without relying on dense view transformation nor global attention, and is more friendly to edge devices deployment. Furthermore, we introduce an instance-level depth reweight module to alleviate the ill-posed issue in 3D-to-2D projection. In experiment, our method outperforms all sparse based methods and most BEV based methods on detection task in the nuScenes dataset.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Robust Camera Only 3D Object Detection nuScenes-C Sparse4D(r101) mean Corruption Error (mCE) 100.01 # 10
mean Resilience Rate (mRR) 55.04 # 16
Robust Camera Only 3D Object Detection nuScenes-C Sparse4D (r101) mean Corruption Error (mCE) 100.01 # 10
mean Resilience Rate (mRR) 55.04 # 16

Methods


No methods listed for this paper. Add relevant methods here