Home

Awesome

Multi-Camera Unified Pre-training via 3D Scene Reconstruction

(for DETR3D, BEVFormer, BEVDet, BEVDepth and Semantic Occupancy Prediction)

Paper in arXiv

Abstract

Multi-camera 3D perception has emerged as a prominent research field in autonomous driving, offering a viable and cost-effective alternative to LiDAR-based solutions. The existing multi-camera algorithms primarily rely on monocular 2D pre-training. However, the monocular 2D pre-training overlooks the spatial and temporal correlations among the multi-camera system. To address this limitation, we propose the first multi-camera unified pre-training framework, called UniScene, which involves initially reconstructing the 3D scene as the foundational stage and subsequently fine-tuning the model on downstream tasks. Specifically, we employ Occupancy as the general representation for the 3D scene, enabling the model to grasp geometric priors of the surrounding world through pre-training.

Methods

method

Getting Started

Model Zoo

BackboneMethodPre-trainingLr SchdNDSmAPConfig
R101-DCNBEVFormerImageNet24ep47.737.7config/[model]
R101-DCNBEVFormerImageNet + UniScene24ep50.039.7config/model
R101-DCNBEVFormerFCOS3D24ep51.741.6config/model
R101-DCNBEVFormerFCOS3D + UniScene24ep53.443.8config/pre-trained model/fine-tune-model

Bibtex

If this work is helpful for your research, please consider citing the following BibTeX entry.

@article{UniScene,
  title={Multi-Camera Unified Pre-training via 3D Scene Reconstruction},
  author={Chen Min, Liang Xiao, Dawei Zhao, Yiming Nie, and Bin Dai}
  journal={arXiv preprint},
  year={2023}
}

Acknowledgement

Many thanks to these excellent open source projects: