Home

Awesome

CoCa3D

License: MIT

This repository contains the official PyTorch implementation of

Collaboration Helps Camera Overtake LiDAR in 3D Detection</a> <br> <a href="https://scholar.google.com/citations?user=XBbwb78AAAAJ&hl=zh-CN"> Yue Hu, <a href="https://github.com/yifanlu0227"> Yifan Lu, <a href="https://derrickxunu.github.io//">Runsheng Xu, <a href="https://weidixie.github.io/"> Weidi Xie, <a href="https://siheng-chen.github.io/">Siheng Chen, <a href="https://mediabrain.sjtu.edu.cn/members/">Yanfeng wang</a> <br> Presented at CVPR 2023

Where2comm

<div align='center' ><font size='2'>LiDAR detection v.s. collaborative camera perception</font></div>

Main idea

Abstract: Camera-only 3D detection provides an economical solution with a simple configuration for localizing objects in 3D space compared to LiDAR-based detection systems. However, a major challenge lies in precise depth estimation due to the lack of direct 3D measurements in the input. Many previous methods attempt to improve depth estimation through network designs, e.g., deformable layers and larger receptive fields. This work proposes an orthogonal direction, improving the camera-only 3D detection by introducing multi-agent collaborations. Our preliminary results show a potential that with sufficient collaboration, the camera might overtake LiDAR in some practical scenarios.

Where2comm

Features

Quick Start

Install

Please refer to the INSTALL.md for detailed documentations.

Download dataset DAIR-V2X

  1. Download raw data of DAIR-V2X.
  2. Download complemented annotation from dair-v2x-c-complemented.

Train your model

We adopt the same setting as OpenCOOD which uses yaml file to configure all the parameters for training. To train your own model from scratch or a continued checkpoint, run the following commonds:

python opencood/tools/train.py --hypes_yaml ${CONFIG_FILE} [--model_dir  ${CHECKPOINT_FOLDER}]

Arguments Explanation:

Test the model

Before you run the following command, first make sure the validation_dir in config.yaml under your checkpoint folder refers to the testing dataset path, e.g. opv2v_data_dumping/test.

python opencood/tools/inference.py --model_dir ${CHECKPOINT_FOLDER} --fusion_method ${FUSION_STRATEGY} --save_vis_n ${amount}

Arguments Explanation:

The evaluation results will be dumped in the model directory.

Acknowledgements

Thank for the excellent cooperative perception codebases OpenCOOD and CoPerception.

Thank for the excellent cooperative perception datasets DAIR-V2X, OPV2V and V2X-SIM.

Thanks for the insightful previous works in cooperative perception field. Where2comm(NeruIPS22), CoAlign(ICRA23), V2VNet(ECCV20), When2com(CVPR20), Who2com(ICRA20), DiscoNet(NeurIPS21), V2X-ViT(ECCV2022), STAR(CoRL2022), CoBEVT(CoRL2022).

Contact

If you have any problem with this code, please feel free to contact 18671129361@sjtu.edu.cn.

Citation

If you find this code useful in your research then please cite

@inproceedings{CoCa3D:23,
  author    = {Yue Hu, Yifan Lu, Runsheng Xu, Weidi Xie, Siheng Chen, Yanfeng Wang},
  title     = {Collaboration Helps Camera Overtake LiDAR in 3D Detection},
  booktitle = {The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  year      = {2023}
}