Home

Awesome

<div align="center">

Planning-oriented Autonomous Driving

</div> <h3 align="center"> <a href="https://arxiv.org/abs/2212.10156">arXiv</a> | <a href="https://www.youtube.com/watch?v=cyrxJJ_nnaQ">Video</a> | <a href="sources/cvpr23_uniad_poster.png">Poster</a> | <a href="https://opendrivelab.com/e2ead/UniAD_plenary_talk_slides.pdf">Slides</a> </h3>

https://github.com/OpenDriveLab/UniAD/assets/48089846/bcf685e4-2471-450e-8b77-e028a46bd0f7

<br><br>

teaser

Table of Contents:

  1. Highlights
  2. News
  3. Getting Started
  4. Results and Models
  5. TODO List
  6. License
  7. Citation
  8. 🔥 See Also: GenAD & Vista

Highlights <a name="high"></a>

News <a name="news"></a>

Getting Started <a name="start"></a>

Results and Pre-trained Models <a name="models"></a>

UniAD is trained in two stages. Pretrained checkpoints of both stages will be released and the results of each model are listed in the following tables.

Stage1: Perception training

We first train the perception modules (i.e., track and map) to obtain a stable weight initlization for the next stage. BEV features are aggregated with 5 frames (queue_length = 5).

MethodEncoderTracking<br>AMOTAMapping<br>IoU-laneconfigDownload
UniAD-BR1010.3900.297base-stage1base-stage1

Stage2: End-to-end training

We optimize all task modules together, including track, map, motion, occupancy and planning. BEV features are aggregated with 3 frames (queue_length = 3).

<!-- Pre-trained models and results under main metrics are provided below. We refer you to the [paper](https://arxiv.org/abs/2212.10156) for more details. -->
MethodEncoderTracking<br>AMOTAMapping<br>IoU-laneMotion<br>minADEOccupancy<br>IoU-n.Planning<br>avg.Col.configDownload
UniAD-BR1010.3630.3130.70563.70.29base-stage2base-stage2

Checkpoint Usage

Model Structure

The overall pipeline of UniAD is controlled by uniad_e2e.py which coordinates all the task modules in UniAD/projects/mmdet3d_plugin/uniad/dense_heads. If you are interested in the implementation of a specific task module, please refer to its corresponding file, e.g., motion_head.

License <a name="license"></a>

All assets and code are under the Apache 2.0 license unless specified otherwise.

Citation <a name="citation"></a>

If you find our project useful for your research, please consider citing our paper and codebase with the following BibTeX:

@inproceedings{hu2023_uniad,
 title={Planning-oriented Autonomous Driving}, 
 author={Yihan Hu and Jiazhi Yang and Li Chen and Keyu Li and Chonghao Sima and Xizhou Zhu and Siqi Chai and Senyao Du and Tianwei Lin and Wenhai Wang and Lewei Lu and Xiaosong Jia and Qiang Liu and Jifeng Dai and Yu Qiao and Hongyang Li},
 booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
 year={2023},
}
@misc{contributors2023_uniadrepo,
  title={Planning-oriented Autonomous Driving},
  author={UniAD contributors},
  howpublished={\url{https://github.com/OpenDriveLab/UniAD}},
  year={2023}
}

Related Resources

Awesome

🔥 See Also <a name="see"></a>

We are thrilled to launch our recent line of works: GenAD and Vista, to advance driving world models with the largest driving video dataset collected from the web - OpenDV.

GenAD: Generalized Predictive Model for Autonomous Driving (CVPR'24, Highlight ⭐)

<div id="top" align="center"> <p align="center"> <img src="sources/opendv_dataset.png" width="1000px" > </p> </div>

Vista: A Generalizable Driving World Model with High Fidelity and Versatile Controllability 🌏

<div id="top" align="center"> <p align="center"> <img src="sources/vista.gif" width="1000px" > </p> </div>