Home

Awesome

WoVoGen: World Volume-aware Diffusion for Controllable Multi-camera Driving Scene Generation

[Paper]

WoVoGen: World Volume-aware Diffusion for Controllable Multi-camera Driving Scene Generation,
Jiachen Lu, Ze Huang, Jiahui Zhang, Zeyu Yang, Li Zhang
Fudan University

Official implementation of "WoVoGen: World Volume-aware Diffusion for Controllable Multi-camera Driving Scene Generation".

<div align="center"> <img src="assets/scene_2.png"/> </div><br/>

Abstract

Generating multi-camera street-view videos is critical for augmenting autonomous driving datasets, addressing the urgent demand for extensive and varied data. Due to the limitations in diversity and challenges in handling lighting conditions, traditional rendering-based methods are increasingly being supplanted by diffusion-based methods. However, a significant challenge in diffusion-based methods is ensuring that the generated sensor data preserve both intra-world consistency and inter-sensor coherence. To address these challenges, we combine an additional explicit world volume and propose the World Volume-aware Multi-camera Driving Scene Generator (WoVoGen). This system is specifically designed to leverage 4D world volume as a foundational element for video generation. Our model operates in two distinct phases: (i) envisioning the future 4D temporal world volume based on vehicle control sequences, and (ii) generating multi-camera videos, informed by this envisioned 4D temporal world volume and sensor interconnectivity. The incorporation of the 4D world volume empowers WoVoGen not only to generate high-quality street-view videos in response to vehicle control inputs but also to facilitate scene editing tasks.

🛠️ Pipeline

<div align="center"> <img src="assets/architecture.png"/> </div><br/>

🎞️ Controlling and Editing

<div align="center"> <img src="assets/control_gen.png"/> </div><br/> <div align="center"> <img src="assets/control_edit.png"/> </div><br/> <div align="center"> <img src="assets/addtional_single-frame.png"/> </div><br/>

🚗 Scene Generation

<div align="center"> <img src="assets/scene_0.png"/> </div><br/> <div align="center"> <img src="assets/scene_1.png"/> </div><br/>

Get started

Requirements

Generate world Volumes

python tools/nuscenes_convertor.py \
    --nusc_root=$NUSC_ROOT \
    --nusc_occ_root=$NUSC_OCC_ROOT \
    --out_root=$OUT_ROOT \
    --vae  # add image latent to world volumes

Generate object guidance

python tools/clip_convertor.py

Single frame generation

# Add weights to the model
python tools/add_weight.py \
    models/v2-1_512-ema-pruned.ckpt \
    models/wovogen_single.ckpt

# Train the single frame model
python train_single_frame.py \
    --config_path models/cldm_v21_c64_256x448_6cat_clip_local_high_dim.yaml \
    --resume-path=models/wovogen_single.ckpt

📜 BibTeX

@article{lu2023wovogen,
  title={WoVoGen: World Volume-aware Diffusion for Controllable Multi-camera Driving Scene Generation},
  author={Lu, Jiachen and Huang, Ze and Zhang, Jiahui and Yang, Zeyu and Zhang, Li},
  journal={arXiv preprint arXiv:2312.02934},
  year={2023},
}

Acknowledgements

Here's how you can add a URL link to your README:

This includes a clickable link to the Stability AI GitHub repository. You can use this format to add more links or modify as needed!