Home

Awesome

MT-DETR

This project is the code of WACV 2023 paper: MT-DETR: Robust End-to-end Multimodal Detection with Confidence Fusion by Shih-Yun Chu, Ming-Sui Lee. You can find more visualized result and details in supplementary material.

<!-- ### Visualized Results in different weathers and times --> <div align=center> <img src='figure/snow_day.png' width="47%"> <img src='figure/densefog_day.png' width="47%"> <img src='figure/lightfog_night.png' width="47%"> <img src='figure/clear_night.png' width="47%"> </div>

Brief Introduction

In the application of autonomous driving, there are times when unexpected and severe weather (fog, snow, night) occurs in outdoor environments, making the detection tasks less effective. Therefore, this paper proposes a novel multimodal object detection network called MT-DETR. It achieves state-of-the-art performance using the camera, lidar and radar, and additional time information. The experimental results demonstrate that the MT-DETR is robust and performs well in various weather conditions. The good generalization and scalability confirm future applicability to different multimodal tasks.

<div align=center> <img src='figure/architecture.png' width="80%"> </div> <!-- ### Comparison Tables <div align=center> <img src='figure/comparison.png' width="47%"> <img src='figure/input_sensor.png' width="47%"> </div> -->

Getting Started

The repository is based on mmdetection and cbnetv2. Many thanks for their awesome open-source project.

To run the code:

  1. To construct an environment first, please follow the cbnetv2 (https://github.com/VDIGPKU/CBNetV2) and mmdetection (https://github.com/open-mmlab/mmdetection) tutorial.
  2. Download the dataset and model checkpoints. Please go to data/ and checkpoint/ and read the instructions there to download.
  3. After preparation, type the following command in your terminal:
bash run_script/${script_name}

You can comment training/inference block in shell scripts if you want.

The following are the important directories of this project:

BibTeX

If you find our work useful in your research, please consider citing our paper.

@InProceedings{Chu_2023_WACV,
    author    = {Chu, Shih-Yun and Lee, Ming-Sui},
    title     = {MT-DETR: Robust End-to-End Multimodal Detection With Confidence Fusion},
    booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)},
    month     = {January},
    year      = {2023},
    pages     = {5252-5261}
}