Home

Awesome

TITAN-Next

Depth- and semantics-aware multi-modal domain translation: Generating 3D panoramic color images from LiDAR point clouds

YouTube Demo

TITAN-Next builds upon the foundation of our previous work, TITAN-Net, enhancing its capabilities to achieve state-of-the-art results in multi-modal domain translation by transforming raw 3D LiDAR point clouds into detailed RGB-D panoramic images. This unique framework leverages depth and semantic segmentation, offering groundbreaking applications for autonomous vehicles and beyond.

πŸ“„ Read the Full Paper

🌟 Key Features

πŸ”— Pretrained Models

This project utilizes SalsaNext for LiDAR segmentation. You can download the pretrained weights for SalsaNext here:

Additionally, the pretrained weights for TITAN-Next are available for download:

🧠 Methodology

TITAN-Next is a modular framework that combines semantic segmentation and depth estimation. The key components include:

πŸ“Š Results

TITAN-Next has been extensively evaluated on the Semantic-KITTI dataset, achieving significant improvements in:

Sample Results

ModelMean IoU (%)FID ↓SWD ↓
Pix2Pix12.5261.282.59
TITAN-Net31.161.912.38
TITAN-Next54.829.561.82

πŸ† Citation

If you use TITAN-Next in your research, please consider citing:

@article{Cortinhal2024,
  title={Depth- and Semantics-Aware Multi-Modal Domain Translation: Generating 3D Panoramic Color Images from LiDAR Point Clouds},
  author={Tiago Cortinhal and Eren Erdal Aksoy},
  journal={Robotics and Autonomous Systems},
  year={2024},
  volume={171},
  pages={104583},
  doi={10.1016/j.robot.2023.104583}
}