Home

Awesome

3D Mapping

Welcome! This repository contains all software used to create CarlaSC (our dynamic scene completion data set) and MotionSC (our local semantic mapping network), and all baselines. Below is an overview of our project. For more information on the data and method, see our website. For more information on the software and files involved, see the Wiki.

Table of Contents

Data: CarlaSC

A novel data set with accurate, complete dynamic outdoor scenes for semantic scene completion or mapping. We use the CARLA simulator (1) to gather clean ground truth information, where we randomly place a large number of sensors in each scene, guaranteeing theoretical convergence to the true scene and network generalization to the full scene. Dynamic maps are difficult due to traces left behind by moving objects, and occlusions. Below is a comparison of a frame from our data set with a similar frame from Semantic KITTI (4).

More introduction about CarlaSC dataset and download links are avialable on our website.

<p align="center"> <img width="740" src="https://user-images.githubusercontent.com/91337470/175836944-f4c91add-95bc-490e-aaee-248005bf1351.png"> </p>

Networks: MotionSC

We create a network using ideas from MotionNet (2) and LMSCNet (3), which we call MotionSC. MotionSC extends semantic scene completion to achieve a higher level of scene understanding by integrating the structure of a view-volume scene completion network with the temporal backbone of an object tracking network. It is built off the idea that semantic scene completion is fundamentally a similar task to 3D semantic mapping, where the major difference is the use of temporal information, readily available in robotic applications. We compare with scene completion baselines LMSCNet (3), JS3CNet (5), and SSCNet (6). We show that the number of past scans (T) is correlated with improvements in semantic and geometric completeness quantitatively and qualitatively. Note that in the video below, the ego vehicle is the stationary Tesla in the bottom right, and is excluded from the complete semantic scenes.

<p align="center"> <img width="740" src="https://user-images.githubusercontent.com/91337470/175836958-8e08f5d7-2017-476e-a826-803e983cffa9.gif"> </p>

Use 3DMapping

Dependencies

We also provide an environment.yml which you can use to create a conda environment

conda env create -f environment.yml
conda activate MotionSC

Data Generation and Visualization

We provide script to generate the dataset. We also provide scripts for visualizing the data. We have detailed explanation on the useful scripts and parameters in our wiki. You can check the TODO block in scripts for parameters that could be adjusted.

Training and Testing

We provide synthetic scripts of training and testing MotionSC, LMSCNet and SSCNet in our repo. You can check the TODO block in scripts for parameters that could be adjusted. You can find how to use them on our wiki. Due to special dependencies of JS3CNet, you can check our forked JS3CNet to run it.

Results

We trained our model on the CarlaSC dataset and the SemanticKitti dataset. Note that while improved semantic scene completion networks are being released with better results, our method of incorporating temporal information may be applied to these networks for enhanced mapping capabilities.

CarlaSC

For the MotionSC model on CarlaSC we also list its performance dependent on the amount of temporal information provided, T.

Inference Time (Measured on NVIDIA GeForce RTX 3090)

MethodLatency (ms)
LMSCNet SS4.86
SSCNet Full2.18
JS3C-Net166.2
MotionSC (T=1)5.72

Semantic Completeness

MethodMean IoUAccuracyFreeBuildingBarrierOtherPedestrianPoleRoadGroundSidewalkVegetationVehicles
LMSCNet SS42.5394.6497.4125.613.3511.3133.7643.5485.9621.1552.6439.9953.09
SSCNet Full41.9194.1196.0227.041.8213.6529.6927.0288.4525.8965.3633.2952.78
JS3C-Net48.9595.4896.7834.683.0322.9443.6444.5093.3130.9075.1534.3559.21
MotionSC (T=1)46.3195.1197.4231.592.6314.7739.8742.1190.5725.8960.7742.4161.37
MotionSC (T=5)45.3595.0097.4329.482.5417.4841.8743.4390.9022.0858.4335.7959.41
MotionSC (T=10)47.0195.1597.4432.292.3519.8244.0645.4790.1927.3562.4836.9258.80
MotionSC (T=16)47.4595.5797.6034.912.6622.8637.7843.8790.1228.3166.2041.5956.08

Geometric Completeness

MethodPrecisionRecallIoU
LMSCNet SS95.6298.9585.98
SSCNet Full85.8793.0580.69
JS3C-Net89.4393.0283.80
MotionSC (T=1)93.3292.1686.46
MotionSC (T=5)94.7690.5786.25
MotionSC (T=10)93.1792.4386.56
MotionSC (T=16)94.6191.7787.21

SemanticKITTI

For the SemanticKITTI dataset, we only compare with T=1 as the rules of the competion require a single input frame. Our results may be found on the Semantic KITTI leader board under user "tigeriv4." The results of other models are collected from their papers.

Semantic Completeness

MethodMean IoUAccuracyRoadSidewalkParkingOther-groundBuildingCarTruckBicycleMotorcycleOther-vehiclevegetationTrunkTerrainPersonBicyclistMotorcyclistFencePoleTraffic-sign
LMSCNet SS17.6256.7264.8034.6829.024.6238.0830.891.47000.8141.3119.8932.0500021.3215.010.84
SSCNet Full16.1449.9851.1530.7627.126.4434.5324.261.180.540.784.3435.2518.1729.010.250.250.0319.8713.106.73
JS3C-Net23.856.664.739.934.914.137.433.37.214.48.812.743.119.640.58.05.10.430.418.915.9
MotionSC (T=1)18.456.966.036.529.67.039.031.41.0003.640.019.030.000023.420.03.4

Acknowledgement

We utilize data and code from:

Reference

If you find our work useful in your research work, consider citing our paper!

@misc{https://doi.org/10.48550/arxiv.2203.07060,
  doi = {10.48550/ARXIV.2203.07060},
  
  url = {https://arxiv.org/abs/2203.07060},
  
  author = {Wilson, Joey and Song, Jingyu and Fu, Yuewei and Zhang, Arthur and Capodieci, Andrew and Jayakumar, Paramsothy and Barton, Kira and Ghaffari, Maani},
  
  keywords = {Computer Vision and Pattern Recognition (cs.CV), Robotics (cs.RO), FOS: Computer and information sciences, FOS: Computer and information sciences},
  
  title = {MotionSC: Data Set and Network for Real-Time Semantic Mapping in Dynamic Environments},
  
  publisher = {arXiv},
  
  year = {2022},
  
  copyright = {Creative Commons Attribution 4.0 International}
}