Home

Awesome

nuInsSeg

provide comprehensive information for nuInsSeg dataset nuInsSeg instance segmentation annotaion demo

1.Introduction

Through investigation, we found that although existing large-scale autonomous driving datasets provide multi-modal data, they lack multi-modal and consistent annotations. So on the basis of the nuScenes dataset, we supplement the image instance segmentation annotation, and call it nuInsSeg. As far as we know, this is the first one that contains both: LiDAR point cloud, RGB image, Open source autonomous driving datasets with manual 2D annotations (2D detection boxes, 2D instance segmentation), 3D annotations (3D detection boxes...), and these 2D and 3D annotations have instance-level consistency. Added 947,072 2D annotations (box detection and instance segmentation) to the 204,894 images in the nuScenes-trainval data.

It is hoped that nuInsSeg can help researchers to better study large-scale autonomous driving scenarios, especially from the perspective of multimodal sensor fusion.

2.Statistics

The nuScenes training and validation dataset contains 1,166,187 3D detection box annotations, but because many objects are occluded or unclear in the camera's line of sight, the corresponding 2D annotations will be less. The following are the data statistics of the nuInsSeg data training set and validation set, and the target category is consistent with the nuScenes detection task.

nuInsSeg taining set class distribution

categorycarbusmotorcyclebarriertruckconst.vehpedestraintrailerbicyletraff.conetotal
instances35700612917748310302863532841314385714824724170892789193

nuInsSeg validation set class distribution

categorycarbusmotorcyclebarriertruckconst.vehpedestraintrailerbicyletraff.conetotal
instances685783084193421964141791836277782695194213889157879

3.Benchmark

Here we provide the benchmark for instance segmentation and Weakly supervised instance segmentation on nuInsSeg dataset. The category of instance segmentation is the same as nuScenes detection tasks.

We adopt a standard evaluation metric of instance segmentation which is the same as that used in Mask R-CNN. It includes AP (average precision over IoU thresholds), AP50, AP75 and APs, APm, APl (AP at different scale). In our experiment, models are implemented using the official codebase without modification.

Task1: Instance Segmentation

Instance Segmentation aims to recognize distinct instances of objects in an image by predicting pixel-level category and instance identity, which has benefited a wide range of applications such as autonomous driving. Here we provide the results of several compatitable models on nuInsSeg dataset(official code).

ModelBackboneLr_scheduleMask_APDownloadLog
Mask R-CNNR-501x47.55linklink
Mask R-CNNR-1011x49.13linklink
PointRendR-501x48.03linklink
PointRendR-1011x49.12linklink
CondInstR-501x44.88linklink
CondInstR-1011x46.83linklink

R-50 and R-101 is pretrained on imagenet. 'Lr_schedule' means the training iterations and '1x' is 7w iterations.

Task2: Weakly Supervised Instance Segmentation

Weakly supervised instance segmentation aims to extract objects with simple and cheap annotation such as image-level tags, scribble and bounding box instead of expensive pixel-level annotations.

ModelBackboneAnnotationsLr_scheduleMask_APDownloadLog
BoxInstR-50box1x33.65linklink
BoxInstR-101box1x34.39linklink
PointSupR-50box+point1x43.80linklink
PointSupR-101box+point1x44.72linklink
LWSIS+BoxInstR-503dbox+pc1x35.65linklink
LWSIS+BoxInstR-1013dbox+pc1x36,22linklink
LWSIS+PointSupR-503dbox+pc1x45.46linklink
LWSIS+PointSupR-1013dbox+pc1x46.17linklink

Here we explain different annotations used in the exp. 'box' means only using the 2D bounding box annotation for each instance, 'point' means using a specific number of points with human annotation indicating the background/foreground, '3dbox' means using the 3D bounding box annotations for each instance and 'pc' means the original point cloud.

LWSIS is the method used in our paper "LWSIS: LiDAR-guided Weakly Supervised Instance Segmentation for Autonomous Driving."

4.Tutorial

We provide development tookit to access the instance segmentation results easily: nuInsSeg-devkit.

5.Download

Here we provide the download link of the nuInsSeg dataset.

  1. nuInsSeg

  2. nuInsSeg-mini

6.Acknowledgement

It is stated here that we only add the annotation of instance segmentation on the basis of the nuscenes dataset, in order to better conduct research on multimodal perception for researchers. Many thanks to nuscnes for contributions to open source data.