Home

Awesome

LV-VIS: Large-Vocabulary Video Instance Segmentation dataset

📄[arXiv]   📄[ICCV(Oral)]   🔥[Dataset Download]   🔥[Evaluation Server]

This repo is the official implementation of Towards Open Vocabulary Video Instance Segmentation (ICCV2023 oral)

News

We are working on the final revision of the annotations. The Codalab set will be released in November.

Towards Open Vocabulary Video Instance Segmentation

Haochen Wang<sup>1</sup>, Cilin Yan<sup>2</sup>, Shuai Wang <sup>1</sup>, Xiaolong Jiang <sup>3</sup>, Xu Tang<sup>3</sup>, Yao Hu<sup>3</sup>, Weidi Xie <sup>4</sup>,Efstratios Gavves <sup>1</sup>

<sup>1</sup>University of Amsterdam, <sup>2</sup>Beihang University, <sup>3</sup>Xiaohongshu Inc, <sup>4</sup> Shanghai Jiao Tong University.

LV-VIS dataset

LV-VIS is a dataset/benchmark for Open-Vocabulary Video Instance Segmentation. It contains a total of 4,828 videos with pixel-level segmentation masks for 26,099 objects from 1,196 unique categories.

LV-VIS is licensed under a CC BY-NC-SA 4.0 License. The data of LV-VIS is released for non-commercial research purpose only.

<img src="assert/dataset/visualizations/00000.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00001.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00005.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00012.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00013.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00018.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00028.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00035.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00058.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00066.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00078.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00087.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00119.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00129.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00199.gif" alt="Demo" width="240" height="140"> <img src="assert/dataset/visualizations/00203.gif" alt="Demo" width="240" height="140">

Dataset Download

<!-- - [Training Videos](https://drive.google.com/file/d/1er2lBQLF75TI5O4wzGyur0YYoohMK6C3/view?usp=sharing) - [Validation Videos](https://drive.google.com/file/d/1vTYUz_XLOBnYb9e7upJsZM-nQz2S6wDn/view?usp=drive_link) - [Test Videos](https://drive.google.com/file/d/13Hgz2hxOPbe4_yTiUpwWb2ZWphaP06AF/view?usp=drive_link) - [Training Annotations](https://drive.google.com/file/d/18ifd40HuXbjKBtwpUzmboucmOSuAzD1n/view?usp=sharing) - [Validation Annotations](https://drive.google.com/file/d/1hvZHShzVNmxIQrGGB1chZTV2nqGShi6X/view?usp=drive_link) - [Validation Annotations (oracle)](https://drive.google.com/file/d/1jmD4aQoP98nOdo_mjtV1eiCAsFGORN0R/view?usp=sharing) # iccv23 | | Videos | Annotations | Annotations (oracle) | | :---: | :---: | :---: | :---: | | Training | [Download](https://drive.google.com/file/d/1er2lBQLF75TI5O4wzGyur0YYoohMK6C3/view?usp=sharing) | [Download](https://drive.google.com/file/d/18ifd40HuXbjKBtwpUzmboucmOSuAzD1n/view?usp=sharing) | - | | Validation | [Download](https://drive.google.com/file/d/1vTYUz_XLOBnYb9e7upJsZM-nQz2S6wDn/view?usp=drive_link) | [Download](https://drive.google.com/file/d/1hvZHShzVNmxIQrGGB1chZTV2nqGShi6X/view?usp=drive_link) | [Download](https://drive.google.com/file/d/1jmD4aQoP98nOdo_mjtV1eiCAsFGORN0R/view?usp=sharing) | | Test | [Download](https://drive.google.com/file/d/13Hgz2hxOPbe4_yTiUpwWb2ZWphaP06AF/view?usp=drive_link) | - | - | -->
VideosAnnotationsAnnotations (oracle)Submission Example
TrainingDownloadDownload--
ValidationDownloadDownloadDownload-
TestDownload--Download

Dataset Structure

## JPEGImages

|- train
  |- 00000
    |- 00000.jpg
    |- 00001.jpg
       ...
  |- 00001
    |- 00000.jpg
    |- 00001.jpg
       ...
    ...
|- val
    ...
|- test
    ...

## Annotations
train_instances.json
val_instances.json
image_val_instances.json

The annotation files have the same formation as Youtube-VIS 2019.

Annotation Tool

We used this platform for the annotation of LV-VIS. This platform is a smart video segmentation annotation tool based on Lableme, SAM, and STCN. See segment-anything-annotator.

Baseline

We provide our baseline OV2Seg code for LV-VIS. Please check Baseline.md for more details.

TODO

NOTE:

Cite

@inproceedings{wang2023towards,
  title={Towards Open-Vocabulary Video Instance Segmentation},
  author={Wang, Haochen and Yan, Cilin and Wang, Shuai and Jiang, Xiaolong and Tang, XU and Hu, Yao and Xie, Weidi and Gavves, Efstratios},
  booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
  year={2023}
}

Acknowledgement

This repo is built based on Mask2Former and Detic, thanks for those excellent projects.