Awesome
ECLIPSE (CVPR 2024)
ECLIPSE: Efficient Continual Learning in Panoptic Segmentation with Visual Prompt Tuning <br /> Beomyoung Kim<sup>1,2</sup>, Joonsang Yu<sup>1</sup>, Sung Ju Hwang<sup>2</sup><br>
<sup>1</sup> <sub>NAVER Cloud, ImageVision</sub><br /> <sup>2</sup> <sub>KAIST</sub><br />
</div>Introduction
Panoptic segmentation, combining semantic and instance segmentation, stands as a cutting-edge computer vision task. Despite recent progress with deep learning models, the dynamic nature of real-world applications necessitates continual learning, where models adapt to new classes (plasticity) over time without forgetting old ones (catastrophic forgetting). Current continual segmentation methods often rely on distillation strategies like knowledge distillation and pseudo-labeling, which are effective but result in increased training complexity and computational overhead. In this paper, we introduce a novel and efficient method for continual panoptic segmentation based on Visual Prompt Tuning, dubbed ECLIPSE. Our approach involves freezing the base model parameters and fine-tuning only a small set of prompt embeddings, addressing both catastrophic forgetting and plasticity and significantly reducing the trainable parameters. To mitigate inherent challenges such as error propagation and semantic drift in continual segmentation, we propose logit manipulation to effectively leverage common knowledge across the classes. Experiments on ADE20K continual panoptic segmentation benchmark demonstrate the superiority of ECLIPSE, notably its robustness against catastrophic forgetting and its reasonable plasticity, achieving a new state-of-the-art.
Updates
2024-04-29 First Commit, We release the official implementation of ECLIPSE.
Installation
Our implementation is based on CoMFormer and Mask2Former.
Please check the installation instructions and dataset preparation.
You can see our core implementation from
mask2former/maskformer_model.py
mask2former/modeling/transformer_decoder/mask2former_transformer_decoder.py
Quick Start
- Step t=0: Training the model for base classes (you can skip this process if you use pre-trained weights.)
- Step t>1: Training the model for novel classes with ECLIPSE
Scenario | Script | Step-0 Weight | Final Weight |
---|---|---|---|
ADE20K-Panoptic 100-5 | bash script/ade_ps/100_5.sh | step0 | step10 |
ADE20K-Panoptic 100-10 | bash script/ade_ps/100_10.sh | step0 | step5 |
ADE20K-Panoptic 100-50 | bash script/ade_ps/100_50.sh | step0 | step1 |
ADE20K-Panoptic 50-10 | bash script/ade_ps/50_10.sh | step0 | step10 |
ADE20K-Panoptic 50-20 | bash script/ade_ps/50_20.sh | step0 | step5 |
ADE20K-Panoptic 50-50 | bash script/ade_ps/50_50.sh | step0 | step2 |
ADE20K-Semantic 100-5 | bash script/ade_ss/100_5.sh | step0 | step10 |
ADE20K-Semantic 100-10 | bash script/ade_ss/100_10.sh | step0 | step5 |
ADE20K-Semantic 100-50 | reproduce error | step0 | step1 |
COCO-Panoptic 83-5 | bash script/coco_ps/83_5.sh | step0 | step10 |
COCO-Panoptic 83-10 | bash script/coco_ps/83_10.sh | step0 | step5 |
How to Cite
@InProceedings{Kim_2024_CVPR,
author = {Kim, Beomyoung and Yu, Joonsang and Hwang, Sung Ju},
title = {ECLIPSE: Efficient Continual Learning in Panoptic Segmentation with Visual Prompt Tuning},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2024},
pages = {3346-3356}
}
License
ECLIPSE
Copyright (c) 2024-present NAVER Cloud Corp.
CC BY-NC 4.0 (https://creativecommons.org/licenses/by-nc/4.0/)