Home

Awesome

ASI-Seg: Audio-Driven Surgical Instrument Segmentation with Surgeon Intention Understanding

Overview

<div align=center> <img src="./docs/framework.png"> </div>

Installation

The code requires python>=3.8, as well as pytorch>=1.7 and torchvision>=0.8. For this project, python=3.8, pytorch=1.11.0, and torchvision=0.12.0 are used; run the following command in the root directory of this project:

cd ./ASI
pip install -r requirements.txt

Dataset

The datasets we used in our experiments are endoivs 2018 and endoivs 2017.

For EndoVis2017, we use robot-surgery-segmentation as our pre-processing strategies and cross-validation splits.

For EndoVis2018, we use ISINet as the instrument type segmentation annotation.

Checkpoints

In ASI-Seg, we used vit_h for SAM (Segmentation Anything Model), and CLIP (Contrastive Language-Image Pre-Training).

Please see the download link of the checkpoint of SAM in the vit_h version here.

Please run the following command in the root directory of this project to download CLIP:

cd ./ASI
git clone https://github.com/openai/CLIP.git

Train

Run the following command in the root directory:

cd ./ASI
python train.py

Inference

Run the following command in the root directory:

cd ./ASI
python inference.py

Citation

If you use our code or paper in your work, please cite our paper.

@inproceedings{chen2024iros,
	title={{ASI-Seg: Audio-Driven Surgical Instrument Segmentation with Surgeon Intention Understanding}},
	author={Zhen Chen, Zongming Zhang, Wenwu Guo, Xingjian Luo, Long Bai, Jinlin Wu,
Hongliang Ren, Hongbin Liu},
	booktitle={IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)},
	year={2024}
}