Home

Awesome

Learning Motion-Appearance Co-Attention for Zero-Shot Video Object Segmentation (AMC-Net)

GitHub License GitHub last commit GitHub issues

Official implementation of 'Learning Motion-Appearance Co-Attention for Zero-Shot Video Object Segmentation', ICCV-2021

in Pytorch

image Paper

Installation

Enviroment

If you have some problems installing pydensecrf, you can visit pydensecrf for more information.

Datasets

In datasets/dataloader_list:

adaptor_dataset.py and adaptor_dataset_.py are used to assemble the DAVIS and Youtube-VOS datasets to build the model dataset, in practice we only use DAVIS for training.

davis.py and tyb_vos.py represent the data set construction rules, which can be modified to get the desired data list and stored in the cache (e.g. datasets/DAVIS/cache/tran2016.pkl and datasets/DAVIS/cache/val2016.pkl).

custom_transforms_f.py and transform.py contain some functions for data augmentation.

Avaliable flow maps from PWCNet can be find in baidu(mvuv)

Training

Testing

Metric

Please use the files in the EVALVOS folder to measure metrics.

Taking test_for_davis.py as an example:

Line 13: Setup db_info.yml

Line 14: Set the folder of groundtruth

Line 254: Set the folder of images

Line 255 & 256: Whether to discard the first frame and the last frame

Line 257: Save output in .pkl format

python test_for_davis.py

Tools

I have provided some tools, which can be seen in README