Home

Awesome

Learning Self-Similarity in Space and Time as Generalized Motion for Video Action Recognition


<img src="./img/SELFYBlock.jpg" width="100%" height="100%" alt="SELFY_block"></img>


This is the official implementation of the paper "Learning Self-Similarity in Space and Time as Generalized Motion for Video Action Recognition" by H.Kwon, M.Kim, S.Kwak, and M.Cho. For more information, checkout the project website and the paper on arXiv.

Environment:

Anaconda environment setting

git clone https://github.com/arunos728/SELFY.git
cd selfy
conda env create -f environment.yml
conda activate selfy

Installing Correlation sampler

cd Pytorch-Correlation-extension
python setup.py install

# check whether SpatialCorrelationSampler is installed correctly.
python check.py forward
python check.py backward
python checkCorrelationSampler.py

Please check this repo for the detailed instructions.

Dataset preparation

Please refer to TSM repo for the detailed data preparation instructions.

File lists (.txt files in ./data) specify configurations of each video clips (path, #frames, class). We upload our Something-Something-V1 & V2 video file lists in ./data. The path of the file lists should be added into the scripts for training (or testing).

Training & Testing

    ./scripts/train_SELFY_Something.sh
    ./scripts/test_SELFY_Something.sh

Citation

If you use this code or ideas from the paper for your research, please cite our paper:

@inproceedings{kwon2021learning,
  title={Learning self-similarity in space and time as generalized motion for video action recognition},
  author={Kwon, Heeseung and Kim, Manjin and Kwak, Suha and Cho, Minsu},
  booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
  pages={13065--13075},
  year={2021}
}

Contact

Heeseung Kwon(aruno@postech.ac.kr), Manjin Kim(mandos@postech.ac.kr)

Questions can also be left as issues in the repository. We will be happy to answer them.