Home

Awesome

<div align="center"> <h1>Local All-Pair Correspondence for Point Tracking</h1>

Seokju Cho<sup>1</sup> · Jiahui Huang<sup>2</sup> · Jisu Nam<sup>1</sup> · Honggyu An<sup>1</sup> · Seungryong Kim<sup>1</sup> · Joon-Young Lee<sup>2</sup>

<sup>1</sup>Korea University    <sup>2</sup>Adobe Research

ECCV 2024

<a href="https://arxiv.org/abs/2407.15420"><img src='https://img.shields.io/badge/arXiv-LocoTrack-red' alt='Paper PDF'></a> <a href='https://ku-cvlab.github.io/locotrack'><img src='https://img.shields.io/badge/Project_Page-LocoTrack-green' alt='Project Page'></a> <a href='https://huggingface.co/spaces/hamacojr/LocoTrack'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Spaces-blue'></a>

<p float='center'><img src="assets/teaser.png" width="80%" /></p> <span style="color: green; font-size: 1.3em; font-weight: bold;">LocoTrack is an incredibly efficient model,</span> enabling near-dense point tracking in real-time. It is <span style="color: red; font-size: 1.3em; font-weight: bold;">6x faster</span> than the previous state-of-the-art models. </div>

📰 News

Please stay tuned for an easy-to-use API for LocoTrack, coming soon!

🎮 Interactive Demo

Try our interactive demo on Huggingface. To run the demo locally, please follow these steps:

  1. Install Dependencies: Ensure you have all the necessary packages by running:

    pip install -r demo/requirements.txt
    
  2. Run the Demo: Launch the interactive Gradio demo with:

    python demo/demo.py
    

Training and Evaluation

For detailed instructions on training and evaluation, please refer to the README file for your chosen implementation:

Evaluation Dataset Preparation

First, download the evaluation datasets:

# TAP-Vid-DAVIS dataset
wget https://storage.googleapis.com/dm-tapnet/tapvid_davis.zip
unzip tapvid_davis.zip

# TAP-Vid-RGB-Stacking dataset
wget https://storage.googleapis.com/dm-tapnet/tapvid_rgb_stacking.zip
unzip tapvid_rgb_stacking.zip

# RoboTAP dataset
wget https://storage.googleapis.com/dm-tapnet/robotap/robotap.zip
unzip robotap.zip

For downloading TAP-Vid-Kinetics, please refer to official TAP-Vid repository.

Training Dataset Preparation

Download the panning-MOVi-E dataset used for training (approximately 273GB) from Huggingface using the following script. Git LFS should be installed to download the dataset. To install Git LFS, please refer to this link. Additionally, downloading instructions for the Huggingface dataset are available at this link

git clone git@hf.co:datasets/hamacojr/LocoTrack-panning-MOVi-E

📚 Citing this Work

Please use the following bibtex to cite our work:

@article{cho2024local,
  title={Local All-Pair Correspondence for Point Tracking},
  author={Cho, Seokju and Huang, Jiahui and Nam, Jisu and An, Honggyu and Kim, Seungryong and Lee, Joon-Young},
  journal={arXiv preprint arXiv:2407.15420},
  year={2024}
}

🙏 Acknowledgement

This project is largely based on the TAP repository. Thanks to the authors for their invaluable work and contributions.