Awesome
<div align="center"> <h1>Local All-Pair Correspondence for Point Tracking</h1>Seokju Cho<sup>1</sup> · Jiahui Huang<sup>2</sup> · Jisu Nam<sup>1</sup> · Honggyu An<sup>1</sup> · Seungryong Kim<sup>1</sup> · Joon-Young Lee<sup>2</sup>
<sup>1</sup>Korea University <sup>2</sup>Adobe Research
ECCV 2024
<a href="https://arxiv.org/abs/2407.15420"><img src='https://img.shields.io/badge/arXiv-LocoTrack-red' alt='Paper PDF'></a> <a href='https://ku-cvlab.github.io/locotrack'><img src='https://img.shields.io/badge/Project_Page-LocoTrack-green' alt='Project Page'></a> <a href='https://huggingface.co/spaces/hamacojr/LocoTrack'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Spaces-blue'></a>
<p float='center'><img src="assets/teaser.png" width="80%" /></p> <span style="color: green; font-size: 1.3em; font-weight: bold;">LocoTrack is an incredibly efficient model,</span> enabling near-dense point tracking in real-time. It is <span style="color: red; font-size: 1.3em; font-weight: bold;">6x faster</span> than the previous state-of-the-art models. </div>📰 News
- 2024-07-22: LocoTrack is released.
- 2024-08-03: PyTorch inference and training code released.
- 2024-08-05: Interactive demo released.
Please stay tuned for an easy-to-use API for LocoTrack, coming soon!
🎮 Interactive Demo
Try our interactive demo on Huggingface. To run the demo locally, please follow these steps:
-
Install Dependencies: Ensure you have all the necessary packages by running:
pip install -r demo/requirements.txt
-
Run the Demo: Launch the interactive Gradio demo with:
python demo/demo.py
Training and Evaluation
For detailed instructions on training and evaluation, please refer to the README file for your chosen implementation:
Evaluation Dataset Preparation
First, download the evaluation datasets:
# TAP-Vid-DAVIS dataset
wget https://storage.googleapis.com/dm-tapnet/tapvid_davis.zip
unzip tapvid_davis.zip
# TAP-Vid-RGB-Stacking dataset
wget https://storage.googleapis.com/dm-tapnet/tapvid_rgb_stacking.zip
unzip tapvid_rgb_stacking.zip
# RoboTAP dataset
wget https://storage.googleapis.com/dm-tapnet/robotap/robotap.zip
unzip robotap.zip
For downloading TAP-Vid-Kinetics, please refer to official TAP-Vid repository.
Training Dataset Preparation
Download the panning-MOVi-E dataset used for training (approximately 273GB) from Huggingface using the following script. Git LFS should be installed to download the dataset. To install Git LFS, please refer to this link. Additionally, downloading instructions for the Huggingface dataset are available at this link
git clone git@hf.co:datasets/hamacojr/LocoTrack-panning-MOVi-E
📚 Citing this Work
Please use the following bibtex to cite our work:
@article{cho2024local,
title={Local All-Pair Correspondence for Point Tracking},
author={Cho, Seokju and Huang, Jiahui and Nam, Jisu and An, Honggyu and Kim, Seungryong and Lee, Joon-Young},
journal={arXiv preprint arXiv:2407.15420},
year={2024}
}
🙏 Acknowledgement
This project is largely based on the TAP repository. Thanks to the authors for their invaluable work and contributions.