Awesome
<div align=center> <img width="100%" src="./assets/logo.png"> </div> <div align="center"> <a href="https://pytorch.org"><img src="https://img.shields.io/badge/Framework-Pytorch-orange.svg" alt="Framework"></a> <a href="https://github.com/jiawen-zhu/TrackGPT/blob/main/LICENSE"><img src="assets/LICENSE-Apache License-blue.svg" alt="License"></a> <a href="https://arxiv.org/abs/2312.17448"><img src="https://img.shields.io/badge/%F0%9F%93%96-Arxiv_2312.17448-red.svg?style=flat-squre"></a> </div>InsTrack dataset is avaliable at google drive
<div align=center> <img width="100%" src="./assets/introduction.png"/> </div>TrackGPT is a new tracking architecture that is capable of performing complex reasoning-based tracking by injecting (Large Vision-Language Model) LVLM's comprehension of the multi-modal world. A new tracking task, Instruction Tracking (InsT), is proposed simultaneously in a context where perception and comprehension tasks are no longer clearly demarcated. We undertake this modest attempt to advance next-generation object tracking with more intelligence.
<div align=center> <td><center><img width="90%" alt="" src=./assets/video_demo1.gif/></center></td> <td><center><img width="90%" alt="" src=./assets/video_demo2.gif/></center></td> </div>:loudspeaker:News
- [2023/12/31] We make TrackGPT public.
:fire: Highlight
-
A new task called instruction tracking (InsT) is proposed, where a tracker must have the self-reasoning capability, autonomously interpret implicit instruction and track the target object. This human-tracker interaction paradigm aligns better with the way humans ask questions.
-
A benchmark, InsTrack, is also constructed for instruction tuning and evaluation.
-
This work present TrackGPT, a tracker that can comprehend human intent by leveraging the reasoning capability of LVLM. TrackGPT is designed sticking to a principle of simple yet effective, we hope this work could catalyze more compelling research in the future.
:memo: Results
- Referring Tracking
- Instruction Tracking
:bookmark_tabs:Installation
- Install the conda environment
conda create -n trackgpt python=3.9
conda activate trackgpt
- Install the required packages:
pip install -r requirements.txt
pip install flash-attn --no-build-isolation
:car:Run TrackGPT
sh TrackGPT_demo.sh
For example,
Please input your tracking instrcution: I'd like to focus on the protagonist of this street event. Please track the object.
Please input the video path: test_videos/breakdance
:hearts: Acknowledgment
This project is based on LISA and LLaVA. Thanks for these excellent works.
:book: Citation
If you find TrackGPT useful for you, please consider citing :mega:
@misc{trackgpt,
Title={Tracking with Human-Intent Reasoning},
Author = {Jiawen Zhu and Zhi-Qi Cheng and Jun-Yan He and Chenyang Li and Bin Luo and Huchuan Lu and Yifeng Geng and Xuansong Xie},
Year = {2023},
Eprint = {arXiv:2312.17448},
PrimaryClass={cs.CV}
}