Home

Awesome

Multi-direction and Multi-scale Pyramid in Transformer for Video-based Pedestrian Retrieval

LICENSE Python pytorch PWC PWC

Implementation of the proposed PiT. For the preprint version, please refer to [Arxiv].

framework

Getting Started

Requirements

Here is a brief instruction for installing the experimental environment.

# install virtual envs
$ conda create -n PiT python=3.6 -y
$ conda activate PiT
# install pytorch 1.8.1/1.6.0 (other versions may also work)
$ pip install timm scipy einops yacs opencv-python tensorboard pandas

Download pre-trained model

The pre-trained vit model can be downloaded in this link and should be put in the /home/[USER]/.cache/torch/checkpoints/ directory.

Dataset Preparation

For iLIDS-VID, please refer to this issue.

Training and Testing

# This command below includes the training and testing processes.
$ python train.py --config_file configs/MARS/pit.yml MODEL.DEVICE_ID "('0')" 
# For testing only
$ python train.py --config_file configs/MARS/pit-test.yml MODEL.DEVICE_ID "('0')" 

Results in the Paper

The results of MARS and iLIDS-VID are trained using one 24G NVIDIA GPU and provided below. You can change the parameter DATALOADER.P in yml file to decrease the GPU memory cost.

ModelRank-1@MARSRank-1@iLIDS-VID
PiT90.22 (code:wqxv)92.07 (code: quci)

You can download these models and put them in the ../logs/[DATASET]_PiT_1x210_3x70_105x2_6p directory. Then use the command below to evaluate them.

$ python test.py --config_file configs/MARS/pit.yml MODEL.DEVICE_ID "('0')" 

Acknowledgement

This repository is built upon the repository TranReID.

Citation

If you find this project useful for your research, please kindly cite:

@ARTICLE{9714137,
  author={Zang, Xianghao and Li, Ge and Gao, Wei},
  journal={IEEE Transactions on Industrial Informatics}, 
  title={Multidirection and Multiscale Pyramid in Transformer for Video-Based Pedestrian Retrieval}, 
  year={2022},
  volume={18},
  number={12},
  pages={8776-8785},
  doi={10.1109/TII.2022.3151766}
}

License

This repository is released under the GPL-2.0 License as found in the LICENSE file.