Home

Awesome

VisDrone2018-SOT Tooklit for Single-Object Tracking

Introduction

This is the documentation of the VisDrone2018 competitions development kit for single-object tracking (SOT) challenge.

This code library is for research purpose only, which is modified based on the visual benchmark platform of Wu et al. [1].

The code is tested on the Windows 10 and macOS Sierra 10.12.6 systems, with the Matlab 2013a/2014b/2016b/2017b platforms.

If you have any questions, please contact us (email:tju.drone.vision@gmail.com).

Citation

If you use our toolkit or dataset, please cite our paper as follows:

@article{zhuvisdrone2018,

title={Vision Meets Drones: A Challenge},

author={Zhu, Pengfei and Wen, Longyin and Bian, Xiao and Haibin, Ling and Hu, Qinghua},

journal={arXiv preprint:1804.07437},

year={2018}

}

Dataset

For SOT competition, there are three sets of data and labels: training data, validation data, and test-challenge data. There is no overlap between the three sets.

                                                     Number of snippets
----------------------------------------------------------------------------------------------
Dataset                            Training              Validation            Test-Challenge
----------------------------------------------------------------------------------------------
Signle object tracking             86 clips                  11 clips               35 clips
                                 69,941 frames              7,046 frames         29,367 frames
----------------------------------------------------------------------------------------------

For an input video sequence and the initial bounding box of the target object in the first frame, the challenge requires a participating algorithm to locate the target bounding boxes in the subsequent video frames. The objects to be tracked are of various types including pedestrians, cars, and animals. We manually annotate the bounding boxes of different objects in each video frame. Annotations on the training and validation sets are publicly available.

The link for downloading the data can be obtained by registering for the challenge at

http://www.aiskyeye.com/

Evaluation Routines

The notes for the folders:

SOT submission format

Submission of the results will consist of TXT files with one line per predicted object or MAT files as same as that in [1]. For txt submission, it looks as follows:

<bbox_left>,<bbox_top>,<bbox_width>,<bbox_height>


   Name	                                Description
--------------------------------------------------------------------------------------------------
  <bbox_left>	    The x coordinate of the top-left corner of the predicted bounding box

  <bbox_top>	    The y coordinate of the top-left corner of the predicted object bounding box
  
  <bbox_width>      The width in pixels of the predicted object bounding box 
  
  <bbox_height>     The height in pixels of the predicted object bounding box.

For mat submission, it looks as follows:

 < results: {type = 'rect', res, fps, len, annoBegin = 1, startFrame = 1} >


  Variable	                                   Description
---------------------------------------------------------------------------------------------------------
   <type>	     The representation type of the predicted bounding box representation. 
                     It should be set as 'rect'.

   <res>	     The tracking results in the video clip. Notably, each row includes the frame index, 
                     the x and y coordinates of the top-left corner of the predicted bounding box, 
	             and the width and height in pixels of the predicted bounding box.

   <fps>	     The running speed of the evaluated tracker, namely frame-per-second.

   <len>	     The length of the evaluated sequence.

<annoBegin>	     The start frame index for tracking. The default value is 1.

<startFrame>         The start frame index of the video. The default value is 1.

The sample submission of the tracker can be found in our website.

References

[1] Y. Wu, J. Lim, and M.-H. Yang, "Online Object Tracking: A Benchmark", in CVPR 2013.

[2] M. Mueller, N. Smith, B. Ghanem, "A Benchmark and Simulator for UAV Tracking", in ECCV 2016.


Version History

1.0.2 - May 7, 2018

1.0.1 - May 3, 2018

1.0.0 - Apr 19, 2018