Home

Awesome

JTA Dataset

License: CC BY-NC 4.0

JTA (Joint Track Auto) is a huge dataset for pedestrian pose estimation and tracking in urban scenarios created by exploiting the highly photorealistic video game Grand Theft Auto V. We collected a set of 512 full-HD videos (256 for training and 256 for testing), 30 seconds long, recorded at 30 fps. The dataset was created with this tool.

banner

Obtain the Dataset

You can download JTA here. By downloading the dataset you agree on the following statement: "I declare that I will use the JTA Dataset for research and educational purposes only, since I am aware that commercial use is prohibited. I also undertake to purchase a copy of Grand Theft Auto V."

JTA-Dataset Contents

After the data download, your JTA-Dataset directory will contain the following files:

Annotations

Each annotation file refers to a specific sequence (e.g. seq_42.json is the annotation file of seq_42.mp4). An annotation file consists of a list of lists, that is a matrix having N rows and 10 columns. Each row of the matrix contains the data of a joint; these data are organized as follows:

ElementNameDescription
row[0]frame numbernumber of the frame to which the joint belongs
row[1]person IDunique identifier of the person to which the joint belongs
row[2]joint typeidentifier of the type of joint; see 'Joint Types' subsection
row[3]x2D2D x coordinate of the joint in pixel
row[4]y2D2D y coordinate of the joint in pixel
row[5]x3D3D x coordinate of the joint in meters
row[6]y3D3D y coordinate of the joint in meters
row[7]z3D3D z coordinate of the joint in meters
row[8]occluded1 if the joint is occluded; 0 otherwise
row[9]self-occluded1 if the joint is occluded by its owner; 0 otherwise

Camera

Each sequence has been recorded with the same camera with the followng intrinsic matrix:

<a href="https://www.codecogs.com/eqnedit.php?latex=K&space;=&space;\begin{pmatrix}&space;1158&space;&&space;0&space;&&space;960\\&space;0&space;&&space;1158&space;&&space;540\\&space;0&space;&&space;0&space;&&space;1&space;\end{pmatrix}" target="_blank"><img src="https://latex.codecogs.com/gif.latex?K&space;=&space;\begin{pmatrix}&space;1158&space;&&space;0&space;&&space;960\\&space;0&space;&&space;1158&space;&&space;540\\&space;0&space;&&space;0&space;&&space;1&space;\end{pmatrix}" title="K = \begin{pmatrix} 1158 & 0 & 960\\ 0 & 1158 & 540\\ 0 & 0 & 1 \end{pmatrix}" /></a>

Joint Types

The associations between numerical identifier and type of joint are the following:

 0: head_top
 1: head_center
 2: neck
 3: right_clavicle
 4: right_shoulder
 5: right_elbow
 6: right_wrist
 7: left_clavicle
 8: left_shoulder
 9: left_elbow
10: left_wrist
11: spine0
12: spine1
13: spine2
14: spine3
15: spine4
16: right_hip
17: right_knee
18: right_ankle
19: left_hip
20: left_knee
21: left_ankle

Annotation Management - Example

COCO-Style Annotations

If you want, you can convert our annotations to COCO format using the coco_style_convert.py script, but note that occlusion, tracking and 3D informations are not available in that format.

Important Note

This dataset was introduced in the paper "Learning to Detect and Track Visible and Occluded Body Joints in a Virtual World" (ECCV 2018). In the experimental section of the paper, when referring to the "test set", we mean the set consisting of the test and val directories of the JTA Dataset.

Citation

We believe in open research and we are happy if you find this data useful.
If you use it, please cite our work.

@inproceedings{fabbri2018learning,
   title     = {Learning to Detect and Track Visible and Occluded Body Joints in a Virtual World},
   author    = {Fabbri, Matteo and Lanzi, Fabio and Calderara, Simone and Palazzi, Andrea and Vezzani, Roberto and Cucchiara, Rita},
   booktitle = {European Conference on Computer Vision (ECCV)},
   year      = {2018}
 }

License

JTA-Dataset</span> is licensed under a <a rel="license" href="http://creativecommons.org/licenses/by-nc/4.0/">Creative Commons Attribution-NonCommercial 4.0 International License</a>.