Home

Awesome

Joint learning of object and action detectors

By Vicky Kalogeiton, Philippe Weinzaepfel, Vittorio Ferrari, Cordelia Schmid

Introduction

In our ICCV17 paper we jointly detect objects performing various avtions in videos. Here, we provide the evaluation code used in our experiments.

Citing Joint learning of object and action detectors

If you find our evaluation code useful in your research, please cite:

@inproceedings{kalogeiton17biccv,
  TITLE = {Joint learning of object and action detectors},
  AUTHOR = {Kalogeiton, Vicky and Weinzaepfel, Philippe and Ferrari, Vittorio and Schmid, Cordelia},
  YEAR = {2017},
  BOOKTITLE = {ICCV},
}

Evalutation of all architectures

You can download our detection results (multitask, hierarchical and cartesian):

curl http://pascal.inrialpes.fr/data2/joint-objects-actions/JointLearningDetections.tar.gz | tar xz 

To run the mAP evaluation function for the multitask, hierarchical and cartesian cases (Table 4 in our paper), run:

vic_A2D_eval(learning_case) # learning_case: 1, 2 or 3 for multitask, hierarchical and cartesian, respectively

Zero shot learning

To run the mAP evaluation function for the zero shot learning (Table 5 in our paper), run:

vic_A2D_eval_zeroshot 

Object-action semantic segmentation

You can download our semantic segmentation images:

curl http://pascal.inrialpes.fr/data2/joint-objects-actions/JointLearningSegmentationResults.tar.gz | tar xz 

To run the semantic segmentation evaluation function (Table 6 in our paper), run:

vic_eval_segmentation