Home

Awesome

PWC PWC

<div align="center"> <h1>Semi-Supervised Temporal Action Detection with Proposal-Free Masking</h1> <div> <a href='https://sauradip.github.io/' target='_blank'>Sauradip Nag</a><sup>1,2,+</sup>&emsp; <a href='https://scholar.google.co.uk/citations?hl=en&user=ZbA-z1cAAAAJ&view_op=list_works&sortby=pubdate' target='_blank'>Xiatian Zhu</a><sup>1,3</sup>&emsp; <a href='https://scholar.google.co.uk/citations?user=irZFP_AAAAAJ&hl=en' target='_blank'>Yi-Zhe Song</a><sup>1,2</sup>&emsp; <a href='https://scholar.google.co.uk/citations?hl=en&user=MeS5d4gAAAAJ&view_op=list_works&sortby=pubdate' target='_blank'>Tao Xiang</a><sup>1,2</sup>&emsp; </div> <div> <sup>1</sup>CVSSP, University of Surrey, UK&emsp; <sup>2</sup>iFlyTek-Surrey Joint Research Center on Artificial Intelligence, UK&emsp; <br> <sup>3</sup>Surrey Institute for People-Centred Artificial Intelligence, UK </div> <div> <sup>+</sup>corresponding author </div> <h3><strong>Accepted to <a href='https://eccv2022.ecva.net/' target='_blank'>ECCV 2022</a></strong></h3> <h3 align="center"> <a href="https://arxiv.org/abs/2207.07059" target='_blank'>Paper</a> | <a href="https://sauradip.github.io/project_pages/SPOT/" target='_blank'>Project Page</a> </h3> <table style="border:none;border-color:white;"> <tr style="border:0px;border-color:white;"> <td style="border:0px;border-color:white;"><img src="asserts/SPOT_Combo.gif" width="100%"/></td> </tr> </table> </div>

Updates

Summary

Abstract

Existing temporal action detection (TAD) methods rely on a large number of training data with segment-level annotations. Collecting and annotating such a training set is thus highly expensive and unscalable. Semi-supervised TAD (SS-TAD) alleviates this problem by leveraging unlabeled videos freely available at scale. However, SS-TAD is also a much more challenging problem than supervised TAD, and consequently much under-studied. Prior SS-TAD methods directly combine an existing proposal-based TAD method and a SSL method. Due to their sequential localization (e.g, proposal generation) and classification design, they are prone to proposal error propagation. To overcome this limitation, in this work we propose a novel Semi-supervised Temporal action detection model based on PropOsal-free Temporal mask (SPOT) with a parallel localization (mask generation) and classification architecture. Such a novel design effectively eliminates the dependence between localization and classification by cutting off the route for error propagation in-between. We further introduce an interaction mechanism between classification and localization for prediction refinement, and a new pretext task for self-supervised model pre-training. Extensive experiments on two standard benchmarks show that our SPOT outperforms state-of-the-art alternatives, often by a large margin.

Architecture

Getting Started

Requirements

Environment Setup

It is suggested to create a Conda environment and install the following requirements

pip3 install -r requirements.txt

Download Features

Download the video features and update the Video paths/output paths in config/anet.yaml file. For now ActivityNetv1.3 dataset config is available. We are planning to release the code for THUMOS14 dataset soon.

DatasetFeature BackbonePre-TrainingLink
ActivityNetTSNKinetics-400Google Drive
THUMOSTSNKinetics-400Google Drive
ActivityNetI3DKinetics-400Google Drive
THUMOSI3DKinetics-400Google Drive

Model Training

To train SPOT from scratch run the following command. The training configurations can be adjusted from config/anet.yaml file. This training includes both Pre-training and the fine-tuning stages.

python spot_train.py

Model Inference

We provide the pretrained models containing the checkpoint for I3D features on ActivityNetv1.3 . It can be found in the Link

After downloading the checkpoints, the checkpoints path can be saved in config/anet.yaml file. The model inference can be then performed using the following command

python spot_inference.py

Model Evaluation

To evaluate our TAGS model run the following command.

python eval.py

Performance

Qualitative Results

TO-DO Checklist

Acknowlegement

This code repository has borrowed some parts of SSTAP and BMN. We thank the author for open-sourcing their codes and clarifying the doubts.

Citation

If you find this project useful for your research, please use the following BibTeX entry.

@article{nag2022temporal,
  title={Temporal Action Detection with Global Segmentation Mask Learning},
  author={Nag, Sauradip and Zhu, Xiatian and Song, Yi-Zhe and Xiang, Tao},
  journal={arXiv preprint arXiv:2207.06580},
  year={2022}
}
}