Home

Awesome

THU<sup>E-ACT</sup>-50: A Real-World Event-Based Action Recognition Benchmark

📢 Update: We are excited to announce the release of a larger and more comprehensive dataset, THU<sup>MV-EACT</sup>-50, which extends the THU<sup>E-ACT</sup>-50 to include multi-view action recognition. For more details, please visit THU-MV-EACT-50.

Introduced by the paper "Action Recognition and Benchmark Using Event Cameras" in TPAMI 2023, THU<sup>E-ACT</sup>-50 stands as a large-scale, real-world event-specific action recognition dataset with more than 4 times the size of the current largest event-based action recognition dataset. It contains 50 action categories and is primarily designed for whole-body motions and indoor healthcare applications. This repository provides access to the dataset, alongside detailed information about its contents and structure.

<img src="figures/sample-sequences.jpg" alt="Sample-sequences" style="zoom: 33%;" />

Dataset Overview

THU<sup>E-ACT</sup>-50 is designed to address the limitations of existing event-based action recognition datasets, which are often too small and limited in the range of actions they cover. The dataset consists of two parts: the standard THU<sup>E-ACT</sup>-50 and a more challenging version,THU<sup>E-ACT</sup>-50 CHL, which is designed to test the robustness of algorithms under challenging conditions.

The dataset comprises a diverse set of action categories, including whole-body motions, indoor healthcare applications, detail-oriented actions, confusing actions, human-object interactions, and two-player interactive movements. With a total of 10,500 video recordings for the standard THU<sup>E-ACT</sup>-50 and 2,330 recordings for the challenging THU<sup>E-ACT</sup>-50 CHL, this dataset provides an extensive and varied collection of action sequences for researchers to explore and evaluate their models.

Dataset Description

Standard THU<sup>E-ACT</sup>-50

Challenging THU<sup>E-ACT</sup>-50 CHL

List of Actions

IDActionIDActionIDActionIDActionIDAction
A0WalkingA10Cross armsA20Calling with phoneA30FanA40Check time
A1RunningA11SaluteA21ReadingA31Open umbrellaA41Drink water
A2Jump upA12Squat downA22Tai chiA32Close umbrellaA42Wipe face
A3Running in circlesA13Sit downA23Swing objectsA33Put on glassesA43Long jump
A4Falling downA14Stand upA24ThrowA34Take off glassesA44Push up
A5Waving one handA15Sit and standA25StaggeringA35Pick upA45Sit up
A6Waving two handsA16Knead faceA26HeadacheA36Put on bagA46Shake hands (two-players)
A7ClapA17Nod headA27StomachacheA37Take off bagA47Fighting (two-players)
A8Rub handsA18Shake headA28Back painA38Put object into bagA48Handing objects (two-players)
A9PunchA19Thumb upA29VomitA39Take object out of bagA49Lifting chairs (two-players)

Evaluation Criteria

To evaluate the performance of event-based action recognition methods on the THU<sup>E-ACT</sup>-50 and THU<sup>E-ACT</sup>-50 CHL datasets, we divided the subjects in a ratio of 8:2 to create disjoint identity sets for training and testing. The training and test sets of the THU<sup>E-ACT</sup>-50 dataset contain 85 and 20 persons, respectively, while the training and test sets of the THU<sup>E-ACT</sup>-50 CHL dataset contain 14 and 4 persons, respectively.

We report the following evaluation metrics for each dataset:

Dataset Download

We're pleased to announce the release of the THU<sup>E-ACT</sup>-50 and THU<sup>E-ACT</sup>-50 CHL datasets.

THU<sup>E-ACT</sup>-50

Note: After decompression, the dataset will require about 332GB of storage space.

THU<sup>E-ACT</sup>-50 CHL

Note: After decompression, the dataset will occupy approximately 4.6GB of storage space.

Dataset Format

In the two datasets, the division for training and test sets can be found in the train.txt and test.txt files, respectively. Each line consists of File Name and Action ID.

The preprocessing operations for the 2 datasets can be found in dataset.py.

THU<sup>E-ACT</sup>-50

In the THU-EACT-50 dataset, which is provided in the .csv format, the data is structured with 5 columns as follows:

THU<sup>E-ACT</sup>-50 CHL

For the THU-EACT-50-CHL dataset, which is available in the .npy format, each line contains 4 elements:

Acknowledgements

We would like to express our sincere gratitude to Tsinghua University, partner companies, and organizations for their invaluable support and collaboration in making this dataset possible. Additionally, we extend our thanks to all the volunteers who participated in the data collection process. Their contributions have been instrumental in the development and evaluation of this benchmark.

License

This dataset is licensed under the MIT License.

Citing Our Work

If you find this dataset beneficial for your research, please cite our works:

@article{gao2023action,
  title={Action Recognition and Benchmark Using Event Cameras},
  author={Gao, Yue and Lu, Jiaxuan and Li, Siqi and Ma, Nan and Du, Shaoyi and Li, Yipeng and Dai, Qionghai},
  journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
  year={2023},
  volume={45},
  number={12},
  pages={14081-14097},
  publisher={IEEE}
}

@article{gao2024hypergraph,
  title={Hypergraph-Based Multi-View Action Recognition Using Event Cameras},
  author={Gao, Yue and Lu, Jiaxuan and Li, Siqi and Li, Yipeng and Du, Shaoyi},
  journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
  year={2024},
  publisher={IEEE}
}