SkillAgentSearch skills...

OpenTAD

OpenTAD is an open-source temporal action detection (TAD) toolbox based on PyTorch.

Install / Use

/learn @sming256/OpenTAD
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

OpenTAD: An Open-Source Temporal Action Detection Toolbox.

<p align="left"> <a href="https://arxiv.org/abs/2502.20361" alt="arXiv"> <img src="https://img.shields.io/badge/arXiv-2502.20361-b31b1b.svg?style=flat" /></a> <a href="https://github.com/sming256/opentad/blob/main/LICENSE" alt="license"> <img src="https://img.shields.io/badge/License-Apache_2.0-blue.svg" /></a> <a href="https://github.com/sming256/OpenTAD/issues" alt="docs"> <img src="https://img.shields.io/github/issues-raw/sming256/OpenTAD?color=%23FF9600" /></a> <a href="https://img.shields.io/github/stars/sming256/opentad" alt="arXiv"> <img src="https://img.shields.io/github/stars/sming256/opentad" /></a> </p>

OpenTAD is an open-source temporal action detection (TAD) toolbox based on PyTorch.

🥳 What's New

📖 Major Features

  • Support SoTA TAD methods with modular design. We decompose the TAD pipeline into different components, and implement them in a modular way. This design makes it easy to implement new methods and reproduce existing methods.
  • Support multiple TAD datasets. We support 9 TAD datasets, including ActivityNet-1.3, THUMOS-14, HACS, Ego4D-MQ, EPIC-Kitchens-100, FineAction, Multi-THUMOS, Charades, and EPIC-Sounds Detection datasets.
  • Support feature-based training and end-to-end training. The feature-based training can easily be extended to end-to-end training with raw video input, and the video backbone can be easily replaced.
  • Release various pre-extracted features. We release the feature extraction code, as well as many pre-extracted features on each dataset.

🌟 Model Zoo

<table align="center"> <tbody> <tr align="center" valign="bottom"> <td> <b>One Stage</b> </td> <td> <b>Two Stage</b> </td> <td> <b>DETR</b> </td> <td> <b>End-to-End Training</b> </td> </tr> <tr valign="top"> <td> <ul> <li><a href="configs/actionformer">ActionFormer (ECCV'22)</a></li> <li><a href="configs/tridet">TriDet (CVPR'23)</a></li> <li><a href="configs/temporalmaxer">TemporalMaxer (arXiv'23)</a></li> <li><a href="configs/videomambasuite">VideoMambaSuite (arXiv'24)</a></li> <li><a href="configs/dyfadet">DyFADet (ECCV'24)</a></li> <li><a href="configs/causaltad">CausalTAD (arXiv'24)</a></li> </ul> </td> <td> <ul> <li><a href="configs/bmn">BMN (ICCV'19)</a></li> <li><a href="configs/gtad">GTAD (CVPR'20)</a></li> <li><a href="configs/tsi">TSI (ACCV'20)</a></li> <li><a href="configs/vsgn">VSGN (ICCV'21)</a></li> </ul> </td> <td> <ul> <li><a href="configs/tadtr">TadTR (TIP'22)</a></li> </ul> </td> <td> <ul> <li><a href="configs/afsd">AFSD (CVPR'21)</a></li> <li><a href="configs/tadtr">E2E-TAD (CVPR'22)</a></li> <li><a href="configs/etad">ETAD (CVPRW'23)</a></li> <li><a href="configs/re2tal">Re2TAL (CVPR'23)</a></li> <li><a href="configs/adatad">AdaTAD (CVPR'24)</a></li> </ul> </td> </tr> </td> </tr> </tbody> </table>

The detailed configs, results, and pretrained models of each method can be found in above folders.

🛠️ Installation

Please refer to install.md for installation.

📝 Data Preparation

Please refer to data.md for data preparation.

🚀 Usage

Please refer to usage.md for details of training and evaluation scripts.

📄 Updates

Please refer to changelog.md for update details.

🤝 Roadmap

All the things that need to be done in the future is in roadmap.md.

🖊️ Citation

[Acknowledgement] This repo is inspired by OpenMMLab project, and we give our thanks to their contributors.

If you think this repo is helpful, please cite us:

@article{liu2025opentad,
  title={OpenTAD: A Unified Framework and Comprehensive Study of Temporal Action Detection},
  author={Liu, Shuming and Zhao, Chen and Zohra, Fatimah and Soldan, Mattia and Pardo, Alejandro and Xu, Mengmeng and Alssum, Lama and Ramazanova, Merey and Alcázar, Juan León and Cioppa, Anthony and Giancola, Silvio and Hinojosa, Carlos and Ghanem, Bernard},
  journal={arXiv preprint arXiv:2502.20361},
  year={2025}
}

If you have any questions, please contact: shuming.liu@kaust.edu.sa.

Related Skills

View on GitHub
GitHub Stars325
CategoryContent
Updated3d ago
Forks25

Languages

Python

Security Score

100/100

Audited on Mar 31, 2026

No findings