SkillAgentSearch skills...

Mocap

Code to reproduce experiments in 'LSTM-based real-time action detection and prediction in human motion streams'

Install / Use

/learn @fabiocarrara/Mocap
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

Real-Time Action Detection and Prediction in Human Motion Streams

<p align=center> <img src="images/3144.gif" width="50%"><img src="images/3292.gif" width="50%"> </p>

This code runs experiments for a real-time action detection in motion capture data implemented with LSTMs. It reproduces experiments presented in the following paper:

Carrara, F., Elias, P., Sedmidubsky, J., & Zezula, P. (2019).
LSTM-based real-time action detection and prediction in human motion streams.
Multimedia Tools and Applications, 78(19), 27309-27331.

Experiments are conducted on the HDM-05 dataset. NOTE: Few sequences from the HDM05 dataset are partially missing labels. The above videos show two sequences of this kind. The prediction of our model is on top, while the (wrong) groundtruth is on the bottom.

How to reproduce

  1. Download the proprocessed data archive and extract it in the repo root folder: hdm05-mocap-data.tar.gz (~1GB, the original HDM05 dataset is available here)

  2. Run parse_HDM05_data.sh to generate data splits.

  3. See train_classify.py and train_segment.py if you want to train single models for classification or segmentation, respectively.

    To train all the segmentation models of the paper in batch or check out some examples of invocation, see train_segmentation_models.sh.

Related Skills

View on GitHub
GitHub Stars29
CategoryDevelopment
Updated1y ago
Forks7

Languages

Python

Security Score

65/100

Audited on Oct 23, 2024

No findings