Mocap
Code to reproduce experiments in 'LSTM-based real-time action detection and prediction in human motion streams'
Install / Use
/learn @fabiocarrara/MocapREADME
Real-Time Action Detection and Prediction in Human Motion Streams
<p align=center> <img src="images/3144.gif" width="50%"><img src="images/3292.gif" width="50%"> </p>This code runs experiments for a real-time action detection in motion capture data implemented with LSTMs. It reproduces experiments presented in the following paper:
Carrara, F., Elias, P., Sedmidubsky, J., & Zezula, P. (2019).
LSTM-based real-time action detection and prediction in human motion streams.
Multimedia Tools and Applications, 78(19), 27309-27331.
Experiments are conducted on the HDM-05 dataset. NOTE: Few sequences from the HDM05 dataset are partially missing labels. The above videos show two sequences of this kind. The prediction of our model is on top, while the (wrong) groundtruth is on the bottom.
How to reproduce
-
Download the proprocessed data archive and extract it in the repo root folder: hdm05-mocap-data.tar.gz (~1GB, the original HDM05 dataset is available here)
-
Run
parse_HDM05_data.shto generate data splits. -
See
train_classify.pyandtrain_segment.pyif you want to train single models for classification or segmentation, respectively.To train all the segmentation models of the paper in batch or check out some examples of invocation, see
train_segmentation_models.sh.
Related Skills
node-connect
351.4kDiagnose OpenClaw node connection and pairing failures for Android, iOS, and macOS companion apps
frontend-design
110.7kCreate distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.
openai-whisper-api
351.4kTranscribe audio via OpenAI Audio Transcriptions API (Whisper).
qqbot-media
351.4kQQBot 富媒体收发能力。使用 <qqmedia> 标签,系统根据文件扩展名自动识别类型(图片/语音/视频/文件)。
