LoCO
This repository contains the source code related to the paper Compressed Volumetric Heatmaps for Multi-Person 3D Pose Estimation
Install / Use
/learn @fabbrimatteo/LoCOREADME
Learning on Compressed Output (LoCO)
Accepted to CVPR 2020
This repo contains the code related to the paper Compressed Volumetric Heatmaps for Multi-Person 3D Pose Estimation accepted to CVPR 2020 with the instructions for training and testing our models on the JTA dataset. Here you can also find the code for training the Volumetric Heatmap Autoencoder.
Some Results
<table> <tr> <th>Input</th> <th>Prediction</th> </tr> <tr> <th><img src=imgs/sample_1.jpg width=400></th> <th><img src=imgs/sample_1.gif width=400></th> </tr> <tr> <th><img src=imgs/sample_2.jpg width=400></th> <th><img src=imgs/sample_2.gif width=400></th> </tr> <tr> <th><img src=imgs/sample_3.jpg width=400></th> <th><img src=imgs/sample_3.gif width=400></th> </tr> <tr> <th><img src=imgs/sample_5.jpg width=400></th> <th><img src=imgs/sample_5.gif width=400></th> </tr> <tr> <th><img src=imgs/sample_4.jpg width=400></th> <th><img src=imgs/sample_4.gif width=400></th> </tr> </table>Quick Demo
- run
python demo.py --ex=1(python >= 3.6)- please wait some seconds: it will display some precomputed results. You can change the
exnumber from 1 to 3 to see different results
- please wait some seconds: it will display some precomputed results. You can change the
Compile Cuda Kernel
cdinto the foldernms3dand runpython setup.py install(python >= 3.6). Make sure to add your cuda directory to your environment variables.
Intructions
- Download the JTA dataset
in
<your_jta_path> - Run
python to_poses.py --out_dir_path='poses' --format='torch'(link) to generate the<your_jta_path>/posesdirectory - Run
python to_imgs.py --out_dir_path='frames' --img_format='jpg'(link) to generate the<your_jta_path>/framesdirectory - Download our precomputed codes from here
and unzip them into
<your_jta_path> - Modify the
conf/default.yamlconfiguration file specifying the path to the JTA dataset directoryJTA_PATH: <your_jta_path>
Train
- run
python main.py default(python >= 3.6)
Show Visual Results
- run
python show.py default(python >= 3.6)- Note that, before showing the results, you must have completed at least one training epoch; however, to achieve results comparable to those reported in the paper, it is advisable to carry out a training of at least 100 epochs
Show Paper Results
- Download the pretrained weights and extract them into the project folder
- Modify the
conf/pretrained.yamlconfiguration file specifying the path to the JTA dataset directoryJTA_PATH: <your_jta_path>
- run
python show.py pretrainedto show qualitative results (python >= 3.6) - run
python eval.py pretrainedto obtain the results reported in the paper (python >= 3.6)
Citation
We believe in open research and we are happy if you find this data useful.
If you use it, please cite our work.
@inproceedings{fabbri2020compressed,
title = {Compressed Volumetric Heatmaps for Multi-Person 3D Pose Estimation},
author = {Fabbri, Matteo and Lanzi, Fabio and Calderara, Simone and Alletto, Stefano and Cucchiara, Rita},
booktitle = {Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2020}
}
License
LoCO</span> is licensed under a <a rel="license" href="http://creativecommons.org/licenses/by-nc/4.0/">Creative Commons Attribution-NonCommercial 4.0 International License</a>.
Related Skills
node-connect
349.7kDiagnose OpenClaw node connection and pairing failures for Android, iOS, and macOS companion apps
frontend-design
109.7kCreate distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.
openai-whisper-api
349.7kTranscribe audio via OpenAI Audio Transcriptions API (Whisper).
qqbot-media
349.7kQQBot 富媒体收发能力。使用 <qqmedia> 标签,系统根据文件扩展名自动识别类型(图片/语音/视频/文件)。
