PixelwiseRegression
PyTorch release for paper "Pixel-wise Regression: 3D Hand Pose Estimation via Spatial-form Representation and Differentiable Decoder"
Install / Use
/learn @IcarusWizard/PixelwiseRegressionREADME
Pixel-wise Regression for 3D hand pose estimation
PyTroch release of our paper:
Pixel-wise Regression: 3D Hand Pose Estimation via Spatial-form Representation and Differentiable Decoder
Xingyuan Zhang, Fuhai Zhang
If you find this repository useful, please make a reference in your paper:
@ARTICLE{zhang2022srnet,
author={Zhang, Xingyuan and Zhang, Fuhai},
journal={IEEE Transactions on Multimedia},
title={Differentiable Spatial Regression: A Novel Method for 3D Hand Pose Estimation},
year={2022},
volume={24},
number={},
pages={166-176},
doi={10.1109/TMM.2020.3047552}
}
Update: The paper has been acceptted at TMM! Title has changed as suggested by one of the reviewers. Please consider cite the new version. I did not upload the new version to Arxiv since I am not sure if it is allowed. If you know it is ok to do so, please contact me and I am glad to do the update.
Setup
conda env create -f env.yml
conda activate pixelwise
Dataset
All datasets should be placed in ./Data folder. After placing datasets correctly, run python check_dataset.py --dataset <dataset_name> to build the data files used to train.
NYU
- Download the dataset from website.
- Unzip the files to
./Dataand rename the folder asNYU.
MSRA
- Download the dataset from dropbox.
- Unzip the files to
./Dataand rename the folder asMSRA.
ICVL
- Download the dataset from here.
- Extract
Training.tar.gzandTesting.tar.gzto./Data/ICVL/Trainingand./Data/ICVL/Testingrespectively.
HAND17
- Ask for the permission from the website and download.
- Download center files from github release, and put them in
Data/HAND17/. - Extract
frame.zipandimages.zipto./Data/HAND17/. Your should end with a folder look like below:
HAND17/
|
|-- hands17_center_train.txt
|
|-- hands17_center_test.txt
|
|-- training/
| |
| |-- images/
| |
| |-- Training_Annotation.txt
|
|-- frame/
| |
| |-- images/
| |
| |-- BoundingBox.txt
Train
Run python train.py --dataset <dataset_name>, dataset_name can be chose from NYU, ICVL and HAND17.
For MSRA dataset, you should run python train_msra.py --subject <subject_id>.
Test
Run python test.py --dataset <dataset_name>.
For MSRA dataset, you should run python test_msra.py --subject <subject_id>.
Results
Results and pretrained models are available in github release. These pretrained models are under a CC BY 4.0 license.
Related Skills
node-connect
345.9kDiagnose OpenClaw node connection and pairing failures for Android, iOS, and macOS companion apps
frontend-design
106.4kCreate distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.
openai-whisper-api
345.9kTranscribe audio via OpenAI Audio Transcriptions API (Whisper).
qqbot-media
345.9kQQBot 富媒体收发能力。使用 <qqmedia> 标签,系统根据文件扩展名自动识别类型(图片/语音/视频/文件)。
