SkillAgentSearch skills...

KeypointNeRF

KeypointNeRF Generalizing Image-based Volumetric Avatars using Relative Spatial Encoding of Keypoints

Install / Use

/learn @facebookresearch/KeypointNeRF
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

<p align="center"> <p align="center"> <h1 align="center">KeypointNeRF: Generalizing Image-based Volumetric Avatars using Relative Spatial Encoding of Keypoints</h1> </p> <p align="center" style="font-size:16px"> <a target="_blank" href="https://markomih.github.io/"><strong>Marko Mihajlovic</strong></a> · <a target="_blank" href="https://www.aayushbansal.xyz/"><strong>Aayush Bansal</strong></a> · <a target="_blank" href="https://zollhoefer.com/"><strong>Michael Zollhoefer</strong></a> . <a target="_blank" href="https://inf.ethz.ch/people/person-detail.MjYyNzgw.TGlzdC8zMDQsLTg3NDc3NjI0MQ==.html"><strong>Siyu Tang</strong></a> · <a target="_blank" href="https://shunsukesaito.github.io/"><strong>Shunsuke Saito</strong></a> </p> <h2 align="center">ECCV 2022</h2> <p> KeypointNeRF leverages human keypoints to instantly generate volumetric radiance representation from 2-3 input images without retraining or fine-tuning. It can represent human faces and full bodies. </p> <div align="center"></div> <img src="./assets/keynerf.gif" alt="Logo" width="100%"> <div align="center"></div> <img src="./assets/teaser.gif" alt="Logo" width="100%"> <p align="center"> <a href="https://pytorch.org/get-started/locally/"><img alt="PyTorch" src="https://img.shields.io/badge/PyTorch-ee4c2c?logo=pytorch&logoColor=white"></a> <a href="https://pytorchlightning.ai/"><img alt="Lightning" src="https://img.shields.io/badge/-Lightning-792ee5?logo=pytorchlightning&logoColor=white"></a> <br> <a href='https://arxiv.org/abs/2205.04992'> <img src='https://img.shields.io/badge/Paper-PDF-green?style=for-the-badge&logo=arXiv&logoColor=green' alt='Paper PDF'> </a> <a href='https://markomih.github.io/KeypointNeRF/' style='padding-left: 0.5rem;'> <img src='https://img.shields.io/badge/KeypointNeRF-Page-orange?style=for-the-badge&logo=Google%20chrome&logoColor=orange' alt='Project Page'> <a href="https://youtu.be/RMs1S5k9vrk"><img alt="youtube views" title="Subscribe to my YouTube channel" src="https://img.shields.io/youtube/views/RMs1S5k9vrk?logo=youtube&labelColor=ce4630&style=for-the-badge"/></a> </p> <p align="center"><a href='https://paperswithcode.com/sota/generalizable-novel-view-synthesis-on-zju?p=keypointnerf-generalizing-image-based'> <img src='https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/keypointnerf-generalizing-image-based/generalizable-novel-view-synthesis-on-zju' alt='Generalizable Novel View Synthesis'></a> <a href='https://paperswithcode.com/sota/3d-human-reconstruction-on-cape?p=keypointnerf-generalizing-image-based'> <img src='https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/keypointnerf-generalizing-image-based/3d-human-reconstruction-on-cape' alt='Generalizable Novel View Synthesis'></a></p> </p>

News :new:

  • [2022/10/01] Combine ICON with our relative spatial keypoint encoding for fast and convenient monocular reconstruction, without requiring the expensive SMPL feature. More details are here.

Installation

Please install python dependencies specified in environment.yml:

conda env create -f environment.yml
conda activate KeypointNeRF

Data preparation

Please see DATA_PREP.md to setup the ZJU-MoCap dataset.

After this step the data directory follows the structure:

./data/zju_mocap
├── CoreView_313
├── CoreView_315
├── CoreView_377
├── CoreView_386
├── CoreView_387
├── CoreView_390
├── CoreView_392
├── CoreView_393
├── CoreView_394
└── CoreView_396

Train your own model on the ZJU dataset

Execute train.py script to train the model on the ZJU dataset.

python train.py --config ./configs/zju.json --data_root ./data/zju_mocap

After the training, the model checkpoint will be stored under ./EXPERIMENTS/zju/ckpts/last.ckpt, which is equivalent to the one provided here.

Evaluation

To extract render and evaluate images, execute:

python train.py --config ./configs/zju.json --data_root ./data/zju_mocap --run_val
python eval_zju.py --src_dir ./EXPERIMENTS/zju/images_v3

To visualize the dynamic results, execute:

python render_dynamic.py --config ./configs/zju.json --data_root ./data/zju_mocap --model_ckpt ./EXPERIMENTS/zju/ckpts/last.ckpt
<div align="center"></div> <img src="./assets/zju_result_sub393.gif" alt="Logo" width="100%"> <p align="center" style="font-size:12px"> (The first three views of an unseen subject are the input to KeypointNeRF; the last image is a rendered novel view) </p>

We compare KeypointNeRF with recent state-of-the-art methods. The evaluation metric is SSIM and PSNR. | Models | PSNR ↑ | SSIM ↑ | |---|---|---| | pixelNeRF <font size="1">(Yu et al., CVPR'21)</font> | 23.17 | 86.93 | | PVA <font size="1">(Raj et al., CVPR'21)</font> | 23.15 | 86.63 | | NHP <font size="1">(Kwon et al., NeurIPS'21)</font> | 24.75 | 90.58 | | KeypointNeRF* <font size="1">(Mihajlovic et al., ECCV'22)</font> | 25.86 | 91.07 |

<p align="left" style="font-size:10px"> (*Note that results of KeypointNeRF are slightly higher compared to the numbers reported in the original paper due to training views not beeing shuffled during training.) </p>

Reconstruction from a Single Image

Our relative spatial encoding can be used to reconstruct humans from a single image. As a example, we leverage ICON and replace its expensive SDF feature with our relative spatial encoding.

<div align="center"><img src="./assets/icon_vs_kpts.png" alt="Logo" width="100%"> While it achieves comparable quality to ICON, it's much <strong>faster</strong> and more <strong>convinient</strong> to use <span align="left" style="font-size:8px">(*displayed image taken from pinterest.com)</span>. </div>

3D Human Reconstruction on CAPE

| Models | Chamfer ↓ (cm) | P2S ↓ (cm) | |---|---|---| | PIFu <font size="1">(Saito et al., ICCV'19)</font> | 3.573 | 1.483 | | ICON <font size="1">(Xiu et al., CVPR'22)</font> | 1.424 | 1.351 | | KeypointICON <font size="1">(Mihajlovic et al., ECCV'22; Xiu et al., CVPR'22)</font> | 1.539 | 1.358 |

Check the benchmark here and more details here.

Publication

If you find our code or paper useful, please consider citing:

@inproceedings{Mihajlovic:ECCV2022,
  title = {{KeypointNeRF}: Generalizing Image-based Volumetric Avatars using Relative Spatial Encoding of Keypoints},
  author = {Mihajlovic, Marko and Bansal, Aayush and Zollhoefer, Michael and Tang, Siyu and Saito, Shunsuke},
  booktitle={European conference on computer vision},
  year={2022},
}

License

CC-BY-NC 4.0. See the LICENSE file.

View on GitHub
GitHub Stars376
CategoryDevelopment
Updated3d ago
Forks27

Languages

Python

Security Score

80/100

Audited on Mar 30, 2026

No findings