SkillAgentSearch skills...

Tri2plane

The repository for 'Tri$^{2}$-plane: Volumetric Avatar Reconstruction with Feature Pyramid'

Install / Use

/learn @Songluchuan/Tri2plane
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

Tri2Plane

The repository for Tri<sup>2</sup>-plane: Thinking Head Avatar via Feature Pyramid

[Tri<sup>2</sup>-plane: Thinking Head Avatar via Feature Pyramid],
Arxiv Link: https://arxiv.org/abs/2401.09386,
Luchuan Song, Pinxin Liu, Lele Chen, Guojun Yin and Chenliang Xu,
Project Page: https://songluchuan.github.io/Tri2Plane.github.io/,
Arxiv 2024

<div align="center"> <!-- <a href="https://github.com/Songluchuan/Tri2plane" target="_blank"> --> <img src="media/Teaser.jpg" alt="" width="1120" style="height: auto;" /> </a> </div> <h2 align="center">Abstract</h2> Recent years have witnessed considerable achievements in facial avatar reconstruction with neural volume rendering. Despite notable advancements, the reconstruction of complex and dynamic head movements from monocular videos still suffers from capturing and restoring fine-grained details. In this work, we propose a novel approach, named Tri$^2$-plane, for monocular photo-realistic volumetric head avatar reconstructions. Distinct from the existing works that rely on a single tri-plane deformation field for dynamic facial modeling, the proposed Tri<sup>^2</sup>-Plane leverages the principle of feature pyramids and three top-to-down lateral connections tri-planes for details improvement. It samples and renders facial details at multiple scales, transitioning from the entire face to specific local regions and then to even more refined sub-regions. Moreover, we incorporate a camera-based geometry-aware sliding window method as an augmentation in training, which improves the robustness beyond the canonical space, with a particular improvement in cross-identity generation capabilities. Experimental outcomes indicate that the Tri<sup>^2</sup>-Plane not only surpasses existing methodologies but also achieves superior performance across quantitative and qualitative assessments. The project website is located on: https://songluchuan.github.io/Tri2Plane.github.io/. <h2 align="center">Code</h2>

Requirements

Python 3.8+ and PyTorch 1.13.0+cu116.

Note: The code has been tested only with PyTorch 1.13.0, there are no guarantees that it is compatible with other versions.


Installation

You can install the package by running

conda env create -f environments.yml

If you run into any issues installing PyTorch3D, try compiling and installing it manually.


Data Preprocessing


Training


Reenactment

<h2 align="center">Video</h2> <div align="center"> <a href="https://www.youtube.com/watch?v=qU7zijI2DCQ" target="_blank"> <img src="media/video.jpg" alt="youtube video" width="1120" style="height: auto;" /> </a> </div> <h2 align="center">Citation</h2>
@article{song2024tri,
  title={Tri $\^{}$\{$2$\}$ $-plane: Volumetric Avatar Reconstruction with Feature Pyramid},
  author={Song, Luchuan and Liu, Pinxin and Chen, Lele and Yin, Guojun and Xu, Chenliang},
  journal={arXiv preprint arXiv:2401.09386},
  year={2024}
}
View on GitHub
GitHub Stars141
CategoryDevelopment
Updated8mo ago
Forks16

Languages

Python

Security Score

72/100

Audited on Jul 28, 2025

No findings