SkillAgentSearch skills...

LatentEditor

LatentEditor: Text Driven Local Editing of 3D Scenes

Install / Use

/learn @umarkhalidAI/LatentEditor
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

<h2 align="center"> <a href="https://github.com/umarkhalidAI/LatentEditor">LatentEditor: Text Driven Local Editing of 3D Scenes</a></h2> <h5 align="center"> If you like our project, please give us a star ⭐ on GitHub for latest update. </h2> <h5 align="center">

webpage arXiv License: MIT

</h5>

Project page | Paper

<img src="assets/teaser_final.jpg"/>

😮 Highlights

LatentEditor allows you to edit your 3D scenes in the matter of minutes!.

💡 View consistent, high quality and fast speed

  • Stable Diffusion for high quality and controllable repainting for reference alignment --> view-consistent high-quality image generation.
  • View-consistent high-quality images with simple MSE loss --> fast high-quality 3D content reconstruction.

🚩 Updates

Welcome to watch 👀 this repository for the latest updates.

[2023.12.18] : We have released our paper, LatentEditor on arXiv.

[2023.12.18] : Release project page.

  • [ ] Code release.

🛠️ Methodology

<img src="assets/main_fig.jpg"/> Overall pipeline of LatentEditor for model initialization and editing. (a) We initialize the NeRF model within the latent domain, guided by the latent features of the original dataset. Our refinement adapter mitigates the misalignment in the latent space and encompasses a trainable adapter with residual and self-attention mechanisms. (b) Upon initialization, LatentEditor iteratively refines the model within the latent space for a predetermined number of iterations, while consistently updating the training set with the edited latents, Ze. (c) The Delta Module is adept at interpreting prompts and produces the mask for targeted editing. Additionally, it integrates the denoising U-Net from IP2P. However, ∆t used in calculating delta scores ∆e is a hyperparameter and is different from t which is randomly selected. An RGB image can be obtained by feeding the edited latent to the stable diffusion (SD) decoder D whereas E represents SD encoder.

🚀 3D-Editing Results

Qualitative comparison

<img src="assets/compare_qual.jpg"/>

Quantitative comparison

<img src="assets/compare_quan.png"/> Quantitative evaluation of scene edits in terms of text alignment and frame consistency in CLIP space where our approach demonstrates the highest consistency.

👍 Acknowledgement

This work is built on many amazing research works and open-source projects, thanks a lot to all the authors for sharing!

✏️ Citation

If you find our paper and code useful in your research, please consider giving a star :star: and citation :pencil:.

@misc{khalid2023latenteditor,
      title={LatentEditor: Text Driven Local Editing of 3D Scenes}, 
      author={Umar Khalid and Hasan Iqbal and Nazmul Karim and Jing Hua and Chen Chen},
      year={2023},
      eprint={2312.09313},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}
<!---->
View on GitHub
GitHub Stars67
CategoryDevelopment
Updated2mo ago
Forks2

Security Score

95/100

Audited on Jan 27, 2026

No findings