SpatialGen
[3DV 2026] SpatialGen: Layout-guided 3D Indoor Scene Generation
Install / Use
/learn @manycore-research/SpatialGenREADME
SpatialGen: Layout-guided 3D Indoor Scene Generation
<!-- markdownlint-disable first-line-h1 --> <!-- markdownlint-disable html --> <!-- markdownlint-disable no-duplicate-header --> <div align="center"> <img src="assets/logo_light.png#gh-light-mode-only" width="60%" alt="SpatialLM" /> <img src="assets/logo_dark.png#gh-dark-mode-only" width="60%" alt="SpatialLM" /> </div> <hr style="margin-top: 0; margin-bottom: 8px;"> <div align="center" style="margin-top: 0; padding-top: 0; line-height: 1;"> <a href="https://manycore-research.github.io/SpatialGen" target="_blank" style="margin: 2px;"><img alt="Project" src="https://img.shields.io/badge/🌐%20Project-SpatialGen-ffc107?color=42a5f5&logoColor=white" style="display: inline-block; vertical-align: middle;"/></a> <a href="https://arxiv.org/abs/2509.14981" target="_blank" style="margin: 2px;"><img alt="arXiv" src="https://img.shields.io/badge/arXiv-SpatialGen-b31b1b?logo=arxiv&logoColor=white" style="display: inline-block; vertical-align: middle;"/></a> <a href="https://github.com/manycore-research/SpatialGen" target="_blank" style="margin: 2px;"><img alt="GitHub" src="https://img.shields.io/badge/GitHub-SpatialGen-24292e?logo=github&logoColor=white" style="display: inline-block; vertical-align: middle;"/></a> <a href="https://huggingface.co/manycore-research/SpatialGen-1.0" target="_blank" style="margin: 2px;"><img alt="Hugging Face" src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-SpatialGen-ffc107?color=ffc107&logoColor=white" style="display: inline-block; vertical-align: middle;"/></a> </div> <div align="center">Chuan Fang, Heng Li, Yixun Liang, Jia Zheng, Yongsen Mao, Yuan Liu, Rui Tang, Zihan Zhou, Ping Tan
HKUST Spatial Artificial Intelligence Lab; Manycore Tech Inc
</div> <div align="center">| Image-to-Scene Results | Text-to-Scene Results |
| :--------------------------------------: | :----------------------------------------: |
|
|
|
✨ News
- [Jan, 2026] The offical SpatialGen Dataset will be released soon, it is undergoing the data releasing approval process of Manycore Tech Inc.
- [Jan, 2026] We release the training code of SpatialGen, incuding the multi-view diffusion training and the gassuain optimization.
- [Nov, 2025] SpatialGen is accepted to 3DV 2026!
- [Sep, 2025] We release the paper of SpatialGen!
- [Aug, 2025] Initial release of SpatialGen-1.0!
📋 Release Plan
- [x] Provide inference code of SpatialGen.
- [x] Provide training instruction for SpatialGen.
- [ ] Release SpatialGen dataset.
SpatialGen Models
<div align="center">| Model | Download | | :-----------------------: | -------------------------------------------------------------------------------------| | SpatialGen-1.0 | 🤗 HuggingFace | | FLUX.1-Wireframe-dev-lora | 🤗 HuggingFace |
</div>Usage
🔧 Installation
Tested with the following environment:
- Python 3.10
- PyTorch 2.3.1
- CUDA Version 12.1
# clone the repository
git clone --recursive https://github.com/manycore-research/SpatialGen.git
cd SpatialGen
python -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt
pip install src/recons/Sparse-RaDeGS/submodules/diff-gaussian-rasterization
# Optional: fix the [flux inference bug](https://github.com/vllm-project/vllm/issues/4392)
pip install nvidia-cublas-cu12==12.4.5.8
📊 Dataset
We provide SpatialGen-Testset with 48 rooms, which labeled with 3D layout and 4.8K rendered images (48 x 100 views, including RGB, normal, depth maps and semantic maps) for MVD inference.
Inference
# download the pretrain weights
huggingface-cli download --resume-download manycore-research/SpatialGen-1.0 --local-dir spatialgen_ckpts
# Single image-to-3D Scene
bash scripts/infer_spatialgen_i2s.sh
# Text-to-image-to-3D Scene
# step 1. prepare controlnet conditional images, save to /path/to/your/spatialgen-testset/scene_xxxx/condition
python3 preprocess/prepare_flux_ctrlnet_conditions.py --dataset_dir /path/to/your/spatialgen-testset
# step 2. run spatialgen text2scene
# in captions/spatialgen_testset_captions.jsonl, we provide text prompts of different styles for each room,
# choose a pair of scene_id and prompt to run the text2scene experiment
bash scripts/infer_spatialgen_t2s.sh
Training
# TODO: preprocess the dataset
# run SCM_VAE training
bash scripts/train_scm_vae.sh
# run multi-view diffusion training
bash scripts/train_spatialgen_mvd.sh
License
SpatialGen-1.0 is derived from Stable-Diffusion-v2.1, which is licensed under the CreativeML Open RAIL++-M License. FLUX.1-Wireframe-dev-lora is licensed under the FLUX.1-dev Non-Commercial License.
Citation
@inproceedings{SpatialGen,
title = {SpatialGen: Layout-guided 3D Indoor Scene Generation},
author = {Fang, Chuan and Li, Heng and Liang, Yixu and Zheng, Jia and Mao, Yongsen and Liu, Yuan and Tang, Rui and Zhou, Zihan and Tan, Ping},
booktitle = {International Conference on 3D Vision},
year = {2026}
}
Acknowledgements
We would like to thank the following projects that made this work possible:
Related Skills
qqbot-channel
353.3kQQ 频道管理技能。查询频道列表、子频道、成员、发帖、公告、日程等操作。使用 qqbot_channel_api 工具代理 QQ 开放平台 HTTP 接口,自动处理 Token 鉴权。当用户需要查看频道、管理子频道、查询成员、发布帖子/公告/日程时使用。
docs-writer
100.7k`docs-writer` skill instructions As an expert technical writer and editor for the Gemini CLI project, you produce accurate, clear, and consistent documentation. When asked to write, edit, or revie
model-usage
353.3kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
project-overview
FlightPHP Skeleton Project Instructions This document provides guidelines and best practices for structuring and developing a project using the FlightPHP framework. Instructions for AI Coding A
