SPIRAL
[NeurIPS 2025] SPIRAL: Semantic-Aware Progressive LiDAR Scene Generation and Understanding
Install / Use
/learn @worldbench/SPIRALREADME
| <img src="./images/teaser.png" alt="Teaser" width="100%"> | | :-: |
Existing LiDAR generative models are limited to producing unlabeled LiDAR scenes, lacking any semantic annotations. Performing post-hoc labeling on these generated scenes requires additional pretrained segmentation models, which introduces extra computational overhead. Moreover, such after-the-fact annotation yields suboptimal segmentation quality.
To address this issue, we make the following contributions:
- We propose a novel state-of-the-art semantic-aware range-view LiDAR diffusion model, Spiral, which jointly produces depth and reflectance images along with semantic labels.
- We introduce unified evaluation metrics that comprehensively evaluate the geometric, physical, and semantic quality of generated labeled LiDAR scenes.
- We demonstrate the effectiveness of the generated LiDAR scenes for training segmentation models, highlighting Spiral's potential for generative data augmentation.
:books: Citation
If you find this work helpful for your research, please kindly consider citing our paper:
@inproceedings{zhu2025spiral,
title = {Spiral: Semantic-Aware Progressive LiDAR Scene Generation and Understanding},
author = {Zhu, Dekai and Hu, Yixuan and Liu, Youquan and Lu, Dongyue and Kong, Lingdong and Ilic, Slobodan},
booktitle = {The Thirty-ninth Annual Conference on Neural Information Processing Systems},
year = {2025}
}
Updates
- [11/2025] - The code for Spiral is released. :rocket:
- [10/2025] - The project page is online. :rocket:
- [09/2025] - This work has been accepted to NeurIPS 2025.
:gear: Installation
For details related to installation and environment setups, please run:
conda env create -f environment.yaml
conda activate spiral
If you are stuck with an endless installation, try:
mamba env create -f environment.yaml
conda activate spiral
:hotsprings: Data Preparation
We use the official SemanticKITTI API to preprocess the data by projecting the LiDAR data from Cartesian coordinates into range images. You can download the preprocessed data here. :hugs:
:rocket: Getting Started
First, specify the data_path in utils/option.py to point to the directory of the preprocessed data. Then simply run:
python train.py
to start the training.
<!-- ## License This work is under the <a rel="license" href="https://www.apache.org/licenses/LICENSE-2.0">Apache License Version 2.0</a>, while some specific implementations in this codebase might be with other licenses. Kindly refer to [LICENSE.md](docs/LICENSE.md) for a more careful check, if you are using our code for commercial matters. -->Acknowledgements
This work is developed based on the R2DM codebase.
Related Skills
node-connect
349.2kDiagnose OpenClaw node connection and pairing failures for Android, iOS, and macOS companion apps
frontend-design
109.5kCreate distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.
openai-whisper-api
349.2kTranscribe audio via OpenAI Audio Transcriptions API (Whisper).
qqbot-media
349.2kQQBot 富媒体收发能力。使用 <qqmedia> 标签,系统根据文件扩展名自动识别类型(图片/语音/视频/文件)。
