OmniSplat
[CVPR 2025] OmniSplat: Taming Feed-Forward 3D Gaussian Splatting for Omnidirectional Images with Editable Capabilities
Install / Use
/learn @esw0116/OmniSplatREADME
<div align="center">
</div>
<p align="center">
<h1 align="center">OmniSplat: Taming Feed-Forward 3D Gaussian Splatting for </br> Omnidirectional Images with Editable Capabilities </h1>
<p align="center">
<a href="https://esw0116.github.io/">Suyoung Lee</a><sup>1*</sup>
·
<a href="https://robot0321.github.io/">Jaeyoung Chung</a><sup>1*</sup>
·
Kihoon Kim<sup>1</sup>
·
Jaeyoo Huh<sup>1</sup>
</br>
Gunhee Lee<sup>2</sup>
·
Minsoo Lee<sup>2</sup>
·
<a href="https://cv.snu.ac.kr/index.php/~kmlee/">Kyoung Mu Lee</a><sup>1</sup>
</br>
1: Seoul National Univiersity 2: LG AI Research Center
</br>
(* denotes equal contribution)
</p>
<h3 align="center">CVPR 2025, Highlight</h3>
</p>
<!-- <div align="center">
[![ArXiv]()]()
[](https://github.com/luciddreamer-cvlab/LucidDreamer)
[](https://github.com/luciddreamer-cvlab/LucidDreamer/blob/master/LICENSE)
</div> -->
<p align="center">
<img src="figures/logo_cvlab.png" height=60 style="margin-right:40px;">
<img src="figures/logo_lgai.jpg" height=60>
</p>
This is an official implementation of "OmniSplat: Taming Feed-Forward 3D Gaussian Splatting for Omnidirectional Images with Editable Capabilities."
<p align="center"> <img src="figures/framework.png" height=270> </p>Update Log
25.06.06: First code upload
Installation
git clone https://github.com/esw0116/OmniSplat.git --recursive
cd OminSplat
# Set Environment
conda env create --file environment.yaml
conda activate omnisplat
pip install submodules/simple-knn
pip install submodules/diff-gaussian-yin-rasterization
pip install submodules/diff-gaussian-yang-rasterization
Benchmark Dataset
We evaluate 6 datasets by adjusting their resolutions and performing Structure-from-Motion using OpenMVG.
For your convenience, we provide :star:links to the converted datasets:star: used in our paper.
The reference and target indices for each dataset is described in the supplementary material of the paper.
For reference, we provide the links to the original datasets.
OmniBlender & Ricoh360 / OmniPhotos / 360Roam / OmniScenes / 360VO
Running OmniSplat
- OmniSplat runs based on MVSplat, without fine-tuning any parameters.
Preparation
- Get the pretrained model (re10k.ckpt) from MVSplat repo, and save the model in
./checkpointsfolder - Put the downloaded datasets in the
./datasetsfolder
Evaluation Scripts
python -m src.main +experiment=[dataset_name]
- The config files are listed in
./config/experiment - The results will be saved in
./outputs/test
Note
- There will be a pixel misalignment during the omnidirectional image warping.
- To solve the issue, please go to the
equi2equi/torch.pyinpyequiliblibrary, and comment the two lines (L33-34)ui += 0.5; uj += 0.5 - We will modify the code to resolve the issue without changing the function in the library.
