SkillAgentSearch skills...

OmniSplat

[CVPR 2025] OmniSplat: Taming Feed-Forward 3D Gaussian Splatting for Omnidirectional Images with Editable Capabilities

Install / Use

/learn @esw0116/OmniSplat
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

<div align="center">

Project ArXiv

</div> <p align="center"> <h1 align="center">OmniSplat: Taming Feed-Forward 3D Gaussian Splatting for </br> Omnidirectional Images with Editable Capabilities </h1> <p align="center"> <a href="https://esw0116.github.io/">Suyoung Lee</a><sup>1*</sup> &nbsp;·&nbsp; <a href="https://robot0321.github.io/">Jaeyoung Chung</a><sup>1*</sup> &nbsp;·&nbsp; Kihoon Kim<sup>1</sup> &nbsp;·&nbsp; Jaeyoo Huh<sup>1</sup> </br> Gunhee Lee<sup>2</sup> &nbsp;·&nbsp; Minsoo Lee<sup>2</sup> &nbsp;·&nbsp; <a href="https://cv.snu.ac.kr/index.php/~kmlee/">Kyoung Mu Lee</a><sup>1</sup> </br> 1: Seoul National Univiersity &nbsp;&nbsp; 2: LG AI Research Center </br> (* denotes equal contribution) </p> <h3 align="center">CVPR 2025, Highlight</h3> </p> <!-- <div align="center"> [![ArXiv]()]() [![Github](https://img.shields.io/github/stars/luciddreamer-cvlab/LucidDreamer)](https://github.com/luciddreamer-cvlab/LucidDreamer) [![LICENSE](https://img.shields.io/badge/license-MIT-lightgrey)](https://github.com/luciddreamer-cvlab/LucidDreamer/blob/master/LICENSE) </div> --> <p align="center"> <img src="figures/logo_cvlab.png" height=60 style="margin-right:40px;"> <img src="figures/logo_lgai.jpg" height=60> </p>

This is an official implementation of "OmniSplat: Taming Feed-Forward 3D Gaussian Splatting for Omnidirectional Images with Editable Capabilities."

<p align="center"> <img src="figures/framework.png" height=270> </p>

Update Log

25.06.06: First code upload

Installation

git clone https://github.com/esw0116/OmniSplat.git --recursive
cd OminSplat

# Set Environment
conda env create --file environment.yaml
conda activate omnisplat
pip install submodules/simple-knn
pip install submodules/diff-gaussian-yin-rasterization
pip install submodules/diff-gaussian-yang-rasterization

Benchmark Dataset

We evaluate 6 datasets by adjusting their resolutions and performing Structure-from-Motion using OpenMVG.
For your convenience, we provide :star:links to the converted datasets:star: used in our paper. The reference and target indices for each dataset is described in the supplementary material of the paper.

For reference, we provide the links to the original datasets.
OmniBlender & Ricoh360 / OmniPhotos / 360Roam / OmniScenes / 360VO

Running OmniSplat

  • OmniSplat runs based on MVSplat, without fine-tuning any parameters.

Preparation

  • Get the pretrained model (re10k.ckpt) from MVSplat repo, and save the model in ./checkpoints folder
  • Put the downloaded datasets in the ./datasets folder

Evaluation Scripts

python -m src.main +experiment=[dataset_name]
  • The config files are listed in ./config/experiment
  • The results will be saved in ./outputs/test

Note

  • There will be a pixel misalignment during the omnidirectional image warping.
  • To solve the issue, please go to the equi2equi/torch.py in pyequilib library, and comment the two lines (L33-34) ui += 0.5; uj += 0.5
  • We will modify the code to resolve the issue without changing the function in the library.
<section class="section" id="BibTeX"> <div class="container is-max-desktop content"> <h2 class="title">Citation</h2> <pre><code>@InProceedings{Lee2025OmniSplat, author = {Lee, Suyoung and Chung, Jaeyoung and Kim, Kihoon and Huh, Jaeyoo and Lee, Gunhee and Lee, Minsoo and Lee, Kyoung Mu}, title = {OmniSplat: Taming Feed-Forward 3D Gaussian Splatting for Omnidirectional Images with Editable Capabilities}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {2025}, pages = {16356-16365} } }</code></pre> </div> </section>
View on GitHub
GitHub Stars36
CategoryDevelopment
Updated13h ago
Forks0

Languages

Python

Security Score

95/100

Audited on Mar 27, 2026

No findings