SkillAgentSearch skills...

BiFuse

[CVPR2020] BiFuse: Monocular 360 Depth Estimation via Bi-Projection Fusion

Install / Use

/learn @yuhsuanyeh/BiFuse

README

[CVPR2020] BiFuse: Monocular 360 Depth Estimation via Bi-Projection Fusion

<p align='center'> <img src='src/1690-teaser.gif'> </p>

[Paper] [Project Page]

Getting Started

Requirements

  • Python (tested on 3.7.4)
  • PyTorch (tested on 1.4.0)
  • Other dependencies
pip install -r requirements.txt

Usage

First clone our repo:

git clone https://github.com/Yeh-yu-hsuan/BiFuse.git
cd BiFuse

Step1

Download our pretrained Model and create a save folder:

mkdir save

then put the BiFuse_Pretrained.pkl into save folder.

Step2

My_Test_Data folder has contained a Sample.jpg RGB image as an example. <br> If you want to test your own data, please put your own rgb images into My_Test_Data folder and run:

python main.py --path './My_Test_Data'

Our argument: <br> --path is the folder path of your own testing images. <br> --nocrop if you don't want to crop the original images. <br>

After testing, you can see the results in My_Test_Result folder! <br>

  • Here shows some sample results
<p float="left"> <img src="src/007.jpg" width="30%" align="middle"/> <img src="src/146.jpg" width="30%" align="middle"/> <img src="src/147.jpg" width="30%" align="middle"/> </p> <p float="left"> <img src="src/200.jpg" width="30%" align="middle"/> <img src="src/232.jpg" width="30%" align="middle"/> <img src="src/246.jpg" width="30%" align="middle"/> </p> <p float="left"> <img src="src/260.jpg" width="30%" align="middle" /> <img src="src/272.jpg" width="30%" align="middle" /> <img src="src/236.jpg" width="30%" align="middle" /> </p>

The Restuls contain Combine.jpg, Depth.jpg, and Data.npy. <br> Combine.jpg is concatenating rgb image with its corresponding depth map prediction. <br> Depth.jpg is only depth map prediction. <br> Data.npy is the original data of both RGB and predicted depth value. <br>

Point Cloud Visualization

If you also want to visualize the point cloud of predicted depth, we also provide the script to render it. You can have a look at tools/.

License

This work is licensed under MIT License. See LICENSE for details.

If you find our code/models useful, please consider citing our paper:

@InProceedings{BiFuse20,
author = {Wang, Fu-En and Yeh, Yu-Hsuan and Sun, Min and Chiu, Wei-Chen and Tsai, Yi-Hsuan},
title = {BiFuse: Monocular 360 Depth Estimation via Bi-Projection Fusion},
booktitle = {The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2020}
}

Related Skills

View on GitHub
GitHub Stars181
CategoryEducation
Updated2mo ago
Forks31

Languages

Python

Security Score

100/100

Audited on Jan 19, 2026

No findings