PASSRnet
[CVPR 2019] Learning Parallax Attention for Stereo Image Super-Resolution
Install / Use
/learn @SYSU-SAIL/PASSRnetREADME
PASSRnet: Parallax Attention Stereo Super-Resolution Network
Pytorch implementation of "Learning Parallax Attention for Stereo Image Super-Resolution", CVPR 2019
Overview

Figure 1. Overview of our PASSRnet network.
<img width="400" src="https://github.com/LongguangWang/PASSRnet/blob/master/Figs/Parallax-attention.png"/></div>
Figure 2. Illustration of our parallax-attention mechanism.
<img width="500" src="https://github.com/LongguangWang/PASSRnet/blob/master/Figs/Toy-example.png"/></div>
Figure 3. A toy example illustration of the parallax-attention and cycle-attention maps generated by our PAM. The attention maps (30×30) correspond to the regions (1×30) marked by a yellow stroke. In (a) and (b), the first row represents left/right stereo images, the second row stands for parallax-attention maps, and the last row represents cycle-attention maps.
Flickr1024 Dataset
<img width="500" src="https://github.com/LongguangWang/PASSRnet/blob/master/Figs/Flickr1024.jpg"/></div>
Figure 4. The Flickr1024 dataset.
Requirements
- pytorch (0.4), torchvision (0.2) (Note: The code is tested with
python=3.6, cuda=9.0) - Matlab (For training/test data generation)
Train
Prepare training data
- Download the Flickr1024 dataset and put the images in
data/train/Flickr1024(Note: In our paper, we also use 60 images in the Middlebury dataset as the training set.) - Cd to
data/trainand rungenerate_trainset.mto generate training data.
Begin to train
python train.py --scale_factor 4 --device cuda:0 --batch_size 32 --n_epochs 80 --n_steps 30
Test
Prepare test data
- Download the KITTI2012 dataset and put folders
testing/colored_0andtesting/colored_1indata/test/KITTI2012/original - Cd to
data/testand rungenerate_testset.mto generate test data. - (optional) You can also download KITTI2015, Middlebury or other stereo datasets and prepare test data in
data/testas below:
data
└── test
├── dataset_1
├── hr
├── scene_1
├── hr0.png
└── hr1.png
├── ...
└── scene_M
└── lr_x4
├── scene_1
├── lr0.png
└── lr1.png
├── ...
└── scene_M
├── ...
└── dataset_N
Demo
python demo_test.py --scale_factor 4 --device cuda:0 --dataset KITTI2012
Results

Figure 5. Visual comparison for 2× SR. These results are achieved on “test_image_013” of the KITTI 2012 dataset and “test_image_019” of the KITTI 2015 dataset.

Figure 6. Visual comparison for 4× SR. These results are achieved on “test_image_004” of the KITTI 2015 dataset.

Figure 7. Visual comparison for 2× SR. These results are achieved on a stereo image pair acquired in our laboratory.
Citation
@InProceedings{Wang2019Learning,
author = {Longguang Wang and Yingqian Wang and Zhengfa Liang and Zaiping Lin and Jungang Yang and Wei An and Yulan Guo},
title = {Learning Parallax Attention for Stereo Image Super-Resolution},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2019},
}
Contact
For questions, please send an email to wanglongguang15@nudt.edu.cn
Related Skills
proje
Interactive vocabulary learning platform with smart flashcards and spaced repetition for effective language acquisition.
YC-Killer
2.7kA library of enterprise-grade AI agents designed to democratize artificial intelligence and provide free, open-source alternatives to overvalued Y Combinator startups. If you are excited about democratizing AI access & AI agents, please star ⭐️ this repository and use the link in the readme to join our open source AI research team.
best-practices-researcher
The most comprehensive Claude Code skills registry | Web Search: https://skills-registry-web.vercel.app
research_rules
Research & Verification Rules Quote Verification Protocol Primary Task "Make sure that the quote is relevant to the chapter and so you we want to make sure that we want to have it identifie
