SkillAgentSearch skills...

UCAN

[CVPR 2026] Unified Convolutional Attention Network for Expansive Receptive Fields in Lightweight Super-Resolution

Install / Use

/learn @hokiyoshi/UCAN
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

<p align="center"> <h1 align="center"> <img width="200" src="logo.png"> <br> UCAN: Unified Convolutional Attention Network <br> for Expansive Receptive Fields in Lightweight Super-Resolution</h1> <p align="center"> Cao Thien Tan<sup>1,2,3</sup> · Phan Thi Thu Trang <sup>3,5</sup> · Do Nghiem Duc</a><sup>6</sup> · Ho Ngoc Anh</a><sup>5</sup> · Hanyang Zhuang</a><sup>4</sup> . <br> Nguyen Duc Dung</a><sup>2</sup> </p> <p align="center"> <sup>1</sup>Ho Chi Minh City Open University &nbsp; &nbsp; <sup>2</sup>AI Tech Lab, Ho Chi Minh City University of Technology <br> <sup>3</sup>Code Mely AI Research Team &nbsp; &nbsp; <sup>4</sup>Global College, Shanghai Jiao Tong University <br> <sup>5</sup>Ha Noi University of Science and Technology &nbsp; &nbsp; <sup>6</sup>University of Manitoba </p> <h3 align="center"> CVPR 2026 - Poster </h3> <h3 align="center"><a href="https://arxiv.org/abs/2603.11680">[Paper]</a></h3> </p>

Abstract: Hybrid CNN-Transformer architectures achieve strong results in image super-resolution, but scaling attention windows or convolution kernels significantly increases computational cost, limiting deployment on resource-constrained devices. We present UCAN, a lightweight network that unifies convolution and attention to expand the effective receptive field efficiently. UCAN combines window-based spatial attention with a Hedgehog Attention mechanism to model both local texture and long-range dependencies, and introduces a distillation-based large-kernel module to preserve high-frequency structure without heavy computation. In addition, we employ cross-layer parameter sharing to further reduce complexity. On Manga109 (4×), UCAN-L achieves 31.63 dB PSNR with only 48.4G MACs, surpassing recent lightweight models. On BSDS100, UCAN attains 27.79 dB, outperforming methods with significantly larger models. Extensive experiments show that UCAN achieves a superior trade-off between accuracy, efficiency, and scalability, making it well-suited for practical high-resolution image restoration.

Paper will be updated soon.

🔥 News

  • 2026-02: 🎉UCAN is accepted by CVPR 2026! This repo is released.

🛠️ Setup

🗃️ Datasets

The datasets used in our training and testing are orgnized as follows:

Training and testing sets can be downloaded as follows:

| Training Set | Testing Set | Visual Results | | :-----------------------------------------------------------: | :----------------------------------------------------------: | :----------------------------------------------------------: | | DIV2K (800 training images, 100 validation images) [organized training dataset DIV2K: One Drive] | Set5 + Set14 + BSD100 + Urban100 + Manga109 [complete testing dataset: One Drive] | Updating |

💡 Models

🚀 Training

🧪 Testing

📚 Results

🙇 Citation

🥂 Acknowledgements

This work is based on BasicSR, HiT-SR, ESC and MambaIRv2. We thank them for their great work and for sharing the code.

Related Skills

View on GitHub
GitHub Stars18
CategoryDevelopment
Updated1d ago
Forks0

Security Score

75/100

Audited on Apr 7, 2026

No findings