SkillAgentSearch skills...

LGTD

[IEEE TCSVT 2023] Local-Global Temporal Difference Learning for Satellite Video Super-Resolution

Install / Use

/learn @XY-boy/LGTD
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

LGTD (IEEE TCSVT 2023)

📖Paper | 🖼️PDF | 🎁Dataset

PyTorch codes for "Local-Global Temporal Difference Learning for Satellite Video Super-Resolution", IEEE Transactions on Circuits and Systems for Video Technology (TCSVT), 2023.

Abstract

Optical-flow-based and kernel-based approaches have been extensively explored for temporal compensation in satellite Video Super-Resolution (VSR). However, these techniques are less generalized in large-scale or complex scenarios, especially in satellite videos. In this paper, we propose to exploit the well-defined temporal difference for efficient and effective temporal compensation. To fully utilize the local and global temporal information within frames, we systematically modeled the short-term and long-term temporal discrepancies since we observed that these discrepancies offer distinct and mutually complementary properties. Specifically, we devise a Short-term Temporal Difference Module (S-TDM) to extract local motion representations from RGB difference maps between adjacent frames, which yields more clues for accurate texture representation. To explore the global dependency in the entire frame sequence, a Long-term Temporal Difference Module (L-TDM) is proposed, where the differences between forward and backward segments are incorporated and activated to guide the modulation of the temporal feature, leading to a holistic global compensation. Moreover, we further propose a Difference Compensation Unit (DCU) to enrich the interaction between the spatial distribution of the target frame and temporal compensated results, which helps maintain spatial consistency while refining the features to avoid misalignment. Rigorous objective and subjective evaluations conducted across five mainstream video satellites demonstrate that our method performs favorably against state-of-the-art approaches.

Network

image

🧩Install

git clone https://github.com/XY-boy/LGTD.git

Environment

  • CUDA 11.1
  • PyTorch 1.9.1
  • build DCNv2

Dataset Preparation

Please download our dataset in

  • Baidu Netdisk Jilin-189 Code:31ct
  • Zenodo: <a href="https://doi.org/10.5281/zenodo.6969604"><img src="https://zenodo.org/badge/DOI/10.5281/zenodo.6969604.svg" alt="DOI"></a>

You can also train your dataset following the directory sturture below!

Data directory structure

trainset--
 | train--
  | LR4x---
   | 000.png
   | ···.png
   | 099.png
  | GT---
  | Bicubic4x---

testset--
 | eval--
  | LR4x---
   | 000.png
   | ···.png
   | 099.png
  | GT---
  | Bicubic4x---

Training

python main.py

Test

python eval.py

Qualitative results

image

Quantitative results

image

More details can be found in our paper!

Acknowledgement

The temporal difference modules are built upon the awesome video architecture "Temporal Difference Networks for Efficient Action Recognition". We thank this excellent work for inspiring our research. For more technical details and references, please refer to the official repository TDN.

Contact

If you have any questions or suggestions, feel free to contact me. 😊
Email: xiao_yi@whu.edu.cn; xy574475@gmail.com

Citation

If you find our work helpful in your research, please consider citing it. Thank you! 😊😊

@ARTICLE{xiao2023lgtd,
  author={Xiao, Yi and Yuan, Qiangqiang and Jiang, Kui and Jin, Xianyu and He, Jiang and Zhang, Liangpei and Lin, Chia-wen},
  journal={IEEE Transactions on Circuits and Systems for Video Technology}, 
  title={Local-Global Temporal Difference Learning for Satellite Video Super-Resolution}, 
  year={2023},
  volume={},
  number={},
  pages={1-14},
  doi={10.1109/TCSVT.2023.3312321}
}
View on GitHub
GitHub Stars45
CategoryContent
Updated4d ago
Forks1

Languages

Python

Security Score

75/100

Audited on Mar 19, 2026

No findings