LSTMConvSR: Joint Long–Short-Range Modeling via LSTM-First–CNN-Next Architecture for Remote Sensing Image Super-Resolution
Please use BasicSR, it is an open-source image super-resolution toolbox based on PyTorch. If you encounter any issues during use, please refer to MambaIRv2 — it is an excellent work and may provide helpful guidance.
If you find this work useful, please consider citing:
@Article{LSTMConvSR,
AUTHOR = {Zhu, Qiwei and Zhang, Guojing and Wang, Xiaoying and Huang, Jianqiang},
TITLE = {LSTMConvSR: Joint Long–Short-Range Modeling via LSTM-First–CNN-Next Architecture for Remote Sensing Image Super-Resolution},
JOURNAL = {Remote Sensing},
VOLUME = {17},
YEAR = {2025},
NUMBER = {15},
ARTICLE-NUMBER = {2745},
URL = {https://www.mdpi.com/2072-4292/17/15/2745},
ISSN = {2072-4292},
ABSTRACT = {The inability of existing super-resolution methods to jointly model short-range and long-range spatial dependencies in remote sensing imagery limits reconstruction efficacy. To address this, we propose LSTMConvSR, a novel framework inspired by top-down neural attention mechanisms. Our approach pioneers an LSTM-first–CNN-next architecture. First, an LSTM-based global modeling stage efficiently captures long-range dependencies via downsampling and spatial attention, achieving 80.3% lower FLOPs and 11× faster speed. Second, a CNN-based local refinement stage, guided by the LSTM’s attention maps, enhances details in critical regions. Third, a top-down fusion stage dynamically integrates global context and local features to generate the output. Extensive experiments on Potsdam, UAVid, and RSSCN7 benchmarks demonstrate state-of-the-art performance, achieving 33.94 dB PSNR on Potsdam with 2.4× faster inference than MambaIRv2.},
DOI = {10.3390/rs17152745}
}Some of the codes in this repo are borrowed from:
Thanks to their great work.