Remote Sensing Image Super-Resolution for Imbalanced Textures: A Texture-Aware Diffusion Framework
Enzhuo Zhang, Sijie Zhao, Dilxat Muhtar, Zhenshi Li, Xueliang Zhang + 1 more
TLDR
TexADiff is a diffusion framework that improves remote sensing image super-resolution by addressing imbalanced textures with a texture-aware approach.
Key contributions
- Proposes TexADiff, a novel diffusion framework for remote sensing image super-resolution.
- Estimates a Relative Texture Density Map (RTDM) to explicitly represent texture distribution.
- Leverages RTDM for spatial conditioning, loss modulation, and dynamic sampling schedule.
- Generates faithful high-frequency details, suppresses hallucinations, and boosts downstream tasks.
Why it matters
Remote sensing images have unique texture challenges that hinder current super-resolution models. TexADiff provides a crucial advancement by explicitly addressing these imbalanced textures, leading to more accurate and reliable high-resolution images. This improves downstream applications reliant on precise remote sensing data.
Original Abstract
Generative diffusion priors have recently achieved state-of-the-art performance in natural image super-resolution, demonstrating a powerful capability to synthesize photorealistic details. However, their direct application to remote sensing image super-resolution (RSISR) reveals significant shortcomings. Unlike natural images, remote sensing images exhibit a unique texture distribution where ground objects are globally stochastic yet locally clustered, leading to highly imbalanced textures. This imbalance severely hinders the model's spatial perception. To address this, we propose TexADiff, a novel framework that begins by estimating a Relative Texture Density Map (RTDM) to represent the texture distribution. TexADiff then leverages this RTDM in three synergistic ways: as an explicit spatial conditioning to guide the diffusion process, as a loss modulation term to prioritize texture-rich regions, and as a dynamic adapter for the sampling schedule. These modifications are designed to endow the model with explicit texture-aware capabilities. Experiments demonstrate that TexADiff achieves superior or competitive quantitative metrics. Furthermore, qualitative results show that our model generates faithful high-frequency details while effectively suppressing texture hallucinations. This improved reconstruction quality also results in significant gains in downstream task performance. The source code of our method can be found at https://github.com/ZezFuture/TexAdiff.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.