ArXiv TLDR

Hero-Mamba: Mamba-based Dual Domain Learning for Underwater Image Enhancement

🐦 Tweet
2604.16266

Tejeswar Pokuri, Shivarth Rai

cs.CV

TLDR

Hero-Mamba is a novel Mamba-based network that efficiently enhances underwater images by processing spatial and spectral domains.

Key contributions

  • Proposes Hero-Mamba, a Mamba-based network for efficient underwater image enhancement.
  • Utilizes dual-domain learning (spatial RGB & spectral FFT) to decouple degradation factors.
  • Employs Mamba-based SS2D blocks for global dependencies with linear computational complexity.
  • Introduces a ColorFusion block guided by a background light prior for accurate color restoration.

Why it matters

Underwater image enhancement is crucial but challenging due to severe degradation and limitations of existing methods like CNNs and Transformers. Hero-Mamba offers an efficient solution by leveraging Mamba's linear complexity and a novel dual-domain approach. This significantly advances the field, enabling clearer underwater vision for various applications.

Original Abstract

Underwater images often suffer from severe degradation, such as color distortion, low contrast, and blurred details, due to light absorption and scattering in water. While learning-based methods like CNNs and Transformers have shown promise, they face critical limitations: CNNs struggle to model the long-range dependencies needed for non-uniform degradation, and Transformers incur quadratic computational complexity, making them inefficient for high-resolution images. To address these challenges, we propose Hero-Mamba, a novel Mamba-based network that achieves efficient dual-domain learning for underwater image enhancement. Our approach uniquely processes information from both the spatial domain (RGB image) and the spectral domain (FFT components) in parallel. This dual-domain input allows the network to decouple degradation factors, separating color/brightness information from texture/noise. The core of our network utilizes Mamba-based SS2D blocks to capture global receptive fields and long-range dependencies with linear complexity, overcoming the limitations of both CNNs and Transformers. Furthermore, we introduce a ColorFusion block, guided by a background light prior, to restore color information with high fidelity. Extensive experiments on the LSUI and UIEB benchmark datasets demonstrate that Hero-Mamba outperforms state-of-the-art methods. Notably, our model achieves a PSNR of 25.802 and an SSIM of 0.913 on LSUI, validating its superior performance and generalization capabilities.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.