ArXiv TLDR

A unified Benchmark for Multi-Frame Image Restoration under Severe Refractive Warping

🐦 Tweet
2605.05079

Maxim V. Shugaev, Md Reshad Ul Hoque, Bridget Kennedy, Joseph T. Riley, Fiona Hwang + 6 more

cs.CV

TLDR

A new unified benchmark evaluates multi-frame image restoration methods for severe refractive warping, covering mild to extreme distortions.

Key contributions

  • Introduces a comprehensive benchmark for geometric distortion removal in videos under severe refractive warping.
  • Includes both laboratory-captured real data and physics-based synthetic sequences across four distortion levels.
  • Evaluates diverse methods, from classical to learning-based (including V-cache), using pixel and perceptual metrics.

Why it matters

Existing benchmarks lack systematic evaluation under strong, nonuniform refractive conditions. This paper provides a crucial foundation for developing and assessing algorithms capable of reconstructing video from highly distorted optical environments, pushing the boundaries of image restoration.

Original Abstract

Video sequence capturing through refractive dynamic media, such as a turbulent air or water surface, often suffer from severe geometric distortions and temporal instability. While recent advances address mild atmospheric turbulence, no existing benchmarks systematically evaluate restoration methods under strong and highly nonuniform refractive conditions. We present a comprehensive benchmark for geometric distortion removal in video, covering a range from turbulence-like mild warping to strong discontinuous refractive deformations. The benchmark includes both laboratory-captured real data and synthetic sequences generated for static scenes via physics-based light refraction modeling across four distortion levels and multiple surface wave types. We evaluate a spectrum of methods from simple baselines and classical registration algorithms to advanced learning-based approaches including DATUM and our proposed diffusion based V-cache for high and extreme distortions regimes. Evaluation uses both pixel-level (PSNR, SSIM), and perceptual (LPIPS, DINO, CLIP) metrics providing the first large scale analysis of geometric distortion removal. Our benchmark establishes a new foundation for developing and evaluating algorithms capable of reconstructing video from highly distorted optical environments. Our code and datasets are available at https://github.com/iafoss/refractive-mfir-benchmark.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.