Bridging the Training-Deployment Gap: Gated Encoding and Multi-Scale Refinement for Efficient Quantization-Aware Image Enhancement
Dat To-Thanh, Nghia Nguyen-Trong, Hoang Vo, Hieu Bui-Minh, Tinh-Anh Nguyen-Nhu
TLDR
An efficient mobile image enhancement model uses gated encoding, multi-scale refinement, and QAT to achieve high quality with low computational overhead.
Key contributions
- Proposes an efficient image enhancement model specifically for mobile deployment.
- Utilizes a hierarchical network with gated encoder blocks and multi-scale refinement for feature preservation.
- Incorporates Quantization-Aware Training (QAT) to prevent quality degradation from low-precision formats.
- Achieves high-fidelity visual output while maintaining low computational overhead on standard mobile devices.
Why it matters
Mobile image enhancement often struggles with quality loss during quantization for deployment. This paper introduces a QAT-enabled model with a hierarchical network to bridge this gap. It delivers high-fidelity output with low overhead, making advanced mobile image enhancement practical.
Original Abstract
Image enhancement models for mobile devices often struggle to balance high output quality with the fast processing speeds required by mobile hardware. While recent deep learning models can enhance low-quality mobile photos into high-quality images, their performance is often degraded when converted to lower-precision formats for actual use on mobile phones. To address this training-deployment mismatch, we propose an efficient image enhancement model designed specifically for mobile deployment. Our approach uses a hierarchical network architecture with gated encoder blocks and multiscale refinement to preserve fine-grained visual features. Moreover, we incorporate Quantization-Aware Training (QAT) to simulate the effects of low-precision representation during the training process. This allows the network to adapt and prevents the typical drop in quality seen with standard post-training quantization (PTQ). Experimental results demonstrate that the proposed method produces high-fidelity visual output while maintaining the low computational overhead needed for practical use on standard mobile devices. The code will be available at https://github.com/GenAI4E/QATIE.git.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.