z-logo
open-access-imgOpen Access
Efficient Multi-Scale Transformer with Convolutional Attention for High-Definition Image Dehazing
Author(s) -
Zhirui Wang
Publication year - 2025
Publication title -
ieee access
Language(s) - English
Resource type - Magazines
SCImago Journal Rank - 0.587
H-Index - 127
eISSN - 2169-3536
DOI - 10.1109/access.2025.3588390
Subject(s) - aerospace , bioengineering , communication, networking and broadcast technologies , components, circuits, devices and systems , computing and processing , engineered materials, dielectrics and plasmas , engineering profession , fields, waves and electromagnetics , general topics for engineers , geoscience , nuclear engineering , photonics and electrooptics , power, energy and industry applications , robotics and control systems , signal processing and analysis , transportation
Image dehazing aims to recover clear images from hazy inputs, crucial for advanced computer vision tasks. While convolutional neural networks (CNNs) have traditionally been used, transformer-based methods have recently demonstrated superior performance in capturing global information. However, their computational complexity scales quadratically with image resolution, posing challenges for high-resolution tasks. To address these limitations, we propose Mulsormer, a novel transformer-based dehazing model tailored for high-definition images. Mulsormer incorporates Multi-Head Convolutional Attention (MHCA) and a Gated Feed-Forward Network (GFFN) to reduce computational complexity and control critical information flow, respectively. A multi-scale loss function combining perceptual and patch-based losses is utilized to improve detail restoration. Extensive experiments on the SOTS and RTTS datasets show that Mulsormer achieves state-of-the-art performance, with a PSNR of 37.49 dB and an SSIM of 0.993 on the SOTS Indoor dataset, significantly outperforming previous methods. Our approach demonstrates the potential of transformer-based models for efficient high-definition image dehazing. Additionally, our code is publicly available at https://github.com/Wang1666570/Efficient-Multi-Scale-Transformer.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom