TY - GEN
T1 - Efficient Lightweight Image Denoising with Triple Attention Transformer
AU - Zhou, Yubo
AU - Lin, Jin
AU - Ye, Fangchen
AU - Qu, Yanyun
AU - Xie, Yuan
N1 - Publisher Copyright:
Copyright © 2024, Association for the Advancement of Artificial Intelligence (www.aaai.org).All rights reserved.
PY - 2024/3/25
Y1 - 2024/3/25
N2 - Transformer has shown outstanding performance on image denoising, but the existing Transformer methods for image denoising are with large model sizes and high computational complexity, which is unfriendly to resource-constrained devices.In this paper, we propose a Lightweight Image Denoising Transformer method (LIDFormer) based on Triple Multi-Dconv Head Transposed Attention (TMDTA) to boost computational efficiency.LIDFormer first implements Discrete Wavelet Transform (DWT), which transforms the input image into a low-frequency space, greatly reducing the computational complexity of image denoising.However, the low-frequency image lacks fine-feature information, which degrades the denoising performance.To handle this problem, we introduce the Complementary Periodic Feature Reusing (CPFR) scheme for aggregating the shallow-layer features and the deep-layer features.Furthermore, TMDTA is proposed to integrate global context along three dimensions, thereby enhancing the ability of global feature representation.Note that our method can be applied as a pipeline for both convolutional neural networks and Transformers.Extensive experiments on several benchmarks demonstrate that the proposed LIDFormer achieves a better trade-off between high performance and low computational complexity on real-world image denoising tasks.
AB - Transformer has shown outstanding performance on image denoising, but the existing Transformer methods for image denoising are with large model sizes and high computational complexity, which is unfriendly to resource-constrained devices.In this paper, we propose a Lightweight Image Denoising Transformer method (LIDFormer) based on Triple Multi-Dconv Head Transposed Attention (TMDTA) to boost computational efficiency.LIDFormer first implements Discrete Wavelet Transform (DWT), which transforms the input image into a low-frequency space, greatly reducing the computational complexity of image denoising.However, the low-frequency image lacks fine-feature information, which degrades the denoising performance.To handle this problem, we introduce the Complementary Periodic Feature Reusing (CPFR) scheme for aggregating the shallow-layer features and the deep-layer features.Furthermore, TMDTA is proposed to integrate global context along three dimensions, thereby enhancing the ability of global feature representation.Note that our method can be applied as a pipeline for both convolutional neural networks and Transformers.Extensive experiments on several benchmarks demonstrate that the proposed LIDFormer achieves a better trade-off between high performance and low computational complexity on real-world image denoising tasks.
UR - https://www.scopus.com/pages/publications/85189524702
U2 - 10.1609/aaai.v38i7.28604
DO - 10.1609/aaai.v38i7.28604
M3 - 会议稿件
AN - SCOPUS:85189524702
T3 - Proceedings of the AAAI Conference on Artificial Intelligence
SP - 7704
EP - 7712
BT - Technical Tracks 14
A2 - Wooldridge, Michael
A2 - Dy, Jennifer
A2 - Natarajan, Sriraam
PB - Association for the Advancement of Artificial Intelligence
T2 - 38th AAAI Conference on Artificial Intelligence, AAAI 2024
Y2 - 20 February 2024 through 27 February 2024
ER -