Residential College | false |
Status | 已發表Published |
QFormer: An Efficient Quaternion Transformer for Image Denoising | |
Jiang, Bo1; Lu, Yao2; Lu, Guangming2; Zhang, Bob3 | |
2024 | |
Conference Name | 33rd International Joint Conference on Artificial Intelligence, IJCAI 2024 |
Source Publication | Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) |
Pages | 4237-4245 |
Conference Date | 3-9 August 2024 |
Conference Place | Jeju, South Korea |
Country | Republic of Korea |
Publisher | International Joint Conferences on Artificial Intelligence |
Abstract | Since Deep Convolutional Neural Networks (DCNNs) and Vision Transformer perform well in learning generalizable image priors from large-scale data, these models have been widely used in image denoising tasks. However, vanilla DCNNs and Transformer suffer from two problems. First, the vanilla DCNNs and Transformer only accumulate the output along the channel axis, ignoring the internal relationship among channels. This results in the severely inadequate color structure representation retrieved from color images. Secondly, the DCNNs or Transformer-based image denoising models usually have a large number of parameters, high computational complexity, and slow inference speed. To resolve these issues, this paper proposes a highly-efficient Quaternion Transformer (QFormer) for image denoising. Specifically, the proposed Quaternion Transformer Block (QTB) simplifies the typical Transformer from a multi-branch structure to an elaborately sequential structure mainly with quaternion transformations, to alternately capture both long-range dependencies and local contextual features with color structure information. Furthermore, the proposed QTB can also avoid considerable element-wise multiplications of computing the self-attention matrices. Thus, our QTB can significantly reduce the computational complexity and its sequential structure can further improve the practical inference speed. Comprehensive experiments demonstrate that the proposed QFormer produces state-of-the-art results in both denoising performance and efficiency. We hope that our work will encourage further research to explore the Quaternion Transformer architecture for image denoising tasks. |
Keyword | Machine Learning Ml Knowledge-aided Learning |
DOI | 10.24963/ijcai.2024/468 |
URL | View the original |
Language | 英語English |
Scopus ID | 2-s2.0-85204305282 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | Faculty of Science and Technology DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Corresponding Author | Lu, Yao; Lu, Guangming |
Affiliation | 1.College of Mechanical and Electronic Engineering, Northwest A&F University, China 2.Department of Computer Science, Harbin Institute of Technology, Shenzhen, China 3.Department of Computer and Information Science, University of Macau, Macao |
Recommended Citation GB/T 7714 | Jiang, Bo,Lu, Yao,Lu, Guangming,et al. QFormer: An Efficient Quaternion Transformer for Image Denoising[C]:International Joint Conferences on Artificial Intelligence, 2024, 4237-4245. |
APA | Jiang, Bo., Lu, Yao., Lu, Guangming., & Zhang, Bob (2024). QFormer: An Efficient Quaternion Transformer for Image Denoising. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24), 4237-4245. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment