UM  > Faculty of Science and Technology
Residential Collegefalse
Status已發表Published
QFormer: An Efficient Quaternion Transformer for Image Denoising
Jiang, Bo1; Lu, Yao2; Lu, Guangming2; Zhang, Bob3
2024
Conference Name33rd International Joint Conference on Artificial Intelligence, IJCAI 2024
Source PublicationProceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24)
Pages4237-4245
Conference Date3-9 August 2024
Conference PlaceJeju, South Korea
CountryRepublic of Korea
PublisherInternational Joint Conferences on Artificial Intelligence
Abstract

Since Deep Convolutional Neural Networks (DCNNs) and Vision Transformer perform well in learning generalizable image priors from large-scale data, these models have been widely used in image denoising tasks. However, vanilla DCNNs and Transformer suffer from two problems. First, the vanilla DCNNs and Transformer only accumulate the output along the channel axis, ignoring the internal relationship among channels. This results in the severely inadequate color structure representation retrieved from color images. Secondly, the DCNNs or Transformer-based image denoising models usually have a large number of parameters, high computational complexity, and slow inference speed. To resolve these issues, this paper proposes a highly-efficient Quaternion Transformer (QFormer) for image denoising. Specifically, the proposed Quaternion Transformer Block (QTB) simplifies the typical Transformer from a multi-branch structure to an elaborately sequential structure mainly with quaternion transformations, to alternately capture both long-range dependencies and local contextual features with color structure information. Furthermore, the proposed QTB can also avoid considerable element-wise multiplications of computing the self-attention matrices. Thus, our QTB can significantly reduce the computational complexity and its sequential structure can further improve the practical inference speed. Comprehensive experiments demonstrate that the proposed QFormer produces state-of-the-art results in both denoising performance and efficiency. We hope that our work will encourage further research to explore the Quaternion Transformer architecture for image denoising tasks.

KeywordMachine Learning Ml Knowledge-aided Learning
DOI10.24963/ijcai.2024/468
URLView the original
Language英語English
Scopus ID2-s2.0-85204305282
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionFaculty of Science and Technology
DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
Corresponding AuthorLu, Yao; Lu, Guangming
Affiliation1.College of Mechanical and Electronic Engineering, Northwest A&F University, China
2.Department of Computer Science, Harbin Institute of Technology, Shenzhen, China
3.Department of Computer and Information Science, University of Macau, Macao
Recommended Citation
GB/T 7714
Jiang, Bo,Lu, Yao,Lu, Guangming,et al. QFormer: An Efficient Quaternion Transformer for Image Denoising[C]:International Joint Conferences on Artificial Intelligence, 2024, 4237-4245.
APA Jiang, Bo., Lu, Yao., Lu, Guangming., & Zhang, Bob (2024). QFormer: An Efficient Quaternion Transformer for Image Denoising. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24), 4237-4245.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Jiang, Bo]'s Articles
[Lu, Yao]'s Articles
[Lu, Guangming]'s Articles
Baidu academic
Similar articles in Baidu academic
[Jiang, Bo]'s Articles
[Lu, Yao]'s Articles
[Lu, Guangming]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Jiang, Bo]'s Articles
[Lu, Yao]'s Articles
[Lu, Guangming]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.