Residential Collegefalse
Status已發表Published
Spectrum-irrelevant fine-grained representation for visible–infrared person re-identification
Gong,Jiahao1; Zhao,Sanyuan1,3; Lam,Kin Man2; Gao,Xin4; Shen,Jianbing5
2023-04-21
Source PublicationComputer Vision and Image Understanding
ISSN1077-3142
Volume232Pages:103703
Abstract

Visible–infrared person re-identification (VI-ReID) is an important and practical task for full-time intelligent surveillance systems. Compared to visible person re-identification, it is more challenging due to the large cross-modal discrepancy. Existing VI-ReID methods suffer from heterogeneous structures and the different spectra of visible and infrared images. In this work, we propose the Spectrum-Insensitive Data Augmentation (SIDA) strategy, which effectively alleviates the disturbance in the visible and infrared spectra and forces the network to learn spectrum-irrelevant features. The network also compares samples with both global and local features. We devise a Feature Relation Reasoning (FRR) module to learn discriminative fine-grained representations according to the graph reasoning principle. Compared to the most commonly used uniform partition, our FRR better adopts to the case of VI-ReID, in which human bodies are difficult to align. Furthermore, we design the dual center loss for learning the global feature in order to maintain the intra-modality relations, while learning the cross-modal similarities. Our method achieves better convergence in training. Extensive experiments demonstrate that our method achieves state-of-the-art performance on two visible–infrared cross-modal Re-ID datasets.

KeywordVisible–infrared Person Re-identification
DOI10.1016/j.cviu.2023.103703
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaComputer Science ; Engineering
WOS SubjectComputer Science, Artificial Intelligence ; Engineering, Electrical & Electronic
WOS IDWOS:000988823400001
PublisherACADEMIC PRESS INC ELSEVIER SCIENCE, 525 B ST, STE 1900, SAN DIEGO, CA 92101-4495
Scopus ID2-s2.0-85153504159
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionDEPARTMENT OF COMPUTER AND INFORMATION SCIENCE
THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU)
Corresponding AuthorZhao,Sanyuan
Affiliation1.School of Computer Science,Beijing Institute of Technology,100081,China
2.The Department of Electronic and Information Engineering,The Hong Kong Polytechnic University,Kowloon,Hung Hom,Hong Kong
3.Yangtze Delta Region Academy of Beijing Institute of Technology,Jiaxing,China
4.King Abdullah University of Science and Technology,Thuwal,23955-6900,Saudi Arabia
5.the State Key Laboratory of Internet of Things for Smart City,Department of Computer and Information Science,University of Macau,Macau,China
Recommended Citation
GB/T 7714
Gong,Jiahao,Zhao,Sanyuan,Lam,Kin Man,et al. Spectrum-irrelevant fine-grained representation for visible–infrared person re-identification[J]. Computer Vision and Image Understanding, 2023, 232, 103703.
APA Gong,Jiahao., Zhao,Sanyuan., Lam,Kin Man., Gao,Xin., & Shen,Jianbing (2023). Spectrum-irrelevant fine-grained representation for visible–infrared person re-identification. Computer Vision and Image Understanding, 232, 103703.
MLA Gong,Jiahao,et al."Spectrum-irrelevant fine-grained representation for visible–infrared person re-identification".Computer Vision and Image Understanding 232(2023):103703.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Gong,Jiahao]'s Articles
[Zhao,Sanyuan]'s Articles
[Lam,Kin Man]'s Articles
Baidu academic
Similar articles in Baidu academic
[Gong,Jiahao]'s Articles
[Zhao,Sanyuan]'s Articles
[Lam,Kin Man]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Gong,Jiahao]'s Articles
[Zhao,Sanyuan]'s Articles
[Lam,Kin Man]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.