Residential College | false |
Status | 已發表Published |
Multi-threshold deep metric learning for facial expression recognition | |
Yang, Wenwu1; Yu, Jinyi1; Chen, Tuo1; Liu, Zhenguang2; Wang, Xun1; Shen, Jianbing3 | |
2024-12-01 | |
Source Publication | Pattern Recognition |
ISSN | 0031-3203 |
Volume | 156Pages:110711 |
Abstract | Feature representations generated through triplet-based deep metric learning offer significant advantages for facial expression recognition (FER). Each threshold in triplet loss inherently shapes a distinct distribution of inter-class variations, leading to unique representations of expression features. Nonetheless, pinpointing the optimal threshold for triplet loss presents a formidable challenge, as the ideal threshold varies not only across different datasets but also among classes within the same dataset. In this paper, we propose a novel multi-threshold deep metric learning approach that bypasses the complex process of threshold validation and markedly improves the effectiveness in creating expression feature representations. Instead of choosing a single optimal threshold from a valid range, we comprehensively sample thresholds throughout this range, which ensures that the representation characteristics exhibited by the thresholds within this spectrum are fully captured and utilized for enhancing FER. Specifically, we segment the embedding layer of the deep metric learning network into multiple slices, with each slice representing a specific threshold sample. We subsequently train these embedding slices in an end-to-end fashion, applying triplet loss at its associated threshold to each slice, which results in a collection of unique expression features corresponding to each embedding slice. Moreover, we identify the issue that the traditional triplet loss may struggle to converge when employing the widely-used Batch Hard strategy for mining informative triplets, and introduce a novel loss termed dual triplet loss to address it. Extensive evaluations demonstrate the superior performance of the proposed approach on both posed and spontaneous facial expression datasets. |
Keyword | Facial Expression Recognition Multiple Thresholds Triplet Loss Learning |
DOI | 10.1016/j.patcog.2024.110711 |
URL | View the original |
Language | 英語English |
Publisher | Elsevier Ltd |
Scopus ID | 2-s2.0-85197750107 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU) |
Affiliation | 1.Zhejiang Gongshang University, Hangzhou, 310018, China 2.Zhejiang University, Hangzhou, 310012, China 3.University of Macau, 999078, Macao |
Recommended Citation GB/T 7714 | Yang, Wenwu,Yu, Jinyi,Chen, Tuo,et al. Multi-threshold deep metric learning for facial expression recognition[J]. Pattern Recognition, 2024, 156, 110711. |
APA | Yang, Wenwu., Yu, Jinyi., Chen, Tuo., Liu, Zhenguang., Wang, Xun., & Shen, Jianbing (2024). Multi-threshold deep metric learning for facial expression recognition. Pattern Recognition, 156, 110711. |
MLA | Yang, Wenwu,et al."Multi-threshold deep metric learning for facial expression recognition".Pattern Recognition 156(2024):110711. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment