Residential College | false |
Status | 已發表Published |
Hyperspectral image classification via discriminative convolutional neural network with an improved triplet loss | |
Huang, Ke Kun1,2; Ren, Chuan Xian3; Liu, Hui1; Lai, Zhao Rong4; Yu, Yu Feng5; Dai, Dao Qing3 | |
2021-04-01 | |
Source Publication | Pattern Recognition |
ISSN | 0031-3203 |
Volume | 112Pages:107744 |
Abstract | Hyper-Spectral Image (HSI) classification is an important task because of its wide range of applications. With the remarkable success from the Convolutional Neural Network (CNN), the performance of HSI classification has been significantly improved. However, two main challenges remained. One is that the samples of HSI have dramatic intra-class diversity and inter-class similarity, and the conventional cross-entropy loss is not good enough to learn discriminative features. The other is that the number of the training samples is so limited that the network is easy to overfit. To address the first challenge, we develop an improved triplet loss in order to make samples from the same class close to each other and make samples from different classes further apart. The proposed loss function considers all the possible positive pairs and negative pairs in a training batch, filters many trivial pairs, and prevents the impact of the outliers at the same time. To deal with the second challenge, we design an appropriate network architecture with less learnable parameters. We train the designed network based on the proposed loss with randomly initialized network weights using only hundreds of training samples, and attain quite good results. The experimental results show that the proposed method significantly surpasses other state-of-the-art methods, especially with less training samples. Furthermore, being less complex, the training process only takes a few minutes on a single GPU, which is faster than other state-of-the-art CNN-based methods. |
Keyword | Convolutional Neural Network Discriminative Learning Hyper-spectral Image Classification Metric Learning Triplet Loss |
DOI | 10.1016/j.patcog.2020.107744 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Computer Science ; Engineering |
WOS Subject | Computer Science, Artificial Intelligence ; Engineering, Electrical & Electronic |
WOS ID | WOS:000615938100005 |
Publisher | ELSEVIER SCI LTD, THE BOULEVARD, LANGFORD LANE, KIDLINGTON, OXFORD OX5 1GB, OXON, ENGLAND |
Scopus ID | 2-s2.0-85096445310 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE Faculty of Science and Technology |
Corresponding Author | Dai, Dao Qing |
Affiliation | 1.School of Mathematics, JiaYing University, Meizhou, 514015, China 2.Guangdong Provincial Key Laboratory of Conservation and Precision Utilization of Characteristic Agricultural Resources in Mountainous Areas, JiaYing University, Meizhou, 514015, China 3.Intelligent Data Center and Department of Mathematics, Sun Yat-Sen University, Guangzhou, 510275, China 4.Department of Mathematics, JiNan University, Guangzhou, 510632, China 5.Department of Computer and Information Science, University of Macau, Macau, 999078, China |
Recommended Citation GB/T 7714 | Huang, Ke Kun,Ren, Chuan Xian,Liu, Hui,et al. Hyperspectral image classification via discriminative convolutional neural network with an improved triplet loss[J]. Pattern Recognition, 2021, 112, 107744. |
APA | Huang, Ke Kun., Ren, Chuan Xian., Liu, Hui., Lai, Zhao Rong., Yu, Yu Feng., & Dai, Dao Qing (2021). Hyperspectral image classification via discriminative convolutional neural network with an improved triplet loss. Pattern Recognition, 112, 107744. |
MLA | Huang, Ke Kun,et al."Hyperspectral image classification via discriminative convolutional neural network with an improved triplet loss".Pattern Recognition 112(2021):107744. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment