Residential College | false |
Status | 已發表Published |
COMO: Efficient Deep Neural Networks Expansion With COnvolutional MaxOut | |
Zhao,Baoxin1; Xiong,Haoyi2; Bian,Jiang3; Guo,Zhishan4; Xu,Cheng Zhong5; Dou,Dejing6 | |
2020-06-15 | |
Source Publication | IEEE Transactions on Multimedia |
ISSN | 1520-9210 |
Volume | 23Pages:1722-1730 |
Abstract | In this paper, we extend the classic MaxOut strategy, originally designed for Multiple Layer Preceptors (MLPs), into COnvolutional MaxOut (COMO) --- a new strategy making deep convolutional neural networks wider with parameter efficiency. Compared to the existing solutions, such as ResNeXt for ResNet or Inception for VGG-alikes, COMO works well on both linear architectures and the ones with skipped connections and residual blocks. More specifically, COMO adopts a novel split-transform-merge paradigm that extends the layers with spatial resolution reduction into multiple parallel splits. For the layer with COMO, each split passes the input feature maps through a 4D convolution operator with independent batch normalization operators for transformation, then merge into the aggregated output of the original sizes through max-pooling. Such a strategy is expected to tackle the potential classification accuracy degradation due to the spatial resolution reduction, by incorporating the multiple splits and max-pooling-based feature selection. Our experiment using a wide range of deep architectures shows that COMO can significantly improve the classification accuracy of ResNet/VGG-alike networks based on a large number of benchmark datasets. COMO further outperforms the existing solutions, e.g., Inceptions, ResNeXts, SE-ResNet, and Xception, that make networks wider, and it dominates in the comparison of accuracy versus parameter sizes. |
Keyword | Computer Architecture Convolution Convolutional Neural Networks Deep Learning Spatial Resolution Transforms |
DOI | 10.1109/TMM.2020.3002614 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Computer Science ; Telecommunications |
WOS Subject | Computer Science, Information Systems ; Computer Science, Software Engineering ; Telecommunications |
WOS ID | WOS:000655830300019 |
Scopus ID | 2-s2.0-85087478813 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | Faculty of Science and Technology |
Corresponding Author | Xiong,Haoyi |
Affiliation | 1.Cloud Computing Center, Shenzhen Institutes of Advanced Technology Chinese Academy of Sciences, 85411 Shenzhen, Guangdong China 2.Big Data Lab, Baidu Inc, 438127 Beijing China 100085 3.ECE, University of Central Florida, 6243 Orlando, Florida United States 4.ECE, University of Central Florida, 6243 Orlando, Florida United States 5.CIS, University of Macau, 59193 Taipa, Macau Macao 6.Big Data Lab, Baidu Inc, 438127 Beijing China |
Recommended Citation GB/T 7714 | Zhao,Baoxin,Xiong,Haoyi,Bian,Jiang,et al. COMO: Efficient Deep Neural Networks Expansion With COnvolutional MaxOut[J]. IEEE Transactions on Multimedia, 2020, 23, 1722-1730. |
APA | Zhao,Baoxin., Xiong,Haoyi., Bian,Jiang., Guo,Zhishan., Xu,Cheng Zhong., & Dou,Dejing (2020). COMO: Efficient Deep Neural Networks Expansion With COnvolutional MaxOut. IEEE Transactions on Multimedia, 23, 1722-1730. |
MLA | Zhao,Baoxin,et al."COMO: Efficient Deep Neural Networks Expansion With COnvolutional MaxOut".IEEE Transactions on Multimedia 23(2020):1722-1730. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment