Residential Collegefalse
Status已發表Published
Robust and Efficient Memory Network for Video Object Segmentation
Chen, Yadang1; Zhang, Dingwei2; Yang, Zhi Xin3; Wu, Enhua4
2023-11
Conference NameInternational Conference on Multimedia and Expo (ICME)
Source PublicationProceedings - IEEE International Conference on Multimedia and Expo
Volume2023-July
Pages1769-1774
Conference Date10-14 July 2023
Conference PlaceBrisbane, Australia
CountryAustralia
PublisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 445 HOES LANE, PISCATAWAY, NJ 08855-4141
Abstract

This paper proposes a Robust and Efficient Memory Network, referred to as REMN, for studying semi-supervised video object segmentation (VOS). Memory-based methods have recently achieved outstanding VOS performance by performing non-local pixel-wise matching between the query and memory. However, these methods have two limitations. 1) Non-local matching could cause distractor objects in the background to be incorrectly segmented. 2) Memory features with high temporal redundancy consume significant computing resources. For limitation 1, we introduce a local attention mechanism that tackles the background distraction by enhancing the features of foreground objects with the previous mask. For limitation 2, we first adaptively decide whether to update the memory features depending on the variation of foreground objects to reduce temporal redundancy. Second, we employ a dynamic memory bank, which uses a lightweight and differentiable soft modulation gate to decide how many memory features need to be removed in the temporal dimension. Experiments demonstrate that our REMN achieves state-of-the-art results on DAVIS 2017, with a J & F score of 86.3% and on YouTube-VOS 2018, with a G over mean of 85.5%. Furthermore, our network shows a high inference speed of 25+ FPS and uses relatively few computing resources.

KeywordBackground Distraction Space-time Memory Network Temporal Redundancy Video Object Segmentation
DOI10.1109/ICME55011.2023.00304
URLView the original
Indexed ByCPCI-S
Language英語English
WOS Research AreaComputer Science
WOS SubjectComputer Science, Artificial Intelligence ; Computer Science, Information Systems ; Computer Science, Theory & Methods
WOS IDWOS:001062707300287
Scopus ID2-s2.0-85171143019
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionTHE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU)
Faculty of Science and Technology
DEPARTMENT OF ELECTROMECHANICAL ENGINEERING
Corresponding AuthorChen, Yadang
Affiliation1.Nanjing University of Information Science and Technology, School of Computer Science, Nanjing, China
2.Nanjing University of Information Science and Technology, School of Software, Nanjing, China
3.University of Macau, State Key Laboratory of Internet of Things for Smart City, Macao
4.University of Chinese Academy of Sciences, State Key Laboratory of Computer Science, Beijing, China
Recommended Citation
GB/T 7714
Chen, Yadang,Zhang, Dingwei,Yang, Zhi Xin,et al. Robust and Efficient Memory Network for Video Object Segmentation[C]:IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 445 HOES LANE, PISCATAWAY, NJ 08855-4141, 2023, 1769-1774.
APA Chen, Yadang., Zhang, Dingwei., Yang, Zhi Xin., & Wu, Enhua (2023). Robust and Efficient Memory Network for Video Object Segmentation. Proceedings - IEEE International Conference on Multimedia and Expo, 2023-July, 1769-1774.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Chen, Yadang]'s Articles
[Zhang, Dingwei]'s Articles
[Yang, Zhi Xin]'s Articles
Baidu academic
Similar articles in Baidu academic
[Chen, Yadang]'s Articles
[Zhang, Dingwei]'s Articles
[Yang, Zhi Xin]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Chen, Yadang]'s Articles
[Zhang, Dingwei]'s Articles
[Yang, Zhi Xin]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.