Residential College | false |
Status | 已發表Published |
LWSIS: LiDAR-Guided Weakly Supervised Instance Segmentation for Autonomous Driving | |
Li, Xiang1; Yin, Junbo1; Shi, Botian2; Li, Yikang2; Yang, Ruigang3; Shen, Jianbing4 | |
2023-06-26 | |
Conference Name | 37th AAAI Conference on Artificial Intelligence |
Source Publication | Proceedings of the 37th AAAI Conference on Artificial Intelligence, AAAI 2023 |
Volume | 37 |
Issue | 2 |
Pages | 1433-1441 |
Conference Date | 2023/02/07-2023/02/14 |
Conference Place | Washington DC |
Country | USA |
Publisher | AAAI Press |
Abstract | Image instance segmentation is a fundamental research topic in autonomous driving, which is crucial for scene understanding and road safety. Advanced learning-based approaches often rely on the costly 2D mask annotations for training. In this paper, we present a more artful framework, LiDAR-guided Weakly Supervised Instance Segmentation (LWSIS), which leverages the off-the-shelf 3D data, i.e., Point Cloud, together with the 3D boxes, as natural weak supervisions for training the 2D image instance segmentation models. Our LWSIS not only exploits the complementary information in multimodal data during training, but also significantly reduces the annotation cost of the dense 2D masks. In detail, LWSIS consists of two crucial modules, Point Label Assignment (PLA) and Graph-based Consistency Regularization (GCR). The former module aims to automatically assign the 3D point cloud as 2D point-wise labels, while the latter further refines the predictions by enforcing geometry and appearance consistency of the multimodal data. Moreover, we conduct a secondary instance segmentation annotation on the nuScenes, named nuInsSeg, to encourage further research on multimodal perception tasks. Extensive experiments on the nuInsSeg, as well as the large-scale Waymo, show that LWSIS can substantially improve existing weakly supervised segmentation models by only involving 3D data during training. Additionally, LWSIS can also be incorporated into 3D object detectors like PointPainting to boost the 3D detection performance for free. The code and dataset are available at https://github.com/Serenos/LWSIS. |
Keyword | Vision For Robotics & Autonomous Driving Multi-modal Vision Segmentation |
DOI | 10.1609/aaai.v37i2.25228 |
URL | View the original |
Language | 英語English |
Scopus ID | 2-s2.0-85167681603 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE THE STATE KEY LABORATORY OF INTERNET OF THINGS FOR SMART CITY (UNIVERSITY OF MACAU) |
Co-First Author | Li, Xiang |
Corresponding Author | Shen, Jianbing |
Affiliation | 1.School of Computer Science, Beijing Institute of Technology, China 2.Shanghai AI Laboratory, China 3.Inceptio, United States 4.SKL-IOTSC, CIS, University of Macau, Macao |
Corresponding Author Affilication | University of Macau |
Recommended Citation GB/T 7714 | Li, Xiang,Yin, Junbo,Shi, Botian,et al. LWSIS: LiDAR-Guided Weakly Supervised Instance Segmentation for Autonomous Driving[C]:AAAI Press, 2023, 1433-1441. |
APA | Li, Xiang., Yin, Junbo., Shi, Botian., Li, Yikang., Yang, Ruigang., & Shen, Jianbing (2023). LWSIS: LiDAR-Guided Weakly Supervised Instance Segmentation for Autonomous Driving. Proceedings of the 37th AAAI Conference on Artificial Intelligence, AAAI 2023, 37(2), 1433-1441. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment