Residential College | false |
Status | 已發表Published |
LVI-ExC: A Target-free LiDAR-Visual-Inertial Extrinsic Calibration Framework | |
Wang, Zhong1; Zhang, Lin1; Shen, Ying1; Zhou, Yicong2 | |
2022-10-10 | |
Conference Name | 30th ACM International Conference on Multimedia, MM 2022 |
Source Publication | MM 2022 - Proceedings of the 30th ACM International Conference on Multimedia |
Pages | 3319-3327 |
Conference Date | 10 October 2022through 14 October 2022 |
Conference Place | Lisboa |
Abstract | Recently, the multi-modal fusion with 3D LiDAR, camera, and IMU has shown great potential in applications of automation-related fields. Yet a prerequisite for a successful fusion is that the geometric relationships among the sensors are accurately determined, which is called an extrinsic calibration problem. To date, the existing target-based approaches to deal with this problem rely on sophisticated calibration objects (sites) and well-trained operators, which is time-consuming and inflexible in practical applications. Contrarily, a few target-free methods can overcome these shortcomings, while they only focus on the calibrations of two types of the sensors. Although it is possible to obtain LiDAR-visual-inertial extrinsics by chained calibrations, problems such as cumbersome operations, large cumulative errors, and weak geometric consistency still exist. To this end, we propose LVI-ExC, an integrated LiDAR-Visual-Inertial Extrinsic Calibration framework, which takes natural multi-modal data as input and yields sensor-to-sensor extrinsics end-to-end without any auxiliary object (site) or manual assistance. To fuse multi-modal data, we formulate the LiDAR-visual-inertial extrinsic calibration as a continuous-time simultaneous localization and mapping problem, in which the extrinsics, trajectories, time differences, and map points are jointly estimated by establishing sensor-to-sensor and sensor-to-trajectory constraints. Extensive experiments show that LVI-ExC can produce precise results. With LVI-ExC's outputs, the LiDAR-visual reprojection results and the reconstructed environment map are all highly consistent with the actual natural scenes, demonstrating LVI-ExC's outstanding performance. To ensure that our results are fully reproducible, all the relevant data and codes have been released publicly at https://cslinzhang.github.io/LVI-ExC/. |
Keyword | Extrinsic Calibration Multi-modal Fusion Target-free Calibration |
DOI | 10.1145/3503161.3547878 |
URL | View the original |
Language | 英語English |
Scopus ID | 2-s2.0-85151060358 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | Faculty of Science and Technology DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Affiliation | 1.School of Software Engineering, Tongji University, Shanghai, China 2.Department of Computer and Information Science, University of Macau, Macao |
Recommended Citation GB/T 7714 | Wang, Zhong,Zhang, Lin,Shen, Ying,et al. LVI-ExC: A Target-free LiDAR-Visual-Inertial Extrinsic Calibration Framework[C], 2022, 3319-3327. |
APA | Wang, Zhong., Zhang, Lin., Shen, Ying., & Zhou, Yicong (2022). LVI-ExC: A Target-free LiDAR-Visual-Inertial Extrinsic Calibration Framework. MM 2022 - Proceedings of the 30th ACM International Conference on Multimedia, 3319-3327. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment