Fast Semantic-Aware Motion State Detection for Visual SLAM in Dynamic Environment

Page view(s)
95
Checked on Oct 06, 2024
Fast Semantic-Aware Motion State Detection for Visual SLAM in Dynamic Environment
Title:
Fast Semantic-Aware Motion State Detection for Visual SLAM in Dynamic Environment
Journal Title:
IEEE Transactions on Intelligent Transportation Systems
Publication Date:
25 October 2022
Citation:
Singh, G., Wu, M., Do, M. V., & Lam, S.-K. (2022). Fast Semantic-Aware Motion State Detection for Visual SLAM in Dynamic Environment. IEEE Transactions on Intelligent Transportation Systems, 1–17. https://doi.org/10.1109/tits.2022.3213694
Abstract:
Existing visual SLAM (vSLAM) systems fail to perform well in dynamic environments as they cannot effectively ignore moving objects during pose estimation and mapping. We propose a lightweight approach to improve the robustness of existing feature based RGB-D and stereo vSLAM by accurately removing dynamic outliers in the scene that contribute to failures in pose estimation and mapping. First, a novel motion state detection algorithm using the depth and feature flow information is presented to identify regions in the scene with high moving probability. This information is then fused with semantic cues via a probability framework to enable accurate and robust moving object extraction to retain the useful features for pose estimation and mapping. To reduce the computational complexity of extracting semantic information in every frame, we propose to extract semantics only on keyframes with significant changes in image content. Semantic propagation is used to compensate for the changes in the intermediate frames (i.e., non-keyframes). This is achieved by computing the dense transformation map using the available feature flow vectors. The proposed techniques can be integrated into existing vSLAM systems to increase their robustness in dynamic environments without incurring much computation cost. Our work highlights the importance of distinguishing between motion states of potential moving objects for vSLAM in highly dynamic environments. We provide extensive experimental results on four well-known RGB-D and stereo datasets to show that the proposed technique outperforms existing vSLAM methods in indoor and outdoor environments under various dynamic scenarios including crowded scenes. We also perform our experiments on a low-cost embedded platform, i.e., Jetson TX1, to demonstrate the computational efficiency of our method.
License type:
Publisher Copyright
Funding Info:
This research / project is supported by the Ministry of Education - Academic Research Fund Tier 1
Grant Reference no. : RG78/21

This research / project is supported by the A*STAR - RIE2020 Industry Alignment Fund – Industry Collaboration Projects (IAF-ICP)
Grant Reference no. : 2019-1079

Cash and in-kind contribution from Singapore Telecommunications Limited (Singtel), through Singtel Cognitive and Artificial Intelligence Lab for Enterprises (SCALE@NTU)
Description:
© 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
ISSN:
1558-0016
1524-9050
Files uploaded: