Sensors | Free Full-Text | BY-SLAM: Dynamic Visual SLAM System Based on BEBLID and Semantic Information Extraction
Next Article in Journal
Clustering Runners’ Response to Different Midsole Stack Heights: A Field Study
Previous Article in Journal
Ultrasonic Levitation for Airway Humidification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

BY-SLAM: Dynamic Visual SLAM System Based on BEBLID and Semantic Information Extraction

by
Daixian Zhu
*,
Peixuan Liu
*,
Qiang Qiu
,
Jiaxin Wei
and
Ruolin Gong
College of Communication and Information Engineering, Xi’an University of Science and Technology, Xi’an 710054, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(14), 4693; https://doi.org/10.3390/s24144693
Submission received: 18 June 2024 / Revised: 17 July 2024 / Accepted: 18 July 2024 / Published: 19 July 2024
(This article belongs to the Section Navigation and Positioning)

Abstract

SLAM is a critical technology for enabling autonomous navigation and positioning in unmanned vehicles. Traditional visual simultaneous localization and mapping algorithms are built upon the assumption of a static scene, overlooking the impact of dynamic targets within real-world environments. Interference from dynamic targets can significantly degrade the system’s localization accuracy or even lead to tracking failure. To address these issues, we propose a dynamic visual SLAM system named BY-SLAM, which is based on BEBLID and semantic information extraction. Initially, the BEBLID descriptor is introduced to describe Oriented FAST feature points, enhancing both feature point matching accuracy and speed. Subsequently, FasterNet replaces the backbone network of YOLOv8s to expedite semantic information extraction. By using the results of DBSCAN clustering object detection, a more refined semantic mask is obtained. Finally, by leveraging the semantic mask and epipolar constraints, dynamic feature points are discerned and eliminated, allowing for the utilization of only static feature points for pose estimation and the construction of a dense 3D map that excludes dynamic targets. Experimental evaluations are conducted on both the TUM RGB-D dataset and real-world scenarios and demonstrate the effectiveness of the proposed algorithm at filtering out dynamic targets within the scenes. On average, the localization accuracy for the TUM RGB-D dataset improves by 95.53% compared to ORB-SLAM3. Comparative analyses against classical dynamic SLAM systems further corroborate the improvement in localization accuracy, map readability, and robustness achieved by BY-SLAM.
Keywords: visual SLAM; BEBLID; YOLOv8s; FasterNet; epipolar constraint; clustering visual SLAM; BEBLID; YOLOv8s; FasterNet; epipolar constraint; clustering

Share and Cite

MDPI and ACS Style

Zhu, D.; Liu, P.; Qiu, Q.; Wei, J.; Gong, R. BY-SLAM: Dynamic Visual SLAM System Based on BEBLID and Semantic Information Extraction. Sensors 2024, 24, 4693. https://doi.org/10.3390/s24144693

AMA Style

Zhu D, Liu P, Qiu Q, Wei J, Gong R. BY-SLAM: Dynamic Visual SLAM System Based on BEBLID and Semantic Information Extraction. Sensors. 2024; 24(14):4693. https://doi.org/10.3390/s24144693

Chicago/Turabian Style

Zhu, Daixian, Peixuan Liu, Qiang Qiu, Jiaxin Wei, and Ruolin Gong. 2024. "BY-SLAM: Dynamic Visual SLAM System Based on BEBLID and Semantic Information Extraction" Sensors 24, no. 14: 4693. https://doi.org/10.3390/s24144693

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop