To read this content please select one of the options below:

Visual SLAM algorithm in dynamic environment based on deep learning

Yingjie Yu (School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing, China)
Shuai Chen (Kunlun Digital Technology Co., Ltd, Beijing, China)
Xinpeng Yang (School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing, China)
Changzhen Xu (School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing, China)
Sen Zhang (School of Automation and Electrical Engineering and Shunde Innovation School, University of Science and Technology Beijing, Beijing, China)
Wendong Xiao (School of Automation and Electrical Engineering and Shunde Innovation School, University of Science and Technology Beijing, Beijing, China)

Industrial Robot

ISSN: 0143-991X

Article publication date: 6 August 2024

40

Abstract

Purpose

This paper proposes a self-supervised monocular depth estimation algorithm under multiple constraints, which can generate the corresponding depth map end-to-end based on RGB images. On this basis, based on the traditional visual simultaneous localisation and mapping (VSLAM) framework, a dynamic object detection framework based on deep learning is introduced, and dynamic objects in the scene are culled during mapping.

Design/methodology/approach

Typical SLAM algorithms or data sets assume a static environment and do not consider the potential consequences of accidentally adding dynamic objects to a 3D map. This shortcoming limits the applicability of VSLAM in many practical cases, such as long-term mapping. In light of the aforementioned considerations, this paper presents a self-supervised monocular depth estimation algorithm based on deep learning. Furthermore, this paper introduces the YOLOv5 dynamic detection framework into the traditional ORBSLAM2 algorithm for the purpose of removing dynamic objects.

Findings

Compared with Dyna-SLAM, the algorithm proposed in this paper reduces the error by about 13%, and compared with ORB-SLAM2 by about 54.9%. In addition, the algorithm in this paper can process a single frame of image at a speed of 15–20 FPS on GeForce RTX 2080s, far exceeding Dyna-SLAM in real-time performance.

Originality/value

This paper proposes a VSLAM algorithm that can be applied to dynamic environments. The algorithm consists of a self-supervised monocular depth estimation part under multiple constraints and the introduction of a dynamic object detection framework based on YOLOv5.

Keywords

Acknowledgements

This work is supported in part by the National Natural Science Foundation of China under Grant 62173032, in part by Guangdong Basic and Applied Basic Research Foundation Grant 2022A1515140109, Grant BK22BF005.

Citation

Yu, Y., Chen, S., Yang, X., Xu, C., Zhang, S. and Xiao, W. (2024), "Visual SLAM algorithm in dynamic environment based on deep learning", Industrial Robot, Vol. ahead-of-print No. ahead-of-print. https://doi.org/10.1108/IR-04-2024-0166

Publisher

:

Emerald Publishing Limited

Copyright © 2024, Emerald Publishing Limited

Related articles