To read this content please select one of the options below:

Development of vision–based SLAM: from traditional methods to multimodal fusion

Zengrui Zheng (School of Automation Science and Engineering, South China University of Technology, Guangzhou, China)
Kainan Su (School of Automation Science and Engineering, South China University of Technology, Guangzhou, China)
Shifeng Lin (School of Automation Science and Engineering, South China University of Technology, Guangzhou, China)
Zhiquan Fu (Zhejiang VIE Science and Technology Co., Ltd, Zhuji, China)
Chenguang Yang (Bristol Robotics Laboratory, University of the West of England, Bristol, UK and Department of Computer Science, University of Liverpool, Liverpool, UK)

Robotic Intelligence and Automation

ISSN: 2754-6969

Article publication date: 9 July 2024

Issue publication date: 18 July 2024

230

Abstract

Purpose

Visual simultaneous localization and mapping (SLAM) has limitations such as sensitivity to lighting changes and lower measurement accuracy. The effective fusion of information from multiple modalities to address these limitations has emerged as a key research focus. This study aims to provide a comprehensive review of the development of vision-based SLAM (including visual SLAM) for navigation and pose estimation, with a specific focus on techniques for integrating multiple modalities.

Design/methodology/approach

This paper initially introduces the mathematical models and framework development of visual SLAM. Subsequently, this paper presents various methods for improving accuracy in visual SLAM by fusing different spatial and semantic features. This paper also examines the research advancements in vision-based SLAM with respect to multi-sensor fusion in both loosely coupled and tightly coupled approaches. Finally, this paper analyzes the limitations of current vision-based SLAM and provides predictions for future advancements.

Findings

The combination of vision-based SLAM and deep learning has significant potential for development. There are advantages and disadvantages to both loosely coupled and tightly coupled approaches in multi-sensor fusion, and the most suitable algorithm should be chosen based on the specific application scenario. In the future, vision-based SLAM is evolving toward better addressing challenges such as resource-limited platforms and long-term mapping.

Originality/value

This review introduces the development of vision-based SLAM and focuses on the advancements in multimodal fusion. It allows readers to quickly understand the progress and current status of research in this field.

Keywords

Citation

Zheng, Z., Su, K., Lin, S., Fu, Z. and Yang, C. (2024), "Development of vision–based SLAM: from traditional methods to multimodal fusion", Robotic Intelligence and Automation, Vol. 44 No. 4, pp. 529-548. https://doi.org/10.1108/RIA-10-2023-0142

Publisher

:

Emerald Publishing Limited

Copyright © 2024, Emerald Publishing Limited

Related articles