3D SLAM based on NDT matching and ground constraints for ground robots in complex environments
ISSN: 0143-991X
Article publication date: 24 August 2022
Issue publication date: 2 January 2023
Abstract
Purpose
In large-scale environments and unstructured scenarios, the accuracy and robustness of traditional light detection and ranging (LiDAR) simultaneous localization and mapping (SLAM) algorithms are reduced, and the algorithms might even be completely ineffective. To overcome these problems, this study aims to propose a 3D LiDAR SLAM method for ground-based mobile robots, which uses a 3D LiDAR fusion inertial measurement unit (IMU) to establish an environment map and realize real-time localization.
Design/methodology/approach
First, we use a normal distributions transform (NDT) algorithm based on a local map with a corresponding motion prediction model for point cloud registration in the front-end. Next, point cloud features are tightly coupled with IMU angle constraints, ground constraints and gravity constraints for graph-based optimization in the back-end. Subsequently, the cumulative error is reduced by adding loop closure detection.
Findings
The algorithm is tested using a public data set containing indoor and outdoor scenarios. The results confirm that the proposed algorithm has high accuracy and robustness.
Originality/value
To improve the accuracy and robustness of SLAM, this method proposed in the paper introduced the NDT algorithm in the front-end and designed ground constraints and gravity constraints in the back-end. The proposed method has a satisfactory performance when applied to ground-based mobile robots in complex environments experiments.
Keywords
Acknowledgements
National Natural Science Foundation of China U20A20201. Key R&D projects of Liaoning Province 2020JH2/10300104.
Citation
Jiang, Y., Wang, T., Shao, S. and Wang, L. (2023), "3D SLAM based on NDT matching and ground constraints for ground robots in complex environments", Industrial Robot, Vol. 50 No. 1, pp. 174-185. https://doi.org/10.1108/IR-05-2022-0128
Publisher
:Emerald Publishing Limited
Copyright © 2022, Emerald Publishing Limited