학술논문

EKF-LOAM: An Adaptive Fusion of LiDAR SLAM With Wheel Odometry and Inertial Data for Confined Spaces With Few Geometric Features
Document Type
Periodical
Source
IEEE Transactions on Automation Science and Engineering IEEE Trans. Automat. Sci. Eng. Automation Science and Engineering, IEEE Transactions on. 19(3):1458-1471 Jul, 2022
Subject
Robotics and Control Systems
Power, Energy and Industry Applications
Components, Circuits, Devices and Systems
Laser radar
Simultaneous localization and mapping
Robots
Sensors
Wheels
Cameras
Mobile robots
Mobile service robots
adaptive extended Kalman filtering
simultaneous localization and mapping in confined spaces
LiDAR SLAM
wheel odometry
vector field based navigation
Language
ISSN
1545-5955
1558-3783
Abstract
A precise localization system and a map that properly represents the environment are fundamental for several robotic applications. Traditional LiDAR SLAM algorithms are particularly susceptible to underestimating the distance covered by real robots in environments with few geometric features. Common industrial confined spaces, such as ducts and galleries, have long and homogeneous structures, which are difficult to map. In this paper, we propose a novel approach, the EKF-LOAM , which fuses wheel odometry and IMU (Inertial Measurement Unit) data into the LeGO-LOAM algorithm using an Extended Kalman Filter. For that, the EKF-LOAM uses a simple and lightweight adaptive covariance matrix based on the number of detected geometric features. Simulated and real-world experiments with the EspeleoRobô, a service robot designed to inspect confined places, show that the EKF-LOAM method reduces the underestimating problem, with improvements greater than 50% when compared to the original LeGO-LOAM algorithm. Note to Practitioners —This paper is motivated by the challenges of autonomous navigation for mobile ground robots within confined and unstructured environments. Here, we propose a data fusion framework that uses common sensors (such as LiDARs, wheel odometry, and inertial devices) to improve the simultaneous localization and mapping (SLAM) capabilities of a robot without GPS and compass. This approach does not need artificial landmarks nor ideal light and, in scenarios with few geometric features, increases the performance of LiDAR SLAM techniques based on edge and planar features. We also provide a robust controller for the autonomous navigation of the robot during the mapping of a tunnel. Experiments carried out in simulation and real-world confined places show the effectiveness of our approach. In future work, we shall incorporate other sensors, such as cameras, to improve the SLAM process.