Search results
1 – 10 of 24Zengrui Zheng, Kainan Su, Shifeng Lin, Zhiquan Fu and Chenguang Yang
Visual simultaneous localization and mapping (SLAM) has limitations such as sensitivity to lighting changes and lower measurement accuracy. The effective fusion of information…
Abstract
Purpose
Visual simultaneous localization and mapping (SLAM) has limitations such as sensitivity to lighting changes and lower measurement accuracy. The effective fusion of information from multiple modalities to address these limitations has emerged as a key research focus. This study aims to provide a comprehensive review of the development of vision-based SLAM (including visual SLAM) for navigation and pose estimation, with a specific focus on techniques for integrating multiple modalities.
Design/methodology/approach
This paper initially introduces the mathematical models and framework development of visual SLAM. Subsequently, this paper presents various methods for improving accuracy in visual SLAM by fusing different spatial and semantic features. This paper also examines the research advancements in vision-based SLAM with respect to multi-sensor fusion in both loosely coupled and tightly coupled approaches. Finally, this paper analyzes the limitations of current vision-based SLAM and provides predictions for future advancements.
Findings
The combination of vision-based SLAM and deep learning has significant potential for development. There are advantages and disadvantages to both loosely coupled and tightly coupled approaches in multi-sensor fusion, and the most suitable algorithm should be chosen based on the specific application scenario. In the future, vision-based SLAM is evolving toward better addressing challenges such as resource-limited platforms and long-term mapping.
Originality/value
This review introduces the development of vision-based SLAM and focuses on the advancements in multimodal fusion. It allows readers to quickly understand the progress and current status of research in this field.
Details
Keywords
Sixing Liu, Yan Chai, Rui Yuan and Hong Miao
Simultaneous localization and map building (SLAM), as a state estimation problem, is a prerequisite for solving the problem of autonomous vehicle motion in unknown environments…
Abstract
Purpose
Simultaneous localization and map building (SLAM), as a state estimation problem, is a prerequisite for solving the problem of autonomous vehicle motion in unknown environments. Existing algorithms are based on laser or visual odometry; however, the lidar sensing range is small, the amount of data features is small, the camera is vulnerable to external conditions and the localization and map building cannot be performed stably and accurately using a single sensor. This paper aims to propose a laser three dimensions tightly coupled map building method that incorporates visual information, and uses laser point cloud information and image information to complement each other to improve the overall performance of the algorithm.
Design/methodology/approach
The visual feature points are first matched at the front end of the method, and the mismatched point pairs are removed using the bidirectional random sample consensus (RANSAC) algorithm. The laser point cloud is then used to obtain its depth information, while the two types of feature points are fed into the pose estimation module for a tightly coupled local bundle adjustment solution using a heuristic simulated annealing algorithm. Finally, the visual bag-of-words model is fused in the laser point cloud information to establish a threshold to construct a loopback framework to further reduce the cumulative drift error of the system over time.
Findings
Experiments on publicly available data sets show that the proposed method in this paper can match its real trajectory well. For various scenes, the map can be constructed by using the complementary laser and vision sensors, with high accuracy and robustness. At the same time, the method is verified in a real environment using an autonomous walking acquisition platform, and the system loaded with the method can run well for a long time and take into account the environmental adaptability of multiple scenes.
Originality/value
A multi-sensor data tight coupling method is proposed to fuse laser and vision information for optimal solution of the positional attitude. A bidirectional RANSAC algorithm is used for the removal of visual mismatched point pairs. Further, oriented fast and rotated brief feature points are used to build a bag-of-words model and construct a real-time loopback framework to reduce error accumulation. According to the experimental validation results, the accuracy and robustness of the single-sensor SLAM algorithm can be improved.
Details
Keywords
Yong Qin and Haidong Yu
This paper aims to provide a better understanding of the challenges and potential solutions in Visual Simultaneous Localization and Mapping (SLAM), laying the foundation for its…
Abstract
Purpose
This paper aims to provide a better understanding of the challenges and potential solutions in Visual Simultaneous Localization and Mapping (SLAM), laying the foundation for its applications in autonomous navigation, intelligent driving and other related domains.
Design/methodology/approach
In analyzing the latest research, the review presents representative achievements, including methods to enhance efficiency, robustness and accuracy. Additionally, the review provides insights into the future development direction of Visual SLAM, emphasizing the importance of improving system robustness when dealing with dynamic environments. The research methodology of this review involves a literature review and data set analysis, enabling a comprehensive understanding of the current status and prospects in the field of Visual SLAM.
Findings
This review aims to comprehensively evaluate the latest advances and challenges in the field of Visual SLAM. By collecting and analyzing relevant research papers and classic data sets, it reveals the current issues faced by Visual SLAM in complex environments and proposes potential solutions. The review begins by introducing the fundamental principles and application areas of Visual SLAM, followed by an in-depth discussion of the challenges encountered when dealing with dynamic objects and complex environments. To enhance the performance of SLAM algorithms, researchers have made progress by integrating different sensor modalities, improving feature extraction and incorporating deep learning techniques, driving advancements in the field.
Originality/value
To the best of the authors’ knowledge, the originality of this review lies in its in-depth analysis of current research hotspots and predictions for future development, providing valuable references for researchers in this field.
Details
Keywords
Zhiyu Li, Hongguang Li, Yang Liu, Lingyun Jin and Congqing Wang
Autonomous flight of unmanned aerial vehicles (UAVs) in global position system (GPS)-denied environments has become an increasing research hotspot. This paper aims to realize the…
Abstract
Purpose
Autonomous flight of unmanned aerial vehicles (UAVs) in global position system (GPS)-denied environments has become an increasing research hotspot. This paper aims to realize the indoor fixed-point hovering control and autonomous flight for UAVs based on visual inertial simultaneous localization and mapping (SLAM) and sensor fusion algorithm based on extended Kalman filter.
Design/methodology/approach
The fundamental of the proposed method is using visual inertial SLAM to estimate the position information of the UAV and position-speed double-loop controller to control the UAV. The motion and observation models of the UAV and the fusion algorithm are given. Finally, experiments are performed to test the proposed algorithms.
Findings
A position-speed double-loop controller is proposed, by fusing the position information obtained by visual inertial SLAM with the data of airborne sensors. The experiment results of the indoor fixed-points hovering show that UAV flight control can be realized based on visual inertial SLAM in the absence of GPS.
Originality/value
A position-speed double-loop controller for UAV is designed and tested, which provides a more stable position estimation and enabled UAV to fly autonomously and hover in GPS-denied environment.
Details
Keywords
Xiangdi Yue, Yihuan Zhang, Jiawei Chen, Junxin Chen, Xuanyi Zhou and Miaolei He
In recent decades, the field of robotic mapping has witnessed widespread research and development in light detection and ranging (LiDAR)-based simultaneous localization and…
Abstract
Purpose
In recent decades, the field of robotic mapping has witnessed widespread research and development in light detection and ranging (LiDAR)-based simultaneous localization and mapping (SLAM) techniques. This paper aims to provide a significant reference for researchers and engineers in robotic mapping.
Design/methodology/approach
This paper focused on the research state of LiDAR-based SLAM for robotic mapping as well as a literature survey from the perspective of various LiDAR types and configurations.
Findings
This paper conducted a comprehensive literature review of the LiDAR-based SLAM system based on three distinct LiDAR forms and configurations. The authors concluded that multi-robot collaborative mapping and multi-source fusion SLAM systems based on 3D LiDAR with deep learning will be new trends in the future.
Originality/value
To the best of the authors’ knowledge, this is the first thorough survey of robotic mapping from the perspective of various LiDAR types and configurations. It can serve as a theoretical and practical guide for the advancement of academic and industrial robot mapping.
Details
Keywords
Yingjie Yu, Shuai Chen, Xinpeng Yang, Changzhen Xu, Sen Zhang and Wendong Xiao
This paper proposes a self-supervised monocular depth estimation algorithm under multiple constraints, which can generate the corresponding depth map end-to-end based on RGB…
Abstract
Purpose
This paper proposes a self-supervised monocular depth estimation algorithm under multiple constraints, which can generate the corresponding depth map end-to-end based on RGB images. On this basis, based on the traditional visual simultaneous localisation and mapping (VSLAM) framework, a dynamic object detection framework based on deep learning is introduced, and dynamic objects in the scene are culled during mapping.
Design/methodology/approach
Typical SLAM algorithms or data sets assume a static environment and do not consider the potential consequences of accidentally adding dynamic objects to a 3D map. This shortcoming limits the applicability of VSLAM in many practical cases, such as long-term mapping. In light of the aforementioned considerations, this paper presents a self-supervised monocular depth estimation algorithm based on deep learning. Furthermore, this paper introduces the YOLOv5 dynamic detection framework into the traditional ORBSLAM2 algorithm for the purpose of removing dynamic objects.
Findings
Compared with Dyna-SLAM, the algorithm proposed in this paper reduces the error by about 13%, and compared with ORB-SLAM2 by about 54.9%. In addition, the algorithm in this paper can process a single frame of image at a speed of 15–20 FPS on GeForce RTX 2080s, far exceeding Dyna-SLAM in real-time performance.
Originality/value
This paper proposes a VSLAM algorithm that can be applied to dynamic environments. The algorithm consists of a self-supervised monocular depth estimation part under multiple constraints and the introduction of a dynamic object detection framework based on YOLOv5.
Details
Keywords
Minghao Wang, Ming Cong, Dong Liu, Yu Du, Xiaojing Tian and Bing Li
The purpose of this study is to designed a robot odometry based on three dimensional (3D) laser point cloud data, inertial measurement unit (IMU) data and real-time kinematic…
Abstract
Purpose
The purpose of this study is to designed a robot odometry based on three dimensional (3D) laser point cloud data, inertial measurement unit (IMU) data and real-time kinematic (RTK) data in underground spatial features and gravity fluctuations environment. This method improves the mapping accuracy in two types of underground space: multi-layer space and large-scale scenarios.
Design/methodology/approach
An IMU–Laser–RTK fusion mapping algorithm based on Iterative Kalman Filter was proposed, and the observation equation and Jacobian matrix were derived. Aiming at the problem of inaccurate gravity estimation, the optimization of gravity is transformed into the optimization of SO(3), which avoids the problem of gravity over-parameterization.
Findings
Compared with the optimization method, the computational cost is reduced. Without relying on the wheel speed odometer, the robot synchronization localization and 3D environment modeling for multi-layer space are realized. The performance of the proposed algorithm is tested and compared in two types of underground space, and the robustness and accuracy in multi-layer space and large-scale scenarios are verified. The results show that the root mean square error of the proposed algorithm is 0.061 m, which achieves higher accuracy than other algorithms.
Originality/value
Based on the problem of large loop and low feature scale, this algorithm can better complete the map loop and self-positioning, and its root mean square error is more than double compared with other methods. The method proposed in this paper can better complete the autonomous positioning of the robot in the underground space with hierarchical feature degradation, and at the same time, an accurate 3D map can be constructed for subsequent research.
Details
Keywords
Krystian Borodacz and Cezary Szczepański
Before designing a navigation system, it is necessary to analyse possible approaches in terms of expected accuracy, existing limitations and economic justification to select the…
Abstract
Purpose
Before designing a navigation system, it is necessary to analyse possible approaches in terms of expected accuracy, existing limitations and economic justification to select the most advantageous solution. This paper aims to collect possible navigation methods that can provide correction for inertial navigation and to evaluate their suitability for use on a manoeuvring tactical missile.
Design/methodology/approach
The review of existing munitions was based on data collected from the literature and online databases. The data collected included dimensions, performance, applied navigation and guidance methods and their achievable accuracy. The requirements and limitations identified were confronted with the range of sensor parameters available on the market. Based on recent literature, navigation methods were reviewed and evaluated for applicability to inertial navigation system (INS) correction in global navigation satellite system-denied space.
Findings
The performance analysis of existing munition shows that small and relatively inexpensive micro-electro-mechanical system-type inertial sensors are required. A review of the parameters of existing devices of this type has shown that they are subject to measurement errors that do not allow them to achieve the delivery accuracy expected of precision missiles. The most promising navigation correction methods for manoeuvring flying objects have been identified.
Originality/value
The information presented in this paper is the result of the first phase of a project and presents the results of the requirements selection, initial sizing and preliminary design of the navigation system. This paper combines a review of the current state of the art in missile systems and an analysis of INS accuracy including the selection of sensor parameters.
Details
Keywords
Ruoxing Wang, Shoukun Wang, Junfeng Xue, Zhihua Chen and Jinge Si
This paper aims to investigate an autonomous obstacle-surmounting method based on a hybrid gait for the problem of crossing low-height obstacles autonomously by a six wheel-legged…
Abstract
Purpose
This paper aims to investigate an autonomous obstacle-surmounting method based on a hybrid gait for the problem of crossing low-height obstacles autonomously by a six wheel-legged robot. The autonomy of obstacle-surmounting is reflected in obstacle recognition based on multi-frame point cloud fusion.
Design/methodology/approach
In this paper, first, for the problem that the lidar on the robot cannot scan the point cloud of low-height obstacles, the lidar is driven to rotate by a 2D turntable to obtain the point cloud of low-height obstacles under the robot. Tightly-coupled Lidar Inertial Odometry via Smoothing and Mapping algorithm, fast ground segmentation algorithm and Euclidean clustering algorithm are used to recognize the point cloud of low-height obstacles and obtain low-height obstacle in-formation. Then, combined with the structural characteristics of the robot, the obstacle-surmounting action planning is carried out for two types of obstacle scenes. A segmented approach is used for action planning. Gait units are designed to describe each segment of the action. A gait matrix is used to describe the overall action. The paper also analyzes the stability and surmounting capability of the robot’s key pose and determines the robot’s surmounting capability and the value scheme of the surmounting control variables.
Findings
The experimental verification is carried out on the robot laboratory platform (BIT-6NAZA). The obstacle recognition method can accurately detect low-height obstacles. The robot can maintain a smooth posture to cross low-height obstacles, which verifies the feasibility of the adaptive obstacle-surmounting method.
Originality/value
The study can provide the theory and engineering foundation for the environmental perception of the unmanned platform. It provides environmental information to support follow-up work, for example, on the planning of obstacles and obstacles.
Details
Keywords
Weak repeatability is observed in handcrafted keypoints, leading to tracking failures in visual simultaneous localization and mapping (SLAM) systems under challenging scenarios…
Abstract
Purpose
Weak repeatability is observed in handcrafted keypoints, leading to tracking failures in visual simultaneous localization and mapping (SLAM) systems under challenging scenarios such as illumination change, rapid rotation and large angle of view variation. In contrast, learning-based keypoints exhibit higher repetition but entail considerable computational costs. This paper proposes an innovative algorithm for keypoint extraction, aiming to strike an equilibrium between precision and efficiency. This paper aims to attain accurate, robust and versatile visual localization in scenes of formidable complexity.
Design/methodology/approach
SiLK-SLAM initially refines the cutting-edge learning-based extractor, SiLK, and introduces an innovative postprocessing algorithm for keypoint homogenization and operational efficiency. Furthermore, SiLK-SLAM devises a reliable relocalization strategy called PCPnP, leveraging progressive and consistent sampling, thereby bolstering its robustness.
Findings
Empirical evaluations conducted on TUM, KITTI and EuRoC data sets substantiate SiLK-SLAM’s superior localization accuracy compared to ORB-SLAM3 and other methods. Compared to ORB-SLAM3, SiLK-SLAM demonstrates an enhancement in localization accuracy even by 70.99%, 87.20% and 85.27% across the three data sets. The relocalization experiments demonstrate SiLK-SLAM’s capability in producing precise and repeatable keypoints, showcasing its robustness in challenging environments.
Originality/value
The SiLK-SLAM achieves exceedingly elevated localization accuracy and resilience in formidable scenarios, holding paramount importance in enhancing the autonomy of robots navigating intricate environments. Code is available at https://github.com/Pepper-FlavoredChewingGum/SiLK-SLAM.
Details