1
|
Chang Z, Wu H, Li C. YOLOv4‐tiny‐based robust RGB‐D SLAM approach with point and surface feature fusion in complex indoor environments. J FIELD ROBOT 2022. [DOI: 10.1002/rob.22145] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
Affiliation(s)
- Zhanyuan Chang
- College of Information, Mechanical and Electrical Engineering Shanghai Normal University Shanghai China
| | - Honglin Wu
- College of Information, Mechanical and Electrical Engineering Shanghai Normal University Shanghai China
| | - Chuanjiang Li
- College of Information, Mechanical and Electrical Engineering Shanghai Normal University Shanghai China
| |
Collapse
|
2
|
Yuan H, Wu C, Deng Z, Yin J. Robust Visual Odometry Leveraging Mixture of Manhattan Frames in Indoor Environments. SENSORS (BASEL, SWITZERLAND) 2022; 22:8644. [PMID: 36433239 PMCID: PMC9698556 DOI: 10.3390/s22228644] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/10/2022] [Revised: 11/05/2022] [Accepted: 11/07/2022] [Indexed: 06/16/2023]
Abstract
We propose a robust RGB-Depth (RGB-D) Visual Odometry (VO) system to improve the localization performance of indoor scenes by using geometric features, including point and line features. Previous VO/Simultaneous Localization and Mapping (SLAM) algorithms estimate the low-drift camera poses with the Manhattan World (MW)/Atlanta World (AW) assumption, which limits the applications of such systems. In this paper, we divide the indoor environments into two different scenes: MW and non-MW scenes. The Manhattan scenes are modeled as a Mixture of Manhattan Frames, in which each Manhattan Frame in itself defines a Manhattan World of a specific orientation. Moreover, we provide a method to detect Manhattan Frames (MFs) using the dominant directions extracted from the parallel lines. Our approach is designed with lower computational complexity than existing techniques using planes to detect Manhattan Frame (MF). For MW scenes, we separately estimate rotational and translational motion. A novel method is proposed to estimate the drift-free rotation using MF observations, unit direction vectors of lines, and surface normal vectors. Then, the translation part is recovered from point-line tracking. In non-MW scenes, the tracked and matched dominant directions are combined with the point and line features to estimate the full 6 degree of freedom (DoF) camera poses. Additionally, we exploit the rotation constraints generated from the multi-view dominant directions observations. The constraints are combined with the reprojection errors of points and lines to refine the camera pose through local map bundle adjustment. Evaluations on both synthesized and real-world datasets demonstrate that our approach outperforms state-of-the-art methods. On synthesized datasets, average localization accuracy is 1.5 cm, which is equivalent to state-of-the-art methods. On real-world datasets, the average localization accuracy is 1.7 cm, which outperforms the state-of-the-art methods by 43%. Our time consumption is reduced by 36%.
Collapse
Affiliation(s)
- Huayu Yuan
- School of Electronic Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
| | - Chengfeng Wu
- Science and Technology on Complex System Control and Intelligent Agent Cooperation Laboratory, Beijing 100074, China
| | - Zhongliang Deng
- School of Electronic Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
| | - Jiahui Yin
- School of Electronic Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
| |
Collapse
|
3
|
Wang Y, Bu H, Zhang X, Cheng J. YPD-SLAM: A Real-Time VSLAM System for Handling Dynamic Indoor Environments. SENSORS (BASEL, SWITZERLAND) 2022; 22:8561. [PMID: 36366259 PMCID: PMC9656896 DOI: 10.3390/s22218561] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/17/2022] [Revised: 10/28/2022] [Accepted: 11/03/2022] [Indexed: 06/16/2023]
Abstract
Aiming at the problem that Simultaneous localization and mapping (SLAM) is greatly disturbed by many dynamic elements in the actual environment, this paper proposes a real-time Visual SLAM (VSLAM) algorithm to deal with a dynamic indoor environment. Firstly, a lightweight YoloFastestV2 deep learning model combined with NCNN and Mobile Neural Network (MNN) inference frameworks is used to obtain preliminary semantic information of images. The dynamic feature points are removed according to epipolar constraint and dynamic properties of objects between consecutive frames. Since reducing the number of feature points after rejection affects the pose estimation, this paper innovatively combines Cylinder and Plane Extraction (CAPE) planar detection. We generate planes from depth maps and then introduce planar and in-plane point constraints into the nonlinear optimization of SLAM. Finally, the algorithm is tested on the publicly available TUM (RGB-D) dataset, and the average improvement in localization accuracy over ORB-SLAM2, DS-SLAM, and RDMO-SLAM is about 91.95%, 27.21%, and 30.30% under dynamic sequences, respectively. The single-frame tracking time of the whole system is only 42.68 ms, which is 44.1%, being 14.6-34.33% higher than DS-SLAM, RDMO-SLAM, and RDS-SLAM respectively. The system that we proposed significantly increases processing speed, performs better in real-time, and is easily deployed on various platforms.
Collapse
|
4
|
Li X, Wang W, Chen J, Zhang X. DR-SLAM: drift rejection SLAM with Manhattan regularity for indoor environments. Adv Robot 2022. [DOI: 10.1080/01691864.2022.2129032] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/01/2022]
Affiliation(s)
- Xiuzhi Li
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
- Engineering Research Center of Digital Community, Ministry of Education, Beijing, People's Republic of China
| | - Wen Wang
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
- Engineering Research Center of Digital Community, Ministry of Education, Beijing, People's Republic of China
| | - Jiahao Chen
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
- Engineering Research Center of Digital Community, Ministry of Education, Beijing, People's Republic of China
| | - Xiangyin Zhang
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
- Engineering Research Center of Digital Community, Ministry of Education, Beijing, People's Republic of China
| |
Collapse
|
5
|
Zhou B, Gilles M, Meng Y. Structure SLAM with points, planes and objects. Adv Robot 2022. [DOI: 10.1080/01691864.2022.2123253] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/01/2022]
Affiliation(s)
- Benchun Zhou
- Institute for Material Handling and Logistics, Karlsruhe Institute of Technogy, Karlsruhe, Germany
| | - Maximilian Gilles
- Institute for Material Handling and Logistics, Karlsruhe Institute of Technogy, Karlsruhe, Germany
| | - Yongqi Meng
- Institute for Material Handling and Logistics, Karlsruhe Institute of Technogy, Karlsruhe, Germany
| |
Collapse
|
6
|
Influence of the Stiffness of the Robotic Arm on the Position of the Effector of an EOD Robot. ELECTRONICS 2022. [DOI: 10.3390/electronics11152355] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
Terrestrial robots are being employed in a variety of sectors and for a variety of objectives. The purpose of this paper is to analyze and validate an analytical–numerical model of a robotic arm’s behavior. The proposed robot was designed to replace human personnel who remove ammunition or explosive devices. At the same time, the influence of the stiffness of the EOD robotic arm on the position of the effector in a variety of geometric task configurations was investigated. In order to obtain results relevant to the investigation, the angles of rotation under the load of each component of the arm’s composition and the vertical movement of the effector were measured. The main conclusions emphasize that a lower stiffness comes from the components of linear motors, which act on the elements of the robotic arm, and they substantially influence the elastic behavior of the arm. In addition, the constructive components of the arm have high rigidity compared to those of the linear actuators.
Collapse
|
7
|
Company-Corcoles JP, Garcia-Fidalgo E, Ortiz A. MSC-VO: Exploiting Manhattan and Structural Constraints for Visual Odometry. IEEE Robot Autom Lett 2022. [DOI: 10.1109/lra.2022.3142900] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
8
|
Aguiar AS, Neves dos Santos F, Sobreira H, Boaventura-Cunha J, Sousa AJ. Localization and Mapping on Agriculture Based on Point-Feature Extraction and Semiplanes Segmentation From 3D LiDAR Data. Front Robot AI 2022; 9:832165. [PMID: 35155589 PMCID: PMC8831384 DOI: 10.3389/frobt.2022.832165] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2021] [Accepted: 01/10/2022] [Indexed: 11/13/2022] Open
Abstract
Developing ground robots for agriculture is a demanding task. Robots should be capable of performing tasks like spraying, harvesting, or monitoring. However, the absence of structure in the agricultural scenes challenges the implementation of localization and mapping algorithms. Thus, the research and development of localization techniques are essential to boost agricultural robotics. To address this issue, we propose an algorithm called VineSLAM suitable for localization and mapping in agriculture. This approach uses both point- and semiplane-features extracted from 3D LiDAR data to map the environment and localize the robot using a novel Particle Filter that considers both feature modalities. The numeric stability of the algorithm was tested using simulated data. The proposed methodology proved to be suitable to localize a robot using only three orthogonal semiplanes. Moreover, the entire VineSLAM pipeline was compared against a state-of-the-art approach considering three real-world experiments in a woody-crop vineyard. Results show that our approach can localize the robot with precision even in long and symmetric vineyard corridors outperforming the state-of-the-art algorithm in this context.
Collapse
Affiliation(s)
- André Silva Aguiar
- INESC TEC—INESC Technology and Science, Porto, Portugal
- School of Science and Technology, University of Trás-os-Montes e Alto Douro, Vila Real, Portugal
- *Correspondence: André Silva Aguiar,
| | | | | | - José Boaventura-Cunha
- INESC TEC—INESC Technology and Science, Porto, Portugal
- School of Science and Technology, University of Trás-os-Montes e Alto Douro, Vila Real, Portugal
| | - Armando Jorge Sousa
- INESC TEC—INESC Technology and Science, Porto, Portugal
- FEUP, University of Porto, Porto, Portugal
| |
Collapse
|
9
|
Long R, Rauch C, Zhang T, Ivan V, Lam TL, Vijayakumar S. RGB-D SLAM in Indoor Planar Environments with Multiple Large Dynamic Objects. IEEE Robot Autom Lett 2022. [DOI: 10.1109/lra.2022.3186091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Affiliation(s)
- Ran Long
- Institute of Perception, Action and Behaviour, School of Informatics, University of Edinburgh, Edinburgh, U.K
| | - Christian Rauch
- Institute of Perception, Action and Behaviour, School of Informatics, University of Edinburgh, Edinburgh, U.K
| | - Tianwei Zhang
- Shenzhen Institute of Artificial Intelligence and Robotics for Society (AIRS), China
| | - Vladimir Ivan
- Institute of Perception, Action and Behaviour, School of Informatics, University of Edinburgh, Edinburgh, U.K
| | - Tin Lun Lam
- Shenzhen Institute of Artificial Intelligence and Robotics for Society (AIRS), China
| | - Sethu Vijayakumar
- Institute of Perception, Action and Behaviour, School of Informatics, University of Edinburgh, Edinburgh, U.K
| |
Collapse
|
10
|
Wu J, Xiong J, Guo H. Enforcing Regularities between Planes Using Key Plane for Monocular Mesh-based VIO. J INTELL ROBOT SYST 2021. [DOI: 10.1007/s10846-021-01529-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
11
|
Low-Bandwidth and Compute-Bound RGB-D Planar Semantic SLAM. SENSORS 2021; 21:s21165400. [PMID: 34450841 PMCID: PMC8399848 DOI: 10.3390/s21165400] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/17/2021] [Revised: 08/07/2021] [Accepted: 08/07/2021] [Indexed: 11/16/2022]
Abstract
Visual simultaneous location and mapping (SLAM) using RGB-D cameras has been a necessary capability for intelligent mobile robots. However, when using point-cloud map representations as most RGB-D SLAM systems do, limitations in onboard compute resources, and especially communication bandwidth can significantly limit the quantity of data processed and shared. This article proposes techniques that help address these challenges by mapping point clouds to parametric models in order to reduce computation and bandwidth load on agents. This contribution is coupled with a convolutional neural network (CNN) that extracts semantic information. Semantics provide guidance in object modeling which can reduce the geometric complexity of the environment. Pairing a parametric model with a semantic label allows agents to share the knowledge of the world with much less complexity, opening a door for multi-agent systems to perform complex tasking, and human–robot cooperation. This article takes the first step towards a generalized parametric model by limiting the geometric primitives to a planar surface and providing semantic labels when appropriate. Two novel compression algorithms for depth data and a method to independently fit planes to RGB-D data are provided, so that plane data can be used for real-time odometry estimation and mapping. Additionally, we extend maps with semantic information predicted from sparse geometries (planes) by a CNN. In experiments, the advantages of our approach in terms of computational and bandwidth resources savings are demonstrated and compared with other state-of-the-art SLAM systems.
Collapse
|
12
|
Depth-Image Segmentation Based on Evolving Principles for 3D Sensing of Structured Indoor Environments. SENSORS 2021; 21:s21134395. [PMID: 34198980 PMCID: PMC8271552 DOI: 10.3390/s21134395] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/20/2021] [Revised: 06/22/2021] [Accepted: 06/24/2021] [Indexed: 11/23/2022]
Abstract
This paper presents an approach of depth image segmentation based on the Evolving Principal Component Clustering (EPCC) method, which exploits data locality in an ordered data stream. The parameters of linear prototypes, which are used to describe different clusters, are estimated in a recursive manner. The main contribution of this work is the extension and application of the EPCC to 3D space for recursive and real-time detection of flat connected surfaces based on linear segments, which are all detected in an evolving way. To obtain optimal results when processing homogeneous surfaces, we introduced two-step filtering for outlier detection within a clustering framework and considered the noise model, which allowed for the compensation of characteristic uncertainties that are introduced into the measurements of depth sensors. The developed algorithm was compared with well-known methods for point cloud segmentation. The proposed approach achieves better segmentation results over longer distances for which the signal-to-noise ratio is low, without prior filtering of the data. On the given database, an average rate higher than 90% was obtained for successfully detected flat surfaces, which indicates high performance when processing huge point clouds in a non-iterative manner.
Collapse
|
13
|
Yu S, Fu C, Gostar AK, Hu M. A Review on Map-Merging Methods for Typical Map Types in Multiple-Ground-Robot SLAM Solutions. SENSORS 2020; 20:s20236988. [PMID: 33297376 PMCID: PMC7730201 DOI: 10.3390/s20236988] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/12/2020] [Revised: 11/28/2020] [Accepted: 12/04/2020] [Indexed: 11/29/2022]
Abstract
When multiple robots are involved in the process of simultaneous localization and mapping (SLAM), a global map should be constructed by merging the local maps built by individual robots, so as to provide a better representation of the environment. Hence, the map-merging methods play a crucial rule in multi-robot systems and determine the performance of multi-robot SLAM. This paper looks into the key problem of map merging for multiple-ground-robot SLAM and reviews the typical map-merging methods for several important types of maps in SLAM applications: occupancy grid maps, feature-based maps, and topological maps. These map-merging approaches are classified based on their working mechanism or the type of features they deal with. The concepts and characteristics of these map-merging methods are elaborated in this review. The contents summarized in this paper provide insights and guidance for future multiple-ground-robot SLAM solutions.
Collapse
Affiliation(s)
- Shuien Yu
- State Key Laboratory of Mechanical Transmissions, School of Automotive Engineering, Chongqing University, Chongqing 400044, China; (S.Y.); (M.H.)
| | - Chunyun Fu
- State Key Laboratory of Mechanical Transmissions, School of Automotive Engineering, Chongqing University, Chongqing 400044, China; (S.Y.); (M.H.)
- Correspondence:
| | - Amirali K. Gostar
- School of Engineering, RMIT University, Melbourne, VIC 3001, Australia;
| | - Minghui Hu
- State Key Laboratory of Mechanical Transmissions, School of Automotive Engineering, Chongqing University, Chongqing 400044, China; (S.Y.); (M.H.)
| |
Collapse
|
14
|
Li X, Li Y, Ornek EP, Lin J, Tombari F. Co-Planar Parametrization for Stereo-SLAM and Visual-Inertial Odometry. IEEE Robot Autom Lett 2020. [DOI: 10.1109/lra.2020.3027230] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
15
|
Liao Z, Wang W, Qi X, Zhang X. RGB-D Object SLAM Using Quadrics for Indoor Environments. SENSORS 2020; 20:s20185150. [PMID: 32917023 PMCID: PMC7571184 DOI: 10.3390/s20185150] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/31/2020] [Revised: 08/29/2020] [Accepted: 09/07/2020] [Indexed: 11/18/2022]
Abstract
Indoor service robots need to build an object-centric semantic map to understand and execute human instructions. Conventional visual simultaneous localization and mapping (SLAM) systems build a map using geometric features such as points, lines, and planes as landmarks. However, they lack a semantic understanding of the environment. This paper proposes an object-level semantic SLAM algorithm based on RGB-D data, which uses a quadric surface as an object model to compactly represent the object’s position, orientation, and shape. This paper proposes and derives two types of RGB-D camera-quadric observation models: a complete model and a partial model. The complete model combines object detection and point cloud data to estimate a complete ellipsoid in a single RGB-D frame. The partial model is activated when the depth data is severely missing because of illuminations or occlusions, which uses bounding boxes from object detection to constrain objects. Compared with the state-of-the-art quadric SLAM algorithms that use a monocular observation model, the RGB-D observation model reduces the requirements of the observation number and viewing angle changes, which helps improve the accuracy and robustness. This paper introduces a nonparametric pose graph to solve data associations in the back end, and innovatively applies it to the quadric surface model. We thoroughly evaluated the algorithm on two public datasets and an author-collected mobile robot dataset in a home-like environment. We obtained obvious improvements on the localization accuracy and mapping effects compared with two state-of-the-art object SLAM algorithms.
Collapse
Affiliation(s)
| | - Wei Wang
- Correspondence: ; Tel.: +86-010-8231-4554
| | | | | |
Collapse
|
16
|
Object Semantic Grid Mapping with 2D LiDAR and RGB-D Camera for Domestic Robot Navigation. APPLIED SCIENCES-BASEL 2020. [DOI: 10.3390/app10175782] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Occupied grid maps are sufficient for mobile robots to complete metric navigation tasks in domestic environments. However, they lack semantic information to endow the robots with the ability of social goal selection and human-friendly operation modes. In this paper, we propose an object semantic grid mapping system with 2D Light Detection and Ranging (LiDAR) and RGB-D sensors to solve this problem. At first, we use a laser-based Simultaneous Localization and Mapping (SLAM) to generate an occupied grid map and obtain a robot trajectory. Then, we employ object detection to get an object’s semantics of color images and use joint interpolation to refine camera poses. Based on object detection, depth images, and interpolated poses, we build a point cloud with object instances. To generate object-oriented minimum bounding rectangles, we propose a method for extracting the dominant directions of the room. Furthermore, we build object goal spaces to help the robots select navigation goals conveniently and socially. We have used the Robot@Home dataset to verify the system; the verification results show that our system is effective.
Collapse
|
17
|
|