1
|
Song Y, Romero A, Müller M, Koltun V, Scaramuzza D. Reaching the limit in autonomous racing: Optimal control versus reinforcement learning. Sci Robot 2023; 8:eadg1462. [PMID: 37703383 DOI: 10.1126/scirobotics.adg1462] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2022] [Accepted: 05/23/2023] [Indexed: 09/15/2023]
Abstract
A central question in robotics is how to design a control system for an agile mobile robot. This paper studies this question systematically, focusing on a challenging setting: autonomous drone racing. We show that a neural network controller trained with reinforcement learning (RL) outperformed optimal control (OC) methods in this setting. We then investigated which fundamental factors have contributed to the success of RL or have limited OC. Our study indicates that the fundamental advantage of RL over OC is not that it optimizes its objective better but that it optimizes a better objective. OC decomposes the problem into planning and control with an explicit intermediate representation, such as a trajectory, that serves as an interface. This decomposition limits the range of behaviors that can be expressed by the controller, leading to inferior control performance when facing unmodeled effects. In contrast, RL can directly optimize a task-level objective and can leverage domain randomization to cope with model uncertainty, allowing the discovery of more robust control responses. Our findings allowed us to push an agile drone to its maximum performance, achieving a peak acceleration greater than 12 times the gravitational acceleration and a peak velocity of 108 kilometers per hour. Our policy achieved superhuman control within minutes of training on a standard workstation. This work presents a milestone in agile robotics and sheds light on the role of RL and OC in robot control.
Collapse
|
2
|
Kaufmann E, Bauersfeld L, Loquercio A, Müller M, Koltun V, Scaramuzza D. Champion-level drone racing using deep reinforcement learning. Nature 2023; 620:982-987. [PMID: 37648758 PMCID: PMC10468397 DOI: 10.1038/s41586-023-06419-4] [Citation(s) in RCA: 15] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2023] [Accepted: 07/10/2023] [Indexed: 09/01/2023]
Abstract
First-person view (FPV) drone racing is a televised sport in which professional competitors pilot high-speed aircraft through a 3D circuit. Each pilot sees the environment from the perspective of their drone by means of video streamed from an onboard camera. Reaching the level of professional pilots with an autonomous drone is challenging because the robot needs to fly at its physical limits while estimating its speed and location in the circuit exclusively from onboard sensors1. Here we introduce Swift, an autonomous system that can race physical vehicles at the level of the human world champions. The system combines deep reinforcement learning (RL) in simulation with data collected in the physical world. Swift competed against three human champions, including the world champions of two international leagues, in real-world head-to-head races. Swift won several races against each of the human champions and demonstrated the fastest recorded race time. This work represents a milestone for mobile robotics and machine intelligence2, which may inspire the deployment of hybrid learning-based solutions in other physical systems.
Collapse
Affiliation(s)
- Elia Kaufmann
- Robotics and Perception Group, University of Zurich, Zürich, Switzerland.
| | - Leonard Bauersfeld
- Robotics and Perception Group, University of Zurich, Zürich, Switzerland
| | - Antonio Loquercio
- Robotics and Perception Group, University of Zurich, Zürich, Switzerland
| | | | | | - Davide Scaramuzza
- Robotics and Perception Group, University of Zurich, Zürich, Switzerland
| |
Collapse
|
3
|
Pham HX, Sarabakha A, Odnoshyvkin M, Kayacan E. PencilNet: Zero-Shot Sim-to-Real Transfer Learning for Robust Gate Perception in Autonomous Drone Racing. IEEE Robot Autom Lett 2022. [DOI: 10.1109/lra.2022.3207545] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Affiliation(s)
- Huy Xuan Pham
- Artificial Intelligence in Robotics Laboratory (Air Lab), Department of Electrical and Computer Engineering, Aarhus University, Aarhus N, Denmark
| | - Andriy Sarabakha
- School of Electrical and Electronic Engineering (EEE), Nanyang Technological University (NTU), Singapore
| | - Mykola Odnoshyvkin
- School of Computation, Information and Technology, Technical University of Munich (TUM), Munich, Germany
| | - Erdal Kayacan
- Artificial Intelligence in Robotics Laboratory (Air Lab), Department of Electrical and Computer Engineering, Aarhus University, Aarhus N, Denmark
| |
Collapse
|
4
|
Wang Z, Chen C, Dong D. Lifelong Incremental Reinforcement Learning With Online Bayesian Inference. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022; 33:4003-4016. [PMID: 33571098 DOI: 10.1109/tnnls.2021.3055499] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
A central capability of a long-lived reinforcement learning (RL) agent is to incrementally adapt its behavior as its environment changes and to incrementally build upon previous experiences to facilitate future learning in real-world scenarios. In this article, we propose lifelong incremental reinforcement learning (LLIRL), a new incremental algorithm for efficient lifelong adaptation to dynamic environments. We develop and maintain a library that contains an infinite mixture of parameterized environment models, which is equivalent to clustering environment parameters in a latent space. The prior distribution over the mixture is formulated as a Chinese restaurant process (CRP), which incrementally instantiates new environment models without any external information to signal environmental changes in advance. During lifelong learning, we employ the expectation-maximization (EM) algorithm with online Bayesian inference to update the mixture in a fully incremental manner. In EM, the E-step involves estimating the posterior expectation of environment-to-cluster assignments, whereas the M-step updates the environment parameters for future learning. This method allows for all environment models to be adapted as necessary, with new models instantiated for environmental changes and old models retrieved when previously seen environments are encountered again. Simulation experiments demonstrate that LLIRL outperforms relevant existing methods and enables effective incremental adaptation to various dynamic environments for lifelong learning.
Collapse
|
5
|
Penicka R, Song Y, Kaufmann E, Scaramuzza D. Learning Minimum-Time Flight in Cluttered Environments. IEEE Robot Autom Lett 2022. [DOI: 10.1109/lra.2022.3181755] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Affiliation(s)
- Robert Penicka
- Robotics and Perception Group, Department of Informatics, University of Zurich, and Department of Neuroinformatics, University of Zurich and ETH Zurich, Switzerland
| | - Yunlong Song
- Robotics and Perception Group, Department of Informatics, University of Zurich, and Department of Neuroinformatics, University of Zurich and ETH Zurich, Switzerland
| | - Elia Kaufmann
- Robotics and Perception Group, Department of Informatics, University of Zurich, and Department of Neuroinformatics, University of Zurich and ETH Zurich, Switzerland
| | - Davide Scaramuzza
- Robotics and Perception Group, Department of Informatics, University of Zurich, and Department of Neuroinformatics, University of Zurich and ETH Zurich, Switzerland
| |
Collapse
|
6
|
Cheng Y, Zhao P, Wang F, Block DJ, Hovakimyan N. Improving the Robustness of Reinforcement Learning Policies With ${\mathcal {L}_{1}}$ Adaptive Control. IEEE Robot Autom Lett 2022. [DOI: 10.1109/lra.2022.3169309] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Affiliation(s)
- Yikun Cheng
- Mechanical Science and Engineering Department, University of Illinois at Urbana-Champaign, IL, USA
| | - Pan Zhao
- Mechanical Science and Engineering Department, University of Illinois at Urbana-Champaign, IL, USA
| | - Fanxin Wang
- Mechanical Science and Engineering Department, University of Illinois at Urbana-Champaign, IL, USA
| | - Daniel J. Block
- Electrical and Computer Engineering Department, University of Illinois at Urbana-Champaign, IL, USA
| | - Naira Hovakimyan
- Mechanical Science and Engineering Department, University of Illinois at Urbana-Champaign, IL, USA
| |
Collapse
|
7
|
Lopez M, Martinez-Carranza J. A CNN-based Approach for Cable-Suspended Load Lifting with an Autonomous MAV. J INTELL ROBOT SYST 2022. [DOI: 10.1007/s10846-022-01637-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
8
|
Schlachter K, Felsner K, Zambal S. Training neural networks on domain randomized simulations for ultrasonic inspection. OPEN RESEARCH EUROPE 2022; 2:43. [PMID: 37645298 PMCID: PMC10446096 DOI: 10.12688/openreseurope.14358.2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 05/09/2022] [Indexed: 08/31/2023]
Abstract
To overcome the data scarcity problem of machine learning for nondestructive testing, data augmentation is a commonly used strategy. We propose a method to enable training of neural networks exclusively on simulated data. Simulations not only provide a scalable way to generate and access training data, but also make it possible to cover edge cases which rarely appear in the real world. However, simulating data acquired from complex nondestructive testing methods is still a challenging task. Due to necessary simplifications and a limited accuracy of parameter identification, statistical models trained solely on simulated data often generalize poorly to the real world. Some effort has been made in the field to adapt pre-trained classifiers with a small set of real world data. A different approach for bridging the reality gap is domain randomization which was recently very successfully applied in different fields of autonomous robotics. In this study, we apply this approach for ultrasonic testing of carbon-fiber-reinforced plastics. Phased array captures of virtual specimens are simulated by approximating sound propagation via ray tracing. In addition to a variation of the geometric model of the specimen and its defects, we vary simulation parameters. Results indicate that this approach allows a generalization to the real world without applying any domain adaptation. Further, the trained network distinguishes correctly between ghost artifacts and defects. Although this study is tailored towards evaluation of ultrasound phased array captures, the presented approach generalizes to other nondestructive testing methods.
Collapse
|
9
|
Pfeiffer C, Wengeler S, Loquercio A, Scaramuzza D. Visual attention prediction improves performance of autonomous drone racing agents. PLoS One 2022; 17:e0264471. [PMID: 35231038 PMCID: PMC8887736 DOI: 10.1371/journal.pone.0264471] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2021] [Accepted: 02/10/2022] [Indexed: 11/18/2022] Open
Abstract
Humans race drones faster than neural networks trained for end-to-end autonomous flight. This may be related to the ability of human pilots to select task-relevant visual information effectively. This work investigates whether neural networks capable of imitating human eye gaze behavior and attention can improve neural networks' performance for the challenging task of vision-based autonomous drone racing. We hypothesize that gaze-based attention prediction can be an efficient mechanism for visual information selection and decision making in a simulator-based drone racing task. We test this hypothesis using eye gaze and flight trajectory data from 18 human drone pilots to train a visual attention prediction model. We then use this visual attention prediction model to train an end-to-end controller for vision-based autonomous drone racing using imitation learning. We compare the drone racing performance of the attention-prediction controller to those using raw image inputs and image-based abstractions (i.e., feature tracks). Comparing success rates for completing a challenging race track by autonomous flight, our results show that the attention-prediction based controller (88% success rate) outperforms the RGB-image (61% success rate) and feature-tracks (55% success rate) controller baselines. Furthermore, visual attention-prediction and feature-track based models showed better generalization performance than image-based models when evaluated on hold-out reference trajectories. Our results demonstrate that human visual attention prediction improves the performance of autonomous vision-based drone racing agents and provides an essential step towards vision-based, fast, and agile autonomous flight that eventually can reach and even exceed human performances.
Collapse
Affiliation(s)
- Christian Pfeiffer
- Robotics and Perception Group, Department of Informatics, University of Zurich, Zurich, Switzerland
- Department of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
- * E-mail:
| | - Simon Wengeler
- Robotics and Perception Group, Department of Informatics, University of Zurich, Zurich, Switzerland
- Department of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
| | - Antonio Loquercio
- Robotics and Perception Group, Department of Informatics, University of Zurich, Zurich, Switzerland
- Department of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
| | - Davide Scaramuzza
- Robotics and Perception Group, Department of Informatics, University of Zurich, Zurich, Switzerland
- Department of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland
| |
Collapse
|
10
|
Adamkiewicz M, Chen T, Caccavale A, Gardner R, Culbertson P, Bohg J, Schwager M. Vision-Only Robot Navigation in a Neural Radiance World. IEEE Robot Autom Lett 2022. [DOI: 10.1109/lra.2022.3150497] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
11
|
Time-Optimal Online Replanning for Agile Quadrotor Flight. IEEE Robot Autom Lett 2022. [DOI: 10.1109/lra.2022.3185772] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
12
|
Song Y, Scaramuzza D. Policy Search for Model Predictive Control With Application to Agile Drone Flight. IEEE T ROBOT 2022. [DOI: 10.1109/tro.2022.3141602] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
13
|
Rojas-Perez LO, Martinez-Carranza J. Towards Autonomous Drone Racing without GPU Using an OAK-D Smart Camera. SENSORS (BASEL, SWITZERLAND) 2021; 21:7436. [PMID: 34833511 PMCID: PMC8620925 DOI: 10.3390/s21227436] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/26/2021] [Revised: 10/27/2021] [Accepted: 11/02/2021] [Indexed: 11/24/2022]
Abstract
Recent advances have shown for the first time that it is possible to beat a human with an autonomous drone in a drone race. However, this solution relies heavily on external sensors, specifically on the use of a motion capture system. Thus, a truly autonomous solution demands performing computationally intensive tasks such as gate detection, drone localisation, and state estimation. To this end, other solutions rely on specialised hardware such as graphics processing units (GPUs) whose onboard hardware versions are not as powerful as those available for desktop and server computers. An alternative is to combine specialised hardware with smart sensors capable of processing specific tasks on the chip, alleviating the need for the onboard processor to perform these computations. Motivated by this, we present the initial results of adapting a novel smart camera, known as the OpenCV AI Kit or OAK-D, as part of a solution for the ADR running entirely on board. This smart camera performs neural inference on the chip that does not use a GPU. It can also perform depth estimation with a stereo rig and run neural network models using images from a 4K colour camera as the input. Additionally, seeking to limit the payload to 200 g, we present a new 3D-printed design of the camera's back case, reducing the original weight 40%, thus enabling the drone to carry it in tandem with a host onboard computer, the Intel Stick compute, where we run a controller based on gate detection. The latter is performed with a neural model running on an OAK-D at an operation frequency of 40 Hz, enabling the drone to fly at a speed of 2 m/s. We deem these initial results promising toward the development of a truly autonomous solution that will run intensive computational tasks fully on board.
Collapse
Affiliation(s)
| | - Jose Martinez-Carranza
- Department of Computational Science, Instituto Nacional de Astrofísica, Óptica y Electrónica (INAOE), Puebla 72840, Mexico;
| |
Collapse
|
14
|
Loquercio A, Kaufmann E, Ranftl R, Müller M, Koltun V, Scaramuzza D. Learning high-speed flight in the wild. Sci Robot 2021; 6:eabg5810. [PMID: 34613820 DOI: 10.1126/scirobotics.abg5810] [Citation(s) in RCA: 26] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
Abstract
Quadrotors are agile. Unlike most other machines, they can traverse extremely complex environments at high speeds. To date, only expert human pilots have been able to fully exploit their capabilities. Autonomous operation with onboard sensing and computation has been limited to low speeds. State-of-the-art methods generally separate the navigation problem into subtasks: sensing, mapping, and planning. Although this approach has proven successful at low speeds, the separation it builds upon can be problematic for high-speed navigation in cluttered environments. The subtasks are executed sequentially, leading to increased processing latency and a compounding of errors through the pipeline. Here, we propose an end-to-end approach that can autonomously fly quadrotors through complex natural and human-made environments at high speeds with purely onboard sensing and computation. The key principle is to directly map noisy sensory observations to collision-free trajectories in a receding-horizon fashion. This direct mapping drastically reduces processing latency and increases robustness to noisy and incomplete perception. The sensorimotor mapping is performed by a convolutional network that is trained exclusively in simulation via privileged learning: imitating an expert with access to privileged information. By simulating realistic sensor noise, our approach achieves zero-shot transfer from simulation to challenging real-world environments that were never experienced during training: dense forests, snow-covered terrain, derailed trains, and collapsed buildings. Our work demonstrates that end-to-end policies trained in simulation enable high-speed autonomous flight through challenging environments, outperforming traditional obstacle avoidance pipelines.
Collapse
|
15
|
Rezende AMC, Miranda VRF, Rezeck PAF, Azpúrua H, Santos ERS, Gonçalves VM, Macharet DG, Freitas GM. An integrated solution for an autonomous drone racing in indoor environments. INTEL SERV ROBOT 2021. [DOI: 10.1007/s11370-021-00385-4] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
16
|
Valtchev SZ, Wu J. Domain randomization for neural network classification. JOURNAL OF BIG DATA 2021; 8:94. [PMID: 34760433 PMCID: PMC8570318 DOI: 10.1186/s40537-021-00455-5] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/14/2020] [Accepted: 05/03/2021] [Indexed: 06/13/2023]
Abstract
Large data requirements are often the main hurdle in training neural networks. Convolutional neural network (CNN) classifiers in particular require tens of thousands of pre-labeled images per category to approach human-level accuracy, while often failing to generalized to out-of-domain test sets. The acquisition and labelling of such datasets is often an expensive, time consuming and tedious task in practice. Synthetic data provides a cheap and efficient solution to assemble such large datasets. Using domain randomization (DR), we show that a sufficiently well generated synthetic image dataset can be used to train a neural network classifier that rivals state-of-the-art models trained on real datasets, achieving accuracy levels as high as 88% on a baseline cats vs dogs classification task. We show that the most important domain randomization parameter is a large variety of subjects, while secondary parameters such as lighting and textures are found to be less significant to the model accuracy. Our results also provide evidence to suggest that models trained on domain randomized images transfer to new domains better than those trained on real photos. Model performance appears to remain stable as the number of categories increases.
Collapse
Affiliation(s)
- Svetozar Zarko Valtchev
- Laboratory of Industrial and Applied Mathematics, York University, 4700 Keele St, M3J 1P3 Toronto, ON Canada
| | - Jianhong Wu
- Laboratory of Industrial and Applied Mathematics, York University, 4700 Keele St, M3J 1P3 Toronto, ON Canada
| |
Collapse
|
17
|
Abstract
With the rise of Deep Learning approaches in computer vision applications, significant strides have been made towards vehicular autonomy. Research activity in autonomous drone navigation has increased rapidly in the past five years, and drones are moving fast towards the ultimate goal of near-complete autonomy. However, while much work in the area focuses on specific tasks in drone navigation, the contribution to the overall goal of autonomy is often not assessed, and a comprehensive overview is needed. In this work, a taxonomy of drone navigation autonomy is established by mapping the definitions of vehicular autonomy levels, as defined by the Society of Automotive Engineers, to specific drone tasks in order to create a clear definition of autonomy when applied to drones. A top–down examination of research work in the area is conducted, focusing on drone navigation tasks, in order to understand the extent of research activity in each area. Autonomy levels are cross-checked against the drone navigation tasks addressed in each work to provide a framework for understanding the trajectory of current research. This work serves as a guide to research in drone autonomy with a particular focus on Deep Learning-based solutions, indicating key works and areas of opportunity for development of this area in the future.
Collapse
|
18
|
Hayat S, Jung R, Hellwagner H, Bettstetter C, Emini D, Schnieders D. Edge Computing in 5G for Drone Navigation: What to Offload? IEEE Robot Autom Lett 2021. [DOI: 10.1109/lra.2021.3062319] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
19
|
Cohen MR, Abdulrahim K, Forbes JR. Finite-Horizon LQR Control of Quadrotors on $SE_2(3)$. IEEE Robot Autom Lett 2020. [DOI: 10.1109/lra.2020.3010214] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
20
|
Cazzato D, Cimarelli C, Sanchez-Lopez JL, Voos H, Leo M. A Survey of Computer Vision Methods for 2D Object Detection from Unmanned Aerial Vehicles. J Imaging 2020; 6:jimaging6080078. [PMID: 34460693 PMCID: PMC8321148 DOI: 10.3390/jimaging6080078] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2020] [Revised: 07/27/2020] [Accepted: 07/31/2020] [Indexed: 11/16/2022] Open
Abstract
The spread of Unmanned Aerial Vehicles (UAVs) in the last decade revolutionized many applications fields. Most investigated research topics focus on increasing autonomy during operational campaigns, environmental monitoring, surveillance, maps, and labeling. To achieve such complex goals, a high-level module is exploited to build semantic knowledge leveraging the outputs of the low-level module that takes data acquired from multiple sensors and extracts information concerning what is sensed. All in all, the detection of the objects is undoubtedly the most important low-level task, and the most employed sensors to accomplish it are by far RGB cameras due to costs, dimensions, and the wide literature on RGB-based object detection. This survey presents recent advancements in 2D object detection for the case of UAVs, focusing on the differences, strategies, and trade-offs between the generic problem of object detection, and the adaptation of such solutions for operations of the UAV. Moreover, a new taxonomy that considers different heights intervals and driven by the methodological approaches introduced by the works in the state of the art instead of hardware, physical and/or technological constraints is proposed.
Collapse
Affiliation(s)
- Dario Cazzato
- Interdisciplinary Center for Security, Reliability and Trust (SnT), University of Luxembourg, 1855 Luxembourg, Luxembourg; (C.C.); (J.L.S.-L.); (H.V.)
- Correspondence:
| | - Claudio Cimarelli
- Interdisciplinary Center for Security, Reliability and Trust (SnT), University of Luxembourg, 1855 Luxembourg, Luxembourg; (C.C.); (J.L.S.-L.); (H.V.)
| | - Jose Luis Sanchez-Lopez
- Interdisciplinary Center for Security, Reliability and Trust (SnT), University of Luxembourg, 1855 Luxembourg, Luxembourg; (C.C.); (J.L.S.-L.); (H.V.)
| | - Holger Voos
- Interdisciplinary Center for Security, Reliability and Trust (SnT), University of Luxembourg, 1855 Luxembourg, Luxembourg; (C.C.); (J.L.S.-L.); (H.V.)
| | - Marco Leo
- Institute of Applied Sciences and Intelligent Systems, National Research Council of Italy, 73100 Lecce, Italy;
| |
Collapse
|
21
|
UAV-Based Structural Damage Mapping: A Review. ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION 2019. [DOI: 10.3390/ijgi9010014] [Citation(s) in RCA: 66] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Structural disaster damage detection and characterization is one of the oldest remote sensing challenges, and the utility of virtually every type of active and passive sensor deployed on various air- and spaceborne platforms has been assessed. The proliferation and growing sophistication of unmanned aerial vehicles (UAVs) in recent years has opened up many new opportunities for damage mapping, due to the high spatial resolution, the resulting stereo images and derivatives, and the flexibility of the platform. This study provides a comprehensive review of how UAV-based damage mapping has evolved from providing simple descriptive overviews of a disaster science, to more sophisticated texture and segmentation-based approaches, and finally to studies using advanced deep learning approaches, as well as multi-temporal and multi-perspective imagery to provide comprehensive damage descriptions. The paper further reviews studies on the utility of the developed mapping strategies and image processing pipelines for first responders, focusing especially on outcomes of two recent European research projects, RECONASS (Reconstruction and Recovery Planning: Rapid and Continuously Updated Construction Damage, and Related Needs Assessment) and INACHUS (Technological and Methodological Solutions for Integrated Wide Area Situation Awareness and Survivor Localization to Support Search and Rescue Teams). Finally, recent and emerging developments are reviewed, such as recent improvements in machine learning, increasing mapping autonomy, damage mapping in interior, GPS-denied environments, the utility of UAVs for infrastructure mapping and maintenance, as well as the emergence of UAVs with robotic abilities.
Collapse
|