1
|
Xi M, Yang J, Wen J, Li Z, Lu W, Gao X. An Information-Assisted Deep Reinforcement Learning Path Planning Scheme for Dynamic and Unknown Underwater Environment. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:842-853. [PMID: 37988205 DOI: 10.1109/tnnls.2023.3332172] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/23/2023]
Abstract
An autonomous underwater vehicle (AUV) has shown impressive potential and promising exploitation prospects in numerous marine missions. Among its various applications, the most essential prerequisite is path planning. Although considerable endeavors have been made, there are several limitations. A complete and realistic ocean simulation environment is critically needed. As most of the existing methods are based on mathematical models, they suffer from a large gap with reality. At the same time, the dynamic and unknown environment places high demands on robustness and generalization. In order to overcome these limitations, we propose an information-assisted reinforcement learning path planning scheme. First, it performs numerical modeling based on real ocean current observations to establish a complete simulation environment with the grid method, including 3-D terrain, dynamic currents, local information, and so on. Next, we propose an information compression (IC) scheme to trim the mutual information (MI) between reinforcement learning neural network layers to improve generalization. A proof based on information theory provides solid support for this. Moreover, for the dynamic characteristics of the marine environment, we elaborately design a confidence evaluator (CE), which evaluates the correlation between two adjacent frames of ocean currents to provide confidence for the action. The performance of our method has been evaluated and proven by numerical results, which demonstrate a fair sensitivity to ocean currents and high robustness and generalization to cope with the dynamic and unknown underwater environment.
Collapse
|
2
|
Zhao C, Yan H, Gao D. ADHDP-based robust self-learning 3D trajectory tracking control for underactuated UUVs. PeerJ Comput Sci 2024; 10:e2605. [PMID: 39896383 PMCID: PMC11784788 DOI: 10.7717/peerj-cs.2605] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2024] [Accepted: 11/21/2024] [Indexed: 02/04/2025]
Abstract
In this work, we propose a robust self-learning control scheme based on action-dependent heuristic dynamic programming (ADHDP) to tackle the 3D trajectory tracking control problem of underactuated uncrewed underwater vehicles (UUVs) with uncertain dynamics and time-varying ocean disturbances. Initially, the radial basis function neural network is introduced to convert the compound uncertain element, comprising uncertain dynamics and time-varying ocean disturbances, into a linear parametric form with just one unknown parameter. Then, to improve the tracking performance of the UUVs trajectory tracking closed-loop control system, an actor-critic neural network structure based on ADHDP technology is introduced to adaptively adjust the weights of the action-critic network, optimizing the performance index function. Finally, an ADHDP-based robust self-learning control scheme is constructed, which makes the UUVs closed-loop system have good robustness and control performance. The theoretical analysis demonstrates that all signals in the UUVs trajectory tracking closed-loop control system are bounded. The simulation results for the UUVs validate the effectiveness of the proposed control scheme.
Collapse
Affiliation(s)
- Chunbo Zhao
- Merchant Marine College, Shanghai Maritime University, Shanghai, China
| | - Huaran Yan
- Merchant Marine College, Shanghai Maritime University, Shanghai, China
| | - Deyi Gao
- Merchant Marine College, Shanghai Maritime University, Shanghai, China
| |
Collapse
|
3
|
Treesatayapun C. Discrete-Time Reinforcement Learning Adaptive Control for Non-Gaussian Distribution of Sampling Intervals. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:13453-13460. [PMID: 37204951 DOI: 10.1109/tnnls.2023.3269441] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/21/2023]
Abstract
This article proposes an optimal controller based on reinforcement learning (RL) for a class of unknown discrete-time systems with non-Gaussian distribution of sampling intervals. The critic and actor networks are implemented using the MiFRENc and MiFRENa architectures, respectively. The learning algorithm is developed with learning rates determined through convergence analysis of internal signals and tracking errors. Experimental systems with a comparative controller are conducted to validate the proposed scheme, and comparative results show superior performance for non-Gaussian distributions, with weight transfer for the critic network omitted. Additionally, the proposed learning laws, using the estimated co-state, significantly improve dead-zone compensation and nonlinear variation.
Collapse
|
4
|
Chen L, Dai SL, Dong C. Adaptive Optimal Tracking Control of an Underactuated Surface Vessel Using Actor-Critic Reinforcement Learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:7520-7533. [PMID: 36449582 DOI: 10.1109/tnnls.2022.3214681] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/17/2023]
Abstract
In this article, we present an adaptive reinforcement learning optimal tracking control (RLOTC) algorithm for an underactuated surface vessel subject to modeling uncertainties and time-varying external disturbances. By integrating backstepping technique with the optimized control design, we show that the desired optimal tracking performance of vessel control is guaranteed due to the fact that the virtual and actual control inputs are designed as optimized solutions of every subsystem. To enhance the robustness of vessel control systems, we employ neural network (NN) approximators to approximate uncertain vessel dynamics and present adaptive control technique to estimate the upper boundedness of external disturbances. Under the reinforcement learning framework, we construct actor-critic networks to solve the Hamilton-Jacobi-Bellman equations corresponding to subsystems of surface vessel to achieve the optimized control. The optimized control algorithm can synchronously train the adaptive parameters not only for actor-critic networks but also for NN approximators and adaptive control. By Lyapunov stability theorem, we show that the RLOTC algorithm can ensure the semiglobal uniform ultimate boundedness of the closed-loop systems. Compared with the existing reinforcement learning control results, the presented RLOTC algorithm can compensate for uncertain vessel dynamics and unknown disturbances, and obtain the optimized control performance by considering optimization in every backstepping design. Simulation studies on an underactuated surface vessel are given to illustrate the effectiveness of the RLOTC algorithm.
Collapse
|
5
|
Ding F, Wang R, Zhang T, Zheng G, Wu Z, Wang S. Real-time Trajectory Planning and Tracking Control of Bionic Underwater Robot in Dynamic Environment. CYBORG AND BIONIC SYSTEMS 2024; 5:0112. [PMID: 38725972 PMCID: PMC11079444 DOI: 10.34133/cbsystems.0112] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2023] [Accepted: 03/14/2024] [Indexed: 05/12/2024] Open
Abstract
In this article, we study the trajectory planning and tracking control of a bionic underwater robot under multiple dynamic obstacles. We first introduce the design of the bionic leopard cabinet underwater robot developed in our lab. Then, we model the trajectory planning problem of the bionic underwater robot by combining its dynamics and physical constraints. Furthermore, we conduct global trajectory planning for bionic underwater robots based on the temporal-spatial Bezier curves. In addition, based on the improved proximal policy optimization, local dynamic obstacle avoidance trajectory replanning is carried out. In addition, we design the fuzzy proportional-integral-derivative controller for tracking control of the planned trajectory. Finally, the effectiveness of the real-time trajectory planning and tracking control method is verified by comparative simulation in dynamic environment and semiphysical simulation of UWSim. Among them, the real-time trajectory planning method has advantages in trajectory length, trajectory smoothness, and planning time. The error of trajectory tracking control method is controlled around 0.2 m.
Collapse
Affiliation(s)
- Feng Ding
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation,
Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence,
University of Chinese Academy of Sciences, Beijing, China
| | - Rui Wang
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation,
Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence,
University of Chinese Academy of Sciences, Beijing, China
| | - Tiandong Zhang
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation,
Chinese Academy of Sciences, Beijing, China
| | - Gang Zheng
- Centrale Lille, CRIStAL-Centre de Recherche en Informatique Signal et Automatique de Lille,
University of Lille, 59000 Lille, France
| | - Zhengxing Wu
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation,
Chinese Academy of Sciences, Beijing, China
| | - Shuo Wang
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation,
Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence,
University of Chinese Academy of Sciences, Beijing, China
- Center for Excellence in Brain Science and Intelligence Technology,
Chinese Academy of Sciences, Shanghai, China
| |
Collapse
|
6
|
Yu Z, Li J, Xu Y, Zhang Y, Jiang B, Su CY. Reinforcement Learning-Based Fractional-Order Adaptive Fault-Tolerant Formation Control of Networked Fixed-Wing UAVs With Prescribed Performance. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:3365-3379. [PMID: 37310817 DOI: 10.1109/tnnls.2023.3281403] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
This article investigates the fault-tolerant formation control (FTFC) problem for networked fixed-wing unmanned aerial vehicles (UAVs) against faults. To constrain the distributed tracking errors of follower UAVs with respect to neighboring UAVs in the presence of faults, finite-time prescribed performance functions (PPFs) are developed to transform the distributed tracking errors into a new set of errors by incorporating user-specified transient and steady-state requirements. Then, the critic neural networks (NNs) are developed to learn the long-term performance indices, which are used to evaluate the distributed tracking performance. Based on the generated critic NNs, actor NNs are designed to learn the unknown nonlinear terms. Moreover, to compensate for the reinforcement learning errors of actor-critic NNs, nonlinear disturbance observers (DOs) with skillfully constructed auxiliary learning errors are developed to facilitate the FTFC design. Furthermore, by using the Lyapunov stability analysis, it is shown that all follower UAVs can track the leader UAV with predesigned offsets, and the distributed tracking errors are finite-time convergent. Finally, comparative simulation results are presented to show the effectiveness of the proposed control scheme.
Collapse
|
7
|
Guo F, Xu H, Xu P, Guo Z. Design of a reinforcement learning-based intelligent car transfer planning system for parking lots. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2024; 21:1058-1081. [PMID: 38303454 DOI: 10.3934/mbe.2024044] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/03/2024]
Abstract
In this study, a car transfer planning system for parking lots was designed based on reinforcement learning. The car transfer planning system for parking lots is an intelligent parking management system that is designed by using reinforcement learning techniques. The system features autonomous decision-making, intelligent path planning and efficient resource utilization. And the problem is solved by constructing a Markov decision process and using a dynamic planning-based reinforcement learning algorithm. The system has the advantage of looking to the future and using reinforcement learning to maximize its expected returns. And this is in contrast to manual transfer planning which relies on traditional thinking. In the context of this paper on parking lots, the states of the two locations form a finite set. The system ultimately seeks to find a strategy that is beneficial to the long-term development of the operation. It aims to prioritize strategies that have positive impacts in the future, rather than those that are focused solely on short-term benefits. To evaluate strategies, as its basis the system relies on the expected return of a state from now to the future. This approach allows for a more comprehensive assessment of the potential outcomes and ensures the selection of strategies that align with long-term goals. Experimental results show that the system has high performance and robustness in the area of car transfer planning for parking lots. By using reinforcement learning techniques, parking lot management systems can make autonomous decisions and plan optimal paths to achieve efficient resource utilization and reduce parking time.
Collapse
Affiliation(s)
- Feng Guo
- School of Management Science and Engineering, Chongqing Technology and Business University, Chongqing 400067, China
| | - Haiyu Xu
- School of Artificial Intelligence, Chongqing Technology and Business University, Chongqing 400067, China
| | - Peng Xu
- School of Artificial Intelligence, Chongqing Technology and Business University, Chongqing 400067, China
| | - Zhiwei Guo
- School of Artificial Intelligence, Chongqing Technology and Business University, Chongqing 400067, China
| |
Collapse
|
8
|
Cheng Y, Huang L, Chen CLP, Wang X. Robust Actor-Critic With Relative Entropy Regulating Actor. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:9054-9063. [PMID: 35286268 DOI: 10.1109/tnnls.2022.3155483] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
The accurate estimation of Q-function and the enhancement of agent's exploration ability have always been challenges of off-policy actor-critic algorithms. To address the two concerns, a novel robust actor-critic (RAC) is developed in this article. We first derive a robust policy improvement mechanism (RPIM) by using the local optimal policy about the current estimated Q-function to guide policy improvement. By constraining the relative entropy between the new policy and the previous one in policy improvement, the proposed RPIM can enhance the stability of the policy update process. The theoretical analysis shows that the incentive to increase the policy entropy is endowed when the policy is updated, which is conducive to enhancing the exploration ability of agents. Then, RAC is developed by applying the proposed RPIM to regulate the actor improvement process. The developed RAC is proven to be convergent. Finally, the proposed RAC is evaluated on some continuous-action control tasks in the MuJoCo platform and the experimental results show that RAC outperforms several state-of-the-art reinforcement learning algorithms.
Collapse
|
9
|
Li Z, Wang M, Ma G. Adaptive optimal trajectory tracking control of AUVs based on reinforcement learning. ISA TRANSACTIONS 2023; 137:122-132. [PMID: 36522214 DOI: 10.1016/j.isatra.2022.12.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/27/2022] [Revised: 11/16/2022] [Accepted: 12/02/2022] [Indexed: 06/04/2023]
Abstract
In this paper, an adaptive model-free optimal reinforcement learning (RL) neural network (NN) control scheme based on filter error is proposed for the trajectory tracking control problem of an autonomous underwater vehicle (AUV) with input saturation. Generally, the optimal control is realized by solving the Hamilton-Jacobi-Bellman (HJB) equation. However, due to its inherent nonlinearity and complexity, the HJB equation of AUV dynamics is challenging to solve. To deal with this problem, an RL strategy based on an actor-critic framework is proposed to approximate the solution of the HJB equation, where actor and critic NNs are used to perform control behavior and evaluate control performance, respectively. In addition, for the AUV system with the second-order strict-feedback dynamic model, the optimal controller design method based on filtering errors is proposed for the first time to simplify the controller design and accelerate the response speed of the system. Then, to solve the model-dependent problem, an extended state observer (ESO) is designed to estimate the unknown nonlinear dynamics, and an adaptive law is designed to estimate the unknown model parameters. To deal with the input saturation, an auxiliary variable system is utilized in the control law. The strict Lyapunov analysis guarantees that all signals of the system are semi-global uniformly ultimately bounded (SGUUB). Finally, the superiority of the proposed method is verified by comparative experiments.
Collapse
Affiliation(s)
- Zhifu Li
- School of Mechanical and Electrical Engineering, Guangzhou University, Guangzhou, 510006, China.
| | - Ming Wang
- School of Mechanical and Electrical Engineering, Guangzhou University, Guangzhou, 510006, China
| | - Ge Ma
- School of Mechanical and Electrical Engineering, Guangzhou University, Guangzhou, 510006, China
| |
Collapse
|
10
|
Wang N, Chen T, Liu S, Wang R, Karimi HR, Lin Y. Deep Learning-based Visual Detection of Marine Organisms: A Survey. Neurocomputing 2023. [DOI: 10.1016/j.neucom.2023.02.018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/19/2023]
|
11
|
Oh K, Seo J. Development of a Sliding-Mode-Control-Based Path-Tracking Algorithm with Model-Free Adaptive Feedback Action for Autonomous Vehicles. SENSORS (BASEL, SWITZERLAND) 2022; 23:405. [PMID: 36617002 PMCID: PMC9824019 DOI: 10.3390/s23010405] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/11/2022] [Revised: 12/26/2022] [Accepted: 12/27/2022] [Indexed: 06/17/2023]
Abstract
This paper presents a sliding mode control (SMC)-based path-tracking algorithm for autonomous vehicles by considering model-free adaptive feedback actions. In autonomous vehicles, safe path tracking requires adaptive and robust control algorithms because driving environment and vehicle conditions vary in real time. In this study, the SMC was adopted as a robust control method to adjust the switching gain, taking into account the sliding surface and unknown uncertainty to make the control error zero. The sliding surface can be designed mathematically, but it is difficult to express the unknown uncertainty mathematically. Information of priori bounded uncertainties is needed to obtain closed-loop stability of the control system, and the unknown uncertainty can vary with changes in internal and external factors. In the literature, ongoing efforts have been made to overcome the limitation of losing control stability due to unknown uncertainty. This study proposes an integrated method of adaptive feedback control (AFC) and SMC that can adjust a bounded uncertainty. Some illustrative and representative examples, such as autonomous driving scenarios, are also provided to show the main properties of the designed integrated controller. The examples show superior control performance, and it is expected that the integrated controller could be widely used for the path-tracking algorithms of autonomous vehicles.
Collapse
Affiliation(s)
- Kwangseok Oh
- School of ICT, Robotics & Mechanical Engineering, Hankyong National University, Anseong-si 17579, Republic of Korea
| | - Jaho Seo
- Department of Automotive and Mechatronics Engineering, Ontario Tech University, Oshawa, ON L1G 0C5, Canada
| |
Collapse
|
12
|
Jiang H, He X, Dong Shen QS. Decentralized Learning Control for Large-Scale Systems with Gain-Adaptation Mechanisms. Inf Sci (N Y) 2022. [DOI: 10.1016/j.ins.2022.12.043] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
|
13
|
Song D, Gan W, Yao P. Search and tracking strategy of autonomous surface underwater vehicle in oceanic eddies based on deep reinforcement learning. Appl Soft Comput 2022. [DOI: 10.1016/j.asoc.2022.109902] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/05/2022]
|
14
|
Tu Vu V, Pham TL, Dao PN. Disturbance observer-based adaptive reinforcement learning for perturbed uncertain surface vessels. ISA TRANSACTIONS 2022; 130:277-292. [PMID: 35450728 DOI: 10.1016/j.isatra.2022.03.027] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/03/2021] [Revised: 03/12/2022] [Accepted: 03/27/2022] [Indexed: 06/14/2023]
Abstract
This article considers a problem of tracking, convergence of disturbance observer (DO) based optimal control design for uncertain surface vessels (SVs) with external disturbance. The advantage of proposed optimal control using adaptive/approximate reinforcement learning (ARL) is that consideration for whole SVs with only one dynamic equation and without conventional separation technique. Additionally, thanks to appropriate disturbance observer, the attraction region of tracking error is remarkably reduced. On the other hand, the particular case of optimal control problem is presented by directly solving for the purpose of choosing the suitable activation functions of ARL. Furthermore, the proposed ARL based optimal control also deals with non-autonomous property of closed tracking error SV model by considering the equivalent system. Based on the Lyapunov function candidate using optimal function and quadratic form of estimated error of actor/critic weight, the stability and convergence of the closed system are proven. Some examples are given to verify and demonstrate the effectiveness of the new control strategy.
Collapse
Affiliation(s)
- Van Tu Vu
- Haiphong University, Haiphong, Viet Nam
| | - Thanh Loc Pham
- School of Electrical Engineering, Hanoi University of Science and Technology, Hanoi, Viet Nam
| | - Phuong Nam Dao
- School of Electrical Engineering, Hanoi University of Science and Technology, Hanoi, Viet Nam.
| |
Collapse
|
15
|
Yan L, Liu Z, Philip Chen C, Zhang Y, Wu Z. Optimized Adaptive Consensus Control for Multi-agent Systems with Prescribed Performance. Inf Sci (N Y) 2022. [DOI: 10.1016/j.ins.2022.08.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
16
|
Reinforcement-Learning-Based Tracking Control with Fixed-Time Prescribed Performance for Reusable Launch Vehicle under Input Constraints. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12157436] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
This paper proposes a novel reinforcement learning (RL)-based tracking control scheme with fixed-time prescribed performance for a reusable launch vehicle subject to parametric uncertainties, external disturbances, and input constraints. First, a fixed-time prescribed performance function is employed to restrain attitude tracking errors, and an equivalent unconstrained system is derived via an error transformation technique. Then, a hyperbolic tangent function is incorporated into the optimal performance index of the unconstrained system to tackle the input constraints. Subsequently, an actor-critic RL framework with super-twisting-like sliding mode control is constructed to establish a practical solution for the optimal control problem. Benefiting from the proposed scheme, the robustness of the RL-based controller against unknown dynamics is enhanced, and the control performance can be qualitatively prearranged by users. Theoretical analysis shows that the attitude tracking errors converge to a preset region within a preassigned fixed time, and the weight estimation errors of the actor-critic networks are uniformly ultimately bounded. Finally, comparative numerical simulation results are provided to illustrate the effectiveness and improved performance of the proposed control scheme.
Collapse
|
17
|
Wang N, Gao Y, Yang C, Zhang X. Reinforcement learning-based finite-time tracking control of an unknown unmanned surface vehicle with input constraints. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.04.133] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
18
|
A Novel Fixed-Time Trajectory Tracking Strategy of Unmanned Surface Vessel Based on the Fractional Sliding Mode Control Method. ELECTRONICS 2022. [DOI: 10.3390/electronics11050726] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
A novel sliding mode control method is proposed to achieve the trajectory tracking of the Unmanned Surface Vessel (USV) and effectively deal with the unmodeled dynamics and external unknown disturbances. First, a fixed-time fractional-order sliding mode control (FTFOSMC) strategy is proposed, combined with the fixed-time control theory and fractional-order control theory based on the sliding mode control method. The FTFOSMC strategy can improve the convergence velocity of the system, and effectively track the desired path, weakening the “chattering” effect in sliding mode control systems. Second, a fixed-time fractional-order sliding mode control strategy combined with the radial basis function neural network (RBF-FTFOSMC) was designed, which can effectively estimate the lumped uncertainties, such as the disturbance of external wind, wave, and current, and the unmodeled dynamics of the USV model. Then, the stability and effectiveness of the designed control strategy are guaranteed by the Lyapunov theory and the corresponding lemmas. Finally, a rigorous simulation experiment is designed to validate the effectiveness and stability of the proposed control strategy. The simulation results show that the control strategy can effectively achieve trajectory tracking of the USV, reduce the “chattering” phenomenon of sliding mode, and effectively estimate the lumped uncertainties.
Collapse
|