• Reference Citation Analysis
  • v
  • v
  • Find an Article
Find an Article PDF (4711324)   Today's Articles (5434)
For: Modares H, Lewis FL. Optimal Output-Feedback Control of Unknown Continuous-Time Linear Systems Using Off-policy Reinforcement Learning. IEEE Trans Cybern 2016;46:2401-2410. [PMID: 28113995 DOI: 10.1109/tcyb.2015.2477810] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [What about the content of this article? (0)] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Number Cited by Other Article(s)
1
Wu H, Hu Q, Zheng J, Dong F, Ouyang Z, Li D. Discounted Inverse Reinforcement Learning for Linear Quadratic Control. IEEE TRANSACTIONS ON CYBERNETICS 2025;55:1995-2007. [PMID: 40036510 DOI: 10.1109/tcyb.2025.3540967] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2025]
2
Wang J, Wu J, Cao J, Chadli M, Shen H. Nonfragile Output Feedback Tracking Control for Markov Jump Fuzzy Systems Based on Integral Reinforcement Learning Scheme. IEEE TRANSACTIONS ON CYBERNETICS 2023;53:4521-4530. [PMID: 36194715 DOI: 10.1109/tcyb.2022.3203795] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
3
Arogeti SA, Lewis FL. Static Output-Feedback H Control Design Procedures for Continuous-Time Systems With Different Levels of Model Knowledge. IEEE TRANSACTIONS ON CYBERNETICS 2023;53:1432-1446. [PMID: 34570712 DOI: 10.1109/tcyb.2021.3103148] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
4
Mazouchi M, Yang Y, Modares H. Data-Driven Dynamic Multiobjective Optimal Control: An Aspiration-Satisfying Reinforcement Learning Approach. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2022;33:6183-6193. [PMID: 33886483 DOI: 10.1109/tnnls.2021.3072571] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
5
Cheng Y, Huang L, Wang X. Authentic Boundary Proximal Policy Optimization. IEEE TRANSACTIONS ON CYBERNETICS 2022;52:9428-9438. [PMID: 33705327 DOI: 10.1109/tcyb.2021.3051456] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
6
Wang N, Gao Y, Yang C, Zhang X. Reinforcement learning-based finite-time tracking control of an unknown unmanned surface vehicle with input constraints. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.04.133] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
7
Duan K, Fong S, Chen CP. Reinforcement learning based model-free optimized trajectory tracking strategy design for an AUV. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.10.056] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
8
Zhou P, Zhao W, Li J, Li A, Du W, Wen S. Massive Maritime Path Planning: A Contextual Online Learning Approach. IEEE TRANSACTIONS ON CYBERNETICS 2021;51:6262-6273. [PMID: 32112685 DOI: 10.1109/tcyb.2019.2959543] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
9
Integral reinforcement learning-based optimal output feedback control for linear continuous-time systems with input delay. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.06.073] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
10
Hao Y, Wang T, Li G, Wen C. Linear Quadratic Optimal Control of Time-Invariant Linear Networks With Selectable Input Matrix. IEEE TRANSACTIONS ON CYBERNETICS 2021;51:4743-4754. [PMID: 31804949 DOI: 10.1109/tcyb.2019.2953218] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
11
Calafiore GC, Possieri C. Output Feedback Q-Learning for Linear-Quadratic Discrete-Time Finite-Horizon Control Problems. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2021;32:3274-3281. [PMID: 32745011 DOI: 10.1109/tnnls.2020.3010304] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
12
Na J, Zhao J, Gao G, Li Z. Output-Feedback Robust Control of Uncertain Systems via Online Data-Driven Learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2021;32:2650-2662. [PMID: 32706646 DOI: 10.1109/tnnls.2020.3007414] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
13
Yang X, He H, Zhong X. Approximate Dynamic Programming for Nonlinear-Constrained Optimizations. IEEE TRANSACTIONS ON CYBERNETICS 2021;51:2419-2432. [PMID: 31329149 DOI: 10.1109/tcyb.2019.2926248] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
14
Zhu P, Zeng J. Observer-based control for nonlinear parameter-varying systems: A sum-of-squares approach. ISA TRANSACTIONS 2021;111:121-131. [PMID: 33220944 DOI: 10.1016/j.isatra.2020.11.010] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/23/2020] [Revised: 10/21/2020] [Accepted: 11/06/2020] [Indexed: 06/11/2023]
15
Ballesteros M, Chairez I, Poznyak A. Robust optimal feedback control design for uncertain systems based on artificial neural network approximation of the Bellman’s value function. Neurocomputing 2020. [DOI: 10.1016/j.neucom.2020.06.085] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
16
Huang M, Liu C, He X, Ma L, Lu Z, Su H. Reinforcement Learning-Based Control for Nonlinear Discrete-Time Systems with Unknown Control Directions and Control Constraints. Neurocomputing 2020. [DOI: 10.1016/j.neucom.2020.03.061] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
17
Zhou Y, Wang H, Li L, Lian J. Bench calibration method for automotive electric motors based on deep reinforcement learning. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2020. [DOI: 10.3233/jifs-191567] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
18
Fuzzy Reinforcement Learning and Curriculum Transfer Learning for Micromanagement in Multi-Robot Confrontation. INFORMATION 2019. [DOI: 10.3390/info10110341] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]  Open
19
Banerjee S, Chatterjee A. ALERA. ACM T INTEL SYST TEC 2019. [DOI: 10.1145/3338123] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
20
Rizvi SAA, Lin Z. Output Feedback Q-Learning Control for the Discrete-Time Linear Quadratic Regulator Problem. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2019;30:1523-1536. [PMID: 30296242 DOI: 10.1109/tnnls.2018.2870075] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
21
Rizvi SAA, Lin Z. Reinforcement Learning-Based Linear Quadratic Regulation of Continuous-Time Systems Using Dynamic Output Feedback. IEEE TRANSACTIONS ON CYBERNETICS 2019;50:4670-4679. [PMID: 30605117 DOI: 10.1109/tcyb.2018.2886735] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
22
Training a robust reinforcement learning controller for the uncertain system based on policy gradient method. Neurocomputing 2018. [DOI: 10.1016/j.neucom.2018.08.007] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
23
Li X, Xue L, Sun C. Linear quadratic tracking control of unknown discrete-time systems using value iteration algorithm. Neurocomputing 2018. [DOI: 10.1016/j.neucom.2018.05.111] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
24
Narayanan V, Jagannathan S. Event-Triggered Distributed Control of Nonlinear Interconnected Systems Using Online Reinforcement Learning With Exploration. IEEE TRANSACTIONS ON CYBERNETICS 2018;48:2510-2519. [PMID: 28885167 DOI: 10.1109/tcyb.2017.2741342] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
25
Qi Q, Zhang H. Output Feedback Control and Stabilization for Multiplicative Noise Systems With Intermittent Observations. IEEE TRANSACTIONS ON CYBERNETICS 2018;48:2128-2138. [PMID: 28767382 DOI: 10.1109/tcyb.2017.2728078] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
26
Wang Z, Liu L, Wu Y, Zhang H. Optimal Fault-Tolerant Control for Discrete-Time Nonlinear Strict-Feedback Systems Based on Adaptive Critic Design. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018;29:2179-2191. [PMID: 29771670 DOI: 10.1109/tnnls.2018.2810138] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
27
Yang Y, Modares H, Wunsch DC, Yin Y. Leader-Follower Output Synchronization of Linear Heterogeneous Systems With Active Leader Using Reinforcement Learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018;29:2139-2153. [PMID: 29771667 DOI: 10.1109/tnnls.2018.2803059] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
28
Yang X, He H. Adaptive critic designs for optimal control of uncertain nonlinear systems with unmatched interconnections. Neural Netw 2018;105:142-153. [PMID: 29843095 DOI: 10.1016/j.neunet.2018.05.005] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2018] [Revised: 04/13/2018] [Accepted: 05/04/2018] [Indexed: 10/16/2022]
PrevPage 1 of 1 1Next
© 2004-2025 Baishideng Publishing Group Inc. All rights reserved. 7041 Koll Center Parkway, Suite 160, Pleasanton, CA 94566, USA