Policy iterations for reinforcement learning problems in continuous time and space -- fundamental theory and methods
DOI10.1016/j.automatica.2020.109421zbMath1461.93143arXiv1705.03520OpenAlexW3128350768MaRDI QIDQ2664203
Richard S. Sutton, Jae Young Lee
Publication date: 20 April 2021
Published in: Automatica (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1705.03520
adaptive systemsreinforcement learningpolicy iterationiterative schemesoptimization under uncertaintiescontinuous time and space
Existence of optimal solutions belonging to restricted classes (Lipschitz controls, bang-bang controls, etc.) (49J30) Control/observation systems governed by ordinary differential equations (93C15) Iterative learning control (93B47)
Related Items (2)
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Neural network approach to continuous-time direct adaptive optimal control for partially unknown nonlinear systems
- Adaptive dynamic programming and optimal control of nonlinear nonaffine systems
- Stabilization with discounted optimal control
- Galerkin approximations of the generalized Hamilton-Jacobi-Bellman equation
- Nearly optimal control laws for nonlinear systems with saturating actuators using a neural network HJB approach
- Policy iterations for reinforcement learning problems in continuous time and space -- fundamental theory and methods
- Linear Quadratic Tracking Control of Partially-Unknown Continuous-Time Systems Using Reinforcement Learning
- An Approximation Theory of Optimal Control for Trainable Manipulators
- Approximate Dynamic Programming
- Construction of Suboptimal Control Sequences
This page was built for publication: Policy iterations for reinforcement learning problems in continuous time and space -- fundamental theory and methods