Neural networks for control
Connectionist learning for control: an overview
Neural networks for control
A menu of designs for reinforcement learning over time
Neural networks for control
Neural networks for control
Practical Issues in Temporal Difference Learning
Machine Learning
Technical Note: \cal Q-Learning
Machine Learning
The Convergence of TD(λ) for General λ
Machine Learning
Galerkin approximations of the generalized Hamilton-Jacobi-Bellman equation
Automatica (Journal of IFAC)
Introduction to Reinforcement Learning
Introduction to Reinforcement Learning
Oscillations in Neural Systems
Oscillations in Neural Systems
Neuro-Dynamic Programming
Learning to Predict by the Methods of Temporal Differences
Machine Learning
Handbook of Learning and Approximate Dynamic Programming (IEEE Press Series on Computational Intelligence)
Reinforcement Learning in Continuous Time and Space
Neural Computation
Stochastic Learning and Optimization: A Sensitivity-Based Approach (International Series on Discrete Event Dynamic Systems)
Approximate Dynamic Programming: Solving the Curses of Dimensionality (Wiley Series in Probability and Statistics)
Brief paper: Adaptive optimal control for continuous-time linear systems based on policy iteration
Automatica (Journal of IFAC)
2009 Special Issue: Language and cognition
Neural Networks
IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews
Neural dynamic optimization for control systems. I. Background
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
Neural dynamic optimization for control systems.II. Theory
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
Neural dynamic optimization for control systems.III. Applications
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
Discrete-Time Nonlinear HJB Solution Using Approximate Dynamic Programming: Convergence Proof
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
Issues on Stability of ADP Feedback Controllers for Dynamical Systems
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
Hamilton–Jacobi–Bellman Equations and Approximate Dynamic Programming on Time Scales
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
Adaptive-critic based optimal neuro control synthesis for distributed parameter systems
Automatica (Journal of IFAC)
Neuro-controller design for nonlinear fighter aircraft maneuver using fully tuned RBF networks
Automatica (Journal of IFAC)
Automatica (Journal of IFAC)
IEEE Transactions on Neural Networks
Online learning control by association and reinforcement
IEEE Transactions on Neural Networks
IEEE Transactions on Neural Networks
Continuous-Time Adaptive Critics
IEEE Transactions on Neural Networks
Robust/Optimal Temperature Profile Control of a High-Speed Aerospace Vehicle Using Neural Networks
IEEE Transactions on Neural Networks
Neural net robot controller with guaranteed tracking performance
IEEE Transactions on Neural Networks
Optimal control for a class of unknown nonlinear systems via the iterative GDHP algorithm
ISNN'11 Proceedings of the 8th international conference on Advances in neural networks - Volume Part II
Automatica (Journal of IFAC)
Automatica (Journal of IFAC)
Automatica (Journal of IFAC)
Constrained adaptive optimal control using a reinforcement learning agent
Automatica (Journal of IFAC)
Automatica (Journal of IFAC)
Information Sciences: an International Journal
Information Sciences: an International Journal
Data-based stability analysis of a class of nonlinear discrete-time systems
Information Sciences: an International Journal
Reinforcement learning algorithms with function approximation: Recent advances and applications
Information Sciences: an International Journal
On integral generalized policy iteration for continuous-time linear quadratic regulations
Automatica (Journal of IFAC)
Hi-index | 0.01 |
Living organisms learn by acting on their environment, observing the resulting reward stimulus, and adjusting their actions accordingly to improve the reward. This action-based or Reinforcement Learning can capture notions of optimal behavior occurring in natural systems. We describe mathematical formulations for Reinforcement Learning and a practical implementation method known as Adaptive Dynamic Programming. These give us insight into the design of controllers for man-made engineered systems that both learn and exhibit optimal behavior.