In this paper, we place deep Q-learning into a control-oriented perspective and study its learning dynamics with well-established techniques from robust control. We formulate an uncertain linear time-invariant model by means of the neural tangent kernel to describe learning. We show the instability of learning and analyze the agent's behavior in frequency-domain. Then, we ensure convergence via robust controllers acting as dynamical rewards in the loss function. We synthesize three controllers: state-feedback gain scheduling H2, dynamic Hinf, and constant gain Hinf controllers. Setting up the learning agent with a control-oriented tuning methodology is more transparent and has well-established literature compared to the heuristics in reinfo...
In the past decade, machine learning strategies centered on the use of Deep Neural Networks (DNNs) h...
International audienceDespite remarkable successes, Deep Reinforcement Learning (DRL) is not robust ...
Department Head: Stephen B. Seidman.2000 Summer.Includes bibliographical references (pages 227-231)....
This work aims at constructing a bridge between robust control theory and reinforcement learning. Al...
In this article, we address two key challenges in deep reinforcement learning (DRL) setting, sample ...
Reinforcement learning (RL) offers powerful algorithms to search for optimal controllers of systems ...
In this paper the two reinforcement learning algorithmsQ-learning and deep Q-learning (DQN) are used...
Abstract. Q-learning can be used to learn a control policy that max-imises a scalar reward through i...
This is the version of record. It originally appeared on arXiv at http://arxiv.org/abs/1603.00748.Mo...
© 2018, the Authors. Reinforcement learning (RL) aims to resolve the sequential decision-making unde...
In Reinforcement learning, Q-learning is the best-known algorithm but it suffers from overestimation...
If reinforcement learning (RL) techniques are to be used for "real world" dynamic system c...
Traditional feedback control methods are often model-based and the mathematical system models need t...
This paper proposes a neural network based reinforcement learning controller that is able to learn c...
We present a variant of the Q-learning algorithm with automatic control of the exploration rate by a...
In the past decade, machine learning strategies centered on the use of Deep Neural Networks (DNNs) h...
International audienceDespite remarkable successes, Deep Reinforcement Learning (DRL) is not robust ...
Department Head: Stephen B. Seidman.2000 Summer.Includes bibliographical references (pages 227-231)....
This work aims at constructing a bridge between robust control theory and reinforcement learning. Al...
In this article, we address two key challenges in deep reinforcement learning (DRL) setting, sample ...
Reinforcement learning (RL) offers powerful algorithms to search for optimal controllers of systems ...
In this paper the two reinforcement learning algorithmsQ-learning and deep Q-learning (DQN) are used...
Abstract. Q-learning can be used to learn a control policy that max-imises a scalar reward through i...
This is the version of record. It originally appeared on arXiv at http://arxiv.org/abs/1603.00748.Mo...
© 2018, the Authors. Reinforcement learning (RL) aims to resolve the sequential decision-making unde...
In Reinforcement learning, Q-learning is the best-known algorithm but it suffers from overestimation...
If reinforcement learning (RL) techniques are to be used for "real world" dynamic system c...
Traditional feedback control methods are often model-based and the mathematical system models need t...
This paper proposes a neural network based reinforcement learning controller that is able to learn c...
We present a variant of the Q-learning algorithm with automatic control of the exploration rate by a...
In the past decade, machine learning strategies centered on the use of Deep Neural Networks (DNNs) h...
International audienceDespite remarkable successes, Deep Reinforcement Learning (DRL) is not robust ...
Department Head: Stephen B. Seidman.2000 Summer.Includes bibliographical references (pages 227-231)....