We study the problem of learning controllers for discrete-time non-linear stochastic dynamical systems with formal reach-avoid guarantees. This work presents the first method for providing formal reach-avoid guarantees, which combine and generalize stability and safety guarantees, with a tolerable probability threshold $p\in[0,1]$ over the infinite time horizon. Our method leverages advances in machine learning literature and it represents formal certificates as neural networks. In particular, we learn a certificate in the form of a reach-avoid supermartingale (RASM), a novel notion that we introduce in this work. Our RASMs provide reachability and avoidance guarantees by imposing constraints on what can be viewed as a stochastic extension ...
Providing non-trivial certificates of safety for non-linear stochastic systems is an important open ...
Reinforcement Learning (RL) is a widely employed machine learning architecture that has been applied...
One of the most fundamental problems in Markov decision processes is analysis and control synthesis ...
We study the problem of learning controllers for discrete-time non-linear stochastic dynamical syste...
We study the problem of learning controllers for discrete-time non-linear stochastic dynamical syste...
We study the problem of learning controllers for discrete-time non-linear stochastic dynamical syste...
We consider the problem of learning control policies in discrete-time stochastic systems which guara...
We consider the problem of learning control policies in discrete-time stochastic systems which guara...
We consider the problem of formally verifying almost-sure (a.s.) asymptotic stability in discrete-ti...
Stability and safety are crucial in safety-critical control of dynamical systems. The reach-avoid-st...
Reinforcement learning (RL) is promising for complicated stochastic nonlinear control problems. With...
In this work, we address the problem of learning provably stable neural network policies for stochas...
The problem of synthesizing stochastic explicit model predictive control policies is known to be qui...
In this thesis, we study the related problems of reinforcement learning and optimal adaptive control...
In this work, the model predictive control problem is extended to include not only open-loop control...
Providing non-trivial certificates of safety for non-linear stochastic systems is an important open ...
Reinforcement Learning (RL) is a widely employed machine learning architecture that has been applied...
One of the most fundamental problems in Markov decision processes is analysis and control synthesis ...
We study the problem of learning controllers for discrete-time non-linear stochastic dynamical syste...
We study the problem of learning controllers for discrete-time non-linear stochastic dynamical syste...
We study the problem of learning controllers for discrete-time non-linear stochastic dynamical syste...
We consider the problem of learning control policies in discrete-time stochastic systems which guara...
We consider the problem of learning control policies in discrete-time stochastic systems which guara...
We consider the problem of formally verifying almost-sure (a.s.) asymptotic stability in discrete-ti...
Stability and safety are crucial in safety-critical control of dynamical systems. The reach-avoid-st...
Reinforcement learning (RL) is promising for complicated stochastic nonlinear control problems. With...
In this work, we address the problem of learning provably stable neural network policies for stochas...
The problem of synthesizing stochastic explicit model predictive control policies is known to be qui...
In this thesis, we study the related problems of reinforcement learning and optimal adaptive control...
In this work, the model predictive control problem is extended to include not only open-loop control...
Providing non-trivial certificates of safety for non-linear stochastic systems is an important open ...
Reinforcement Learning (RL) is a widely employed machine learning architecture that has been applied...
One of the most fundamental problems in Markov decision processes is analysis and control synthesis ...