In this dissertation, we show a number of new results relating to stability, optimal control, and value iteration algorithms for discrete-time Markov decision processes (MDPs). First, we adapt two recent results in controlled diffusion processes to suit countable state MDPs by making assumptions that approximate continuous behavior. We show that if the MDP is stable under any stationary policy, then it must be uniformly so under all policies. This abstract result is very useful in the analysis of optimal control problems, and extends the characterization of uniform stability properties for MDPs. Then we derive two useful local bounds on the discounted value functions for a large class of MDPs, facilitating analysis of the ergodic cost probl...
In a nutshell, this thesis studies discrete-time Markov decision processes (MDPs) on Borel Spaces, w...
The thesis develops methods to solve discrete-time finite-state partially observable Markov decision...
This paper is concerned with the links between the Value Iteration algorithm and the Rolling Horizon...
In this dissertation, we show a number of new results relating to stability, optimal control, and va...
The intent of this book is to present recent results in the control theory for the long run average ...
We propose various computational schemes for solving Partially Observable Markov Decision Processes...
Using the value iteration procedure for discrete-time Markov con-trol processes on general Borel spa...
summary:In this note we focus attention on identifying optimal policies and on elimination suboptima...
AbstractThis paper concerns a discrete-time Markov decision model with an infinite planning horizon....
International audienceThe intent of this book is to present recent results in the control theory for...
A Markov decision problem is called reversible if the stationary controlled Markov chain is reversib...
This work considers denumerable state Markov Decision Chains endowed with a long-run expected averag...
summary:In a Discounted Markov Decision Process (DMDP) with finite action sets the Value Iteration A...
This paper investigates the criterion of long-term average costs for a Markov decision process (MDP)...
summary:In a Discounted Markov Decision Process (DMDP) with finite action sets the Value Iteration A...
In a nutshell, this thesis studies discrete-time Markov decision processes (MDPs) on Borel Spaces, w...
The thesis develops methods to solve discrete-time finite-state partially observable Markov decision...
This paper is concerned with the links between the Value Iteration algorithm and the Rolling Horizon...
In this dissertation, we show a number of new results relating to stability, optimal control, and va...
The intent of this book is to present recent results in the control theory for the long run average ...
We propose various computational schemes for solving Partially Observable Markov Decision Processes...
Using the value iteration procedure for discrete-time Markov con-trol processes on general Borel spa...
summary:In this note we focus attention on identifying optimal policies and on elimination suboptima...
AbstractThis paper concerns a discrete-time Markov decision model with an infinite planning horizon....
International audienceThe intent of this book is to present recent results in the control theory for...
A Markov decision problem is called reversible if the stationary controlled Markov chain is reversib...
This work considers denumerable state Markov Decision Chains endowed with a long-run expected averag...
summary:In a Discounted Markov Decision Process (DMDP) with finite action sets the Value Iteration A...
This paper investigates the criterion of long-term average costs for a Markov decision process (MDP)...
summary:In a Discounted Markov Decision Process (DMDP) with finite action sets the Value Iteration A...
In a nutshell, this thesis studies discrete-time Markov decision processes (MDPs) on Borel Spaces, w...
The thesis develops methods to solve discrete-time finite-state partially observable Markov decision...
This paper is concerned with the links between the Value Iteration algorithm and the Rolling Horizon...