A Markov decision process (MDP) relies on the notions of state, describing the current situation of the agent, action affecting the dynamics of the process, and reward, observed for each transition between states. This chapter presents the basics of MDP theory and optimization, in the case of an agent having a perfect knowledge of the decision process and of its state at every time step, when the agent’s goal is to maximize its global revenue over time. Solving a Markov decision problem implies searching for a policy, in a given set, which optimizes a performance criterion for the considered MDP. The main criteria studied in the theory of MDPs are: finite criterion, discounted criterion, total reward criterion and average criterion. The cha...
Let (Xn) be a Markov process (in discrete time) with I state space E, I transition kernel Qn(·|x). L...
Optimality criteria for Markov decision processes have historically been based on a risk neutral for...
In many real-world applications of Markov Decision Processes (MPDs), the number of states is so larg...
We consider multistage decision processes where criterion function is an expectation of minimum func...
Abstract. State-based systems with discrete or continuous time are of-ten modelled with the help of ...
A short tutorial introduction is given to Markov decision processes (MDP), including the latest acti...
Bounded parameter Markov Decision Processes (BMDPs) address the issue of dealing with uncertainty in...
AbstractThe following optimality principle is established for finite undiscounted or discounted Mark...
Markov Decision Problems (MDPs) are the foundation for many problems that are of interest to researc...
The goal of this paper is two-fold: First, we present a sensitivity point of view on the optimizatio...
Markov Decision Processes (MDPs) are a mathematical framework for modeling sequential decision probl...
Abstract. The goal of this paper is two-fold: First, we present a sensitivity point of view on the o...
The running time of the classical algorithms of the Markov Decision Process (MDP) typically grows li...
Markov decision processes (MDP) [1] provide a mathe-matical framework for studying a wide range of o...
In this paper we consider a constrained optimization of discrete time Markov Decision Processes (MDP...
Let (Xn) be a Markov process (in discrete time) with I state space E, I transition kernel Qn(·|x). L...
Optimality criteria for Markov decision processes have historically been based on a risk neutral for...
In many real-world applications of Markov Decision Processes (MPDs), the number of states is so larg...
We consider multistage decision processes where criterion function is an expectation of minimum func...
Abstract. State-based systems with discrete or continuous time are of-ten modelled with the help of ...
A short tutorial introduction is given to Markov decision processes (MDP), including the latest acti...
Bounded parameter Markov Decision Processes (BMDPs) address the issue of dealing with uncertainty in...
AbstractThe following optimality principle is established for finite undiscounted or discounted Mark...
Markov Decision Problems (MDPs) are the foundation for many problems that are of interest to researc...
The goal of this paper is two-fold: First, we present a sensitivity point of view on the optimizatio...
Markov Decision Processes (MDPs) are a mathematical framework for modeling sequential decision probl...
Abstract. The goal of this paper is two-fold: First, we present a sensitivity point of view on the o...
The running time of the classical algorithms of the Markov Decision Process (MDP) typically grows li...
Markov decision processes (MDP) [1] provide a mathe-matical framework for studying a wide range of o...
In this paper we consider a constrained optimization of discrete time Markov Decision Processes (MDP...
Let (Xn) be a Markov process (in discrete time) with I state space E, I transition kernel Qn(·|x). L...
Optimality criteria for Markov decision processes have historically been based on a risk neutral for...
In many real-world applications of Markov Decision Processes (MPDs), the number of states is so larg...