In this paper we present a new algorithm for policy iteration for Markov decision processes (MDP) skip-free in one direction. This algorithm, which is based on matrix analytic methods, is in the same spirit as the algorithm of White (Stochastic Models, 21:785-797, 2005) which was limited to matrices that are skip-free in both directions. Optimization problems that can be solved using Markov decision processes arise in the domain of optical buffers, when trying to improve loss rates of fibre delay line (FDL) buffers. Based on the analysis of such an FDL buffer we present a comparative study between the different techniques available to solve an MDP. The results illustrate that the exploitation of the structure of the transition matrices...
Markov Decision Processes (MDP) are a widely used model including both non-deterministic and probabi...
In this paper, we propose an approximate policy iteration (API) algorithm for asemiconductor fab-lev...
We study the problem of computing the optimal value function for a Markov decision process with posi...
Markov decision processes (MDP) [1] provide a mathe-matical framework for studying a wide range of o...
This research focuses on Markov Decision Processes (MDP). MDP is one of the most important and chall...
summary:In this note we focus attention on identifying optimal policies and on elimination suboptima...
We present a technique for speeding up the convergence of value iteration for partially observable M...
An efficient algorithm for solving Markov decision problems is proposed. The value iteration method ...
In this paper we study a class of modified policy iteration algorithms for solving Markov decision p...
Abstract. Markov Decision Processes (MDP) are a widely used model including both non-deterministic a...
We consider Howard's policy iteration algorithm for multichained finite state and action Markov deci...
The Markov decision process is treated in a variety of forms or cases: finite or infinite horizon, w...
summary:In a Discounted Markov Decision Process (DMDP) with finite action sets the Value Iteration A...
Solving Markov Decision Processes is a recurrent task in engineering which can be performed efficien...
The thesis deals with linear approaches to the Markov Decision Process (MDP). In particular, we desc...
Markov Decision Processes (MDP) are a widely used model including both non-deterministic and probabi...
In this paper, we propose an approximate policy iteration (API) algorithm for asemiconductor fab-lev...
We study the problem of computing the optimal value function for a Markov decision process with posi...
Markov decision processes (MDP) [1] provide a mathe-matical framework for studying a wide range of o...
This research focuses on Markov Decision Processes (MDP). MDP is one of the most important and chall...
summary:In this note we focus attention on identifying optimal policies and on elimination suboptima...
We present a technique for speeding up the convergence of value iteration for partially observable M...
An efficient algorithm for solving Markov decision problems is proposed. The value iteration method ...
In this paper we study a class of modified policy iteration algorithms for solving Markov decision p...
Abstract. Markov Decision Processes (MDP) are a widely used model including both non-deterministic a...
We consider Howard's policy iteration algorithm for multichained finite state and action Markov deci...
The Markov decision process is treated in a variety of forms or cases: finite or infinite horizon, w...
summary:In a Discounted Markov Decision Process (DMDP) with finite action sets the Value Iteration A...
Solving Markov Decision Processes is a recurrent task in engineering which can be performed efficien...
The thesis deals with linear approaches to the Markov Decision Process (MDP). In particular, we desc...
Markov Decision Processes (MDP) are a widely used model including both non-deterministic and probabi...
In this paper, we propose an approximate policy iteration (API) algorithm for asemiconductor fab-lev...
We study the problem of computing the optimal value function for a Markov decision process with posi...