We prove a central limit theorem for a class of additive processes that arise naturally in the theory of finite horizon Markov decision problems. The main theorem generalizes a classic result of Dobrushin (1956) for temporally non-homogeneous Markov chains, and the principal innovation is that here the summands are permitted to depend on both the current state and a bounded number of future states of the chain. We show through several examples that this added flexibility gives one a direct path to asymptotic normality of the optimal total reward of finite horizon Markov decision problems. The same examples also explain why such results are not easily obtained by alternative Markovian techniques such as enlargement of the state space
We study the optimization of average rewards of discrete time nonhomogeneous Markov chains, in which...
Staudigl M. A limit theorem for Markov decision processes. Center for Mathematical Economics Working...
AbstractSuppose that {Xt,t≥0} is a non-stationary Markov process, taking values in a Polish metric s...
We prove a central limit theorem for a class of additive processes that arise naturally in the theor...
Abstract. We prove a central limit theorem for a class of additive processes that arise naturally in...
Abstract. We prove a central limit theorem for a class of additive processes that arise naturally in...
In this paper we started by explaining what a Markov chain is. After this we defined some key concep...
International audienceBifurcating Markov chains (BMC) are Markov chains indexed by a full binary tre...
AbstractA new type of central limit theorems for random evolutions with semi-Markov switch-overs in ...
The theory of Markov Decision Processes is the theory of controlled Markov chains. Its origins can b...
Bifurcating Markov chains (BMC) are Markov chains indexed by a full binary tree representing the evo...
International audienceWe establish self-norming central limit theorems for non-stationary time serie...
AbstractA simple sufficient condition for the Central Limit Theorem for functionals of Harris ergodi...
In this paper, we study a Markov decision process with a non-linear discount function and with a Bor...
The main objective of this article is to establish a central limit theorem for additive three-variab...
We study the optimization of average rewards of discrete time nonhomogeneous Markov chains, in which...
Staudigl M. A limit theorem for Markov decision processes. Center for Mathematical Economics Working...
AbstractSuppose that {Xt,t≥0} is a non-stationary Markov process, taking values in a Polish metric s...
We prove a central limit theorem for a class of additive processes that arise naturally in the theor...
Abstract. We prove a central limit theorem for a class of additive processes that arise naturally in...
Abstract. We prove a central limit theorem for a class of additive processes that arise naturally in...
In this paper we started by explaining what a Markov chain is. After this we defined some key concep...
International audienceBifurcating Markov chains (BMC) are Markov chains indexed by a full binary tre...
AbstractA new type of central limit theorems for random evolutions with semi-Markov switch-overs in ...
The theory of Markov Decision Processes is the theory of controlled Markov chains. Its origins can b...
Bifurcating Markov chains (BMC) are Markov chains indexed by a full binary tree representing the evo...
International audienceWe establish self-norming central limit theorems for non-stationary time serie...
AbstractA simple sufficient condition for the Central Limit Theorem for functionals of Harris ergodi...
In this paper, we study a Markov decision process with a non-linear discount function and with a Bor...
The main objective of this article is to establish a central limit theorem for additive three-variab...
We study the optimization of average rewards of discrete time nonhomogeneous Markov chains, in which...
Staudigl M. A limit theorem for Markov decision processes. Center for Mathematical Economics Working...
AbstractSuppose that {Xt,t≥0} is a non-stationary Markov process, taking values in a Polish metric s...