It is showed in this paper that quasi-(super)martingales play an important role in the theory of Markov decision processes. For excessive functions (with respect to a charge) it is proved that the value of the state at time t converges almost surely under each Markov strategy, which implies that the value function in the state at time t converges to zero (a.s), if an optimal strategy is used. At last a characterization of the conserving and equalizing properties is formulated using martingale theory
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In this paper the following result is proved. In any total reward countable state Markov decision pr...
This paper gives a systematic treatment of results about the existence of various types of nearly-op...
In this paper the following result is proved. In any total reward countable state Markov decision pr...
In this paper the following result is proved. In any total reward countable state Markov decision pr...
In this paper we consider a Markov decision process with countable state and time spaces. Rewards ha...
This paper gives a systematic treatment of results about the existence of various types of nearly-op...
This paper gives a systematic treatment of results about the existence of various types of nearly-op...
In this paper we consider a Markov decision process with countable state and time spaces. Rewards ha...
In this paper we consider a Markov decision process with countable state and time spaces. Rewards ha...
Strongly excessive functions play an important role in the theory of Markov decision processes and M...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In this paper the following result is proved. In any total reward countable state Markov decision pr...
This paper gives a systematic treatment of results about the existence of various types of nearly-op...
In this paper the following result is proved. In any total reward countable state Markov decision pr...
In this paper the following result is proved. In any total reward countable state Markov decision pr...
In this paper we consider a Markov decision process with countable state and time spaces. Rewards ha...
This paper gives a systematic treatment of results about the existence of various types of nearly-op...
This paper gives a systematic treatment of results about the existence of various types of nearly-op...
In this paper we consider a Markov decision process with countable state and time spaces. Rewards ha...
In this paper we consider a Markov decision process with countable state and time spaces. Rewards ha...
Strongly excessive functions play an important role in the theory of Markov decision processes and M...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...
In several standard models of dynamic programming (gambling houses, Markov decision processes (MDPs)...