In every finite-state leavable gambling problem and in every finite-state Markov decision process with discounted, negative or positive reward criteria there exists a Markov strategy which is monotonically improving and optimal in the limit along every history. An example is given to show that for the positive and gambling cases such strategies cannot be constructed by simply switching to a "better" action or gamble at each successive return to a state. Key words and phrases: gambling problem, Markov decision process, strategy, stationary strategy, monotonically improving strategy, limit-optimal strategy
In this report the same situation will be considered as in Hordijk, Dynamic programrrdng and Markov ...
Markov decision processes (MDPs) are a standard model for dynamic systems that exhibit both stochast...
We examine the use of stationary and Markov strategies in zero-sum stochastic games with finite stat...
In every finite-state leavable gambling problem and in every finite-state Markov decision process wi...
In every finite-state leavable gambling problem and in every finite-state Markov decision process wi...
Suppose you are in a casino with a number of dollars you wish to gamble. You may quit whenever you p...
In this no te we consider the finite-stage Markov game with finitely many states and actions as desc...
In several standard models of dynamic programming (gambling houses, MDPs, POMDPs), we prove the exis...
AbstractIn a decision process (gambling or dynamic programming problem) with finite state space and ...
A short introduction to goal problems in abstract gambling theory is given, along with statementes o...
This paper considers the two-person zero-sum Markov game with finite state and action spaces at the ...
The notion of persistently optimal strategy in gambling theory is analogous to that of subgame-perfe...
In a zero-sum limiting average stochastic game, we evaluate a\nstrategy π for the maximizing ...
This paper gives a systematic treatment of results about the existence of various types of nearly-op...
In the paper it is demonstrated, how a dynamic programming approach may be useful for the analysis o...
In this report the same situation will be considered as in Hordijk, Dynamic programrrdng and Markov ...
Markov decision processes (MDPs) are a standard model for dynamic systems that exhibit both stochast...
We examine the use of stationary and Markov strategies in zero-sum stochastic games with finite stat...
In every finite-state leavable gambling problem and in every finite-state Markov decision process wi...
In every finite-state leavable gambling problem and in every finite-state Markov decision process wi...
Suppose you are in a casino with a number of dollars you wish to gamble. You may quit whenever you p...
In this no te we consider the finite-stage Markov game with finitely many states and actions as desc...
In several standard models of dynamic programming (gambling houses, MDPs, POMDPs), we prove the exis...
AbstractIn a decision process (gambling or dynamic programming problem) with finite state space and ...
A short introduction to goal problems in abstract gambling theory is given, along with statementes o...
This paper considers the two-person zero-sum Markov game with finite state and action spaces at the ...
The notion of persistently optimal strategy in gambling theory is analogous to that of subgame-perfe...
In a zero-sum limiting average stochastic game, we evaluate a\nstrategy π for the maximizing ...
This paper gives a systematic treatment of results about the existence of various types of nearly-op...
In the paper it is demonstrated, how a dynamic programming approach may be useful for the analysis o...
In this report the same situation will be considered as in Hordijk, Dynamic programrrdng and Markov ...
Markov decision processes (MDPs) are a standard model for dynamic systems that exhibit both stochast...
We examine the use of stationary and Markov strategies in zero-sum stochastic games with finite stat...