Suppose you are in a casino with a number of dollars you wish to gamble. You may quit whenever you please, and your objective is to find a strategy which will maximize the probability that you reach some goal, say $1000. In formal gambling-theoretic terminology, since there are only a finite number of dollars in the world, and since you may quit and leave whenever you wish, this is a finite-state leavable gambling problem [4], and the classical result of Dubins and Savage [4, Theorem 3.9.2.] says that for each e \u3e 0 there is always a stationary strategy which is uniformly e-optimal. That is, there is always a strategy for betting in which the bet you place at each play depends only on your current fortune, and using this strategy your ex...
(Extended version of Memorandum COSOR 81-11) This paper deals with total reward Markov decision proc...
This paper lays down conceptual groundwork for optimal choice in infinite-horizon finite-state Marko...
We study countably infinite Markov decision processes with B\"uchi objectives, which ask to visit a ...
In every finite-state leavable gambling problem and in every finite-state Markov decision process wi...
In every finite-state leavable gambling problem and in every finite-state Markov decision process wi...
AbstractIn a decision process (gambling or dynamic programming problem) with finite state space and ...
In several standard models of dynamic programming (gambling houses, MDPs, POMDPs), we prove the exis...
The notion of persistently optimal strategy in gambling theory is analogous to that of subgame-perfe...
A short introduction to goal problems in abstract gambling theory is given, along with statementes o...
AbstractThis paper deals with a discrete time Markov decision model with a finite state space, arbit...
In this no te we consider the finite-stage Markov game with finitely many states and actions as desc...
Markov decision processes (MDPs) are a standard model for dynamic systems that exhibit both stochast...
AbstractThe game n-bet is to bet n times with each bet between 0 and 1 inclusive. To win the game me...
A decision maker observes the evolving state of the world while constantly trying to predict the nex...
We consider the relationship between the reward function and the existence of (nearly) optimal Marko...
(Extended version of Memorandum COSOR 81-11) This paper deals with total reward Markov decision proc...
This paper lays down conceptual groundwork for optimal choice in infinite-horizon finite-state Marko...
We study countably infinite Markov decision processes with B\"uchi objectives, which ask to visit a ...
In every finite-state leavable gambling problem and in every finite-state Markov decision process wi...
In every finite-state leavable gambling problem and in every finite-state Markov decision process wi...
AbstractIn a decision process (gambling or dynamic programming problem) with finite state space and ...
In several standard models of dynamic programming (gambling houses, MDPs, POMDPs), we prove the exis...
The notion of persistently optimal strategy in gambling theory is analogous to that of subgame-perfe...
A short introduction to goal problems in abstract gambling theory is given, along with statementes o...
AbstractThis paper deals with a discrete time Markov decision model with a finite state space, arbit...
In this no te we consider the finite-stage Markov game with finitely many states and actions as desc...
Markov decision processes (MDPs) are a standard model for dynamic systems that exhibit both stochast...
AbstractThe game n-bet is to bet n times with each bet between 0 and 1 inclusive. To win the game me...
A decision maker observes the evolving state of the world while constantly trying to predict the nex...
We consider the relationship between the reward function and the existence of (nearly) optimal Marko...
(Extended version of Memorandum COSOR 81-11) This paper deals with total reward Markov decision proc...
This paper lays down conceptual groundwork for optimal choice in infinite-horizon finite-state Marko...
We study countably infinite Markov decision processes with B\"uchi objectives, which ask to visit a ...