We study countably infinite Markov decision processes with Büchi objectives, which ask to visit a given subset F of states infinitely often. A question left open by T.P. Hill in 1979 [10] is whether there always exist ε-optimal Markov strategies, i.e., strategies that base decisions only on the current state and the number of steps taken so far. We provide a negative answer to this question by constructing a non-trivial counterexample. On the other hand, we show that Markov strategies with only 1 bit of extra memory are sufficient.</p
In this paper the following result is proved. In any total reward countable state Markov decision pr...
AbstractFor countable-state decision processes (dynamic programming problems), a general class of ob...
We consider discrete-time Markov decision processes in which the decision maker is interested in lon...
We study countably infinite Markov decision processes with B\"uchi objectives, which ask to visit a ...
The Transience objective is not to visit any state infinitely often. While this is not possible in a...
Markov decision processes (MDPs) are a standard model for dynamic systems that exhibit both stochast...
The Transience objective is not to visit any state infinitely often. While this is not possible in f...
We study countably infinite Markov decision processes (MDPs) with real-valued transition rewards. Ev...
We study countably infinite Markov decision processes (MDPs) with real-valuedtransition rewards. Eve...
We study countably infinite MDPs with parity objectives, and special cases with a bounded number of ...
We study countably infinite MDPs with parity objectives. Unlike in finite MDPs, optimal strategies n...
We study countably infinite MDPs with parity objectives, and special cases with a bounded number of ...
We study countably infinite MDPs with parity objectives. Unlike in finite MDPs, optimal strategies n...
This paper gives a systematic treatment of results about the existence of various types of nearly-op...
We study countably infinite Markov decision processes (MDPs) with real-valued transition rewards. Ev...
In this paper the following result is proved. In any total reward countable state Markov decision pr...
AbstractFor countable-state decision processes (dynamic programming problems), a general class of ob...
We consider discrete-time Markov decision processes in which the decision maker is interested in lon...
We study countably infinite Markov decision processes with B\"uchi objectives, which ask to visit a ...
The Transience objective is not to visit any state infinitely often. While this is not possible in a...
Markov decision processes (MDPs) are a standard model for dynamic systems that exhibit both stochast...
The Transience objective is not to visit any state infinitely often. While this is not possible in f...
We study countably infinite Markov decision processes (MDPs) with real-valued transition rewards. Ev...
We study countably infinite Markov decision processes (MDPs) with real-valuedtransition rewards. Eve...
We study countably infinite MDPs with parity objectives, and special cases with a bounded number of ...
We study countably infinite MDPs with parity objectives. Unlike in finite MDPs, optimal strategies n...
We study countably infinite MDPs with parity objectives, and special cases with a bounded number of ...
We study countably infinite MDPs with parity objectives. Unlike in finite MDPs, optimal strategies n...
This paper gives a systematic treatment of results about the existence of various types of nearly-op...
We study countably infinite Markov decision processes (MDPs) with real-valued transition rewards. Ev...
In this paper the following result is proved. In any total reward countable state Markov decision pr...
AbstractFor countable-state decision processes (dynamic programming problems), a general class of ob...
We consider discrete-time Markov decision processes in which the decision maker is interested in lon...