Several researchers have shown that the efficiency of value iteration, a dynamic programming algorithm for Markov decision processes, can be improved by prioritizing the order of Bellman backups to focus computation on states where the value function can be improved the most. In previous work, a priority queue has been used to order backups. Although this incurs overhead for maintaining the priority queue, previous work has argued that the overhead is usually much less than the benefit from prioritization. However this conclusion is usually based on a comparison to a non-prioritized approach that performs Bellman backups on states in an arbitrary order. In this paper, we show that the overhead for maintaining the priority queue can be great...
In this paper, we study the M_n/M_n/c/(K_1+K_2)+M_n system with two finite-size queues where underly...
In this paper, we explore parallel implementations of the abstract data type priority queue. We use ...
This research focuses on Markov Decision Processes (MDP). MDP is one of the most important and chall...
We address the problem of computing an optimal value func-tion for Markov decision processes. Since ...
The problem of solving large Markov decision processes accurately and quickly is challenging. Since ...
Prioritisation of Bellman backups or updating only a small subset of actions represent important tec...
Abstract. Recent scaling up of POMDP solvers towards realistic applications is largely due to point-...
The performance of value and policy iteration can be dramatically improved by eliminating redundant ...
Abstract—Recent scaling up of POMDP solvers towards re-alistic applications is largely due to point-...
Consider the following random process: we are given n queues, into which elements of increasing labe...
We give a closed-form expression for the discounted weighted queue length and switching costs of a t...
Although partially observable Markov decision processes (POMDPs) have received significant attention...
We introduce a framework for reducing the number of element comparisons performed in priority-queue ...
In this paper we present a new algorithm for policy iteration for Markov decision processes (MDP) sk...
Relative priorities in an n-class queueing system can reduce server and customer costs. This propert...
In this paper, we study the M_n/M_n/c/(K_1+K_2)+M_n system with two finite-size queues where underly...
In this paper, we explore parallel implementations of the abstract data type priority queue. We use ...
This research focuses on Markov Decision Processes (MDP). MDP is one of the most important and chall...
We address the problem of computing an optimal value func-tion for Markov decision processes. Since ...
The problem of solving large Markov decision processes accurately and quickly is challenging. Since ...
Prioritisation of Bellman backups or updating only a small subset of actions represent important tec...
Abstract. Recent scaling up of POMDP solvers towards realistic applications is largely due to point-...
The performance of value and policy iteration can be dramatically improved by eliminating redundant ...
Abstract—Recent scaling up of POMDP solvers towards re-alistic applications is largely due to point-...
Consider the following random process: we are given n queues, into which elements of increasing labe...
We give a closed-form expression for the discounted weighted queue length and switching costs of a t...
Although partially observable Markov decision processes (POMDPs) have received significant attention...
We introduce a framework for reducing the number of element comparisons performed in priority-queue ...
In this paper we present a new algorithm for policy iteration for Markov decision processes (MDP) sk...
Relative priorities in an n-class queueing system can reduce server and customer costs. This propert...
In this paper, we study the M_n/M_n/c/(K_1+K_2)+M_n system with two finite-size queues where underly...
In this paper, we explore parallel implementations of the abstract data type priority queue. We use ...
This research focuses on Markov Decision Processes (MDP). MDP is one of the most important and chall...