Markov decision processes (MDPs) with large number of states are of high practical interest. However, conventional algorithms to solve MDP are computationally infeasible in this scenario. Approximate dynamic programming (ADP) methods tackle this issue by computing approximate solutions. A widely applied ADP method is approximate linear program (ALP) which makes use of linear function approximation and offers theoretical performance guarantees. Nevertheless, the ALP is difficult to solve due to the presence of a large number of constraints and in practice, a reduced linear program (RLP) is solved instead. The RLP has a tractable number of constraints sampled from the original constraints of the ALP. Though the RLP is known to perform well in...
We study reinforcement learning (RL) with linear function approximation. For episodic time-inhomogen...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
Address email We present an approximation scheme for solving Markov Decision Processes (MDPs) in whi...
Problems involving optimal sequential making in uncertain dynamic systems arise in domains such as e...
In many situations, it is desirable to optimize a sequence of decisions by maximizing a primary obje...
A weakness of classical Markov decision processes (MDPs) is that they scale very poorly due to the f...
In many situations, it is desirable to optimize a sequence of decisions by maximizing a primary obje...
Abstract Approximate linear programming (ALP) has emerged recently as one ofthe most promising metho...
We introduce a new algorithm based on linear programming for optimization of average-cost Markov dec...
We describe an approximate dynamic programming al-gorithm for partially observable Markov decision p...
We introduce a class of Markov decision problems (MDPs) which greatly simplify Reinforcement Learnin...
We introduce a class of MPDs which greatly simplify Reinforcement Learning. They have discrete state...
We introduce a class of MPDs which greatly simplify Reinforcement Learning. They have discrete state...
Approximate dynamic programming has been used successfully in a large variety of domains, but it rel...
I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, includ...
We study reinforcement learning (RL) with linear function approximation. For episodic time-inhomogen...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
Address email We present an approximation scheme for solving Markov Decision Processes (MDPs) in whi...
Problems involving optimal sequential making in uncertain dynamic systems arise in domains such as e...
In many situations, it is desirable to optimize a sequence of decisions by maximizing a primary obje...
A weakness of classical Markov decision processes (MDPs) is that they scale very poorly due to the f...
In many situations, it is desirable to optimize a sequence of decisions by maximizing a primary obje...
Abstract Approximate linear programming (ALP) has emerged recently as one ofthe most promising metho...
We introduce a new algorithm based on linear programming for optimization of average-cost Markov dec...
We describe an approximate dynamic programming al-gorithm for partially observable Markov decision p...
We introduce a class of Markov decision problems (MDPs) which greatly simplify Reinforcement Learnin...
We introduce a class of MPDs which greatly simplify Reinforcement Learning. They have discrete state...
We introduce a class of MPDs which greatly simplify Reinforcement Learning. They have discrete state...
Approximate dynamic programming has been used successfully in a large variety of domains, but it rel...
I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, includ...
We study reinforcement learning (RL) with linear function approximation. For episodic time-inhomogen...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
Address email We present an approximation scheme for solving Markov Decision Processes (MDPs) in whi...