We consider infinite horizon stochastic dynamic programs with discounted costs and study how to use information relaxations to calculate lower bounds on the performance of an optimal policy. We develop a general framework that allows for reformulations of the underlying state transition function. These reformulations can simplify the information relaxation calculations, both in leading to finite horizon subproblems and by reducing the number of states in these subproblems. We study as important special cases both “weak formulations ” in which states are independent of actions and “strong formulations” which retain the original dependence on actions. Our reformulations incorporate penalties for information in a direct way via control variate...
In Chapter 2, we propose several two-timescale simulation-based actor-critic algorithms for solution...
Partially observable decision processes (POMDP) can be used as a model for planning in stochastic do...
Partially observable Markov decision processes (POMDPs) allow one to model complex dynamic decision ...
Dynamic programming is a principal method for analyzing stochastic optimal control problems. However...
We consider the information relaxation approach for calculating performance bounds for stochastic dy...
Information relaxation and duality in Markov decision processes have been studied recently by severa...
summary:This paper is related to Markov Decision Processes. The optimal control problem is to minimi...
Information relaxation and duality in Markov decision processes have been studied recently by severa...
We analyze the per unit-time infinite horizon average cost Markov control model, subject to a total ...
This paper addresses the optimality of stochastic control strategies based on the infinite horizon a...
We consider terminating Markov decision processes with imperfect state information. We first assume ...
We consider approximate dynamic programming for the infinite-horizon stationary γ-discounted optimal...
47 pages, 3 figuresThis paper introduces algorithms for problems where a decision maker has to contr...
The thesis develops methods to solve discrete-time finite-state partially observable Markov decision...
In this study, we consider the infinite-horizon, discounted cost, optimal control of stochastic nonl...
In Chapter 2, we propose several two-timescale simulation-based actor-critic algorithms for solution...
Partially observable decision processes (POMDP) can be used as a model for planning in stochastic do...
Partially observable Markov decision processes (POMDPs) allow one to model complex dynamic decision ...
Dynamic programming is a principal method for analyzing stochastic optimal control problems. However...
We consider the information relaxation approach for calculating performance bounds for stochastic dy...
Information relaxation and duality in Markov decision processes have been studied recently by severa...
summary:This paper is related to Markov Decision Processes. The optimal control problem is to minimi...
Information relaxation and duality in Markov decision processes have been studied recently by severa...
We analyze the per unit-time infinite horizon average cost Markov control model, subject to a total ...
This paper addresses the optimality of stochastic control strategies based on the infinite horizon a...
We consider terminating Markov decision processes with imperfect state information. We first assume ...
We consider approximate dynamic programming for the infinite-horizon stationary γ-discounted optimal...
47 pages, 3 figuresThis paper introduces algorithms for problems where a decision maker has to contr...
The thesis develops methods to solve discrete-time finite-state partially observable Markov decision...
In this study, we consider the infinite-horizon, discounted cost, optimal control of stochastic nonl...
In Chapter 2, we propose several two-timescale simulation-based actor-critic algorithms for solution...
Partially observable decision processes (POMDP) can be used as a model for planning in stochastic do...
Partially observable Markov decision processes (POMDPs) allow one to model complex dynamic decision ...