We consider stochastic dynamic programming problems with high-dimensional, discrete state-spaces and finite, discrete-time horizons that prohibit direct computation of the value function from a given Bellman equation for all states and time steps due to the “curse of dimensionality”. For the case where the value function of the dynamic program is concave extensible and submodular in its state-space, we present a new algorithm that computes deterministic upper and stochastic lower bounds of the value function in the realm of dual dynamic programming. We show that the proposed algorithm terminates after a finite number of iterations. Furthermore, we derive probabilistic guarantees on the value accumulated under the associated policy for a sin...
We consider the revenue management problem of finding profit-maximising prices for delivery time slo...
Recent advances in algorithms for solving large linear programs, specifically constraint generation,...
We consider dynamic programming problems with a large time horizon, and give suf- ficient conditions...
We consider dynamic programming problems with finite, discrete-time horizons and prohibitively high-...
We consider the problem of finding profit-maximising prices for delivery time slots in the context o...
We introduce a class of algorithms, called Trajectory Following Dynamic Programming (TFDP) algorithm...
This paper explores sufficient conditions for a continuous stationary Markov optimal policy and a co...
We introduce a class of algorithms, called Trajectory Following Dynamic Programming (TFDP) algorithm...
23 pages, 3 figuresWe present an algorithm called Tropical Dynamic Programming (TDP) which builds up...
The Stochastic Dual Dynamic Programming (SDDP) algorithm has become one of the main tools to address...
We consider the information relaxation approach for calculating performance bounds for stochastic dy...
Several attempt to dampen the curse of dimensionnality problem of the Dynamic Programming approach f...
We study both the value function and Q-function formulation of the Linear Programming approach to Ap...
We propose an algorithm, which we call 'Fast Value Iteration' (FVI), to compute the value function o...
We investigate in this paper submodular properties of the value function arrizing in complex Dynamic...
We consider the revenue management problem of finding profit-maximising prices for delivery time slo...
Recent advances in algorithms for solving large linear programs, specifically constraint generation,...
We consider dynamic programming problems with a large time horizon, and give suf- ficient conditions...
We consider dynamic programming problems with finite, discrete-time horizons and prohibitively high-...
We consider the problem of finding profit-maximising prices for delivery time slots in the context o...
We introduce a class of algorithms, called Trajectory Following Dynamic Programming (TFDP) algorithm...
This paper explores sufficient conditions for a continuous stationary Markov optimal policy and a co...
We introduce a class of algorithms, called Trajectory Following Dynamic Programming (TFDP) algorithm...
23 pages, 3 figuresWe present an algorithm called Tropical Dynamic Programming (TDP) which builds up...
The Stochastic Dual Dynamic Programming (SDDP) algorithm has become one of the main tools to address...
We consider the information relaxation approach for calculating performance bounds for stochastic dy...
Several attempt to dampen the curse of dimensionnality problem of the Dynamic Programming approach f...
We study both the value function and Q-function formulation of the Linear Programming approach to Ap...
We propose an algorithm, which we call 'Fast Value Iteration' (FVI), to compute the value function o...
We investigate in this paper submodular properties of the value function arrizing in complex Dynamic...
We consider the revenue management problem of finding profit-maximising prices for delivery time slo...
Recent advances in algorithms for solving large linear programs, specifically constraint generation,...
We consider dynamic programming problems with a large time horizon, and give suf- ficient conditions...