Abstract—The multi-armed bandit problem and one of its most interesting extensions, the restless bandits problem, are frequently encountered in various stochastic control problems. We present a linear programming relaxation for the restless bandits problem with discounted rewards, where only one project can be activated at each period but with additional costs penalizing switching between projects. The relaxation can be efficiently computed and provides a bound on the achievable performance. We describe several heuristic policies; in particular, we show that a policy adapted from the primal-dual heuristic of Bertsimas and Niño-Mora [1] for the classical restless bandits problem is in fact equivalent to a one-step lookahead policy; thus, th...
International audienceIn this paper we study a Multi-Armed Restless Bandit Problem (MARBP) subject t...
This paper develops a framework based on convex optimization and economic ideas to formulate and sol...
Abstract—We consider two variants of the standard multi-armed bandit problem, namely, the multi-arme...
Abstract—We consider a task assignment problem for a fleet of UAVs in a surveillance/search mission....
We propose a mathematical programming approach for the classical PSPACE - hard problem of n restless...
We consider the multi-armed restless bandit problem (RMABP) with an infinite horizon average cost ob...
Includes bibliographical references (p. 19-21).Supported by a Presidential Young Investigator Award....
We provide a framework to analyse control policies for the restless Markovian bandit model, under bo...
155 pagesWe consider multi-action restless bandits with multiple resource constraints, also referred...
We study a resource allocation problem with varying requests and with resources of limited capacity ...
We explore the scheduling rules and the hedging levels that can be obtained by using a Restless Band...
We propose an asymptotically optimal heuristic, which we termed the Randomized Assignment Control (R...
Restless multi-armed bandits (RMABs) are an important model to optimize allocation of limited resour...
International audienceWe develop a unifying framework to obtain efficient index policies for restles...
International audienceThis paper studies optimal control subject to changing conditions. This is an ...
International audienceIn this paper we study a Multi-Armed Restless Bandit Problem (MARBP) subject t...
This paper develops a framework based on convex optimization and economic ideas to formulate and sol...
Abstract—We consider two variants of the standard multi-armed bandit problem, namely, the multi-arme...
Abstract—We consider a task assignment problem for a fleet of UAVs in a surveillance/search mission....
We propose a mathematical programming approach for the classical PSPACE - hard problem of n restless...
We consider the multi-armed restless bandit problem (RMABP) with an infinite horizon average cost ob...
Includes bibliographical references (p. 19-21).Supported by a Presidential Young Investigator Award....
We provide a framework to analyse control policies for the restless Markovian bandit model, under bo...
155 pagesWe consider multi-action restless bandits with multiple resource constraints, also referred...
We study a resource allocation problem with varying requests and with resources of limited capacity ...
We explore the scheduling rules and the hedging levels that can be obtained by using a Restless Band...
We propose an asymptotically optimal heuristic, which we termed the Randomized Assignment Control (R...
Restless multi-armed bandits (RMABs) are an important model to optimize allocation of limited resour...
International audienceWe develop a unifying framework to obtain efficient index policies for restles...
International audienceThis paper studies optimal control subject to changing conditions. This is an ...
International audienceIn this paper we study a Multi-Armed Restless Bandit Problem (MARBP) subject t...
This paper develops a framework based on convex optimization and economic ideas to formulate and sol...
Abstract—We consider two variants of the standard multi-armed bandit problem, namely, the multi-arme...