We study the problem of decision-making under uncertainty in the bandit setting. This thesis goes beyond the well-studied multi-armed bandit model to consider structured bandit settings and their applications. In particular, we learn to make better decisions by leveraging the application-specific problem-structure in the form of features or graph information. We investigate the use of structured bandits in two practical applications: online recommender systems with an available network of users and viral marketing in social networks. For each of these applications, we design efficient bandit algorithms and theoretically characterize their performance. We experimentally evaluate the efficiency and effectiveness of these algorithms on real-wo...
University of Technology Sydney. Faculty of Engineering and Information Technology.The sequential de...
We study structured multi-armed bandits, which is the problem of online decision-making under uncert...
How do humans search for rewards? This question is commonly studied using multi-armed bandit tasks, ...
We study the problem of decision-making under uncertainty in the bandit setting. This thesis goes be...
Inspired by advertising markets, we consider large-scale sequential decision making problems in whic...
In a bandit problem there is a set of arms, each of which when played by an agent yields some reward...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
Abstract—We present a formal model of human decision-making in explore-exploit tasks using the conte...
Inspired by advertising markets, we consider large-scale sequential decision making problems in whic...
We study bandit problems in which a decision-maker gets reward-or-failure feedback when choosing rep...
This paper introduces the Banditron, a vari-ant of the Perceptron [Rosenblatt, 1958], for the multic...
The bandit problem models a sequential decision process between a player and an environment. In the ...
Multi-Armed bandit (MAB) framework is a widely used sequential decision making framework in which a ...
In this thesis we address the multi-armed bandit (MAB) problem with stochastic rewards and correlate...
University of Technology Sydney. Faculty of Engineering and Information Technology.The sequential de...
We study structured multi-armed bandits, which is the problem of online decision-making under uncert...
How do humans search for rewards? This question is commonly studied using multi-armed bandit tasks, ...
We study the problem of decision-making under uncertainty in the bandit setting. This thesis goes be...
Inspired by advertising markets, we consider large-scale sequential decision making problems in whic...
In a bandit problem there is a set of arms, each of which when played by an agent yields some reward...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
Abstract—We present a formal model of human decision-making in explore-exploit tasks using the conte...
Inspired by advertising markets, we consider large-scale sequential decision making problems in whic...
We study bandit problems in which a decision-maker gets reward-or-failure feedback when choosing rep...
This paper introduces the Banditron, a vari-ant of the Perceptron [Rosenblatt, 1958], for the multic...
The bandit problem models a sequential decision process between a player and an environment. In the ...
Multi-Armed bandit (MAB) framework is a widely used sequential decision making framework in which a ...
In this thesis we address the multi-armed bandit (MAB) problem with stochastic rewards and correlate...
University of Technology Sydney. Faculty of Engineering and Information Technology.The sequential de...
We study structured multi-armed bandits, which is the problem of online decision-making under uncert...
How do humans search for rewards? This question is commonly studied using multi-armed bandit tasks, ...