Contextual bandits with linear payoffs, which are also known as linear bandits, provide a powerful alternative for solving practical problems of sequential decisions, e.g., online advertisements. In the era of big data, contextual data usually tend to be high-dimensional, which leads to new challenges for traditional linear bandits mostly designed for the setting of low-dimensional contextual data. Due to the curse of dimensionality, there are two challenges in most of the current bandit algorithms: the first is high time-complexity; and the second is extreme large upper regret bounds with high-dimensional data. In this paper, in order to attack the above two challenges effectively, we develop an algorithm of Contextual Bandits via RAndo...
We consider an adversarial variant of the classic $K$-armed linear contextual bandit problem where t...
We propose a novel algorithm for generalized linear contextual bandits (GLBs) with a regret bound su...
We consider a novel variant of the contextual bandit problem (i.e., the multi-armed bandit with side...
The bandit problem models a sequential decision process between a player and an environment. In the ...
Machine and Statistical Learning techniques are used in almost all online advertisement systems. The...
We analyze the K-armed bandit problem where the reward for each arm is a noisy realization based on ...
Contextual multi-armed bandit algorithms are widely used in sequential decision tasks such as news a...
We study the contextual bandits with knapsack (CBwK) problem under the high-dimensional setting wher...
In contextual continuum-armed bandits, the contexts $x$ and the arms $y$ are both continuous and dra...
Inspired by advertising markets, we consider large-scale sequential decision making problems in whic...
Contextual multi-armed bandit (MAB) algorithms have been shown promising for maximizing cumulative r...
Abstract To address the contextual bandit problem, we propose an online random forest algorithm. The...
In the contextual linear bandit setting, algorithms built on the optimism principle fail to exploit ...
Contextual bandit algorithms are sensitive to the estimation method of the outcome model as well as ...
Upper confidence bound (UCB) based contextual bandit algorithms require one to know the tail propert...
We consider an adversarial variant of the classic $K$-armed linear contextual bandit problem where t...
We propose a novel algorithm for generalized linear contextual bandits (GLBs) with a regret bound su...
We consider a novel variant of the contextual bandit problem (i.e., the multi-armed bandit with side...
The bandit problem models a sequential decision process between a player and an environment. In the ...
Machine and Statistical Learning techniques are used in almost all online advertisement systems. The...
We analyze the K-armed bandit problem where the reward for each arm is a noisy realization based on ...
Contextual multi-armed bandit algorithms are widely used in sequential decision tasks such as news a...
We study the contextual bandits with knapsack (CBwK) problem under the high-dimensional setting wher...
In contextual continuum-armed bandits, the contexts $x$ and the arms $y$ are both continuous and dra...
Inspired by advertising markets, we consider large-scale sequential decision making problems in whic...
Contextual multi-armed bandit (MAB) algorithms have been shown promising for maximizing cumulative r...
Abstract To address the contextual bandit problem, we propose an online random forest algorithm. The...
In the contextual linear bandit setting, algorithms built on the optimism principle fail to exploit ...
Contextual bandit algorithms are sensitive to the estimation method of the outcome model as well as ...
Upper confidence bound (UCB) based contextual bandit algorithms require one to know the tail propert...
We consider an adversarial variant of the classic $K$-armed linear contextual bandit problem where t...
We propose a novel algorithm for generalized linear contextual bandits (GLBs) with a regret bound su...
We consider a novel variant of the contextual bandit problem (i.e., the multi-armed bandit with side...