In search advertising, the search engine needs to select the most profitable adver-tisements to display, which can be formulated as an instance of online learning with partial feedback, also known as the stochastic multi-armed bandit (MAB) problem. In this paper, we show that the naive application of MAB algorithms to search advertising for advertisement selection will produce sample selection bias that harms the search engine by decreasing expected revenue and “estima-tion of the largest mean ” (ELM) bias that harms the advertisers by increasing game-theoretic player-regret. We then propose simple bias-correction methods with benefits to both the search engine and the advertisers.
International audienceWe investigate multi-armed bandits with budgets, a natural model for ad-displa...
Sequential decision making is central to a range of marketing problems. Both firms and consumers aim...
Multi-armed bandit (MAB) problem is derived from slot machines in the casino. It is about how a gamb...
In search advertising, the search engine needs to select the most profitable adver-tisements to disp...
The multi-armed bandit (MAB) problem is a widely studied problem in machine learning literature in t...
We consider how a search engine should select advertisements to display with search results, in orde...
Nowadays it is important to be able to solve problems which involve multiple agents. In particular ...
We investigate a number of multi-armed bandit problems that model different aspects of online advert...
In pay-per click sponsored search auctions which are currently extensively used by search engines, t...
We present a data-driven algorithm that advertisers can use to automate their digital ad-campaigns a...
In this work, we explore an online reinforcement learning problem called the multi-armed bandit for ...
In pay-per-click sponsored search auctions which are currently extensively used by search engines, t...
Firms using online advertising regularly run experiments with multiple versions of their ads since t...
Sequential decision making is central to a range of marketing problems. Both firms and consumers aim...
The multi-armed bandit (MAB) problem is a mathematical formulation of the exploration-exploitation t...
International audienceWe investigate multi-armed bandits with budgets, a natural model for ad-displa...
Sequential decision making is central to a range of marketing problems. Both firms and consumers aim...
Multi-armed bandit (MAB) problem is derived from slot machines in the casino. It is about how a gamb...
In search advertising, the search engine needs to select the most profitable adver-tisements to disp...
The multi-armed bandit (MAB) problem is a widely studied problem in machine learning literature in t...
We consider how a search engine should select advertisements to display with search results, in orde...
Nowadays it is important to be able to solve problems which involve multiple agents. In particular ...
We investigate a number of multi-armed bandit problems that model different aspects of online advert...
In pay-per click sponsored search auctions which are currently extensively used by search engines, t...
We present a data-driven algorithm that advertisers can use to automate their digital ad-campaigns a...
In this work, we explore an online reinforcement learning problem called the multi-armed bandit for ...
In pay-per-click sponsored search auctions which are currently extensively used by search engines, t...
Firms using online advertising regularly run experiments with multiple versions of their ads since t...
Sequential decision making is central to a range of marketing problems. Both firms and consumers aim...
The multi-armed bandit (MAB) problem is a mathematical formulation of the exploration-exploitation t...
International audienceWe investigate multi-armed bandits with budgets, a natural model for ad-displa...
Sequential decision making is central to a range of marketing problems. Both firms and consumers aim...
Multi-armed bandit (MAB) problem is derived from slot machines in the casino. It is about how a gamb...