In collective decision-making (CDM) a group of experts with a shared set of values and a common goal must combine their knowledge to make a collectively optimal decision. Whereas existing research on CDM primarily focuses on making binary decisions, we focus here on CDM applied to solving contextual multi-armed bandit (CMAB) problems, where the goal is to exploit contextual information to select the best arm among a set. To address the limiting assumptions of prior work, we introduce confidence estimates and propose a novel approach to deciding with expert advice which can take advantage of these estimates. We further show that, when confidence estimates are imperfect, the proposed approach is more robust than the classical confidence-weigh...
Presented as part of the ARC11 lecture on October 30, 2017 at 10:00 a.m. in the Klaus Advanced Compu...
Abstract—The contextual bandit problem is typically used to model online applications such as articl...
International audienceThis paper introduces a general multi-agent bandit model in which each agent i...
Collective decision-making (CDM) processes – wherein the knowledge of a group of individuals with a ...
Quite some real-world problems can be formulated as decision-making problems wherein one must repeat...
Quite some real-world problems can be formulated as decision-making problems wherein one must repeat...
Most important decisions in our society are made by groups, from cabinets and commissions to boards ...
There are numerous situations when a service requester wishes to expertsource a series of identical ...
© 2016 EUCA. We study the explore-exploit tradeoff in distributed cooperative decision-making using ...
Abstract—We present a formal model of human decision-making in explore-exploit tasks using the conte...
Consider a requester who wishes to crowdsource a series of identical binary labeling tasks from a po...
When humans collaborate with each other, they often make decisions by observing others and consideri...
Adversarial multiarmed bandits with expert advice is one of the fundamental problems in studying the...
The failure of groups to make optimal decisions is an important topic in human sciences. Recently th...
The total knowledge contained within a collective supersedes the knowledge of even its most intellig...
Presented as part of the ARC11 lecture on October 30, 2017 at 10:00 a.m. in the Klaus Advanced Compu...
Abstract—The contextual bandit problem is typically used to model online applications such as articl...
International audienceThis paper introduces a general multi-agent bandit model in which each agent i...
Collective decision-making (CDM) processes – wherein the knowledge of a group of individuals with a ...
Quite some real-world problems can be formulated as decision-making problems wherein one must repeat...
Quite some real-world problems can be formulated as decision-making problems wherein one must repeat...
Most important decisions in our society are made by groups, from cabinets and commissions to boards ...
There are numerous situations when a service requester wishes to expertsource a series of identical ...
© 2016 EUCA. We study the explore-exploit tradeoff in distributed cooperative decision-making using ...
Abstract—We present a formal model of human decision-making in explore-exploit tasks using the conte...
Consider a requester who wishes to crowdsource a series of identical binary labeling tasks from a po...
When humans collaborate with each other, they often make decisions by observing others and consideri...
Adversarial multiarmed bandits with expert advice is one of the fundamental problems in studying the...
The failure of groups to make optimal decisions is an important topic in human sciences. Recently th...
The total knowledge contained within a collective supersedes the knowledge of even its most intellig...
Presented as part of the ARC11 lecture on October 30, 2017 at 10:00 a.m. in the Klaus Advanced Compu...
Abstract—The contextual bandit problem is typically used to model online applications such as articl...
International audienceThis paper introduces a general multi-agent bandit model in which each agent i...