We tackle the problem of an agent interacting with humans in a general-sum environment, i.e., a non-zero sum, non-fully cooperative setting, where the agent's goal is to increase its own utility. We show that when data is limited, building an accurate human model is very challenging, and that a reinforcement learning agent, which is based on this data, does not perform well in practice. Therefore, we propose that the agent should try maximizing a linear combination of the human's utility and its own utility rather than simply trying to maximize only its own utility
Human communication usually exhibits two fundamental and essential characteristics under environment...
A longstanding problem in the area of reinforcement learning is human-agent col- laboration. As past...
I Multi-agent Reinforcement Learning (RL) arises in many applications ranging from networked control...
We tackle the problem of an agent interacting with humans in a general-sum environment, i.e., a non-...
Interaction of humans and AI systems is becoming ubiquitous. Specifically, recent advances in machin...
Can artificial agents learn to assist others in achieving their goals without knowing what those goa...
Abstract. We examine ultraintelligent reinforcement learning agents. Reinforcement learning can only...
Abstract. This paper addresses the problem of cooperation between learning situated agents. We prese...
One of the complexities of social systems is the emergence of behavior norms that are costly for ind...
Reinforcement learning is the task of learning to act well in a variety of unknown environments. The...
While we would like agents that can coordinate with humans, current algorithms such as self-play and...
While we would like agents that can coordinate with humans, current algorithms such as self-play and...
Agents (humans, mice, computers) need to constantly make decisions to survive and thrive in their e...
The current reward learning from human preferences could be used to resolve complex reinforcement le...
We introduce learning in a principal-agent model of stochastic output sharing under moral haz-ard. W...
Human communication usually exhibits two fundamental and essential characteristics under environment...
A longstanding problem in the area of reinforcement learning is human-agent col- laboration. As past...
I Multi-agent Reinforcement Learning (RL) arises in many applications ranging from networked control...
We tackle the problem of an agent interacting with humans in a general-sum environment, i.e., a non-...
Interaction of humans and AI systems is becoming ubiquitous. Specifically, recent advances in machin...
Can artificial agents learn to assist others in achieving their goals without knowing what those goa...
Abstract. We examine ultraintelligent reinforcement learning agents. Reinforcement learning can only...
Abstract. This paper addresses the problem of cooperation between learning situated agents. We prese...
One of the complexities of social systems is the emergence of behavior norms that are costly for ind...
Reinforcement learning is the task of learning to act well in a variety of unknown environments. The...
While we would like agents that can coordinate with humans, current algorithms such as self-play and...
While we would like agents that can coordinate with humans, current algorithms such as self-play and...
Agents (humans, mice, computers) need to constantly make decisions to survive and thrive in their e...
The current reward learning from human preferences could be used to resolve complex reinforcement le...
We introduce learning in a principal-agent model of stochastic output sharing under moral haz-ard. W...
Human communication usually exhibits two fundamental and essential characteristics under environment...
A longstanding problem in the area of reinforcement learning is human-agent col- laboration. As past...
I Multi-agent Reinforcement Learning (RL) arises in many applications ranging from networked control...