In settings without well-defined goals, methods for reward learning allow reinforcement learning agents to infer goals from human feedback. Existing work has discussed the problem that such agents may manipulate humans, or the reward learning process, in order to gain higher reward. We introduce the neglected problem that, in multi-agent settings, agents may have incentives to manipulate one another’s reward functions in order to change each other’s behav- ioral policies. We focus on the setting with humans acting alongside assistive (artificial) agents who must learn the reward function by interacting with these humans. We propose a possible solution to manipulation of human feedback in this setting: the Shared Value Prior (SVP). The SVP e...
In multi-agent systems (MAS), agents rarely act in isolation but tend to achieve their goals through...
Learning from rewards generated by a human trainer observing an agent in action has proven to be a p...
Problems where agents wish to cooperate for a common goal, but disagree on their view of reality ar...
An important goal in artificial intelligence is to create agents that can both interact naturally wi...
University of Technology Sydney. Faculty of Engineering and Information Technology.A promising metho...
In the future, artificial learning agents are likely to become increasingly widespread in our societ...
Thesis (Ph.D.)--University of Washington, 2021Rapid strides made in the development of computing inf...
Interaction of humans and AI systems is becoming ubiquitous. Specifically, recent advances in machin...
In this work, we address a relatively unexplored aspect of designing agents that learn from human re...
In the future, artificial learning agents are likely to become increasingly widespread in our societ...
Can artificial agents learn to assist others in achieving their goals without knowing what those goa...
Cooperation is a widespread phenomenon in nature that has also been a cornerstone in the development...
Potential-based reward shaping has previously been proven to both be equivalent to Q-table initialis...
textRobots and other computational agents are increasingly becoming part of our daily lives. They wi...
Learning from rewards generated by a human trainer observing an agent in action has been proven to b...
In multi-agent systems (MAS), agents rarely act in isolation but tend to achieve their goals through...
Learning from rewards generated by a human trainer observing an agent in action has proven to be a p...
Problems where agents wish to cooperate for a common goal, but disagree on their view of reality ar...
An important goal in artificial intelligence is to create agents that can both interact naturally wi...
University of Technology Sydney. Faculty of Engineering and Information Technology.A promising metho...
In the future, artificial learning agents are likely to become increasingly widespread in our societ...
Thesis (Ph.D.)--University of Washington, 2021Rapid strides made in the development of computing inf...
Interaction of humans and AI systems is becoming ubiquitous. Specifically, recent advances in machin...
In this work, we address a relatively unexplored aspect of designing agents that learn from human re...
In the future, artificial learning agents are likely to become increasingly widespread in our societ...
Can artificial agents learn to assist others in achieving their goals without knowing what those goa...
Cooperation is a widespread phenomenon in nature that has also been a cornerstone in the development...
Potential-based reward shaping has previously been proven to both be equivalent to Q-table initialis...
textRobots and other computational agents are increasingly becoming part of our daily lives. They wi...
Learning from rewards generated by a human trainer observing an agent in action has been proven to b...
In multi-agent systems (MAS), agents rarely act in isolation but tend to achieve their goals through...
Learning from rewards generated by a human trainer observing an agent in action has proven to be a p...
Problems where agents wish to cooperate for a common goal, but disagree on their view of reality ar...