We study reward design strategies for incentivizing a reinforcement learning agent to adopt a policy from a set of admissible policies. The goal of the reward designer is to modify the underlying reward function cost-efficiently while ensuring that any approximately optimal deterministic policy under the new reward function is admissible and performs well under the original reward function. This problem can be viewed as a dual to the problem of optimal reward poisoning attacks: instead of forcing an agent to adopt a specific policy, the reward designer incentivizes an agent to avoid taking actions that are inadmissible in certain states. Perhaps surprisingly, and in contrast to the problem of optimal reward poisoning attacks, we first show ...
In safe Reinforcement Learning (RL), the agent attempts to find policies which maximize the expectat...
Reward shaping (RS) is a powerful method in reinforcement learning (RL) for overcoming the problem o...
We expose the danger of reward poisoning in offline multi-agent reinforcement learning (MARL), where...
We study reward design strategies for incentivizing a reinforcement learning agent to adopt a policy...
We study defense strategies against reward poisoning attacks in reinforcement learning. As a threat ...
Recent work has defined an optimal reward problem (ORP) in which an agent designer, with an objectiv...
To convey desired behavior to a Reinforcement Learning (RL) agent, a designer must choose a reward f...
Reward design is a fundamental problem in reinforcement learning (RL). A misspecified or poorly desi...
In reinforcement learning (RL), a reward function that aligns exactly with a task's true performance...
In the field of reinforcement learning, agent designers build agents which seek to maximize reward. ...
Reinforcement learning involves the study of how to solve sequential decision-making problems using ...
Many situations arise in which an interested party's utility is dependent on the actions of an agent...
How should a reinforcement learning agent act if its sole purpose is to efficiently learn an optimal...
Learning near-optimal behaviour from an expert's demonstrations typically relies on the assumption t...
In safe Reinforcement Learning (RL), the agent attempts to find policies which maximize the expectat...
In safe Reinforcement Learning (RL), the agent attempts to find policies which maximize the expectat...
Reward shaping (RS) is a powerful method in reinforcement learning (RL) for overcoming the problem o...
We expose the danger of reward poisoning in offline multi-agent reinforcement learning (MARL), where...
We study reward design strategies for incentivizing a reinforcement learning agent to adopt a policy...
We study defense strategies against reward poisoning attacks in reinforcement learning. As a threat ...
Recent work has defined an optimal reward problem (ORP) in which an agent designer, with an objectiv...
To convey desired behavior to a Reinforcement Learning (RL) agent, a designer must choose a reward f...
Reward design is a fundamental problem in reinforcement learning (RL). A misspecified or poorly desi...
In reinforcement learning (RL), a reward function that aligns exactly with a task's true performance...
In the field of reinforcement learning, agent designers build agents which seek to maximize reward. ...
Reinforcement learning involves the study of how to solve sequential decision-making problems using ...
Many situations arise in which an interested party's utility is dependent on the actions of an agent...
How should a reinforcement learning agent act if its sole purpose is to efficiently learn an optimal...
Learning near-optimal behaviour from an expert's demonstrations typically relies on the assumption t...
In safe Reinforcement Learning (RL), the agent attempts to find policies which maximize the expectat...
In safe Reinforcement Learning (RL), the agent attempts to find policies which maximize the expectat...
Reward shaping (RS) is a powerful method in reinforcement learning (RL) for overcoming the problem o...
We expose the danger of reward poisoning in offline multi-agent reinforcement learning (MARL), where...