We present an instance-based, online method for learning action models in unanticipated, relational domains. Our algorithm memorizes pre- and post-states of transitions an agent encounters while experiencing the environment, and makes predictions by using analogy to map the recorded transitions to novel situations. Our algorithm is implemented in the Soar cognitive architecture, integrating its task-independent episodic memory module and analogical reasoning implemented in procedural memory. We evaluate this algorithm’s prediction performance in a modified version of the blocks world domain and the taxi domain. We also present a reinforcement learning agent that uses our model learning algorithm to significantly speed up its convergen...
In this paper we describe Icarus, an architecture for physical agents that uses hierarchical skills ...
In this dissertation, we investigate learning by observation , a machine learning approach to create...
We present an algorithm that derives actions ’ effects and preconditions in partially observable, re...
The exploration-exploitation tradeoff is crucial to reinforcement-learning (RL) agents, and a signi...
The exploration-exploitation tradeoff is crucial to reinforcement-learning (RL) agents, and a signif...
Action modeling is an important skill for agents that must perform tasks in novel domains. Previous ...
In recent years, there has been a growing interest in using rich representations such as relational...
International audienceA number of recent works have designed algorithms that allow an agent to revis...
In recent years, there has been a growing interest in using rich representations such as relational ...
The largest project at the AICG lab at Linköping University, Cognitive models for virtual characters...
Abstract—We present a method for predicting action outcomes in unstructured environments with variab...
Agents (humans, mice, computers) need to constantly make decisions to survive and thrive in their e...
We present new algorithms for learning a logical model of actions' effects and preconditions in part...
In this paper we present a new method for reinforcement learning in relational domains. A logical la...
Abstract. In reinforcement learning, an agent tries to learn a policy, i.e., how to select an action...
In this paper we describe Icarus, an architecture for physical agents that uses hierarchical skills ...
In this dissertation, we investigate learning by observation , a machine learning approach to create...
We present an algorithm that derives actions ’ effects and preconditions in partially observable, re...
The exploration-exploitation tradeoff is crucial to reinforcement-learning (RL) agents, and a signi...
The exploration-exploitation tradeoff is crucial to reinforcement-learning (RL) agents, and a signif...
Action modeling is an important skill for agents that must perform tasks in novel domains. Previous ...
In recent years, there has been a growing interest in using rich representations such as relational...
International audienceA number of recent works have designed algorithms that allow an agent to revis...
In recent years, there has been a growing interest in using rich representations such as relational ...
The largest project at the AICG lab at Linköping University, Cognitive models for virtual characters...
Abstract—We present a method for predicting action outcomes in unstructured environments with variab...
Agents (humans, mice, computers) need to constantly make decisions to survive and thrive in their e...
We present new algorithms for learning a logical model of actions' effects and preconditions in part...
In this paper we present a new method for reinforcement learning in relational domains. A logical la...
Abstract. In reinforcement learning, an agent tries to learn a policy, i.e., how to select an action...
In this paper we describe Icarus, an architecture for physical agents that uses hierarchical skills ...
In this dissertation, we investigate learning by observation , a machine learning approach to create...
We present an algorithm that derives actions ’ effects and preconditions in partially observable, re...