Originally developed for natural language problems, transformer models have recently been widely used in offline reinforcement learning tasks. This is because the agent's history can be represented as a sequence, and the whole task can be reduced to the sequence modeling task. However, the quadratic complexity of the transformer operation limits the potential increase in context. Therefore, different versions of the memory mechanism are used to work with long sequences in a natural language. This paper proposes the Recurrent Memory Decision Transformer (RMDT), a model that uses a recurrent memory mechanism for reinforcement learning problems. We conduct thorough experiments on Atari games and MuJoCo control problems and show that our propos...
Learning to solve sequential tasks with recurrent models requires the ability to memorize long seque...
Reinforcement Learning (RL) is a popular method in machine learning. In RL, an agent learns a policy...
In partially observable (PO) environments, deep reinforcement learning (RL) agents often suffer from...
Transformer-based models show their effectiveness across multiple domains and tasks. The self-attent...
Memory is an important component of effective learning systems and is crucial in non-Markovian as we...
The transformer architecture and variants presented remarkable success across many machine learning ...
Reinforcement learning (RL) algorithms face two distinct challenges: learning effective representati...
Thesis: S.M., Massachusetts Institute of Technology, Department of Electrical Engineering and Comput...
Modern deep reinforcement learning (RL) algorithms, despite being at the forefront of artificial int...
Reinforcement Learning (RL) can be considered as a sequence modeling task: given a sequence of past ...
Pour être capable d'apprendre et reconnaître des séquences, un agent robotique doit être équipé d'un...
This past year, RNNs have seen a lot of attention as powerful models that are able to decode sequenc...
Reinforcement Learning (RL) algorithms allow artificial agents to improve their action selection pol...
We introduce the Block-Recurrent Transformer, which applies a transformer layer in a recurrent fashi...
In this paper, we study transformers for text-based games. As a promising replacement of recurrent m...
Learning to solve sequential tasks with recurrent models requires the ability to memorize long seque...
Reinforcement Learning (RL) is a popular method in machine learning. In RL, an agent learns a policy...
In partially observable (PO) environments, deep reinforcement learning (RL) agents often suffer from...
Transformer-based models show their effectiveness across multiple domains and tasks. The self-attent...
Memory is an important component of effective learning systems and is crucial in non-Markovian as we...
The transformer architecture and variants presented remarkable success across many machine learning ...
Reinforcement learning (RL) algorithms face two distinct challenges: learning effective representati...
Thesis: S.M., Massachusetts Institute of Technology, Department of Electrical Engineering and Comput...
Modern deep reinforcement learning (RL) algorithms, despite being at the forefront of artificial int...
Reinforcement Learning (RL) can be considered as a sequence modeling task: given a sequence of past ...
Pour être capable d'apprendre et reconnaître des séquences, un agent robotique doit être équipé d'un...
This past year, RNNs have seen a lot of attention as powerful models that are able to decode sequenc...
Reinforcement Learning (RL) algorithms allow artificial agents to improve their action selection pol...
We introduce the Block-Recurrent Transformer, which applies a transformer layer in a recurrent fashi...
In this paper, we study transformers for text-based games. As a promising replacement of recurrent m...
Learning to solve sequential tasks with recurrent models requires the ability to memorize long seque...
Reinforcement Learning (RL) is a popular method in machine learning. In RL, an agent learns a policy...
In partially observable (PO) environments, deep reinforcement learning (RL) agents often suffer from...