The lottery ticket hypothesis questions the role of overparameterization in supervised deep learning. But how is the performance of winning lottery tickets affected by the distributional shift inherent to reinforcement learning problems? In this work, we address this question by comparing sparse agents who have to address the non-stationarity of the exploration-exploitation problem with supervised agents trained to imitate an expert. We show that feed-forward networks trained with behavioural cloning compared to reinforcement learning can be pruned to higher levels of sparsity without performance degradation. This suggests that in order to solve the RL-specific distributional shift agents require more degrees of freedom. Using a set of care...
Network pruning is an effective approach to reduce network complexity with acceptable performance co...
Learning optimal policies in sparse rewards settings is difficult as the learning agent has little t...
Large neural networks can be pruned to a small fraction of their original size, with little loss in ...
In the era of foundation models with huge pre-training budgets, the downstream tasks have been shift...
The reinforcement learning (RL) problem is rife with sources of non-stationarity, making it a notori...
The lottery ticket hypothesis has sparked the rapid development of pruning algorithms that perform s...
peer reviewedWe study the generalization properties of pruned models that are the winners of the lot...
Reinforcement learning (RL) algorithms have proven transformative in a range of domains. To tackle r...
Marginalized importance sampling (MIS), which measures the density ratio between the state-action oc...
One of the most challenging types of environments for a Deep Reinforcement Learning agent to learn i...
Deep reinforcement learning (DRL) agents are trained through trial-and-error interactions with the e...
Recent advances in deep learning optimization showed that just a subset of parameters are really nec...
We investigate sparse representations for control in reinforcement learning. While these representat...
Pruning deep neural networks is a widely used strategy to alleviate the computational burden in mach...
The lottery ticket hypothesis conjectures the existence of sparse subnetworks of large randomly init...
Network pruning is an effective approach to reduce network complexity with acceptable performance co...
Learning optimal policies in sparse rewards settings is difficult as the learning agent has little t...
Large neural networks can be pruned to a small fraction of their original size, with little loss in ...
In the era of foundation models with huge pre-training budgets, the downstream tasks have been shift...
The reinforcement learning (RL) problem is rife with sources of non-stationarity, making it a notori...
The lottery ticket hypothesis has sparked the rapid development of pruning algorithms that perform s...
peer reviewedWe study the generalization properties of pruned models that are the winners of the lot...
Reinforcement learning (RL) algorithms have proven transformative in a range of domains. To tackle r...
Marginalized importance sampling (MIS), which measures the density ratio between the state-action oc...
One of the most challenging types of environments for a Deep Reinforcement Learning agent to learn i...
Deep reinforcement learning (DRL) agents are trained through trial-and-error interactions with the e...
Recent advances in deep learning optimization showed that just a subset of parameters are really nec...
We investigate sparse representations for control in reinforcement learning. While these representat...
Pruning deep neural networks is a widely used strategy to alleviate the computational burden in mach...
The lottery ticket hypothesis conjectures the existence of sparse subnetworks of large randomly init...
Network pruning is an effective approach to reduce network complexity with acceptable performance co...
Learning optimal policies in sparse rewards settings is difficult as the learning agent has little t...
Large neural networks can be pruned to a small fraction of their original size, with little loss in ...