A growing body of research in continual learning focuses on the catastrophic forgetting problem. While many attempts have been made to alleviate this problem, the majority of the methods assume a single model in the continual learning setup. In this work, we question this assumption and show that employing ensemble models can be a simple yet effective method to improve continual performance. However, ensembles' training and inference costs can increase significantly as the number of models grows. Motivated by this limitation, we study different ensemble models to understand their benefits and drawbacks in continual learning scenarios. Finally, to overcome the high compute cost of ensembles, we leverage recent advances in neural network subs...
Learning and adapting to new distributions or learning new tasks sequentially without forgetting the...
Ensemble methods can deliver surprising performance gains but also bring significantly higher comput...
Learning continuously during all model lifetime is fundamental to deploy machine learning solutions ...
This paper argues that continual learning methods can benefit by splitting the capacity of the learn...
Deep neural networks have shown remarkable performance when trained on independent and identically d...
Continual learning (CL) is a setting in which an agent has to learn from an incoming stream of data ...
Continual learning with an increasing number of classes is a challenging task. The difficulty rises ...
Deep learning has enjoyed tremendous success over the last decade, but the training of practically u...
In continual learning (CL), the goal is to design models that can learn a sequence of tasks without ...
Recently, continual learning (CL) has gained significant interest because it enables deep learning m...
For decades research has pursued the ambitious goal of designing computer models that learn to solve...
Using task-specific components within a neural network in continual learning (CL) is a compelling st...
Bayesian Flow Networks (BFNs) has been recently proposed as one of the most promising direction to u...
The ability of a model to learn continually can be empirically assessed in different continual learn...
The ability of a model to learn continually can be empirically assessed in different continual learn...
Learning and adapting to new distributions or learning new tasks sequentially without forgetting the...
Ensemble methods can deliver surprising performance gains but also bring significantly higher comput...
Learning continuously during all model lifetime is fundamental to deploy machine learning solutions ...
This paper argues that continual learning methods can benefit by splitting the capacity of the learn...
Deep neural networks have shown remarkable performance when trained on independent and identically d...
Continual learning (CL) is a setting in which an agent has to learn from an incoming stream of data ...
Continual learning with an increasing number of classes is a challenging task. The difficulty rises ...
Deep learning has enjoyed tremendous success over the last decade, but the training of practically u...
In continual learning (CL), the goal is to design models that can learn a sequence of tasks without ...
Recently, continual learning (CL) has gained significant interest because it enables deep learning m...
For decades research has pursued the ambitious goal of designing computer models that learn to solve...
Using task-specific components within a neural network in continual learning (CL) is a compelling st...
Bayesian Flow Networks (BFNs) has been recently proposed as one of the most promising direction to u...
The ability of a model to learn continually can be empirically assessed in different continual learn...
The ability of a model to learn continually can be empirically assessed in different continual learn...
Learning and adapting to new distributions or learning new tasks sequentially without forgetting the...
Ensemble methods can deliver surprising performance gains but also bring significantly higher comput...
Learning continuously during all model lifetime is fundamental to deploy machine learning solutions ...