In this paper, we investigate the continual learning of Vision Transformers (ViT) for the challenging exemplar-free scenario, with special focus on how to efficiently distill the knowledge of its crucial self-attention mechanism (SAM). Our work takes an initial step towards a surgical investigation of SAM for designing coherent continual learning methods in ViTs. We first carry out an evaluation of established continual learning regularization techniques. We then examine the effect of regularization when applied to two key enablers of SAM: (a) the contextualized embedding layers, for their ability to capture well-scaled representations with respect to the values, and (b) the prescaled attention maps, for carrying value-independent global co...
This paper investigates two techniques for developing efficient self-supervised vision transformers ...
Vision transformers (ViTs) have recently obtained success in many applications, but their intensive ...
Although transformer networks are recently employed in various vision tasks with outperforming perfo...
Recent studies show that Vision Transformers(ViTs) exhibit strong robustness against various corrupt...
In contrast to batch learning where all training data is available at once, continual learning repre...
In this paper we propose a new method for exemplar-free class incremental training of ViTs. The main...
Transformers have recently shown superior performances on various vision tasks. The large, sometimes...
Recent work has shown that the attention maps of Vision Transformers (VTs), when trained with self-s...
Vision transformers (ViT) have demonstrated impressive performance across numerous machine vision ta...
International audienceSelf-supervised models have been shown to produce comparable or better visual ...
Vision transformers (ViTs) have pushed the state-of-the-art for various visual recognition tasks by ...
Though modern deep learning based approaches have achieved remarkable progress in computer vision co...
In this short paper, we propose a baseline (off-the-shelf) for Continual Learning of Computer Vision...
I first review the existing methods based on regularization for continual learning. While regularizi...
International audienceIn this paper, we question if self-supervised learning provides new properties...
This paper investigates two techniques for developing efficient self-supervised vision transformers ...
Vision transformers (ViTs) have recently obtained success in many applications, but their intensive ...
Although transformer networks are recently employed in various vision tasks with outperforming perfo...
Recent studies show that Vision Transformers(ViTs) exhibit strong robustness against various corrupt...
In contrast to batch learning where all training data is available at once, continual learning repre...
In this paper we propose a new method for exemplar-free class incremental training of ViTs. The main...
Transformers have recently shown superior performances on various vision tasks. The large, sometimes...
Recent work has shown that the attention maps of Vision Transformers (VTs), when trained with self-s...
Vision transformers (ViT) have demonstrated impressive performance across numerous machine vision ta...
International audienceSelf-supervised models have been shown to produce comparable or better visual ...
Vision transformers (ViTs) have pushed the state-of-the-art for various visual recognition tasks by ...
Though modern deep learning based approaches have achieved remarkable progress in computer vision co...
In this short paper, we propose a baseline (off-the-shelf) for Continual Learning of Computer Vision...
I first review the existing methods based on regularization for continual learning. While regularizi...
International audienceIn this paper, we question if self-supervised learning provides new properties...
This paper investigates two techniques for developing efficient self-supervised vision transformers ...
Vision transformers (ViTs) have recently obtained success in many applications, but their intensive ...
Although transformer networks are recently employed in various vision tasks with outperforming perfo...