International audienceThe use of pretrained deep neural networks represents an attractive alternative to achieve strong results with few data available. When specialized in dense problems such as object detection, learning local rather than global information in images has proven to be much more efficient. However, for unsupervised pretraining, the popular contrastive learning requires a large batch size and, therefore, a lot of resources. To address this problem, we are interested in Transformer-based object detectors that have recently gained traction in the community with good performance by generating many diverse object proposals.In this work, we propose ProSeCo, a novel unsupervised end-to-end pretraining approach to leverage this pro...
This thesis focuses on large-scale visual pretraining in computer vision and addresses various limi...
In a weakly-supervised scenario object detectors need to be trained using image-level annotation alo...
This paper explores a better prediction target for BERT pre-training of vision transformers. We obs...
International audienceThe use of pretrained deep neural networks represents an attractive alternativ...
DETR-based object detectors have achieved remarkable performance but are sample-inefficient and exhi...
Object detection, which aims to recognize and locate objects within images using bounding boxes, is ...
Weakly supervised object detection (WSOD) enables object detectors to be trained using image-level c...
Object detection is a fundamental computer vision task that estimates object classification labels a...
Combining simple architectures with large-scale pre-training has led to massive improvements in imag...
We explore the plain, non-hierarchical Vision Transformer (ViT) as a backbone network for object det...
Our purpose in this work is to boost the performance of object classifiers learned using the self-tr...
In this article, a novel real-time object detector called Transformers Only Look Once (TOLO) is prop...
This thesis concerns the problem of object detection, which is defined as finding all instances of a...
Visual object detection has seen substantial improvements during the last years due to the possibili...
In this paper, we present a novel methodology based on machine learning for identifying the most app...
This thesis focuses on large-scale visual pretraining in computer vision and addresses various limi...
In a weakly-supervised scenario object detectors need to be trained using image-level annotation alo...
This paper explores a better prediction target for BERT pre-training of vision transformers. We obs...
International audienceThe use of pretrained deep neural networks represents an attractive alternativ...
DETR-based object detectors have achieved remarkable performance but are sample-inefficient and exhi...
Object detection, which aims to recognize and locate objects within images using bounding boxes, is ...
Weakly supervised object detection (WSOD) enables object detectors to be trained using image-level c...
Object detection is a fundamental computer vision task that estimates object classification labels a...
Combining simple architectures with large-scale pre-training has led to massive improvements in imag...
We explore the plain, non-hierarchical Vision Transformer (ViT) as a backbone network for object det...
Our purpose in this work is to boost the performance of object classifiers learned using the self-tr...
In this article, a novel real-time object detector called Transformers Only Look Once (TOLO) is prop...
This thesis concerns the problem of object detection, which is defined as finding all instances of a...
Visual object detection has seen substantial improvements during the last years due to the possibili...
In this paper, we present a novel methodology based on machine learning for identifying the most app...
This thesis focuses on large-scale visual pretraining in computer vision and addresses various limi...
In a weakly-supervised scenario object detectors need to be trained using image-level annotation alo...
This paper explores a better prediction target for BERT pre-training of vision transformers. We obs...