Semi-supervised learning (SSL) is a popular setting aiming to effectively utilize unlabelled data to improve model performance in downstream natural language processing (NLP) tasks. Currently, there are two popular approaches to make use of unlabelled data: Self-training (ST) and Task-adaptive pre-training (TAPT). ST uses a teacher model to assign pseudo-labels to the unlabelled data, while TAPT continues pre-training on the unlabelled data before fine-tuning. To the best of our knowledge, the effectiveness of TAPT in SSL tasks has not been systematically studied, and no previous work has directly compared TAPT and ST in terms of their ability to utilize the pool of unlabelled data. In this paper, we provide an extensive empirical ...
Pseudo labeling and consistency regularization approaches based on confidencethresholding have made ...
Natural Language Inference (NLI) or Recognizing Textual Entailment (RTE) aims at predicting the rela...
The key to semi-supervised learning (SSL) is to explore adequate information to leverage the unlabel...
Following the success of supervised learning, semi-supervised learning (SSL) is now becoming increas...
Author's accepted manuscriptGood performance in supervised text classification is usually obtained w...
Given an unlabeled dataset and an annotation budget, we study how to selectively label a fixed numbe...
Semi-supervised learning (SSL) addresses the lack of labeled data by exploiting large unlabeled data...
Göpfert C, Ben-David S, Bousquet O, Gelly S, Tolstikhin I, Urner R. When can unlabeled data improve ...
Following the success of supervised learning, semi-supervised learning (SSL) is now becoming increas...
As an effective way to alleviate the burden of data annotation, semi-supervised learning (SSL) provi...
While semi-supervised learning (SSL) algorithms provide an efficient way to make use of both labelle...
Obtaining labeled data to train natural language machine learning algorithms is often expen...
Semi-supervised learning (SSL) is a learning framework that enables the use of unlabeled data with l...
Self-supervised learning (SSL) is a scalable way to learn general visual representations since it le...
Semi-supervised learning (SSL) aims to avoid the need for collecting prohibitively expensive labelle...
Pseudo labeling and consistency regularization approaches based on confidencethresholding have made ...
Natural Language Inference (NLI) or Recognizing Textual Entailment (RTE) aims at predicting the rela...
The key to semi-supervised learning (SSL) is to explore adequate information to leverage the unlabel...
Following the success of supervised learning, semi-supervised learning (SSL) is now becoming increas...
Author's accepted manuscriptGood performance in supervised text classification is usually obtained w...
Given an unlabeled dataset and an annotation budget, we study how to selectively label a fixed numbe...
Semi-supervised learning (SSL) addresses the lack of labeled data by exploiting large unlabeled data...
Göpfert C, Ben-David S, Bousquet O, Gelly S, Tolstikhin I, Urner R. When can unlabeled data improve ...
Following the success of supervised learning, semi-supervised learning (SSL) is now becoming increas...
As an effective way to alleviate the burden of data annotation, semi-supervised learning (SSL) provi...
While semi-supervised learning (SSL) algorithms provide an efficient way to make use of both labelle...
Obtaining labeled data to train natural language machine learning algorithms is often expen...
Semi-supervised learning (SSL) is a learning framework that enables the use of unlabeled data with l...
Self-supervised learning (SSL) is a scalable way to learn general visual representations since it le...
Semi-supervised learning (SSL) aims to avoid the need for collecting prohibitively expensive labelle...
Pseudo labeling and consistency regularization approaches based on confidencethresholding have made ...
Natural Language Inference (NLI) or Recognizing Textual Entailment (RTE) aims at predicting the rela...
The key to semi-supervised learning (SSL) is to explore adequate information to leverage the unlabel...