To obtain high-quality sentence embeddings from pretrained language models (PLMs), they must either be augmented with additional pretraining objectives or finetuned on a large set of labeled text pairs. While the latter approach typically outperforms the former, it requires great human effort to generate suitable datasets of sufficient size. In this paper, we show how PLMs can be leveraged to obtain high-quality sentence embeddings without the need for labeled data, finetuning or modifications to the pretraining objective: We utilize the generative abilities of large and high-performing PLMs to generate entire datasets of labeled text pairs from scratch, which we then use for finetuning much smaller and more efficient models. Our fully unsu...
The increasing size of generative Pre-trained Language Models (PLMs) has greatly increased the deman...
Modern large-scale Pre-trained Language Models (PLMs) have achieved tremendous success on a wide ran...
The binary nature of grammaticality judgments and their use to access the structure of syntax are a ...
To obtain high-quality sentence embeddings from pretrained language models (PLMs), they must either ...
The task of data-to-text generation amounts to describing structured data in fluent natural language...
With the recent advances in deep learning, different approaches to improving pre-trained language mo...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Text Generation aims to produce plausible and readable text in a human language from input data. The...
The recent tremendous success of unsupervised word embeddings in a multitude of applications raises ...
Natural language processing (NLP) is one of the most important technologies of the information age. ...
In recent years, neural machine translation (NMT) has become the dominant approach in automated tran...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Current state of the art systems in NLP heavily rely on manually annotated datasets, which are expen...
Many modern NLP systems rely on word embeddings, previously trained in an unsupervised manner on lar...
A variety of contextualised language models have been proposed in the NLP community, which are train...
The increasing size of generative Pre-trained Language Models (PLMs) has greatly increased the deman...
Modern large-scale Pre-trained Language Models (PLMs) have achieved tremendous success on a wide ran...
The binary nature of grammaticality judgments and their use to access the structure of syntax are a ...
To obtain high-quality sentence embeddings from pretrained language models (PLMs), they must either ...
The task of data-to-text generation amounts to describing structured data in fluent natural language...
With the recent advances in deep learning, different approaches to improving pre-trained language mo...
Providing pretrained language models with simple task descriptions in natural language enables them ...
Text Generation aims to produce plausible and readable text in a human language from input data. The...
The recent tremendous success of unsupervised word embeddings in a multitude of applications raises ...
Natural language processing (NLP) is one of the most important technologies of the information age. ...
In recent years, neural machine translation (NMT) has become the dominant approach in automated tran...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Current state of the art systems in NLP heavily rely on manually annotated datasets, which are expen...
Many modern NLP systems rely on word embeddings, previously trained in an unsupervised manner on lar...
A variety of contextualised language models have been proposed in the NLP community, which are train...
The increasing size of generative Pre-trained Language Models (PLMs) has greatly increased the deman...
Modern large-scale Pre-trained Language Models (PLMs) have achieved tremendous success on a wide ran...
The binary nature of grammaticality judgments and their use to access the structure of syntax are a ...