Large Language Models (LLMs) have demonstrated impressive zero shot performance on a wide range of NLP tasks, demonstrating the ability to reason and apply commonsense. A relevant application is to use them for creating high quality synthetic datasets for downstream tasks. In this work, we probe whether GPT-4 can be used to augment existing extractive reading comprehension datasets. Automating data annotation processes has the potential to save large amounts of time, money and effort that goes into manually labelling datasets. In this paper, we evaluate the performance of GPT-4 as a replacement for human annotators for low resource reading comprehension tasks, by comparing performance after fine tuning, and the cost associated with annotati...
In low resource settings, data augmentation strategies are commonly leveraged to improve performance...
Large Language Models (LLMs) have achieved significant success across various natural language proce...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
Standard Full-Data classifiers in NLP demand thousands of labeled examples, which is impractical in ...
This paper investigates very low resource language model pretraining, when less than 100 thousand se...
Most NLG systems generate texts for readers with good reading ability, but SkillSum adapts its outpu...
Pretrained multilingual language models have become a common tool in transferring NLP capabilities t...
Alongside huge volumes of research on deep learning models in NLP in the recent years, there has bee...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
Large language models (LLMs) have significantly advanced the field of natural language processing, w...
Pretrained large language models (LLMs) are widely used in many sub-fields of natural language proce...
Text augmentation is a technique for constructing synthetic data from an under-resourced corpus to i...
Pretrained language models (PLMs) have made remarkable progress in table-to-text generation tasks. H...
Scaling language models with more data, compute and parameters has driven significant progress in na...
In recent years, there has been significant progress in developing pre-trained language models for N...
In low resource settings, data augmentation strategies are commonly leveraged to improve performance...
Large Language Models (LLMs) have achieved significant success across various natural language proce...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...
Standard Full-Data classifiers in NLP demand thousands of labeled examples, which is impractical in ...
This paper investigates very low resource language model pretraining, when less than 100 thousand se...
Most NLG systems generate texts for readers with good reading ability, but SkillSum adapts its outpu...
Pretrained multilingual language models have become a common tool in transferring NLP capabilities t...
Alongside huge volumes of research on deep learning models in NLP in the recent years, there has bee...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
Large language models (LLMs) have significantly advanced the field of natural language processing, w...
Pretrained large language models (LLMs) are widely used in many sub-fields of natural language proce...
Text augmentation is a technique for constructing synthetic data from an under-resourced corpus to i...
Pretrained language models (PLMs) have made remarkable progress in table-to-text generation tasks. H...
Scaling language models with more data, compute and parameters has driven significant progress in na...
In recent years, there has been significant progress in developing pre-trained language models for N...
In low resource settings, data augmentation strategies are commonly leveraged to improve performance...
Large Language Models (LLMs) have achieved significant success across various natural language proce...
Domain-specific text classification faces the challenge of scarce labeled data due to the high cost ...