In many cases of machine learning, research suggests that the development of training data might have a higher relevance than the choice and modelling of classifiers themselves. Thus, data augmentation methods have been developed to improve classifiers by artificially created training data. In NLP, there is the challenge of establishing universal rules for text transformations which provide new linguistic patterns. In this paper, we present and evaluate a text generation method suitable to increase the performance of classifiers for long and short texts. We achieved promising improvements when evaluating short as well as long text tasks with the enhancement by our text generation method. Especially with regard to small data analytics, addit...
Thanks to increases in computing power and the growing availability of large datasets, neural netwo...
Data augmentation is widely used in text classification, especially in the low-resource regime where...
This study discusses the effect of semi-supervised learning in combination with pretrained language ...
In many cases of machine learning, research suggests that the development of training data might hav...
Based on recent advances in natural language modeling and those in text generation capabilities, we ...
Data augmentation, the artificial creation of training data for machine learning by transformations,...
We study the effect of different approaches to text augmentation. To do this we use three datasets t...
Text has traditionally been used to train automated classifiers for a multitude of purposes, such as...
Text classification typically performs best with large training sets, but short texts are very commo...
Data Augmentation approaches often use Language Models, pretrained on large quantities of unlabeled ...
In Natural Language Processing (NLP), applications trained on downstream tasks for text classificati...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
Text augmentation is a technique for constructing synthetic data from an under-resourced corpus to i...
International audienceWe study the effect of different approaches to text augmentation. To do this w...
As an effective strategy, data augmentation (DA) alleviates data scarcity scenarios where deep learn...
Thanks to increases in computing power and the growing availability of large datasets, neural netwo...
Data augmentation is widely used in text classification, especially in the low-resource regime where...
This study discusses the effect of semi-supervised learning in combination with pretrained language ...
In many cases of machine learning, research suggests that the development of training data might hav...
Based on recent advances in natural language modeling and those in text generation capabilities, we ...
Data augmentation, the artificial creation of training data for machine learning by transformations,...
We study the effect of different approaches to text augmentation. To do this we use three datasets t...
Text has traditionally been used to train automated classifiers for a multitude of purposes, such as...
Text classification typically performs best with large training sets, but short texts are very commo...
Data Augmentation approaches often use Language Models, pretrained on large quantities of unlabeled ...
In Natural Language Processing (NLP), applications trained on downstream tasks for text classificati...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
Text augmentation is a technique for constructing synthetic data from an under-resourced corpus to i...
International audienceWe study the effect of different approaches to text augmentation. To do this w...
As an effective strategy, data augmentation (DA) alleviates data scarcity scenarios where deep learn...
Thanks to increases in computing power and the growing availability of large datasets, neural netwo...
Data augmentation is widely used in text classification, especially in the low-resource regime where...
This study discusses the effect of semi-supervised learning in combination with pretrained language ...