This study discusses the effect of semi-supervised learning in combination with pretrained language models for data-to-text generation. It is not known whether semi-supervised learning is still helpful when a large-scale language model is also supplemented. This study aims to answer this question by comparing a data-to-text system only supplemented with a language model, to two data-to-text systems that are additionally enriched by a data augmentation or a pseudo-labeling semi-supervised learning approach. Results show that semi-supervised learning results in higher scores on diversity metrics. In terms of output quality, extending the training set of a data-to-text system with a language model using the pseudo-labeling approach did increas...
Multi-view learning makes use of diverse models arising from multiple sources of input or different ...
In many cases of machine learning, research suggests that the development of training data might hav...
We present a comparison of word-based and character-based sequence-to sequence models for data-to-te...
This study discusses the effect of semi-supervised learning in combination with pretrained language ...
Within a situation where Semi-Supervised Learning (SSL) is available to exploit unlabeled data, this...
International audienceIn Natural Language Generation (NLG), End-to-End (E2E) systems trained through...
We introduce the problems of data-to-text generation and the current state of the art, i.e. pretrain...
Supervised learning algorithms employ labeled training data for classification purposes while obtain...
Current research state-of-the-art in automatic data-to-text generation, a major task in natural lang...
Data annotation is the process of labeling text, images, or other types of content for machine learn...
Author's accepted manuscriptGood performance in supervised text classification is usually obtained w...
Based on recent advances in natural language modeling and those in text generation capabilities, we ...
Obtaining labeled data to train natural language machine learning algorithms is often expen...
The current success of deep neural networks (DNNs) in an increasingly broad range of tasks involving...
The task of data-to-text generation amounts to describing structured data in fluent natural language...
Multi-view learning makes use of diverse models arising from multiple sources of input or different ...
In many cases of machine learning, research suggests that the development of training data might hav...
We present a comparison of word-based and character-based sequence-to sequence models for data-to-te...
This study discusses the effect of semi-supervised learning in combination with pretrained language ...
Within a situation where Semi-Supervised Learning (SSL) is available to exploit unlabeled data, this...
International audienceIn Natural Language Generation (NLG), End-to-End (E2E) systems trained through...
We introduce the problems of data-to-text generation and the current state of the art, i.e. pretrain...
Supervised learning algorithms employ labeled training data for classification purposes while obtain...
Current research state-of-the-art in automatic data-to-text generation, a major task in natural lang...
Data annotation is the process of labeling text, images, or other types of content for machine learn...
Author's accepted manuscriptGood performance in supervised text classification is usually obtained w...
Based on recent advances in natural language modeling and those in text generation capabilities, we ...
Obtaining labeled data to train natural language machine learning algorithms is often expen...
The current success of deep neural networks (DNNs) in an increasingly broad range of tasks involving...
The task of data-to-text generation amounts to describing structured data in fluent natural language...
Multi-view learning makes use of diverse models arising from multiple sources of input or different ...
In many cases of machine learning, research suggests that the development of training data might hav...
We present a comparison of word-based and character-based sequence-to sequence models for data-to-te...