We present a comparison of word-based and character-based sequence-to sequence models for data-to-text natural language generation, which generate natural language descriptions for structured inputs. On the datasets of two recent generation challenges, our models achieve comparable or better automatic evaluation results than the best challenge submissions. Subsequent detailed statistical and human analyses shed light on the differences between the two input representations and the diversity of the generated texts. In a controlled experiment with synthetic training data generated from templates, we demonstrate the ability of neural models to learn novel combinations of the templates and thereby generalize beyond the linguistic structures the...
End-to-end encoder-decoder approaches to data-to-text generation are often black boxes whose predict...
Current approaches that generate text from linked data for complex real-world domains can face probl...
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data repres...
We present a comparison of word-based and character-based sequence-to sequence models for data-to-te...
We introduce the problems of data-to-text generation and the current state of the art, i.e. pretrain...
Text Generation is a pressing topic of Natural Language Processing that involves the prediction of u...
I propose a representation formalism and algorithms to be used in a new language generation mechanis...
Current research state-of-the-art in automatic data-to-text generation, a major task in natural lang...
Traditionally, most data-to-text applications have been designed using a modular pipeline architectu...
Current approaches that generate text from linked data for complex real-world domains can face probl...
2018-08-01Recurrent neural networks (RNN) have been successfully applied to various Natural Language...
The task of data-to-text generation amounts to describing structured data in fluent natural language...
Traditional statistical natural language generation (NLG) systems require substantial hand-engineeri...
Large-scale neural language models have made impressive strides in natural language generation. Howe...
Recurrent neural networks (RNNs) are exceptionally good models of distributions over natural languag...
End-to-end encoder-decoder approaches to data-to-text generation are often black boxes whose predict...
Current approaches that generate text from linked data for complex real-world domains can face probl...
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data repres...
We present a comparison of word-based and character-based sequence-to sequence models for data-to-te...
We introduce the problems of data-to-text generation and the current state of the art, i.e. pretrain...
Text Generation is a pressing topic of Natural Language Processing that involves the prediction of u...
I propose a representation formalism and algorithms to be used in a new language generation mechanis...
Current research state-of-the-art in automatic data-to-text generation, a major task in natural lang...
Traditionally, most data-to-text applications have been designed using a modular pipeline architectu...
Current approaches that generate text from linked data for complex real-world domains can face probl...
2018-08-01Recurrent neural networks (RNN) have been successfully applied to various Natural Language...
The task of data-to-text generation amounts to describing structured data in fluent natural language...
Traditional statistical natural language generation (NLG) systems require substantial hand-engineeri...
Large-scale neural language models have made impressive strides in natural language generation. Howe...
Recurrent neural networks (RNNs) are exceptionally good models of distributions over natural languag...
End-to-end encoder-decoder approaches to data-to-text generation are often black boxes whose predict...
Current approaches that generate text from linked data for complex real-world domains can face probl...
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data repres...