In this thesis, we consider the task of data-to-text generation, which takes non-linguistic structures as input and produces textual output. The inputs can take the form of database tables, spreadsheets, charts, and so on. The main application of data-to-text generation is to present information in a textual format which makes it accessible to a layperson who may otherwise find it problematic to understand numerical figures. The task can also automate routine document generation jobs, thus improving human efficiency. We focus on generating long-form text, i.e., documents with multiple paragraphs. Recent approaches to data-to-text generation have adopted the very successful encoder-decoder architecture or its variants. These models ge...
Current approaches that generate text from linked data for complex real-world domains can face probl...
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data repres...
The current sequence-to-sequence with attention models, despite being successful, are inherently lim...
Recent neural models have shown significant progress on the problem of generating short descriptive ...
Recent advances in data-to-text generation have led to the use of large-scale datasets and neural ne...
Much of the data found on the world wide web is in numeric, tabular, or other nontextual format (e....
The problem of Data-to-Text Generation (D2T) is usually solved using a modular approach by breaking ...
International audienceIn this paper, we focus on how to create data-to-text corpora which can suppor...
International audienceA generation system can only be as good as the data it is trained on. In this ...
A number of researchers have recently questioned the necessity of increasingly complex neural netwo...
This work is funded by the Engineering and Physical Sciences Research Council (EPSRC), under a Natio...
Current research state-of-the-art in automatic data-to-text generation, a major task in natural lang...
The task of data-to-text generation amounts to describing structured data in fluent natural language...
Numerical tables are widely employed to communicate or report the classification performance of mach...
International audienceEnd-to-end encoder-decoder approaches to data-to-text generation are often bla...
Current approaches that generate text from linked data for complex real-world domains can face probl...
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data repres...
The current sequence-to-sequence with attention models, despite being successful, are inherently lim...
Recent neural models have shown significant progress on the problem of generating short descriptive ...
Recent advances in data-to-text generation have led to the use of large-scale datasets and neural ne...
Much of the data found on the world wide web is in numeric, tabular, or other nontextual format (e....
The problem of Data-to-Text Generation (D2T) is usually solved using a modular approach by breaking ...
International audienceIn this paper, we focus on how to create data-to-text corpora which can suppor...
International audienceA generation system can only be as good as the data it is trained on. In this ...
A number of researchers have recently questioned the necessity of increasingly complex neural netwo...
This work is funded by the Engineering and Physical Sciences Research Council (EPSRC), under a Natio...
Current research state-of-the-art in automatic data-to-text generation, a major task in natural lang...
The task of data-to-text generation amounts to describing structured data in fluent natural language...
Numerical tables are widely employed to communicate or report the classification performance of mach...
International audienceEnd-to-end encoder-decoder approaches to data-to-text generation are often bla...
Current approaches that generate text from linked data for complex real-world domains can face probl...
In data-to-text (D2T) generation, training on in-domain data leads to overfitting to the data repres...
The current sequence-to-sequence with attention models, despite being successful, are inherently lim...