In real-world scenarios with naturally occurring datasets, reference summaries are noisy and may contain information that cannot be inferred from the source text. On large news corpora, removing low quality samples has been shown to reduce model hallucinations. Yet, for smaller, and/or noisier corpora, filtering is detrimental to performance. To improve reference quality while retaining all data, we propose a new approach: to selectively re-write unsupported reference sentences to better reflect source data. We automatically generate a synthetic dataset of positive and negative revisions by corrupting supported sentences and learn to revise reference sentences with contrastive learning. The intensity of revisions is treated as a controllabl...
The fluency and creativity of large pre-trained language models (LLMs) have led to their widespread ...
Textual knowledge bases such as Wikipedia require considerable effort to keep up to date and consist...
Factual inconsistencies in generated summaries severely limit the practical applications of abstract...
Document-level models for information extraction tasks like slot-filling are flexible: they can be a...
Large Language Models (LLMs) like the GPT and LLaMA families have demonstrated exceptional capabilit...
Canonical automatic summary evaluation metrics, such as ROUGE, focus on lexical similarity which can...
Missing information is a common issue of dialogue summarization where some information in the refere...
Despite the recent progress in language generation models, their outputs may not always meet user ex...
We study the robustness of machine reading comprehension (MRC) models to entity renaming -- do model...
We call into question the recently popularized method of direct model editing as a means of correcti...
Adopting a two-stage paradigm of pretraining followed by fine-tuning, Pretrained Language Models (PL...
Despite the recent advances in abstractive text summarization, current summarization models still su...
Hallucination is a known issue for neural abstractive summarization models. Recent work suggests tha...
Current abstractive summarization systems present important weaknesses which prevent their deploymen...
Language models, given their black-box nature, often exhibit sensitivity to input perturbations, lea...
The fluency and creativity of large pre-trained language models (LLMs) have led to their widespread ...
Textual knowledge bases such as Wikipedia require considerable effort to keep up to date and consist...
Factual inconsistencies in generated summaries severely limit the practical applications of abstract...
Document-level models for information extraction tasks like slot-filling are flexible: they can be a...
Large Language Models (LLMs) like the GPT and LLaMA families have demonstrated exceptional capabilit...
Canonical automatic summary evaluation metrics, such as ROUGE, focus on lexical similarity which can...
Missing information is a common issue of dialogue summarization where some information in the refere...
Despite the recent progress in language generation models, their outputs may not always meet user ex...
We study the robustness of machine reading comprehension (MRC) models to entity renaming -- do model...
We call into question the recently popularized method of direct model editing as a means of correcti...
Adopting a two-stage paradigm of pretraining followed by fine-tuning, Pretrained Language Models (PL...
Despite the recent advances in abstractive text summarization, current summarization models still su...
Hallucination is a known issue for neural abstractive summarization models. Recent work suggests tha...
Current abstractive summarization systems present important weaknesses which prevent their deploymen...
Language models, given their black-box nature, often exhibit sensitivity to input perturbations, lea...
The fluency and creativity of large pre-trained language models (LLMs) have led to their widespread ...
Textual knowledge bases such as Wikipedia require considerable effort to keep up to date and consist...
Factual inconsistencies in generated summaries severely limit the practical applications of abstract...