Earlier research has shown that few studies in Natural Language Generation (NLG) evaluate their system outputs using an error analysis, despite known limitations of automatic evaluation metrics and human ratings. This position paper takes the stance that error analyses should be encouraged, and discusses several ways to do so. This paper is based on our shared experience as authors as well as a survey we distributed as a means of public consultation. We provide an overview of existing barriers to carrying out error analyses, and propose changes to improve error reporting in the NLG literature
With the fast-growing popularity of current large pre-trained language models (LLMs), it is necessar...
We consider the evaluation problem in Natural Language Generation (NLG) and present results for eval...
Human assessment remains the most trusted form of evaluation in NLG, but highly diverse approaches a...
Earlier research has shown that few studies in Natural Language Generation (NLG) evaluate their syst...
We observe a severe under-reporting of the different kinds of errors that Natural Language Generatio...
While automatically computing numerical scores remains the dominant paradigm in NLP system evaluatio...
ABSTRACT Many evaluation issues for grammatical error detection have previously been overlooked, mak...
Natural language generation (nlg) systems are computer software systems that pro-duce texts in Engli...
The Natural Language Generation (NLG) community relies on shared evaluation techniques to understand...
Rating and Likert scales are widely used in evaluation experiments to measure the quality of Natural...
Recently, there has been an increased interest in demographically grounded bias in natural language ...
This study evaluates a natural language generation system that creates literacy assessment reports i...
As Natural Language Processing (NLP) technology rapidly develops and spreads into daily life, it bec...
One of the advantages of deep grammars, such as those based on HPSG, is that they can be used for ge...
In this paper, we present the results of two re- production studies for the human evaluation origina...
With the fast-growing popularity of current large pre-trained language models (LLMs), it is necessar...
We consider the evaluation problem in Natural Language Generation (NLG) and present results for eval...
Human assessment remains the most trusted form of evaluation in NLG, but highly diverse approaches a...
Earlier research has shown that few studies in Natural Language Generation (NLG) evaluate their syst...
We observe a severe under-reporting of the different kinds of errors that Natural Language Generatio...
While automatically computing numerical scores remains the dominant paradigm in NLP system evaluatio...
ABSTRACT Many evaluation issues for grammatical error detection have previously been overlooked, mak...
Natural language generation (nlg) systems are computer software systems that pro-duce texts in Engli...
The Natural Language Generation (NLG) community relies on shared evaluation techniques to understand...
Rating and Likert scales are widely used in evaluation experiments to measure the quality of Natural...
Recently, there has been an increased interest in demographically grounded bias in natural language ...
This study evaluates a natural language generation system that creates literacy assessment reports i...
As Natural Language Processing (NLP) technology rapidly develops and spreads into daily life, it bec...
One of the advantages of deep grammars, such as those based on HPSG, is that they can be used for ge...
In this paper, we present the results of two re- production studies for the human evaluation origina...
With the fast-growing popularity of current large pre-trained language models (LLMs), it is necessar...
We consider the evaluation problem in Natural Language Generation (NLG) and present results for eval...
Human assessment remains the most trusted form of evaluation in NLG, but highly diverse approaches a...