While automatically computing numerical scores remains the dominant paradigm in NLP system evaluation, error analysis is receiving increasing attention, with numerous error annotation schemes being proposed for automatically generated text. However, there is little agreement about what error annotation schemes should look like, how many different types of errors should be distinguished and at what level of granularity. In this paper, our aim is to map out recent work on annotating errors in automatically generated text, with a particular focus on error taxonomies. We describe our systematic paper selection process, and survey the error annotation schemes reported in the papers, drawing out similarities and differences between them. Finally,...
Quality Estimation (QE) and error analysis of Machine Translation (MT) output remain active areas in...
Since the emergence of the first fully automatic machine translation (MT) systems over fifty years a...
This work proposes a new method for manual evaluation of Machine Translation (MT) output based on ma...
While automatically computing numerical scores remains the dominant paradigm in NLP system evaluatio...
This is the accompanying data for our paper "Annotation Error Detection: Analyzing the Past and Pres...
Earlier research has shown that few studies in Natural Language Generation (NLG) evaluate their syst...
ABSTRACT Many evaluation issues for grammatical error detection have previously been overlooked, mak...
We observe a severe under-reporting of the different kinds of errors that Natural Language Generatio...
In this thesis, we investigate methods for automatic detection, and to some extent correction, of gr...
This work examines different ways of aggregating scores for error annotation in MT outputs: raw erro...
Grammatical Error Correction (GEC) is the task of automatically detecting and correcting grammatical...
We introduce a method for error detection in automatically annotated text, aimed at supporting the c...
Shortage of available training data is holding back progress in the area of automated error detectio...
PEst-OE/EEI/LA0021/2013 SFRH/BD/85737/2012In this paper we describe a corpus of automatic translatio...
We describe the creation of an annotation layer for word-based writing errors for a corpus of studen...
Quality Estimation (QE) and error analysis of Machine Translation (MT) output remain active areas in...
Since the emergence of the first fully automatic machine translation (MT) systems over fifty years a...
This work proposes a new method for manual evaluation of Machine Translation (MT) output based on ma...
While automatically computing numerical scores remains the dominant paradigm in NLP system evaluatio...
This is the accompanying data for our paper "Annotation Error Detection: Analyzing the Past and Pres...
Earlier research has shown that few studies in Natural Language Generation (NLG) evaluate their syst...
ABSTRACT Many evaluation issues for grammatical error detection have previously been overlooked, mak...
We observe a severe under-reporting of the different kinds of errors that Natural Language Generatio...
In this thesis, we investigate methods for automatic detection, and to some extent correction, of gr...
This work examines different ways of aggregating scores for error annotation in MT outputs: raw erro...
Grammatical Error Correction (GEC) is the task of automatically detecting and correcting grammatical...
We introduce a method for error detection in automatically annotated text, aimed at supporting the c...
Shortage of available training data is holding back progress in the area of automated error detectio...
PEst-OE/EEI/LA0021/2013 SFRH/BD/85737/2012In this paper we describe a corpus of automatic translatio...
We describe the creation of an annotation layer for word-based writing errors for a corpus of studen...
Quality Estimation (QE) and error analysis of Machine Translation (MT) output remain active areas in...
Since the emergence of the first fully automatic machine translation (MT) systems over fifty years a...
This work proposes a new method for manual evaluation of Machine Translation (MT) output based on ma...