This paper describes a methodology for supporting the task of annotating sentiment in natural language by detecting borderline cases and inconsistencies. Inspired by the co-training strategy, a number of machine learning models are trained on different views of the same data. The predic-tions obtained by these models are then automatically compared in order to bring to light highly uncertain annotations and systematic mistakes. We tested the methodology against an English corpus annotated according to a fine-grained sentiment analysis annotation schema (SentiML). We detected that 153 instances (35%) classified differently from the gold standard were accept-able and further 69 instances (16%) suggested that the gold standard should have been...
This work describes an analysis of inter-annotator disagreements in human evaluation of machine tran...
International audienceComputing inter-annotator agreement measures on a manually annotated corpus is...
International audienceComputing inter-annotator agreement measures on a manually annotated corpus is...
Annotated data is an essential ingredient in natural language processing for training and evaluating...
This is the accompanying data for our paper "Annotation Error Detection: Analyzing the Past and Pres...
We investigate how disagreement in natural language inference (NLI) annotation arises. We developed ...
We investigate how disagreement in natural language inference (NLI) annotation arises. We developed ...
The availability of annotated data is an important prerequisite for the development of machine learn...
Natural language inference (NLI) is the task of determining whether a piece of text is entailed, con...
Natural language inference (NLI) is the task of determining whether a piece of text is entailed, con...
A common way to express sentiment about some product is by comparing it to a different product. The ...
The usual practice in assessing whether a multimodal annotated corpus is fit for purpose is to calcu...
Neural methods for sentiment analysis have led to quantitative improvements over previous approaches...
This work describes an analysis of inter-annotator disagreements in human evaluation of machine tran...
Sentiments are positive and negative emotions, evaluations and stances. This dissertation focuses on...
This work describes an analysis of inter-annotator disagreements in human evaluation of machine tran...
International audienceComputing inter-annotator agreement measures on a manually annotated corpus is...
International audienceComputing inter-annotator agreement measures on a manually annotated corpus is...
Annotated data is an essential ingredient in natural language processing for training and evaluating...
This is the accompanying data for our paper "Annotation Error Detection: Analyzing the Past and Pres...
We investigate how disagreement in natural language inference (NLI) annotation arises. We developed ...
We investigate how disagreement in natural language inference (NLI) annotation arises. We developed ...
The availability of annotated data is an important prerequisite for the development of machine learn...
Natural language inference (NLI) is the task of determining whether a piece of text is entailed, con...
Natural language inference (NLI) is the task of determining whether a piece of text is entailed, con...
A common way to express sentiment about some product is by comparing it to a different product. The ...
The usual practice in assessing whether a multimodal annotated corpus is fit for purpose is to calcu...
Neural methods for sentiment analysis have led to quantitative improvements over previous approaches...
This work describes an analysis of inter-annotator disagreements in human evaluation of machine tran...
Sentiments are positive and negative emotions, evaluations and stances. This dissertation focuses on...
This work describes an analysis of inter-annotator disagreements in human evaluation of machine tran...
International audienceComputing inter-annotator agreement measures on a manually annotated corpus is...
International audienceComputing inter-annotator agreement measures on a manually annotated corpus is...