In natural language processing (NLP) an-notation projects, we use inter-annotator agreement measures and annotation guide-lines to ensure consistent annotations. However, annotation guidelines often make linguistically debatable and even somewhat arbitrary decisions, and inter-annotator agreement is often less than perfect. While annotation projects usu-ally specify how to deal with linguisti-cally debatable phenomena, annotator dis-agreements typically still stem from these “hard ” cases. This indicates that some er-rors are more debatable than others. In this paper, we use small samples of doubly-annotated part-of-speech (POS) data for Twitter to estimate annotation reliability and show how those metrics of likely inter-annotator agreemen...
This article presents our work on constructing a corpus of news articles in which events are annotat...
We commonly use agreement measures to assess the utility of judgements made by human annotators in N...
Abstract We present a first analysis of interannotator agreement for the DIT ++ tagset of dialogue a...
In natural language processing (NLP) annotation projects, we use inter-annotator agreement measures ...
International audienceLinguistic annotation underlies many successful approaches in Natural Language...
International audienceIn this abstract we present a methodology to improve Argument annotation guide...
Standard agreement measures for interannota-tor reliability are neither necessary nor suffi-cient to...
Corpus linguistic and language technological research needs empirical corpus data with nearly correc...
Several works in Natural Language Processing have recently looked into part-of-speech (POS) annotati...
In NLP annotation, it is common to have multiple annotators label the text and then obtain the groun...
For one aspect of grammatical annotation, part-of-speech tagging, we investigate experimentally whet...
Computing inter-annotator agreement measures on a manually annotated corpus is necessary to evaluate...
© 2005 Andrew MacKinlayIn natural language processing (NLP), a crucial subsystem in a wide range of ...
This article details a series of carefully de-signed experiments aiming at evaluating the influence ...
Since state-of-the-art approaches to offensive language detection rely on supervised learning, it is...
This article presents our work on constructing a corpus of news articles in which events are annotat...
We commonly use agreement measures to assess the utility of judgements made by human annotators in N...
Abstract We present a first analysis of interannotator agreement for the DIT ++ tagset of dialogue a...
In natural language processing (NLP) annotation projects, we use inter-annotator agreement measures ...
International audienceLinguistic annotation underlies many successful approaches in Natural Language...
International audienceIn this abstract we present a methodology to improve Argument annotation guide...
Standard agreement measures for interannota-tor reliability are neither necessary nor suffi-cient to...
Corpus linguistic and language technological research needs empirical corpus data with nearly correc...
Several works in Natural Language Processing have recently looked into part-of-speech (POS) annotati...
In NLP annotation, it is common to have multiple annotators label the text and then obtain the groun...
For one aspect of grammatical annotation, part-of-speech tagging, we investigate experimentally whet...
Computing inter-annotator agreement measures on a manually annotated corpus is necessary to evaluate...
© 2005 Andrew MacKinlayIn natural language processing (NLP), a crucial subsystem in a wide range of ...
This article details a series of carefully de-signed experiments aiming at evaluating the influence ...
Since state-of-the-art approaches to offensive language detection rely on supervised learning, it is...
This article presents our work on constructing a corpus of news articles in which events are annotat...
We commonly use agreement measures to assess the utility of judgements made by human annotators in N...
Abstract We present a first analysis of interannotator agreement for the DIT ++ tagset of dialogue a...