This is the data obtained from crowdsourcing tasks which ask workers to provide similarity metrics between pairs of documents. Each document, as well as each pair, has a unique ID. We provide crowd workers with the pairs through three different task variations: Variation 1: We showed workers 5 pairs of documents and, for each, asked them to rate their similarity in a 4-level Likert scale (None, Low, Medium, High), tell us a confidence level of how sure they were (from 0 to 4) and a written reason as to why they chose that similarity level. For quality reasons, two of the 5 pairs were golden-standards, which means we knew their ratings already and checked the workers' responses. They had to give the golden pair with the higher similarity ...
For many applications measuring the similarity between documents is essential. However, little is kn...
Many important data management and analytics tasks cannot be completely addressed by automated proce...
ABSTRACT How can we best use crowdsourcing to perform a subjective labeling task with low inter-rate...
Understanding natural language is an inherently complex task for computer algorithms. Crowdsourcing ...
Document similarity measures are crucial components of many text-analysis tasks, including informati...
This paper explores a crowdsourcing approach to the evaluation of a document recommender system inte...
Crowdsourcing has become an alternative approach to collect relevance judgments at scale thanks to t...
htmlabstractThe performance of information retrieval (IR) systems is commonly evaluated using a test...
Crowdsourcing has become an alternative approach to collect relevance judgments at large scale. In t...
Crowdsourcing has become an alternative approach to collect relevance judgments at scale thanks to t...
Social (or folksonomic) tagging has become a very popular way to describe content within Web 2.0 web...
This paper explores a crowdsourcing approach to the evalua-tion of a document recommender system int...
The internet enables us to collect and store unprecedented amounts of data. We need better models fo...
Abstract. We consider the problem of acquiring relevance judgements for in-formation retrieval (IR) ...
This survey deals with the problem of evaluating the submissions to crowdsourcing websites on which ...
For many applications measuring the similarity between documents is essential. However, little is kn...
Many important data management and analytics tasks cannot be completely addressed by automated proce...
ABSTRACT How can we best use crowdsourcing to perform a subjective labeling task with low inter-rate...
Understanding natural language is an inherently complex task for computer algorithms. Crowdsourcing ...
Document similarity measures are crucial components of many text-analysis tasks, including informati...
This paper explores a crowdsourcing approach to the evaluation of a document recommender system inte...
Crowdsourcing has become an alternative approach to collect relevance judgments at scale thanks to t...
htmlabstractThe performance of information retrieval (IR) systems is commonly evaluated using a test...
Crowdsourcing has become an alternative approach to collect relevance judgments at large scale. In t...
Crowdsourcing has become an alternative approach to collect relevance judgments at scale thanks to t...
Social (or folksonomic) tagging has become a very popular way to describe content within Web 2.0 web...
This paper explores a crowdsourcing approach to the evalua-tion of a document recommender system int...
The internet enables us to collect and store unprecedented amounts of data. We need better models fo...
Abstract. We consider the problem of acquiring relevance judgements for in-formation retrieval (IR) ...
This survey deals with the problem of evaluating the submissions to crowdsourcing websites on which ...
For many applications measuring the similarity between documents is essential. However, little is kn...
Many important data management and analytics tasks cannot be completely addressed by automated proce...
ABSTRACT How can we best use crowdsourcing to perform a subjective labeling task with low inter-rate...