Crowdsourcing is a popular technique to collect large amounts of human-generated labels, such as relevance judgments used to create information retrieval (IR) evaluation collections. Previous research has shown how collecting high quality labels from a crowdsourcing platform can be challenging. Existing quality assurance techniques focus on answer aggregation or on the use of gold questions where ground-truth data allows to check for the quality of the responses.In this paper, we present qualitative and quantitative results, revealing how different crowd workers adopt different work strategies to complete relevance judgment tasks efficiently and their consequent impact on quality. We delve into the techniques and tools that highly experienced crowd workers use to be more efficient in completing crowdsourcing micro-tasks. To this end, we use both qualitative results from worker interviews and surveys, as well as the results of a data-driven study of behavioral log data (i.e., clicks, keystrokes and keyboard shortcuts) collected from crowd workers performing relevance judgment tasks. Our results highlight the presence of frequently used shortcut patterns that can speed-up task completion, thus increasing the hourly wage of efficient workers. We observe how crowd work experiences result in different types of working strategies, productivity levels, quality and diversity of the crowdsourced judgments.

Crowd Worker Strategies in Relevance Judgment Tasks / Han, Lei; Maddalena, Eddy; Checco, Alessandro; Sarasua, Cristina; Gadiraju, Ujwal; Roitero, Kevin; Demartini, Gianluca. - (2020), pp. 241-249. (Intervento presentato al convegno WSDM '20: The Thirteenth ACM International Conference on Web Search and Data Mining tenutosi a Houston TX; USA) [10.1145/3336191.3371857].

Crowd Worker Strategies in Relevance Judgment Tasks

Alessandro Checco;
2020

Abstract

Crowdsourcing is a popular technique to collect large amounts of human-generated labels, such as relevance judgments used to create information retrieval (IR) evaluation collections. Previous research has shown how collecting high quality labels from a crowdsourcing platform can be challenging. Existing quality assurance techniques focus on answer aggregation or on the use of gold questions where ground-truth data allows to check for the quality of the responses.In this paper, we present qualitative and quantitative results, revealing how different crowd workers adopt different work strategies to complete relevance judgment tasks efficiently and their consequent impact on quality. We delve into the techniques and tools that highly experienced crowd workers use to be more efficient in completing crowdsourcing micro-tasks. To this end, we use both qualitative results from worker interviews and surveys, as well as the results of a data-driven study of behavioral log data (i.e., clicks, keystrokes and keyboard shortcuts) collected from crowd workers performing relevance judgment tasks. Our results highlight the presence of frequently used shortcut patterns that can speed-up task completion, thus increasing the hourly wage of efficient workers. We observe how crowd work experiences result in different types of working strategies, productivity levels, quality and diversity of the crowdsourced judgments.
2020
WSDM '20: The Thirteenth ACM International Conference on Web Search and Data Mining
crowdsourincg; relevance judgment; information retrieval
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Crowd Worker Strategies in Relevance Judgment Tasks / Han, Lei; Maddalena, Eddy; Checco, Alessandro; Sarasua, Cristina; Gadiraju, Ujwal; Roitero, Kevin; Demartini, Gianluca. - (2020), pp. 241-249. (Intervento presentato al convegno WSDM '20: The Thirteenth ACM International Conference on Web Search and Data Mining tenutosi a Houston TX; USA) [10.1145/3336191.3371857].
File allegati a questo prodotto
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1696316
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 25
  • ???jsp.display-item.citation.isi??? ND
social impact