Published:
2018-07-09
Proceedings:
Proceedings of the AAAI Conference on Human Computation and Crowdsourcing, 6
Volume
Issue:
Vol. 6 (2018): Sixth AAAI Conference on Human Computation and Crowdsourcing
Track:
Technical Papers
Downloads:
Abstract:
While peer-agreement and gold checks are well-established methods for ensuring quality in crowdsourced data collection, we explore a relatively new direction for quality control: estimating work quality directly from workers’ behavioral traces collected during annotation. We propose three behavior-based models to predict label correctness and worker accuracy, then further apply model predictions to label aggregation and optimization of label collection. As part of this work, we collect and share a new Mechanical Turk dataset of behavioral signals judging the relevance of search results. Results show that behavioral data can be effectively used to predict work quality, which could be especially useful with single labeling or in a cold start scenario in which individuals’ prior work history is unavailable. We further show improvement in label aggregation and reducing labeling cost while ensuring data quality.
DOI:
10.1609/hcomp.v6i1.13331
HCOMP
Vol. 6 (2018): Sixth AAAI Conference on Human Computation and Crowdsourcing
ISBN 978-1-57735-799-5