Conference Paper

Enabling Uneven Task Difficulty in Micro-Task Crowdsourcing

Fulltext URI

Document type

Text/Conference Paper

Additional Information

Date

2018

Journal Title

Journal ISSN

Volume Title

Publisher

Association for Computing Machinery

Abstract

In micro-task crowdsourcing markets such as Amazon's Mechanical Turk, how to obtain high quality result without exceeding the limited budgets is one main challenge. The existing theory and practice of crowdsourcing suggests that uneven task difficulty plays a crucial role to task quality. Yet, it lacks a clear identifying method to task difficulty, which hinders effective and efficient execution of micro-task crowdsourcing. This paper explores the notion of task difficulty and its influence to crowdsourcing, and presents a difficulty-based crowdsourcing method to optimize the crowdsourcing process. We firstly identify task difficulty feature based on a local estimation method in the real crowdsourcing context, followed by proposing an optimization method to improve the accuracy of results, while reducing the overall cost. We conduct a series of experimental studies to evaluate our method, which show that our difficulty-based crowdsourcing method can accurately identify the task difficulty feature, improve the quality of task performance and reduce the cost significantly, and thus demonstrate the effectiveness of task difficulty as task modeling property.

Description

Jiang, Yu; Sun, Yuling; Yang, Jing; Lin, Xin; He, Liang (2018): Enabling Uneven Task Difficulty in Micro-Task Crowdsourcing. Proceedings of the 2018 ACM International Conference on Supporting Group Work. DOI: 10.1145/3148330.3148342. Association for Computing Machinery. pp. 12–21. Sanibel Island, Florida, USA

Citation

Tags

Collections