Crowd Anatomy Beyond the Good and Bad: Behavioral Traces for Crowd Worker Modeling and Pre-selection
Fulltext URI
Document type
Additional Information
Journal Title
Journal ISSN
Volume Title
Publisher
Abstract
The suitability of crowdsourcing to solve a variety of problems has been investigated widely. Yet, there is still a lack of understanding about the distinct behavior and performance of workers within microtasks. In this paper, we first introduce a fine-grained data-driven worker typology based on different dimensions and derived from behavioral traces of workers. Next, we propose and evaluate novel models of crowd worker behavior and show the benefits of behavior-based worker pre-selection using machine learning models. We also study the effect of task complexity on worker behavior. Finally, we evaluate our novel typology-based worker pre-selection method in image transcription and information finding tasks involving crowd workers completing 1,800 HITs. Our proposed method for worker pre-selection leads to a higher quality of results when compared to the standard practice of using qualification or pre-screening tests. For image transcription tasks our method resulted in an accuracy increase of nearly 7% over the baseline and of almost 10% in information finding tasks, without a significant difference in task completion time. Our findings have important implications for crowdsourcing systems where a worker’s behavioral type is unknown prior to participation in a task. We highlight the potential of leveraging worker types to identify and aid those workers who require further training to improve their performance. Having proposed a powerful automated mechanism to detect worker types, we reflect on promoting fairness, trust and transparency in microtask crowdsourcing platforms.
Description
Keywords
Citation
URI
Collections
Endorsement
Review
Supplemented By
Referenced By
Number of citations to item: 24
- Sidney T Scott-Sharoni, Shashank Mehrotra, Kevin Salubre, Miao Song, Teruhisa Misu, Kumar Akash (2024): Can we enhance prosocial behavior? Using post-ride feedback to improve micromobility interactions, In: Proceedings of the 16th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, doi:10.1145/3640792.3675739
- Xicheng Yin, Hongwei Wang, Wei Wang, Kevin Zhu (2020): Task recommendation in crowdsourcing systems: A bibliometric analysis, In: Technology in Society, doi:10.1016/j.techsoc.2020.101337
- Chu Li, Zhihan Zhang, Michael Saugstad, Esteban Safranchik, Chaitanyashareef Kulkarni, Xiaoyu Huang, Shwetak Patel, Vikram Iyer, Tim Althoff, Jon E. Froehlich (2024): LabelAId: Just-in-time AI Interventions for Improving Human Labeling Quality and Domain Knowledge in Crowdsourcing Systems, In: Proceedings of the CHI Conference on Human Factors in Computing Systems, doi:10.1145/3613904.3642089
- Bhavya Bhavya, Si Chen, Zhilin Zhang, Wenting Li, Chengxiang Zhai, Lawrence Angrave, Yun Huang (2022): Exploring collaborative caption editing to augment video-based learning, In: Educational technology research and development 5(70), doi:10.1007/s11423-022-10137-5
- Esra Cemre Su de Groot, Ujwal Gadiraju (2024): "Are we all in the same boat?" Customizable and Evolving Avatars to Improve Worker Engagement and Foster a Sense of Community in Online Crowd Work, In: Proceedings of the CHI Conference on Human Factors in Computing Systems, doi:10.1145/3613904.3642429
- Ujwal Gadiraju, Gianluca Demartini (2019): Understanding Worker Moods and Reactions to Rejection in Crowdsourcing, In: Proceedings of the 30th ACM Conference on Hypertext and Social Media, doi:10.1145/3342220.3343644
- Danula Hettiachchi, Vassilis Kostakos, Jorge Goncalves (2022): A Survey on Task Assignment in Crowdsourcing, In: ACM Computing Surveys 3(55), doi:10.1145/3494522
- Run Luo, Song Huang, MeiJuan Wang, JinChang Hu, JinHu Du (2021): Unit Crowdsourcing Software Testing of Go Program, In: 2021 IEEE 21st International Conference on Software Quality, Reliability and Security (QRS), doi:10.1109/qrs54544.2021.00045
- Dennis Paulino, António Correia, João Barroso, Hugo Paredes (2023): Cognitive personalization for online microtask labor platforms: A systematic literature review, In: User Modeling and User-Adapted Interaction 3(34), doi:10.1007/s11257-023-09383-w
- António Correia, Andrea Grover, Daniel Schneider, Ana Paula Pimentel, Ramon Chaves, Marcos Antonio de Almeida, Benjamim Fonseca (2023): Designing for Hybrid Intelligence: A Taxonomy and Survey of Crowd-Machine Interaction, In: Applied Sciences 4(13), doi:10.3390/app13042198
- Guglielmo Faggioli, Laura Dietz, Charles L. A. Clarke, Gianluca Demartini, Matthias Hagen, Claudia Hauff, Noriko Kando, Evangelos Kanoulas, Martin Potthast, Benno Stein, Henning Wachsmuth (2023): Perspectives on Large Language Models for Relevance Judgment, In: Proceedings of the 2023 ACM SIGIR International Conference on Theory of Information Retrieval, doi:10.1145/3578337.3605136
- Sihang Qiu, Alessandro Bozzon, Max V. Birk, Ujwal Gadiraju (2021): Using Worker Avatars to Improve Microtask Crowdsourcing, In: Proceedings of the ACM on Human-Computer Interaction CSCW2(5), doi:10.1145/3476063
- Run Luo, Song Huang, Hao Chen, MingYu Chen (2021): Code Confusion in White Box Crowdsourced Software Testing, In: International Journal of Performability Engineering 3(17), doi:10.23940/ijpe.21.03.p3.276288
- Anne Rother, Gunther Notni, Alexander Hasse, Benjamin Noack, Christian Beyer, Jan Reißmann, Chen Zhang, Marco Ragni, Julia Arlinghaus, Myra Spiliopoulou (2023): Productive teaming under uncertainty: when a human and a machine classify objects together, In: 2023 IEEE International Conference on Advanced Robotics and Its Social Impacts (ARSO), doi:10.1109/arso56563.2023.10187430
- Donglai Fu, Yanhua Liu (2021): Remote Attestation on Behavioral Traces for Crowd Quality Control Based on Trusted Platform Module, In: Security and Communication Networks, doi:10.1155/2021/8859618
- Margeret Hall, Mohammad Farhad Afzali, Markus Krause, Simon Caton (2022): What Quality Control Mechanisms do We Need for High-Quality Crowd Work?, In: IEEE Access, doi:10.1109/access.2022.3207292
- Joseph G. Szmerekovsky, Raghavan Srinivasan, Satpal S. Wadhwa (2023): Aggregate delivery capacity planning with a crowdsourcing option, In: Computers & Industrial Engineering, doi:10.1016/j.cie.2023.109609
- Yongjun Huang, Shah Nazir, Jiyu Wu, Fida Hussain Khoso, Farhad Ali, Habib Ullah Khan (2021): An Efficient Decision Support System for the Selection of Appropriate Crowd in Crowdsourcing, In: Complexity 1(2021), doi:10.1155/2021/5518878
- Anne Rother, Uli Niemann, Tommy Hielscher, Henry Völzke, Till Ittermann, Myra Spiliopoulou (2021): Assessing the difficulty of annotating medical data in crowdworking with help of experiments, In: PLOS ONE 7(16), doi:10.1371/journal.pone.0254764
- Dennis Paulino, Diogo Guimarães, António Correia, José Ribeiro, João Barroso, Hugo Paredes (2023): A Model for Cognitive Personalization of Microtask Design, In: Sensors 7(23), doi:10.3390/s23073571
- Danzhao Cheng, Eugene Ch’ng (2022): Facilitating Situated Crowdsourcing of 3D Cultural Heritage via Asynchronous Virtual Collaboration, doi:10.21203/rs.3.rs-2245108/v1
- Olena Skrynnyk, Tetiana Vasylieva (2022): The Prediction of Leadership Degree Based on Machine Learning, In: Communications in Computer and Information Science, doi:10.1007/978-3-031-14841-5_6
- Ujwal Gadiraju, Mengdie Zhuang (2019): What You Sow, So Shall You Reap! Toward Preselection Mechanisms for Macrotask Crowdsourcing, In: Human–Computer Interaction Series, doi:10.1007/978-3-030-12334-5_6
- Michael A. Hedderich, Antti Oulasvirta (2024): Explaining crowdworker behaviour through computational rationality, In: Behaviour & Information Technology 3(44), doi:10.1080/0144929x.2024.2329616