Crowd Anatomy Beyond the Good and Bad: Behavioral Traces for Crowd Worker Modeling and Pre-selection

dc.contributor.authorGadiraju, Ujwal
dc.contributor.authorDemartini, Gianluca
dc.contributor.authorKawase, Ricardo
dc.contributor.authorDietze, Stefan
dc.date.accessioned2020-06-06T13:06:08Z
dc.date.available2020-06-06T13:06:08Z
dc.date.issued2019
dc.date.issued2019
dc.description.abstractThe suitability of crowdsourcing to solve a variety of problems has been investigated widely. Yet, there is still a lack of understanding about the distinct behavior and performance of workers within microtasks. In this paper, we first introduce a fine-grained data-driven worker typology based on different dimensions and derived from behavioral traces of workers. Next, we propose and evaluate novel models of crowd worker behavior and show the benefits of behavior-based worker pre-selection using machine learning models. We also study the effect of task complexity on worker behavior. Finally, we evaluate our novel typology-based worker pre-selection method in image transcription and information finding tasks involving crowd workers completing 1,800 HITs. Our proposed method for worker pre-selection leads to a higher quality of results when compared to the standard practice of using qualification or pre-screening tests. For image transcription tasks our method resulted in an accuracy increase of nearly 7% over the baseline and of almost 10% in information finding tasks, without a significant difference in task completion time. Our findings have important implications for crowdsourcing systems where a worker’s behavioral type is unknown prior to participation in a task. We highlight the potential of leveraging worker types to identify and aid those workers who require further training to improve their performance. Having proposed a powerful automated mechanism to detect worker types, we reflect on promoting fairness, trust and transparency in microtask crowdsourcing platforms.de
dc.identifier.doi10.1007/s10606-018-9336-y
dc.identifier.pissn1573-7551
dc.identifier.urihttp://dx.doi.org/10.1007/s10606-018-9336-y
dc.identifier.urihttps://dl.eusset.eu/handle/20.500.12015/3746
dc.publisherSpringer
dc.relation.ispartofComputer Supported Cooperative Work (CSCW): Vol. 28, No. 5
dc.relation.ispartofseriesComputer Supported Cooperative Work (CSCW)
dc.subjectBehavioral traces
dc.subjectCrowdsourcing
dc.subjectMicrotasks
dc.subjectPre-screening
dc.subjectPre-selection
dc.subjectWorker typology
dc.subjectWorkers
dc.titleCrowd Anatomy Beyond the Good and Bad: Behavioral Traces for Crowd Worker Modeling and Pre-selectionde
dc.typeText/Journal Article
gi.citation.endPage841
gi.citation.startPage815
gi.citations.count24
gi.citations.elementSidney T Scott-Sharoni, Shashank Mehrotra, Kevin Salubre, Miao Song, Teruhisa Misu, Kumar Akash (2024): Can we enhance prosocial behavior? Using post-ride feedback to improve micromobility interactions, In: Proceedings of the 16th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, doi:10.1145/3640792.3675739
gi.citations.elementXicheng Yin, Hongwei Wang, Wei Wang, Kevin Zhu (2020): Task recommendation in crowdsourcing systems: A bibliometric analysis, In: Technology in Society, doi:10.1016/j.techsoc.2020.101337
gi.citations.elementChu Li, Zhihan Zhang, Michael Saugstad, Esteban Safranchik, Chaitanyashareef Kulkarni, Xiaoyu Huang, Shwetak Patel, Vikram Iyer, Tim Althoff, Jon E. Froehlich (2024): LabelAId: Just-in-time AI Interventions for Improving Human Labeling Quality and Domain Knowledge in Crowdsourcing Systems, In: Proceedings of the CHI Conference on Human Factors in Computing Systems, doi:10.1145/3613904.3642089
gi.citations.elementBhavya Bhavya, Si Chen, Zhilin Zhang, Wenting Li, Chengxiang Zhai, Lawrence Angrave, Yun Huang (2022): Exploring collaborative caption editing to augment video-based learning, In: Educational technology research and development 5(70), doi:10.1007/s11423-022-10137-5
gi.citations.elementEsra Cemre Su de Groot, Ujwal Gadiraju (2024): "Are we all in the same boat?" Customizable and Evolving Avatars to Improve Worker Engagement and Foster a Sense of Community in Online Crowd Work, In: Proceedings of the CHI Conference on Human Factors in Computing Systems, doi:10.1145/3613904.3642429
gi.citations.elementUjwal Gadiraju, Gianluca Demartini (2019): Understanding Worker Moods and Reactions to Rejection in Crowdsourcing, In: Proceedings of the 30th ACM Conference on Hypertext and Social Media, doi:10.1145/3342220.3343644
gi.citations.elementDanula Hettiachchi, Vassilis Kostakos, Jorge Goncalves (2022): A Survey on Task Assignment in Crowdsourcing, In: ACM Computing Surveys 3(55), doi:10.1145/3494522
gi.citations.elementRun Luo, Song Huang, MeiJuan Wang, JinChang Hu, JinHu Du (2021): Unit Crowdsourcing Software Testing of Go Program, In: 2021 IEEE 21st International Conference on Software Quality, Reliability and Security (QRS), doi:10.1109/qrs54544.2021.00045
gi.citations.elementDennis Paulino, António Correia, João Barroso, Hugo Paredes (2023): Cognitive personalization for online microtask labor platforms: A systematic literature review, In: User Modeling and User-Adapted Interaction 3(34), doi:10.1007/s11257-023-09383-w
gi.citations.elementAntónio Correia, Andrea Grover, Daniel Schneider, Ana Paula Pimentel, Ramon Chaves, Marcos Antonio de Almeida, Benjamim Fonseca (2023): Designing for Hybrid Intelligence: A Taxonomy and Survey of Crowd-Machine Interaction, In: Applied Sciences 4(13), doi:10.3390/app13042198
gi.citations.elementGuglielmo Faggioli, Laura Dietz, Charles L. A. Clarke, Gianluca Demartini, Matthias Hagen, Claudia Hauff, Noriko Kando, Evangelos Kanoulas, Martin Potthast, Benno Stein, Henning Wachsmuth (2023): Perspectives on Large Language Models for Relevance Judgment, In: Proceedings of the 2023 ACM SIGIR International Conference on Theory of Information Retrieval, doi:10.1145/3578337.3605136
gi.citations.elementSihang Qiu, Alessandro Bozzon, Max V. Birk, Ujwal Gadiraju (2021): Using Worker Avatars to Improve Microtask Crowdsourcing, In: Proceedings of the ACM on Human-Computer Interaction CSCW2(5), doi:10.1145/3476063
gi.citations.elementRun Luo, Song Huang, Hao Chen, MingYu Chen (2021): Code Confusion in White Box Crowdsourced Software Testing, In: International Journal of Performability Engineering 3(17), doi:10.23940/ijpe.21.03.p3.276288
gi.citations.elementAnne Rother, Gunther Notni, Alexander Hasse, Benjamin Noack, Christian Beyer, Jan Reißmann, Chen Zhang, Marco Ragni, Julia Arlinghaus, Myra Spiliopoulou (2023): Productive teaming under uncertainty: when a human and a machine classify objects together, In: 2023 IEEE International Conference on Advanced Robotics and Its Social Impacts (ARSO), doi:10.1109/arso56563.2023.10187430
gi.citations.elementDonglai Fu, Yanhua Liu (2021): Remote Attestation on Behavioral Traces for Crowd Quality Control Based on Trusted Platform Module, In: Security and Communication Networks, doi:10.1155/2021/8859618
gi.citations.elementMargeret Hall, Mohammad Farhad Afzali, Markus Krause, Simon Caton (2022): What Quality Control Mechanisms do We Need for High-Quality Crowd Work?, In: IEEE Access, doi:10.1109/access.2022.3207292
gi.citations.elementJoseph G. Szmerekovsky, Raghavan Srinivasan, Satpal S. Wadhwa (2023): Aggregate delivery capacity planning with a crowdsourcing option, In: Computers & Industrial Engineering, doi:10.1016/j.cie.2023.109609
gi.citations.elementYongjun Huang, Shah Nazir, Jiyu Wu, Fida Hussain Khoso, Farhad Ali, Habib Ullah Khan (2021): An Efficient Decision Support System for the Selection of Appropriate Crowd in Crowdsourcing, In: Complexity 1(2021), doi:10.1155/2021/5518878
gi.citations.elementAnne Rother, Uli Niemann, Tommy Hielscher, Henry Völzke, Till Ittermann, Myra Spiliopoulou (2021): Assessing the difficulty of annotating medical data in crowdworking with help of experiments, In: PLOS ONE 7(16), doi:10.1371/journal.pone.0254764
gi.citations.elementDennis Paulino, Diogo Guimarães, António Correia, José Ribeiro, João Barroso, Hugo Paredes (2023): A Model for Cognitive Personalization of Microtask Design, In: Sensors 7(23), doi:10.3390/s23073571
gi.citations.elementDanzhao Cheng, Eugene Ch’ng (2022): Facilitating Situated Crowdsourcing of 3D Cultural Heritage via Asynchronous Virtual Collaboration, doi:10.21203/rs.3.rs-2245108/v1
gi.citations.elementOlena Skrynnyk, Tetiana Vasylieva (2022): The Prediction of Leadership Degree Based on Machine Learning, In: Communications in Computer and Information Science, doi:10.1007/978-3-031-14841-5_6
gi.citations.elementUjwal Gadiraju, Mengdie Zhuang (2019): What You Sow, So Shall You Reap! Toward Preselection Mechanisms for Macrotask Crowdsourcing, In: Human–Computer Interaction Series, doi:10.1007/978-3-030-12334-5_6
gi.citations.elementMichael A. Hedderich, Antti Oulasvirta (2024): Explaining crowdworker behaviour through computational rationality, In: Behaviour & Information Technology 3(44), doi:10.1080/0144929x.2024.2329616

Files