skip to main content
10.1145/3531146.3534629acmotherconferencesArticle/Chapter ViewAbstractPublication PagesfacctConference Proceedingsconference-collections
research-article
Open Access

The Effects of Crowd Worker Biases in Fact-Checking Tasks

Published:20 June 2022Publication History

ABSTRACT

Due to the increasing amount of information shared online every day, the need for sound and reliable ways of distinguishing between trustworthy and non-trustworthy information is as present as ever. One technique for performing fact-checking at scale is to employ human intelligence in the form of crowd workers. Although earlier work has suggested that crowd workers can reliably identify misinformation, cognitive biases of crowd workers may reduce the quality of truthfulness judgments in this context. We performed a systematic exploratory analysis of publicly available crowdsourced data to identify a set of potential systematic biases that may occur when crowd workers perform fact-checking tasks. Following this exploratory study, we collected a novel data set of crowdsourced truthfulness judgments to validate our hypotheses. Our findings suggest that workers generally overestimate the truthfulness of statements and that different individual characteristics (i.e., their belief in science) and cognitive biases (i.e., the affect heuristic and overconfidence) can affect their annotations. Interestingly, we find that, depending on the general judgment tendencies of workers, their biases may sometimes lead to more accurate judgments.

References

  1. Jennifer Allen, Antonio A Arechar, Gordon Pennycook, and David G Rand. 2021. Scaling up fact-checking using the wisdom of crowds. Science advances 7, 36 (2021), eabf4393.Google ScholarGoogle Scholar
  2. Ricardo Baeza-Yates. 2018. Bias on the Web. Commun. ACM 61, 6 (2018), 54–61.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Momen Bhuiyan, Amy Zhang, Connie Sehat, and Tanushree Mitra. 2020. Investigating “Who” in the Crowdsourcing of News Credibility. In Computational Journalism Symposium.Google ScholarGoogle Scholar
  4. Praveen Chandar and Ben Carterette. 2018. Estimating Clickthrough Bias in the Cascade Model. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management (Torino, Italy) (CIKM ’18). ACM, New York, NY, USA, 1587–1590.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Jiawei Chen, Hande Dong, Xiang Wang, Fuli Feng, Meng Wang, and Xiangnan He. 2020. Bias and Debias in Recommender System: A Survey and Future Directions. arXiv preprint arXiv:2010.03240(2020).Google ScholarGoogle Scholar
  6. Neil Dagnall, Andrew Denovan, Kenneth Graham Drinkwater, and Andrew Parker. 2019. An Evaluation of the Belief in Science Scale. Frontiers in Psychology 10 (2019), 861.Google ScholarGoogle ScholarCross RefCross Ref
  7. Gianluca Demartini, Stefano Mizzaro, and Damiano Spina. 2020. Human-in-the-loop Artificial Intelligence for Fighting Online Misinformation: Challenges and Opportunities. Bulletin of IEEE Computer Society 43, 3 (2020), 65–74.Google ScholarGoogle Scholar
  8. Tim Draws, Oana Inel, Nava Tintarev, Christian Baden, and Benjamin Timmermans. 2022. Comprehensive Viewpoint Representations for a Deeper Understanding of User Interactions With Debated Topics. In ACM SIGIR Conference on Human Information Interaction and Retrieval (Regensburg, Germany) (CHIIR ’22). Association for Computing Machinery, New York, NY, USA, 135–145. https://doi.org/10.1145/3498366.3505812Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Tim Draws, Alisa Rieger, Oana Inel, Ujwal Gadiraju, and Nava Tintarev. 2021. A Checklist to Combat Cognitive Biases in Crowdsourcing. Proceedings of the AAAI Conference on Human Computation and Crowdsourcing 9, 1 (Oct. 2021), 48–59. https://ojs.aaai.org/index.php/HCOMP/article/view/18939Google ScholarGoogle ScholarCross RefCross Ref
  10. Tim Draws, Nava Tintarev, Ujwal Gadiraju, Alessandro Bozzon, and Benjamin Timmermans. 2021. This Is Not What We Ordered: Exploring Why Biased Search Result Rankings Affect User Attitudes on Debated Topics. ACM, New York, NY, USA, 295–305.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Carsten Eickhoff. 2018. Cognitive Biases in Crowdsourcing. In Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining (WSDM). ACM, 162–170.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Robert Epstein and Ronald E. Robertson. 2015. The search engine manipulation effect (SEME) and its possible impact on the outcomes of elections. Proceedings of the National Academy of Sciences 112, 33(2015), E4512–E4521.Google ScholarGoogle ScholarCross RefCross Ref
  13. Ziv Epstein, Gordon Pennycook, and David Rand. 2020. Will the Crowd Game the Algorithm? Using Layperson Judgments to Combat Misinformation on Social Media by Downranking Distrusted Sources. ACM, New York, NY, USA, 1–11.Google ScholarGoogle Scholar
  14. Edgar Erdfelder, Franz Faul, and Axel Buchner. 1996. GPOWER: A general power analysis program. Behavior Research Methods, Instruments, & Computers 28, 1 (01 Mar 1996), 1–11.Google ScholarGoogle Scholar
  15. Shane Frederick. 2005. Cognitive Reflection and Decision Making. Journal of Economic Perspectives 19, 4 (12 2005), 25–42.Google ScholarGoogle ScholarCross RefCross Ref
  16. Amira Ghenai and Yelena Mejova. 2017. Catching Zika Fever: Application of Crowdsourcing and Machine Learning for Tracking Health Misinformation on Twitter. In 2017 IEEE International Conference on Healthcare Informatics (ICHI). 518–518.Google ScholarGoogle Scholar
  17. Anastasia Giachanou and Paolo Rosso. 2020. The Battle Against Online Harmful Information: The Cases of Fake News and Hate Speech. Association for Computing Machinery, New York, NY, USA, 3503–3504.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Stephan Grimmelikhuijsen and Eva Knies. 2017. Validating a scale for citizen trust in government organizations. International Review of Administrative Sciences 83, 3 (2017), 583–601.Google ScholarGoogle ScholarCross RefCross Ref
  19. Matthew Haigh. 2016. Has the Standard Cognitive Reflection Test Become a Victim of Its Own Success?Advances in cognitive psychology 12, 3 (30 Sep 2016), 145–149.Google ScholarGoogle Scholar
  20. Lei Han, Kevin Roitero, Ujwal Gadiraju, Cristina Sarasua, Alessandro Checco, Eddy Maddalena, and Gianluca Demartini. 2019. The Impact of Task Abandonment in Crowdsourcing. IEEE Transactions on Knowledge and Data Engineering (2019), 1–1.Google ScholarGoogle ScholarCross RefCross Ref
  21. Christoph Hube, Besnik Fetahu, and Ujwal Gadiraju. 2019. Understanding and Mitigating Worker Biases in the Crowdsourced Collection of Subjective Judgments. In Proceedings of CHI. 12 pages.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Klaus Krippendorff. 2011. Computing Krippendorff’s Alpha-Reliability. (2011).Google ScholarGoogle Scholar
  23. David La Barbera, Kevin Roitero, Damiano Spina, Stefano Mizzaro, and Gianluca Demartini. 2020. Crowdsourcing Truthfulness: The Impact of Judgment Scale and Assessor Bias. In Proceedings of ECIR. Springer, New York, NY, USA, 207–214.Google ScholarGoogle Scholar
  24. Eun-Ju Lee. 2012. That’s Not the Way It Is: How User-Generated Comments on the News Affect Perceived Media Bias. Journal of Computer-Mediated Communication 18, 1 (2012), 32–45.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Kevin G. Love. 1981. Comparison of peer assessment methods: Reliability, validity, friendship bias, and user reaction.Journal of Applied Psychology 66, 4 (1981), 451.Google ScholarGoogle Scholar
  26. Eddy Maddalena, Davide Ceolin, and Stefano Mizzaro. 2018. Multidimensional News Quality: A Comparison of Crowdsourcing and Nichesourcing.. In CIKM Workshops. ACM, New York, NY, USA.Google ScholarGoogle Scholar
  27. Saif Mohammad, Svetlana Kiritchenko, Parinaz Sobhani, Xiaodan Zhu, and Colin Cherry. 2016. Semeval-2016 task 6: Detecting stance in tweets. In Proceedings of the 10th international workshop on semantic evaluation (SemEval-2016). 31–41.Google ScholarGoogle ScholarCross RefCross Ref
  28. Susan Morgan. 2018. Fake news, disinformation, manipulation and online tactics to undermine democracy. Journal of Cyber Policy 3, 1 (2018), 39–43. https://doi.org/10.1080/23738871.2018.1462395 arXiv:https://doi.org/10.1080/23738871.2018.1462395Google ScholarGoogle ScholarCross RefCross Ref
  29. Lev Muchnik, Sinan Aral, and Sean J Taylor. 2013. Social influence bias: A randomized experiment. Science 341, 6146 (2013), 647–651.Google ScholarGoogle Scholar
  30. Gordon Pennycook and David G Rand. 2018. Crowdsourcing judgments of news source quality. SSRN. com (2018).Google ScholarGoogle Scholar
  31. Gordon Pennycook and David G Rand. 2019. Fighting Misinformation on Social Media Using Crowdsourced Judgments of News Source Quality. Proceedings of the National Academy of Sciences 116, 7 (2019), 2521–2526.Google ScholarGoogle ScholarCross RefCross Ref
  32. Marcos Rodrigues Pinto, Yuri Oliveira de Lima, Carlos Eduardo Barbosa, and Jano Moreira de Souza. 2019. Towards fact-checking through crowdsourcing. In 2019 IEEE 23rd International Conference on Computer Supported Cooperative Work in Design (CSCWD). IEEE, 494–499.Google ScholarGoogle ScholarCross RefCross Ref
  33. Frances A. Pogacar, Amira Ghenai, Mark D. Smucker, and Charles L.A. Clarke. 2017. The Positive and Negative Influence of Search Results on People’s Decisions about the Efficacy of Medical Treatments. In Proceedings of the ACM SIGIR International Conference on Theory of Information Retrieval (Amsterdam, The Netherlands) (ICTIR ’17). Association for Computing Machinery, New York, NY, USA, 209–216.Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. Nicolas Pröllochs. 2021. Community-Based Fact-Checking on Twitter’s Birdwatch Platform. arXiv preprint arXiv:2104.07175(2021).Google ScholarGoogle Scholar
  35. Alisa Rieger, Tim Draws, Mariët Theune, and Nava Tintarev. 2021. This Item Might Reinforce Your Opinion: Obfuscation and Labeling of Search Results to Mitigate Confirmation Bias. In Proceedings of the 32nd ACM Conference on Hypertext and Social Media. 189–199.Google ScholarGoogle ScholarDigital LibraryDigital Library
  36. Kevin Roitero, Gianluca Demartini, Stefano Mizzaro, and Damiano Spina. 2018. How Many Truth Levels? Six? One Hundred? Even More? Validating Truthfulness of Statements via Crowdsourcing. In Proceedings of the CIKM 2018 Workshops co-located with 27th ACM International Conference on Information and Knowledge Management (CIKM 2018), Torino, Italy, October 22, 2018. ACM, New York, NY, USA, 6 pages.Google ScholarGoogle Scholar
  37. Kevin Roitero, Michael Soprano, Shaoyang Fan, Damiano Spina, Stefano Mizzaro, and Gianluca Demartini. 2020. Can The Crowd Identify Misinformation Objectively? The Effects of Judgment Scale and Assessor’s Background. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM, New York, NY, USA, 439–448.Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. Kevin Roitero, Michael Soprano, Beatrice Portelli, Massimiliano De Luise, Damiano Spina, Vincenzo Della Mea, Giuseppe Serra, Stefano Mizzaro, and Gianluca Demartini. 2021. Can the crowd judge truthfulness? A longitudinal study on recent misinformation about COVID-19. Personal and Ubiquitous Computing 1, 1 (2021), 1–31.Google ScholarGoogle Scholar
  39. Kevin Roitero, Michael Soprano, Beatrice Portelli, Damiano Spina, Vincenzo Della Mea, Giuseppe Serra, Stefano Mizzaro, and Gianluca Demartini. 2020. The COVID-19 infodemic: Can the crowd judge recent misinformation objectively?. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management. 1305–1314.Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Ricky J Sethi. 2017. Crowdsourcing the verification of fake news and alternative facts. In Proceedings of the 28th ACM Conference on Hypertext and Social Media. 315–316.Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. Shaban Shabani and Maria Sokhn. 2018. Hybrid machine-crowd approach for fake news detection. In 2018 IEEE 4th International Conference on Collaboration and Internet Computing (CIC). IEEE, 299–306.Google ScholarGoogle ScholarCross RefCross Ref
  42. Michael Soprano, Kevin Roitero, Francesco Bombassei De Bona, and Stefano Mizzaro. 2022. Crowd_Frame: A Simple and Complete Framework to Deploy Complex Crowdsourcing Tasks Off-the-Shelf. In Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining (Virtual Event, AZ, USA) (WSDM ’22). Association for Computing Machinery, New York, NY, USA, 1605–1608. https://doi.org/10.1145/3488560.3502182Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. Michael Soprano, Kevin Roitero, David La Barbera, Davide Ceolin, Damiano Spina, Stefano Mizzaro, and Gianluca Demartini. 2021. The many dimensions of truthfulness: Crowdsourcing misinformation assessments on a multidimensional scale. Information Processing & Management 58, 6 (2021), 102710.Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. Sebastian Tschiatschek, Adish Singla, Manuel Gomez Rodriguez, Arpit Merchant, and Andreas Krause. 2017. Detecting fake news in social networks via crowdsourcing. arXiv preprint arXiv:1711.09025(2017).Google ScholarGoogle Scholar
  45. Jacky Visser, John Lawrence, and Chris Reed. 2020. Reason-Checking Fake News. Commun. ACM 63, 11 (oct 2020), 38–40.Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. Silvio Waisbord. 2018. Truth is what happens to news: On journalism, fake news, and post-truth. Journalism studies 19, 13 (2018), 1866–1878.Google ScholarGoogle ScholarCross RefCross Ref
  47. William Yang Wang. 2017. ” Liar, Liar Pants on Fire”: A new benchmark dataset for fake news detection. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). ACL, Vancouver, Canada, 422–426.Google ScholarGoogle ScholarCross RefCross Ref
  48. William Yang Wang. 2017. ”Liar, Liar Pants on Fire”: A New Benchmark Dataset for Fake News Detection. In Proceedings of ACL. 422–426.Google ScholarGoogle Scholar
  49. Ryen W White. 2014. Belief dynamics in Web search. Journal of the Association for Information Science and Technology 65, 11(2014), 2165–2178.Google ScholarGoogle ScholarDigital LibraryDigital Library
  50. Pinar Yildirim, Esther Gal-Or, and Tansev Geylani. 2013. User-Generated Content and Bias in News Media. Management Science 59, 12 (2013), 2655–2666.Google ScholarGoogle ScholarDigital LibraryDigital Library
  51. Yisong Yue, Rajan Patel, and Hein Roehrig. 2010. Beyond Position Bias: Examining Result Attractiveness as a Source of Presentation Bias in Clickthrough Data. In Proceedings of the 19th International Conference on World Wide Web (Raleigh, North Carolina, USA) (WWW ’10). Association for Computing Machinery, New York, NY, USA, 1011–1018.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. The Effects of Crowd Worker Biases in Fact-Checking Tasks
        Index terms have been assigned to the content through auto-classification.

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Other conferences
          FAccT '22: Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency
          June 2022
          2351 pages
          ISBN:9781450393522
          DOI:10.1145/3531146

          Copyright © 2022 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 20 June 2022

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article
          • Research
          • Refereed limited

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        HTML Format

        View this article in HTML Format .

        View HTML Format