ABSTRACT
The increasing adoption of smart and interactive technologies, based on Artificial Intelligence (AI), in business and society transforms humans’ perception of these technologies towards social actors. However, since there are also reservations of humans to interact with artificial agents, it is fundamental to explore the social perception of this new technology and the implications thereof. Social perception governs social behavior and subsequently affects interactions with AI-agents. Moral responsibility plays a pivotal role in functioning societies and should thus be taken into account when implementing AI. This qualitative study examines workers in an industrial setting, who perform a decision-making task together with an AI-agent possessing no visual human-like characteristics. Based on the model of Responsibility in Hybrid Societies we show that workers assign capacities of social perception and moral responsibility to their AI-based counterpart. We present first evidence that theoretically derived models of social perception and moral responsibility are applicable in industrial settings.
Supplemental Material
- S. Berretta, A. Tausch, and C. Peifer. 2022. Humanzentrierte Arbeitsgestaltung im Zeitalter von KI [human-centered work design in the age of AI]. Arbeitsmedizin Sozialmedizin Umweltmedizin 10–2022, (2022), 614–619. DOI:https://doi.org/10.17147/asu-1-225905Google ScholarCross Ref
- M. Brandt. 2020. Robotisierung: So viele Roboter kommen auf 10.000 Beschäftigte. Statista. Retrieved April 7, 2021 from https://de.statista.com/infografik/13676/roboterdichte-in-der-fertigungsindustrie/Google Scholar
- Maximilian Bretschneider, Sarah Mandl, Anja Strobel, Frank Asbrock, and Bertolt Meyer. 2022. Social perception of embodied digital technologies—a closer look at bionics and social robotics. Gr Interakt Org (September 2022). DOI:https://doi.org/10.1007/s11612-022-00644-7Google ScholarCross Ref
- Amy J. C. Cuddy, Susan T. Fiske, and Peter Glick. 2007. The BIAS map: Behaviors from intergroup affect and stereotypes. Journal of Personality and Social Psychology 92, 4 (2007), 631–648. DOI:https://doi.org/10.1037/0022-3514.92.4.631Google ScholarCross Ref
- Amy J.C. Cuddy, Susan T. Fiske, and Peter Glick. 2008. Warmth and Competence as Universal Dimensions of Social Perception: The Stereotype Content Model and the BIAS Map. Advances in Experimental Social Psychology 40, (2008), 61–149. DOI:https://doi.org/10.1016/S0065-2601(07)00002-0Google ScholarCross Ref
- Dominik Dellermann, Adrian Calma, Nikolaus Lipusch, Thorsten Weber, Sascha Weigel, and Philipp Ebel. 2019. The Future of Human-AI Collaboration: A Taxonomy of Design Knowledge for Hybrid Intelligence Systems. In Proceedings of the 52nd Hawaii International Conference on System Sciences, 274–283. DOI:https://doi.org/10.24251/HICSS.2019.034Google ScholarCross Ref
- Kadir Alpaslan Demir, Gözde Döven, and Bülent Sezen. 2019. Industry 5.0 and Human-Robot Co-working. Procedia Computer Science 158, (2019), 688–695. DOI:https://doi.org/10.1016/j.procs.2019.09.104Google ScholarDigital Library
- Susan T. Fiske, Amy J. C. Cuddy, Peter Glick, and Jun Xu. 2002. A model of (often mixed) stereotype content: Competence and warmth respectively follow from perceived status and competition. Journal of Personality and Social Psychology 82, 6 (June 2002), 878–902. DOI:https://doi.org/10.1037/0022-3514.82.6.878Google ScholarCross Ref
- Carl Benedikt Frey and Michael A. Osborne. 2017. The future of employment: How susceptible are jobs to computerisation? Technological Forecasting and Social Change 114, (January 2017), 254–280. DOI:https://doi.org/10.1016/j.techfore.2016.08.019Google ScholarCross Ref
- Raul Hakli and Pekka Mäkelä. 2016. Robots, Autonomy, and Responsibility. Frontiers in Artificial Intelligence and Applications 290, (2016), 145–154. DOI:https://doi.org/10.3233/978-1-61499-708-5-145Google ScholarCross Ref
- Peter A. Hancock, Deborah R. Billings, Kristin E. Schaefer, Jessie Y. C. Chen, Ewart J. de Visser, and Raja Parasuraman. 2011. A Meta-Analysis of Factors Affecting Trust in Human-Robot Interaction. Hum Factors 53, 5 (October 2011), 517–527. DOI:https://doi.org/10.1177/0018720811417254Google ScholarCross Ref
- Hsiu-Fang Hsieh and Sarah E. Shannon. 2005. Three Approaches to Qualitative Content Analysis. Qualitative Health Research 15, 9 (2005), 1277–1288. DOI:https://doi.org/10.1177/1049732305276687Google ScholarCross Ref
- Nicolas Kervyn, Susan Fiske, and Vincent Yzerbyt. 2015. Forecasting the Primary Dimension of Social Perception: Symbolic and Realistic Threats Together Predict Warmth in the Stereotype Content Model. Social Psychology 46, 1 (January 2015), 36–45. DOI:https://doi.org/10.1027/1864-9335/a000219Google ScholarCross Ref
- Colin Wayne Leach, Naomi Ellemers, and Manuela Barreto. 2007. Group virtue: The importance of morality (vs. competence and sociability) in the positive evaluation of in-groups. Journal of Personality and Social Psychology 93, 2 (2007), 234–249. DOI:https://doi.org/10.1037/0022-3514.93.2.234Google ScholarCross Ref
- Sarah Mandl, Jennifer Brade, Maximilian Bretschneider, Anabel Skripcak, Frank Asbrock, Bertolt Meyer, Georg Jahn, Philipp Klimant, and Anja Strobel. 2023. Social perception of Embodied Digital Technologies interacting with humans. AHFE International, USA. DOI:https://doi.org/10.54941/ahfe1002836Google ScholarCross Ref
- Sarah Mandl, Maximilian Bretschneider, Frank Asbrock, Bertolt Meyer, and Anja Strobel. 2022. The Social Perception of Robots Scale (SPRS): Developing and Testing a Scale for Successful Interaction Between Humans and Robots. In Collaborative Networks in Digitalization and Society 5.0, Luis M. Camarinha-Matos, Angel Ortiz, Xavier Boucher and A. Luís Osório (eds.). Springer International Publishing, Cham, 321–334. DOI:https://doi.org/10.1007/978-3-031-14844-6_26Google ScholarCross Ref
- Sarah Mandl, Maximilian Bretschneider, Stefanie Meyer, Dagmar Gesmann-Nuissl, Frank Asbrock, Bertolt Meyer, and Anja Strobel. 2022. Embodied Digital Technologies: First insights in the Social and Legal Perception of Robots and Users of Prostheses. Front. Robot. AI 9:787970, (2022). DOI:https://doi.org/10.3389/frobt.2022.787970Google ScholarCross Ref
- Kevin R. McKee, Xuechunzi Bai, and Susan T. Fiske. 2022. Warmth and competence in human-agent cooperation. Retrieved November 23, 2022 from http://arxiv.org/abs/2201.13448Google Scholar
- Stefanie Meyer, Sarah Mandl, Dagmar Gesmann-Nuissl, and Anja Strobel. 2022. Responsibility in Hybrid Societies: concepts and terms. AI Ethics (June 2022). DOI:https://doi.org/10.1007/s43681-022-00184-2Google ScholarCross Ref
- Stefanie Meyer, Michael R. Müller, Anne Sonnenmoser, Sarah Mandl, Anja Strobel, and Dagmar Gesmann-Nuissl. accepted. Towards Hybrid Personae? In Proceedings of the 1st International Conference on Hybrid Societies, March 15-17, 2023, Chemnitz, Germany.Google Scholar
- Matthew B. Miles, A. Michael Huberman, and Johnny Saldana. 2020. Qualitative Data Analysis: A Methods Sourcebook (4th ed. ed.). SAGE Publications, Thousand Oaks.Google Scholar
- A.B. Moniz, M. Candeias, and N. Boavida. 2022. Changes in productivity and labour relations: Artificial intelligence in the automotive sector in Portugal. International Journal of Automotive Technology and Mangement 22, 2 (2022), 222–244. DOI:https://doi.org/10.1504/IJATM.2022.124366Google ScholarCross Ref
- Clifford Nass, Jonathan Steuer, and Ellen R. Tauber. 1994. Computers are social actors. In Conference companion on Human factors in computing systems - CHI ’94, ACM Press, Boston, Massachusetts, United States, 204. DOI:https://doi.org/10.1145/259963.260288Google ScholarDigital Library
- Allison Sauppé and Bilge Mutlu. 2015. The Social Impact of a Robot Co-Worker in Industrial Settings. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems - CHI ’15, ACM Press, Seoul, Republic of Korea, 3613–3622. DOI:https://doi.org/10.1145/2702123.2702181Google ScholarDigital Library
- Isabella Seeber, Eva Bittner, Robert O Briggs, Gert-Jan de Vreede, Triparna de Vreede, Douglas Druckenmiller, Alexander B Merz, Sarah Oeste-Reiß, Nils Randrup, Gerhard Schwabe, and Matthias Söllner. 2018. Machines as Teammates: A Collaboration Research Agenda. In Proceedings of the 51st Hawaii International Conference on System Sciences, 420–429. DOI:https://doi.org/10.24251/HICSS.2018.055Google ScholarCross Ref
- VERBI Software. 2021. MAXQDA 2022 [computer software].Google Scholar
- 2021. Statistisches Bundesamt. Statista. Retrieved July 22, 2021 from https://de.statista.com/statistik/daten/studie/445223/umfrage/produtkionsmenge-von-mehrzweck-industrierobotern-in-deutschland/Google Scholar
Index Terms
- Exploring key categories of social perception and moral responsibility of AI-based agents at work: Findings from a case study in an industrial setting
Recommendations
Human Perceptions on Moral Responsibility of AI: A Case Study in AI-Assisted Bail Decision-Making
CHI '21: Proceedings of the 2021 CHI Conference on Human Factors in Computing SystemsHow to attribute responsibility for autonomous artificial intelligence (AI) systems’ actions has been widely debated across the humanities and social science disciplines. This work presents two experiments (N=200 each) that measure people’s perceptions ...
Moral luck and computer ethics: Gauguin in cyberspace
I argue that the problem of `moral luck' is an unjustly neglected topic within Computer Ethics. This is unfortunate given that the very nature of computer technology, its `logical malleability', leads to ever greater levels of complexity, unreliability ...
Narrative responsibility and artificial intelligence: How AI challenges human responsibility and sense-making
AbstractMost accounts of responsibility focus on one type of responsibility, moral responsibility, or address one particular aspect of moral responsibility such as agency. This article outlines a broader framework to think about responsibility that ...
Comments