ABSTRACT
The recent months have seen an explosion of interest, hype, and concern about generative AI, driven by the release of ChatGPT. In this article I seek to explicate some potential and actual harms of the engineering and use of generative AI such as ChatGPT. With this I also suggest a reframing for researchers with an interest in interaction. With this reframing I seek to provoke researchers to consider studying the settings of ChatGPT development and use as active sites of production. Research should focus on the organisational, technological and interactional practices and contexts in and through which generative AI and its outputs—harmful and otherwise—are produced, by whom, to what end, and with what consequences on societies.
- Andrew Arsht and Daniel Etcovitch. 2018. The human cost of online content moderation. Harvard Journal of Law and Technology (2018).Google Scholar
- Gagan Bansal, Alison Marie Smith-Renner, Zana Buçinca, Tongshuang Wu, Kenneth Holstein, Jessica Hullman, and Simone Stumpf. 2022. Workshop on Trust and Reliance in AI-Human Teams (TRAIT). In Extended Abstracts of the 2022 CHI Conference on Human Factors in Computing Systems (New Orleans, LA, USA) (CHI EA ’22). Association for Computing Machinery, New York, NY, USA, Article 116, 6 pages. https://doi.org/10.1145/3491101.3503704Google ScholarDigital Library
- Emily M. Bender, Timnit Gebru, Angelina McMillan-Major, and Shmargaret Shmitchell. 2021. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (Virtual Event, Canada) (FAccT ’21). Association for Computing Machinery, New York, NY, USA, 610–623. https://doi.org/10.1145/3442188.3445922Google ScholarDigital Library
- Tolga Bolukbasi, Kai-Wei Chang, James Y Zou, Venkatesh Saligrama, and Adam T Kalai. 2016. Man is to computer programmer as woman is to homemaker? debiasing word embeddings. Advances in neural information processing systems 29 (2016).Google Scholar
- Barry Brown and Eric Laurier. 2017. The trouble with autopilots: Assisted and autonomous driving on the social road. In Proceedings of the 2017 CHI conference on human factors in computing systems. 416–429.Google ScholarDigital Library
- Barry Brown, Moira McGregor, and Eric Laurier. 2013. iPhone in vivo: video analysis of mobile device use. In Proceedings of the SIGCHI conference on Human Factors in computing systems. 1031–1040.Google ScholarDigital Library
- Debby R. E. Cotton, Peter A. Cotton, and J. Reuben Shipway. 2023. Chatting and cheating: Ensuring academic integrity in the era of ChatGPT. Innovations in Education and Teaching International 0, 0 (2023), 1–12. https://doi.org/10.1080/14703297.2023.2190148 arXiv:https://doi.org/10.1080/14703297.2023.2190148Google ScholarCross Ref
- Andrew Crabtree, Tom Rodden, Peter Tolmie, and Graham Button. 2009. Ethnography Considered Harmful. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Boston, MA, USA) (CHI ’09). Association for Computing Machinery, New York, NY, USA, 879–888. https://doi.org/10.1145/1518701.1518835Google ScholarDigital Library
- Kate Crawford and Vladan Joler. 2018. Anatomy of an AI system: the Amazon Echo as an anatomical map of human labor, data and planetary resources. AI now institute and share lab. https://anatomyof.ai/Google Scholar
- Mateusz Dubiel, Sylvain Daronnat, and Luis A. Leiva. 2022. Conversational Agents Trust Calibration: A User-Centred Perspective to Design. In Proceedings of the 4th Conference on Conversational User Interfaces (Glasgow, United Kingdom) (CUI ’22). Association for Computing Machinery, New York, NY, USA, Article 30, 6 pages. https://doi.org/10.1145/3543829.3544518Google ScholarDigital Library
- Justin Edwards and Elaheh Sanoubari. 2019. A Need for Trust in Conversational Interface Research. In Proceedings of the 1st International Conference on Conversational User Interfaces (Dublin, Ireland) (CUI ’19). Association for Computing Machinery, New York, NY, USA, Article 21, 3 pages. https://doi.org/10.1145/3342775.3342809Google ScholarDigital Library
- Joel E. Fischer, Stuart Reeves, Martin Porcheron, and Rein Ove Sikveland. 2019. Progressivity for Voice Interface Design. In Proceedings of the 1st International Conference on Conversational User Interfaces (Dublin, Ireland) (CUI ’19). Association for Computing Machinery, New York, NY, USA, Article 26, 8 pages. https://doi.org/10.1145/3342775.3342788Google ScholarDigital Library
- Lilly Irani. 2016. The hidden faces of automation. XRDS: Crossroads, The ACM Magazine for Students 23, 2 (2016), 34–37.Google ScholarDigital Library
- Lilly C. Irani and M. Six Silberman. 2013. Turkopticon: Interrupting Worker Invisibility in Amazon Mechanical Turk. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Paris, France) (CHI ’13). Association for Computing Machinery, New York, NY, USA, 611–620. https://doi.org/10.1145/2470654.2470742Google ScholarDigital Library
- Razan Jaber and Donald McMillan. 2022. Cross-Modal Repair: Gaze and Speech Interaction for List Advancement. In Proceedings of the 4th Conference on Conversational User Interfaces (Glasgow, United Kingdom) (CUI ’22). Association for Computing Machinery, New York, NY, USA, Article 25, 11 pages. https://doi.org/10.1145/3543829.3543833Google ScholarDigital Library
- Minha Lee. 2020. Speech Acts Redux: Beyond Request-Response Interactions. In Proceedings of the 2nd Conference on Conversational User Interfaces (Bilbao, Spain) (CUI ’20). Association for Computing Machinery, New York, NY, USA, Article 13, 10 pages. https://doi.org/10.1145/3405755.3406124Google ScholarDigital Library
- Donald A. Norman. 2005. Human-Centered Design Considered Harmful. Interactions 12, 4 (jul 2005), 14–19. https://doi.org/10.1145/1070960.1070976Google ScholarDigital Library
- Martin Porcheron, Joel E. Fischer, and Stuart Reeves. 2021. Pulling Back the Curtain on the Wizards of Oz. Proc. ACM Hum.-Comput. Interact. 4, CSCW3, Article 243 (jan 2021), 22 pages. https://doi.org/10.1145/3432942Google ScholarDigital Library
- Martin Porcheron, Joel E Fischer, Stuart Reeves, and Sarah Sharples. 2018. Voice Interfaces in Everyday Life. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (Montreal QC, Canada) (CHI ’18). ACM, New York, NY, USA, Article 640, 12 pages. https://doi.org/10.1145/3173574.3174214Google ScholarDigital Library
- Stuart Reeves. 2019. Conversation Considered Harmful?. In Proceedings of the 1st International Conference on Conversational User Interfaces (Dublin, Ireland) (CUI ’19). Association for Computing Machinery, New York, NY, USA, Article 10, 3 pages. https://doi.org/10.1145/3342775.3342796Google ScholarDigital Library
- Yunhan Wu, Martin Porcheron, Philip Doyle, Justin Edwards, Daniel Rough, Orla Cooney, Anna Bleakley, Leigh Clark, and Benjamin Cowan. 2022. Comparing Command Construction in Native and Non-Native Speaker IPA Interaction through Conversation Analysis. In Proceedings of the 4th Conference on Conversational User Interfaces (Glasgow, United Kingdom) (CUI ’22). Association for Computing Machinery, New York, NY, USA, Article 10, 12 pages. https://doi.org/10.1145/3543829.3543839Google ScholarDigital Library
Index Terms
- Generative AI Considered Harmful
Recommendations
Exploring the Transformative Impact of Generative AI on Higher Education
New Sustainable Horizons in Artificial Intelligence and Digital SolutionsAbstractThe launch and subsequent rapid adoption of ChatGPT has initiated significant debate within the academic and practice-based community generating both fear and anxiety that is also contrasted with a sense of opportunity and excitement within the ...
Opinion Paper: “So what if ChatGPT wrote it?” Multidisciplinary perspectives on opportunities, challenges and implications of generative conversational AI for research, practice and policy
AbstractTransformative artificially intelligent tools, such as ChatGPT, designed to generate sophisticated text indistinguishable from that produced by a human, are applicable across a wide range of contexts. The technology presents opportunities as well ...
Highlights- Generative AI can enhance productivity but may also lead to replacement of human employees.
- Teaching, learning, and academic research will experience some of the most transformative impacts.
- Biases, out of date training data, and ...
Ethnography considered harmful
CHI '09: Proceedings of the SIGCHI Conference on Human Factors in Computing SystemsWe review the current status of ethnography in systems design. We focus particularly on new approaches to and understandings of ethnography that have emerged as the computer has moved out of the workplace. These seek to implement a different order of ...
Comments