1 Introduction

Memorialization of mass atrocities, such as war crimes and genocides, has a special meaning to humankind. Memorialization involves the practice of preserving the memory of those who suffered, the experiences of what they endured, and the mass atrocity event itself. As such, the memorialization of collective suffering addresses many ethical and moral obligations that may be specific to a particular culture and society and reflect ideas drawn from transnational human rights and transitional justice. These obligations range from respecting and tending to the dead [66, 68] and honoring those who fought against or resisted the perpetrators [31, 42]Footnote 1 to countering the distortion of historical facts and preventing the repetition of atrocities [17]. The construction of social memory of atrocities via memorialization is thus integral to the moral reconstitution and social repair of societies in the aftermath of atrocities [18, 99]. However, these obligations are not easy to fulfill. Challenges include contestation over individual and collective memories with the imperative of accounting for various perspectives of the past, but without legitimizing patently false ones [68]; balancing individual narratives versus collective narratives and local versus national and international ones [44]; temporal, political, and other situational factors that may influence different narrative constructions among individuals or groups [80, 100]; limited access to human and financial resources required for commemoration [68]; and the need to resist efforts to instrumentalize memorialization efforts, which subordinate social needs to short-term political gains [47].

As with other societal practices, digital technologies have transformed the memorialization of mass atrocities [4, 20, 62, 63, 89, 91, 101]. While mass media had a major influence on memorialization already in the analog era [90], the rise of digital platforms caused a profound disruption in memory practices. The unprecedented amount of digitized and digital-born content, the increased connectivity between consumers and producers of memories [35], and the fluidity between various digital formats, including images, text, and video [39] have created space for less top-down memory practices and greater autonomy and creativity in memorial works. These practices offer the prospect of giving voice to under-represented communities and enable possibilities for engaging with the difficult pasts that might not otherwise have been accessible via heritage institutions and archives [77]. At the same time, digital technologies create challenges for memorialization, potentially facilitating denialism, and distortion [32, 55], while contributing to the trivialization of past suffering through the amplification of cynical attitudes toward it [36]. This latter phenomenon emerges from the potent combination of the unprecedented availability of atrocities-related content [10] and the ease with which such evidence of violence can be fabricated. Furthermore, the plethora of online memorial activities may obscure some memorial efforts in the noise of others or spark the banality that accompanies surface-level repetition in a manner reminiscent of Walter Benjamin’s thoughts on the risks of mechanical reproduction [3].

The extensive volume of atrocities-related digital content prompts the need to help individuals navigate it. Since the 2000s, this task has been increasingly delegated to artificial intelligence (AI)-driven systems, such as search engines and recommendation systems, which filter content items in response to user input (e.g., search queries) and retrieve items that are viewed by the system as the most relevant for the users [48]. Despite the increasing use of such systems by the heritage institutions (e.g., USHMM digital collections’ search; https://collections.ushmm.org/search/) and commercial platforms (e.g., Google search), the discussion of the opportunities and risks that AI systems create for the memorialization of mass atrocities remains limited (for some exceptions, see [6, 43, 56, 88]). For example, recent research has exposed the role of different search engines in memorialization by analyzing the search results in relation to the Holocaust [59] and the Ukrainian HolodomorFootnote 2 [60] and potential distortions in the representation of mass atrocities which these systems may cause.

The situation becomes more complex with the emergence of generative AIs, which represent a profound transformation in the context of the memorialization of mass atrocities. Instead of retrieving information about atrocities and prioritizing certain content sources, generative AIs produce the textual/visual content themselves [12], as illustrated in Fig. 1. The implications of this change for the memorialization of mass atrocities are substantial: we can expect not only a rapid increase of content engaging with memory about past suffering but also significant challenges in distinguishing between human- and AI-made content. The shift to generative AI also implies major changes in how information about mass atrocities is retrieved and how humans interact with such content in various contexts (e.g., education, community NGOs, international development and humanitarian entities, and government).

Fig. 1
figure 1

a ChatGPT (https://chat.openai.com/) produced in a few seconds the piece of text responding to the prompt inquiring to explain in 500 characters “what happened at Sobibor concentration camp”. b The AI platform Rytr (https://rytr.me/), also in seconds, produced a digital rendition of an oil painting using the prompt “digital memorial of Sobibor concentration camp with representation of victims in oil paint”

The efficiency of generative AIs in producing text and illustrations can facilitate the processes of compiling and analyzing memorialization data and make efforts to memorialize the past more accessible for individuals worldwide. Yet the long-term consequences of adopting generative AIs in the context of memorializing both historical (e.g., the Holocaust) and recent atrocities (e.g., Russian war crimes in Ukraine) are uncertain. Some of the related concerns include the non-transparency of sources used to automatically generate narratives about the atrocities, the possibilities for misrepresentations of historical facts in rendered content, and the ethical problems arising from the generation of fake evidence or the manipulation of real one in a manner that changes the meaning of personal testimonies or artifacts. Additionally, it is unclear how the generated content might influence beliefs and attitudes toward a given atrocity, or how a growing distrust of some forms of digital mediaFootnote 3 might interact with an increasing awareness of AI capabilities and the pervasiveness of AI-rendered content in highly sensitive contexts such as the ones dealing with the memorialization of suffering.

Through our interdisciplinary perspective and experience of working in the field of memorialization, we aim to address some of these concerns and questions by reviewing the possibilities and challenges generative AI brings to the online memorialization of mass atrocities. As AI-related technologies continue to approach and exceed human capabilities in pattern recognition and processing large volumes of data in an increasing number of areasFootnote 4, it is crucial to evaluate their potential impact. Such evaluations are integral for deciding whether (and how) to incorporate generative AI in an area where technologies have to be adopted particularly carefully due to the ethical implications or whether recommending against AI use in this field is more beneficial due to its inadequacy to account for multiple ethical considerations guiding the process of atrocities’ memorialization.

2 Multiple facets of AI: from information retrieval to information generation

The use of AI in the context of mass atrocity memorialization started long before the rise of generative AI. Defined as the ability of human-made artifacts to engage in intellectual behavior [102], AI has been applied to a broad range of memorialization-related tasks both by heritage institutions and commercial platforms. Most of the tasksFootnote 5 involving these non-generative forms of AI concerned information retrieval, a prominent field of computer science focused on information identification and retrieval for user information needs in response to user input [105]. Initially inheriting many features of the information organization systems used by human practitioners to curate collections in the analogue archives and museums, non-generative AI systems, such as recommender systems and search engines, brought the ability to scale the retrieval tasks and automatically update databases with information about new sources and items (e.g., as in the case of web search engines constantly re-indexing the Internet with the help of crawlers [106]). Over time, non-generative AI systems started including additional mechanisms aiming to customize information delivery for individual users via personalization mechanisms (e.g., to select more relevant sources depending on the language of the query [107]) or introducing new formats for information retrieval (e.g., conversational exchanges as in the case of the Holocaust holograms [85]).

As a portal to information relevant for memorialization for the general public, non-generative AI systems have become the gatekeepers of information [108]. However, the expanding role of these systems in information retrieval and knowledge production highlights a number of shortcomings associated with their performance, in particular in relation to ontologically contested subjects, that is, subjects for which there are multiple viewpoints and a lack of general consensus. Existing research shows that the exact ways in which non-generative AI can malperform vary from non-systematic errors (e.g., search engines retrieving irrelevant or factually incorrect information [59]) to systematic bias resulting in a skewed representation of social reality [109, 110]. Researchers have also demonstrated the potential of non-generative AI systems to reiterate and amplify stereotypes, in particular in relation to discriminated gender and racial groups [110,111,112,113,114].

The evidence of limitations of non-generative AI systems together with their fragmented regulation [115, 116] has triggered a response from the scholarly community and also steps from the AI companies themselves (e.g., staged rollouts of products and calls for state regulation). The growing interest in non-generative AI (mal)performance resulted in the establishment of a new set of research techniques used to systematically audit the performance of non-generative AI systems [117, 118]. The exact implementations of these auditing approaches (for a comprehensive review, see [119] vary from sock puppet approaches using software simulating human behavior to generate inputs for the AI system [120], to techniques collecting system outputs from crowdworkers [121] to system code audits [122]. While the research on AI bias in the context of memorialization-related issues so far remains limited, a few studies [56, 60, 123] highlighted the tendency of non-generative AIs for focusing on a limited set of memorialization-related sources (often, the ones reiterating predominantly state narratives [123]), or prioritizing only a few selected aspects of the atrocity (e.g., prioritizing information drawn from a single atrocity site and downplaying the significance of other sites [59]).

Despite the lack of transparency of non-generative AIs, these systems would provide sources and excerpts that, although potentially subject to decontextualization, could still be traced back to their origins. This has changed with the use of generative AI which produce new content items, instead of finding and retrieving already existing onesFootnote 6. For example, to generate text, language models (LMs), which power generative AI-systems (e.g., chatgGPT), are trained to predict the likelihood of the next token (e.g., a word) given information about the preceding or surrounding text [124]. In principle, the mathematical procedures allowing generative AI to make content do not allow for the emergence of understanding of the meaning behind such content [2]. Instead, generative AI can achieve only a shallow definition of meaning, known as distributional semantics—i.e., the description of the contexts in which a specific token appears [125, 126]. However, some evidence suggests that AI is capable of generating content of substantial internal complexity that goes beyond simple statistical modeling, although it may be just the result of parroting the gargantuan amount of data used for the training [127]. In either case, the ability of generative AI to imitate human capacities for generating content and its implications for different sectors has sparked heated academic and societal debates (e.g., [128]).

3 Possibilities of generative AI for digital memorialization of mass atrocities

Generative AI enables new possibilities for creating content that can be used for memorialization purposes. Such content can be either visual (motion or still) or textual and can reproduce multiple formats which are currently used in the context of digital grassroots memorialization of mass atrocities; examples of these formats range from video tributes commemorating specific instances of mass violence [55], to the Internet memes reinforcing [28] or challenging established genocide narratives [7], to the images imitating photos made at the sites of violence [52], to the drawings aiming to attract the attention of the general public to the past and present suffering [81], to the online encyclopedia entries aiming to find consensus between the different viewpoints on the past [54]. As a result, AI might induce a variety of positive developments, including capacities for creation and participation in the context of memorialization, expanding public access to knowledge about the past, helping researchers learn more about mass atrocities and the memorialization thereof, and serving as a tool to help detect and counter false narratives.

  • AI can expand the capacity to memorialize

    Generative AIs can easily produce representations of mass atrocities. The ease of access and use of these platforms enables new possibilities for expressing mourning, producing and analyzing testimony, and communicating loss—and for a fraction of the cost it would take to hire a human to create the textual and image content to fulfill these purposes. The outputs of AI models can then be employed on their own or combined with other digital- or analog-born materials to memorialize past and present atrocities (e.g., by populating digital memorials with AI-generated content where such content is lacking, or in cases where an AI-derived image or voice might serve to protect the privacy or safety of the victims or their relatives). This is particularly true in relation to survivors engaging with an AI platform to tell their stories or in cases where AI is employed as a therapist assistant [21].

    Similarly, the textual descriptions of atrocities or commemorative statements generated by text-based AIs may augment international participation and engagement by enabling individuals to participate in online memorial practices in different languages without the need for translators, copyeditors, or other production assistants. Under these circumstances, image- and text-focused AIs can empower individuals who want to create, contribute to, or expand existing grassroots memorialization practices with almost limitless creativity providing their projects remain within the accepted use requirements of the platforms (e.g., Midjourney’s use guidelines aiming to prevent the creation of graphic or potentially harmful images through an explicit list of banned prompts [30, 65]).

  • AI can enable new ways for the general public to learn about mass atrocities

    Engaging with generative AIs to learn about mass atrocities can be a form of memorialization practice itself. Earlier, non-generative versions of AI-driven systems focused on information retrieval tasks [1]. For example, the Let Them SpeakFootnote 7 platform [133] incorporated databases of Holocaust textual testimonies that could be queried for specific words and used to create lists of testimonial fragments [69]. The Shoah Foundation-produced holograms of Holocaust survivors [85, 88] is a technologically similar form of AI that uses a finite database of questions with pre-recorded answers from Holocaust survivors and is employed to enable an interactive experience for individuals and groups engaging with the holograms.

    A key distinction of generative AI is the ability to produce new narratives, store these narratives, and use them as additional data for iterative training. It allows generative AIs not only to retrieve content in response to an (often limited) set of user inputs, but to generate content and sustain a broader range of interactions with individuals interested in exploring the past. Thus a user may have a vested interest in continuing the dialogue with the generative AI chatbot on a given topic rather than having to start at the beginning each time as in the case of more traditional conversational agents. These interactions, which occur on one-on-one user and machine bases over private accounts, can be understood as a form of memorialization in itself: for instance, when individuals engage with AIs to generate a text- or image-based narrative about a mass atrocity.

  • AI can be a means for researchers to collect/analyze data

    The advancements in the field of AI also transform the processes for collecting and analyzing data concerning both historical and present mass atrocities. Instead of simply retrieving and ranking information sources relevant to the user queries (as non-generative AIs used to do), generative AI platforms can directly respond to user requests while also providing suggestions of information sources for further exploration (e.g., as in the case of generative AI interface for Bing search). Such functionality can potentially accelerate the process of analyzing data on mass atrocities and lead to profound transformations in how institutions provide access to this data. For example, instead of using a conventional archive search, a generative AI can be integrated into the interface to provide recommendations to the researcher in a conversational format; similarly, generative AI can be used to generate visualizations based on text descriptions to offer additional perspectives for the researcher. In addition to retrieving information in a more digestible format, generative AIs can also be used for automated content labeling that can help identify content related to mass atrocities, given AIs’ capacity for content identification can be comparable or superior to that of humans [26, 86].

    There are also other capabilities of generative AI which can facilitate the analysis of data related to mass atrocities. Similar to earlier AI-driven information retrieval systems, generative AI can personalize outputs (e.g., textual content generation prompted by individual user needs and interests). For example, AI can summarize data and present several summaries emphasizing various elements to let the researcher think through the data from various perspectives. Furthermore, both text- and image-focused generative AIs can be used for creating synthetic data (see, for instance [75]), which can then be employed for training computational approaches for automated content analysis. For example, the Iraq war crimes investigation initiative UNITAD and the Germany-based non-governmental organization Mnemonic have reportedly started using AI to scan hundreds of thousands of hours of video to identify evidence of war crimes in the context of recent armed conflicts in Iraq and Syria [84].

  • AI can be a tool for detecting distortion and denialism

    In addition to the above-mentioned possibilities, the ability of generative AIs to interprete and engage with user input makes them an effective means of content classification. Content classification involves the assignment of specific labels to content items to detect specific attributes of such content (e.g., whether it is related to a specific subject, such as politics, or expresses a specific sentiment, for instance, positive or negative one). In the case of content related to mass atrocities, such classification can take multiple forms, including whether the statement that AI is asked to evaluate comes from a certain source or whether it might contain denialist claims. While the degree to which generative AI can perform these functions, specifically in the case of content dealing with mass atrocities, is yet to be studied, generative AI platforms, such as chatGPT, have demonstrated the ability to identify whether claims entered by the users are false in relation to diverse subjects [34] and health-related matters in particular [41], and to evaluate the degree of news outlets’ credibility [94]. Under these circumstances, generative AI can potentially be used to filter out content promoting distortion or denialism of mass atrocities, with the possibility that individuals will be more open to engaging with views they find contrasting to their own if these views are expressed via a human-like conversational interface (for examples related to conventional chatbots, see [95]).

4 Threats of generative AI for digital memorialization of mass atrocities

Despite the many advantages generative AI presents for the memorialization of mass atrocities, there are also risks. An amoral machine entity, AI does not assign specific meaning to the data it processes. Rather, AI is more akin to Bender and Kohler’s “octopus test” [2], where a clever octopus fools people into believing it is a sentient human by communicating in the English language through statistical patterns until it fails to comprehend the implications of an unfamiliar context, or Searle’s “Chinese Room” proposition [76] in which he posits he could create coherent text in Chinese by learning to manipulate symbols without actually understanding the language itself. Both exercises demonstrate how the capacity to impart information should not be conflated with the ability to understand the meaning—nor should it, beyond that point, be confused with the capability to make context-driven moral judgments. Yet memorialization is a realm of human thought and activity for which the abilities to make fine distinctions in meaning and to exercise moral judgment are essential.

  • AI can serve as a means of enforcing hegemonic narratives and practices

    Generative AIs are trained on specific sets of data. Unless their training process is diversified, AIs can take prevalent patterns in data for granted and then reiterate them in content which AIs generate. In the case of atrocities-related content, it may result in the enforcement of hegemonic narratives and representation practices, for instance, by prioritizing Western-centric views on how mass atrocities shall be remembered or interpreted and uncritically translating these views to other contexts. Such reinforcement of memory hegemonies might result in the silencing or erasure of experiences of minority groups (including the ones which were disproportionately affected by the mass atrocities) and suppression of alternative practices of memorialization [19]. It is additionally concerning in the case of authoritarian states, where hegemonic memory practices often serve as an integral means of state ideology and propaganda and where generative AIs can become an effective tool for consolidating the national memory regimes, such as in the case of recent efforts by China [16].

    The capabilities of generative AIs to enforce hegemonic narratives are further amplified by the risks of keeping users in information bubbles, which was a common concern about the non-generative AI-driven systems dealing with information retrieval tasks (e.g., [9, 71]). While the existence of AI-amplified information bubbles has so far found little empirical support [9] and, in some cases, such bubbles can be beneficial for society by nurturing independent thinking (e.g., [58]), the possibility of generative AIs leading their users into the “rabbit holes"Footnote 8 [8] of memory hegemonies can not be currently excluded. Further support to such concerns is provided by the observations indicating that some generative AI models can remember the previous interactions with a particular user [96], thus contouring their answers to what the system perceives as the desired type of response that the specific user is seeking.

    Self-reinforcement or feedback loops wherein model outputs become increasingly less diverse over time [82] are another aspect of generative AIs which can potentially serve to amplify hegemonic narratives of atrocities. While the possibility of collecting user input (e.g., positive/negative evaluations of outputs of chatGPT) is important for improving the system functionality, in the case of memorialization, it might also result in prioritization of small sets of possible outputs (e.g., standardized textual or visual narratives of atrocities). Given the highly political battles over narrative creation that are often embedded in memorialization efforts (see, for instance, the case of the Holodomor [45, 72] or the genocide in Rwanda [40]), AI platforms—assiduously apolitical by design—might nevertheless have profound political and social consequences.

  • AI can introduce bias in analysis and make some research practices more obscure/obsolete

    The popularity of AI can entice (or even force) researchers towards relying on AI-driven tools instead of traditional qualitative practices of studying the memorialization of mass atrocities. While it can enable new possibilities for research in the field of memorialization, it might diminish the role of the human in the loop by making obsolete interactions and forms of knowledge that require human embodiment and, by doing so, allow for a wider range of communication beyond what the digital systems can produce or relay. Such a shift can be concerning for a multitude of reasons, including the possibility of generative AIs being subjected to certain forms of bias that, in the case of earlier AI systems, varied from the retrieval of factually incorrect information to systematic skewness of outputs (e.g., in terms of visibility of specific memorial sites and practices [59]).

  • AI can be used to facilitate and enhance censorship

    The ability to program what AI can (and cannot) generate creates not only possibilities but also risks for memorialization. It might improve efforts to detect and combat denial and distortion, but it could just as easily inhibit individuals or groups in vulnerable situations, such as in authoritarian countries, from participating in memorial activities that counter the state's narrative. For instance, AIs can be prevented from generating information about particular atrocities, thus limiting memorial efforts of communities who wish to remember these events publicly. Authoritarian-leaning states seeking to censure memorialization activities may use AI to reinforce a preferred narrative of the past while stifling expressions that stray from that norm.Footnote 9

    Alternatively, AI can potentially produce watered-down or even irrelevant generalities about a given atrocity at the expense of more nuanced complexities that make memorialization meaningful, thus enabling a different form of algorithm-driven “masked censorship” [57, p. 38]. In some cases, the motivation for censorship can be benign—i.e., in the case of commercial companies behind AIs forbidding the generation of atrocities-related content to avoid potential misuse of AIs for trivialization and denialism—but it nevertheless can interfere with the use of technology for memorialization.

  • AI can be used to generate narratives that support distortion and denialism

    Similar to the possibilities for amplifying genuine memory practices, generative AIs can be abused to amplify distortion and denial in the context of mass atrocities. Such abuses can be intentional and unintentional. In the former case, AIs can be used to produce a large volume of diverse content promoting claims that deny that specific mass atrocities happened (e.g., via text-focused AIs) or promoting a distorted representation of the atrocities (e.g., via image-focused AIs generating sexualized images of Holocaust victims, such as Anne Frank, or images glorifying the Holocaust perpetrators). Distorted representations could even be quite nuanced, for example, by producing generative images of atrocities that look similar to the authentic ones but lack important details or by replacing the faces of perpetrators to confuse the public.

    Aside from the risks of intentional technology abuse, generative AIs are prone to hallucinations [5], in which the systems fill up information voids with generated content that is not factually supported. An example of such hallucinations in the case of mass atrocities can be the invention of fictional details about a particular perpetrator or a victim to respond to the user prompt in response to which the AI has no factual information. In this case, the risk is intrinsic to the technology itself (i.e., the ability of AI to generate content): without establishing a system of controls (e.g., informing the user that there is no information available about their inquiry), the use of AI can lead to the distortion of historical facts.

  • AI can undermine trust in the use of technology for memorializing mass atrocities

    The humanlike responses of AI chat systems to users’ questions make these systems inherently more trustworthy to the humans who engage with them [27]. Science fiction films like Spike Jonze’s Her [132] and Alex Garland’s Ex Machina  [131] have contributed to the popular perception of AI as compelling—but also potentially dangerous and untrustworthy. The inherent uncertainty that accompanies trust [67] can be integral to engaging in new experiences and, in some cases, goes so far as to be a ‘leap of faith'Footnote 10 [98]—for example, in extremis in Garland’s film or potentially for some of today’s generative AI users. With respect to the transmission of memory about genocides and other mass atrocity events, however, the impacts of a potential ‘leap of faith’ by users engaging with generative AIs(e.g., chat systems such as chatGPT) could have negative results. For example, users may fail to check questionable claims generated by AIs and potentially repeat them.

    The unsuccessful leaps of faith in relation to the use of generative AI can interfere with the successful integration of these technologies in memorialization practices due to such integration being predicated (as it is for other tasks [22]) on a degree of trust in the technology. There needs to be some level of certainty that the AI system is conforming to expectations both in its use and its output. When applied to subjects as delicate as mass atrocity memorialization, accuracy is often viewed as a key imperative. The guiding transnational belief is that the atrocities should be remembered, and the suffering of victims should be honored and not forgotten [19], and for this purpose to be achieved, the accuracy of representation of atrocities and suffering must be trusted. With this in mind, if a specific form of generative AI was to acquire a reputation for inventing its facts or promoting skewed interpretations of mass atrocities, it would likely lead to an erosion of that underlying trust.

    Given that generative AI’s output is (at least partially) determined by the training data selected by its creators, critics of the technology sector might seek to disparage AI based on the trope that “big tech” routinely censors knowledge production in the interests of a hidden agenda—which, per the logic of conspiracy theory, need not even be specified to be believed. Moreover, irrespective of the systems themselves, individuals may still distrust the AI-produced text and narratives due to unverified claims made about the content's veracity. Under these circumstances, the malperformance of certain forms of generative AIs could create a sort of moral panic, causing severe distrust towards new forms of innovation similar to what had happened when earlier non-generative AI systems spawned fears of filter bubbles [9], and led to distrust towards their use in societally relevant sectors, such as journalism.

  • AI can disclose sensitive information about individuals

    Generative AI models are trained on vast amounts of training data collected from the Internet, and such data may include personal information. Since 2021, research has demonstrated that it is possible to extract such information from earlier AI models [13], and ChatGPT is not exempt from this issue [49]. Similarly, it is also possible to retrieve training images (or sections of them) from models that generate visual content [14]. The sudden launch of these models to the market has left governments scrambling to take action against generative AI providers, for instance, in the case of Italian data protection regulators demanding OpenAI stop using training data of Italian citizens as it is not compliant with the General Data Protection Regulation (GDPR). In response, OpenAI has stopped providing access to chatGPT in Italy [11].

    While the disclosure of sensitive information per se is problematic, in the case of the memorialization of mass atrocities, it raises additional concerns due to the care and sensitivity integral to respecting and honoring victims. Sensitive details about the lives of atrocity victims, including, for instance, images of them being murdered or tortured, prisoner intake photographs, or other visual representations (including images taken before the atrocities, such as victims’ childhood photos) may be problematic in the context of the production of new types of content for public (or even commercialFootnote 11) purposes. However, the exclusion of this material is challenging given the volume of sensitive content available for AI model training as well as the difficulty of drawing boundaries regarding exclusion of such content.

  • AI can create non-authentic content in relation to mass atrocities

    The ability of AI to generate content that can be hardly differentiated from the content produced by a human (e.g., [83]) raises concerns that generative AIs could produce inauthentic content about a mass atrocity which might then be presented and treated as authentic. Such content can vary from fabricated historical documents assigning the blame to specific actors or whitewashing the actual perpetrators to fake visual evidence offering a distorted representation of the atrocities (e.g., to scapegoat a specific group of victims or bystanders). While companies behind AIs will likely try to limit such uses, the susceptibility of models to blindly follow user instructions can result in users circumventing these restrictions in a process known as “jailbreaking” as demonstrated in studies examining the potential of generative AIs to produce extremist content (e.g., [64]). If people believe potentially false statements made by an AI, then they can fail to check the results against other sources and it can lead to perpetuating false claims and exacerbate the problems of denial and distortion already present. Even if users of generative AIs remain capable of detecting and rejecting misleading or extremist claims, there could still be a cumulative effect of undermining the trust in historical sources through the generation of more subtle forms of non-authentic content that steers the narrative toward interests beyond the genuine purpose of atrocity memorialization itself.Footnote 12

5 Other considerations of using generative AI for digital memorialization of mass atrocities

Beyond the possibilities and potential risks identified in the previous sections, there are additional considerations and dimensions related to applying generative AI technologies to mass atrocity memorialization practices. Some of these considerations emanate from challenges that may arise in research efforts to understand the impacts of AI on memorialization, whereas others relate to the long-term implications of integrating generative AIs in the broad field of digital memorialization.

  • Access to data about mass atrocities. The widely used AI models usually rely on information available on the Internet, but most of the content related to mass atrocities is stored in traditional archives, which are often only partially digitized. Therefore, the knowledge that AIs possess about atrocities is limited to already digitized or digital-born content (e.g., PDFs, images, and other content items, including ones coming from non-institutional sites such as Pinterest), which the model can access. Another aspect of data access concerns the tendency of some generative AIs (e.g., chatGPT-3) to rely on a snapshot of data produced at some time point (e.g., only up to 2021). Time-range boundaries can have implications for content generation in relation to mass atrocities, especially concerning the atrocities which happened after the model underlying the AI has been trained and deployed. At the same time, many AI models integrate possibilities for retrieving data from the Web to overcome these time limits; however, doing so also increases the possibility of using unverified or false data.

  • Matters of user privacy. Currently, there is limited understanding of how generative AI platforms use the data provided by their users (e.g., whether they track information requests generated by individual users) and what applications of these data they might consider. In the case of data about mass atrocities, certain requests might be rather sensitive (for instance, inquiries about the war in Ukraine contradicting the official state narrative in Russia for Russian users, in particular when using Russia-based AI platforms).

  • Matters of representation. With respect to the Let Them Speak project, Naron and Toth [69] assert that the aim of the project is to make the fragmented collection of Holocaust testimonies representative of those who perished and, therefore, could not speak. Applying generative AI could take such projects a step further by using separate testimonies to give AI the power to speak and tell coherent stories about victims’ experiences. On the positive side, such a system arguably permits a representation of those who have no voices anymore. However, this move begins to approach the uncanny valley where the chatbot speaker could become a victim of a sort itself or, interpreted through a supernatural lens, as a medium for the ghosts and spirits of the deceased. A chatbot could conceivably become seen as an entity unto itself, distinct from the inputs used to construct its persona (and memories, of a sort). Representing those who have perished in a mass atrocity through a technology that uses existing testimonies to fabricate new ones may also raise ethical concerns about atrocity remembrance along with epistemological ones about how we understand and represent the past.

  • Replacement of the human working force involved in memorialization. The ability of generative AI to produce a large volume of textual/visual content related to mass atrocities enables new possibilities for decreasing the costs of digital memorialization activities. However, these possibilities also create risks for humans working in memorialization-related fields. For example, visual artists, copywriters, data analysts, and certain types of archivists could potentially be reduced or replaced by generative AI technologies. Additionally, generative AIs can perform some functions associated with the roles of curators working in heritage institutions connected with mass atrocities. Not only shall we consider the economic implications of job loss or displacement, but the fact that the involved individuals, directly or indirectly, increase awareness of mass atrocities and propagate the sensitivities of the past suffering to groups that would otherwise not engage with it. Under these circumstances, it is imperative to consider integrating generative AIs into memorialization practices while prioritizing the input and expertise professionals in these sectors have and the importance of social interaction in informing individuals in other societal sectors.

6 An agenda for future research on the role of generative AI in the context of atrocity memorialization

In light of the trends, potentials, and uncertainties in AI-generated atrocity memorialization that we have identified, we now turn to the question of what we, as a scholarly community, can do to realize the possibilities and counter the risks of using generative AI for digital memorialization. Below, we share four recommendations that, in our view, are integral for steering the use of these technologies toward a more ethical and sustainable direction.

Experimenting with AI in the context of mass atrocities. We recognize threats related to genocide distortion/denialism, which can be amplified by AI, but we should not ignore the possibilities provided for the memorialization of mass atrocities. To examine these possibilities, we need to keep experimenting with generative AI and also keep options for experimentation open for other stakeholders, including ordinary citizens. The ability to experiment with AI is essential for understanding the full range of its possible uses for memorialization. While banning the atrocities-related prompts (as done by certain AIs, such as Bing Image Creator) can be a safe option, it could also, problematically, contribute to erasing memories about past suffering by undermining memorialization (and prevention and rescue) efforts. For example, human rights groups seeking to collect and display evidence of war crimes and genocide may have the content of their projects censored out from AI training data (and, consequently, AI system outputs), thus making it difficult for them to show the crimes being committed with the hope of some action [37, p. 10]. Several studies [23, 37, 38] have observed a similar phenomenon concerning human and algorithmic moderation systems used by social media platforms to block potentially offensive content (including evidence of mass atrocities) and then delete such content before perpetrators can be held accountable. Finally, even if major AI companies ban queries dealing with mass atrocities, it might be hard to do it comprehensively (e.g., so users would not find a way to circumvent the limitations), whereas such bans can push individuals towards using generative AI models deployed in fringe environments like the dark web.

Studying how generative AI is used for preserving and distorting memories of mass atrocities in and outside academia. There are indications (e.g., [79]) that generative AIs could become another source of moral panic, similar to earlier forms of AI-driven technology (e.g., recommender systems and the associated filter bubble fears [9]). To counter it, it is important to understand how generative AIs are used in the memorialization of mass atrocities and whether the concerns about the risks of their adoption in the field of memorialization are justified. The complexity of memorialization, which includes a broad range of individual and group practices focused both on remembering past suffering and on researching such practices, adds to the challenging nature of this task.

How to achieve such an understanding is an open question. Unlike content released via websites, blog posts, or social media, memorial activity relying on AI platforms is often confined to private accounts. The inputs and outputs of AI systems are hidden from public view unless or until the user shares them in a public (presumably digital) forum. While there are exceptions (e.g., the rendering of images on Discord servers using models like Midjourney), even in these cases, the degree of communal engagement with generative AI appears to be more limited than with more conventional digital memorialization practices such as commentaries for statements of memorial institutions published on social media.

Under these circumstances, a first step can be to examine cases where individuals or groups integrate AI-rendered text, image, video, or voice content into established memorialization practices (e.g., by adding them to memorial websites devoted to a mass atrocity) and, where possible, conduct interviews with practitioners involved in such integration. Though few currently, such cases are likely to proliferate as individuals and groups seek the capabilities that AI can provide to enhance the message they want to convey about past suffering. Another alternative can be to use experimental research designs to examine how individuals use AIs to generate memorialization-related content or engage with such content in a controlled environment. Similarly, one might track the extent to which heritage institutions and human rights organizations adopt generative AIs, along with their respective motivations and expectations.

Arguing for the importance of integrating transparency principles in the generative AI design and functionality. Generative AI design must follow the principles of transparency both in data acquisition for training the underlying models (e.g., to ensure that data was acquired without breaching intellectual rights) and in providing possibilities for analyzing the performance of generative AIs and their potential bias. Similarly, greater transparency is needed in the functionality of generative AIs: for instance, by providing sources for data that are used to generate outputs (when possible) so potential privacy or copyright breaches would be easy to identify. Transparency is especially important in the context of using generative AIs for producing content related to mass atrocities due to associated ethical and moral considerations (e.g., the importance of treating the memory of victims respectfully) as well as the need for public cooperation in the effort to identify potential risks of AI misuse. Under these circumstances, transparency of AI design and functionality becomes a main prerequisite for AIs becoming reliably truthful and establishing a reputation as such.

Watch AI developments beyond the West and Silicon Valley. The development and application of generative AIs to mass atrocity memorial practices are not confined to the West and Silicon Valley. Generative AIs are employed, refined, and redesigned to meet the needs and tastes of various cultures and nationalities around the globe. For instance, the Cambodian government is currently creating its chatGPT system for the public [51], and another version was created in Saudi Arabia (miniGPT-4 [61]). While it is not clear yet when or how these platforms will be used for memorial purposes, there is little question they will eventually be adopted to memorialize past and present suffering. In China, for example, the government has sought to place guardrails on the range of sources a chatbot might use to discuss particular events, such as the Tiananmen Square massacres [16], and the Cambodian government is evaluating the risks of the public using the Khmer chatGPT as a primary source of information [93]. Moreover, the speed at which these technologies are being developed worldwide is astounding. Cambodia, which is considered a developing nation and has been struggling for decades to recover from the devastating Khmer Rouge era, has already produced a robot that speaks using the model underlying the ChatGPT platform [74]. What do these developments suggest for memorialization? How will these varying forms of ChatGPT impact the memorial narratives produced? We can expect the unexpected: developments that Silicon Valley could not have anticipated as the AI models are increasingly trained implicitly and explicitly based on different core ideologies, cultures, and political agendas.

7 Discussion

Generative AIs defy conceptual premises regarding the relationship between digital technologies and memorialization by surpassing the information retrieval functionalities which are commonly associated with the use of AI for digital memorialization. Instead of ranking existing memorialization-related content in response to human queries, generative AIs utilize complex models to produce responses that to a certain degree emulate the human ability to interpret and comprehend information. While the emulation capabilities of generative AI may fall short at times, they can expand the analytical capacities of researchers studying memorialization and amplify possibilities for the general public to learn about past suffering. However, the possibilities enabled by generative AI also come with a price (as demonstrated by the discussion of shortcomings of its use in the context of memorialization above). Therefore there is an urgent need for careful critical assessment of appropriateness of using generative AI and its capacity in this sensitive context. Adding to considerations of its application to mass atrocity memorialization is the importance of extensive contextualization of each unique instance of atrocity and the loss of life and suffering that these events entail.

The major challenge of enabling such contextualization relates to the difference between human and AI capacities of understanding semantics of articulations and representations of specific instances of mass atrocities and their memorialization. For generative AI to be meaningful for memorialization, it needs to achieve a certain understanding of the meaning of texts, images, and utterances involved in the production of stories about suffering as well as the ethical and moral dimensions of these stories. The ability to understand these complexities is also crucial for how AI-generated content about mass atrocities will be received by the human audience and what meanings embedded in memorialization practices will be rendered or lost. In this context, Walter Benjamin’s [3] observations regarding the aura of an artwork that is grounded in its originality and immediate connection to its human maker are particularly relevant. Benjamin argues that the mechanical reproduction of a work of art—e.g., by AI models which reuse elements of existing texts or images to create new ones—severs the art from its human maker, effectively removing the work of art’s aura.

Applying this to generative AI memorialization of mass atrocities a similar and more poignant conclusion can be drawn when the “work of art” is witness testimonies or other representations of collective trauma and suffering. Even if a generative AI can produce a relatively convincing version of a textual or visual narrative of a mass atrocity, it begs the question (beyond the obvious ethical issues) whether the human element was lost along the way. Turning to the human audience, there is then the question of perception. Beyond perceiving the AI-generated story (as film, image, or narration) as passable by objective standards, is there still something that is lost in its reading? Roland Barthes [129, p. 27] identifies what he calls the punctum that photographic images can evoke: “that accident which pricks” that are the details in a photograph that elicit human meaning and connection—a missing button on a jacket, a particular look of the subject, a wilted flower. These are the bits that connect the viewer to the subject by being subjective and born from human experience as in the case of. the small familiarities or notable absences. Bender’s Octopus [2] lacks the experience and imagination to make these connections and inferences. Arguably, neither can generative AI, and even if it could, what meaning or purpose would that serve?

These considerations are integral for critically assessing the implications of using generative AI in the context of memorialization in different areas, including public policy and international relations. In the context of public policy, generative AI can enable new modes of accessing information about mass atrocity (e.g., as part of school education and life-long learning) that can increase societal awareness about the mass atrocities which is essential for their prevention. Similarly, generative AI can shape new standards of using atrocity-related content for memorialization purposes (e.g., by determining what can and can not be generated). This can lead to harmonizing knowledge about the past and sustaining the production of content that can be used for grassroots memory practices while countering the spread of distortion. However, without careful understanding of the context in which the atrocity occurred and the semantic complexities of accounting for the diversity of individual experiences, such knowledge harmonization can result in the simplified representation of the atrocity that can undermine human-led memorialization efforts or even silence victim voices (e.g., in the case they belong to the groups which are marginalized in the present).

The implications of relying on generative AI for memorialization in the context of international politics can also be rather ambiguous. The ongoing rollout of generative AI systems in different parts of the world suggests that the national agendas regarding the use of AI for mass atrocity memorialization will not necessarily be driven by international consensus. Such a consensus can be even more difficult to achieve considering that different countries may rely on their own generative AI platforms together with country-specific datasets used for training models powering the respective platforms and distinct sets of AI guardrails. Under these circumstances, the possibility of generative AIs being used to fuel cross-state online memory wars—i.e., the discursive battles dealing with contrasting interpretations of the past [130]—becomes more likely with AI serving as a source of content promoting specific narratives and subjugating the alternative interpretations.

The discussion of the implications of the use of generative AI for mass atrocity memorializations in the context of public policy and international relations guides us to the crucial question about how such use will affect the role of human practitioners in the processes of atrocity memorialization and the ways to keep the human in the loop. While it is too early to know how exactly generative AI will add value to human memorialization practices, our work suggests that it can enable a more multidimensional and comprehensive understanding of the past, for instance, by summarizing insights coming from the large volume of testimonies, legal documents, and other historical materials relevant to memorialization. It can also provide representation of personal experiences that are difficult to convey and where in the interest of protecting the identities of victims is essential.

At the same time, any benefits of generative AI for memorialization must be considered with the potential harms it may cause and this is why it is imperative that humans guide, analyze, and review any memorialization-related process incorporating AI. Without such supervision, generative AI can easily create confusion and distrust by producing false representations of past suffering that can make individuals doubt the accuracy of related atrocity narratives. Such uncertainty (or false certainty) can exacerbate concerns about the rise of denialism and distortion in relation to mass atrocity which are associated with social media and other digital developments. To address this problem, it is essential to implement mechanisms for human supervision of AI with a particular emphasis on transparency of using AI-driven platforms as well as data and procedures used for training AI models. Such transparency mechanisms shall be supplemented by capacities for monitoring the performance of generative AI and nurturing digital literacies for different demographic groups (and not only teenagers/young adults) to prevent the potential erosion of trust towards AI-enhanced memorialization practices.

The implementation of the above-mentioned mechanisms is essential for integrating generative AI with human memorialization practices, in particular considering the limited capacities of AI to comprehend the full scope of human semantics in the context of memorializing mass atrocities as it was discussed earlier. Any memorialization process needs to be human-driven rather than machine-driven for it to be a true representation of the human experiences it is intended to convey and to respect the dignity of those who perished. Generative AIs can augment human understanding provided it does not directly shape that understanding (e.g., by representing only particular subsets of information or by highlighting or repressing certain experiences). The practical procedures that might be employed by humans are difficult to predict in their entirety, however, the ability of human practitioners to stay informed about how generative AI works in the context of memorialization and also be able to guide its design and usage may be considered as a central requirement to the process. For this, establishing national and international guidance and policies for its use, especially pertaining to mass atrocity cases is fundamental.

8 Conclusion

The foregoing insights about the impact of generative AI on mass atrocity memorialization must come, as is the case for any future-looking discussion on AI, with the caveat that there are more unknowns than there are knowns. As the developments in the field of AI accelerate, academic research and societal discussions must remain responsive to the ongoing developments. This article offers some guidelines in terms of what scholars, as well as practitioners in both the AI and memorialization fields, may need to consider at this moment. As is the case with other rapidly evolving subjects, the discussion of possibilities and risks of relying on generative AI in the context of memorialization, as well as considerations associated with such reliance, can help to establish a framework of reference for anticipating practical and conceptual issues and guiding future debates on the topic. Transparent and collaborative efforts drawing from various academic disciplines can help to address the challenges presented by this emergent technology to support its ethical integration with existing memorialization practices.