Introduction

“A groundbreaking fusion of data-driven creativity and artificial intelligence, poised to redefine the boundaries of innovation and transform the future of digital landscapes.”

-ChatGPT on GPT-4, Mar 23 Version, on Generative AI

In an era where applications like ChatGPT set records for the fastest-growing user base by demonstrating unprecedented domain-independent expertise (Hu, 2023), the concept of “Generative Artificial Intelligence” (GAI) emerges as a disruptor in the digital landscape (Dwivedi et al., 2023; Teubner et al., 2023; Wessel et al., 2023). With capabilities to generate high-quality, contextually relevant content almost indistinguishable from human-created work, discussions arise on whether this new technology even holds early signs of artificial general intelligence (Bubeck et al., 2023; The Washington Post, 2022). Regardless of discussions on AI’s sentiency, the latest advancements in machine learning (ML) and deep learning (DL) have extended traditional, data-driven AI tasks such as predictions, classifications, or recommendations toward the generation of unique, realistic, and creative content. Prevalent collaborations between humans and intelligent systems in businesses and private life have been shaped by the adoption of AI in various ways, especially in the field of process optimization and decision-making (Brynjolfsson & McAfee, 2016; Burström et al., 2021; Moussawi et al., 2021). GAI addresses this development by providing novel augmentation and automation prospects in form of innovative services and business models (Huang & Grady, 2022; Mondal et al., 2023), e.g., by assisting customer support employees with suggestions of appropriate conversation responses (Brynjolfsson et al., 2023). The fast-changing and continuously evolving landscape of GAI calls for an extensive conceptualization of the properties and capabilities to fathom the phenomenon (Dwivedi et al., 2023; Strobel et al., 2024).

While GAI research and development is continuing to invest toward better, faster, and more capable models (e.g., Microsoft, 2023), studies on the fundamental principles, applications, and socio-economic impact remain largely unexplored in the academic discourse (Strobel et al., 2024; Susarla et al., 2023; Wessel et al., 2023). GAI provides innovation opportunities for various domains (e.g., networked businesses and digital platforms) but also comes with challenges (e.g., transparency, biases, and misuse) that need to be addressed for successful implementations (Houde et al., 2020; Schramowski et al., 2022; van Slyke et al., 2023). However, an examination of the key concepts is yet to be conducted, leaving a clear image and understanding of generative AI undefined. To overcome that shortcoming, this article provides an introduction to the fundamentals of generative AI, with its concepts, applications, and challenges. To do so, we exploratively synthesize recent literature on the technical foundation leading toward generative AI in combination with contemporary empirical examples of generative AI models and applications. Our aim is to conceptualize the key properties of GAI and differentiate them from ML and DL methods, to foster the understanding of the theoretical foundations of generative AI, and to guide further endeavors in examining as well as designing generative AI-based systems.

The remainder of the article is structured as follows: In the next section, we conceptualize generative AI and provide a distinction between related AI methods as well as outline the technological foundations. Afterward, we elaborate on the prospects and applications of the generative AI value chain and examine the impact of different generation modalities. Then, we address the potential challenges of adopting generative AI before concluding with a brief research outlook.

Conceptualization

The field of artificial intelligence (AI) has taken a prominent place in research and practice across various disciplines for the past decades. Especially in information systems (IS) research, the socio-technical impact of AI as a phenomenon is at the core of investigation (Ågerfalk et al., 2022; Berente et al., 2021). These technologies have transformed the way we interact with data and make decisions, leading to uncharted ways in society as well as economy (Fügener et al., 2021; J. Li et al., 2021; van den Broek et al., 2021). However, as we venture into the new era of generative AI, it becomes increasingly crucial to understand the core concepts and distinctions within GAI as a rapidly evolving technology. To differentiate GAI from other AI concepts and provide a fundamental conceptualization, we will present a brief overview of AI and its subfields, machine learning (ML) and deep learning (DL), in the next section (Fig. 1). Afterward, we continue to elaborate on how DL has driven the development of deep generative models to enable distinct GAI characteristics and capabilities, ultimately leading to an even broader array of AI applications and opportunities for various fields.

Fig. 1
figure 1

Generative AI and other AI concepts (inspired by Goodfellow et al., 2016, p. 9; Janiesch et al., 2021, p. 687)

From artificial intelligence to deep learning

Artificial intelligence is considered an umbrella term, spanning over different computational algorithms capable of performing tasks that typically require human intelligence, such as understanding natural language, recognizing patterns, making decisions, and learning from experience (Castelvecchi, 2016; Winston, 1993). Early AI systems, such as expert systems and knowledge bases, were rule based and aimed at supporting users and businesses in decision-making (Harmon, 1985; Patterson, 1990). Machine learning as a subfield of AI deals with the development of algorithms capable to autonomously solve tasks through exposure to data without being explicitly programmed—i.e., learning (Brynjolfsson & Mitchell, 2017). In the realm of ML, there are several types of learning approaches based on the nature of the data and the desired outcome. Supervised learning is a common approach, e.g., for applications in commercial contexts, as algorithms are trained on labeled datasets to classify or forecast (business) data (Janiesch et al., 2021). The algorithm learns to map inputs to outputs and, thus, is capable of making predictions on new, unseen data. Moreover, unsupervised learning (i.e., discovering hidden structures or patterns within unlabeled data) and reinforcement learning (i.e., learning optimal decision-making by interacting with an environment and maximizing cumulative rewards over time through trial and error) are further learning strategies in ML (Kühl et al., 2022). What ML algorithms share in common are their discriminative properties, i.e., the goal of processing data to conduct classification, regression, or cluster and determine decision boundaries. Exemplary algorithms include decision trees, k-nearest neighbors, or support vector machines (Ray, 2019). Deep learning is a more advanced subset of ML and leverages artificial neural networks to model complex data representations and automatically detect correlations and patterns in large datasets (Janiesch et al., 2021; Samtani et al., 2023). Neural networks are computational models inspired by the structure and function of the human brain, consisting of interconnected layers of artificial neurons (Goodfellow et al., 2016). In DL, neural networks comprise of multiple hidden layers in a nested architecture to learn hierarchical feature representations from the data, leading to improved performance on various tasks. Thus, DL is capable of processing high-dimensional data in various domains, ranging from one-dimensional data like signals and texts to multidimensional data such as images, video, or audio (LeCun et al., 2015). These advances have enabled a plethora of use cases across different domains, from societal good, such as improving healthcare and environmental sustainability (Piccialli et al., 2021; Schoormann et al., 2023; Strobel et al., 2023), to electronic markets, where DL can optimize pricing, serve as recommendation systems, forecast demands, and detect fake consumer reviews (Ferreira et al., 2016; M. Li et al., 2022; Zhang et al., 2023b).

Toward generative AI

Fueled by advancements in DL techniques, deep generative models (DGMs) have emerged as a class of DL models to generate new content based on existing data, creating a variety of new possibilities for AI applications (Lehmann & Buschek, 2020; Tomczak, 2022). These models are trained to understand complex data distributions, which allows them to produce outputs that closely resemble real-world data. By leveraging statistics, the goal of DGM training is to learn high-dimensional probability distributions from a finite training dataset and create new, similar samples that resemble an approximation to the underlying class of training data (Ruthotto & Haber, 2021). While discriminative models focus on modeling the relationship between input features and output labels, generative models learn the inherent data structure and generation processes (Jebara, 2004). Generative models have been around for decades, with, for example, hidden Markov models or Bayesian networks aiming to model statistical problems involving time series or sequences (Gm et al., 2020). Nonetheless, DGMs relying on neural networks have paved the way for significantly higher-quality generated content in recent advancements in the field of so-called generative AI. Thus, the goals of DGMs differ from traditional discriminative AI models (e.g., in ML) because the focus lies on the probabilistic generation of new data instead of determining extant data’s decision boundaries (e.g., classification, regression, or clustering) (Tomczak, 2022; Weisz et al., 2023). In the following, we will focus on DGMs as the underpinning of GAI and give an overview of four core DGMs that have shaped the evolution of GAI in Table 1.

Table 1 Overview of core deep generative models

To leverage DGMs in GAI applications, they can be trained to generate new data and enable a variety of use cases (we refer to DGMs implemented in GAI applications as GAI models). Training a GAI model can be different than a discriminative AI model due to semi-supervised learning, a combination of learning techniques leveraging a small amount of labeled data (i.e., supervised) followed by extensive unlabeled data (i.e., unsupervised) (Kingma et al., 2014). For instance, recent GAI models apply techniques like supervised fine-tuning (SFT), reward models, and reinforcement learning via proximal policy optimization (PPO) to achieve an alignment of the model with the developers’ intentions and values (OpenAI, 2023; Ouyang et al., 2022). This unique approach allows the training of very large datasets required for GAI models without the need for difficult complete labeling.

The application system functions as an interface for the user to interact with a GAI model. Prompting is an interaction technique and unique GAI property that enables end users using natural language to engage with and instruct GAI application (e.g., LLMs) to create desired output such as text, images, or other types (Dang et al., 2022; Liu & Chilton, 2022). Depending on the application, prompts vary in their modality and directly influence the mode of operation. For instance, text-to-image applications use textual prompts describing the visuals of the desired image, while image-to-image applications rely on an input image to steer the generation process.

By design, the outputs of generative AI models are probabilistic and not replicable compared to the deterministic outcomes of discriminative AI—i.e., variance (Weisz et al., 2023). For one exact input prompt, a GAI application will generate varying outputs each time it is prompted, but the results remain valid and prompt fulfilling. On the other hand, different input prompts can lead to the same goal. Hence, formulating a meaningful prompt that leads to the desired outcome is based on trial-and-error process, e.g., by rephrasing textual prompts with the same keywords. The field of prompt engineering deals with systematically constructing prompts to improve the generated outputs (Liu & Chilton, 2022).

Based on the heuristic approach of prompt engineering and the inherent variance in the generated content, GAI users continuously and iteratively specify their desired tasks as input prompts to generate outputs until their task is solved. The primary goal of generating new, probabilistically produced data (i.e., content) with varying outputs based on the same input distinguishes generative AI from discriminative AI, which pursues boundary determination by analyzing data and making a decision (see Fig. 2). Hence, a primary difference lies in the role of data, as GAI leverages very large datasets in its generative model to produce diverse content, while discriminative AI processes user data based on a (pre-trained) algorithm.

Fig. 2
figure 2

Procedural differences of generative AI and discriminative AI

Prospects and applications of generative AI

Complementing discriminative AI, GAI has recently emerged as a novel tool with a wide range of new possibilities impacting multiple sectors, from education and healthcare (Brand et al., 2023; Burger et al., 2023; Cooper, 2023) to networked businesses (Dwivedi et al., 2023; Wessel et al., 2023). These emerging applications inherit the generativity and variance properties of GAI and, therefore, are capable of producing unique and creative content, going beyond mere assistance. Hence, GAI becomes increasingly multidisciplinary, enabling disruptive innovations and automating even traditionally creative tasks, e.g., by generating customized contextual texts or images, facilitating new opportunities for businesses to innovate and differentiate themselves in the competitive economic landscape (Dwivedi et al., 2021; Lund et al., 2023; Pavlik, 2023).

Generative AI finds its utility across various modalities, including the generation of text, image, video, code, sound, and other produced content, such as molecules or 3D renderings (see Table 2). For example, GAI applications aim to create tailored marketing content, generate realistic (product) images or videos, and even assist in software development by generating code (Bakpayev et al., 2022; Elasri et al., 2022; Kowalczyk et al., 2023). Several modalities can serve as the input for GAI models. Distinguishing the different modality types, unimodal models generate the same output type as their input type, e.g., text-to-text or image-to-image generation, whereas multi-modal models combine different input and output types, for instance, in a text-to-image or code-to-text scenario. Different multi-modal models can subsume as x-to-modality models (e.g., x-to-text or x-to-image).

Table 2 Overview of different output modalities for generative AI applications

Examining the architecture of GAI-based systems, three major component layers can be identified: model layer, connection layer, and application layer (see Fig. 3). These parts embed generative AI in its information systems context and draw a boundary from external entities that can be interacted with its environment (e.g., users, organizations) and data (i.e., public and enterprise data) (Samtani et al., 2023). Inside the boundaries of GAI-based systems, the prevalent characteristics of generativity and variance persist and affect all layers and processes. The model layer comprises the pre-trained, deployable GAI artifact (i.e., a DGM) for application systems. Depending on the training dataset, general purpose models aim at solving a wide range of tasks in multiple domains (e.g., GPT-4), whereas customized models are designed for domain-specific tasks and were, therefore, trained on highly specific data (e.g., CodeBERT).

Fig. 3
figure 3

Conceptual framework of generative AI

Integrating these models into a system environment that affects people and organizations leads to the application layer of generative AI. By providing a proper context for the artifact, users are able to leverage the capabilities of GAI models for a specific application use case. Observing the trend of various recently emerging GAI applications building on top of existing models, it becomes apparent to further distinguish between end-to-end applications that are based on undisclosed, proprietary models (e.g., Midjourney) and open applications that are built around open-source models or leverage publicly accessible pre-trained models (e.g., Jasper and Elicit using OpenAI’s GPT-3 (Elicit, 2022; Jasper, 2022)). Huang and Grady (2022) describe GAI applications as a “UI layer and ‘little brain’ [i.e., application layer] that sits on top of the ‘big brain’ that is the large general-purpose models [i.e., model layer].” This perspective emphasizes that new business models and applications can be developed without the need to train large GAI models from scratch by leveraging publicly available application programming interfaces (API) or AI-as-a-service platforms (Burström et al., 2021; Janiesch et al., 2021; Lins et al., 2021). Indeed, the accessibility and availability of pre-trained GAI models foster value co-creation and can be leveraged via a connection layer (e.g., Hugging Face). Fully integrated GAI systems, on the other hand, employ their own, custom-trained proprietary models. In many cases, end-to-end GAI applications represent fully integrated systems (e.g., GitHub Copilot), while open GAI applications leverage external models via APIs (e.g., Stable Diffusion). Overall, GAI models may aim for general purposes or customized tasks regardless their connection or application characteristics (see Fig. 3). To enrich GAI-based systems with additional data beyond their training state and the GAI boundary, external data sources can be connected. Enterprise data (e.g., internal documents, enterprise resource planning (ERP) systems, knowledge bases) and public data (e.g., the Internet, libraries, social media) may serve as complementary, contextual data that GAI applications can further draw upon for more relevant and personalized results.

Employing GAI in enterprises can extend the level of assistance for workers and open up opportunities for augmentation and automation of the job, leading to new forms of collaborations between humans and machines (Einola & Khoreva, 2023). Furthermore, GAI transforms the way businesses operate in their daily tasks, innovate, and interact with their customers (Brynjolfsson et al., 2023; Mondal et al., 2023). Thus, the prospects of value co-creation come in hand with potential changes in human work roles, requiring workforces in various domains to adapt their tasks as a diverse set of tasks could be impacted by generative AI (Brynjolfsson & McAfee, 2016; Eloundou et al., 2023). The ongoing diffusion of AI into businesses gets accelerated by GAI applications, resulting in a possible replacement of human jobs on the one hand but also the creation of new jobs (e.g., for prompt engineers or with new business models) on the other hand (Einola & Khoreva, 2023). Hence, the effect on the labor market by the disruption needs to be discussed, and businesses should seek to understand and embrace the potential of generative AI (Eloundou et al., 2023; Willcocks, 2020).

Challenges for generative AI-based systems

While generative AI holds transformative potential for individuals, organizations, and society due to its vast possible application space, the technology also inherits various challenges that parallel those of traditional ML and DL systems. The domain of electronic markets is a prime example that moved into the center of transformation due to its latest focus on data-driven efforts (Selz, 2020). Outlining and emphasizing these challenges relevant for research and practice helps to raise awareness of the constraints as well as supports future efforts in developing, implementing, and improving GAI-based systems.

Bias

Because of GAI’s data-driven nature, data quality plays an essential role in how GAI-based systems perform and, thus, how feasible their adoption for real-world scenarios in business contexts is. Similar to their traditional discriminative AI relatives, GAI models are prone to bias causing biased decisions, disadvantages, and discriminations (Ferrara, 2023; Schramowski et al., 2022). Biases manifest in different ways and evolve primarily during two development phases of an AI-based system: training and inference.

Data bias gets injected during the model’s training phase and leads to biased results because of faulty datasets. Factors such as non-representative, imbalanced sampling, incorrect labeling, and mismeasured features during the selection and processing of datasets hinder an unbiased training of the GAI model, ultimately leading to biased algorithmic outcomes (Mehrabi et al., 2022; Ntoutsi et al., 2020). The development of large-scale training datasets is especially important for GAI models and often involves strategies of scraping public-available data on the Internet (Schuhmann et al., 2022). This approach is usually performed unsupervised and autonomously, which complicates the dataset’s quality assurance because of its large quantity of unstructured data. Since GAI models are often trained to be general-purpose and multi-modal, they require and rely even more on such training datasets. Hence, moderating potential data bias is crucial for applications in business contexts like electronic markets due to the closeness to customers (e.g., points of contact via advertisements, social media, or customer support). Furthermore, social bias as a form of data bias can cause distorted views in generated texts or images and should be considered as well as mitigated (Baeza-Yates, 2018).

Algorithmic bias is introduced during the inference phase, independent from the model’s training dataset (Mehrabi et al., 2022). In this case, the models have been trained on diverse, unbiased input data, and either the model’s algorithm or the application around it introduces biases affecting users. Overfitting is a typical phenomenon that originates from the chosen learning strategies or optimization functions and causes biased algorithmic outcomes (Danks & London, 2017; Hooker, 2021). In this case, GAI models might introduce biases not reflected in the data because they fail to learn the data distribution correctly. Likewise, the presentation of and the user interaction with GAI-based systems can cause biases, such as when only selected generated content (e.g., one image out of multiple variants) is shown to the user (Baeza-Yates, 2018).

Thus, generative AI applications exerting biased results influence users’ opinions and judgement and require control mechanisms (Jakesch et al., 2023a). Strategies should be developed to prevent, detect, and mitigate biases in order to safeguard users and ensure the service quality and reputations of a company. One approach to steer the quality of outputs from GAI models is via reinforcement learning from human feedback (RLHF) (Christiano et al., 2017; Griffith et al., 2013). The technique involves feedback from human evaluators to guide the model’s training process, with evaluators assessing and comparing the quality of generated outputs. This approach enables generative models to refine their output generation process, aiming for better alignment with human expectations and objectives. Nevertheless, determining what content is “good” or “right” remains a difficult and bias-prone task (Teubner et al., 2023).

Transparency

The need for explainability arises with the unpredictability of the inherent generative nature of GAI models and the overall functionality of ML models as “black boxes” (Janiesch et al., 2021; Meske et al., 2022). While the impact of GAI may not reach as far as discriminative AI use cases (e.g., decision-making or dynamic pricing), research on explainable generative AI is still in its infancy, and the justification for more transparency is without a doubt (Brasse et al., 2023; Sun et al., 2022). Governments are already discussing the enforcement of AI regulations that include explainable AI to protect the general society and mitigate risks tied to the technology (Hamon et al., 2020). Interpretability (i.e., the human capability to understand the AI system’s processes and decisions) is key, especially for GAI-based systems employed in large-scale information systems that affect large user groups, such as in networked businesses and digital platforms. In these cases, generated content has the potential to impact individuals and society, for instance, when generative AI serves as an advisor based on user questions and provides unsophisticated answers that are difficult to verify. Inaccuracy in generated product recommendations can have varying consequences depending on the situation, ranging from selecting the wrong product to taking the wrong medication. Early studies have shown how the chatbot ChatGPT performed surprisingly well in medical exams, suggesting inherent knowledge similar to medical students (Bhayana et al., 2023; Gilson et al., 2023). However, the seemingly omniscient capabilities may be restricted because, in the case of the medical exams, the GAI model might have been trained on the exam data and can reproduce its answers but is not able to comprehend the contextual state of an individual relevant for medical assessment. Therefore, understanding how the system performs sensemaking and generates its data helps users and businesses to achieve their goals responsibly and effectively, satisfying stakeholders’ needs and expectations (Miller, 2019; Sun et al., 2022). Particularly for autonomous systems in critical business applications that interact with human beings, supervision and explainability of the GAI-generated content remain vital to ensure reliable, safe, and trustworthy outputs (Brasse et al., 2023; Hamm et al., 2023).

Another angle of transparency concerns the debate between open-source and closed-source models. Legal issues revolving around copyright, licenses, and intellectual property make it difficult for individuals and enterprises to deploy GAI-based systems, especially when the large training data of closed-sourced GAI models is procured through Internet scraping (Jin et al., 2023; Smits & Borghuis, 2022). Research initiatives revolving around open-source datasets (e.g., Schuhmann et al., 2022) and open-source models (e.g., Stability.ai, 2023) aim at increasing the transparency on data provenance and highlight, for instance, the data sources as well as the presence of watermarks on images (Schuhmann et al., 2022). Although most system engineers will rely on pre-trained models and perform fine-tuning for their specific use case, open-source efforts are facilitating a legally safer route for businesses to deploy GAI models, ensuring legal compliance, and mitigating associated risks.

Hallucinations

Due to the probabilistic variance property of GAI, generative models are not immune to output errors, or so-called hallucinations, which manifest themselves in confidently generated results that seem plausible but are unreasonable with respect to the source of information (Ji et al., 2023; Susarla et al., 2023). The underlying causes of hallucinations are still being researched, with early findings suggesting that the training data that might contain contradictory or fictional content besides factual information (Dziri et al., 2022). This combination of varied inputs can lead to the generation of outputs that deviate from reality and introduce false information. As a result, the uncertainty of generation quality is further fueled by closed-source models that do not disclose any information on their training, making it crucial to carefully select appropriate datasets and models to mitigate the risk of hallucinations.

To illustrate the occurrence of hallucinations, studies have identified GAI-based image generators that facilitate anatomical inaccuracies in the generated images of humans (Choi et al., 2022; Hughes, 2023). These inaccuracies suggest that GAI models require further refinement and improvement before they can be reliably used for unsupervised production tasks (e.g., advertisement production, automated social media posts). Additionally, errors in simple arithmetic operations have also been observed (Bubeck et al., 2023), highlighting the limitations and potential shortcomings of current generative models in performing even basic computations accurately. Due to the seemingly realistic data produced by GAI models, the detection and evaluation of hallucinations are a challenging task. Current automatic evaluation includes statistical methods to measure discrepancies between ground-truth references and generated data and model-based metrics that leverage additional DL models to detect content inconsistencies (Ji et al., 2023). However, both approaches can be subjected to errors and are still inferior to cumbersome human evaluation. These instances emphasize the importance of appropriate hallucination mitigation methods, such as human supervision, to ensure the quality and accuracy of generated content.

Moving forward, addressing the issue of hallucinations in generative AI requires ongoing research and development efforts. Enhancing the transparency of training data and computation processes as well as promoting the adoption of open-source models can help mitigate the risk of generating misleading or flawed results. Furthermore, refining the underlying algorithms and incorporating robust error-checking mechanisms can contribute to the overall reliability and trustworthiness of GAI models (Zhou et al., 2023).

Misuse

The access to GAI-based content creation tools with realistic outputs does not only enable new creative opportunities for good (e.g., novel automation and innovation prospects across multiple domains for businesses and users), but can also be leveraged for malicious purposes to intentionally cause risk and harm to society (Weidinger et al., 2022). Deepfakes have become increasingly sophisticated over the past decade as a result of the low cost and ease of creating such media using x-to-image, x-to-video, and x-to-sound GAI models (Mirsky & Lee, 2022; Vasist & Krishnan, 2022). They are authentic media content designed to impersonate individuals, such as celebrities or politicians, and are created to entertain or manipulate the viewers. For example, deepfakes depicting Trump’s arrest have circulated on social media and in the news, causing misinformation due to their hyperrealistic, almost unrecognizable appearance (BBC, 2023). This is just one example of how generative AI carries potential for abuse and can be extended to other areas of daily life, such as fraudulent service offers, identity theft, or fake shops (Houde et al., 2020; Weidinger et al., 2022). The availability of GAI models provides starting points for new applications and business models for misuse and criminals, ultimately being leveraged to spread misinformation and influence the media and politics, or to defraud individuals and businesses (Hartmann et al., 2023; Kreps et al., 2022; Mirsky & Lee, 2022). Such issues, combined with low data quality and bias, provide a preview of potential social and ethical harms that may result in discrimination, exclusion, toxicity, and information hazards (Weidinger et al., 2022). This increases the urgency for disclosure and transparency of GAI models, along with the aforementioned calls for explainability (Brasse et al., 2023; Horneber & Laumer, 2023; Raj et al., 2023).

Generative AI researchers seek to develop measures for safer and more responsible use (van Slyke et al., 2023). RLHF and carefully crafted open-source datasets are first attempts at improvement, besides input filters restricting user prompts to harmless content. However, applications can still be tricked into bypassing filters and safeguards of GAI models, for instance, through prompt injections that insert malicious prompts to achieve misaligned outputs of generative AI applications (Perez & Ribeiro, 2022). The collaborative efforts between researchers, organizations, and regulators (e.g., by initiatives such as the European Union AI Act or US National AI Initiative) serve as promising initial steps toward opening pathways for future research to effectively address these issues, ensuring that AI-generated content is morally, ethically, and legally appropriate and cannot be misused (Hacker et al., 2023).

Societal impact

With its ability to produce novel and diverse content, generative AI has significant implications for society in several areas. Besides the aforementioned risks of misuse and misinformation, the environmental footprint of developing and operating any AI-based system is an emerging concern and gets amplified for GAI-based systems due to the demanding technical requirements for training and running large DGMs and their high implicit energy consumption (Schneider et al., 2023; Schoormann et al., 2023). Consequently, the public acceptance of AI on a societal but also individual level inevitably guides the future development and use of GAI. Individually, researchers have investigated how AI can be perceived as a threat by some users (Johnson & Verdicchio, 2017; Lysyakov & Viswanathan, 2022). For example, a fear of employees is losing their jobs due to AI’s automation potential (Mirbabaie et al., 2022). With GAI’s capabilities of producing high-quality, customized realistic content, the frontier toward machines taking over human tasks moves even closer. Hence, trust plays a pivotal role in the adoption of emerging technologies and comes with direct impact between users and application providers (Lukyanenko et al., 2022; Riedl, 2022; Wanner et al., 2022; Yang & Wibowo, 2022). The promotion of explainable and responsible AI becomes paramount once more (Meske et al., 2022), especially in a technological state where distinguishing between generated and real content by GAI models is becoming increasingly more challenging (Jakesch et al., 2023b; Lehmann & Buschek, 2020). Early studies on generative AI suggest that generated content is perceived as less trustworthy, resulting in aversion toward generated products (Longoni et al., 2022; Rix & Hess, 2023). Nonetheless, GAI applications offer a variety of value propositions that revolve around building a trusting relationship toward such systems (Tomitza et al., 2023). For instance, social chatbots like Replika aim to realistically mimic or replace humans, leading to feelings of attachment toward machines among users (Pentina et al., 2023), whereas assistive GAI applications intent to support their users in the best way possible (Burger et al., 2023; van Dun et al., 2023).

While generative AI promises significant advances and has the potential to revolutionize domains, such as marketing, arts and culture, or even electronic markets as a whole, by supporting, augmenting, and automating a wide range of operations and offering novel services, it also underlines the importance of fully understanding its inherent challenges and mitigating possible risks. The far-reaching implications of GAI, ranging from poor data quality and hallucinations to cases of societal misuse, require a proactive and sustained effort by researchers and practitioners to develop appropriate solutions ensuring the responsible and beneficial integration of generative AI technology into our digital society. In Table 3, we summarize the aforementioned challenges of GAI-based systems and suggest future research questions revolving around the GAI system, environment, and data perspective.

Table 3 Future research questions regarding the challenges of generative AI-based systems

Conclusion

With this fundamentals article, we provide an introduction to generative artificial intelligence. Drawing on ML and DL as the underlying technologies, we conceptualized generative AI and differentiated it from traditional AI methods. We outlined the most common DGMs as the theoretical foundation and described the core principles of generativity and variance. Moreover, we discussed the potential applications, acknowledging the distinct generation modalities and layers along the value chain. The paradigm shift of AI applications from discriminative to generative is leading to unique use cases and promising opportunities in various domains, including those traditionally resistant to automation. Therefore, researchers and practitioners need to understand the inherent properties of generative AI to effectively leverage its potential while also mitigating associated risks. We also considered five core challenges by elaborating on bias, transparency, hallucinations, misuse, and societal impact.

Generative AI carries the potential to significantly impact various industries. Taking the different layers of the value chain into account, this progression could further lead to the creation of new platforms and services centered around whole GAI ecosystems. It remains to be seen how these forms of unprecedented artificial creativity and generativity will find a place in the industry and everyday life. Future research will need to address the challenges for safety and responsibility measures, especially when employed in highly regulated or autonomous scenarios. For example, ensuring technological transparency, increasing public trust in GAI, and developing process models for employing GAI-based systems can contribute to the body of knowledge on generative AI. We believe this article provides an entry point to this novel type of technology and guides other researcher in their efforts in examining generative AI. Closing with the initial quote by ChatGPT, generative AI indeed holds transformative the potential to redefine innovation boundaries of the digital landscapes.