1. Introduction
The recent advancements in the field of artificial intelligence (AI) opened the opportunity to delegate fractions of activities to artificial systems. Specifically, with large language models (LLMs), the generation of human language can be executed by AI systems designed for this purpose. The outcomes of these systems range from written language to images and videos. The effectiveness of the contents artificially generated by LLMs is related to their abilities to process natural language, synthesize and reason about facts, as well as problem-solve within abstract topics. AI systems that use LLMs “understand” the (contextual, however probabilistic) meaning of natural language (Surden, Reference Surden2024).
Equipped with this new set of abilities, the artificial generation of content entails delegating significant fragments of socially relevant behaviors to AI systems. Their autonomous execution of the tasks assigned implies that these systems can actively intervene in the world. They manifest a form of “behavior.” However, it is a behavior without intelligence, without understanding or cognitive abilities (Floridi, Reference Floridi2023). As “actors”, Generative AI autonomous functioning can generate relevant harms, raising the question of who will answer for algorithmic misbehavior. However, given the shared-agency context where Generative AI systems operate, which include multiple components and actors, Generative AI-related harms are not simply the result of a behavior, but rather of an interaction. To mention a few: interactions with designers, with users, with the social environment where the AI is supposed to operate.
When stating that the harm is the result of an interaction, we are not simply looking at the sum of multiple behaviors, but on the qualities of the relations that keep the behaviors interacting with each other. These interactions can be malicious if they are purposely designed or actualized to cause intentional harm. Instead, the interactions can be dangerous if they are designed or actualized lawfully, but they remain capable of causing unintentional harm.
Both malicious and dangerous interactions, and the harms they cause, might be relevant under criminal law. This paper suggests a possible categorization of Generative AI-related harms based on the distinction between intentional and unintentional harm. It then outlines the consequences they might raise in responsibility attribution. The paper’s focus is primarily on the issues related to unintentional harm, since they create the most challenging scenario when it comes to criminal law. Can we find guilt when harms are the result of a lawful interaction gone bad? If so, can or should criminal law have a role in the set of tech policies related to Generative AI?
2. Generative AI and criminal law
In the case of Generative AI-related harms, the areas where criminal law might play a role can be categorized based on whether a human agent acted with criminal intent or not. The first area of interest includes all the interactions purposely directed to committing a crime, which this paper will label as “malicious interactions.” The second area of interest encompasses interactions that occur within a lawful context but have the potential to cause significant harm, which this paper will label as “dangerous interactions.”
2.1 Generative AI-related malicious interactions
Considering that criminal phenomena involving Generative AI systems have started to appear only in recent times, rather than verifying which specific criminal offences might apply in this context, the paper will highlight, from a criminological perspective, the categories of interactions and the areas of crime that might find, in theory, a connection with Generative AI. This approach will allow us to maintain a cross-border perspective, without needing to base the analysis on a specific criminal justice system (which will remain the next needed step in the future developments of this field of study).Footnote 1 Without presuming to be exhaustive, when there is a human agent who participates with criminal intent, the following malicious interactions with Generative AI can be outlined:
(a) Support to realize content-related criminal offences, by intentionally producing or disseminating artificially generated illegal contents.
(b) Support to realize other crimes, by aiding or facilitating criminal activities.
Generative AI can be used or designed to create harmful contents, whose creation itself, or whose dissemination might be criminalized in most criminal justice systems. The range of contents that can be created is significantly wide. It includes terroristic contents, image-based sexual abuse, child pornography, and harmful speech, which can be relevant under criminal law, for instance, in case of some categories of discrimination-based discourse, incitement to crime, defamation (Henderson et al., Reference Henderson, Hashimoto and Lemley2023; Ousidhoum et al., Reference Ousidhoum2021; Weidinger et al., Reference Weidinger2022). Taking as a case-study a well-known chat-based LLM, GPT-4, OpenAI published a report on the safety challenges presented by the model, where they described the results of their adversarial testing processes, called “red teaming.”Footnote 2 The list of harmful contents that might be generated by LLMs include: advice or encouragement for self-harm behaviors; graphic material such as erotic or violent content;Footnote 3 harassing, demeaning, hateful content; apology or incitement to terroristic activities (Sivaram et al., Reference Sivaram2023; Weidinger et al., Reference Weidinger2022).
Furthermore, fake artificially generated content is raising significant concerns in the current debate (Chesney & Citron, Reference Chesney and Citron2019), even though it can be based on a different sub-category of Generative AI, called generative adversarial networks, invented by Google researcher Ian Goodfellow (Goodfellow et al., Reference Goodfellow, Pouget-Abadie, Mirza, Xu, Warde-Farley, Ozair, Courville and Bengio2014). In this area, a primary threat regards the so-called deep nudes, namely the use of deep fakes to perpetrate crimes of non-consensual pornography (Franks & Waldman, Reference Franks and Waldman2019). Although deep nudes are undoubtedly a criminal law concern, the broader category of deep fakes outlines a more problematic area from a criminal law standpoint, since it is questionable whether and to what extent behaviors involving disinformation campaigns and propaganda should be criminalized, since such criminal policies can compromise free speech rights.
In the area of artificial content-related offences, distinguishing between private and public harms is pivotal in assessing the criminal relevance of Generative AI content. While private harms – such as individual defamation or non-consensual dissemination of intimate content – raise significant ethical and legal concerns, which grant a sounder legitimation for criminal intervention, public harm must typically involve a higher threshold of risk to society at large. For crimes tied to terrorism or organized crime, the focus is on clear and severe threats, whereas offenses related to freedom of expression or democratic integrity, like electoral interference or incitement to violence, demand a more nuanced approach. Criminalization in these cases is generally limited to actions that demonstrate a significant probability of translating danger into tangible damage, emphasizing the need for scale, coordination and clear harm to the public (Guerini, Reference Guerini2020). As Generative AI continues to evolve, establishing robust frameworks to assess both the technological misuse and its societal ramifications will be integral to balancing innovation with legal accountability.
Generative AI systems, however, enable these crimes to be committed in ways that go beyond the mere creation of false or violent content. More harmful phenomena might be realized, for instance, through manipulative schemes that amplify the severity of the harm. LLMs might become a part of online manipulation schemes (Susser et al., Reference Susser, Roessler and Nissenbaum2019), which can be realized not only through the generation of fake contents, but also through several operations (e.g., spreading of polarized contents). Generative AI systems can be used for disinformation campaigns at scale (Chesney & Citron, Reference Chesney and Citron2019; Kreps et al., Reference Kreps, McCain and Brundage2022), and, more broadly, for covert or deceptive efforts to influence the opinions of a target audience, defined as “influence operations” (Goldstein et al., Reference Goldstein, Sastry, Musser, DiResta, Gentzel and Sedova2023), able to cause manipulative effects, such as cause alteration in the electoral and democratic process, or exploit cognitive vulnerabilities of users (DiPaola & Calo, Reference DiPaola and Calo2024; Neuwirth, Reference Neuwirth2022).
This leads to the second area of intersection between criminal law and malicious interactions with Generative AI, which involves its aiding role as a supporter or facilitator of crime. Considering the GPT-4 system card previously mentioned, two different forms of aid might surface: (a) aid in the preparatory phase of a crime, through the gathering of contents useful for planning a crime (e.g., information for developing unconventional weapons) or through the gathering of instructions for finding illegal content; (b) aid in the execution phase of a crime (e.g., LLMs can be used to find alternative, purchasable chemicals used in criminal activities). A relevant area of interest might be the area of cybercrimes. LLMs can be used to realize or improve social engineering attacks (e.g., drafting phishing emails) or hacking attacks (e.g., explaining systems or code’s vulnerabilities).
The same threats are outlined by Europol in a recent report (Europol, Reference Europol2023), which states how LLMs can be used to learn about crime, ranging from how to break into a home, to cybercrime and child sexual abuse.Footnote 4 A specific concern remains the fight against terroristic attacks, since the model can be used to generally gather more information that may facilitate terrorist activities, such as terrorism financing or anonymous file sharing.
The malicious interactions with Generative AI will determine a direct criminal responsibility of the human agents who committed purposely the crime, either by designing the system to realize one of the activities mentioned or by using it to the same end. Therefore, in terms of responsibility attribution, these cases are less challenging. Although some gray areas remain – especially regarding theories of criminalization (Hörnle, Reference Hörnle, Dubber and Hörnle2015) – legal systems may vary in qualifying malicious interactions either as new, potentially aggravated, ways of committing existing criminal offenses or as new criminal behaviors that require criminalization (Miró-Llinares et al., Reference Miró-Llinares, Duvac and Toader2024).
Beyond the issue of distinguishing between harm to individual victims and harm to public interests, another issue concerns the distinction between the behaviors of creation versus distribution of content. Defining the threshold for criminally relevant behavior is essential in determining when the creation or dissemination of content crosses into punishable territory. Notably, in the European Union (EU), only the creation of child pornography is criminalized in itself (even in the case of artificial or virtual content),Footnote 5 setting a precedent where mere content creation is rarely deemed illicit unless it pertains to certain severe offenses.Footnote 6 Most other content-related offenses, such as terroristic propaganda or non-consensual pornography, require dissemination to meet the threshold of criminal liability, as these materials are not inherently illicit outside of their distribution and impact on the public sphere.
In dealing with Generative AI, criminal policies might consider, on the one hand, to extend criminalization. For instance, there could be arguments in favor of the responsibility of those actors who purposefully design and make available online Generative AI platforms that enable users to realize criminal activities (e.g., platforms to create non-consensual deep nudes). In these cases, where the creation of such platforms is not criminalized itself, the crime would not be directly realized by the providers, but they might be held responsible as accomplice in the crime committed by their users.Footnote 7
On the other hand, to avoid overcriminalization, Courts might interpret existing criminal offenses in an evolutionary manner, so as to include within their scope new forms of conduct involving the abuse or misuse of Generative AI systems. For instance, under certain circumstances, the act of creating terrorist propaganda using Generative AI could be interpreted as a criminal offense if it aligns closely with national laws defining terrorism-related crimes.Footnote 8 However, this requires a case-by-case assessment grounded in the specific offences’ wording provided by each jurisdiction, but it constitutes a more balanced approach, rather than a blanket criminalization of AI-assisted content creation. Similarly, while the use of Generative AI to gather intelligence or to support criminal activities does not constitute a standalone offense, it could be relevant as an aggravating factor, amplifying existing crime’s impact and warranting more severe penalties when technology is misused in such contexts.
In addition to defining criminal thresholds, there is indeed a growing need to address the deterrence of misuse in Generative AI. Protecting the confidentiality, integrity and availability of AI is essential (Flor, Reference Flor, Magagna and Magagna2024), but equally crucial is safeguarding public trust in the technology as it integrates more deeply into society. Generative AI, with its capacity for autonomous content creation, could become a “social actor” that influences public perception and societal norms. Therefore, misuse in the form of large-scale misinformation campaigns, recruitment by terroristic organizations, or criminal campaigns aimed at democratic disruption warrants close scrutiny. The threshold here shifts to focus not only on the quantity of potential harm but also on its nature, which might contribute to the emerging of new interests, such as the public interest to protect the trustworthiness of the interactions with Generative AI.
2.2 Generative AI-related dangerous interactions
The picture becomes very different when considering dangerous interactions and the positions of the different actors involved in the supply chain of a Generative AI system that is lawfully put online or on the market. The umbrella term “operator,” keeping as a reference the definitions provided by the EU AI Act (AIA), includes different categories of actors. Namely, operators are considered as follows: “providers,” those that develop, place on the market and put into service an AI system; “deployers,” those that use an AI system under their authority in a professional activity; “distributors,” all the other remaining actors that participate in the supply chain of an AI system and make it available on the market.Footnote 9
From the perspective of these actors, Generative AI-related harms are unintentional, and the interactions are no longer considered malicious, but dangerous. In other words, they present the probability of the occurrence of severe harms.Footnote 10 Unintentional harms awaken the preventive function of criminal law. The legal response to unintentional harm does not end in punishing the agent that caused the harm, but it widens to providing the means to avoid the harm, on whose violation the responsibility attribution is based. However, these are unstable grounds for criminal law. The need to reassure against the insecurity created by the risks of harm might create a rise in punitiveness and an expansion of the penal state, which clashes with the legitimacy of criminal law as a modern, liberal institution (Ashworth, Reference Ashworth2013; Carvalho, Reference Carvalho2017; Donini & Pavarini, Reference Donini and Pavarini2011). Therefore, we need to stay on the fine line between avoiding broad new criminalization to prevent unintentional harms and filling gaps in the attribution of responsibility for those harms.
In their hybrid forms, theories of punishment combine retributivism and prevention. However, in the field of unintentional harm, the risk is to lean too much on prevention, where the goal shifts from avoiding the commission of crime to assuring social order. Security politics have problematically promoted in the last decades the framing of crime as a social risk and criminal law as a means of risk management (Ashworth & Zedner, Reference Ashworth and Zedner2012; Ashworth, et. al., Reference Ashworth, Zedner and Tomlin2013; Husak, Reference Husak2007). This approach, while influential, encounters significant challenges, particularly when extended into the realm of technology. The modern application of this model of punishment threats overextending its scope, as it must contend with the compounded uncertainties of complex social systems and the rapid evolution of technology as a social reality.
Sociological studies already outlined the limits of the paradigm of “control” when it comes to technology. The responsibility gaps created by technology developments are the result of an old issue, defined as the “dilemma of control,” that society faces when attempting to predict the social consequences of a technology (Collingridge, Reference Collingridge1980). The control of technology has proven to be a very difficult task because of its unanticipated social consequences. Narrowing our focus from the broadly harmful social consequences that a technology might have on society to the consequences that can be qualified as specific harms to a person or to a public interest, the dilemma of control might be addressed using a risk-based approach.
Therefore, assessing the necessity of criminal policies targeting AI-related harms depends on the existence of an objective and collective interest to follow this path (Vassalli, Reference Vassalli1942), which might be, as highlighted, the trustworthiness of AI as a human value. Such policies should require a shift beyond general or negative prevention toward positive prevention, focusing on restoration rather than mere deterrence. Here, culpability emerges not simply as a matter of blame (Kelly, Reference Kelly2018) but as a normative construct which can bring to restorative measures, suggesting a framework for criminal law that emphasizes social co-responsibility in the age of AI.
To establish whether such involvement of criminal policies might find a legitimate space, the objective nature, scope and severeness of AI-related harms must be firstly defined, especially when there are unintentional harms. Although a current analysis must contend with the phenomenon’s young age, that is, the fact that Generative AI systems have only recently been released into society, a framework of dangerous interactions can be suggested.
The paper approximately summarizes dangerous interactions into three main groups:
(a) Risks of harms related to users’ misuse.
(b) Risks of harms related to artificial emergence.
(c) Risks of harms related to overreliance.
The first group corresponds to the misuses listed in the previous paragraph, which become risks from the Generative AI operators’ perspective, whereas the last two groups create the more complex scenarios, where Generative AI-related harms cannot be traced back to a human agent who interacted with criminal intent. Society might face, on the one hand, harms directly caused by the Generative AI system without any user behind it (harmful AI content), and, on the other hand, harms that the Generative AI system contributed to causing with the participation of a human agent who acted without criminal intent (negligent AI content).
Harmful AI contents might be generated by emergent interactions, where “emergence” states the ability of AI to perform unpredictable behaviors (Calo, Reference Calo2015, p. 538). For instance, harmful content might be the result of the hallucinations of LLMs (Gehman, Reference Gehman, Cohn, He and Liu2020). There have already been cases where the LLMs fabricated offensive false claims against a person, falsely accusing him of a crime (Volokh, Reference Volokh2023), or cases where LLMs “degenerate into offensive language from seemingly innocuous prompts” (Weidinger et al., Reference Weidinger2022, p. 216). Furthermore, in cases of Generative AI implemented beyond the task of content-generation, new areas of risks might rise when the model is set to “accomplish goals which may not have been concretely specified or appeared in the training” (GPT-4 System Card, 2023, p. 15).
Closer to the present might be cases of overreliance of users on negligent AI content, where the harm is the result of a dangerous interaction between the Generative AI and an innocent or negligent user. Since Generative AI systems might provide part of the knowledge used in decision-making processes, if it provides manipulative information, it might encourage people to commit self-harm or it might violate their self-determination (Atillah, Reference Atillah2023, March 31; Perrigo, Reference Perrigo2023, February 17), whereas if it provides neutral, but false, misleading, or poor-quality information to a user who acts consequently based on that information, the result of the interaction might cause harm to the user herself or to others (Bambauer, Reference Bambauer2023).Footnote 11 For instance, Generative AI is used to design proteins and inform patients in the healthcare sector (AlQuraishi, Reference AlQuraishi2021; Chakraborty et al., Reference Chakraborty, Hrithik, Sayani, Saroj Kumar, Ankit, Kamred Udham and Mohd Asif2022), which might cause injuries due to the health professionals’ overreliance on the AI outputs.
Each dangerous interaction might raise different modes of responsibility attribution. In scenario (a), where there is an interaction with a malicious user, the possibility to resort to doctrines of complicity (Stewart, Reference Stewart, Dubber and Hörnle2015) might become relevant, holding the operators of the Generative AI responsible for participating in the users’ crimes, whether there is an omissive contribution (i.e., failure to avoid the user’s crime) or an active contribution (e.g., development or putting into service of the Generative AI through which the user committed the crime).Footnote 12
In scenarios (b) and (c), where malicious users’ interactions lack, the operator might be directly responsible for the harm realized, whether there is an omissive interaction (i.e., failure to avoid the harm) or an active interaction (e.g., development or putting into service of the Generative AI which caused the harm). However, all these cases reveal several shortcomings, as the next sections explore.
2.2.1 Omissive interactions
In case of omissive contributions or interactions, the operators cannot be held responsible unless the legal systems provide for duties to act which hold them responsible for preventing harms from happening (either in the case of user misuses or emergence or overreliance). It is questionable whether and which operators might have a duty to protect or a duty to control over a source of danger because of their activity.Footnote 13 Society cannot expect the operators to act to avoid every harm that might be related to the Generative AI system that they develop, place online or deploy. This scenario would hardly comply with a socially expected level of care or diligence. Moreover, in most cases, considering the context of autonomous and active interactions from which the harms originate, the operators might not have foreseen the harm, and they might not have the material possibility to intervene to avoid the harm from occurring. A further argument for distancing from this hypothesis is that omission liability is controversial in many criminal justice systems, given the moral distinction between acts and omissions and the downsides that the commission-by-omission model presents in complying with a strict legal certainty standard (Ambos, Reference Ambos, Ambos, Duff, Roberts, Weigend and Heinze2020).
To follow this path, policies would need to distinguish the positions of the different operators involved (providers, deployers, distributors) and their normatively expected capability to act to avoid AI Generative-related harms. However, to date, neither EU law nor US law regulates positive obligations that might be associated with a duty to act of the Generative AI operators.
Considering the EU context, the AIA qualifies Generative AI within the broader category of general-purpose AI (GPAI),Footnote 14 which is regulated by a specific chapter of the Act (Chapter V). According to Article 53, the legal obligations for GPAI providers mostly entail ensuring detailed technical and informational documentation. In case of GPAI that pose a “systemic risk,” namely if the GPAI has high-impact capabilities (automatically presumed if the model’s training involves more than 10^25 floating-point operations), further obligations must be carried out by the providers. According to article 55 AIA, they must conduct model evaluations, including adversarial testing (red teaming), assess and mitigate systemic risks, document and report incidents to the AI Office, and ensure adequate cybersecurity protection. Generative AI can also fall within the category of “certain AI system,” whose providers are subject to transparency obligations under article 50 of the AIA. However, neither of these obligations entails a duty to act.Footnote 15
The scenario might have been different if the set of obligations required for high-risk AI systems would have been extended to GPAI. Here, an interesting concept is “reasonably foreseeable misuse,” namely “the use of an AI system in a way that is not in accordance with its intended purpose, but which may result from reasonably foreseeable human behavior or interaction with other systems, including other AI systems” (Article 3 n. 13 AIA). The risk of reasonably foreseeable misuse is included in the set of risks that providers of high-risk AI systems must address through the implementation of a risk management system (Article 9 AIA). However, these obligations do not apply to GPAI (except, possibly, when they are part of the AI value chain and can be integrated as components within downstream high-risk AI systems).
A different option might be resorting to another EU regulation, the Digital Service Act (DSA), which regulates intermediaries’ liability. Since the regulation imposes a set of obligations to ensure that intermediaries remove illegal content from their services, the normative behaviors expected of Generative AI providers could be expanded to include those obligations as well. In the context of Generative AI, content moderation obligations and notice-and-take-down procedures might have significant benefits. Nevertheless, it is not clear yet whether Generative AI might fall within the scope of the DSA, depending on whether Generative AI platforms might be qualified as “hosting providers” or “search engines” (Lemoine & Vermeulen, Reference Lemoine and Vermeulen2023).
In the US, Generative AI operators’ liability will have to deal with the debate on Section 230 of the Communications Act and the immunity regime that it provides. Generative AI-related harms might become a new problematic chapter in the extensive use of the defense provided by Section 230 (Bambauer, Reference Bambauer2023; Henderson et al., Reference Henderson, Hashimoto and Lemley2023).
An additional drawback pertains to causation. Establishing a causal link between an operator’s conduct and the harm caused by Generative AI-driven systems presents significant challenges, even when a duty to act and legal obligations can be identified (Beck & Gerndt, Reference Beck and Gerndt2023). The limitations become more evident when using the conditio-sine-qua-non formula and the doctrine of objective attribution. However, beyond the adoption of one model of causation over another, all theories encounter the difficulty of identifying a single person who can be held responsible. The many-hands-problem (Nissenbaum, Reference Nissenbaum1996; Santoni de Sio & Mecacci, Reference Santoni de Sio and Mecacci2021; Thompson, Reference Thompson1980), wherein responsibility is distributed among numerous agents, combined with the “black box” nature of AI, complicates accountability. Specifically, it becomes difficult to determine if the operator fulfilled their obligations in a way that could have prevented a harmful outcome. In this context, the capacity to avoid harm is constrained not only by the opacity of the AI system but also by the complex and shared-agency environment in which decisions are made, where control over outcomes is dispersed among various actors. This fragmentation of agency limits any one individual’s ability to foresee and prevent adverse effects.
2.2.2 Active interactions
Generative AI operators’ responsibility for active contributions or interactions also must face the problem of causation. The harms that result from all three dangerous interactions could be difficult to trace back to a single and specific operator’s behavior, because of the “failures of causation” that characterize the AI context (Bathaee, Reference Bathaee2018; Giannini, Reference Giannini2023). For instance, the action of a health operator using the information provided by a Generative AI system in deciding a treatment that might result wrong and injure the patient’s health is just a fraction of a network of interactions between multiple human and artificial agents.
Considering the role of providers, this path might also lead to overcriminalization of “neutral behaviors,” such as the putting into service of a Generative AI system used to commit crime (Fiorinelli, Reference Fiorinelli2023). The merely material and objective contribution of making available the technical means to the realization of a user’s crime cannot alone legitimize a judgment of blameworthiness.
Moreover, the relevant behaviors might be different depending on the specific Generative AI systems deployed as services or products. An approximate distinction can be drawn between open-source Generative AI models, namely models fully public which can be downloaded and used by malicious actors, and Generative AI platforms, such as chat-based Generative AI. In the first case, harms caused by users’ misuses overstep the designers’ ability to monitor them, whereas in the second case they stay under the operator’s monitoring ability.
In cases of harms related to emergence or overreliance, a direct responsibility of the operator for an active interaction in causing the harm has the same drawbacks. Even if the operator interacted recklessly, it will be highly challenging to ascertain a causality chain between the harm and the operator’s behaviors. However, despite the limits related to causation, the main challenge that characterizes all the hypotheses of responsibility for unintentional harm concerns the problem of guilt.
3. Generative AI and guilt: Walls or doors?
Generative AI-related harms stem from a system dynamic, driven by interactions among various actors. The combination of interactions challenges the paradigm of control by a single actor over the whole actualization of the system’s functioning, as the many-hands-problem outlines. Since the liberal model of criminal law has individual autonomy at its core, when control is challenged, guilt is also challenged.
The field of Generative AI and criminal law, outside of the scope of malicious interactions, needs to address this challenge. The requirement of guilt can be seen in a limiting and confining function (as a “wall”), which is used to claim that where guilt cannot be found, neither criminal responsibility can be attributed. However, in a changing social context, such as the one where artificial and human agents interact with each other, guilt can also be an adapting concept (becoming a “door” instead). On the one hand, in the first perspective, guilt in the context of Generative AI confines criminal law to punishing only intentional malicious interactions (either as uses, developments or deployments of Generative AI), whereas the area of dangerous interactions is left to other branches of law, such as tort law. On the other hand, in the second perspective, if certain dangerous interactions create a too high and diffuse danger to individuals and society, the question of which kind of guilt can be found in those interactions might open new areas for criminal law.
Instead of looking at the issue from two different perspectives (one negative and one affirmative), the field of Generative AI and criminal law might be pictured as a set of closed doors.
3.1 The natural person behind the AI: Closed door
The traditional allocation of guilt in facing AI-related harms is problematic. Scholars analyzing criminal culpability in the AI field, although mostly focused on robotics (Consulich, Reference Consulich2022; Gless et al., Reference Gless, Silverman and Weigend2016; Piergallini, Reference Piergallini2020; Simmler & Markwalder, Reference Simmler and Markwalder2019), outline several collision points related to multiple factors, such as the distance between human behaviors and their results, the failures of causation and the challenged foreseeability of harm. The control of the users and of the operators who stand behind the technology is limited and compromised by the shared-agency dimension. That is if we follow the traditional perspective that devises responsibility attribution starting from the harm instead of from the interactions.
These collision points are not a prerogative of the field of AI-related harms. They are linked to a well-known phenomenon. The use of preventive criminal law to mitigate the risks of our contemporary and complex society is a feature of a longstanding debate related to whether and to what extent criminal law can be used to regulate the harms generated by the risk society (Stortoni & Foffani, Reference Stortoni and Foffani2004), which is now the algorithmic society. The analysis of the broader evolution of criminal law in the risk society provides a lens through which we can examine the inadequacy of the traditional punitive paradigm when applied to risks associated with Generative AI. Specifically, the conventional model of person-action-harm fails to adequately address the complexities inherent in situations where harms arise from systemic processes rather than actions of individual human agents. In other words, assigning direct personal responsibility to a singular human agent for harms resulting from systemic risks poses significant challenges, particularly concerning questions of culpability. Therefore, to regulate the harms related to the risks of the algorithmic society, it is not enough to consider the behavior of the single human agent alone. In these cases, guilt will likely close the door, limiting criminal law only to the punishment of intentional criminal behavior.
Alternatively, it may be necessary to analyze the roots and conditions of the interaction. By disentangling the behaviors of Generative AI operators from the interactions between AI systems and users, scholars (Consulich, Reference Consulich2022; Fiorinelli, Reference Fiorinelli2023; Fragrasso, Reference Fragrasso2024) suggest the possibility of employing criminal law to address novel forms of omission, such as intentional or reckless omission to comply with prescribed safety measures (i.e., failure to comply with specific obligations, such as non-delegable duties, transparency obligations or failure to provide the competent authorities with the required information). However, in addition to the concerns that arise from the use of preventive criminal law, this approach might have collateral effects, disincentivizing progress and investment, as operators might restrict their activities out of fear of facing severe punishments. It would also need to coordinate with the administrative sanctions that might already be regulated in case of violation of legal obligations, for instance by EU law.Footnote 16
Furthermore, the scope and content of the standard of due care in the field of Generative AI have not yet been established, which will depend on the set of safety international standards required by the industry (Ebers, Reference Ebers, DiMatteo, Poncibò and Cannarsa2022), since best practices will likely be insufficient. Resorting only to self-regulation does not appear to be a sufficient alternative, since delegating altogether the process of selecting the policies and measures for harm prevention to the private sector might rise concerns related to legal certainty. Moreover, certain measures (e.g., labeling unlawful content into pre-established categories) might be more efficient if based on an active collaboration with the public authorities.
3.2 Generative AI and criminal law: Whether to open the door
While the door is open for using criminal law to address malicious behavior, this paper has highlighted various setbacks to using criminal law for dangerous behavior. Given this scenario, the question then becomes whether the context of dangerous interactions might need criminal law.
Excluding criminal law completely from the set of policies related to dangerous interactions is an inadequate long-term strategy. Among the multiple theories on the purposes of criminal law (Hörnle, Reference Hörnle, Dubber and Hörnle2015), when viewed through the lens of the harm principle or the theory of the legal goods (Rechtsgüterschutz),Footnote 17 the paper outlined that it becomes essential to assess the nature of harms and risks inherent in regulating Generative AI. These include not only the harms originating from the diffusion of harmful and illicit contents artificially generated by malicious users, but they have a broader scope, which encompass the risks related to the delegation of content-generation to artificial agents in the information society. The ability to generate information in our contemporary society can have significant effects not only on individuals but also on society. This creates a base of legitimacy for criminal law, if limited only to the most serious wrongdoings and the most serious harms (in compliance with the ultima ratio principle). As the paper attempted to demonstrate, engaging in information-generating activities with artificial agents in a shared-agency context can lead to the emergence of social conflicts, manifested in the form of dangerous interactions, which include (it is worth remembering) also the risk of malicious interactions. The normative regulation and protection of a sector of social life (De Francesco, Reference De Francesco2004) highly relevant as the generation of information may ultimately necessitate criminal protection involvement in the long term. However, the condition for such involvement is that social conflicts translate into substantial wrongdoings. The question at issue then becomes whether Generative AI deployments can indeed result in substantial wrongdoing. Although this possibility exists, given that Generative AI opens the possibility for causing widespread and diffuse damage, as well as for facilitating the realization of criminal activities by simplifying their preparation or execution, it is premature to provide a definitive answer. Nevertheless, the question must be posed, and it will require finding an answer by examining the evolution of the sector and the framework of policies aimed at regulating it with an open-ended perspective, which might resort to the blurring boundaries between criminal punishment and administrative or regulatory offences (Duff et al., Reference Duff2010;Fletcher, Reference Fletcher2008).
At this stage, the paper suggests that, in the context of Generative AI and dangerous interactions, following a completely non-criminal approach, which involves civil litigation and administrative regulation, may prove insufficient in the long run. This approach would narrow the legal response to the perspective of harm mitigation, whereas criminal law aims to establish an institutional normative order aimed at protecting social values. The function of censuring and discouraging wrongdoing, which implies “should not do,” entails state intervention into individual freedoms that extends beyond providing a remedy to the victim for the invasion of protected interests (Ashworth, Reference Ashworth2013). The punitive and preventive functions, therefore, require that the distribution of responsibility in cases of serious harms relevant under criminal law follows its stricter standards, both substantively and procedurally, while granting certain protections for the defendant. In the case of AI Generative-related risks, especially in the case of emergence and overreliance, a non-criminal approach may lead to a situation where a crime occurs, and criminal law falls short because it has become challenging to find the guilty mind. However, the result of interactions remains a crime, and possibly, when behaviors become interactions within a system, a “guilty system” can be identified. According to this perspective, the attribution of responsibility must not address a gap but rather a shift (Simmler, Reference Simmler2023).
In addition, there are pragmatic considerations to ponder. Generative AI holds the potential to either facilitate or contribute to relevant harms, which may manifest with increased rapidity upon deployment of artificial agents, challenging investigation activities. Relying solely on retrospective measures to prosecute malicious uses may prove ineffective or unsustainable. The prospect of intervening during stages that precede interactions (e.g., design or online availability) or during interactions themselves (e.g., deployment) presents an opportunity to address social conflicts as they arise, preempting their escalation into broader and diffuse harms. Involvement during these phases could equip criminal law with valuable tools for regulating conflicts. For instance, civil courts lack the authority to order defendants to perform positive acts to avoid or to mitigate punishment, such as implementing reactive compliance efforts. Lastly, given the diffuse nature of risks associated with Generative AI, it prompts inquiry into the adequacy of entrusting the legal response to harms solely to private citizen initiatives, without a structured intervention of the State, especially in cases where there is no victim, and the harm caused is against collective and public interests.
4. Conclusion
Given the possibility of integrating criminal law into the regulation of AI Generative-related interactions, the fundamental question is which policy would best justify criminal coercion. The answer to this question is contingent upon both the theory of attribution and the nature of the punishment that will be selected.
In cases of malicious interactions, the preferred approach should assess whether, and to what extent, existing criminal offenses might apply, with consideration of whether the increased harm caused by the facilitator role of Generative AI may warrant an enhanced punishment. Simultaneously, the evolving landscape of criminal misuse of Generative AI should be monitored to determine whether there are actual gaps in criminal legislation, starting from the harms against a specific victim rather than harms against public interests.Footnote 18
In case of dangerous interactions, instead of dissembling the interactions of the shared-agency context, a way to stay within them is to follow one of the suggestions raised in the literature (Diamantis, Reference Diamantis2021; Mongillo, Reference Mongillo2023), namely, to resort to corporate criminal responsibility instead of human agents’ criminal responsibility in the context of AI. Expanding or amending corporate criminal law might be a possibility to further explore,Footnote 19 in cases where a relevant harm occurs due to the intentional or reckless noncompliance with the standard of care required of the corporation which interacts in the relevant processes of proving and deploying the Generative AI system, and which benefits from its activities. The deterrence and prevention raised from a possible affliction to the economic interests of companies might help in enhancing corporate collaboration in the AI industry, without being solely driven by market pressure (Askell et al., Reference Askell, Brundage and Hadfield2019).
In these instances, legislators may need to ponder whether to open the door of criminal law or not, after assessing the precise nature of harms potentially engendered by Generative AI, and, especially, their severeness from both a qualitative and quantitative point of view. Additionally, a prerequisite for pursuing this course of action involves determining how to preserve culpability in the context of AI-related harms, firstly, by underscoring the possibility to devise frameworks for accountability within shared-agency context, where new socio-digital institutions are emerging (Beckers & Teubner, Reference Beckers and Teubner2021), secondly, by assessing the scope and content of non-criminal regulatory measures as they define a standard of care in the field. Until then, we face a wall with many closed doors. But the doors exist. Whether and which one we want to leave closed or to start opening remains a political choice.
Funding statement
None to declare.
Competing interests
None to declare.