Abstract

Objective

To investigate how the general public trades off explainability versus accuracy of artificial intelligence (AI) systems and whether this differs between healthcare and non-healthcare scenarios.

Materials and Methods

Citizens’ juries are a form of deliberative democracy eliciting informed judgment from a representative sample of the general public around policy questions. We organized two 5-day citizens’ juries in the UK with 18 jurors each. Jurors considered 3 AI systems with different levels of accuracy and explainability in 2 healthcare and 2 non-healthcare scenarios. Per scenario, jurors voted for their preferred system; votes were analyzed descriptively. Qualitative data on considerations behind their preferences included transcribed audio-recordings of plenary sessions, observational field notes, outputs from small group work and free-text comments accompanying jurors’ votes; qualitative data were analyzed thematically by scenario, per and across AI systems.

Results

In healthcare scenarios, jurors favored accuracy over explainability, whereas in non-healthcare contexts they either valued explainability equally to, or more than, accuracy. Jurors’ considerations in favor of accuracy regarded the impact of decisions on individuals and society, and the potential to increase efficiency of services. Reasons for emphasizing explainability included increased opportunities for individuals and society to learn and improve future prospects and enhanced ability for humans to identify and resolve system biases.

Conclusion

Citizens may value explainability of AI systems in healthcare less than in non-healthcare domains and less than often assumed by professionals, especially when weighed against system accuracy. The public should therefore be actively consulted when developing policy on AI explainability.

INTRODUCTION

The use of artificial Intelligence (AI) is on the rise in healthcare as well as other domains, such as agriculture and finance.1 Modern AI systems typically rely on machine learning, a technology that automatically constructs computer systems through real-world experiences captured in large volumes of data.2 Deep learning is at the forefront of this development: a 2019 meta-analysis of 25 studies of AI diagnostic systems in medical imaging and histopathology concluded that the diagnostic performance of deep learning models was equivalent to that of healthcare professionals.3 This may explain why some people anticipate that AI systems will create new roles for doctors as information specialists, or even replace entire medical disciplines.4 However, due to their complex internal structure, deep learning models are generally considered “black boxes” that do not providing information about how exactly they arrive at their decisions.

Some feel uncomfortable with black box AI. They argue that for sensitive decision-making tasks affecting human well-being and health, it is not acceptable to use AI systems that are not transparent and whose reasoning cannot be understood by those affected by the decisions.5 This not only undermines trust but also reduces possibilities for verifying that the reasoning methods are sound and robust and that the resulting decisions are safe and fair.6

However, more explainable AI systems may be less accurate and vice versa.7 Viewed as such, AI explainability becomes a double-edged sword: while more explainability enhances the opportunity to verify and contest a decision, it would also increase the probability of error. The person affected by that decision may therefore not necessarily consider increased explainability a good thing. To date, no studies have explored how people would make that trade-off between accuracy and explainability of AI systems used to make decisions about them.

Currently, there is no clear guidance on whether AI-based decisions should be accompanied by an explanation. There are position statements emphasizing the importance of explainability (or transparency) of AI decision-making,8 and the “right to explanation” is also a subject of ongoing legal debate. Under the United States Equal Credit Opportunity Act, creditors are required to provide specific reasons when denying someone a loan.9 The European Union General Data Protection Regulation, Recital 71, states that when people are subject to decisions based solely on automated processing, they have a right to obtain an explanation of the decision reached.10 But position statements and recitals are not binding, and the right to an explanation has been removed from the binding articles of the text during the legislative process. This has resulted in a scarcity of policy on if and how developers, users, and regulators should ensure AI explainability.

Therefore, we organized 2 citizens’ juries, a method to explore what members of the public think about a policy problem after they have been well informed. Our juries addressed the question: “Should AI systems give an explanation, even if that results in less accurate decisions?” The juries aimed to inform guidance on explaining processes, services, and decisions delivered or assisted by AI to the individuals affected by them. This article reports on the mixed-methods study conducted alongside the juries, which investigated how the public makes the trade-off between accuracy and explainability of AI, their reasoning behind this, and whether this differed between healthcare and other domains.

MATERIALS AND METHODS

Citizens’ juries to engage the public in policy making

Developed in the 1970s by the Jefferson Center,11 citizens’ juries are a form of deliberative democracy: an egalitarian approach to public policy making that encourages mutual recognition and respect and allows public negotiation for the common good.12 In a citizens’ jury, a representative sample of individual citizens (ie, jurors)—from different backgrounds and without having special prior knowledge or expertise—come together for several days, hear expert evidence, deliberate together and reach reasoned conclusions about the public policy questions posed to them. This approach assumes that jurors can answer these questions once they are properly informed about available evidence and are encouraged to deliberate in an environment free of “delusion, deception, power, and strategy.”13,14 Citizens’ juries are particularly effective when there is a values trade-off at the heart of the questions they are being asked. As such, they have been used for informing health policy on ethically complex topics, such as genetic testing,15 screening services,16 and case finding for stigmatized diseases.17 Our research group previously organized 3citizens’ juries on health data sharing for research and commercialuse.18–20

General approach

In February 2019, we organized 2 citizens’ juries in the UK, 1 in Manchester and 1 in Coventry. Each lasted 5 days and involved 18 jurors, without overlap in participants between the juries. This sample size was convenient for smaller and larger group work (eg, 6 groups of 3 and 3 groups of 6) and aimed to enable all jurors to actively take part in the discussions. Both juries followed an identical process, allowing us to systematically compare results between the 2 juries to strengthen the robustness of our findings. The running of the jury was contracted to Citizens’ Juries c.i.c.,21 a social enterprise that managed the project, designed and facilitated the jury process in partnership with the Jefferson Center, and recruited jurors, expert witnesses, and an independent oversight panel. The 3 oversight panel members were chosen for their topic knowledge and lack of conflict of interest in a particular jury outcome. They reviewed all jury materials (scenarios, witness briefs, presentations, questionnaires, planned activities) to ensure scientific rigor and identify bias in the jury design toward either accuracy or explainability. To further minimize risk of bias in our findings, we strictly separated roles within the study team: those who were involved in the design and conduct of the citizens’ juries (NP, KB, SA, MT, MO) were not involved in the qualitative data collection, analysis, and interpretation (LR, SCS, DP, SvdV).

The study was approved by The University of Manchester Research Ethics Committee (Ref: 2019-6023-9065).

Recruitment

We invited citizens to apply as a juror via adverts on a job website (www.indeed.co.uk) for both cities as well as a volunteering website (https://www.vacoventry.org.uk) in Coventry. For each jury, we created a shortlist of 18 applicants through stratified sampling, ensuring that the shortlist matched the latest UK Census Data for England22 in terms of age, gender, ethnicity, educational attainment, and employment status. We also aimed to match their prior views on the topic to the responses of 2074 adults participating in a survey commissioned by the Royal Society for the encouragement of Arts, Manufactures, and Commerce23 to the question “How comfortable, if at all, are you with the following idea? As the accuracy and consistency of automated systems improve over time, more decisions can be fully automated without human intervention required.

Shortlisted applicants were interviewed by telephone to ensure that they: (i) did not have substantial prior knowledge or expertise on the jury topic, or a special or conflicting interest; (ii) understood what was being asked of them; and (iii) were available in the jury week. Participants received £500 for their 5 days “jury service” plus £25 cash expenses.

Jury program

Based on the jury questions, Citizens’ Juries c.i.c. determined jurors’ information needs and developed the program accordingly. This included a brief for expert witnesses, which guided witness selection. The witness brief and slides are publicly available, alongside all other jury materials and documentation.24 Box 1 shows the 5-day jury program, which was a mix of plenary sessions and small group work (see Supplementary Part 1 for detailed jury program). The plenary sessions primarily consisted of presentations by expert witnesses followed by questions from the jurors and of whole group feedback of the small group work. Some expert witnesses argued for a specific position (ie, partial witnesses), while others gave a neutral, balanced account of the topic (ie, impartial witnesses). Witnesses either attended the jury in person or appeared via video link. In both juries, jurors were given the same materials, heard the same presentations, and could ask questions of all expert witnesses. Both juries were led by the same 2 independent facilitators (KB, SA) from the Jefferson Center.

On days 3 and 4, we presented the jurors with 4 scenarios involving applications of AI, 2 of which were related to healthcare (stroke and kidney transplant scenarios) and 2 that were not (recruitment and criminal justice scenarios; see Box 2).

The scenarios were developed by Citizens’ Juries c.i.c. with input from the organizations who commissioned the juries (see Funding statement). The iterative development process was guided by the role of the scenarios within the program, which was to enable jurors to explore the questions and trade-offs of interest. Scenarios reflected realistic settings where AI could be deployed. In each scenario, we described 3 types of automated decision-making systems (systems A, B, and C) with different levels of explainability and accuracy (Box 3). We chose the accuracy levels of these systems so that they offered clear and consistent choices across the scenarios, rather than necessarily reflecting real-world AI performance levels (Supplementary Part 2 presents the complete scenarios). Per scenario, expert witnesses explained the scenario and the systems, after which jurors discussed the (dis)advantages of the systems in small groups and noted these on flip charts. All jurors then indicated on the flipcharts which (dis)advantages they found most important for each system, and anonymously selected their preferred system through electronic voting, including up to 3 reasons in free text to support their preference.

Data collection and analysis

We collected quantitative and qualitative data for analysis:

  • Transcriptions: as part of documenting the jury process, we audio-recorded and transcribed plenary sessions to capture questions raised by jurors, and arranged them by scenario and topic.

  • Electronic voting data: anonymized jurors’ votes by scenario, and up to 3 reasons, recorded as free text, to support their choice.

  • Observational field notes: qualitative researchers (LR, DP, SCS) observed the plenary sessions and group discussions, and captured their observations as field notes.

  • Output from the group discussions on (dis)advantages of each system per scenario: this included jurors’ statements agreed during the small group work, and for each statement how many jurors considered it the most important factor for (not) preferring a particular system.

To investigate how jurors made the trade-off between accuracy and explainability of AI, and whether this differed between healthcare and non-healthcare scenarios, we descriptively summarized the quantitative voting results across the 2 juries per scenario. To understand the reasoning behind their trade-off, we qualitatively analyzed jurors’ free text voting data, researchers’ field notes, and jurors’ statements on (dis)advantages of the systems. For this, LR imported the data into NVivo (QSR International Pty Ltd Version 12, 2018) for management of initial coding and read the data as part of the familiarization process. LR conducted an initial round of conceptual coding using an inductive open coding approach26 and organized codes into themes.27 SCS independently coded a random 10% sample of the data. The qualitative team (LR, SCS, DP) discussed the results from this consensus exercise, resolving coding differences by majority decision. LR then recoded the data in a second coding round according to the agreed revised coding scheme. The qualitative team met to discuss grouping codes into overarching themes by scenario, both per and across AI systems. These themes were iteratively verified and refined a final time through discussions between LR and NP (Supplementary Part 3 presents an example of how we went from initial coding to theme generation).

RESULTS

Jurors

In total, 451 people applied for the juries (271 in Manchester, 180 in Coventry). Thirty-six were selected as jurors (18 per location), of whom most (89%) were recruited via the job website. Table 1 presents juror characteristics, which were largely comparable to those of the English population.

Table 1.

Juror characteristics compared to the population of England22

CharacteristicsJurors (n = 36)England
Value (%)a%
Age (years)
 18–298 (22)21
 30–449 (25)26
 45–599 (25)25
 60 or olderb11 (31)28
Gender
 Male18 (50)49
 Female18 (50)51
Ethnicity
 White27 (75)85
 Non-white9 (25)15
Educational attainmentc
 0 to 4 O levels/GCSEs10 (28)36
 5 or more O levels/GCSEs5 (14)21
 2 or more A levels9 (25)16
 Graduate degree or higher12 (33)27
Employed or self-employed22 (61)76d
Response to pre-jury screening question on the trade-off between accuracy versus explainabilitye,f
 Not at all comfortable8 (22)26
 Not very comfortable15 (42)38
 Fairly comfortable8 (22)23
 Very comfortable2 (6)3
 Don’t know3 (8)9
Recruitment route
 Job website32 (89)n.a.
 Voluntary action website2 (6)n.a.
 Word of mouth2 (6)n.a.
CharacteristicsJurors (n = 36)England
Value (%)a%
Age (years)
 18–298 (22)21
 30–449 (25)26
 45–599 (25)25
 60 or olderb11 (31)28
Gender
 Male18 (50)49
 Female18 (50)51
Ethnicity
 White27 (75)85
 Non-white9 (25)15
Educational attainmentc
 0 to 4 O levels/GCSEs10 (28)36
 5 or more O levels/GCSEs5 (14)21
 2 or more A levels9 (25)16
 Graduate degree or higher12 (33)27
Employed or self-employed22 (61)76d
Response to pre-jury screening question on the trade-off between accuracy versus explainabilitye,f
 Not at all comfortable8 (22)26
 Not very comfortable15 (42)38
 Fairly comfortable8 (22)23
 Very comfortable2 (6)3
 Don’t know3 (8)9
Recruitment route
 Job website32 (89)n.a.
 Voluntary action website2 (6)n.a.
 Word of mouth2 (6)n.a.

Abbreviations: GCSEs, general certificate of secondary education; n.a., not applicable.

a

Percentages may not add up to 100% due to rounding.

b

People of 75 years and older were the lowest users of the internet in the UK in 2019 (47% compared to 91% overall) [28], which means they might have been underrepresented in this age group.

c“O levels” and the examinations that replaced them (GCSEs) are UK academic qualifications in a particular subject, typically taken at age 16. “A levels” are a higher level of academic attainment that give access to university, usually taken at age 18. Students typically take up to 8 GCSEs/O levels, and up to 3 A levels. Many students leave education at aged 16, especially in the 1940s to 1980s. A graduate degree is a university degree.

d% of people in the UK aged 16–64 years who were (self-)employed in the period Oct–Dec 2018.

eThe question was formulated as follows: “How comfortable, if at all, are you with the following idea? As the accuracy and consistency of automated systems improve over time, more decisions can be fully automated without human intervention required.

fPercentages for England refer to responses of 2074 adults to this question as part of a survey commissioned by the Royal Society for the encouragement of Arts, Manufactures, and Commerce [23].

Table 1.

Juror characteristics compared to the population of England22

CharacteristicsJurors (n = 36)England
Value (%)a%
Age (years)
 18–298 (22)21
 30–449 (25)26
 45–599 (25)25
 60 or olderb11 (31)28
Gender
 Male18 (50)49
 Female18 (50)51
Ethnicity
 White27 (75)85
 Non-white9 (25)15
Educational attainmentc
 0 to 4 O levels/GCSEs10 (28)36
 5 or more O levels/GCSEs5 (14)21
 2 or more A levels9 (25)16
 Graduate degree or higher12 (33)27
Employed or self-employed22 (61)76d
Response to pre-jury screening question on the trade-off between accuracy versus explainabilitye,f
 Not at all comfortable8 (22)26
 Not very comfortable15 (42)38
 Fairly comfortable8 (22)23
 Very comfortable2 (6)3
 Don’t know3 (8)9
Recruitment route
 Job website32 (89)n.a.
 Voluntary action website2 (6)n.a.
 Word of mouth2 (6)n.a.
CharacteristicsJurors (n = 36)England
Value (%)a%
Age (years)
 18–298 (22)21
 30–449 (25)26
 45–599 (25)25
 60 or olderb11 (31)28
Gender
 Male18 (50)49
 Female18 (50)51
Ethnicity
 White27 (75)85
 Non-white9 (25)15
Educational attainmentc
 0 to 4 O levels/GCSEs10 (28)36
 5 or more O levels/GCSEs5 (14)21
 2 or more A levels9 (25)16
 Graduate degree or higher12 (33)27
Employed or self-employed22 (61)76d
Response to pre-jury screening question on the trade-off between accuracy versus explainabilitye,f
 Not at all comfortable8 (22)26
 Not very comfortable15 (42)38
 Fairly comfortable8 (22)23
 Very comfortable2 (6)3
 Don’t know3 (8)9
Recruitment route
 Job website32 (89)n.a.
 Voluntary action website2 (6)n.a.
 Word of mouth2 (6)n.a.

Abbreviations: GCSEs, general certificate of secondary education; n.a., not applicable.

a

Percentages may not add up to 100% due to rounding.

b

People of 75 years and older were the lowest users of the internet in the UK in 2019 (47% compared to 91% overall) [28], which means they might have been underrepresented in this age group.

c“O levels” and the examinations that replaced them (GCSEs) are UK academic qualifications in a particular subject, typically taken at age 16. “A levels” are a higher level of academic attainment that give access to university, usually taken at age 18. Students typically take up to 8 GCSEs/O levels, and up to 3 A levels. Many students leave education at aged 16, especially in the 1940s to 1980s. A graduate degree is a university degree.

d% of people in the UK aged 16–64 years who were (self-)employed in the period Oct–Dec 2018.

eThe question was formulated as follows: “How comfortable, if at all, are you with the following idea? As the accuracy and consistency of automated systems improve over time, more decisions can be fully automated without human intervention required.

fPercentages for England refer to responses of 2074 adults to this question as part of a survey commissioned by the Royal Society for the encouragement of Arts, Manufactures, and Commerce [23].

How jurors made the trade-off between accuracy and explainability of AI

Table 2 displays the results of the quantitative analysis of the electronic voting per scenario across the 2 juries. For both healthcare scenarios, a large majority (86% and 92% for the stroke and kidney transplantation scenarios, respectively) voted for system C, implying they favored accuracy over explainability in these contexts. Jurors’ votes for related issues (ie, questions 1 and 2) also reflected this preference, although less starkly for question 1, where, for the kidney transplantation scenario, 36% of jurors found it at least fairly important for an individual to receive an explanation of an automated decision. For the non-healthcare scenarios, the majority of jurors made the trade-off differently: they either weighed accuracy and explainability equally by voting for system B or—particularly in the criminal justice scenario—put more emphasis on explanation. Voting results were comparable between the Manchester and Coventry juries, except for the recruitment scenario; whereas the Manchester jurors distributed their votes equally across the 3 systems, 14/18 of Coventry jurors voted for system B (see Supplementary Part 5 for votes per jury).

Table 2.

Electronic voting results per scenario across the 2 juries; total N = 36, values are numbers (percentages)

Healthcare Scenarios
Non-Healthcare Scenarios
QuestionsStrokeKidney TransplantationRecruitmentCriminal Justice
Question 1: How important is it for an individual to receive an explanation of an automated decision?
 Very important4 (11)1 (3)8 (22)17 (47)
 Fairly important6 (17)12 (33)16 (44)9 (25)
 Not very important21 (58)18 (50)11 (31)8 (22)
 Not at all important4 (11)5 (14)1 (3)2 (6)
 Don’t know1 (3)0 (0)0 (0)0 (0)
Question 2: If system C was chosen [by the NHS], almost no explanation would be provided. How much does this matter?
 Very much4 (11)1 (3)10 (28)17 (47)
 Quite a lot5 (14)3 (8)12 (33)9 (25)
 Not very much24 (67)20 (56)13 (36)7 (19)
 Not at all3 (8)12 (33)1 (3)3 (8)
 Don’t know0 (0)0 (0)0 (0)0 (0)
Question 3: Which automated decision system do you think the NHS should choose/should be chosen?a
 System A—expert system0 (0)1 (3)7 (19)15 (42)
 System B—random forest5 (14)2 (6)20 (56)13 (36)
 System C—deep learning31 (86)33 (92)9 (25)8 (22)
Healthcare Scenarios
Non-Healthcare Scenarios
QuestionsStrokeKidney TransplantationRecruitmentCriminal Justice
Question 1: How important is it for an individual to receive an explanation of an automated decision?
 Very important4 (11)1 (3)8 (22)17 (47)
 Fairly important6 (17)12 (33)16 (44)9 (25)
 Not very important21 (58)18 (50)11 (31)8 (22)
 Not at all important4 (11)5 (14)1 (3)2 (6)
 Don’t know1 (3)0 (0)0 (0)0 (0)
Question 2: If system C was chosen [by the NHS], almost no explanation would be provided. How much does this matter?
 Very much4 (11)1 (3)10 (28)17 (47)
 Quite a lot5 (14)3 (8)12 (33)9 (25)
 Not very much24 (67)20 (56)13 (36)7 (19)
 Not at all3 (8)12 (33)1 (3)3 (8)
 Don’t know0 (0)0 (0)0 (0)0 (0)
Question 3: Which automated decision system do you think the NHS should choose/should be chosen?a
 System A—expert system0 (0)1 (3)7 (19)15 (42)
 System B—random forest5 (14)2 (6)20 (56)13 (36)
 System C—deep learning31 (86)33 (92)9 (25)8 (22)
a

System A, expert system (below human expert-level accuracy, full explanation); System B, random forest system (human expert-level accuracy, partial explanation); System C, deep learning system (beyond human expert-level accuracy, no explanation).

Table 2.

Electronic voting results per scenario across the 2 juries; total N = 36, values are numbers (percentages)

Healthcare Scenarios
Non-Healthcare Scenarios
QuestionsStrokeKidney TransplantationRecruitmentCriminal Justice
Question 1: How important is it for an individual to receive an explanation of an automated decision?
 Very important4 (11)1 (3)8 (22)17 (47)
 Fairly important6 (17)12 (33)16 (44)9 (25)
 Not very important21 (58)18 (50)11 (31)8 (22)
 Not at all important4 (11)5 (14)1 (3)2 (6)
 Don’t know1 (3)0 (0)0 (0)0 (0)
Question 2: If system C was chosen [by the NHS], almost no explanation would be provided. How much does this matter?
 Very much4 (11)1 (3)10 (28)17 (47)
 Quite a lot5 (14)3 (8)12 (33)9 (25)
 Not very much24 (67)20 (56)13 (36)7 (19)
 Not at all3 (8)12 (33)1 (3)3 (8)
 Don’t know0 (0)0 (0)0 (0)0 (0)
Question 3: Which automated decision system do you think the NHS should choose/should be chosen?a
 System A—expert system0 (0)1 (3)7 (19)15 (42)
 System B—random forest5 (14)2 (6)20 (56)13 (36)
 System C—deep learning31 (86)33 (92)9 (25)8 (22)
Healthcare Scenarios
Non-Healthcare Scenarios
QuestionsStrokeKidney TransplantationRecruitmentCriminal Justice
Question 1: How important is it for an individual to receive an explanation of an automated decision?
 Very important4 (11)1 (3)8 (22)17 (47)
 Fairly important6 (17)12 (33)16 (44)9 (25)
 Not very important21 (58)18 (50)11 (31)8 (22)
 Not at all important4 (11)5 (14)1 (3)2 (6)
 Don’t know1 (3)0 (0)0 (0)0 (0)
Question 2: If system C was chosen [by the NHS], almost no explanation would be provided. How much does this matter?
 Very much4 (11)1 (3)10 (28)17 (47)
 Quite a lot5 (14)3 (8)12 (33)9 (25)
 Not very much24 (67)20 (56)13 (36)7 (19)
 Not at all3 (8)12 (33)1 (3)3 (8)
 Don’t know0 (0)0 (0)0 (0)0 (0)
Question 3: Which automated decision system do you think the NHS should choose/should be chosen?a
 System A—expert system0 (0)1 (3)7 (19)15 (42)
 System B—random forest5 (14)2 (6)20 (56)13 (36)
 System C—deep learning31 (86)33 (92)9 (25)8 (22)
a

System A, expert system (below human expert-level accuracy, full explanation); System B, random forest system (human expert-level accuracy, partial explanation); System C, deep learning system (beyond human expert-level accuracy, no explanation).

Jurors’ reasoning when making the trade-off between accuracy and explainability of AI

Three themes emerged from our analysis of the qualitative data. Each consisted of 3 subthemes reflecting factors that jurors considered when voting. We discuss the subthemes in more detail below, with Table 3 providing illustrative quotes from the free-text voting data, researcher field notes, and jurors’ statements on (dis)advantages of systems (see Supplementary Part 6 for more juror statements).

Table 3.

Quotes and statements to illustrate jurors’ considerations when making the trade-off between accuracy and explainability of AI systemsa,b

Healthcare Scenarios
Non-Healthcare Scenarios
(Sub)ThemesStrokeKidney TransplantationRecruitmentCriminal Justice
Theme A: ACCURACY OF DECISION MAKING
1. Impact on the individual
  • “[System C] will save lives” (M, G)

  • “We have a moral and ethical obligation to improve speed and accuracy rates in diagnosis” (C, V, 11)

  • “The explanation comes secondary to accuracy, as without an accurate diagnosis there would be no basis foundation for treatment/aftercare” (C, V, 10)

  • “Higher accuracy level so less chance for an unsuccessful transplant” (M, V, 59)

  • “When you win the lottery you just need to know that you won, not which order the balls come up” (C, G)

  • “[System C would] save valuable time for the patients, as they would not potentially be given a kidney that would be rejected and go back in the waitlist” (C, V, 10)

“[System B] has key points/features for detection.” (C, G)
  • “Accuracy of the decision is paramount, as it may determine the individual’s future” (M, V, 68)

  • “The accuracy level [of system A] is the same as a human, but quicker” (M, V, 59)

2. Impact on society“It would be able to … possibly free up more time for experts to … do more research” (C, G)“It is important to get as accurate a match as possible to eliminate waste of a kidney that could have potentially helped another [recipient]” (M, V, 65)“[System B has] high level of possibility finding correct candidate” (C, V, 14)
  • “[System C has a] lower risk to [the] public” (C, V, 20)

  • “It is important to identify who will reoffend” (M, V, 53)

3. Increased efficiency
  • “[System C could] save money and time in the long term” (M, V, 56)

  • “[System C] frees up doctors for other duties” (C, V, 16)

“[System C] could reduce costs for the NHS in respect of rejection and fewer drugs prescribed post-operative[ly]” (C, V, 21)“[System C] gets through large amounts of CVs in a short space of time” (M, V, 59)“[System A] would require a lot of work from software engineers and policemen” (C, G)
Theme B: EXPLAINABILITY OF DECISION MAKING
1. Learning opportunities for the individual
  • “According to [the evidence], most stroke victims don't usually ask how the stroke was diagnosed” (C, V, 17)

  • “More information would lead to better aftercare [and could support] prevention of repeat of stroke” (C, V, 31)

  • “Patient rights to know how the system computed” (M, V, 67)

  • “Doctors wouldn’t trust themselves when presented with a machine that is almost always right and would sometimes incorrectly follow the machine’s advice especially when the machine offers no explanation at all.” (C, V, 24).

“No transparency to show the criteria used (BMI, blood pressure, lifestyle) factored into the decision” (C, G)
  • “Less transparent feedback [in system B]—eg, could tell you that you didn’t meet essential criteria but not tell you which part you did not meet to improve future chances of success” (C, G)

  • “Gives me motivation to improve and hone my skills for the future” (C, V, 26)

  • “It could rehabilitate a person who is in real need of direction” (M, V, 51)

  • “[System A is the] best system to provide transparent reasoning for individuals where they have specific factors to focus on during rehab/prosecution” (C, V, 18)

2. Learning opportunities for society“[With system C, you] can’t interrogate the system so can’t improve the result” (M, G)n.a.c“Feedback is essential for improving results for candidates, employers, and the economy” (C, V, 16)
  • “[System A allows] clear decision-making process providing transparency for offender, victim, and any potential further victim” (C, V, 11)

  • “No learning from data [with system C]” (M, V, 71)

3. Ability to identify/resolve biasn.a.c“Data going in to this system would be hugely skewed in favour of those privileged with existing wellness … [leading to widening] health inequalities” (C, V, 16)
  • “[System B] can somehow identify biases in the attributes which can be rectified” (C, V, 18)

  • “The factors assessed by systems B and C may be entrenched in historical bias … whether someone is a high flyer precludes factors that make a great employee, like honesty and attitude” (M, V, 55)

  • “Accuracy is less important because the data [are] so subjective anyway” (C, V, 19)

  • “The custody sergeant will have experience and understanding of offenders and possibly what causes them to offend” (M, V, 51)

  • “[System C] removes individual bias made by individual officers in each respective case” (C, V, 10)

Theme C: TRUST IN AUTOMATED SYSTEMS
1. Fairness“There are not specialist doctors available 24 hours a day to reach a decision” (M V, 59)“[System C] is the fairest—it shortlists on organ suitability alone” (M, V, 55)“[System B] is ‘democratic’ in its analysis of past applicants, both successful and unsuccessful, without developing unknowable deep learning iterations” (M, G)“I don’t think something as abstract as potential to reoffend can be discerned by AI” (M, V, 55)
2. Delivery of the decisionn.a.c“[System C] doesn’t take away the sympathetic and empathetic aspect patients need throughout the whole process, but still assists doctors in decision making” (C, V, 18)n.a.c
  • “[With system A] there is more interaction with officers. Feelings and remorse will be taken into account and a full explanation will be given for the end result” (M, G)

  • “[System A] offers better dialogue between the authority and the offender with clearly defined parameters” (C, G)

3. Accountability for decisions“[With system C] who is responsible if the diagnosis is wrong with fatal consequences?” (C, G)
  • “[An explanation is] only relevant for cases that go wrong and an enquiry is launched” (C, V, 11)

  • “There is no recourse when something goes wrong” (C, V, 16)

n.a.cn.a.c
Healthcare Scenarios
Non-Healthcare Scenarios
(Sub)ThemesStrokeKidney TransplantationRecruitmentCriminal Justice
Theme A: ACCURACY OF DECISION MAKING
1. Impact on the individual
  • “[System C] will save lives” (M, G)

  • “We have a moral and ethical obligation to improve speed and accuracy rates in diagnosis” (C, V, 11)

  • “The explanation comes secondary to accuracy, as without an accurate diagnosis there would be no basis foundation for treatment/aftercare” (C, V, 10)

  • “Higher accuracy level so less chance for an unsuccessful transplant” (M, V, 59)

  • “When you win the lottery you just need to know that you won, not which order the balls come up” (C, G)

  • “[System C would] save valuable time for the patients, as they would not potentially be given a kidney that would be rejected and go back in the waitlist” (C, V, 10)

“[System B] has key points/features for detection.” (C, G)
  • “Accuracy of the decision is paramount, as it may determine the individual’s future” (M, V, 68)

  • “The accuracy level [of system A] is the same as a human, but quicker” (M, V, 59)

2. Impact on society“It would be able to … possibly free up more time for experts to … do more research” (C, G)“It is important to get as accurate a match as possible to eliminate waste of a kidney that could have potentially helped another [recipient]” (M, V, 65)“[System B has] high level of possibility finding correct candidate” (C, V, 14)
  • “[System C has a] lower risk to [the] public” (C, V, 20)

  • “It is important to identify who will reoffend” (M, V, 53)

3. Increased efficiency
  • “[System C could] save money and time in the long term” (M, V, 56)

  • “[System C] frees up doctors for other duties” (C, V, 16)

“[System C] could reduce costs for the NHS in respect of rejection and fewer drugs prescribed post-operative[ly]” (C, V, 21)“[System C] gets through large amounts of CVs in a short space of time” (M, V, 59)“[System A] would require a lot of work from software engineers and policemen” (C, G)
Theme B: EXPLAINABILITY OF DECISION MAKING
1. Learning opportunities for the individual
  • “According to [the evidence], most stroke victims don't usually ask how the stroke was diagnosed” (C, V, 17)

  • “More information would lead to better aftercare [and could support] prevention of repeat of stroke” (C, V, 31)

  • “Patient rights to know how the system computed” (M, V, 67)

  • “Doctors wouldn’t trust themselves when presented with a machine that is almost always right and would sometimes incorrectly follow the machine’s advice especially when the machine offers no explanation at all.” (C, V, 24).

“No transparency to show the criteria used (BMI, blood pressure, lifestyle) factored into the decision” (C, G)
  • “Less transparent feedback [in system B]—eg, could tell you that you didn’t meet essential criteria but not tell you which part you did not meet to improve future chances of success” (C, G)

  • “Gives me motivation to improve and hone my skills for the future” (C, V, 26)

  • “It could rehabilitate a person who is in real need of direction” (M, V, 51)

  • “[System A is the] best system to provide transparent reasoning for individuals where they have specific factors to focus on during rehab/prosecution” (C, V, 18)

2. Learning opportunities for society“[With system C, you] can’t interrogate the system so can’t improve the result” (M, G)n.a.c“Feedback is essential for improving results for candidates, employers, and the economy” (C, V, 16)
  • “[System A allows] clear decision-making process providing transparency for offender, victim, and any potential further victim” (C, V, 11)

  • “No learning from data [with system C]” (M, V, 71)

3. Ability to identify/resolve biasn.a.c“Data going in to this system would be hugely skewed in favour of those privileged with existing wellness … [leading to widening] health inequalities” (C, V, 16)
  • “[System B] can somehow identify biases in the attributes which can be rectified” (C, V, 18)

  • “The factors assessed by systems B and C may be entrenched in historical bias … whether someone is a high flyer precludes factors that make a great employee, like honesty and attitude” (M, V, 55)

  • “Accuracy is less important because the data [are] so subjective anyway” (C, V, 19)

  • “The custody sergeant will have experience and understanding of offenders and possibly what causes them to offend” (M, V, 51)

  • “[System C] removes individual bias made by individual officers in each respective case” (C, V, 10)

Theme C: TRUST IN AUTOMATED SYSTEMS
1. Fairness“There are not specialist doctors available 24 hours a day to reach a decision” (M V, 59)“[System C] is the fairest—it shortlists on organ suitability alone” (M, V, 55)“[System B] is ‘democratic’ in its analysis of past applicants, both successful and unsuccessful, without developing unknowable deep learning iterations” (M, G)“I don’t think something as abstract as potential to reoffend can be discerned by AI” (M, V, 55)
2. Delivery of the decisionn.a.c“[System C] doesn’t take away the sympathetic and empathetic aspect patients need throughout the whole process, but still assists doctors in decision making” (C, V, 18)n.a.c
  • “[With system A] there is more interaction with officers. Feelings and remorse will be taken into account and a full explanation will be given for the end result” (M, G)

  • “[System A] offers better dialogue between the authority and the offender with clearly defined parameters” (C, G)

3. Accountability for decisions“[With system C] who is responsible if the diagnosis is wrong with fatal consequences?” (C, G)
  • “[An explanation is] only relevant for cases that go wrong and an enquiry is launched” (C, V, 11)

  • “There is no recourse when something goes wrong” (C, V, 16)

n.a.cn.a.c

Abbreviations: AI, artificial intelligence; BMI, body mass index; n.a., not applicable.

a

For each quote/statement, the information in brackets refers to the jury location where it was captured (M=Manchester, C=Coventry), and whether it was raised during small/plenary group discussions (G) or as free text during electronic voting (V). In case of the latter, we also provided the participant ID.

b

System A, expert system (below human expert-level accuracy, full explanation); System B, random forest system (human expert-level accuracy, partial explanation); System C, deep learning system (beyond human expert-level accuracy, no explanation).

c

Subtheme did not emerge for this scenario.

Table 3.

Quotes and statements to illustrate jurors’ considerations when making the trade-off between accuracy and explainability of AI systemsa,b

Healthcare Scenarios
Non-Healthcare Scenarios
(Sub)ThemesStrokeKidney TransplantationRecruitmentCriminal Justice
Theme A: ACCURACY OF DECISION MAKING
1. Impact on the individual
  • “[System C] will save lives” (M, G)

  • “We have a moral and ethical obligation to improve speed and accuracy rates in diagnosis” (C, V, 11)

  • “The explanation comes secondary to accuracy, as without an accurate diagnosis there would be no basis foundation for treatment/aftercare” (C, V, 10)

  • “Higher accuracy level so less chance for an unsuccessful transplant” (M, V, 59)

  • “When you win the lottery you just need to know that you won, not which order the balls come up” (C, G)

  • “[System C would] save valuable time for the patients, as they would not potentially be given a kidney that would be rejected and go back in the waitlist” (C, V, 10)

“[System B] has key points/features for detection.” (C, G)
  • “Accuracy of the decision is paramount, as it may determine the individual’s future” (M, V, 68)

  • “The accuracy level [of system A] is the same as a human, but quicker” (M, V, 59)

2. Impact on society“It would be able to … possibly free up more time for experts to … do more research” (C, G)“It is important to get as accurate a match as possible to eliminate waste of a kidney that could have potentially helped another [recipient]” (M, V, 65)“[System B has] high level of possibility finding correct candidate” (C, V, 14)
  • “[System C has a] lower risk to [the] public” (C, V, 20)

  • “It is important to identify who will reoffend” (M, V, 53)

3. Increased efficiency
  • “[System C could] save money and time in the long term” (M, V, 56)

  • “[System C] frees up doctors for other duties” (C, V, 16)

“[System C] could reduce costs for the NHS in respect of rejection and fewer drugs prescribed post-operative[ly]” (C, V, 21)“[System C] gets through large amounts of CVs in a short space of time” (M, V, 59)“[System A] would require a lot of work from software engineers and policemen” (C, G)
Theme B: EXPLAINABILITY OF DECISION MAKING
1. Learning opportunities for the individual
  • “According to [the evidence], most stroke victims don't usually ask how the stroke was diagnosed” (C, V, 17)

  • “More information would lead to better aftercare [and could support] prevention of repeat of stroke” (C, V, 31)

  • “Patient rights to know how the system computed” (M, V, 67)

  • “Doctors wouldn’t trust themselves when presented with a machine that is almost always right and would sometimes incorrectly follow the machine’s advice especially when the machine offers no explanation at all.” (C, V, 24).

“No transparency to show the criteria used (BMI, blood pressure, lifestyle) factored into the decision” (C, G)
  • “Less transparent feedback [in system B]—eg, could tell you that you didn’t meet essential criteria but not tell you which part you did not meet to improve future chances of success” (C, G)

  • “Gives me motivation to improve and hone my skills for the future” (C, V, 26)

  • “It could rehabilitate a person who is in real need of direction” (M, V, 51)

  • “[System A is the] best system to provide transparent reasoning for individuals where they have specific factors to focus on during rehab/prosecution” (C, V, 18)

2. Learning opportunities for society“[With system C, you] can’t interrogate the system so can’t improve the result” (M, G)n.a.c“Feedback is essential for improving results for candidates, employers, and the economy” (C, V, 16)
  • “[System A allows] clear decision-making process providing transparency for offender, victim, and any potential further victim” (C, V, 11)

  • “No learning from data [with system C]” (M, V, 71)

3. Ability to identify/resolve biasn.a.c“Data going in to this system would be hugely skewed in favour of those privileged with existing wellness … [leading to widening] health inequalities” (C, V, 16)
  • “[System B] can somehow identify biases in the attributes which can be rectified” (C, V, 18)

  • “The factors assessed by systems B and C may be entrenched in historical bias … whether someone is a high flyer precludes factors that make a great employee, like honesty and attitude” (M, V, 55)

  • “Accuracy is less important because the data [are] so subjective anyway” (C, V, 19)

  • “The custody sergeant will have experience and understanding of offenders and possibly what causes them to offend” (M, V, 51)

  • “[System C] removes individual bias made by individual officers in each respective case” (C, V, 10)

Theme C: TRUST IN AUTOMATED SYSTEMS
1. Fairness“There are not specialist doctors available 24 hours a day to reach a decision” (M V, 59)“[System C] is the fairest—it shortlists on organ suitability alone” (M, V, 55)“[System B] is ‘democratic’ in its analysis of past applicants, both successful and unsuccessful, without developing unknowable deep learning iterations” (M, G)“I don’t think something as abstract as potential to reoffend can be discerned by AI” (M, V, 55)
2. Delivery of the decisionn.a.c“[System C] doesn’t take away the sympathetic and empathetic aspect patients need throughout the whole process, but still assists doctors in decision making” (C, V, 18)n.a.c
  • “[With system A] there is more interaction with officers. Feelings and remorse will be taken into account and a full explanation will be given for the end result” (M, G)

  • “[System A] offers better dialogue between the authority and the offender with clearly defined parameters” (C, G)

3. Accountability for decisions“[With system C] who is responsible if the diagnosis is wrong with fatal consequences?” (C, G)
  • “[An explanation is] only relevant for cases that go wrong and an enquiry is launched” (C, V, 11)

  • “There is no recourse when something goes wrong” (C, V, 16)

n.a.cn.a.c
Healthcare Scenarios
Non-Healthcare Scenarios
(Sub)ThemesStrokeKidney TransplantationRecruitmentCriminal Justice
Theme A: ACCURACY OF DECISION MAKING
1. Impact on the individual
  • “[System C] will save lives” (M, G)

  • “We have a moral and ethical obligation to improve speed and accuracy rates in diagnosis” (C, V, 11)

  • “The explanation comes secondary to accuracy, as without an accurate diagnosis there would be no basis foundation for treatment/aftercare” (C, V, 10)

  • “Higher accuracy level so less chance for an unsuccessful transplant” (M, V, 59)

  • “When you win the lottery you just need to know that you won, not which order the balls come up” (C, G)

  • “[System C would] save valuable time for the patients, as they would not potentially be given a kidney that would be rejected and go back in the waitlist” (C, V, 10)

“[System B] has key points/features for detection.” (C, G)
  • “Accuracy of the decision is paramount, as it may determine the individual’s future” (M, V, 68)

  • “The accuracy level [of system A] is the same as a human, but quicker” (M, V, 59)

2. Impact on society“It would be able to … possibly free up more time for experts to … do more research” (C, G)“It is important to get as accurate a match as possible to eliminate waste of a kidney that could have potentially helped another [recipient]” (M, V, 65)“[System B has] high level of possibility finding correct candidate” (C, V, 14)
  • “[System C has a] lower risk to [the] public” (C, V, 20)

  • “It is important to identify who will reoffend” (M, V, 53)

3. Increased efficiency
  • “[System C could] save money and time in the long term” (M, V, 56)

  • “[System C] frees up doctors for other duties” (C, V, 16)

“[System C] could reduce costs for the NHS in respect of rejection and fewer drugs prescribed post-operative[ly]” (C, V, 21)“[System C] gets through large amounts of CVs in a short space of time” (M, V, 59)“[System A] would require a lot of work from software engineers and policemen” (C, G)
Theme B: EXPLAINABILITY OF DECISION MAKING
1. Learning opportunities for the individual
  • “According to [the evidence], most stroke victims don't usually ask how the stroke was diagnosed” (C, V, 17)

  • “More information would lead to better aftercare [and could support] prevention of repeat of stroke” (C, V, 31)

  • “Patient rights to know how the system computed” (M, V, 67)

  • “Doctors wouldn’t trust themselves when presented with a machine that is almost always right and would sometimes incorrectly follow the machine’s advice especially when the machine offers no explanation at all.” (C, V, 24).

“No transparency to show the criteria used (BMI, blood pressure, lifestyle) factored into the decision” (C, G)
  • “Less transparent feedback [in system B]—eg, could tell you that you didn’t meet essential criteria but not tell you which part you did not meet to improve future chances of success” (C, G)

  • “Gives me motivation to improve and hone my skills for the future” (C, V, 26)

  • “It could rehabilitate a person who is in real need of direction” (M, V, 51)

  • “[System A is the] best system to provide transparent reasoning for individuals where they have specific factors to focus on during rehab/prosecution” (C, V, 18)

2. Learning opportunities for society“[With system C, you] can’t interrogate the system so can’t improve the result” (M, G)n.a.c“Feedback is essential for improving results for candidates, employers, and the economy” (C, V, 16)
  • “[System A allows] clear decision-making process providing transparency for offender, victim, and any potential further victim” (C, V, 11)

  • “No learning from data [with system C]” (M, V, 71)

3. Ability to identify/resolve biasn.a.c“Data going in to this system would be hugely skewed in favour of those privileged with existing wellness … [leading to widening] health inequalities” (C, V, 16)
  • “[System B] can somehow identify biases in the attributes which can be rectified” (C, V, 18)

  • “The factors assessed by systems B and C may be entrenched in historical bias … whether someone is a high flyer precludes factors that make a great employee, like honesty and attitude” (M, V, 55)

  • “Accuracy is less important because the data [are] so subjective anyway” (C, V, 19)

  • “The custody sergeant will have experience and understanding of offenders and possibly what causes them to offend” (M, V, 51)

  • “[System C] removes individual bias made by individual officers in each respective case” (C, V, 10)

Theme C: TRUST IN AUTOMATED SYSTEMS
1. Fairness“There are not specialist doctors available 24 hours a day to reach a decision” (M V, 59)“[System C] is the fairest—it shortlists on organ suitability alone” (M, V, 55)“[System B] is ‘democratic’ in its analysis of past applicants, both successful and unsuccessful, without developing unknowable deep learning iterations” (M, G)“I don’t think something as abstract as potential to reoffend can be discerned by AI” (M, V, 55)
2. Delivery of the decisionn.a.c“[System C] doesn’t take away the sympathetic and empathetic aspect patients need throughout the whole process, but still assists doctors in decision making” (C, V, 18)n.a.c
  • “[With system A] there is more interaction with officers. Feelings and remorse will be taken into account and a full explanation will be given for the end result” (M, G)

  • “[System A] offers better dialogue between the authority and the offender with clearly defined parameters” (C, G)

3. Accountability for decisions“[With system C] who is responsible if the diagnosis is wrong with fatal consequences?” (C, G)
  • “[An explanation is] only relevant for cases that go wrong and an enquiry is launched” (C, V, 11)

  • “There is no recourse when something goes wrong” (C, V, 16)

n.a.cn.a.c

Abbreviations: AI, artificial intelligence; BMI, body mass index; n.a., not applicable.

a

For each quote/statement, the information in brackets refers to the jury location where it was captured (M=Manchester, C=Coventry), and whether it was raised during small/plenary group discussions (G) or as free text during electronic voting (V). In case of the latter, we also provided the participant ID.

b

System A, expert system (below human expert-level accuracy, full explanation); System B, random forest system (human expert-level accuracy, partial explanation); System C, deep learning system (beyond human expert-level accuracy, no explanation).

c

Subtheme did not emerge for this scenario.

Theme A: Accuracy of decision-making

  1. Impact on the individual about whom the decision is being made; in the healthcare scenarios, the impact on individuals included issues such as preservation of life after stroke and reducing the risk of kidney rejection. For the kidney transplantation scenario, jurors felt that patients primarily needed to know that they had been matched for a transplant, with the explanation of why they were matched being secondary. Furthermore, they considered AI accuracy intrinsically linked to the speed with which healthcare professionals could act. Impact on the individual and speed of decision-making were also deemed important in the criminal justice scenario, with many jurors being surprised at how long people had to wait until their case went to court. For the recruitment scenario, jurors placed little emphasis on accuracy because they didn’t see there was necessarily 1 “best person for the job.”

  2. Impact on society; for both healthcare scenarios, jurors felt increased accuracy could reduce waste, for example in terms of freeing up time of experts or fewer rejected kidneys. For the recruitment scenario, jurors thought AI primarily contributed to recruitment companies’ reputation by improving their ability to place the right candidate in the right role—but without this necessarily benefitting job applicants. When considering the impact on society for the criminal justice scenario, jurors linked improved accuracy to a reduced reoffending risk.

  3. Increased efficiency; in the stroke scenario and for both non-healthcare scenarios, jurors viewed fully automated AI systems as a way to save time and resources. For the transplantation scenario, they considered the time gain less critical, and emphasized the potential to reduce costs for the NHS.

Theme B: Explainability of decision-making

  1. Learning opportunities for the individual; for the stroke scenario, jurors felt there was little need for an explanation of the diagnosis because providing an explanation was not part of current clinical practice. However, with regard to a fully automated AI system, they expressed concerns around doctors’ over-relying on the new technology and thereby losing valuable skills. One juror also mentioned patients’ general rights to be informed about their health and care. For both healthcare scenarios, jurors voting for system B felt some degree of explainability would support patients with decreasing their risk of another stroke or increasing their chances of receiving a kidney. Similarly, for the non-healthcare scenarios, jurors emphasized the need for an explanation as a way for individuals to self-improve and enhance their future prospects.

  2. Learning opportunities for society; learning opportunities for society included possibilities to improve the AI system itself through learning from the data and were primarily mentioned as an advantage of explainable AI in the non-healthcare context. Particularly for the criminal justice scenario, jurors discussed how an explanation would allow transparency of the decision-making process and potentially facilitate restorative justice.

  3. Ability to identify and resolve system biases; overall, jurors discussed the risk of inherent bias in AI systems less extensively in the healthcare scenarios compared to the non-healthcare scenarios, with no concerns expressed for the stroke scenario. They did see this as a problem in the other 3 scenarios, mainly because the data used to develop AI systems in these contexts was more likely to be subjective or biased. For example, because people have to be well enough to receive a donor kidney, there would be limited data available on outcomes for those who are less well but may still benefit. Jurors felt that more explainable AI systems would best allow humans to address these biases and challenge decisions if they felt these to be unfair or based on biased data. At the same time, however, some jurors suggested that, for the non-healthcare scenarios, human decisions are often biased, and they saw AI systems as a potential solution to correct this.

Theme C: Trust in automated systems

  1. Fairness; in the healthcare and recruitment scenarios, jurors considered more accurate systems fairer compared to more explainable systems because these contributed to equal access to stroke services or kidneys, or because AI systems would analyze data from all applicants equally, regardless of whether they had been successful or not. In the criminal justice scenario, jurors debated automated decision-making more passionately than for the other scenarios, preferring explainability when it came to fairness. They doubted if AI systems were suitable to fulfil such a complicated task without any human interference and flagged the long-term implications of incorrect decisions.

  2. Delivery of the decision; this subtheme only emerged from juror discussions about the kidney transplantation and criminal justice scenarios. For the former, jurors were comfortable with less explainable decisions, but nevertheless regarded it an integral part of a doctor’s professional role to relay these decisions to patients and answer any questions. In the criminal justice context, jurors stressed that more explainable AI systems had the potential to facilitate a conversation about the decision.

  3. Accountability for decisions; this subtheme only emerged for the healthcare scenarios, where jurors raised questions about who would be held responsible in case of incorrect automated decisions, and how accountability would be determined.

DISCUSSION

Main findings

We conducted two 5-day citizens’ juries in the UK with a representative sample of 36 participants to investigate how the public trades off accuracy and explainability of AI and whether this differs between healthcare and other scenarios. To our knowledge, our study is the first to engage citizens in public policy making on AI explainability in healthcare and compare this to other application domains.

We found that jurors favored accuracy over explainability in healthcare scenarios, but not in non-healthcare contexts where they either valued explainability equally to, or more than, accuracy. Their considerations for preferring accuracy regarded the potential positive impact of decisions on individuals and society and the potential to increase efficiency. Reasons for emphasizing explainability included opportunities for individuals and society to learn and improve, and an enhanced ability for humans to identify and address biases. Other considerations were related to trust in automated systems, including: fairness (with explanations considered crucial for fairness in criminal justice, but less so in other contexts); delivering the decision (primarily in criminal justice); and accountability (only in the healthcare context).

Relation to other studies

In a survey among 170 physicians, Diprose et al29 found strong associations between AI explainability and physicians’ understanding of and trust in AI. They also reported that 88% of responding physicians preferred explainable over non-explainable AI, but without asking respondents to make the trade-off between explainability and accuracy. This may partly account for the stark difference with our findings. Furthermore, patients currently already rely on physicians to make decisions about their health and treatment, so trusting another entity in this context is not new to them. This is different for physicians, whose position as a decision maker changes when introducing decision support systems that do not provide explanations. Lastly, physicians may feel, and often are, accountable for their decisions and, therefore, might want to be able to explain AI system recommendations to patients. However, our results suggest that patients may value explainability differently, especially when weighed against a potential decrease in decision accuracy.

Limitations

One limitation of our study was the difference in materials and witnesses between the healthcare and non-healthcare scenarios, which may have affected their comparability and how jurors made the trade-off for each of them. The healthcare scenarios included perspectives of witnesses who would be directly affected by decisions of the AI systems (ie, kidney patient waiting for a donor and a Stroke Association support worker), whereas the non-healthcare scenario witnesses were system users (ie, recruitment agency staff and criminal justice experts).

Both juries produced a report in their own words as part of the jury process, which are publicly available.24 Although the themes and deliberations in the jurors’ reports strongly aligned with the themes emerging from our post-jury analysis, we did not ask the jurors to confirm whether they thought our themes adequately reflected their perspectives (ie, member checking). This is another limitation of our study.

Implications

The citizens in our juries made it clear that they cared more about accuracy than about explainability in healthcare scenarios compared to non-healthcare scenarios. Therefore, a categorical rule (eg, in legislation) that citizens are always entitled to explanations of AI decisions regardless of the circumstances may not be in keeping with citizens’ preferences and their best interests; particularly in cases of more advanced AI systems where it is difficult or impossible to offer good explanations without compromising accuracy. Our findings suggest that, instead of setting categorical rules around AI explainability, policy makers should consider making these domain-specific.

In the UK, our findings have informed new guidance,30 developed by the Information Commissioner’s Office (the UK’s independent body set up to uphold information rights) and the Alan Turing Institute (the UK’s national institute for data science and AI). This new guidance gives organizations practical advice on explaining the processes, services, and decisions delivered or assisted by AI to the individuals affected by them. It argues that to enhance the explainability of AI decisions, organizations should: be transparent, be accountable, consider the context in which they are operating, and reflect on the impact of the AI system on the individuals affected as well as the wider society.

In our study, we assumed a trade-off between accuracy and explainability of AI models. Generally speaking, more complex systems are capable of modeling complex decision functions more accurately but are less likely to be interpretable by people. However, not every decision problem in healthcare requires modeling of a complex function: sometimes relatively simple models can achieve similar accuracy levels.31 In such cases, there is no trade-off between accuracy and explainability. To account for a range of situations with trade-offs being more or less prominent, developers of healthcare decision support systems could consider incorporating a simpler model alongside a more complex one and present end users with the results of both. Finally, there may be ways to achieve the goal of explainability without end users necessarily having model logic completely explained or to reduce the need for explainability through alternative guarantees (eg, algorithmic fairness32) This is a field of ongoing investigation.33

Unanswered questions

Despite strong consensus on the importance of explainability of AI decision-making, there exists significant variation in what is meant by this.8 Some authors assume it relates to data use, while others link it to access to source code, autonomy of the AI system, the extent to which it achieves causal understanding,34 or contestability of the system’s decision.33 In the context of our study, full explainability (system A) meant that it is always possible to trace the reasoning of a system from start to end. Partial explainability (system B) meant it is known which variables are important for decision-making but not how they are important. Future research may look into the trade-off with accuracy for other types of explainability and compare these with our findings.

AI explainability can be defined relative to different groups of stakeholders (eg, healthcare professionals, patients, regulators, policy makers)35 as well as to healthcare decisions (eg, stroke diagnosis) with and without immediate impact on life and death (eg, prognosis). The trade-off between accuracy and explainability is likely to be relevant to all these stakeholders and in all these decisions, but they might make the trade-off differently—particularly when presented with a variety of scenarios. Future studies should, therefore, investigate the views of these stakeholder groups on a broader range of scenarios.

CONCLUSION

Citizens may value explainability of AI systems in healthcare less than in non-healthcare contexts and less than often assumed by healthcare professionals and researchers, especially when weighed against system accuracy. Our findings thus warrant active consultation of the public when developing policy on AI explainability in healthcare settings.

FUNDING

This work was funded by the National Institute for Health Research Greater Manchester Patient Safety Translational Research Centre (NIHR GM PSTRC) and the Information Commissioner’s Office. The views expressed are those of the author(s) and not necessarily those of the NHS, the NIHR or the Department of Health and Social Care. NIHR had no role in study design, data collection, data analysis, data interpretation, or writing of the report. The corresponding author has full access to all of the data and the final responsibility to submit for publication.

AUTHOR CONTRIBUTIONS

NP and MO conceived the idea for the study. MT, KB, SA, AH, and CW helped to develop the idea. KB and SA designed and facilitated both jury events. MO collected and analyzed quantitative data, while LR, SHS, and DP collected and analyzed qualitative data. SvdV helped to interpret the results. SvdV, LR, and NP drafted the manuscript with important intellectual input from all authors, who approved the final version of the manuscript for publication.

DATA AVAILABILITY STATEMENT

The data underlying this article will be shared on reasonable request to the corresponding author.

SUPPLEMENTARY MATERIAL

Supplementary material is available at Journal of the American Medical Informatics Association online.

ACKNOWLEDGMENTS

We wish to thank the 36 participants in the citizens juries, the expert and scenario witnesses, and the members of the external oversight committee (Ms Reema Patel, programme manager for Data Ethics & AI, Ada Lovelace Institute), Dr David Leslie (ethics fellow, Alan Turing Institute), and Professor Søren Holm (professor of Bioethics, The University of Manchester).

CONFLICT OF INTEREST STATEMENT

MO is director of Citizens’ Juries Community Interest Company, a social enterprise dedicated to designing and running citizens’ juries and was commissioned to deliver these juries. KB (executive director) and SA (design and facilitation specialist) work for the Center for New Democratic Processes (formerly Jefferson Center), which invented the citizens’ jury process. AH (principal policy advisor) and CW (senior policy officer) work for the Information Commissioner’s Office, which partly funded the juries. The other authors have no conflicts of interest to declare.

Box 1.

Five-dayjury program

DayJury ActivityTopic and Expert Witness
Day 1 AM
  • Welcome & process overview (P)

  • Guidelines for deliberation (P)

  • Simulation exercise (G)

Day 1 PM
  • Current perceptions of AI (P)

  • “The Joy of AI” videoa (P)

  • Study time & discussion (P)

  • Wrap-up & daily evaluations (P, V)

  • Balancing AI and explainability

  • Prof. Sofia Olhede (Professor of Statistics, University College London; special interest in how data and algorithms are impacting our daily lives)

Day 2 AM
  • Discussions (P)

  • Decision & explanations exercise (P, G)

  • The law concerning data protection and AI

  • Rhiannon Webster (specialist information lawyer and partner, DAC Beachcroft, London)

Day 2 PM
  • Discussion (P)

  • Wrap-up & daily evaluation (P, G)

  • Making the case for AI accuracy

  • Dr Andre Freitas (lecturer in computer science, University of Manchester; specialist in AI)b

  • Making the case for transparent and explainable AI

  • Prof. Alan Winfield (professor of robot ethics, University of the West of England; researches and develops standards in robot ethics)b

Day 3 AM
  • Context for scenarios (P)

  • Stroke scenario (P)

  • Discussion & voting (P, G, V)

  • Interpreting and helping to explain how AI is being applied in the 4 jury scenarios

  • Dr Allan Tucker (reader in computer science, Brunel University London; special interest in the application of AI in medicine)

Day 3 PM
  • Recruitment scenario (P)

  • Discussion & voting (P, G, V)

  • Wrap-up & daily evaluations (P, V)

Day 4 AM
  • Context for scenarios (P)

  • Kidney transplantation scenario (P)

  • Discussion & voting (P, G, V)

Day 4 PM
  • Criminal justice scenario (P)

  • Discussion & voting (P, G, V)

  • Discussion on scenarios (P)

  • Wrap-up & daily evaluations (P, V)

Day 5 AMDiscussion: General explainability vs accuracy (P)
Day 5 PM
  • Discussion (P, G, V)

  • Presentation of draft jury report (P)

  • Wrap-up & daily evaluations (P, V)

DayJury ActivityTopic and Expert Witness
Day 1 AM
  • Welcome & process overview (P)

  • Guidelines for deliberation (P)

  • Simulation exercise (G)

Day 1 PM
  • Current perceptions of AI (P)

  • “The Joy of AI” videoa (P)

  • Study time & discussion (P)

  • Wrap-up & daily evaluations (P, V)

  • Balancing AI and explainability

  • Prof. Sofia Olhede (Professor of Statistics, University College London; special interest in how data and algorithms are impacting our daily lives)

Day 2 AM
  • Discussions (P)

  • Decision & explanations exercise (P, G)

  • The law concerning data protection and AI

  • Rhiannon Webster (specialist information lawyer and partner, DAC Beachcroft, London)

Day 2 PM
  • Discussion (P)

  • Wrap-up & daily evaluation (P, G)

  • Making the case for AI accuracy

  • Dr Andre Freitas (lecturer in computer science, University of Manchester; specialist in AI)b

  • Making the case for transparent and explainable AI

  • Prof. Alan Winfield (professor of robot ethics, University of the West of England; researches and develops standards in robot ethics)b

Day 3 AM
  • Context for scenarios (P)

  • Stroke scenario (P)

  • Discussion & voting (P, G, V)

  • Interpreting and helping to explain how AI is being applied in the 4 jury scenarios

  • Dr Allan Tucker (reader in computer science, Brunel University London; special interest in the application of AI in medicine)

Day 3 PM
  • Recruitment scenario (P)

  • Discussion & voting (P, G, V)

  • Wrap-up & daily evaluations (P, V)

Day 4 AM
  • Context for scenarios (P)

  • Kidney transplantation scenario (P)

  • Discussion & voting (P, G, V)

Day 4 PM
  • Criminal justice scenario (P)

  • Discussion & voting (P, G, V)

  • Discussion on scenarios (P)

  • Wrap-up & daily evaluations (P, V)

Day 5 AMDiscussion: General explainability vs accuracy (P)
Day 5 PM
  • Discussion (P, G, V)

  • Presentation of draft jury report (P)

  • Wrap-up & daily evaluations (P, V)

Abbreviations: G, group work; P, plenary session; V, voting.

a

Provides a general introduction to artificial intelligence and what it can do, and discusses AI accuracy [25].

b

Partial expert witness who argued for a specific position rather than giving a neutral account of the topic.

Box 1.

Five-dayjury program

DayJury ActivityTopic and Expert Witness
Day 1 AM
  • Welcome & process overview (P)

  • Guidelines for deliberation (P)

  • Simulation exercise (G)

Day 1 PM
  • Current perceptions of AI (P)

  • “The Joy of AI” videoa (P)

  • Study time & discussion (P)

  • Wrap-up & daily evaluations (P, V)

  • Balancing AI and explainability

  • Prof. Sofia Olhede (Professor of Statistics, University College London; special interest in how data and algorithms are impacting our daily lives)

Day 2 AM
  • Discussions (P)

  • Decision & explanations exercise (P, G)

  • The law concerning data protection and AI

  • Rhiannon Webster (specialist information lawyer and partner, DAC Beachcroft, London)

Day 2 PM
  • Discussion (P)

  • Wrap-up & daily evaluation (P, G)

  • Making the case for AI accuracy

  • Dr Andre Freitas (lecturer in computer science, University of Manchester; specialist in AI)b

  • Making the case for transparent and explainable AI

  • Prof. Alan Winfield (professor of robot ethics, University of the West of England; researches and develops standards in robot ethics)b

Day 3 AM
  • Context for scenarios (P)

  • Stroke scenario (P)

  • Discussion & voting (P, G, V)

  • Interpreting and helping to explain how AI is being applied in the 4 jury scenarios

  • Dr Allan Tucker (reader in computer science, Brunel University London; special interest in the application of AI in medicine)

Day 3 PM
  • Recruitment scenario (P)

  • Discussion & voting (P, G, V)

  • Wrap-up & daily evaluations (P, V)

Day 4 AM
  • Context for scenarios (P)

  • Kidney transplantation scenario (P)

  • Discussion & voting (P, G, V)

Day 4 PM
  • Criminal justice scenario (P)

  • Discussion & voting (P, G, V)

  • Discussion on scenarios (P)

  • Wrap-up & daily evaluations (P, V)

Day 5 AMDiscussion: General explainability vs accuracy (P)
Day 5 PM
  • Discussion (P, G, V)

  • Presentation of draft jury report (P)

  • Wrap-up & daily evaluations (P, V)

DayJury ActivityTopic and Expert Witness
Day 1 AM
  • Welcome & process overview (P)

  • Guidelines for deliberation (P)

  • Simulation exercise (G)

Day 1 PM
  • Current perceptions of AI (P)

  • “The Joy of AI” videoa (P)

  • Study time & discussion (P)

  • Wrap-up & daily evaluations (P, V)

  • Balancing AI and explainability

  • Prof. Sofia Olhede (Professor of Statistics, University College London; special interest in how data and algorithms are impacting our daily lives)

Day 2 AM
  • Discussions (P)

  • Decision & explanations exercise (P, G)

  • The law concerning data protection and AI

  • Rhiannon Webster (specialist information lawyer and partner, DAC Beachcroft, London)

Day 2 PM
  • Discussion (P)

  • Wrap-up & daily evaluation (P, G)

  • Making the case for AI accuracy

  • Dr Andre Freitas (lecturer in computer science, University of Manchester; specialist in AI)b

  • Making the case for transparent and explainable AI

  • Prof. Alan Winfield (professor of robot ethics, University of the West of England; researches and develops standards in robot ethics)b

Day 3 AM
  • Context for scenarios (P)

  • Stroke scenario (P)

  • Discussion & voting (P, G, V)

  • Interpreting and helping to explain how AI is being applied in the 4 jury scenarios

  • Dr Allan Tucker (reader in computer science, Brunel University London; special interest in the application of AI in medicine)

Day 3 PM
  • Recruitment scenario (P)

  • Discussion & voting (P, G, V)

  • Wrap-up & daily evaluations (P, V)

Day 4 AM
  • Context for scenarios (P)

  • Kidney transplantation scenario (P)

  • Discussion & voting (P, G, V)

Day 4 PM
  • Criminal justice scenario (P)

  • Discussion & voting (P, G, V)

  • Discussion on scenarios (P)

  • Wrap-up & daily evaluations (P, V)

Day 5 AMDiscussion: General explainability vs accuracy (P)
Day 5 PM
  • Discussion (P, G, V)

  • Presentation of draft jury report (P)

  • Wrap-up & daily evaluations (P, V)

Abbreviations: G, group work; P, plenary session; V, voting.

a

Provides a general introduction to artificial intelligence and what it can do, and discusses AI accuracy [25].

b

Partial expert witness who argued for a specific position rather than giving a neutral account of the topic.

Box 2.

The 4 scenarios involving application of AI as discussed by the juries

HEALTHCARE SCENARIOS

Stroke: In this scenario, diagnostic data are accumulated from previous stroke patients and compared to information about a patient's acute symptoms (eg, paralysis and loss of speech), their medical history, and neuro-radiological images (eg, brain scans) to identify whether they have had a stroke; and its type, location, and severity. Expert witnesses included a support worker from the UK’s Stroke Association. Rapid and accurate diagnosis of stroke by neuroradiologists with many years of training and experience, greatly increases chances of survival and recovery of the patient. However, they are not always available in each hospital, and, in practice, diagnosis is often done by non-specialist emergency medicine doctors (with less accuracy).

Kidney transplantation: In this scenario, medical data were used to categorize patients awaiting kidney transplantation into 1 of 3 groups—low, medium, and high risk of kidney rejection. The expert witness explained that the AI system stratified patients based on their risk of rejection and only those categorized as “low risk” would be prioritized for transplantation. Traditionally, prioritization would only take the patient’s age into acount.

NON-HEALTHCARE SCENARIOS

Recruitment: In this scenario, the proposed AI system screened applicants offering an interview based on their likelihood to become a “high-performing employee” using existing company data. This replaces the need for manual short-listing and telephone pre-interviews. The witness interviewed was a senior recruiting business partner with 15–20 years experience in the insurance and finance sector whose priority was to find the right candidate for the job, with current unsuccessful applicants being offered capacity to receive feedback.

Criminal justice: This scenario discussed an AI system currently being trialled within the UK. People charged with committing low-level (ie, non-violent, non-sexual) crimes would be stratified by the AI system into a group who would be offered a rehabilitation program rather than being dealt with through usual court procedures. This would allow courts to prioritize more serious offenders and provide less serious offenders with an educational intervention.

REFERENCES

1

UK Research and Innovation (UKRI). Transforming our world with AI. UKRI’s role in embracing the opportunity.

2021
. https://www.ukri.org/wp-content/uploads/2021/02/UKRI-120221-TransformingOurWorldWithAI.pdf. Accessed April 14, 2021.

2

Jordan
MI
,
Mitchell
TM.
Machine learning: trends, perspectives, and prospects
.
Science
2015
;
349
(
6245
):
255
60
.

3

Liu
X
,
Faes
L
,
Kale
AU
, et al.
A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis
.
Lancet Digit Heal
2019
;
1
(
6
):
e271
97
.

4

Darcy
AM
,
Louie
AK
,
Roberts
LW.
Machine learning and the profession of medicine
.
JAMA
2016
;
315
:
551
2
.

5

Wang
F
,
Kaushal
R
,
Khullar
D.
Should health care demand interpretable artificial intelligence or accept ‘black box’ medicine?
Ann Intern Med
2020
;
172
:
59
61
. doi:10.7326/M19-2548

6

Samek
W
,
Montavon
G
,
Vedaldi
A
, et al.
Explainable AI: Interpreting, Explaining and Visualizing Deep Learning
. Cham, Switzerland:
Springer
;
2019
.

7

Gunning
D
,
Stefik
M
,
Choi
J
, et al.
XAI—Explainable artificial intelligence
.
Sci Robot
2019
;
4
(
37
):
eaay7120
.

8

Jobin
A
,
Ienca
M
,
Vayena
E.
The global landscape of AI ethics guidelines
.
Nat Mach Intell
2019
;
1
(
9
):
389
99
.

9

United States Federal Government. Code of Federal Regulations, Regulation B: Equal Credit Opportunity Act, Title 12, Chapter X, Part 1002, §1002.9(b)(2).

10

European Union. General Data Protection Regulation, Recital 71.

11

The Jefferson Center. How we work: citizens juries. https://www.jefferson-center.org/about-us/how-we-work/. Accessed December 27,

2020
.

12

Smith
G
,
Wales
C.
Citizens’ juries and deliberative democracy
.
Polit Stud
2000
;
48
(
1
):
51
65
.

13

Armour
A.
The citizens’ jury model of public participation: a critical evaluation. In:
Renn
O
,
Webler
T
,
Wiedemann
P
, eds.
Fairness and Competence in Citizen Participation.
Dordrecht
:
Springer
;
1995
: 175–87. doi: 10.1007/978-94-011-0131-8_9.

14

Wakeford
T.
Citizens juries: a radical alternative for social research
.
Soc Res Updat
2002
;
37
:
1
5
.

15

Iredale
R
,
Longley
M
,
Thomas
C
, et al.
What choices should we be able to make about designer babies? A citizens’ jury of young people in South Wales
.
Health Expect
2006
;
9
:
207
17
.

16

Paul
C
,
Nicholls
R
,
Priest
P
, et al.
Making policy decisions about population screening for breast cancer: the role of citizens’ deliberation
.
Health Policy (New York)
2008
;
85
:
314
20
.

17

Thomas
R
,
Sims
R
,
Beller
E
, et al.
An Australian community jury to consider case-finding for dementia: differences between informed community preferences and general practice guidelines
.
Heal Expect
2019
;
22
:
475
84
.

18

Tully
MP
,
Bozentko
K
,
Clement
S
, et al.
Investigating the extent to which patients should control access to patient records for research: a deliberative process using citizens’ juries
.
J Med Internet Res
2018
;
20
(
3
):
e112
.

19

Tully
MP
,
Hassan
L
,
Oswald
M
, et al.
Commercial use of health data—a public “trial” by citizens’ jury
.
Learn Heal Syst
2019
;
3
:
1
13
.

20

Ford
E
,
Oswald
M
,
Hassan
L
, et al.
Should free-text data in electronic medical records be shared for research? A citizens’ jury study in the UK
.
J Med Ethics
2020
;
46
:
367
77
.

21

Citizens’ Juries c.i.c. Putting the public into public decision making. https://citizensjuries.org. Accessed December 27,

2020
.

22

Office for National Statistics. 2011 Census. Key Stat. Quick Stat. local authorities United Kingdom.

2013
. https://www.ons.gov.uk/census/2011census. Accessed December 27, 2020.

23

Balaram
B
,
Greenham
T
,
Leonard
J.
Artificial Intelligence: Real Public Engagement.
2018
. https://www.thersa.org/reports/artificial-intelligence-real-public-engagement

24

NIHR Greater Manchester Patient Safety Translational Research Centre (Greater Manchester PSTRC). Citizens’ Juries on Decision Making and Explainable Artificial Intelligence. 2019. http://www.patientsafety.manchester.ac.uk/research/themes/safety-informatics/citizens-juries/. Accessed May 4, 2021.

25

BBC Four. The Joy of AI.

2018
. https://www.dailymotion.com/video/x7takq4. Accessed May 4, 2021.

26

Glaser
B
,
Strauss
A.
The Discovery of Grounded Theory. Strategies for Qualitative Research
.
Chicago, IL
:
Aldine de Gruyter
;
1967
.

27

Braun
V
,
Clarke
V.
Using thematic analysis in psychology
.
Qual Res Psychol
2006
;
3
(
2
):
77
101
.

28

Office for National Statistics. Internet users in the UK: 2019. 2019. https://www.ons.gov.uk/releases/internetusersintheuk2019. Accessed April 13, 2021.

29

Diprose
WK
,
Buist
N
,
Hua
N
, et al.
Physician understanding, explainability, and trust in a hypothetical machine learning risk calculator
.
J Am Med Inform Assoc
2020
;
27
(
4
):
592
600
.

30

Information Commissioner’s Office, Alan Turing Institute. Explaining decisions made with AI.

2019
. https://ico.org.uk/for-organisations/guide-to-data-protection/key-data-protection-themes/explaining-decisions-made-with-ai/. Accessed July 9, 2021.

31

Christodoulou
E
,
Ma
J
,
Collins
GS
, et al.
A systematic review shows no performance benefit of machine learning over logistic regression for clinical prediction models
.
J Clin Epidemiol
2019
;
110
:
12
22
.

32

Kearns
M
,
Roth
A.
The Ethical Algorithm. The Science of Socially Aware Algorithm Design
.
Oxford, UK
:
Oxford University Press
;
2020
.

33

Ploug
T
,
Holm
S.
The four dimensions of contestable AI diagnostics—a patient-centric approach to explainable AI
.
Artif Intell Med
2020
;
107
:
101901
.

34

Holzinger
A
,
Langs
G
,
Denk
H
, et al.
Causability and explainability of artificial intelligence in medicine
.
Wiley Interdiscip Rev Data Min Knowl Discov
2019
;
9
:
1
13
.

35

Amann
J
,
Blasimme
A
,
Vayena
E
, et al.
Explainability for artificial intelligence in healthcare: a multidisciplinary perspective
.
BMC Med Inform Decis Mak
2020
;
20
:
310
.

Author notes

*

Authors contributed equally.

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.

Supplementary data