Abstract

Importance

The US healthcare system faces significant challenges, including clinician burnout, operational inefficiencies, and concerns about patient safety. Artificial intelligence (AI), particularly generative AI, has the potential to address these challenges, but its adoption, effectiveness, and barriers to implementation are not well understood.

Objective

To evaluate the current state of AI adoption in US healthcare systems, assess successes and barriers to implementation during the early generative AI era.

Design, setting, and participants

This cross-sectional survey was conducted in Fall 2024, and included 67 health systems members of the Scottsdale Institute, a collaborative of US non-profit healthcare organizations. Forty-three health systems completed the survey (64% response rate). Respondents provided data on the deployment status and perceived success of 37 AI use cases across 10 categories.

Main outcomes and measures

The primary outcomes were the extent of AI use case development, piloting, or deployment, the degree of reported success for AI use cases, and the most significant barriers to adoption.

Results

Across the 43 responding health systems, AI adoption and perceptions of success varied significantly. Ambient Notes, a generative AI tool for clinical documentation, was the only use case with 100% of respondents reporting adoption activities, and 53% reported a high degree of success with using AI for Clinical Documentation. Imaging and radiology emerged as the most widely deployed clinical AI use case, with 90% of organizations reporting at least partial deployment, although successes with diagnostic use cases were limited. Similarly, many organizations have deployed AI for clinical risk stratification such as early sepsis detection, but only 38% report high success in this area. Immature AI tools were identified a significant barrier to adoption, cited by 77% of respondents, followed by financial concerns (47%) and regulatory uncertainty (40%).

Conclusions and relevance

Ambient Notes is rapidly advancing in US healthcare systems and demonstrating early success. Other AI use cases show varying degrees of adoption and success, constrained by barriers such as immature AI tools, financial concerns, and regulatory uncertainty. Addressing these challenges through robust evaluations, shared strategies, and governance models will be essential to ensure effective integration and adoption of AI into healthcare practice.

Introduction

As the US healthcare system grapples with significant challenges in quality, affordability, and labor shortages, Artificial Intelligence (AI) holds promise for transforming the delivery system to become safer, more effective, less wasteful, and more patient-centered.1,2 AI has already demonstrated success in preventing sepsis,3 improving diagnostic accuracy in radiology and pathology,4–6 and reducing clinicians’ documentation burden.7–12 Driven by recent advances in generative AI, investment in health AI continues to grow, with over $30B invested into the healthcare AI companies in the past 3 years.13 This has resulted in a proliferation of solutions that could provide relief to the strained healthcare delivery system.

While AI has been integrated into many aspects of modern life, its extent of use in healthcare delivery remains unclear. The medical literature showcases many examples, but the vast majority have only been demonstrated on select organizations and/or patient populations. As with all forms of innovation, the path to widespread adoption and societal impact is long and by no means certain.14,15 Adoption of AI in healthcare likely faces unique sociological and technical challenges.16 As such, an assessment of the current AI use and impact in healthcare may inform innovators, technology vendors, policy makers, and organizational decision makers.

Factors that influence technology adoption are dynamic. Previous studies that have evaluated AI deployment in US healthcare organizations17–20 are limited in scope and took place before current generative AI has significantly impacted the healthcare industry. It remains unclear how traditional AI, which focuses on predicting clinical or operational outcomes, is being deployed alongside more advanced AI tools that interpret and generate human language and images. In addition, while previous studies have identified various barriers to utilizing AI in healthcare, their relative importance remains unclear.17,18 In light of the new-found promise and enthusiasm surrounding generative AI, our team aimed to assess AI adoption early in the generative AI era, identify the most significant barriers faced by healthcare organizations, and highlight initial examples of successes.

Methods

Instrument development

Our team of healthcare, medical informatics, and organizational management experts developed a survey instrument to (1) assess the state of and drivers behind AI adoption in major US health systems, (2) evaluate successes and barriers for AI, and (3) identify candidates for (future) in-depth case studies.

To develop the survey instrument, we reviewed the literature and analyzed transcripts from various industry meetings, including prior Scottsdale Institute (SI)'s AI surveys17,18 and SI membership AI conferences between Spring 2023 and Spring 2024. First, we identified six key themes related to why health systems may be developing and deploying AI tools: Patient Safety/Quality, Caregiver Burden/Satisfaction, Margin Improvement/Financial, Workflow Efficiency/Productivity, Patient/Consumer Experience, and Market Share/Competitiveness. We asked respondents to rank these 6 organizational goals. We also compiled AI use cases from literature, conference transcripts, and past research,18 identifying 37 use cases in 10 categories related to clinical care, healthcare operations, analytics and research (Appendix S1).

Respondents reported the status of AI adoption in their organizations–developing, piloting, partially deploying, or fully deployed—for each use case. Respondents could also add unlisted use cases via free text. We asked respondents about their perceived degree of success of AI use cases in each of the 10 categories. Based on results of our literature search and prior work, we identified 9 common metrics used to assess the performance of AI tools, and we asked respondents to report how frequently they used each of them. Six key barriers to adopting AI were also identified: financial concerns, regulatory uncertainty, lack of leadership support, low clinician adoption, insufficient expertise or technology, and immature AI tools. Respondents were asked to prioritize these barriers.

To further refine the instrument, we conducted iterative reviews and pilot testing with a small group of health system leaders. This process allowed us to assess the clarity, relevance, and comprehensiveness of the questions, ensuring that they effectively captured meaningful insights. The finalized survey instrument appears in Appendix S2. We used the American Association for Public Opinion Research (AAPOR) Best Practices for Survey Research21 to ensure the accuracy and reliability of our findings.

Participant recruitment

The organizations included in the invitation were members of SI, a not-for-profit organization that assists its 67 US non-profit health systems in accomplishing quality improvement, organizational efficiency, patient safety and transformation through information technology. Invitations were sent to all 67 SI members’ Chief Medical Information Officers, Chief Information/Digital Officers, and/or Chief AI Officers via email from September 27 to October 20, 2024. Each health system was asked to provide one single response via the survey software SurveyMonkey. The Institution Review Board (IRB) at the University of Alabama at Birmingham approved the study (IRB-300009793).

Data analysis

The research team aggregated survey data using Microsoft Excel. Characteristics of health systems represented by survey respondents and non-survey respondents were compared with those of US health systems using online statistical calculators for Chi-Squared test and Fischer’s Exact test.

Results

Of 67 total health systems surveyed, 46 answered at least one question and 43 completed the entire survey (64% survey completion rate). The 3 health systems that attempted but did not complete the survey only responded to the opening question and were discarded from further analysis. Characteristics of (completed) survey respondents and non-respondents in terms of size by net patient revenue (NPR), size by number of hospitals, teaching status, and geography (as sourced from Definitive Healthcare22) are summarized in Table 1. There was no statistically significant difference between survey respondents versus non-respondents, except a higher proportion of our respondents had an NPR between $5B and $9.9B compared to non-respondents. Compared to all US healthcare systems, survey respondents were more heavily represented in the 2 middle categories for NPR ($1-$4.9 billion and $5-9.9 billion). The most common title of the respondents were Chief [Medical | Health | Clinical] Information Officers (50%, n = 23), while 13% (n = 6) held the Chief [Information | Digital] Officer title, and another 13% (n = 6) held the Chief [Analytics | AI] Officer or Chief Data Scientist title.

Table 1.

Characteristics of US hospitals, organizations invited to participate in the survey, and responding organizations.

US health systems n(%)SI members
US health systems vs survey respondents (P value)
Overall n(%)Survey respondents n(%)Survey non-respondents n(%)Respondents vs non-respondents (P value)
Size—NPR.04a<.01b
 Under $1 billion53 (26%)1 (1%)1 (2%)0 (0%)
 $1 billion-$4.9 billion97 (48%)43 (64%)25 (58%)15 (71%)
 $5 billion-$9.9 billion36 (18%)17 (25%)15 (35%)2 (10%)
 $10 billion and above17 (8%)6 (9%)2 (5%)4 (19%)
Size—number of hospitals.79b.93b
 4 or fewer hospitals62 (31%)16 (24%)12 (28%)4 (19%)
 5-10 hospitals57 (28%)23 (34%)14 (33%)6 (29%)
 11-19 hospitals41 (20%)15 (22%)9 (21%)6 (29%)
 20+ hospitals43 (21%)13 (19%)8 (19%)5 (24%)
Teaching status.07a.37b
 Major teaching122 (60%)51 (76%)29 (67%)18 (90%)
 Non-major teaching81 (40%)16 (24%)14 (33%)2 (10%)
Geographic region.16a.15b
 Northeast37 (18%)6 (9%)3 (7%)2 (10%)
 South72 (35%)21 (31%)13 (30%)8 (38%)
 Midwest55 (27%)27 (40%)15 (35%)10 (48%)
 West38 (19%)13 (19%)12 (28%)1 (5%)
 US Territory1 (0%)0 (0%)0 (0%)0 (0%)
US health systems n(%)SI members
US health systems vs survey respondents (P value)
Overall n(%)Survey respondents n(%)Survey non-respondents n(%)Respondents vs non-respondents (P value)
Size—NPR.04a<.01b
 Under $1 billion53 (26%)1 (1%)1 (2%)0 (0%)
 $1 billion-$4.9 billion97 (48%)43 (64%)25 (58%)15 (71%)
 $5 billion-$9.9 billion36 (18%)17 (25%)15 (35%)2 (10%)
 $10 billion and above17 (8%)6 (9%)2 (5%)4 (19%)
Size—number of hospitals.79b.93b
 4 or fewer hospitals62 (31%)16 (24%)12 (28%)4 (19%)
 5-10 hospitals57 (28%)23 (34%)14 (33%)6 (29%)
 11-19 hospitals41 (20%)15 (22%)9 (21%)6 (29%)
 20+ hospitals43 (21%)13 (19%)8 (19%)5 (24%)
Teaching status.07a.37b
 Major teaching122 (60%)51 (76%)29 (67%)18 (90%)
 Non-major teaching81 (40%)16 (24%)14 (33%)2 (10%)
Geographic region.16a.15b
 Northeast37 (18%)6 (9%)3 (7%)2 (10%)
 South72 (35%)21 (31%)13 (30%)8 (38%)
 Midwest55 (27%)27 (40%)15 (35%)10 (48%)
 West38 (19%)13 (19%)12 (28%)1 (5%)
 US Territory1 (0%)0 (0%)0 (0%)0 (0%)
Table 1.

Characteristics of US hospitals, organizations invited to participate in the survey, and responding organizations.

US health systems n(%)SI members
US health systems vs survey respondents (P value)
Overall n(%)Survey respondents n(%)Survey non-respondents n(%)Respondents vs non-respondents (P value)
Size—NPR.04a<.01b
 Under $1 billion53 (26%)1 (1%)1 (2%)0 (0%)
 $1 billion-$4.9 billion97 (48%)43 (64%)25 (58%)15 (71%)
 $5 billion-$9.9 billion36 (18%)17 (25%)15 (35%)2 (10%)
 $10 billion and above17 (8%)6 (9%)2 (5%)4 (19%)
Size—number of hospitals.79b.93b
 4 or fewer hospitals62 (31%)16 (24%)12 (28%)4 (19%)
 5-10 hospitals57 (28%)23 (34%)14 (33%)6 (29%)
 11-19 hospitals41 (20%)15 (22%)9 (21%)6 (29%)
 20+ hospitals43 (21%)13 (19%)8 (19%)5 (24%)
Teaching status.07a.37b
 Major teaching122 (60%)51 (76%)29 (67%)18 (90%)
 Non-major teaching81 (40%)16 (24%)14 (33%)2 (10%)
Geographic region.16a.15b
 Northeast37 (18%)6 (9%)3 (7%)2 (10%)
 South72 (35%)21 (31%)13 (30%)8 (38%)
 Midwest55 (27%)27 (40%)15 (35%)10 (48%)
 West38 (19%)13 (19%)12 (28%)1 (5%)
 US Territory1 (0%)0 (0%)0 (0%)0 (0%)
US health systems n(%)SI members
US health systems vs survey respondents (P value)
Overall n(%)Survey respondents n(%)Survey non-respondents n(%)Respondents vs non-respondents (P value)
Size—NPR.04a<.01b
 Under $1 billion53 (26%)1 (1%)1 (2%)0 (0%)
 $1 billion-$4.9 billion97 (48%)43 (64%)25 (58%)15 (71%)
 $5 billion-$9.9 billion36 (18%)17 (25%)15 (35%)2 (10%)
 $10 billion and above17 (8%)6 (9%)2 (5%)4 (19%)
Size—number of hospitals.79b.93b
 4 or fewer hospitals62 (31%)16 (24%)12 (28%)4 (19%)
 5-10 hospitals57 (28%)23 (34%)14 (33%)6 (29%)
 11-19 hospitals41 (20%)15 (22%)9 (21%)6 (29%)
 20+ hospitals43 (21%)13 (19%)8 (19%)5 (24%)
Teaching status.07a.37b
 Major teaching122 (60%)51 (76%)29 (67%)18 (90%)
 Non-major teaching81 (40%)16 (24%)14 (33%)2 (10%)
Geographic region.16a.15b
 Northeast37 (18%)6 (9%)3 (7%)2 (10%)
 South72 (35%)21 (31%)13 (30%)8 (38%)
 Midwest55 (27%)27 (40%)15 (35%)10 (48%)
 West38 (19%)13 (19%)12 (28%)1 (5%)
 US Territory1 (0%)0 (0%)0 (0%)0 (0%)

Survey results

Among the key organizational goals for deploying AI, the most cited priority was reducing caregiver burden and improving satisfaction, with 72% (n = 31) of organizations ranking it as one of their top two goals. Both patient safety/quality and workflow efficiency/productivity followed closely, with 56% (n = 25) and 53% (n = 23) of organizations identifying them as their top two priorities respectively. Margin improvement/financial, patient/consumer experience, and market share/competitiveness were infrequently cited as the top 2 priorities (Figure 1). The priorities for respondents with NPR greater than $5B (n = 17) were statistically similar to those from health systems with NPR less than $5B (n = 26).

This bar chart describes the proportion of respondents citing each goal as the highest or second-second priority for deploying AI. The most frequently cited goal was ‘Caregiver burden/Satisfaction’, with 72% of respondents citing that as one of the top 2 goals. ‘Patient Safety/Quality’ was the second most frequently cited goal, with 58% of respondents citing as one of the top 2 goals. The 3rd most common cited goal was ‘Workflow efficiency/Productivity’, with 53% of respondents citing as one of the top 2 goals. Margin Improvement/Financial, Patient/Consumer Experience, Market Share/Competitiveness, were infrequently cited as one of the top 2 goals.
Figure 1.

Key goals for deploying AI. Proportion of respondents citing each goal as the highest or second-highest priority.

Health systems demonstrated significant variation in the extent of AI use case development and deployment, as illustrated in Table 2. Among the most commonly deployed use cases, 90% (n = 38) of respondents reported deploying AI for Imaging and Radiology in at least limited areas. This was followed by early detection of sepsis (67%, n = 27), ambient notes (60%, n = 26), risk of clinical deterioration (56%, n = 24), predicting the risk unplanned readmission (52%, n = 22), and in-basket automation (51%, n = 22) (Figure 2). Among the 38 organizations that have deployed the Imaging and Radiology use case in at least limited areas, 17 reported full deployment. Notably, while only 14% (n = 6) of organizations have fully deployed Ambient Notes and another 47% (n = 20) have implemented it in limited areas, 100% of respondents have at least begun developing or piloting this use case, which transforms conversations between the clinician and the patient into draft clinical documentation—this degree of uniform early adoption was not seen with any of the other 36 use cases.

This bar chart describes the top 7 most widely deployed AI use cases.
Figure 2.

Most commonly deployed AI use cases. Proportion of respondents who have deployed each use case in at least limited areas.

Table 2.

AI uses cases and their implementation stage.

AI use case categoryAI use caseNo activity n (%)Developing or piloting n (%)Deploying in limited areas n (%)Deployed fully n (%)
Clinical DocumentationAmbient Notes0 (0%)17 (40%)20 (47%)6 (14%)
Ambient Nursing13 (31%)28 (67%)1 (2%)0 (0%)
Drafting Care Plan Notes12 (29%)28 (68%)1 (2%)0 (0%)
Abstracting Data for Clinical Registries13 (31%)24 (57%)4 (10%)1 (2%)
Clinical Chart ReviewInpatient Chart Summarization9 (21%)30 (71%)3 (7%)0 (0%)
Ambulatory/Clinic Chart Summarization8 (19%)31 (74%)3 (7%)0 (0%)
Clinical Risk StratificationEarly Detection of Sepsis5 (12%)9 (21%)8 (19%)20 (48%)
Risk of Unplanned Admission8 (19%)12 (29%)10 (24%)12 (29%)
Risk of Clinical Deterioration4 (9%)15 (35%)8 (19%)16 (37%)
Risk of Patient Falls9 (21%)18 (43%)7 (17%)8 (19%)
DiagnosisImaging and Radiology2 (5%)2 (5%)21 (50%)17 (40%)
Digital Pathology16 (38%)16 (38%)8 (19%)2 (5%)
Other Diagnostic Tools11 (28%)14 (35%)12 (30%)3 (8%)
Engaging PatientsIn-Basket Automation3 (7%)18 (42%)16 (37%)6 (14%)
Adjust Reading Levels23 (55%)18 (43%)1 (2%)0 (0%)
Language Translation20 (48%)18 (43%)4 (10%)0 (0%)
Remote Patient Monitoring9 (21%)20 (47%)9 (21%)5 (12%)
Companion AI/Conversational Agents10 (24%)23 (55%)7 (17%)2 (5%)
Care Navigation15 (37%)25 (61%)1 (2%)0 (0%)
Patient Access and MarketingPredict Risk of Patient No Shows10 (24%)19 (45%)7 (17%)6 (14%)
Automate Patient Visit Scheduling16 (38%)22 (52%)3 (7%)1 (2%)
AI-Enabled Triage20 (48%)21 (50%)1 (2%)0 (0%)
Revenue CycleMedical Coding7 (17%)16 (38%)9 (21%)10 (24%)
Automate Utilization Review14 (34%)18 (44%)5 (12%)4 (10%)
Streamline Prior Authorization9 (22%)24 (59%)2 (5%)6 (15%)
Business Functions (not Rev Cycle)Forecast Census and Staffing8 (19%)20 (48%)5 (12%)9 (21%)
Optimize Patient Room Utilization15 (36%)19 (45%)5 (12%)3 (7%)
Optimize Operating Room Utilization11 (26%)17 (40%)8 (19%)6 (14%)
Supply Chain Tools18 (45%)16 (40%)4 (10%)2 (5%)
Recruiting and Human Resource Tools16 (40%)15 (38%)6 (15%)3 (8%)
Cash Forecasting (Treasury)25 (63%)12 (30%)2 (5%)1 (3%)
Automating AnalyticsData Analysis3 (8%)23 (58%)12 (30%)2 (5%)
AI-enabled Computer Coding7 (17%)23 (55%)9 (21%)3 (7%)
Supporting ResearchClinical Trials Automation21 (50%)21 (50%)0 (0%)0 (0%)
Drug Discovery25 (61%)16 (39%)0 (0%)0 (0%)
Genetics, Genomics19 (46%)22 (54%)0 (0%)0 (0%)
AI use case categoryAI use caseNo activity n (%)Developing or piloting n (%)Deploying in limited areas n (%)Deployed fully n (%)
Clinical DocumentationAmbient Notes0 (0%)17 (40%)20 (47%)6 (14%)
Ambient Nursing13 (31%)28 (67%)1 (2%)0 (0%)
Drafting Care Plan Notes12 (29%)28 (68%)1 (2%)0 (0%)
Abstracting Data for Clinical Registries13 (31%)24 (57%)4 (10%)1 (2%)
Clinical Chart ReviewInpatient Chart Summarization9 (21%)30 (71%)3 (7%)0 (0%)
Ambulatory/Clinic Chart Summarization8 (19%)31 (74%)3 (7%)0 (0%)
Clinical Risk StratificationEarly Detection of Sepsis5 (12%)9 (21%)8 (19%)20 (48%)
Risk of Unplanned Admission8 (19%)12 (29%)10 (24%)12 (29%)
Risk of Clinical Deterioration4 (9%)15 (35%)8 (19%)16 (37%)
Risk of Patient Falls9 (21%)18 (43%)7 (17%)8 (19%)
DiagnosisImaging and Radiology2 (5%)2 (5%)21 (50%)17 (40%)
Digital Pathology16 (38%)16 (38%)8 (19%)2 (5%)
Other Diagnostic Tools11 (28%)14 (35%)12 (30%)3 (8%)
Engaging PatientsIn-Basket Automation3 (7%)18 (42%)16 (37%)6 (14%)
Adjust Reading Levels23 (55%)18 (43%)1 (2%)0 (0%)
Language Translation20 (48%)18 (43%)4 (10%)0 (0%)
Remote Patient Monitoring9 (21%)20 (47%)9 (21%)5 (12%)
Companion AI/Conversational Agents10 (24%)23 (55%)7 (17%)2 (5%)
Care Navigation15 (37%)25 (61%)1 (2%)0 (0%)
Patient Access and MarketingPredict Risk of Patient No Shows10 (24%)19 (45%)7 (17%)6 (14%)
Automate Patient Visit Scheduling16 (38%)22 (52%)3 (7%)1 (2%)
AI-Enabled Triage20 (48%)21 (50%)1 (2%)0 (0%)
Revenue CycleMedical Coding7 (17%)16 (38%)9 (21%)10 (24%)
Automate Utilization Review14 (34%)18 (44%)5 (12%)4 (10%)
Streamline Prior Authorization9 (22%)24 (59%)2 (5%)6 (15%)
Business Functions (not Rev Cycle)Forecast Census and Staffing8 (19%)20 (48%)5 (12%)9 (21%)
Optimize Patient Room Utilization15 (36%)19 (45%)5 (12%)3 (7%)
Optimize Operating Room Utilization11 (26%)17 (40%)8 (19%)6 (14%)
Supply Chain Tools18 (45%)16 (40%)4 (10%)2 (5%)
Recruiting and Human Resource Tools16 (40%)15 (38%)6 (15%)3 (8%)
Cash Forecasting (Treasury)25 (63%)12 (30%)2 (5%)1 (3%)
Automating AnalyticsData Analysis3 (8%)23 (58%)12 (30%)2 (5%)
AI-enabled Computer Coding7 (17%)23 (55%)9 (21%)3 (7%)
Supporting ResearchClinical Trials Automation21 (50%)21 (50%)0 (0%)0 (0%)
Drug Discovery25 (61%)16 (39%)0 (0%)0 (0%)
Genetics, Genomics19 (46%)22 (54%)0 (0%)0 (0%)

Implementation stage with relative majority of responses bolded and highlighted in blue.

Table 2.

AI uses cases and their implementation stage.

AI use case categoryAI use caseNo activity n (%)Developing or piloting n (%)Deploying in limited areas n (%)Deployed fully n (%)
Clinical DocumentationAmbient Notes0 (0%)17 (40%)20 (47%)6 (14%)
Ambient Nursing13 (31%)28 (67%)1 (2%)0 (0%)
Drafting Care Plan Notes12 (29%)28 (68%)1 (2%)0 (0%)
Abstracting Data for Clinical Registries13 (31%)24 (57%)4 (10%)1 (2%)
Clinical Chart ReviewInpatient Chart Summarization9 (21%)30 (71%)3 (7%)0 (0%)
Ambulatory/Clinic Chart Summarization8 (19%)31 (74%)3 (7%)0 (0%)
Clinical Risk StratificationEarly Detection of Sepsis5 (12%)9 (21%)8 (19%)20 (48%)
Risk of Unplanned Admission8 (19%)12 (29%)10 (24%)12 (29%)
Risk of Clinical Deterioration4 (9%)15 (35%)8 (19%)16 (37%)
Risk of Patient Falls9 (21%)18 (43%)7 (17%)8 (19%)
DiagnosisImaging and Radiology2 (5%)2 (5%)21 (50%)17 (40%)
Digital Pathology16 (38%)16 (38%)8 (19%)2 (5%)
Other Diagnostic Tools11 (28%)14 (35%)12 (30%)3 (8%)
Engaging PatientsIn-Basket Automation3 (7%)18 (42%)16 (37%)6 (14%)
Adjust Reading Levels23 (55%)18 (43%)1 (2%)0 (0%)
Language Translation20 (48%)18 (43%)4 (10%)0 (0%)
Remote Patient Monitoring9 (21%)20 (47%)9 (21%)5 (12%)
Companion AI/Conversational Agents10 (24%)23 (55%)7 (17%)2 (5%)
Care Navigation15 (37%)25 (61%)1 (2%)0 (0%)
Patient Access and MarketingPredict Risk of Patient No Shows10 (24%)19 (45%)7 (17%)6 (14%)
Automate Patient Visit Scheduling16 (38%)22 (52%)3 (7%)1 (2%)
AI-Enabled Triage20 (48%)21 (50%)1 (2%)0 (0%)
Revenue CycleMedical Coding7 (17%)16 (38%)9 (21%)10 (24%)
Automate Utilization Review14 (34%)18 (44%)5 (12%)4 (10%)
Streamline Prior Authorization9 (22%)24 (59%)2 (5%)6 (15%)
Business Functions (not Rev Cycle)Forecast Census and Staffing8 (19%)20 (48%)5 (12%)9 (21%)
Optimize Patient Room Utilization15 (36%)19 (45%)5 (12%)3 (7%)
Optimize Operating Room Utilization11 (26%)17 (40%)8 (19%)6 (14%)
Supply Chain Tools18 (45%)16 (40%)4 (10%)2 (5%)
Recruiting and Human Resource Tools16 (40%)15 (38%)6 (15%)3 (8%)
Cash Forecasting (Treasury)25 (63%)12 (30%)2 (5%)1 (3%)
Automating AnalyticsData Analysis3 (8%)23 (58%)12 (30%)2 (5%)
AI-enabled Computer Coding7 (17%)23 (55%)9 (21%)3 (7%)
Supporting ResearchClinical Trials Automation21 (50%)21 (50%)0 (0%)0 (0%)
Drug Discovery25 (61%)16 (39%)0 (0%)0 (0%)
Genetics, Genomics19 (46%)22 (54%)0 (0%)0 (0%)
AI use case categoryAI use caseNo activity n (%)Developing or piloting n (%)Deploying in limited areas n (%)Deployed fully n (%)
Clinical DocumentationAmbient Notes0 (0%)17 (40%)20 (47%)6 (14%)
Ambient Nursing13 (31%)28 (67%)1 (2%)0 (0%)
Drafting Care Plan Notes12 (29%)28 (68%)1 (2%)0 (0%)
Abstracting Data for Clinical Registries13 (31%)24 (57%)4 (10%)1 (2%)
Clinical Chart ReviewInpatient Chart Summarization9 (21%)30 (71%)3 (7%)0 (0%)
Ambulatory/Clinic Chart Summarization8 (19%)31 (74%)3 (7%)0 (0%)
Clinical Risk StratificationEarly Detection of Sepsis5 (12%)9 (21%)8 (19%)20 (48%)
Risk of Unplanned Admission8 (19%)12 (29%)10 (24%)12 (29%)
Risk of Clinical Deterioration4 (9%)15 (35%)8 (19%)16 (37%)
Risk of Patient Falls9 (21%)18 (43%)7 (17%)8 (19%)
DiagnosisImaging and Radiology2 (5%)2 (5%)21 (50%)17 (40%)
Digital Pathology16 (38%)16 (38%)8 (19%)2 (5%)
Other Diagnostic Tools11 (28%)14 (35%)12 (30%)3 (8%)
Engaging PatientsIn-Basket Automation3 (7%)18 (42%)16 (37%)6 (14%)
Adjust Reading Levels23 (55%)18 (43%)1 (2%)0 (0%)
Language Translation20 (48%)18 (43%)4 (10%)0 (0%)
Remote Patient Monitoring9 (21%)20 (47%)9 (21%)5 (12%)
Companion AI/Conversational Agents10 (24%)23 (55%)7 (17%)2 (5%)
Care Navigation15 (37%)25 (61%)1 (2%)0 (0%)
Patient Access and MarketingPredict Risk of Patient No Shows10 (24%)19 (45%)7 (17%)6 (14%)
Automate Patient Visit Scheduling16 (38%)22 (52%)3 (7%)1 (2%)
AI-Enabled Triage20 (48%)21 (50%)1 (2%)0 (0%)
Revenue CycleMedical Coding7 (17%)16 (38%)9 (21%)10 (24%)
Automate Utilization Review14 (34%)18 (44%)5 (12%)4 (10%)
Streamline Prior Authorization9 (22%)24 (59%)2 (5%)6 (15%)
Business Functions (not Rev Cycle)Forecast Census and Staffing8 (19%)20 (48%)5 (12%)9 (21%)
Optimize Patient Room Utilization15 (36%)19 (45%)5 (12%)3 (7%)
Optimize Operating Room Utilization11 (26%)17 (40%)8 (19%)6 (14%)
Supply Chain Tools18 (45%)16 (40%)4 (10%)2 (5%)
Recruiting and Human Resource Tools16 (40%)15 (38%)6 (15%)3 (8%)
Cash Forecasting (Treasury)25 (63%)12 (30%)2 (5%)1 (3%)
Automating AnalyticsData Analysis3 (8%)23 (58%)12 (30%)2 (5%)
AI-enabled Computer Coding7 (17%)23 (55%)9 (21%)3 (7%)
Supporting ResearchClinical Trials Automation21 (50%)21 (50%)0 (0%)0 (0%)
Drug Discovery25 (61%)16 (39%)0 (0%)0 (0%)
Genetics, Genomics19 (46%)22 (54%)0 (0%)0 (0%)

Implementation stage with relative majority of responses bolded and highlighted in blue.

Among organizations that have started developing, piloting and deploying the 10 use cases categories, 53% of respondents (n = 21) reported a “high degree of success” with use of AI in Clinical Documentation, while 38% (n = 13) reported that for Clinical Risk Stratification use cases, and only 23% (n = 7) reported that for Revenue Cycle use cases. Only 19% (n = 5) of respondents who have started developing, piloting or deploying AI use cases for Clinical Diagnosis reported a high degree of success (Figure 3). Regarding measures of success for AI tools, Use (Uptake) of the AI tool was always measured by 74% (n = 31) of the respondents, while Health Equity/Disparities was always measured only by 17% (n = 7). Of note, Health Equity/Disparities was rarely measured and not measured by 10% (n = 4) and 20% (n = 8) of the respondents, respectively (Appendix S3).

This graph describes the successes reported among AI use case categories. Clinical Documentation was cited by 53% of respondents as enjoying a high degree of success, followed by Clinical Risk Stratification at 38%, Revenue Cycle at 23%, Automate Analytics at 19%, and Diagnosis at 19%.
Figure 3.

Successes reported among AI use case categories. Proportion of organizations reporting a high degree of success within each AI use case category, among those that have started developing, piloting, or deploying it.

Among the 43 respondents, 77% (n = 33) identified lack of AI tool maturity as the biggest or second-biggest barrier to AI development or deployment. This was followed by financial concerns, reported by 47% (n = 20), and regulatory or compliance uncertainty, cited by 40% (n = 17). In contrast, lack of clinician use or adoption (17%, n = 7), insufficient in-house expertise or technology (14%, n = 6), and lack of leadership support (7%, n = 3) were infrequently reported as one of the top 2 barriers (Figure 4). The top barriers were statistically similar between respondents with NPR greater than $5B vs those with NPR less than $5B.

This bar chart describes the key barriers for AI deployment. The top barrier identified was lack of AI tool maturity, with 77% of respondents identifying that as the biggest or second biggest barrier. Financial concerns came second, with 47% of respondents identifying that as either the biggest or second biggest barrier. Regulatory or compliance uncertainty came third.
Figure 4.

Key barriers to developing or deploying AI. Proportion of survey respondents (n = 43) who rated each barrier as the highest or second-biggest.

Discussion

Our 2024 survey assessed the state of AI adoption in leading US health systems less than 2 years following the release of ChatGPT.23 All health system respondents reported adoption activities in Ambient Notes, with over half of them reporting a high degree of success in use of AI for Clinical Documentation, virtually all of which uses generative AI. The remainder of the AI use cases assessed showed variable degrees of adoption, with clinical risk stratification models showing moderate adoption and only 38% of respondents reporting a high degree of success. Similarly, while most organizations have deployed AI in imaging and radiology, only 19% of them report a high degree of success in this area.

Although Ambient Notes solutions were in development by vendors before ChatGPT, recent advances in Generative AI have accelerated their market readiness (Petro J, Chief Technology Officer, Nuance Communications. Personal Communication). In comparison to other use cases that have been in development for much longer, such as imaging and radiology, Ambient Notes using generative AI appears to have swiftly crossed the “chasm” that often hinders new technology products from progressing beyond the early adopters (typically technology enthusiasts and visionaries) to being embraced by the pragmatic early majority.15 This apparent rate of rapid adoption may in part be influenced by concerns about clinician burnout following the COVID-19 pandemic,24 as most respondents identified relieving caregiver burnout/improving satisfaction and improving workflow efficiency/productivity as primary reasons for adopting AI. This rate of adoption compares favorably to that of CTs and MRIs in the 1970s and 1980s, respectively,25 technologies that are now integral to the practice of medicine. However, the swift adoption of Ambient Notes raises concerns about potential unintended consequences, such as technology affordability, workforce readiness, trainee usage, and patient perception, which remain unresolved. In addition, given the evolving landscape of generative AI, future research is necessary to evaluate its impact on organizational and clinical outcomes, including clinician productivity, clinician retention, and patient satisfaction.

Adoption of AI outside of ambient notes is still uneven and incomplete. While several risk stratification AI tools were reported as deployed by about half of our respondents, only approximately one-third of those developing or deploying those use cases cited a high degree of success. Although some of these risk stratification models have been available at relatively modest costs to health systems for more than 5 years, their accuracy, at least for sepsis prediction, has been called into recent question by several recent studies.26 This aligns with the observation made by our respondents that AI tool immaturity is a leading barrier to AI adoption. Busy clinicians who may be too distracted or alert-fatigued to act on these risk predictions could also explain the modest success reported by institutions deploying them. Similarly, only 19% of institutions reported a high degree of success for AI used in clinical diagnosis, despite predictions dating back to 2016 of AI replacing radiologists27,28 and a steady stream of AI-based imaging solutions undergoing FDA approval.29 These data indicate the need for developers of imaging and radiology AI to further collaborate with imaging-based clinicians to improve their workflow and decision-making quality.30,31

Given the current immaturity of AI tools and the need to balance speed with safety, it is clear that the heightened expectations associated with AI32 will need to be managed through robust evaluations to ensure AI deployed in healthcare is safe, effective and equitable.33 This is particularly relevant given mounting cost pressures and uncertainties on health systems,34 which might further constrain resources necessary to deploy, test, and monitor AI and in turn jeopardize patient safety and the long-term value of AI. Fit-for-purpose and practical evaluations are necessary before an AI tool is deployed, during its implementation, and in an ongoing way to ensure that their benefits continue to justify ongoing costs and risks. In addition, local lessons learned throughout the life-cycle of AI deployment will be highly valuable given the varying adoption pace, implementation expertise, and sociotechnical factors across organizations.35 Traditional mechanisms for sharing these evaluations results and lessons learned, including academic journals and conferences may not meet the rapidly evolving needs of the health AI industry, and are subject to biases that tend to favor dissemination of positive findings. Proposals from Coalition for Health AI, VALID AI, Health AI Partnership and others36 are advocating for networks to distribute the evaluation work and disseminate the findings. While this healthcare provider-based approach will bring value, additional measures are needed. First, EHR and AI vendors should strive towards common deployment platforms to enable cost-effective AI tool integration at relevant workflow and decision-making points. This approach has proven effective with “traditional” tools such as medication decision support and standardized medical terminology. Without such an approach, individual clinicians might face multiple user-interfaces, adding cognitive load and potentially diminishing each other’s effectiveness. Second, local configuration decisions, usage policies, implementation strategies, evaluation tools, and user-support resources should be widely shared, perhaps utilizing Patient Safety Organizations to address confidentiality concerns. Such resources would accelerate the maturation speed for healthcare organizations and vendors. Third, professional and membership organizations should encourage members to publish results of programmatic outcome evaluations and case studies so that lessons learned could be applied at scale.

Many use cases assessed in this survey, particularly those in the fields of administration, patient engagement, marketing, and clinical research, are still in the early stages of adoption among our study population. With ongoing cost pressures facing health systems and continued investment in AI, widespread success may become achievable. However, the challenges faced by more mature clinical AI use cases will likely apply here as well. Therefore, it is crucial for national bodies focusing on responsible use of health AI to share their expertise with developers and implementers of these early use cases, many of whom may only be beginning to engage in the responsible AI conversation. Similarly, measures around equity and bias are not used by health systems consistently to monitor the performance of AI tools, in spite of strong recommendations by many health AI consortia to focus on algorithmic bias.33,37,38 This suggests the need for the dissemination of practical approaches that would allow local AI oversight bodies to monitor the differential performance and impact of AI tools across vulnerable populations.

Our study has several limitations. Our study presents only a snapshot of the state of AI adoption across 43 large health systems as of the Fall of 2024. Most SI members represent health systems with NPR of at least $1B and may have access to more resources than the average US health system. In addition, since SI members are actively engaged in quality improvement and learning, their level of AI adoption may be more advanced compared to their average peer. Nonetheless, their activities could indicate what other health systems might focus on in the near future, and challenges faced by our respondents may also affect others unless addressed at the national level. The preponderance of clinicians among our survey respondents could have led us to underestimate the adoption of AI in the business function and revenue cycle domains, and our findings may not generalize immediately to all healthcare organizations in the United States. However, the diverse geographical locations and types of organizations represented by our survey respondents enhance the relevance and applicability of our conclusions. Overall, our study underscores that AI tools, except for Ambient Notes, are deployed inconsistently and with varying degrees of success. As such, efforts to share evaluation results and lessons learned will contribute towards a common good across multiple stakeholders.

Author contributions

Eric G. Poon (Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Resources, Writing—original draft, Writing—review & editing), Christy H. Lemak (Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Supervision, Validation, Writing—review & editing), Juan C. Rojas (Investigation, Methodology, Writing—original draft, Writing—review & editing), Janet Guptill (Conceptualization, Funding acquisition, Investigation, Resources, Writing—review & editing), and David Classen (Conceptualization, Formal analysis, Investigation, Methodology, Supervision, Writing—review & editing)

Supplementary material

Supplementary material is available at Journal of the American Medical Informatics Association online.

Funding

This research received no specific grant from any funding agency in the public, commercial or not-for-profit sectors.

Conflicts of interests

Dr Rojas is supported by NIH/NIDA R01DA051464 and the Robert Wood Johnson Foundation and has received consulting fees from Truveta. Dr Poon has received consulting fees from Triomics.

Data availability

Data available upon request.

Findings

In this cross-sectional survey of 43 US health systems, all respondents reported activity related to Ambient Notes, a generative AI tool for clinical documentation. Other select AI use cases such as imaging/radiology and clinical risk stratification are adopted by many, though successes are modest. Immature AI tools and financial concerns were the most frequently cited barriers.

Meaning

Generative AI tools like Ambient Notes show significant promise and rapid adoption is occurring. However, other AI tools continue to be adopted unevenly. Challenges such as AI tool immaturity and financial constraints must be overcome to ensure broad adoption and impact.

References

1

Ullem
BD
,
Hatlie
MJ
,
Lounsbury
O.
 
AI: promise or peril for patient safety
.
J Patient Saf
.
2025
;
21
:
35
-
37
.

2

Singh
H
,
Sittig
DF
,
Classen
DC.
 
Maximizing the ability of health IT and AI to improve patient safety
.
JAMA Intern Med
.
2025
;
185
:
10
-
12
.

3

Boussina
A
,
Shashikumar
SP
,
Malhotra
A
, et al.  
Impact of a deep learning sepsis prediction model on quality of care and survival
.
NPJ Digit Med
.
2024
;
7
:
14
.

4

Shafi
S
,
Parwani
AV.
 
Artificial intelligence in diagnostic pathology
.
Diagn Pathol
.
2023
;
18
:
109
.

5

Topol
EJ.
 
As artificial intelligence goes multimodal, medical applications multiply
.
Science
.
2023
;
381
:
adk6139
.

6

Rajpurkar
P
,
Chen
E
,
Banerjee
O
,
Topol
EJ.
 
AI in health and medicine
.
Nat Med
.
2022
;
28
:
31
-
38
.

7

Liu
TL
,
Hetherington
TC
,
Stephens
C
, et al.  
AI-powered clinical documentation and clinicians' electronic health record experience: a nonrandomized clinical trial
.
JAMA Netw Open
.
2024
;
7
:
e2432460
.

8

Bundy
H
,
Gerhart
J
,
Baek
S
, et al.  
Can the administrative loads of physicians be alleviated by AI-facilitated clinical documentation?
 
J Gen Intern Med
.
2024
;
39
:
2995
-
3000
.

9

Galloway
JL
,
Munroe
D
,
Vohra-Khullar
PD
, et al.  
Impact of an artificial intelligence-based solution on clinicians' clinical documentation experience: initial findings using ambient listening technology
.
J Gen Intern Med
.
2024
;
39
:
2625
-
2627
.

10

Shreya
SD-SA
,
Ma
S.
 
Ambient AI scribes: physician burnout and perspectives on usability and documentation burden
.
J Am Med Inform Ass
.
2025
;
32

11

Tierney
AGG
,
Hoberman
B
,
Mattern
B
, et al.  
Ambient artificial intelligence scribes to alleviate the burden of clinical documentation
.
NEJM Catal Innov Care Deliv
.
2024
;
5
.

12

Ma
SPLA
,
Shah
SJ
,
Smith
M
, et al.  
Ambient AI scribes: utilization and impact on documentation time
.
J Am Med Inform Assoc
.
2025
;
32
:
381
-
385
.

13

Flare_Capital_Partners
. Where do Healthcare Budgets Match AI Hype? A 10-Year Lookback of Funding Data. Medium. 9/9/24. Accessed April 25, 2025. https://flarecapitalpartners.medium.com/where-do-healthcare-budgets-match-ai-hype-a-10-year-lookback-of-funding-data-783d52010e29

14

Rogers
E.
 
Diffusion of Innovations
. 5th ed.
Simon and Schuster
;
2003
.

15

Moore
GA.
 
Crossing the Chasm: Marketing and Selling Disruptive Products to Mainstream Customers
. 3rd ed.
Harper Business, an imprint of HarperCollins Publishers
;
2014
.

16

Sittig
DF
,
Singh
H.
 A sociotechnical approach to electronic health record related safety.
Key Advances in Clinical Informatics.
 
Academic Press
;
2017
:
197
-
216
.

17

Rojas
JC
,
Rohweder
G
,
Guptill
J
,
Arora
VM
,
Umscheid
CA.
 
Predictive analytics programs at large healthcare systems in the USA: a national survey
.
J Gen Intern Med
.
2022
;
37
:
4015
-
4017
.

18

Guleria
S
,
Guptill
J
,
Kumar
I
, et al.  
Artificial intelligence integration in healthcare: perspectives and trends in a survey of U.S. health system leaders
.
BMC Digital Health
.
2024
;
2
:80.

19

Matheny
ME
,
Whicher
D
,
Thadaney Israni
S.
 
Artificial intelligence in health care: a report from the national academy of medicine
.
JAMA
.
2020
;
323
:
509
-
510
.

20

Nong
P
,
Adler-Milstein
J
,
Apathy
NC
,
Holmgren
AJ
,
Everson
J.
 
Current use and evaluation of artificial intelligence and predictive models in US hospitals
.
Health Aff (Millwood)
.
2025
;
44
:
90
-
98
.

21

AAPOR
. Best Practices for Survey Research. Accessed April 25, 2025. https://aapor.org/standards-and-ethics/best-practices/

22

Definitive Healthcare L. HospitalView
. Accessed April 25, 2025. https://www.definitivehc.com/data-products/hospital-view

23

Wikipedia
. ChatGPT. Accessed April 25, 2025. https://en.wikipedia.org/wiki/ChatGPT

24

Lluch
C
,
Galiana
L
,
Domenech
P
,
Sanso
N.
 
The impact of the COVID-19 pandemic on burnout, compassion fatigue, and compassion satisfaction in healthcare personnel: a systematic review of the literature published during the first year of the pandemic
.
Healthcare (Basel)
.
2022
;
10
:364.

25

Steinberg
EP
,
Sisk
JE
,
Locke
KE.
 
X-ray CT and magnetic resonance imagers. Diffusion patterns and policy issues
.
N Engl J Med
.
1985
;
313
:
859
-
864
.

26

Kamran
F
,
Tjandra
D
,
Heiler
A
, et al.  
Evaluation of sepsis prediction models before onset of treatment
.
NEJM AI
.
2024
;
1
.

27

AI, Radiology and the Future of Work
.
The Economist
 
2018
. Accessed April 25, 2025.

28

Obermeyer
Z
,
Emanuel
EJ.
 
Predicting the future—big data, machine learning, and clinical medicine
.
N Engl J Med
.
2016
;
375
:
1216
-
1219
.

29

Muralidharan
V
,
Adewale
BA
,
Huang
CJ
, et al.  
A scoping review of reporting gaps in FDA-approved AI medical devices
.
NPJ Digit Med
.
2024
;
7
:
273
.

30

Langlotz
CP.
 
Will artificial intelligence replace radiologists?
 
Radiol Artif Intell
.
2019
;
1
:
e190058
.

31

Cacciamani
GE
,
Sanford
DI
,
Chu
TN
, et al.  
Is artificial intelligence replacing our radiology stars? not yet!
 
Eur Urol Open Sci
.
2023
;
48
:
14
-
16
.

32

Gartner
. Explore Beyond GenAI on the 2024 Hype Cycle for Artificial Intelligence. 11/11/2024. Accessed April 25, 2025. https://www.gartner.com/en/articles/hype-cycle-for-artificial-intelligence

33

Coaltion_for_Health_AI
. Assurance Standards Guide. Accessed April 25, 2025. https://chai.org/wp-content/uploads/2024/07/CHAI-Assurance-Standards-Guide-6-26-2024.pdf

34

Trebes
NHM
,
Landis
J.
17 things CEOs need to know in 2025. The Advisory Board,
2025
. Accessed April 25, 2025. https://www.advisory.com/topics/state-of-healthcare-industry/ceo-briefing-2025

35

Sittig
DF
,
Singh
H.
 
Recommendations to ensure safety of AI in real-world clinical care
.
JAMA
.
2025
;
333
:
457
-
458
.

36

Shah
NH
,
Halamka
JD
,
Saria
S
, et al.  
A nationwide network of health AI assurance laboratories
.
JAMA
.
2024
;
331
:
245
-
249
.

37

Kim
JY
,
Hasan
A
,
Kellogg
KC
, et al.  
Development and preliminary testing of Health Equity Across the AI Lifecycle (HEAAL): a framework for healthcare delivery organizations to mitigate the risk of AI solutions worsening health inequities
.
PLOS Digit Health
.
2024
;
3
:
e0000390
.

38

Garba-Sani
ZF-RC
,
Essien
A
,
Yracheta
JM.
 
A.C.C.E.S.S. AI: a new framework for advancing health equity in health care AI
.
Health Affairs Forefront
.
2024
.

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.

Supplementary data