Published on in Vol 4 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/72535, first published .
The Perceived Roles of AI in Clinical Practice: National Survey of 941 Academic Physicians

The Perceived Roles of AI in Clinical Practice: National Survey of 941 Academic Physicians

The Perceived Roles of AI in Clinical Practice: National Survey of 941 Academic Physicians

Original Paper

1Department of Neurosurgery, David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, CA, United States

2David Geffen School of Medicine, University of California, Los Angeles, Los Angeles, CA, United States

Corresponding Author:

Joel Beckett, MD

Department of Neurosurgery

David Geffen School of Medicine

University of California, Los Angeles

300 Stein Plaza Driveway

Suite 420

Los Angeles, CA, 90095

United States

Phone: 1 310 776 3688

Email: joel.beckett@cshs.org


Background: Artificial intelligence (AI) and machine learning models are frequently developed in medical research to optimize patient care, yet they remain rarely used in clinical practice.

Objective: This study aims to understand the disconnect between model development and implementation by surveying physicians of all specialties across the United States.

Methods: The present survey was distributed to residency coordinators at Accreditation Council for Graduate Medical Education–accredited residency programs to disseminate among attending physicians and resident physicians affiliated with their academic institution. Respondents were asked to identify and quantify the extent of their training and specialization, as well as the type and location of their practice. Physicians were then asked follow-up questions regarding AI in their practice, including whether its use is permitted, whether they would use it if made available, primary reasons for using or not using AI, elements that would encourage its use, and ethical concerns.

Results: Of the 941 physicians who responded to the survey, 384 (40.8%) were attending physicians and 557 (59.2%) were resident physicians. The majority of the physicians (651/795, 81.9%) indicated that they would adopt AI in clinical practice if given the opportunity. The most cited intended uses for AI were risk stratification, image analysis or segmentation, and disease prognosis. The most common reservations were concerns about clinical errors made by AI and the potential to replicate human biases.

Conclusions: To date, this study comprises the largest and most diverse dataset of physician perspectives on AI. Our results emphasize that most academic physicians in the United States are open to adopting AI in their clinical practice. However, for AI to become clinically relevant, developers and physicians must work synergistically to design models that are accurate, accessible, and intuitive while thoroughly addressing ethical concerns associated with the implementation of AI in medicine.

JMIR AI 2025;4:e72535

doi:10.2196/72535

Keywords



Background

Artificial intelligence (AI) is a topic of significant interest in recent literature. It has garnered considerable attention as a potential tool to aid in detecting, diagnosing, and managing diseases and demonstrated efficacy in detecting diabetic retinopathy, diagnosing skin cancers, and even predicting patients’ surgical candidacy [1-5]. As artificial intelligence grows increasingly capable, one might expect an associated increase in the clinical uses of such programs; however, AI models are rarely harnessed in medicine today outside the realm of research [6,7].

The disconnect between model development and deployment has been examined previously [8-10]. The lack of clinical implementation of AI and machine learning techniques often stems from a lack of physician trust in model accuracy and the difficulties associated with integrating models into existing clinical paradigms, such as electronic health record platforms [11]. Current articles on the topic are often authored by field experts and frequently present subjective or anecdotal findings rather than a systematic exploration of sentiments toward AI in medicine at large. For AI to realize its potential and become a clinically relevant tool, it is essential to understand how physicians foresee its use and integration.

Objectives

This study aims to quantify key factors relevant to integrating AI into clinical medicine, including physician adoption and use, incentivization strategies, and ethical considerations. To achieve this, we developed and distributed an anonymous survey using the Qualtrics platform (Qualtrics International Inc, Provo, UT) to resident and attending physicians associated with academic institutions across all medical specialties throughout the United States. The diversity and volume of responses to this survey yielded, to our knowledge, the largest and most comprehensive dataset to date of physician perspectives on the implementation of AI. In doing this, we aim to analyze use patterns across a broad range of physicians to better understand the gap between AI model development and clinical implementation.


Ethical Considerations

This study was conducted according to the Declaration of Helsinki, which sets ethical principles for medical research involving human participants [12]. As the present study involved a minimal-risk, de-identified survey, formal Institutional Review Board (IRB) review was not required.

Survey Instrument

The survey was initially conceptualized and developed by 2 authors. It was subsequently reviewed independently by 5 senior coauthors; after a group discussion, relevant modifications were incorporated, and all contributing authors approved the final version of the survey. The survey was piloted at our home institution before being distributed nationally. The final 14-item survey included 7 demographic questions (eg, training status, specialty, years in practice, and practice setting), followed by 7 AI-related questions assessing awareness, intent to use, perceived use cases, incentives, and ethical concerns. Conditional branching logic was used to display specific follow-up questions based on prior responses (Figure 1) to obtain a deeper understanding of the rationale behind response selections. The complete survey is provided in Multimedia Appendix 1. The survey was designed and distributed in accordance with the CHERRIES (Checklist for Reporting Results of Internet E-Surveys) checklist [13].

Figure 1. Flowchart of survey format. Participants were displayed subquestions depending on their selection of the questions shown here, presented as branches. ACGME: Accreditation Council for Graduate Medical Education; AI: artificial intelligence.

Study Design and Distribution

The survey was distributed to residency program coordinators representing 18 medical and surgical specialties in the United States: anesthesiology, dermatology, emergency medicine, family medicine, general surgery, internal medicine, neurology, neurosurgery, obstetrics and gynecology, ophthalmology, orthopedic surgery, otolaryngology, pediatrics, plastic surgery, psychiatry, radiation oncology, diagnostic radiology, and urology. These specialties were chosen to represent a diverse cross-section of core disciplines in US residency training.

This open survey used a nonprobability referral sampling design, relying on program coordinators from Accreditation Council for Graduate Medical Education (ACGME)–accredited residency programs to distribute it to associated resident physicians and attending physicians. Program coordinators’ contact information was manually obtained from the publicly accessible ACGME program search website. If an email address was invalid, the residency program’s website was queried to obtain an updated email address. Correspondence containing the Qualtrics survey was first sent to program coordinators on January 11, 2023, by 1 author. The same author sent follow-up correspondence at 1-month and 2-month intervals after the initial outreach. Program coordinators who confirmed survey distribution were excluded from follow-up communications. The survey remained open for responses until November 3, 2023, allowing for an 11-month data collection period.

Survey participation was voluntary, anonymous, and without incentives. However, cookies, approximate location, and IP address were automatically collected using the Qualtrics platform to identify and remove duplicate submissions. Only complete responses were included in this analysis. As surveys were distributed to program coordinators, who subsequently disseminated them, the true survey response rate was not documented.

Statistical Analysis

Statistical analyses were conducted using the SciPy (version 1.11.4) and Statsmodels Python libraries. For group comparisons, chi-square tests of independence were used to evaluate differences in categorical variables. Phi coefficients were calculated as a measure of effect size for 2x2 contingency tables, and Cramér V was reported for larger contingency tables. A P value of <.05 was considered statistically significant. For proportion comparisons, 95% CIs were calculated using the Wilson score method. All tests were 2-tailed and conducted on complete-case data.


Survey Respondents and Demographics

Of the 4405 programs contacted, coordinators from 645 programs (14.6%) confirmed by email that they had distributed the survey. Figure 1 summarizes the number of responses to the questions and responses within conditional branch points. Table 1 presents participant demographics. A total of 941 physicians completed the survey, including 384 (40.8%) attending physicians and 557 (59.2%) resident physicians. Among attending physicians, the mean time in practice was 15.43 (SD 11.49) years. Responses were received from 43 US states and Washington, DC. Among respondents, 23.7% (225/951) practiced in the West, 21.1% (201/951) in the Midwest, 6.8% (65/951) in the Southwest, 22.9% (219/951) in the Southeast, and 25.3% (241/951) in the Northeast (Figure 2). Of the 941 participants, 10 (1.1%) reported practicing in more than 1 state; in these instances, 1 point was assigned to each state listed. No responses were recorded from Alaska, Hawaii, Idaho, Maine, Montana, Oregon, or Wyoming. Regarding practice setting, 83.2% (768/923) practiced in an urban setting, 11.8% (109/923) in a rural setting, and 5% (46/923) in a suburban setting (Table 1). Of the 923 respondents, 22 (2.4%) used the other free-response option. On manual review of these responses, none could be definitively classified into the aforementioned practice-setting categories; therefore, these were excluded from setting comparisons. Concerning the type of practice, 51.8% (494/954) worked in a university practice, 7% (67/954) in a government practice, 35% (334/954) in a hospital-owned practice, and 6.2% (59/954) in a physician-owned practice (Table 1). Of the 954 respondents, 67 (7%) chose other and were therefore excluded from type-of-practice comparisons because they did not align with the other 4 categories.

Figure 2. Geographic heat map of locations where physician survey participants practiced. Geographic regions were divided based on National Geographic Education outline maps. Physician submissions from California represented the largest group (n=178), while those from Delaware, Nebraska, North Dakota, and South Dakota represented the smallest group (n=1 each). Other responses not shown on the heat map include Washington DC (n=4). No responses were recorded from Alaska, Hawaii, Idaho, Maine, Montana, Oregon, or Wyoming.
Table 1. Participant demographics.
CharacteristicsParticipants, n (%)
Physician status (n=941)

Attending physician384 (40.8)

Resident physician557 (59.2)
Specialty (n=941)

Anesthesiology58 (6.2)

Dermatology44 (4.7)

Emergency medicine91 (9.7)

Family medicine84 (8.9)

General surgery73 (7.8)

Internal medicine147 (15.6)

Neurology22 (2.3)

Neurosurgery32 (3.4)

Obstetrics and gynecology64 (6.8)

Ophthalmology39 (4.1)

Orthopedic surgery37 (3.9)

Otolaryngology37 (3.9)

Pediatrics41 (4.4)

Plastic surgery21 (2.2)

Psychiatry80 (8.5)

Radiation oncology30 (3.2)

Diagnostic radiology20 (2.1)

Urology21 (2.2)
Type of specialty (n=941)

Medicala617 (65.6)

Surgical324 (34.4)
Practice setting (n=923)

Urban768 (83.2)

Rural109 (11.8)

Suburban46 (5)

Other22 (2.4)
Type of practice (n=954)

University494 (51.8)

Government67 (7)

Hospital owned334 (35)

Physician owned59 (6.2)

Other67 (7)
Region (n=951)

West225 (23.7)

Midwest201 (21.1)

Southwest65 (6.8)

Southeast219 (23)

Northeast241 (25.3)

aAnesthesiology, dermatology, emergency medicine, family medicine, internal medicine, neurology, pediatrics, psychiatry, radiation oncology, diagnostic radiology, and oncology.

The Use of AI in Clinical Medicine

Most of the participants (795/941, 84.5%) were either unaware whether using AI was permissible in their practice or could not use AI in their health care setting. Figure 3 summarizes the responses to this survey question by physician specialty. Of the 795 respondents, 651 (81.9%) stated that they would use AI if their health care setting permitted it. Table 2 shows comparisons in responses to this question stratified by training status (resident physician vs attending physician), practice setting, and type of practice. Significantly more resident physicians (407/481, 84.6%, 95% CI 81.1%-87.6%) were open to using AI than attending physicians (244/314, 77.7%, 95% CI 72.8%-82%; χ2=5.7; φ=0.08; P=.02). Similarly, a significantly higher proportion of urban practitioners (538/646, 83.3%) were open to using AI than rural practitioners (68/94, 72.3%, 95% CI 62.6%-80.4%; χ2=5.9; φ=0.09; P=.03). There was no statistically significant difference in willingness to use AI between types of practice (χ2=0.2; Cramér V=0.02; P=.97).

Figure 3. Artificial intelligence (AI) use in clinical practice by specialty. A total of 941 physicians from 18 specialties indicated whether their health care system allowed AI use in clinical practice. This was the first survey question displayed to all participants, and they were required to select only from the 3 options shown.
Table 2. Willingness to adopt artificial intelligence (AI) stratified by participant characteristics.
CharacteristicsWilling to adopt AI, n (%)Not willing to adopt AI, n (%)
Physician status

Attending (n=314)244 (77.7)70 (22.3)

Resident (n=481)407 (84.6)74 (15.4)
Practice settinga

Urban (n=646)538 (83.3)108 (16.7)

Suburban (n=39)31 (79.5)8 (20.5)

Rural (n=94)68 (72.3)26 (27.7)
Type of practicea

Government (n=61)51 (83.6)10 (16.4)

Hospital owned (n= 270)225 (83.3)45 (16.7)

Physician owned (n= 52)43 (82.7)9 (17.3)

University (n= 412)338 (82.0)74 (18)

aPhysicians were able to select multiple options if they practiced in more than 1 state.

Data were also collected regarding how physicians would use AI if given the opportunity and explored factors influencing their present use or nonuse (Figure 4 A-C). Among the respondents who indicated that they would use AI if allowed, the most common applications were risk stratification (513/651,78.8%), image analysis or segmentation (352/651, 54.1%), and disease prognosis (295/651, 45.3%). When stratified by specialty, medicine respondents most often indicated interest in using AI for risk stratification (316/422, 74.9%), followed by image analysis or segmentation (197/422, 46.7%) and clinical decision-making (190/422, 45%), whereas surgical respondents more frequently selected risk stratification (197/229, 86%), followed by image analysis or segmentation (155/229, 67.7%) and disease prognosis (124/229, 54.1%) (Table 3). Although surgical decision-making was the least frequently selected use of AI overall (98/651, 15.1%), surgeons (66/229, 28.8%, 95% CI 23.3%-35%) were significantly more likely than nonsurgeons (32/422, 7.6%, 95% CI 5.4%-10.5%; χ2=50.7; φ=0.28; P<.001) to indicate interest in this application.

Figure 4. Recorded responses to survey questions. Participants were able to select multiple options. (A) Current reported uses of artificial intelligence (AI) in clinical practice (n=77). (B) Desired uses of AI if it were permitted (n=651). (C) Reported reasons given by physicians (n=69) for not using AI in clinical practice. (D) Reported reasons for not using AI if it were permitted (n=144). Physicians who were unable to use AI or who were unsure whether it was permitted in their practice and selected that they would not use AI were displayed this question (n=144). (E) Stated incentives that would increase the physicians’ use of AI in clinical practice (n=941). (F) Stated reservations and ethical concerns regarding AI use in clinical practice (n=941). CPT: current procedural terminology.
Table 3. Intended uses of artificial intelligence (AI).
Intended uses of AIMedicine, n=422, n (%)Surgery, n=229, n (%)Overall (n=651), n (%)
Risk stratification316/422 (74.9)197/229 (86)513 (78.8)
Image analysis or segmentation197/422 (46.7)155/229 (67.7)352 (54.1)
Disease prognosis171/422 (40.5)124/229 (54.1)295 (45.3)
Clinical decision-making190/422 (45)100/229 (43.7)290 (44.5)
Providing objective solutions186/422 (44.1)83/229 (36.2)269 (41.3)
Diagnosis179/422 (42.4)84/229 (36.7)263 (40.4)
Surgical decision-making32/422 (7.6)66/229 (28.8)98 (15.1)

Opinions Regarding the Utility of and Reservations About AI

The most common incentives for wanting to use AI (Figure 4 A-C, E) were proof of utility or success of AI in clinical practice (763/941, 81.1%) and better integration into the clinical workflow (683/941, 72.6%). Among the stated reservations about using AI (Figure 4F), the most commonly listed was clinical errors made by AI (742/941, 78.9%), whereas the second most common was a concern about AI replicating and perpetuating human biases 583/941, 62%).


Principal Findings

This study describes the perspectives of academic physicians on adopting AI in medicine, based on survey responses from 941 physicians across all 18 specialties accredited by the ACGME. Our results demonstrate that, broadly, physicians at academic institutions in the United States are interested in harnessing AI in their clinical practice, particularly for risk stratification, image analysis or segmentation, and disease prognosis. However, this eagerness to adopt AI is tempered by significant and pertinent ethical concerns regarding perpetuating human biases. To our knowledge, this study represents the largest and most diverse dataset to date of physician perspectives on AI.

Given the wide-ranging potential of AI, the term itself carries a high degree of ambiguity in interpretation, both in this study and society at large. [14]; for instance, the term artificial intelligence can be applied to several technologies, ranging from virtual assistants to robotic navigation systems [15]. Essentially, AI refers to the capability of an automated machine to replicate human intelligence, such as understanding and interpreting data to make decisions in ways that approximate human reasoning [16]. Our observation, that 51.6% (486/941) of the respondents were unsure about whether their health care system permitted the use of AI techniques, may be explained by this lack of a concrete definition or understanding of what the term artificial intelligence constitutes. Solidifying this consensus would address at least some of this observed discrepancy [17].

Despite our imperfect understanding of AI in medicine, this survey demonstrates that academic physicians are eager to adopt it in clinical practice. The majority of the physicians (651/795, 81.9%) indicated that they would use AI if given the opportunity. Among the options provided, risk stratification was cited as the most frequent intended use. Patients regularly want to know about their prognosis, how long they might live with a given disease, or their chances of experiencing a complication during a procedure [18]. The ability to input patient-specific data into an algorithm, integrate AI-derived insights, and receive individualized prognostic data would provide immense value for patients in terms of setting expectations and maintaining transparency [19].

The least frequently selected intended use of AI was surgical decision-making (98/651, 15.1%). This is partly a result of fewer surgeons comprising the cohort of respondents (324/941, 34.4%). Unsurprisingly, a significantly greater proportion of surgical respondents indicated that they might use AI to help with surgical decision-making (66/229, 28.8%; P<.001), but the overall proportion still remains relatively low. The results of this study suggest that surgeons foresee using AI in several areas of medicine more readily than in surgical decision-making (Table 3), including risk stratification (197/229, 86%) and image analysis or segmentation (155/229, 67.7%). This observation may in part be due to the phrase “surgical decision-making,” which could be interpreted to mean second-to-second decisions during an active procedure or surgery, in which case it might seem difficult to imagine a real-world application of AI that does not interfere with the pace or rhythm of surgery [20]. Moreover, it could be due to a lack of imagination as to what an AI model can process as input, as preoperative planning often involves synthesizing various information modalities. Without further insight into the respondents’ rationale, we can only speculate why only a few participants (98/651, 15.1%) expressed interest in using AI for surgical decision-making. Ongoing work aims to clarify why, among the options presented, surgical decision-making was perceived as the least applicable AI integration and how this perception might be addressed in future model development.

In our survey, 81.1% (763/941) of the physicians indicated that increased proof of the success of new models in clinical practice would encourage their adoption of AI techniques. This speaks to an important, whether real or perceived, drawback of AI adoption in the current health care system. Realizing the success of AI may not be sufficient for widespread adoption, as 72.6% (683/941) of that physicians felt that improving the integration between AI algorithms and clinical workflow was another important incentive for its use. This should come as no surprise; physicians are unlikely to engage with a technology that is neither trustworthy nor easy to use [21,22].

One concern about the use of AI in medicine expressed by all respondents involves the ethical considerations of using AI algorithms. Respondents seemed most concerned about the possibility of introducing nonhuman error. Such a concern is understandable but fails to appreciate the alternative status quo in which human errors abound [23]. Rather than holding AI models to the impossible standard of perfection, these models should be compared to the current paradigm in which human beings practice medicine and make errors, as the goal of AI should be to provide improved rather than perfect medical treatment [24].

Another focal concern regarding implementing AI in medical practice among respondents was the potential for replicating human biases (583/941, 62%). Several articles have cited bias as an issue that must be resolved before the clinical integration of AI [25,26]. A landmark paper by Obermeyer et al [27] found that racial bias was deeply embedded in a widely used medical algorithm for predicting health care costs. At a given risk level, Black patients were significantly sicker than White patients because the algorithm used health care costs rather than illness severity to assign risk, significantly reducing the likelihood that Black patients were identified as needing additional care. One may argue that health care expenses were—in retrospect—a poor and biased indication of health; however, such bias is well known to persist even in models trained on actual clinical decision-making trends [28,29]. This is largely due to the lack of diversity in age, racial, ethnic, and geographic distributions in the datasets used to train medical machine learning models [30,31]. As it is unknown whether a model’s prognostic accuracy applies equally—or at all—to populations that differ from the training dataset, applying models to populations that have not been validated can generate inaccurate results that mislead physicians [32]; for example, a study testing 5 most commonly used medical machine learning algorithms demonstrated significantly better performance in White patients than in other racial groups despite being trained on a large, publicly available database [33,34]. Models trained on homogeneous data—or those that fail to report the demographic data on which they were trained—are not appropriate for clinical practice, as physicians will be unable to ascertain how deploying such models in clinical practice may impact the care of the patient they are treating. This limits the application of AI in medicine and speaks to the need for both more diverse datasets and data transparency in medical machine learning. These concerns should be thoroughly addressed before widespread clinical adoption of all machine learning models.

To our knowledge, this study represents the largest and most diverse assessment to date of attitudes toward implementing AI in clinical practice. Overall, our results demonstrate that academic physicians are open to implementing AI in clinical practice, especially after specific concerns regarding model evaluation and integration are addressed. However, AI models are vulnerable to any bias present in the data on which they are trained, and thus, the success of AI in medicine will rely as much on the quality as the quantity of training data [35]. For AI to become truly beneficial, we must harness the breadth of objective patient-specific data that are becoming increasingly available, such as through wearable devices, digital phenotyping, and other objective metrics. Even for successful models, widespread adoption may be an uphill battle. We expect that as physicians become more familiar with AI and the general population grows accustomed to its use in other domains, the integration of AI into clinical practice is likely to follow.

Comparison With Prior Work

The scope and utility of AI in medicine continue to expand alongside advances in broader society. Nevertheless, large-scale research studies exploring physician perceptions of AI in medicine are scant and focus on a narrow subset of specialties. Doraiswamy et al [36] surveyed 791 psychiatrists from 22 countries, while Pecqueux et al [37] surveyed 147 German surgeons, and De Simone et al [38] surveyed 200 members of the World Society of Emergency Surgery. To our knowledge, our sample of 941 physicians spanning 18 discrete specialties is the first of its kind in both size and scope.

Implications for Implementation

Essential insight into strengthening the practical utility of AI may come from the two most frequent reasons cited by the minority of physicians (144/795, 18.1%) for not adopting it: (1) an insufficient understanding of AI (84/144, 58.3%) and (2) a distrust in the accuracy of AI models (81/144, 56.3%; Figure 4D). These findings suggest that AI developers still have work to do in explaining how their technology works and validating its efficacy. Some of this understanding may develop naturally over time as AI becomes more common in nonmedical settings, making it feel more familiar to physicians and the public alike. However, there are specific actions that developers could take now that might help them in this effort; for instance, the introduction of AI into medicine would benefit tremendously from explainable AI methods, which provide insight into how a model generates predictions and help counter the perception of AI as a “black box.” In addition, explainable AI methods such as gradient-weighted class activation mapping, occlusion maps, and Shapley additive explanation values should be incorporated into research papers proposing new AI use cases in medicine [39]. Ultimately, for AI to be clinically relevant, models must not only demonstrate success in addressing risk-sensitive decisions but also be intuitive and adaptable to the demands of everyday clinical practice.

Limitations

Our study provides valuable insights into physician perspectives on AI in medicine, but several limitations should be acknowledged. First, the survey targeted only physicians affiliated with ACGME-accredited teaching institutions, which limits the generalizability of our results to the broader population of American physicians, including those in community-based settings; however, given the wide breadth of ACGME-accredited residency programs and the hospitals they serve—including many that cross-cover veterans’ and community hospitals—we believe that the diversity of respondents across specialties and geographic regions is fairly representative of academic physicians overall. Future works are needed to identify whether a similar trend exists among physicians in community-based settings. Second, the survey was designed to be distributed via residency coordinators, making it impossible to determine the number of coordinators who actually forwarded the survey compared to the number of completed responses, limiting the validity of our study. We also cannot determine, therefore, whether certain institutions, for instance, those with an interest in AI, comprised a more significant percentage of the survey responses than other institutions. It is certainly possible that institutions with an interest in AI may have made up a large portion of our response rate, thus influencing the results of our study in favor of the use of AI. Due to the anonymous nature of this study, we are unable to determine whether this bias exists. In addition, only complete responses were included in this analysis, and view, participation, and completion rates were not recorded. Finally, while our survey was designed to explore physicians’ perspectives on AI, we did not formally assess respondents’ understanding of what constitutes AI. This introduces the potential for variation in interpretation, particularly for questions about AI-related clinical applications. Future studies should include a standardized definition of AI to ensure uniform understanding among respondents.

Overall, the success of this study design depended on three key factors: (1) the availability of reliable contact information for program coordinators, (2) the coordinators’ willingness to distribute the survey, and (3) the intended respondents’ willingness to complete it. Although this study does not capture the full spectrum of American physicians and trainees, to our knowledge, it represents one of the most extensive datasets of physician perspectives on AI in medicine. We emphasize that these findings reflect the perspectives of this specific subset of respondents and are not intended to be generalized to the entire physician population.

Conclusions

AI represents an exciting, ever-evolving domain of opportunity for improving clinical practice. However, limited understanding of AI and its utility, uncertainty regarding its performance, and concerns about the perpetuation of human biases remain critical barriers that must be overcome in the process of gaining trust among clinicians who are considering using AI in their practice. This study reflects physicians’ opinions and perceived use cases of AI across a variety of specialties and practice settings, quantifying both the reasons why they currently use or are eager to implement AI as well as the rationale for not using or preferring not to adopt AI in their practice. The most desired avenues for implementing AI in medicine were risk stratification, image analysis or segmentation, and disease prognosis, while the most cited reservations were concerns about clinical errors made by AI and the potential for replicating human biases.

Although this study does not fully represent American physicians and trainees, to our knowledge, it represents one of the largest and most diverse datasets of physician perspectives on AI in medicine. Ultimately, the data and insight provided in this work aim to lay the foundation for how AI may best be modified by scientists and harnessed by physicians to be optimally integrated into clinical practice.

Acknowledgments

The authors would like to dedicate this work in honor of Dr Bilwaj Gaonkar, PhD, a member of our lab who passed away prior to this study. This study would not have been possible without him.

Data Availability

The datasets generated and analyzed during this study are available from the corresponding author on reasonable request.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Full survey distributed to academic attending and resident physicians.

PDF File (Adobe PDF File), 88 KB

  1. Gulshan V, Peng L, Coram M, Stumpe MC, Wu D, Narayanaswamy A, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. Dec 13, 2016;316(22):2402-2410. [CrossRef] [Medline]
  2. Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature. Feb 02, 2017;542(7639):115-118. [FREE Full text] [CrossRef] [Medline]
  3. Wilson B, Gaonkar B, Yoo B, Salehi B, Attiah M, Villaroman D, et al. Predicting spinal surgery candidacy from imaging data using machine learning. Neurosurgery. Jun 15, 2021;89(1):116-121. [CrossRef] [Medline]
  4. Ratnaparkhi A, Wilson B, Zarrin D, Suri A, Yoo B, Salehi B, et al. Deep learning based evaluation of surgical candidacy for cervical spinal cord decompression. Research Square. Preprint posted online on May 30, 2024. [FREE Full text] [CrossRef]
  5. Zhou SK, Greenspan H, Davatzikos C, Duncan JS, van Ginneken B, Madabhushi A, et al. A review of deep learning in medical imaging: imaging traits, technology trends, case studies with progress highlights, and future promises. Proc IEEE Inst Electr Electron Eng. May 2021;109(5):820-838. [FREE Full text] [CrossRef] [Medline]
  6. Daye D, Wiggins WF, Lungren MP, Alkasab T, Kottler N, Allen B, et al. Implementation of clinical artificial intelligence in radiology: who decides and how? Radiology. Dec 2022;305(3):555-563. [FREE Full text] [CrossRef] [Medline]
  7. Yin J, Ngiam KY, Teo HH. Role of artificial intelligence applications in real-life clinical practice: systematic review. J Med Internet Res. Apr 22, 2021;23(4):e25759. [FREE Full text] [CrossRef] [Medline]
  8. Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med. Jan 7, 2019;25(1):44-56. [CrossRef] [Medline]
  9. He J, Baxter SL, Xu J, Xu J, Zhou X, Zhang K. The practical implementation of artificial intelligence technologies in medicine. Nat Med. Jan 2019;25(1):30-36. [FREE Full text] [CrossRef] [Medline]
  10. Balch J, Upchurch GRJ, Bihorac A, Loftus TJ. Bridging the artificial intelligence valley of death in surgical decision-making. Surgery. Apr 2021;169(4):746-748. [FREE Full text] [CrossRef] [Medline]
  11. Chew HS, Achananuparp P. Perceptions and needs of artificial intelligence in health care to increase adoption: scoping review. J Med Internet Res. Jan 14, 2022;24(1):e32939. [FREE Full text] [CrossRef] [Medline]
  12. World Medical Association. World Medical Association Declaration of Helsinki: ethical principles for medical research involving human subjects. JAMA. Nov 27, 2013;310(20):2191-2194. [CrossRef] [Medline]
  13. Eysenbach G. Improving the quality of web surveys: the Checklist for Reporting Results of Internet E-Surveys (CHERRIES). J Med Internet Res. Sep 29, 2004;6(3):e34. [FREE Full text] [CrossRef] [Medline]
  14. Lopes MA, Martins H, Correia T. Artificial intelligence and the future in health policy, planning and management. Int J Health Plann Manage. Jan 25, 2024;39(1):3-8. [CrossRef] [Medline]
  15. Liu PR, Lu L, Zhang JY, Huo TT, Liu SX, Ye ZW. Application of artificial intelligence in medicine: an overview. Curr Med Sci. Dec 2021;41(6):1105-1115. [FREE Full text] [CrossRef] [Medline]
  16. Stubberud A, Langseth H, Nachev P, Matharu MS, Tronvik E. Artificial intelligence and headache. Cephalalgia. Aug 05, 2024;44(8):3331024241268290. [FREE Full text] [CrossRef] [Medline]
  17. Mut NA, Ali O, Zain NM. Artificial intelligence in healthcare: evaluating the knowledge and awareness among healthcare professionals in a hospital setting in Malaysia. J Med Imaging Radiat Sci. Oct 2024;55(3):101677. [CrossRef]
  18. Feldman-Stewart D, Brundage MD, McConnell BA, MacKillop WJ. Practical issues in assisting shared decision-making. Health Expect. Mar 09, 2000;3(1):46-54. [FREE Full text] [CrossRef] [Medline]
  19. Phongpreecha T, Ghanem M, Reiss JD, Oskotsky TT, Mataraso SJ, De Francesco D, et al. AI-guided precision parenteral nutrition for neonatal intensive care units. Nat Med. Jun 2025;31(6):1882-1894. [CrossRef] [Medline]
  20. Hashimoto DA, Rosman G, Rus D, Meireles OR. Artificial intelligence in surgery: promises and perils. Ann Surg. Jul 2018;268(1):70-76. [FREE Full text] [CrossRef] [Medline]
  21. Fritsch SJ, Blankenheim A, Wahl A, Hetfeld P, Maassen O, Deffge S, et al. Attitudes and perception of artificial intelligence in healthcare: a cross-sectional survey among patients. Digit Health. Aug 08, 2022;8:20552076221116772. [FREE Full text] [CrossRef] [Medline]
  22. Maassen O, Fritsch S, Palm J, Deffge S, Kunze J, Marx G, et al. Future medical artificial intelligence application requirements and expectations of physicians in German university hospitals: web-based survey. J Med Internet Res. Mar 05, 2021;23(3):e26646. [FREE Full text] [CrossRef] [Medline]
  23. -. Computers make mistakes and AI will make things worse - the law must recognize that. Nature. Jan 2024;625(7996):631. [CrossRef] [Medline]
  24. Makary MA, Daniel M. Medical error-the third leading cause of death in the US. BMJ. May 03, 2016;353:i2139. [CrossRef] [Medline]
  25. Hornung AL, Hornung CM, Mallow GM, Barajas JN, Espinoza Orías AA, Galbusera F, et al. Artificial intelligence and spine imaging: limitations, regulatory issues and future direction. Eur Spine J. Aug 27, 2022;31(8):2007-2021. [CrossRef] [Medline]
  26. Huang J, Galal G, Etemadi M, Vaidyanathan M. Evaluation and mitigation of racial bias in clinical machine learning models: scoping review. JMIR Med Inform. May 31, 2022;10(5):e36388. [FREE Full text] [CrossRef] [Medline]
  27. Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. Oct 25, 2019;366(6464):447-453. [FREE Full text] [CrossRef] [Medline]
  28. Gao J, Mar PL, Chen G. More generalizable models for sepsis detection under covariate shift. AMIA Jt Summits Transl Sci Proc. 2021;2021:220-228. [FREE Full text] [Medline]
  29. Burlina P, Joshi N, Paul W, Pacheco KD, Bressler NM. Addressing artificial intelligence bias in retinal diagnostics. Transl Vis Sci Technol. Feb 05, 2021;10(2):13. [FREE Full text] [CrossRef] [Medline]
  30. Guo LN, Lee MS, Kassamali B, Mita C, Nambudiri VE. Bias in, bias out: underreporting and underrepresentation of diverse skin types in machine learning research for skin cancer detection-a scoping review. J Am Acad Dermatol. Jul 2022;87(1):157-159. [CrossRef] [Medline]
  31. Kaushal A, Altman R, Langlotz C. Geographic distribution of US cohorts used to train deep learning algorithms. JAMA. Sep 22, 2020;324(12):1212-1213. [FREE Full text] [CrossRef] [Medline]
  32. Celi LA, Cellini J, Charpignon ML, Dee EC, Dernoncourt F, Eber R, et al. Sources of bias in artificial intelligence that perpetuate healthcare disparities-a global review. PLOS Digit Health. Mar 31, 2022;1(3):e0000022. [FREE Full text] [CrossRef] [Medline]
  33. Rolle-Lake L, Robbins E. Behavioral risk factor surveillance system. In: StatPearls [Internet]. Treasure Island, FL. StatPearls Publishing; 2025.
  34. Barton M, Hamza M, Guevel B. Racial equity in healthcare machine learning: illustrating bias in models with minimal bias mitigation. Cureus. Feb 2023;15(2):e35037. [FREE Full text] [CrossRef] [Medline]
  35. Vrudhula A, Kwan AC, Ouyang D, Cheng S. Machine learning and bias in medical imaging: opportunities and challenges. Circ Cardiovasc Imaging. Feb 2024;17(2):e015495. [CrossRef] [Medline]
  36. Doraiswamy PM, Blease C, Bodner K. Artificial intelligence and the future of psychiatry: insights from a global physician survey. Artif Intell Med. Jan 2020;102:101753. [CrossRef] [Medline]
  37. Pecqueux M, Riediger C, Distler M, Oehme F, Bork U, Kolbinger FR, et al. The use and future perspective of artificial intelligence-a survey among German surgeons. Front Public Health. Oct 5, 2022;10:982335. [FREE Full text] [CrossRef] [Medline]
  38. De Simone B, Abu-Zidan FM, Gumbs AA, Chouillard E, Di Saverio S, Sartelli M, et al. Knowledge, attitude, and practice of artificial intelligence in emergency and trauma surgery, the ARIES project: an international web-based survey. World J Emerg Surg. Feb 10, 2022;17(1):10. [FREE Full text] [CrossRef] [Medline]
  39. Chaddad A, Peng J, Xu J, Bouridane A. Survey of explainable AI techniques in healthcare. Sensors (Basel). Jan 05, 2023;23(2):634. [FREE Full text] [CrossRef] [Medline]


ACGME: Accreditation Council for Graduate Medical Education
AI: artificial intelligence
CHERRIES: Checklist for Reporting Results of Internet E-Surveys


Edited by B Malin; submitted 11.Feb.2025; peer-reviewed by F Anupama, M Krishnapatnam, N Mungoli; comments to author 06.Apr.2025; revised version received 16.May.2025; accepted 19.Sep.2025; published 04.Dec.2025.

Copyright

©Anshul Ratnaparkhi, Simon Moore, Abhinav Suri, Bayard Wilson, Jacob Alderete, TJ Florence, David Zarrin, David Berin, Rami Abuqubo, Kirstin Cook, Matiar Jafari, Joseph S Bell, Luke Macyszyn, Andrew C Vivas, Joel Beckett. Originally published in JMIR AI (https://ai.jmir.org), 04.Dec.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR AI, is properly cited. The complete bibliographic information, a link to the original publication on https://www.ai.jmir.org/, as well as this copyright and license information must be included.