Abstract
Background: The introduction of artificial intelligence (AI) in health care holds great promise, offering the potential to alleviate physicians’ workloads and allocate more time for patient interactions. After the emergence of large language models (LLMs), interest in AI has surged in the health care sector, including within primary care. However, patients have expressed concerns about the ethical implications and use of AI in primary care. Understanding patients’ perspectives on using AI in primary care is crucial for its effective integration. Despite this, few studies have addressed patients’ perspectives on using AI in primary care.
Objective: This study aimed to synthesize qualitative research on primary care patients’ perspectives regarding the use of AI, including LLMs, in primary care.
Methods: A qualitative systematic review, using thematic analysis, was performed in accordance with PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines. Databases, including PubMed, Scopus, Web of Science, CINAHL, and PsycINFO, were searched from inception to February 5, 2024. Eligible studies (1) used a qualitative interview research design, (2) explored primary care patients’ perspectives on the use of AI in primary care, (3) were written in English, and (4) were published in peer-reviewed scientific journals. Quantitative studies, gray literature, surveys, and studies lacking depth in qualitative analysis were excluded. The Critical Appraisal Skills Program (CASP) checklist was used for quality assessment.
Results: Of 1004 studies screened, 6 were included, comprising 170 patients aged 13-91 years from 3 countries. Three themes emerged: “The Relationship with and Actions of AI Systems,” “Implementing AI responsibly,” and “Training Physicians and Artificial Minds.” Patients acknowledged AI’s potential benefits but advocated for clinician oversight, safety frameworks, and the preservation of patient autonomy.
Conclusions: This systematic review provides an understanding of patients’ perspectives on AI in primary care. We identified heterogeneity in AI definitions across studies. Further research is needed on patients’ perspectives across different countries. Notably, our synthesis revealed a significant research gap, as none of the included studies particularly explored patients’ perspectives on LLMs, highlighting an important area for future research.
Trial Registration: PROSPERO CRD42024505209; https://www.crd.york.ac.uk/PROSPERO/view/CRD42024505209
doi:10.2196/72211
Keywords
Introduction
Background
The health care sector is facing challenges, with high administrative workloads and patients growing older and becoming more vulnerable. Primary care, often the first point of contact with health care, is especially vulnerable to these challenges []. Introducing artificial intelligence (AI) in primary care is suggested to free up time for patient contact and prevent physician burnout through the automation of administrative tasks [,].
Today, AI-assisted tools are widely used in various clinical settings and can successfully help predict mortality risks [], improve therapy, and assist diagnosis []. Algorithms that are being used are already outperforming clinicians in spotting malignant tumors [-]. AI solutions are predominantly implemented in referral care settings, with limited use in primary care []. The most common AI applications in primary care include diagnostic decision support, treatment decision support, and data extraction [].
Introducing AI in health care, however, is not without risks. Discrimination through biased training data, uncertainty regarding the division of moral and legal responsibilities, and questions of integrity and privacy are all concerns that may limit the implementation of AI in health care [,]. In addition, implementing AI in health care brings ethical obligations, as it both directly and indirectly impacts patient care [].
Patient engagement is increasingly being recognized as vital to health care delivery, improving outcomes, safety, and reducing hospital admissions and costs []. Impactful organizations, including the World Health Organization (WHO) and the Organization for Economic Co-operation and Development (OECD), have identified patients as one of the main stakeholders in the development and assessment of medical AI applications [,]. For AI implementation to be successful, it is essential to consider the needs and perspectives of the end users. While studies have been conducted on clinicians’ perspectives of AI in primary care [-], few studies exist on primary care patients’ perspectives on AI [-].
The emergence of large language models (LLMs) has significantly advanced the capabilities of AI systems, particularly in their ability to interpret and generate human language []. These models, such as ChatGPT (OpenAI), are built on deep learning (DL) and natural language processing (NLP) [], and their conversational interfaces introduce new challenges and opportunities in health care [,]. In primary care, where the patient-doctor consultation is central, the integration of LLMs may influence communication, decision-making, and the overall care experience []. As LLMs become increasingly integrated into health care systems, it is important to consider how patients perceive and interact with these technologies. Understanding patients’ perspectives on AI, including emerging forms, such as LLMs, is essential for responsible and effective implementation.
To our knowledge, this is the first qualitative meta-synthesis aiming to target primary care patient perspectives on AI and LLMs in primary care.
Aim
The primary objective of this qualitative meta-synthesis is to synthesize the current qualitative research on primary care patients’ perspectives on the use of all types of AI, including the use of LLMs, in primary care.
Methods
AI Definitions
AI encompasses a diverse array of technologies rather than being just a single innovation []. The challenge in defining AI technology stems from the lack of a clear, consensus-driven definition. The types of technology associated with the term AI appear to shift over time, creating difficulties in capturing patients’ perspectives []. One influential definition has been presented by the OECD: “An AI system is a machine-based system that, for explicit or implicit objectives, infers, from the input it receives, how to generate outputs, such as predictions, content, recommendations, or decisions that can influence physical or virtual environments. Different AI systems vary in their levels of autonomy and adaptiveness after deployment” [].
The increasing integration of AI in health care has introduced various systems and models with overlapping and complex characteristics, as presented in . To give context to the different types of AI technologies that may be relevant in primary care, we include short descriptions of key concepts, such as machine learning (ML), DL, NLP, LLMs, and clinical decision support systems (CDSS). These definitions are included to help the reader understand how patients might relate to different AI systems and to reflect how the field of AI is developing in health care. ML uses algorithms to learn patterns, and by training on large datasets, it can make predictions based on the training data. DL, a complex form of ML, can predict outcomes based on many different variables. NLP, based on ML (mostly DL), recognizes and analyzes texts and speech []. NLP allows computers to interpret and generate human language []. LLMs involve the use of NLP and DL to interpret and generate human language based on the input they receive []. LLMs may serve as foundation models for developing more sophisticated AI applications, such as ChatGPT (OpenAI). Further technical details and definitions of AI are provided in [,-].

Study Design
This review was prospectively registered with the International Prospective Register for Systematic Reviews (PROSPERO; CRD42024505209) [] and is reported in accordance with PRISMA (Preferred Reporting Items for Systematic Reviews and Meta‐Analyses) statement []. A completed PRISMA checklist is provided in .
This review used a systematic review methodology for qualitative studies, where the authors conducted a secondary qualitative analysis of the published quotes and supplementary information from the reviewed studies, allowing for a deeper exploration of underlying patterns and themes.
Search Strategy
Searches for peer-reviewed literature were conducted in PubMed and Scopus, with the assistance of a research librarian and research colleagues. Broader terms for “primary health care” and “artificial intelligence” ensured an inclusive search string. In PubMed, the search incorporated MeSH (Medical Subject Headings) terms and Boolean operators. The finalized search string was registered in PROSPERO before conducting the final search on February 5, 2024. After feedback from reviewers, an extended search was conducted in Web of Science, CINAHL, and PsycINFO on July 10, 2025. To ensure a comprehensive review, the reference lists of eligible studies were also screened to identify any studies that may have been missed in our initial searches. The search strings are listed in .
Study Selection
The initial search results were imported into Rayyan (Rayyan Systems Inc) software [] for duplicate removal, with all duplicates being manually verified and excluded from the review. Two authors, A Mundzic and RB, evaluated the studies for inclusion, independently screened the studies based on titles and abstracts, and performed a full-text evaluation. Any discrepancies during the inclusion process, whether in the title and abstract screening or full-text evaluations, were resolved through discussion. In this review, 6 studies [,,,,,] were included, with 4 studies [,,,] exclusively focusing on patients’ perspectives and 2 studies [,], including both patients’ and clinicians’ perspectives. Clinicians’ opinions were excluded during coding.
Inclusion and Exclusion Criteria
Inclusion criteria were defined using the SPIDER (sample, phenomenon of interest, design, evaluation, research type) framework []. Studies were eligible if they (1) used a qualitative interview-based design, (2) were written in English, (3) were published in peer-reviewed scientific journals, (4) were conducted in a primary health care setting, and (5) focused on patients’ perspectives on the use of AI in primary care. For this review, “conducted in a primary health care setting,” was defined as studies in which participants were recruited through primary care, had documented contact with primary care, or where the study context explicitly involved primary care (eg, general practice and family medicine). Studies were excluded if they were quantitative, survey-based, gray literature, or lacked sufficient depth in qualitative analysis. Consequently, studies lacking patient quotations (eg, survey-only or author-summarized reports) were excluded for not providing sufficient insight into participants’ experiences.
Data Extraction and Quality Assessment
A Mundzic extracted all data from the eligible studies within the results and supplementary information that contained opinions and direct quotations from patients. A summary of the characteristics of the included studies was compiled by A Mundzic, and the second author, RB, reviewed the table for accuracy. The following data were extracted: number of participants, interview dates, explanation of AI, type of AI in the study, and emerging themes. In case of missing information relating to the study objectives, the corresponding authors of the respective publications were contacted.
The methodological quality of the chosen studies was assessed using the Critical Appraisal Skills Program Qualitative Research Checklist (CASP) []. The authors A Mundzic and RB, independently conducted the quality appraisal, after which they compared their findings and sorted out any discrepancies by discussion. In cases of disagreement, a discussion was held, and if consensus could not be reached, a third researcher, AE, was consulted to finalize the rating as “YES,” “NO” or “Can’t tell.” In cases where information was missing, the authors answered, “Can’t tell.” The CASP checklist is present as .
Data Analysis and Synthesis
Coding was performed in 2 steps. First, the 4 studies [,,,] focusing exclusively on patients’ perspectives were coded. The remaining 2 studies [,], which included both patients’ perspectives and clinicians’ opinions, were coded separately. The emerging themes from these mixed studies were then compared to those from the studies focusing solely on patients to determine whether the presence of clinicians influenced the themes. No differences were found, and all the patients’ perspectives are presented in this report.
The coding was conducted by A Mundzic, with RB double-checking the codes. Any discrepancies were resolved through discussion before proceeding. In cases of disagreement, a third researcher, AE, was consulted. Both authors independently grouped the codes and discussed emerging themes before finalizing them (see for the full codebook). NVivo 14 software (Lumivero) was used to generate initial codes, search for themes, and identify final themes. Thematic analysis was performed using an inductive approach, following Braun and Clarke’s [] 6-step method to identify latent themes. Inductive thematic analysis with a latent approach offers a flexible and in-depth data analysis []. Since patients’ perspectives on AI in primary care are a relatively unexplored area of research [], this method is suitable to answer the research question.
Ethical Considerations
Ethical approval is not required for this systematic review, as it is based solely on previously published studies. All included studies received ethical approval, except for the study conducted in Denmark, where ethical approval was deemed unnecessary according to Danish national regulations, as it involved self-reported data.
Results
Search Results and Selection
A total of 1406 records were retrieved from the electronic databases. After removing duplicates, 1004 studies remained for abstract and title screening. Following title and abstract screening, 53 full-text studies were screened, with 1 additional study identified from reference screening of the included studies. In this review, 6 studies [,,,,,] were included. Among them, 4 studies [,,,] exclusively focused on the patients’ perspectives, and 2 studies [,] included the perspectives of patients, where clinicians’ opinions were present. The process of the study selection is shown in .

Study Characteristics
The characteristics of the 6 included studies [,,,,,] are visualized in , and additional information is presented in .
Four [,,,] of the included studies were conducted in the United States, and one each was conducted in Denmark [] and Canada []. Two studies [,] were conducted by the same authors and appear to include the same participants but with different research questions. The ages of participants ranged from 13 to 91 years. Interviews were held between October 2019 and January 2022. The publication dates of the included studies ranged from September 21, 2021 to January 6, 2024. A total of 170 unique participants were included, excluding those who overlapped across studies.
The 6 included studies [,,,,,] had different approaches to how AI was defined to the participants. In 2 of the studies [,], participants were given a brief definition of AI and case studies focused on using AI for image analysis, optimizing preventative health, in-patient monitoring, diagnostic support, and engagement with patients during primary care appointments. In 1 study [], participants had little knowledge about AI. Different scenarios were used to explore patients’ opinions. In another study [], questions were asked to explore participants’ perspectives on teleophthalmology, AI-based image interpretation, and virtual care. In the study [], teleophthalmology was implemented as a part of a federally qualified health center’s primary care services, and the technology was used in routine diabetes care. The intervention was regarded as an extended service of general practice and therefore fell within the scope of the review. In the study by Davis et al [], AI was described as integrating data from electronic health records and smartphones to predict suicide risk. In the last study [], AI was described to participants as a technology that automates tasks typically requiring human intelligence, such as processing information, reasoning, learning, planning actions, and communicating in natural language, and presented as a general-purpose prediction technology that estimates missing information from available data.
| Study | Age (years) | Sex n (%) | Participants, n | Type of AI | Themes |
| A Framework for Examining Patient Attitudes Regarding Applications of Artificial Intelligence in Healthcare [] | 18-91 |
| 87 | Machine learning, including deep learning |
|
| Patient Perspectives on Data Sharing Regarding Implementing and Using Artificial Intelligence in General Practice: A Qualitative Study [] | 25-74 |
| 10 | Unspecified |
|
| Perspectives of Latinx Patients with Diabetes on Teleophthalmology, Artificial Intelligence-Based Image Interpretation, and Virtual Care: A Qualitative Study [] | 33-79 |
| 20 | Unspecified (AI-based image interpretation) |
|
| Patient Apprehensions About the Use of Artificial Intelligence in Healthcare [] | 18-91 |
| 87 | Natural language processing and machine learning, including deep learning |
|
| Adolescent, Parent, and Provider Perceptions of a Predictive Algorithm to Identify Adolescent Suicide Risk in Primary Care [] | 13-16 |
| 31 in total; 9 adolescents 12 parents 10 providers | Machine learning |
|
| Priorities for Artificial Intelligence Applications in Primary Care: A Canadian Deliberative Dialog with Patients, Providers, and Health System Leaders [] | 23-73 |
| 22 | Unspecified (clinical decision support tools) |
|
aAI: artificial intelligence.
bPCP: primary care physician.
Quality of Included Studies
Based on the quality assessment, all included studies were considered high quality, as they met most of the CASP criteria, with 1 study [] fulfilling all the criteria. Many of the included studies [,,,,] provided limited or no information regarding the CASP 6 item.
Main Findings
The thematic synthesis resulted in 3 main themes (): “The Relationship with and Actions of AI Systems,” “Implementing AI Responsibly,” and “Training Physicians and Artificial Minds.”
The Relationship With and Actions of Artificial Intelligence (AI) Systems
- Accessibility
- Nonverbal communication
- The changing patient-physician relationship
- AI and clinician collaboration
Implementing AI Responsibly
- Data safety
- Data sharing
- Implementation of AI
- Regulations for AI
Training Physicians and Artificial Minds
- Bias
- Diversity in data
- Inaccurate data
- Reinforcing prejudice
Theme 1: The Relationship With and Actions of AI Systems
The included studies highlight patients’ attitudes toward the changing patient-physician relationship, nonverbal communication, AI and clinician collaboration, accessibility, AI accuracy, efficiency, and priority applications of AI in primary care.
Many participants asserted the importance of an emotional connection with their physician [,], noting that body language helped them feel supported [,] and trust their physicians’ advice more []. Concerns were raised about AI’s assumed inability to perceive emotions, potentially affecting care quality when relying solely on nonemotional health data [,], highlighting physicians’ unique ability to identify underlying issues that patients might not verbally express []. Some patients expressed that AI could enhance emotional comfort during times of difficult health challenges and contribute to acceptance [].
Accessibility was a common concern [,,,,]. Many patients considered AI to be beneficial in rural areas []. Patients expressed concerns regarding the accessibility of AI in health care, particularly for individuals with unique characteristics, such as those who use sign language, have strong accents, or have atypical sensory processing []. Some thought AI could increase health care costs [], and overdiagnosis was mentioned as a reason [].
Patients expressed trust in the health system and their physicians [,,]. In 1 study [], all participants said they would follow their physician’s screening recommendations. Primary care physicians were considered authoritative figures [], and many patients trusted them more than AI [,], citing a general distrust of technology and the belief that a trustful relationship isn’t possible with a machine []. Patients believed physicians aimed to avoid mistakes that could harm their professional reputation []. However, concerns have been raised that health care providers might become overly dependent on AI [,], potentially replacing physicians and weakening the patient-physician relationship []. Another concern was that primary care physicians could shift from being health care providers to mere data collectors if AI gained too much influence [].
Opinions on the accuracy of AI systems varied. Some noted that AI’s access to vast medical data could enhance accuracy, stating that machines are more accurate than physicians [,], while others questioned its ability to understand nuanced symptoms []. Many patients doubted that AI could replace physicians, especially in decision-making, believing the patient-physician relationship is inherently human and essential to patient-centered care []. While some admitted AI’s superior accuracy, they still preferred human involvement in treatment [] and viewed AI as a helpful support tool [,].
Patients also discussed AI’s capacity to manage large datasets and hoped it could use existing information effectively. They emphasized the importance of having as much data as possible when making decisions []. Suggested key AI applications in primary care included clinical documentation, practice operations, risk identification, and triage, with the potential to reduce administrative tasks and enhance physician efficiency [].
Theme 2: Implementing AI Responsibly
Patients expressed opinions on regulations for AI, implementation of AI, data sharing, and data safety.
Participants called for AI regulations to prevent potential harm. Many patients viewed physicians as responsible for their care and preferred that physicians ensure the safety of AI. Patients wanted their physician to make treatment decisions and oversee ongoing care []. Patients believed they should have the right to decide if AI tools are used in their care [] but also to correct their recommendations if they believed them to be inaccurate []. Patients were uncomfortable relying entirely on AI-generated recommendations without the ability to independently assess the reasoning behind those suggestions [,]. While some opposed AI acting autonomously, others were open to it, provided their physician verified its recommendations [,].
An urge for carefulness was described when implementing AI in health care. According to patients, the AI tools should be well tested and accurate before being used []. Patients acknowledged that AI would not be acceptable to all [], but that acceptance of new technology comes gradually [,].
Patients’ views on health care and non-AI health technology were shaped by their social context, including identities, communities, and experiences with past illnesses. They supported AI if it aligned with medicine’s goal of curing illnesses [] and if the design, implementation, and use of AI should preserve or strengthen the patient-physician relationship [].
Ethical concerns were raised, such as AI’s ability to detect future illnesses, which could increase anxiety and hinder living in the present []. Other concerns involved the influence of business interests on the development of AI. One patient noted that health care systems have commercial interests, and companies seek to profit from their data []. Opinions on technological advancements were mixed, with some feeling it was beyond their control and a few expressing discomfort with technology, avoiding cell phones or the internet as long as possible [].
Concerns about data safety included technological failures, hacking risks, and the possibility of data being reidentified after anonymization []. Patients identified factors that would make them more comfortable sharing data, including having unremarkable medical records, assurance of anonymization, and trusting their primary care physician as an authority. The purpose of data sharing was also important [].
In one study [], all patients expressed willingness to share data for AI implementation and the use of AI in general practice. However, sensitive data, such as mental illness or early retirement information, made them more skeptical about data sharing. Patients feared data misuse by insurance companies, potentially leading to monitoring or loss of liberty []. They desired transparency about what data were collected and wanted access to the data used for risk assessment [].
Theme 3: Training Physicians and Artificial Minds
Perspectives on various aspects of training AI models, including bias, data diversity, risks of inaccurate data reinforcing prejudice, and the training of physicians in using AI, emerged among the included studies.
Patients expressed concerns about the risk of developers incorporating their biases into the datasets used to train AI []. Concerns were also raised that training AI with data that lacks diversity could harm certain social groups []. An example provided was when medical records contain errors, thereby creating a risk that AI might use incorrect information [].
Participants acknowledged that implementing AI in health care could impact different populations unevenly, potentially reinforcing stigma or restricting access. Some shared instances where they or someone they knew faced discrimination due to mental illness, alcoholism, weight, or profession. They expressed ideas about AI exacerbating or alleviating these stigmas [].
A fear was described, entailing that future generations of physicians might lack essential skills if AI applications or training fail to preserve the core competencies necessary for patient safety and patient-centered care. One patient wondered how AI would be monitored, how physicians would be trained, and how we would ensure that patients still benefit from experience and knowledge, rather than just dependency on AI [].
Discussion
Principal Findings
This review presents newly emerged concepts derived from a synthesis of evidence across several studies, all using different approaches but set in a primary care context. The theme “The Relationship with and Actions of AI Systems” explored changes in the physician-patient relationship, nonverbal communication, AI-physician collaboration, accessibility, AI accuracy, efficiency, and key AI applications in primary care. Through the analysis, it became evident that patients are concerned about the prospect that AI might change the dynamics of the physician-patient relationship and that trusting AI is different from the trust placed in physicians. The part of trust placed in physicians is based on the emotional connection between patient and physician [,,], but also on the knowledge that physicians have something to lose (ie, their professional reputation []. It is not obvious how this sense of trust nor the delegation of responsibility for one’s health could be transferred from physician to AI systems, which is why this should be carefully considered by AI developers and health care organizations implementing AI into clinician workflows. Moreover, many patients expressed feeling supported by seeing their physicians’ body language [,], something that may be hard to replicate using AI. Patients shared their views on AI regulations, implementation, data sharing, and data safety under the theme “Implementing AI Responsibly.” This theme illustrates the varied views patients hold concerning the implementation of AI. Some patients reported that they considered it necessary for health care providers to inform them about the use of AI [], which is not routinely done in current health care. For many years, physicians have received computationally automated decision support on a daily basis, such as when ordering an electrocardiogram for the analysis of heart-related conditions []. Consequently, physicians and health care providers should be aware of the discrepancy between current practice and patients’ expectations and consider that patients’ preferences may differ depending on their social context, identity, and lived experience []. The last theme, “Training Physicians and Artificial Minds,” studies highlighted perspectives on training AI models, focusing on issues such as bias, data diversity, the risk of inaccurate data reinforcing prejudice, and the training of clinicians in AI use. A knowledge gap was identified, as none of the included studies specifically addressed opinions on LLMs.
Comparison With Prior Work
Patients expressed trust in the current health care systems and valued their relationships with their physicians. Trust in physicians stemmed from their professional reputation and experience, qualities that patients felt a machine could not provide. Earlier studies have shown great trust in primary care physicians, aligning with our findings [-].
While AI as a diagnostic and decision support tool was well-accepted by patients, they strongly preferred that physicians be involved in the process. A similar sentiment was observed in a study representative of the Dutch population, where 78% of women did not fully support independent use of AI-based diagnostics in screening mammography without the involvement of a radiologist []. Another study found that patients preferred physicians over AI for all clinical tasks except for treatment planning based on current scientific evidence. When there was a disagreement between physicians and AI regarding diagnosis and treatment planning, most patients favored the physician’s judgment. AI was considered more acceptable when supervised by a physician, especially in diagnosis, than when used independently [].
Patients emphasized the importance of establishing an emotional bond with their physician, which made them feel recognized and supported, thus building trust and easing decisions about whether to trust their physician. However, concerns were raised about AI’s inability to recognize emotions and how this limitation could impact care, especially when it depends solely on nonemotional health data. Concerns about the lack of human touch in AI have been mentioned in previous studies [,]. On the contrary, a study comparing physician and AI chatbot responses showed that chatbot responses were rated as more empathetic and of better quality than physician responses []. In this systematic review, we found that some participants viewed health care AI as a source of relief or hope in cases where a clear diagnosis has been elusive or where standard treatments have failed, a seemingly unique finding.
Patients were concerned that using flawed data for training AI systems could harm certain social groups and lead to unequal care. Evidence of racial bias has indeed been observed in previous studies. One such study found that, despite being assigned the same risk levels by an algorithm, Black patients were actually sicker than their White counterparts, a bias resulting in over half of Black participants being overlooked for additional care [].
Our findings complement previous work, such as the scoping review by Moy et al [], which broadly mapped patients’ perspectives on AI across various health care settings and study designs. While their review provides a valuable overview, our meta-synthesis offers a more focused contribution by synthesizing qualitative interview studies conducted specifically in primary care. This narrower scope enabled a deeper exploration of context-specific themes and patient experiences within primary care. Furthermore, our synthesis identified a notable gap, where none of the included studies addressed LLMs, highlighting an area for future research as these technologies become increasingly integrated into primary care workflows.
Strengths
To our knowledge, this is the first qualitative meta-synthesis that synthesizes primary care patient perspectives on AI.
To ensure the study’s quality, two authors independently handled the selection, quality assessment, and coding. Thematic analysis was conducted in such a way as to identify if nonpatient perspectives influenced the results.
During the inclusion and exclusion process, discussions emerged regarding the definitions of AI and how it was explained to patients in different studies. The varying definitions of AI presented challenges, as the authors’ interpretations influenced the selection of studies and the summaries of the types of AI used. To maintain an inclusive approach, we aimed to include studies focusing on all types of AI. In some cases, studies addressed AI in health care more broadly. In these instances, inclusion was based on whether the recruitment and thematic framing were anchored in primary care. For example, studies that recruited patients following a recent primary care visit and analyzed their perspectives in that context were considered eligible. This approach ensured relevance to primary care while acknowledging the interdisciplinary nature of AI applications in health care.
Limitations
A major limitation of this review is that none of the included studies explicitly addressed primary care patients’ views on LLMs, highlighting a gap in the current qualitative research on this highly relevant subject. Another limitation of this study is that the initial search included only 2 databases, PubMed and Scopus. These were selected based on the assumption that they would cover relevant literature within the study’s scope. To ensure that no eligible studies were overlooked, we later conducted an expanded search on July 10, 2025, including Web of Science, CINAHL, and PsycINFO covering publications from inception up to February 5, 2024. The extended search did not identify any additional relevant studies, which supports the comprehensiveness of the original search strategy. The initial PRISMA flowchart of study selection is listed in .
The definition of primary care varies globally. In Scandinavia, it primarily refers to appointments with general practitioners, while in the United States, it encompasses all non-emergency care. The results might have differed if more studies from general practice clinics were included. However, it is reasonable to assume that patients in general practice clinics share similar views about AI as those in primary care settings, as these clinics also provide care akin to primary care.
The inclusion of teleophthalmology as part of primary care may raise questions about scope. However, in the Pelayo et al [] study, the service was embedded in a primary care setting and used to support chronic disease management. We therefore considered the patients’ perspectives on this AI-supported referral process to be relevant to primary care broadly.
A subgroup analysis of the results was not possible due to missing information about sexes and ages linked to specific quotations. Only 1 study by Upshaw et al [] provided age-related information alongside participant quotes. The number of studies was insufficient to support comparisons by country.
Two of the studies conducted by Richardson et al [,] appeared to have overlapping study participants, which is a potential source of bias. However, when extracting patient quotes and comparing the material, the quotations in the 2 papers were not identical, and the studies focused on different research questions. In qualitative research synthesis, all quotations are considered valuable and important. Repetition of similar statements does not make a particular quote more important than another. Therefore, the risk of bias in this case is expected to be low.
Many of the included studies [,,,,] provided limited or no information regarding CASP 6 item “Has the relationship between researcher and participants been adequately considered?” In contrast to quantitative research, where methodological details, such as double-blinding are commonly reported, qualitative studies less frequently describe the relationship between the interviewer and participants. Our interpretation is that this reflects a general reporting tendency rather than a lack of quality. A “Can’t tell” rating does not imply that a study is of poor quality but rather that the relevant information was vague.
Future Directions
Understanding patients’ perspectives is crucial as AI becomes more integrated into primary care. Future research could investigate how patients interpret various AI concepts and compare their opinions on AI across different health care settings, such as general practice clinics and primary care environments. Conducting studies in diverse countries and cultures is necessary to capture a broad range of patient attitudes toward AI in primary care, acknowledging that cultural and systemic differences may influence perceptions.
Patients have expressed concerns that machines can make mistakes and lack the authority and reputation that physicians possess. These opinions may be consciously or unconsciously influenced by a preference for social status. This raises an intriguing question: Could AI eventually achieve a form of social status and earn professional credibility? This is an important factor in determining patients’ perspectives regarding responsibility and accountability if autonomous AI commits medical malpractice. Currently, no studies have investigated this topic, highlighting a valuable avenue for future research. Finally, a notable gap remains in the literature, as few studies [-] have explored patient perspectives on AI in general practice settings, and none have specifically addressed LLMs.
Conclusions
Primary care patients recognize that AI can enhance efficiency in primary health care by assisting physicians with administrative tasks and offering decision support through access to comprehensive and accurate medical data. Patients value the human elements in their relationship with physicians, considering these essential for patient-centered care. Although some acknowledge AI’s superior accuracy, they still prefer human involvement in treatment. Many patients trust their physician more than AI and doubt that AI can replace physicians. Patients advocate for physicians to ensure AI safety and promote robust frameworks for its implementation. They emphasize the importance of preserving patient choice and autonomy. Understanding patients’ perspectives and needs is vital for designing effective and user-friendly AI solutions. Understanding the risks associated with AI is essential to ensure its safe and ethical implementation in health care settings.
Acknowledgments
We would like to extend our gratitude to Rasmus Arvidsson for his valuable input throughout the process of this study. ChatGPT (OpenAI) was used to proofread and suggest improvements in wording and grammar for this manuscript.
Funding
This study was funded by Region Västra Götaland (reference 2024-00520) and the Swedish Research Council (reference 2023-05744).
Authors' Contributions
AE, A Mundzic, RB, DS, and RG contributed to conceptualization. AE, A Mundzic, RB, DS, RG, and RA developed the methodology. A Mundzic and RB conducted the investigation and data curation. A Mundzic handled writing–original draft of the manuscript, and all authors participated in writing–review and editing. AE provided supervision. A Mundzic and AE acquired funding for the study. A Mundzic handled the visualization.
Conflicts of Interest
AE has been employed since September 1, 2024, by Tandem Health AB, developing artificial intelligence products for health care providers. Tandem Health AB had no role in conceptualization, methodology, investigation, data curation, writing, supervision, funding acquisition, project administration, or visualization.
Artificial intelligence definitions.
DOC File, 221 KBSearch strings.
DOC File, 36 KBCodebooks.
DOC File, 179 KBSociodemographic and study characteristics.
DOC File, 55 KBInitial PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flowchart of study selection.
DOC File, 231 KBPRISMA checklist.
DOC File, 88 KBCASP checklist.
DOC File, 52 KBReferences
- Mikkelsen JG, Sørensen NL, Merrild CH, Jensen MB, Thomsen JL. Patient perspectives on data sharing regarding implementing and using artificial intelligence in general practice - a qualitative study. BMC Health Serv Res. Apr 4, 2023;23(1):335. [CrossRef] [Medline]
- Upshaw TL, Craig-Neil A, Macklin J, et al. Priorities for artificial intelligence applications in primary care: a Canadian deliberative dialogue with patients, providers, and health system leaders. J Am Board Fam Med. Apr 3, 2023;36(2):210-220. [CrossRef] [Medline]
- Chew HSJ, Achananuparp P. Perceptions and needs of artificial intelligence in health care to increase adoption: scoping review. J Med Internet Res. Jan 14, 2022;24(1):e32939. [CrossRef] [Medline]
- Shung DL, Au B, Taylor RA, et al. Validation of a machine learning model that outperforms clinical risk scoring systems for upper gastrointestinal bleeding. Gastroenterology. Jan 2020;158(1):160-167. [CrossRef] [Medline]
- Lam TYT, Cheung MFK, Munro YL, Lim KM, Shung D, Sung JJY. Randomized controlled trials of artificial intelligence in clinical practice: systematic review. J Med Internet Res. Aug 25, 2022;24(8):e37188. [CrossRef] [Medline]
- Davenport T, Kalakota R. The potential for artificial intelligence in healthcare. Future Healthc J. Jun 2019;6(2):94-98. [CrossRef] [Medline]
- Jones OT, Matin RN, van der Schaar M, et al. Artificial intelligence and machine learning algorithms for early detection of skin cancer in community and primary care settings: a systematic review. Lancet Digit Health. Jun 2022;4(6):e466-e476. [CrossRef] [Medline]
- Haenssle HA, Fink C, Schneiderbauer R, et al. Man against machine: diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists. Ann Oncol. Aug 1, 2018;29(8):1836-1842. [CrossRef] [Medline]
- Brinker TJ, Hekler A, Enk AH, et al. Deep neural networks are superior to dermatologists in melanoma image classification. Eur J Cancer. Sep 2019;119:11-17. [CrossRef] [Medline]
- Kueper JK, Terry AL, Zwarenstein M, Lizotte DJ. Artificial intelligence and primary care research: a scoping review. Ann Fam Med. May 2020;18(3):250-258. [CrossRef] [Medline]
- European Commission, Directorate-General for Communications Networks, Content and Technology. Ethics guidelines for trustworthy AI. Publications Office of the European Union. 2019. URL: https://data.europa.eu/doi/10.2759/346720 [Accessed 2024-10-13]
- Richardson JP, Curtis S, Smith C, et al. A framework for examining patient attitudes regarding applications of artificial intelligence in healthcare. Digit Health. 2022;8:20552076221089084. [CrossRef] [Medline]
- Adus S, Macklin J, Pinto A. Exploring patient perspectives on how they can and should be engaged in the development of artificial intelligence (AI) applications in health care. BMC Health Serv Res. Oct 26, 2023;23(1):1163. [CrossRef] [Medline]
- Ethics and governance of artificial intelligence for health: guidance on large multi-modal models. World Health Organization. URL: https://iris.who.int/server/api/core/bitstreams/e9e62c65-6045-481e-bd04-20e206bc5039/content [Accessed 2024-12-27]
- AI in health: huge potential, huge risks. Organisation for Economic Co-operation and Development (OECD). Jan 17, 2024. URL: https://www.oecd.org/content/dam/oecd/en/publications/reports/2024/01/ai-in-health-huge-potential-huge-risks_ff823a24/2f709270-en.pdf [Accessed 2024-11-27]
- Allen MR, Webb S, Mandvi A, Frieden M, Tai-Seale M, Kallenberg G. Navigating the doctor-patient-AI relationship - a mixed-methods study of physician attitudes toward artificial intelligence in primary care. BMC Prim Care. Jan 27, 2024;25(1):42. [CrossRef] [Medline]
- Davis M, Dysart GC, Doupnik SK, et al. Adolescent, parent, and provider perceptions of a predictive algorithm to identify adolescent suicide risk in primary care. Acad Pediatr. 2024;24(4):645-653. [CrossRef] [Medline]
- Fraile Navarro D, Kocaballi AB, Dras M, Berkovsky S. Collaboration, not confrontation: understanding general practitioners’ attitudes towards natural language and text automation in clinical practice. ACM Trans Comput-Hum Interact. Apr 30, 2023;30(2):1-34. [CrossRef]
- Nash DM, Thorpe C, Brown JB, et al. Perceptions of artificial intelligence use in primary care: a qualitative study with providers and staff of Ontario community health centres. J Am Board Fam Med. Apr 3, 2023;36(2):221-228. [CrossRef] [Medline]
- Sangers TE, Wakkee M, Moolenburgh FJ, Nijsten T, Lugtenberg M. Towards successful implementation of artificial intelligence in skin cancer care: a qualitative study exploring the views of dermatologists and general practitioners. Arch Dermatol Res. Jul 2023;315(5):1187-1195. [CrossRef] [Medline]
- Libon J, Ng C, Bailey A, Hareendranathan A, Joseph R, Dulai S. Remote diagnostic imaging using artificial intelligence for diagnosing hip dysplasia in infants: results from a mixed-methods feasibility pilot study. Paediatr Child Health. Aug 2023;28(5):285-290. [CrossRef] [Medline]
- Buck C, Doctor E, Hennrich J, Jöhnk J, Eymann T. General practitioners’ attitudes toward artificial intelligence-enabled systems: interview study. J Med Internet Res. Jan 27, 2022;24(1):e28916. [CrossRef] [Medline]
- Kocaballi AB, Ijaz K, Laranjo L, et al. Envisioning an artificial intelligence documentation assistant for future primary care consultations: a co-design study with general practitioners. J Am Med Inform Assoc. Nov 1, 2020;27(11):1695-1704. [CrossRef] [Medline]
- Shibl R, Lawley M, Debuse J. Factors influencing decision support system acceptance. Decis Support Syst. Jan 2013;54(2):953-961. [CrossRef]
- Litvin CB, Ornstein SM, Wessell AM, Nemeth LS, Nietert PJ. Adoption of a clinical decision support system to promote judicious use of antibiotics for acute respiratory infections in primary care. Int J Med Inform. Aug 2012;81(8):521-526. [CrossRef] [Medline]
- Ahearn MD, Kerr SJ. General practitioners’ perceptions of the pharmaceutical decision-support tools in their prescribing software. Med J Aust. Jul 7, 2003;179(1):34-37. [CrossRef] [Medline]
- Chalutz Ben-Gal H. Artificial intelligence (AI) acceptance in primary care during the coronavirus pandemic: What is the role of patients’ gender, age and health awareness? A two-phase pilot study. Front Public Health. 2022;10:931225. [CrossRef] [Medline]
- Fritsch SJ, Blankenheim A, Wahl A, et al. Attitudes and perception of artificial intelligence in healthcare: a cross-sectional survey among patients. Digit Health. 2022;8:20552076221116772. [CrossRef] [Medline]
- Khullar D, Casalino LP, Qian Y, Lu Y, Krumholz HM, Aneja S. Perspectives of patients about artificial intelligence in health care. JAMA Netw Open. May 2, 2022;5(5):e2210309. [CrossRef] [Medline]
- Beets B, Newman TP, Howell EL, Bao L, Yang S. Surveying public perceptions of artificial intelligence in health care in the United States: systematic review. J Med Internet Res. Apr 4, 2023;25:e40337. [CrossRef] [Medline]
- Freyer O, Wiest IC, Kather JN, Gilbert S. A future role for health applications of large language models depends on regulators enforcing safety standards. Lancet Digit Health. Sep 2024;6(9):e662-e672. [CrossRef] [Medline]
- Ouyang S, Zhang JM, Harman M, Wang M. An empirical study of the non-determinism of ChatGPT in code generation. ACM Trans Softw Eng Methodol. Feb 28, 2025;34(2):1-28. [CrossRef]
- Atil B, Aykent S, Chittams A, Fu L, Passonneau RJ, Radcliffe E, et al. Non-determinism of “deterministic” LLM settings. arXiv. Preprint posted online on Apr 2, 2024. [CrossRef]
- Larsen JH. Patientsamtalet: Handbok i Mötet Mellan Läkare Och Patient [The Patient Conversation: A Handbook on the Physician–Patient Encounter]. 1st ed. Studentlitteratur; 2015. ISBN: 9789144095646
- Kelly S, Kaye SA, Oviedo-Trespalacios O. What factors contribute to the acceptance of artificial intelligence? A systematic review. Telemat Inform. Feb 2023;77:101925. [CrossRef]
- Explanatory memorandum on the updated OECD definition of an AI system. Organisation for Economic Co-operation and Development (OECD). 2024. URL: https://www.oecd.org/en/publications/explanatory-memorandum-on-the-updated-oecd-definition-of-an-ai-system_623da898-en.html [Accessed 2024-03-05]
- Milne-Ives M, de Cock C, Lim E, et al. The effectiveness of artificial intelligence conversational agents in health care: systematic review. J Med Internet Res. Oct 22, 2020;22(10):e20346. [CrossRef] [Medline]
- Tedre M, Toivonen T, Kahila J, Vartiainen H, Valtonen T, Jormanainen I, et al. Teaching machine learning in K–12 classroom: pedagogical and technological trajectories for artificial intelligence education. IEEE Access. 9:110558-110572. [CrossRef]
- Shortliffe EH. Books: computer-based medical consultations. J Clin Eng. Oct 1976;1(1):69-88. [CrossRef]
- Yu P, Xu H, Hu X, Deng C. Leveraging generative AI and large language models: a comprehensive roadmap for healthcare integration. Healthcare (Basel). Oct 20, 2023;11(20):2776. [CrossRef] [Medline]
- Barreto F, Moharkar L, Shirodkar M, Sarode V, Gonsalves S, Johns A. Generative artificial intelligence: opportunities and challenges of large language models. In: Balas VE, Semwal VB, Khandare A, editors. Intelligent Computing and Networking. 2023:545-553. [CrossRef]
- Wasylewicz ATM, Scheepers-Hoeks A. Clinical decision support systems. In: Kubben P, Dumontier M, Dekker A, editors. Fundamentals of Clinical Data Science. Springer; 2019:153-169. URL: https://link.springer.com/book/10.1007/978-3-319-99713-1 [CrossRef]
- Entezarjou A, Mundzic A, Bogdanffy R. Exploring patient and clinician perspectives on artificial intelligence in primary care - a qualitative systematic review. PROSPERO. URL: https://www.crd.york.ac.uk/PROSPERO/view/CRD42024505209 [Accessed 2025-10-16]
- Page MJ, McKenzie JE, Bossuyt PM, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. Mar 29, 2021;372:n71. [CrossRef] [Medline]
- Ouzzani M, Hammady H, Fedorowicz Z, Elmagarmid A. Rayyan-a web and mobile app for systematic reviews. Syst Rev. Dec 5, 2016;5(1):210. [CrossRef] [Medline]
- Richardson JP, Smith C, Curtis S, et al. Patient apprehensions about the use of artificial intelligence in healthcare. NPJ Digit Med. Sep 21, 2021;4(1):140. [CrossRef] [Medline]
- Pelayo C, Hoang J, Mora Pinzón M, et al. Perspectives of latinx patients with diabetes on teleophthalmology, artificial intelligence-based image interpretation, and virtual care: a qualitative study. Telemed Rep. 2023;4(1):317-326. [CrossRef] [Medline]
- Cooke A, Smith D, Booth A. Beyond PICO: the SPIDER tool for qualitative evidence synthesis. Qual Health Res. Oct 2012;22(10):1435-1443. [CrossRef] [Medline]
- Critical Appraisal Skills Programme (2023). CASP (systematic review checklist). URL: https://casp-uk.net/casp-tools-checklists [Accessed 2024-08-08]
- Braun V, Clarke V. Using thematic analysis in psychology. Qual Res Psychol. Jan 2006;3(2):77-101. [CrossRef]
- Thomas J, Harden A. Methods for the thematic synthesis of qualitative research in systematic reviews. BMC Med Res Methodol. Jul 10, 2008;8(45):18616818. [CrossRef] [Medline]
- Adewole KS, Mojeed HA, Ogunmodede JA, et al. Expert system and decision support system for electrocardiogram interpretation and diagnosis: review, challenges and research directions. Appl Sci (Basel). 2022;12(23):12342. [CrossRef]
- Croker JE, Swancutt DR, Roberts MJ, Abel GA, Roland M, Campbell JL. Factors affecting patients’ trust and confidence in GPs: evidence from the English national GP patient survey. BMJ Open. May 28, 2013;3(5):e002762. [CrossRef] [Medline]
- GP patient survey 2023. NHS England. 2023. URL: https://www.england.nhs.uk/statistics/2023/07/13/gp-patient-survey-2023 [Accessed 2025-06-15]
- Boubshait LA, AlShamlan NA, AlShammari RZ, et al. Patient trust in primary care physicians: a mixed methods study of persons with diabetes at university- based clinics in the eastern province of Saudi Arabia. Patient Prefer Adherence. 2022;16:3241-3255. [CrossRef] [Medline]
- Dehghanfard J, Yazdanpanah A, Mahmoodi G. The level of patient trust in family physician. IJMAS. 2016;2(8):18-21. URL: https://ijmas.iraj.in/paper_detail.php?paper_id=5219&name=The_Level_Of_Patient_Trust_In_Family_Physician [Accessed 2025-11-13]
- Kalsingh MJ, Veliah G, Gopichandran V. Psychometric properties of the Trust in Physician Scale in Tamil Nadu, India. J Family Med Prim Care. 2017;6(1):34-38. [CrossRef] [Medline]
- Blendon RJ, Benson JM, Hero JO. Public trust in physicians--U.S. medicine in international perspective. N Engl J Med. Oct 23, 2014;371(17):1570-1572. [CrossRef] [Medline]
- Gardezi SJS, Elazab A, Lei B, Wang T. Breast cancer detection and diagnosis using mammographic data: systematic review. J Med Internet Res. Jul 26, 2019;21(7):e14464. [CrossRef] [Medline]
- Lennartz S, Dratsch T, Zopfs D, et al. Use and control of artificial intelligence in patients across the medical workflow: single-center questionnaire study of patient perspectives. J Med Internet Res. Feb 17, 2021;23(2):e24221. [CrossRef] [Medline]
- Witkowski K, Dougherty RB, Neely SR. Public perceptions of artificial intelligence in healthcare: ethical concerns and opportunities for patient-centered care. BMC Med Ethics. Jun 22, 2024;25(1):74. [CrossRef] [Medline]
- McCradden MD, Sarker T, Paprica PA. Conditionally positive: a qualitative study of public perceptions about using health data for artificial intelligence research. BMJ Open. Oct 28, 2020;10(10):e039798. [CrossRef] [Medline]
- Ayers JW, Poliak A, Dredze M, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. Jun 1, 2023;183(6):589-596. [CrossRef] [Medline]
- Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. Oct 25, 2019;366(6464):447-453. [CrossRef] [Medline]
- Moy S, Irannejad M, Manning SJ, et al. Patient perspectives on the use of artificial intelligence in health care: a scoping review. J Patient Cent Res Rev. 2024;11(1):51-62. [CrossRef] [Medline]
Abbreviations
| AI: artificial intelligence |
| CASP: Critical Appraisal Skills Programme |
| CDSS: clinical decision support system |
| DL: deep learning |
| LLM: large language model |
| MeSH: Medical Subject Headings |
| ML: machine learning |
| NLP: natural language processing |
| OECD: Organization for Economic Co-operation and Development |
| PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses |
| PROSPERO: International Prospective Register for Systematic Reviews |
| SPIDER: sample, phenomenon of interest, design, evaluation, research |
| WHO: World Health Organization |
Edited by Andrew Coristine; submitted 07.Feb.2025; peer-reviewed by Ken Mandl, Sonish Sivarajkumar; final revised version received 04.Oct.2025; accepted 04.Oct.2025; published 19.Nov.2025.
Copyright© Alisa Mundzic, Robin Bogdanffy, David Sundemo, Pär-Daniel Sundvall, Jonathan Widén, Peter Nymberg, Carl Wikberg, Anna Moberg, Ronny Gunnarsson, Artin Entezarjou. Originally published in JMIR AI (https://ai.jmir.org), 19.Nov.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR AI, is properly cited. The complete bibliographic information, a link to the original publication on https://www.ai.jmir.org/, as well as this copyright and license information must be included.

