<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Educ</journal-id><journal-id journal-id-type="publisher-id">mededu</journal-id><journal-id journal-id-type="index">20</journal-id><journal-title>JMIR Medical Education</journal-title><abbrev-journal-title>JMIR Med Educ</abbrev-journal-title><issn pub-type="epub">2369-3762</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v11i1e70079</article-id><article-id pub-id-type="doi">10.2196/70079</article-id><article-categories><subj-group subj-group-type="heading"><subject>Viewpoint</subject></subj-group></article-categories><title-group><article-title>Quo Vadis, AI-Empowered Doctor?</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes" equal-contrib="yes"><name name-style="western"><surname>Takahashi</surname><given-names>Gary</given-names></name><degrees>MS, MD</degrees><xref ref-type="aff" rid="aff1"/><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>von Liechti</surname><given-names>Laurentius</given-names></name><degrees>BS</degrees><xref ref-type="aff" rid="aff1"/><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Tarshizi</surname><given-names>Ebrahim</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1"/><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib></contrib-group><aff id="aff1"><institution>Shiley-Marcos School of Engineering, University of San Diego</institution><addr-line>5998 Alcal&#x00E1; Park</addr-line><addr-line>San Diego</addr-line><addr-line>CA</addr-line><country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Lesselroth</surname><given-names>Blake</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Benitez</surname><given-names>Angel</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Markus</surname><given-names>Elisha</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Ebrahim</surname><given-names>Mansoor Veliyathnadu</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to  Gary Takahashi, MS, MD, Shiley-Marcos School of Engineering, University of San Diego, 5998 Alcal&#x00E1; Park, San Diego, CA, 92110, United States, 1 503-847-3079; <email>gary@garytakahashi.md</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>all authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>15</day><month>8</month><year>2025</year></pub-date><volume>11</volume><elocation-id>e70079</elocation-id><history><date date-type="received"><day>14</day><month>12</month><year>2024</year></date><date date-type="rev-recd"><day>17</day><month>06</month><year>2025</year></date><date date-type="accepted"><day>25</day><month>07</month><year>2025</year></date></history><copyright-statement>&#x00A9; Gary Takahashi, Laurentius von Liechti, Ebrahim Tarshizi. Originally published in JMIR Medical Education (<ext-link ext-link-type="uri" xlink:href="https://mededu.jmir.org">https://mededu.jmir.org</ext-link>), 15.8.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Education, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://mededu.jmir.org/">https://mededu.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://mededu.jmir.org/2025/1/e70079"/><abstract><p>In the first decade of this century, physicians maintained considerable professional autonomy, enabling discretionary evaluation and implementation of new technologies according to individual practice requirements. The past decade, however, has witnessed significant restructuring of medical practice patterns in the United States, with most physicians transitioning to employed status. Concurrently, technological advances and other incentives drove the implementation of electronic systems into the clinic, which these physicians were compelled to integrate. Health care practitioners have now been introduced to applications based on large language models, largely driven by artificial intelligence (AI) developers as well as established electronic health record vendors eager to incorporate these innovations. Although generative AI assistance promises enhanced clinical efficiency and diagnostic precision, its rapid advancement may potentially redefine clinical provider roles and transform workflows, as it has already altered expectations of physician productivity, as well as introduced unprecedented liability considerations. Recognition of the input of physicians and other clinical stakeholders in this nascent stage of AI integration is essential. This requires a more comprehensive understanding of AI as a sophisticated clinical tool. Accordingly, we advocate for its systematic incorporation into standard medical curricula.</p></abstract><kwd-group><kwd>clinical medicine</kwd><kwd>artificial intelligence</kwd><kwd>large language models</kwd><kwd>decision support</kwd><kwd>AI</kwd><kwd>LLM</kwd><kwd>AI in medicine</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Artificial intelligence (AI) has demonstrated long-standing potential to fundamentally transform health care delivery. Prior to the emergence of large language models (LLMs) in the modern era, the implementation and advancement of AI applications were predominantly concentrated in domains such as diagnostic imaging and predictive analytics. These early efforts endeavored to provide decision support for clinicians in critical clinical contexts, such as sepsis identification and management. These implementations were not patient-facing, and these benefits were generally perceived as natural extensions of broader technological progress.</p><p>In contrast, today&#x2019;s interactive chat apps, showcasing advances in LLMs, are able to simulate sentient conversational speech, which has prompted a reconceptualization of AI capabilities. The proficiency of these systems to rapidly process and summarize relevant information from a vast collection of stored knowledge has sparked debates as to the potential of these models to exceed human cognitive performance in tasks requiring sophisticated clinical decision-making and interpretative analysis [<xref ref-type="bibr" rid="ref1">1</xref>].</p><p>Heralded for its transformative potential, AI in medicine has promised to enhance administrative efficiency through the automation of repetitive and time-intensive processes, support doctors through improved diagnostic accuracy, meticulously reduce iatrogenic errors, facilitate personalized medicine tailored to individual patient characteristics, and enable clinicians to navigate the continually expanding corpus of medical research advances and evolving practice guidelines [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>]. However, earlier initiatives to integrate AI into health care frameworks saw limited adoption, as clinicians remained unconvinced as to the technology&#x2019;s capacity to add substantive value in the clinical setting [<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref6">6</xref>]. Technological constraints in computer vision and natural language processing impeded widespread clinical adoption of nascent AI applications, while evolving regulatory frameworks constituted significant barriers to commercialization [<xref ref-type="bibr" rid="ref5">5</xref>].</p><p>Another significant factor impacting the trajectory of health care AI implementation was a shift in professional autonomy. Prior to the preceding decade, the medical profession within the United States operated with greater practitioner independence. Physicians unfamiliar with AI technology, or unconvinced of its practical advantages, had little incentive to incorporate the new technology into their workflow [<xref ref-type="bibr" rid="ref7">7</xref>]. Notably, they were able to determine for themselves when and how best to invest in and implement AI into their medical practice. The contemporary practice landscape has since undergone significant transformation, as the majority of physicians have transitioned from autonomous ownership to employment relationships with hospitals or other corporate health care systems [<xref ref-type="bibr" rid="ref8">8</xref>]. This structural shift has profound implications for the implementation and governance of AI technologies in clinical settings, as employed health care professionals, unable to keep pace with these developments, risk marginalization as key stakeholders [<xref ref-type="bibr" rid="ref9">9</xref>]. Their essential perspectives may be overlooked in critical decisions that will shape clinical workflows, promote work-life balance, and address professional burnout, ultimately redefining their intrinsic role in the health care system [<xref ref-type="bibr" rid="ref10">10</xref>].</p></sec><sec id="s2"><title>What Practicing Physicians Need to Understand Regarding the Role of LLMs</title><p>In the past year, multiple reports have highlighted the remarkable achievements of LLMs on medical knowledge tasks, often claiming accuracy near 100%, which surpasses human capability [<xref ref-type="bibr" rid="ref11">11</xref>]. The benchmark testing panels used to evaluate these models have included datasets of clinical vignettes, urgent care encounters, and medical licensing or board exam datasets [<xref ref-type="bibr" rid="ref12">12</xref>]. Such impressive results, widely publicized in both the general and industry media, have significantly influenced perceptions of medical AI capabilities compared with human practitioners [<xref ref-type="bibr" rid="ref13">13</xref>].</p><p>The inadequacy of standard LLM evaluation metrics as grounds for physician workforce reduction has been comprehensively examined previously [<xref ref-type="bibr" rid="ref14">14</xref>-<xref ref-type="bibr" rid="ref18">18</xref>]. For example, the performance of medical LLMs is still dependent on the provision of pertinent clinical history information and salient features of the physical examination, and it is still not clear that this critical initial step in successfully identifying the nature of a medical condition can be adequately performed by an LLM. Automated techniques to acquire the clinical history by requiring that the user select from a predetermined menu of symptoms and descriptors may fail to capture nuanced empathetic human interaction, such as a sense of advocacy, caring, comfort, and dedication that emerges during genuine patient-provider encounters [<xref ref-type="bibr" rid="ref19">19</xref>,<xref ref-type="bibr" rid="ref20">20</xref>].</p><p>Although LLMs can demonstrate proficiency in tasks involving logic, reasoning, and assimilating large volumes of structured data, these models still lack essential clinical skills such as observation of a patient&#x2019;s demeanor, interpretation of nuanced nonverbal clues, and establishing rapport&#x2014;competencies instinctively performed by a seasoned physician. Such limitations in basic sensorimotor and perceptual processing represent a manifestation of Moravec&#x2019;s paradox, a theoretical conundrum that poses formidable challenges to researchers investigating generative AI [<xref ref-type="bibr" rid="ref21">21</xref>]. Simulated expressions of empathy and clinical judgment can still be perceived as superficial and scripted, precisely because their responses rely on predicted or pretrained responses, rather than authentic and experiential understanding of a patient&#x2019;s lived reality.</p></sec><sec id="s3"><title>Limitations of LLM Capabilities</title><p>Physicians should understand that inference on LLMs is highly dependent on the data on which they have been trained. Details on specific dataset selection for model pretraining are proprietary knowledge, but many have been trained on datasets such as PubMed Central, MIMIC-III clinical notes, sanitized data from electronic health record interactions, and clinical practice guidelines [<xref ref-type="bibr" rid="ref22">22</xref>,<xref ref-type="bibr" rid="ref23">23</xref>]. These models undergo further fine-tuning on additional medical knowledge datasets as well as physician-patient dialog datasets [<xref ref-type="bibr" rid="ref24">24</xref>]. As with any commercial deployments, medical LLMs must adhere to &#x201C;continuous integration/continuous deployment&#x201D; principles in machine learning operations, with monitoring to assure that the application dataset does not drift too far from the training dataset and that regular maintenance fine-tuning and dataset updating are performed [<xref ref-type="bibr" rid="ref25">25</xref>].</p><p>Physicians should also be aware that LLMs, functioning as statistical pattern generators rather than verified information arbiters, generate outputs based on probabilistic distributions within their training data rather than through systematic verification of factual accuracy. Hallucinations remain problematic, afflicting even the latest reasoning models [<xref ref-type="bibr" rid="ref26">26</xref>,<xref ref-type="bibr" rid="ref27">27</xref>]. These confabulatory responses can be difficult for the clinician user to detect, creating a risk for their use in the clinic. Compounding this issue, it has been noted that references cited by LLMs to support their claims may themselves be hallucinatory [<xref ref-type="bibr" rid="ref28">28</xref>].</p><p>Bayesian inference plays a significant role in the clinical application of LLMs in medical decision support. Despite having been trained on extensive medical corpora encompassing comprehensive clinicopathological knowledge, these models may exhibit deficiencies in appropriately weighting disease prevalence. The adage &#x201C;when you hear hoofbeats, think horses, not zebras,&#x201D; reflects the experience of physicians that more common etiologies may present atypically and should still be prioritized. Current LLMs may still struggle in providing reasonable estimates of pretest disease probability, a skill that physicians acquire after years of clinical experience [<xref ref-type="bibr" rid="ref29">29</xref>]. As a consequence, LLMs may disproportionately elevate rare conditions with close symptom concordance over more common diseases with partial clinical alignment [<xref ref-type="bibr" rid="ref30">30</xref>]. LLMs may also fail to understand that the diagnostic process is dynamic and iterative, requiring ongoing refinement in response to emerging patient data revealed in subsequent encounters.</p></sec><sec id="s4"><title>The Importance of Prompting</title><p>The role of system prompt customization in the efficacy of the physician-LLM interaction has been largely unexplored. Physicians may find benefit in interacting with an LLM that behaves like a trusted colleague, rather than a chatbot. Being able to manage the tone of an LLM might encourage a more exploratory and conversational interaction that lowers anxiety and stress, rather than isolated zero-shot querying as with a search engine. Strategic modifications to the system prompt can significantly influence model output, potentially resulting in divergent clinical management recommendations [<xref ref-type="bibr" rid="ref31">31</xref>]. A demonstration of the efficacy of engineered prompting is the use of Medprompt and AutoMedPrompt, which invoke advanced techniques, such as chain-of-thought reasoning, <italic>k</italic>-nearest-neighbor&#x2013;selected few-shot prompting, ensemble voting, and textual gradients, to extract high performance from generalist foundation models in standardized question-answer benchmarks, surpassing that of specialist models [<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref33">33</xref>]. These prompt enhancement techniques can yield impressive scores on multiple-choice question-answer datasets, such as MedQA-USMLE or PubMedQA. However, it is important to recognize that zero-shot (unassisted) performance on unstructured input is the more clinically relevant paradigm, an area where there is a comparative paucity of empirical performance data. A comprehensive study of various open-source models, including several that were fine-tuned on medical corpora, demonstrated that 1- to 3-shot prompting was requisite for optimal clinical language comprehension. The investigators concluded that while LLMs demonstrate proficiency in exam-style question-answer tasks with provided options, they exhibit significant limitations in open-ended scenarios [<xref ref-type="bibr" rid="ref34">34</xref>].</p><p>Public LLMs typically restrict access to system prompting, but domain-specific consultative LLMs should offer this as a customization option. Currently, certain industry stakeholders regard proficiency in prompt engineering as &#x201C;simply an expected skill,&#x201D; exemplifying a troublesome paradigm in which the vast majority of physicians, inadequately trained in this regard, are dependent on software developers to craft tools that physicians poorly understand [<xref ref-type="bibr" rid="ref35">35</xref>]. Physicians should seek training to develop expertise in crafting suitable prompts to obtain the most relevant and suitably formatted information, while minimizing the likelihood of hallucinatory outputs [<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref37">37</xref>].</p></sec><sec id="s5"><title>LLM Performance Compared With Physician Performance</title><p>In addition to reports describing expert-level performance in question-answer multiple-choice testing, LLMs have been touted as being superior in the generation of differential diagnoses when presented with clinical vignettes [<xref ref-type="bibr" rid="ref38">38</xref>]. These capabilities may stem from the models&#x2019; capacity to recall factual information from their training corpora, rather than from any inherent ability to synthesize insight from a panoply of clinical indicators, as with human clinical reasoning [<xref ref-type="bibr" rid="ref38">38</xref>]. For example, the performance of GPT-4 in identifying the diagnosis of published internal medicine cases was significantly decreased when challenged with unpublished clinical vignettes [<xref ref-type="bibr" rid="ref39">39</xref>].</p><p>A recent systematic review and meta-analysis encompassing 83 studies across diverse models (including GPT-4, GPT-4o, PaLM2, and Perplexity, as well as open-source models fine-tuned in the medical domain) found that the pooled accuracy of the generative AI models was 52.1%, demonstrating no overall advantage over physician performance. The models were tested against a variety of clinical vignette datasets, as well as challenges posed in prominent medical journals. Notably, the performance of expert physicians was 15.8% higher, while nonexpert (resident) physicians maintained a marginal 0.6% advantage over LLMs [<xref ref-type="bibr" rid="ref40">40</xref>].</p><p>A counterpoint to these observations, in a commentary highlighting 6 selected studies that examined the effectiveness of LLMs as diagnostic adjuncts, concluded that LLM assistance failed to enhance clinicians&#x2019; diagnostic accuracy, with the models purportedly demonstrating superior performance on various assessment metrics [<xref ref-type="bibr" rid="ref41">41</xref>]. We concur with the contention that claims of physician inferiority in these studies remain inconclusive, given methodological limitations, including an insufficient number of valid datapoints for robust comparison [<xref ref-type="bibr" rid="ref42">42</xref>]. Nevertheless, it is readily apparent that physicians unaccustomed to AI-augmented workflows found LLM assistance unhelpful or counterproductive, especially when resolving discordant or ambiguous model outputs, which consumed valuable clinical time [<xref ref-type="bibr" rid="ref43">43</xref>].</p><p>Physicians should also be cognizant of special legal ramifications regarding the use of AI for clinical decision support. The use of LLMs in patient care potentially exposes a clinician to novel vulnerabilities, broadly including model overreliance, inadequate appreciation of performance limitations, informed consent challenges, and potential bias with ethical ramifications [<xref ref-type="bibr" rid="ref44">44</xref>,<xref ref-type="bibr" rid="ref45">45</xref>]. These risks highlight the need for the robust regulatory oversight of LLM-based technology [<xref ref-type="bibr" rid="ref46">46</xref>]. In litigation, physicians have been required to demonstrate adherence to a reasonable standard of care; however, these norms may evolve in response to transformative technologies [<xref ref-type="bibr" rid="ref47">47</xref>]. In the event of an adverse outcome, physicians also risk penalization by juries whether or not an AI recommendation is accepted or overruled [<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref49">49</xref>]. A rigorous discussion of the legal ramifications of using AI in clinical decision-making is beyond the scope of this viewpoint, but in light of the above considerations, the most prudent use of medical AI may be to confirm an existing medical decision, rather than as a means to augment care [<xref ref-type="bibr" rid="ref50">50</xref>].</p></sec><sec id="s6"><title>The Need for Active Physician Involvement in Shaping the Future of Generative AI in Health Care</title><p>Machine learning and generative AI will undoubtedly catalyze remarkable advancements in health care delivery, especially in clinic settings. These technological advances will undoubtedly exert differential impacts across medical specialties as advances in machine learning are increasingly leveraged to assist in image-processing tasks; however, they are unlikely to wholly replace the clinical expertise of physicians [<xref ref-type="bibr" rid="ref51">51</xref>]. Indeed, Geoffrey Hinton, the &#x201C;godfather of AI,&#x201D; was notoriously inaccurate as to his predictions regarding the demise of diagnostic radiology as a career [<xref ref-type="bibr" rid="ref52">52</xref>]. We feel that health care providers will continue to play essential roles and that AI technology has the potential to augment the capabilities of physicians, nurses, pharmacists, and clinical researchers through the identification of more effective therapeutics and facilitation of novel technological innovations.</p><p>We also wish to emphasize, however, that the notion that a physician empowered by AI may outperform a doctor without this advantage may obscure deeper issues [<xref ref-type="bibr" rid="ref53">53</xref>]. Near-term enhancements in AI-driven productivity gains may ultimately lead to its commoditization and may not necessarily translate to increased compensation, decreased burnout, or even job security [<xref ref-type="bibr" rid="ref54">54</xref>]. In the early stages, physicians may even see an increased demand for their services (Jevons paradox) [<xref ref-type="bibr" rid="ref55">55</xref>]; however, some warn that the augmentation or empowered role of health care providers may ultimately lead to a restructuring of the health care system. Patient intake and flow structures may be eventually redirected to meet the needs of third parties, such as insurers or hospital administrators, to prioritize revenue cycle management, or even to interface with other AI systems, such as those that seek to leverage actionable insights from outcomes data to guide evidence-based treatment recommendations. The adaptation of AI integration may reconfigure key decision-making in health care systems away from the employed physician to those whose priorities put greater weight on economic or political factors.</p><p>Physician input remains critically important in this process, especially in the transformative stages of AI integration into the clinic. We posit that the aforementioned structural shift in the physician employment landscape has significantly attenuated their influence as essential stakeholders and arbiters regarding technological implementation decisions [<xref ref-type="bibr" rid="ref56">56</xref>]. Clinical practitioners should avoid defaulting to passive acceptance as institutionally procured software systems integrate AI technologies into their established clinical workflows.</p><p>Generative AI applications in medicine are still early in development, necessitating an approach that balances technological promotion with the practice-refined workflow of the clinical diagnostic process. The complexities of medical decision-making transcend simplistic evaluation through multiple-choice question-answering from medical datasets. Concern has already been raised that AI-based applications are being adopted too rapidly by hospitals eager to offer the latest in technological innovation, but without the necessary continuous oversight. Relying on the Food and Drug Administration to develop and regulate safeguards is not feasible [<xref ref-type="bibr" rid="ref57">57</xref>]. A different approach, centered on the physician and accommodating the workflow requirements of the practitioner, will better foster physician-AI synergy [<xref ref-type="bibr" rid="ref58">58</xref>,<xref ref-type="bibr" rid="ref59">59</xref>]. Achieving this will require that physicians develop a deeper understanding of the workings of AI technology, comparable to their understanding of more traditional medical tools (<xref ref-type="fig" rid="figure1">Figure 1</xref>). We advocate for research initiatives exploring optimal physician-AI collaboration, potentially including practitioner proficiency in customizing LLM tools to address specific needs. Physicians with such expertise will be better able to advise regulatory bodies on establishing appropriate guardrails against potentially deleterious applications, privacy violations, and the perpetuation of bias and misinformation in health care contexts [<xref ref-type="bibr" rid="ref60">60</xref>].</p><p>Furthermore, clinicians who are well-versed in the limitations of LLMs and related AI applications can provide essential expertise in medicolegal proceedings involving adverse clinical outcomes associated with AI utilization. Enhanced training in AI methodologies will equip physicians to critically evaluate medical research, which increasingly applies advanced data analytics in clinical settings. Such training will also enable physicians to contribute experiential insights and conduct rigorous critiques of machine learning applications designed to enhance predictive analytics. Actualization of these objectives necessitates comprehensive integration of AI education within the pathways of standard medical curricula [<xref ref-type="bibr" rid="ref60">60</xref>].</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Arrows indicate the direction of cause and effect or action initiated to its effect. Green shaded boxes indicate factors where the involvement of AI is direct. AI: artificial intelligence; LLM: large language model; NLM: National Library of Medicine.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v11i1e70079_fig01.png"/></fig></sec><sec id="s7"><title>Proposals for Physician Engagement in AI</title><p>As AI increasingly transforms health care delivery, physicians must proactively expand their expertise to include the following principles, ensuring responsible and effective integration of these technologies into clinical practice:</p><list list-type="bullet"><list-item><p>Physicians should have some understanding of how deep learning models are trained and be aware of factors that can impact accuracy, such as dataset bias, covariate shift, out-of-distribution generalization, and concept drift.</p></list-item><list-item><p>Physicians should understand how deep learning models are evaluated and, when possible, demand from software vendors the provenance of the datasets used for model training as well as performance metrics before they are introduced into the clinic.</p></list-item><list-item><p>Physicians should understand the mechanism underlying LLMs; their intrinsic limitations and vulnerabilities; the impact of prompt engineering on output quality; and how to reduce hallucinatory behavior. Physicians should understand how to evaluate the capabilities of LLM models, as well as whether the information they generate will be exported and used for training purposes. Physicians should understand the ramifications of ambient LLM listening, for example, the custody and retention issues regarding the source recordings generated by AI scribes. These issues pertain to data privacy and confidentiality.</p></list-item><list-item><p>Physicians should understand the potential ethical concerns intrinsic to how LLMs are trained, so as to minimize their perpetuation.</p></list-item><list-item><p>Physicians should understand the legal ramifications of using LLMs as clinical diagnostic support. Physicians should recognize that medical LLMs function best when used adjunctively to validate evidence-based practice, rather than to generate novel treatments or be allowed to operate autonomously.</p></list-item><list-item><p>Physicians should understand how privacy and confidentiality may be breached by incautious use of public LLM models.</p></list-item><list-item><p>Physicians should develop sufficient understanding of clinical AI to be able to critique commercial software.</p></list-item><list-item><p>Physicians should be able to educate and help train ancillary health care staff as to the proper use of AI technology, as well as to instill confidence in patients that such technology will be responsibly deployed.</p></list-item><list-item><p>There should be greater physician participation in the development, validation, and implementation of clinical AI systems, tailored to local deployments.</p></list-item><list-item><p>Physicians should collaborate with clinical informaticians throughout clinical AI implementation to ensure regulatory preparedness and compliance.</p></list-item></list><p>By embracing these essential AI competencies, physicians can maintain their central role in patient care while leveraging this technology to enhance clinical outcomes and preserve the integrity of the medical profession.</p></sec></body><back><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">LLM</term><def><p>large language model</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Castagno</surname><given-names>S</given-names> </name><name name-style="western"><surname>Khalifa</surname><given-names>M</given-names> </name></person-group><article-title>Perceptions of artificial intelligence among healthcare staff: a qualitative survey study</article-title><source>Front Artif Intell</source><year>2020</year><volume>3</volume><issue>578983</issue><fpage>578983</fpage><pub-id pub-id-type="doi">10.3389/frai.2020.578983</pub-id><pub-id pub-id-type="medline">33733219</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bekbolatova</surname><given-names>M</given-names> </name><name name-style="western"><surname>Mayer</surname><given-names>J</given-names> </name><name name-style="western"><surname>Ong</surname><given-names>CW</given-names> </name><name name-style="western"><surname>Toma</surname><given-names>M</given-names> </name></person-group><article-title>Transformative potential of AI in healthcare: definitions, applications, and navigating the ethical landscape and public perspectives</article-title><source>Healthcare (Basel)</source><year>2024</year><month>01</month><day>5</day><volume>12</volume><issue>2</issue><fpage>125</fpage><pub-id pub-id-type="doi">10.3390/healthcare12020125</pub-id><pub-id pub-id-type="medline">38255014</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bajwa</surname><given-names>J</given-names> </name><name name-style="western"><surname>Munir</surname><given-names>U</given-names> </name><name name-style="western"><surname>Nori</surname><given-names>A</given-names> </name><name name-style="western"><surname>Williams</surname><given-names>B</given-names> </name></person-group><article-title>Artificial intelligence in healthcare: transforming the practice of medicine</article-title><source>Future Healthc J</source><year>2021</year><month>07</month><volume>8</volume><issue>2</issue><fpage>e188</fpage><lpage>e194</lpage><pub-id pub-id-type="doi">10.7861/fhj.2021-0095</pub-id><pub-id pub-id-type="medline">34286183</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hirani</surname><given-names>R</given-names> </name><name name-style="western"><surname>Noruzi</surname><given-names>K</given-names> </name><name name-style="western"><surname>Khuram</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Artificial intelligence and healthcare: a journey through history, present innovations, and future possibilities</article-title><source>Life (Basel)</source><year>2024</year><month>04</month><day>26</day><volume>14</volume><issue>5</issue><fpage>557</fpage><pub-id pub-id-type="doi">10.3390/life14050557</pub-id><pub-id pub-id-type="medline">38792579</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Goldfarb</surname><given-names>A</given-names> </name><name name-style="western"><surname>Teodoridis</surname><given-names>F</given-names> </name></person-group><article-title>Why is AI adoption in health care lagging?</article-title><source>Brookings</source><year>2022</year><month>03</month><day>9</day><access-date>2025-06-13</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.brookings.edu/articles/why-is-ai-adoption-in-health-care-lagging/">https://www.brookings.edu/articles/why-is-ai-adoption-in-health-care-lagging/</ext-link></comment></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kelly</surname><given-names>CJ</given-names> </name><name name-style="western"><surname>Karthikesalingam</surname><given-names>A</given-names> </name><name name-style="western"><surname>Suleyman</surname><given-names>M</given-names> </name><name name-style="western"><surname>Corrado</surname><given-names>G</given-names> </name><name name-style="western"><surname>King</surname><given-names>D</given-names> </name></person-group><article-title>Key challenges for delivering clinical impact with artificial intelligence</article-title><source>BMC Med</source><year>2019</year><month>10</month><day>29</day><volume>17</volume><issue>1</issue><fpage>195</fpage><pub-id pub-id-type="doi">10.1186/s12916-019-1426-2</pub-id><pub-id pub-id-type="medline">31665002</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Arvai</surname><given-names>N</given-names> </name><name name-style="western"><surname>Katonai</surname><given-names>G</given-names> </name><name name-style="western"><surname>Mesko</surname><given-names>B</given-names> </name></person-group><article-title>Health care professionals&#x2019; concerns about medical AI and psychological barriers and strategies for successful implementation: scoping review</article-title><source>J Med Internet Res</source><year>2025</year><month>04</month><day>23</day><volume>27</volume><issue>1</issue><fpage>e66986</fpage><pub-id pub-id-type="doi">10.2196/66986</pub-id><pub-id pub-id-type="medline">40267462</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="web"><article-title>PAI-Avalere study on physician employment-practice ownership trends 2019-2023</article-title><source>Physicians Advocacy Institute</source><access-date>2025-05-14</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.physiciansadvocacyinstitute.org/PAI-Research/PAI-Avalere-Study-on-Physician-Employment-Practice-Ownership-Trends-2019-2023">https://www.physiciansadvocacyinstitute.org/PAI-Research/PAI-Avalere-Study-on-Physician-Employment-Practice-Ownership-Trends-2019-2023</ext-link></comment></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hoffman</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wenke</surname><given-names>R</given-names> </name><name name-style="western"><surname>Angus</surname><given-names>RL</given-names> </name><name name-style="western"><surname>Shinners</surname><given-names>L</given-names> </name><name name-style="western"><surname>Richards</surname><given-names>B</given-names> </name><name name-style="western"><surname>Hattingh</surname><given-names>L</given-names> </name></person-group><article-title>Overcoming barriers and enabling artificial intelligence adoption in allied health clinical practice: a qualitative study</article-title><source>Digit Health</source><year>2025</year><volume>11</volume><fpage>20552076241311144</fpage><pub-id pub-id-type="doi">10.1177/20552076241311144</pub-id><pub-id pub-id-type="medline">39906878</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Wolfgruber</surname><given-names>DM</given-names> </name></person-group><article-title>AI&#x2019;s healthcare revolution needs a human touch in 2025</article-title><source>Future Healthcare Today</source><year>2025</year><month>02</month><day>18</day><access-date>2025-05-14</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://futurehealthcaretoday.com/ais-healthcare-revolution-needs-a-human-touch-in-2025/">https://futurehealthcaretoday.com/ais-healthcare-revolution-needs-a-human-touch-in-2025/</ext-link></comment></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>K</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>E</given-names> </name><name name-style="western"><surname>Wei</surname><given-names>K</given-names> </name><etal/></person-group><article-title>An automated framework for assessing how well LLMs cite relevant medical references</article-title><source>Nat Commun</source><year>2025</year><month>04</month><day>16</day><volume>16</volume><issue>1</issue><fpage>3615</fpage><pub-id pub-id-type="doi">10.1038/s41467-025-58551-6</pub-id><pub-id pub-id-type="medline">40240349</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="web"><article-title>Open Medical-LLM leaderboard &#x2013; a Hugging Face space by openlifescienceai</article-title><source>Hugging Face</source><access-date>2025-05-14</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://huggingface.co/spaces/openlifescienceai/open_medical_llm_leaderboard">https://huggingface.co/spaces/openlifescienceai/open_medical_llm_leaderboard</ext-link></comment></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Rajpurkar</surname><given-names>P</given-names> </name><name name-style="western"><surname>Topol</surname><given-names>EJ</given-names> </name></person-group><article-title>Opinion | the robot doctor will see you now</article-title><source>The New York Times</source><year>2025</year><month>02</month><day>2</day><access-date>2025-05-14</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.nytimes.com/2025/02/02/opinion/ai-doctors-medicine.html">https://www.nytimes.com/2025/02/02/opinion/ai-doctors-medicine.html</ext-link></comment></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shah</surname><given-names>NH</given-names> </name><name name-style="western"><surname>Entwistle</surname><given-names>D</given-names> </name><name name-style="western"><surname>Pfeffer</surname><given-names>MA</given-names> </name></person-group><article-title>Creation and adoption of large language models in medicine</article-title><source>JAMA</source><year>2023</year><month>09</month><day>5</day><volume>330</volume><issue>9</issue><fpage>866</fpage><lpage>869</lpage><pub-id pub-id-type="doi">10.1001/jama.2023.14217</pub-id><pub-id pub-id-type="medline">37548965</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bedi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Orr-Ewing</surname><given-names>L</given-names> </name><etal/></person-group><article-title>Testing and evaluation of health care applications of large language models: a systematic review</article-title><source>JAMA</source><year>2025</year><month>01</month><day>28</day><volume>333</volume><issue>4</issue><fpage>319</fpage><lpage>328</lpage><pub-id pub-id-type="doi">10.1001/jama.2024.21700</pub-id><pub-id pub-id-type="medline">39405325</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Raji</surname><given-names>ID</given-names> </name><name name-style="western"><surname>Daneshjou</surname><given-names>R</given-names> </name><name name-style="western"><surname>Alsentzer</surname><given-names>E</given-names> </name></person-group><article-title>It&#x2019;s time to bench the medical exam benchmark</article-title><source>NEJM AI</source><year>2025</year><month>01</month><day>23</day><volume>2</volume><issue>2</issue><fpage>AIe2401235</fpage><pub-id pub-id-type="doi">10.1056/AIe2401235</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hager</surname><given-names>P</given-names> </name><name name-style="western"><surname>Jungmann</surname><given-names>F</given-names> </name><name name-style="western"><surname>Holland</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Evaluation and mitigation of the limitations of large language models in clinical decision-making</article-title><source>Nat Med</source><year>2024</year><month>09</month><volume>30</volume><issue>9</issue><fpage>2613</fpage><lpage>2622</lpage><pub-id pub-id-type="doi">10.1038/s41591-024-03097-1</pub-id><pub-id pub-id-type="medline">38965432</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>F</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>H</given-names> </name><name name-style="western"><surname>Hua</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Rohanian</surname><given-names>O</given-names> </name><name name-style="western"><surname>Clifton</surname><given-names>L</given-names> </name><name name-style="western"><surname>Clifton</surname><given-names>DA</given-names> </name></person-group><article-title>Large language models in healthcare: a comprehensive benchmark</article-title><source>medRxiv</source><comment>Preprint posted online on  Apr 25, 2024</comment><pub-id pub-id-type="doi">10.1101/2024.04.24.24306315</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zakim</surname><given-names>D</given-names> </name></person-group><article-title>Development and significance of automated history-taking software for clinical medicine, clinical research and basic medical science</article-title><source>J Intern Med</source><year>2016</year><month>09</month><volume>280</volume><issue>3</issue><fpage>287</fpage><lpage>299</lpage><pub-id pub-id-type="doi">10.1111/joim.12509</pub-id><pub-id pub-id-type="medline">27071980</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="web"><article-title>AI Patient Actor app &#x2013; Thesen Laboratory</article-title><source>Dartmouth Geisel School of Medicine</source><access-date>2025-05-14</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://geiselmed.dartmouth.edu/thesen/patient-actor-app/">https://geiselmed.dartmouth.edu/thesen/patient-actor-app/</ext-link></comment></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>LoAlza-Bonilla</surname><given-names>A</given-names> </name></person-group><article-title>Moravec&#x2019;s paradox comes to the clinic</article-title><source>LinkedIn</source><year>2024</year><month>12</month><day>31</day><access-date>2025-05-18</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.linkedin.com/pulse/moravecs-paradox-comes-clinic-arturo-loaiza-bonilla-md-lgvee">https://www.linkedin.com/pulse/moravecs-paradox-comes-clinic-arturo-loaiza-bonilla-md-lgvee</ext-link></comment></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Zhou</surname><given-names>H</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>F</given-names> </name><name name-style="western"><surname>Gu</surname><given-names>B</given-names> </name><etal/></person-group><article-title>A survey of large language models in medicine: progress, application, and challenge</article-title><source>arXiv</source><comment>Preprint posted online on  Nov 9, 2023</comment><pub-id pub-id-type="doi">10.48550/arXiv.2311.05112</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>D</given-names> </name><name name-style="western"><surname>Xue</surname><given-names>X</given-names> </name><name name-style="western"><surname>Gao</surname><given-names>P</given-names> </name><etal/></person-group><source>A Survey of Datasets in Medicine for Large Language Models</source><year>2024</year><volume>4</volume><publisher-name>Intell Robot OAE Publishing Inc</publisher-name><fpage>457</fpage><lpage>478</lpage><pub-id pub-id-type="doi">10.20517/ir.2024.27</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Singhal</surname><given-names>K</given-names> </name><name name-style="western"><surname>Azizi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Tu</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Large language models encode clinical knowledge</article-title><source>Nature</source><year>2023</year><month>08</month><volume>620</volume><issue>7972</issue><fpage>172</fpage><lpage>180</lpage><pub-id pub-id-type="doi">10.1038/s41586-023-06291-2</pub-id><pub-id pub-id-type="medline">37438534</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wornow</surname><given-names>M</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Thapa</surname><given-names>R</given-names> </name><etal/></person-group><article-title>The shaky foundations of large language models and foundation models for electronic health records</article-title><source>NPJ Digit Med</source><year>2023</year><month>07</month><day>29</day><volume>6</volume><issue>1</issue><fpage>135</fpage><pub-id pub-id-type="doi">10.1038/s41746-023-00879-8</pub-id><pub-id pub-id-type="medline">37516790</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Jeong</surname><given-names>H</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Medical hallucinations in foundation models and their impact on healthcare</article-title><source>arXiv</source><comment>Preprint posted online on  Feb 26, 2025</comment><pub-id pub-id-type="doi">10.48550/arXiv.2503.05777</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="web"><article-title>OpenAI o3 and o4-mini system card</article-title><source>OpenAI</source><year>2025</year><month>04</month><day>16</day><access-date>2025-05-15</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://openai.com/index/o3-o4-mini-system-card/">https://openai.com/index/o3-o4-mini-system-card/</ext-link></comment></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Ja&#x017A;wi&#x0144;ska</surname><given-names>K</given-names> </name><name name-style="western"><surname>Chandrasekar</surname><given-names>A</given-names> </name></person-group><article-title>AI search has a citation problem</article-title><source>Columbia Journalism Review</source><year>2025</year><month>03</month><day>6</day><access-date>2025-05-15</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.cjr.org/tow_center/we-compared-eight-ai-search-engines-theyre-all-bad-at-citing-news.php">https://www.cjr.org/tow_center/we-compared-eight-ai-search-engines-theyre-all-bad-at-citing-news.php</ext-link></comment></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Gao</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Myers</surname><given-names>S</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Position paper on diagnostic uncertainty estimation from large language models: next-word probability is not pre-test probability</article-title><source>arXiv</source><comment>Preprint posted online on  Nov 7, 2024</comment><pub-id pub-id-type="doi">10.48550/arXiv.2411.04962</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="web"><article-title>A follow up on o1&#x2019;s medical capabilities + major concern about it&#x2019;s utility in medical diagnosis</article-title><source>Substack - Artificial Intelligence Made Simple</source><year>2024</year><month>09</month><day>24</day><access-date>2025-05-15</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://artificialintelligencemadesimple.substack.com/p/a-follow-up-on-o-1s-medical-capabilities">https://artificialintelligencemadesimple.substack.com/p/a-follow-up-on-o-1s-medical-capabilities</ext-link></comment></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>L</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>X</given-names> </name><name name-style="western"><surname>Deng</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Prompt engineering in consistency and reliability with the evidence-based guideline for LLMs</article-title><source>NPJ Digit Med</source><year>2024</year><month>02</month><day>20</day><volume>7</volume><issue>1</issue><fpage>41</fpage><pub-id pub-id-type="doi">10.1038/s41746-024-01029-4</pub-id><pub-id pub-id-type="medline">38378899</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Nori</surname><given-names>H</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>YT</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Can generalist foundation models outcompete special-purpose tuning? Case study in medicine</article-title><source>arXiv</source><comment>Preprint posted online on  Nov 28, 2023</comment><pub-id pub-id-type="doi">10.48550/arXiv.2311.16452</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>S</given-names> </name><name name-style="western"><surname>Koo</surname><given-names>M</given-names> </name><name name-style="western"><surname>Scalzo</surname><given-names>F</given-names> </name><name name-style="western"><surname>Kurtz</surname><given-names>I</given-names> </name></person-group><article-title>AutoMedPrompt: a new framework for optimizing LLM medical prompts using textual gradients</article-title><source>arXiv</source><comment>Preprint posted online on  Feb 21, 2025</comment><pub-id pub-id-type="doi">10.48550/arXiv.2502.15944</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>F</given-names> </name><name name-style="western"><surname>Li</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>H</given-names> </name><etal/></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Al-Onaizan</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Bansal</surname><given-names>M</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>YN</given-names> </name></person-group><article-title>Large language models are poor clinical decision-makers: a comprehensive benchmark</article-title><conf-name>Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing</conf-name><conf-date>Nov 12-16, 2024</conf-date><conf-loc>Miami, FL</conf-loc><fpage>13696</fpage><lpage>13710</lpage><pub-id pub-id-type="doi">10.18653/v1/2024.emnlp-main.759</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Chandonnet</surname><given-names>H</given-names> </name></person-group><article-title>&#x201C;AI is already eating its own&#x201D;: prompt engineering is quickly going extinct</article-title><source>Fast Company</source><year>2025</year><month>06</month><day>5</day><access-date>2025-08-08</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.fastcompany.com/91327911/prompt-engineering-going-extinct">https://www.fastcompany.com/91327911/prompt-engineering-going-extinct</ext-link></comment></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zaghir</surname><given-names>J</given-names> </name><name name-style="western"><surname>Naguib</surname><given-names>M</given-names> </name><name name-style="western"><surname>Bjelogrlic</surname><given-names>M</given-names> </name><name name-style="western"><surname>N&#x00E9;v&#x00E9;ol</surname><given-names>A</given-names> </name><name name-style="western"><surname>Tannier</surname><given-names>X</given-names> </name><name name-style="western"><surname>Lovis</surname><given-names>C</given-names> </name></person-group><article-title>Prompt engineering paradigms for medical applications: scoping review</article-title><source>J Med Internet Res</source><year>2024</year><month>09</month><day>10</day><volume>26</volume><fpage>e60501</fpage><pub-id pub-id-type="doi">10.2196/60501</pub-id><pub-id pub-id-type="medline">39255030</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mesk&#x00F3;</surname><given-names>B</given-names> </name></person-group><article-title>Prompt engineering as an important emerging skill for medical professionals: tutorial</article-title><source>J Med Internet Res</source><year>2023</year><month>10</month><day>4</day><volume>25</volume><fpage>e50638</fpage><pub-id pub-id-type="doi">10.2196/50638</pub-id><pub-id pub-id-type="medline">37792434</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>McDuff</surname><given-names>D</given-names> </name><name name-style="western"><surname>Schaekermann</surname><given-names>M</given-names> </name><name name-style="western"><surname>Tu</surname><given-names>T</given-names> </name><etal/></person-group><article-title>Towards accurate differential diagnosis with large language models</article-title><source>arXiv</source><comment>Preprint posted online on  Nov 30, 2023</comment><pub-id pub-id-type="doi">10.48550/arXiv.2312.00164</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rutledge</surname><given-names>GW</given-names> </name></person-group><article-title>Diagnostic accuracy of GPT-4 on common clinical scenarios and challenging cases</article-title><source>Learn Health Syst</source><year>2024</year><month>07</month><volume>8</volume><issue>3</issue><fpage>e10438</fpage><pub-id pub-id-type="doi">10.1002/lrh2.10438</pub-id><pub-id pub-id-type="medline">39036534</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Takita</surname><given-names>H</given-names> </name><name name-style="western"><surname>Kabata</surname><given-names>D</given-names> </name><name name-style="western"><surname>Walston</surname><given-names>SL</given-names> </name><etal/></person-group><article-title>A systematic review and meta-analysis of diagnostic performance comparison between generative AI and physicians</article-title><source>NPJ Digit Med</source><year>2025</year><month>03</month><day>22</day><volume>8</volume><issue>1</issue><fpage>175</fpage><pub-id pub-id-type="doi">10.1038/s41746-025-01543-z</pub-id><pub-id pub-id-type="medline">40121370</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Topol</surname><given-names>E</given-names> </name><name name-style="western"><surname>Rajpurkar</surname><given-names>P</given-names> </name></person-group><article-title>When doctors with A.I. are outperformed by A.I</article-title><source>Substack - Ground Truths</source><year>2025</year><month>02</month><day>2</day><access-date>2025-05-15</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://erictopol.substack.com/p/when-doctors-with-ai-are-outperformed">https://erictopol.substack.com/p/when-doctors-with-ai-are-outperformed</ext-link></comment></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Polevikov</surname><given-names>S</given-names> </name></person-group><article-title>The &#x201C;AI outperforms doctors&#x201D; claim is false, despite NYT story - a rebuttal (part 2 of 6)</article-title><source>Substack - AI Health Uncut</source><year>2024</year><month>11</month><day>21</day><access-date>2025-05-15</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://sergeiai.substack.com/p/the-ai-outperforms-doctors-claim">https://sergeiai.substack.com/p/the-ai-outperforms-doctors-claim</ext-link></comment></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Agarwal</surname><given-names>N</given-names> </name><name name-style="western"><surname>Moehring</surname><given-names>A</given-names> </name><name name-style="western"><surname>Rajpurkar</surname><given-names>P</given-names> </name><name name-style="western"><surname>Salz</surname><given-names>T</given-names> </name></person-group><article-title>Combining human expertise with artificial intelligence: experimental evidence from radiology</article-title><source>National Bureau of Economic Research</source><year>2023</year><month>07</month><access-date>2025-08-08</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.nber.org/papers/w31422">https://www.nber.org/papers/w31422</ext-link></comment></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Arvai</surname><given-names>N</given-names> </name><name name-style="western"><surname>Katonai</surname><given-names>G</given-names> </name><name name-style="western"><surname>Mesko</surname><given-names>B</given-names> </name></person-group><article-title>Health care professionals&#x2019; concerns about medical AI and psychological barriers and strategies for successful implementation: scoping review</article-title><source>J Med Internet Res</source><year>2025</year><month>04</month><day>23</day><volume>27</volume><fpage>e66986</fpage><pub-id pub-id-type="doi">10.2196/66986</pub-id><pub-id pub-id-type="medline">40267462</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jones</surname><given-names>C</given-names> </name><name name-style="western"><surname>Thornton</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wyatt</surname><given-names>JC</given-names> </name></person-group><article-title>Artificial intelligence and clinical decision support: clinicians&#x2019; perspectives on trust, trustworthiness, and liability</article-title><source>Med Law Rev</source><year>2023</year><month>11</month><day>27</day><volume>31</volume><issue>4</issue><fpage>501</fpage><lpage>520</lpage><pub-id pub-id-type="doi">10.1093/medlaw/fwad013</pub-id><pub-id pub-id-type="medline">37218368</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Weissman</surname><given-names>GE</given-names> </name><name name-style="western"><surname>Mankowitz</surname><given-names>T</given-names> </name><name name-style="western"><surname>Kanter</surname><given-names>GP</given-names> </name></person-group><article-title>Unregulated large language models produce medical device-like output</article-title><source>NPJ Digit Med</source><year>2025</year><month>03</month><day>7</day><volume>8</volume><issue>1</issue><fpage>148</fpage><pub-id pub-id-type="doi">10.1038/s41746-025-01544-y</pub-id><pub-id pub-id-type="medline">40055537</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="web"><article-title>FSMB releases recommendations on the responsible and ethical incorporation of AI into clinical practice</article-title><source>Federation of State Medical Boards</source><year>2024</year><month>05</month><day>2</day><access-date>2025-05-17</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.fsmb.org/advocacy/news-releases/fsmb-releases-recommendations-on-the-responsible-and-ethical-incorporation-of-ai-into-clinical-practice/">https://www.fsmb.org/advocacy/news-releases/fsmb-releases-recommendations-on-the-responsible-and-ethical-incorporation-of-ai-into-clinical-practice/</ext-link></comment></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Appel</surname><given-names>JM</given-names> </name></person-group><article-title>Artificial intelligence in medicine and the negative outcome penalty paradox</article-title><source>J Med Ethics</source><year>2024</year><month>12</month><day>23</day><volume>51</volume><issue>1</issue><fpage>34</fpage><lpage>36</lpage><pub-id pub-id-type="doi">10.1136/jme-2023-109848</pub-id><pub-id pub-id-type="medline">38290853</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Patil</surname><given-names>SV</given-names> </name><name name-style="western"><surname>Myers</surname><given-names>CG</given-names> </name><name name-style="western"><surname>Lu-Myers</surname><given-names>Y</given-names> </name></person-group><article-title>Calibrating AI reliance-a physician&#x2019;s superhuman dilemma</article-title><source>JAMA Health Forum</source><year>2025</year><month>03</month><day>7</day><volume>6</volume><issue>3</issue><fpage>e250106</fpage><pub-id pub-id-type="doi">10.1001/jamahealthforum.2025.0106</pub-id><pub-id pub-id-type="medline">40116804</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Price</surname><given-names>WN</given-names> </name><name name-style="western"><surname>Gerke</surname><given-names>S</given-names> </name><name name-style="western"><surname>Cohen</surname><given-names>IG</given-names> </name></person-group><article-title>Potential liability for physicians using artificial intelligence</article-title><source>JAMA</source><year>2019</year><month>11</month><day>12</day><volume>322</volume><issue>18</issue><fpage>1765</fpage><lpage>1766</lpage><pub-id pub-id-type="doi">10.1001/jama.2019.15064</pub-id><pub-id pub-id-type="medline">31584609</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Wolfe</surname><given-names>D</given-names> </name></person-group><article-title>How physicians are vulnerable to AI</article-title><source>Healthcare Recruiting</source><year>2025</year><month>04</month><day>29</day><access-date>2025-06-15</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.npnow.com/how-physicians-are-vulnerable-to-ai/">https://www.npnow.com/how-physicians-are-vulnerable-to-ai/</ext-link></comment></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Stempniak</surname><given-names>M</given-names> </name></person-group><article-title>NY times revisits nobel prize winner&#x2019;s prediction AI will render radiologists obsolete</article-title><source>Radiology Business</source><year>2025</year><month>05</month><day>15</day><access-date>2025-08-08</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://radiologybusiness.com/topics/artificial-intelligence/ny-times-revisits-nobel-prize-winners-prediction-ai-will-render-radiologists-obsolete">https://radiologybusiness.com/topics/artificial-intelligence/ny-times-revisits-nobel-prize-winners-prediction-ai-will-render-radiologists-obsolete</ext-link></comment></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Choudary</surname><given-names>SP</given-names> </name></person-group><article-title>The many fallacies of &#x201C;AI won&#x2019;t take your job, but someone using AI will&#x201D;</article-title><source>Substack - Platforms, AI, and the Economics of BigTech</source><year>2025</year><month>04</month><day>13</day><access-date>2025-08-08</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://platforms.substack.com/p/the-many-fallacies-of-ai-wont-take">https://platforms.substack.com/p/the-many-fallacies-of-ai-wont-take</ext-link></comment></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>BJ</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>J</given-names> </name></person-group><article-title>The mental health implications of artificial intelligence adoption: the crucial role of self-efficacy</article-title><source>Humanit Soc Sci Commun</source><year>2024</year><month>11</month><day>17</day><volume>11</volume><issue>1</issue><fpage>1</fpage><lpage>15</lpage><pub-id pub-id-type="doi">10.1057/s41599-024-04018-w</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Nguyen</surname><given-names>B</given-names> </name></person-group><article-title>Will AI really lighten the load in allied health? navigating the jevons paradox</article-title><source>LinkedIn</source><year>2025</year><month>01</month><day>15</day><access-date>2025-05-19</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.linkedin.com/pulse/ai-really-lighten-load-allied-health-navigating-jevons-nguyen-pvjnc">https://www.linkedin.com/pulse/ai-really-lighten-load-allied-health-navigating-jevons-nguyen-pvjnc</ext-link></comment></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="web"><article-title>Five key trends driving purchasing decisions in healthcare IT</article-title><source>Signify Research</source><year>2023</year><month>03</month><day>13</day><access-date>2025-05-15</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.signifyresearch.net/insights/five-key-trends-driving-purchasing-decisions-in-healthcare-it/">https://www.signifyresearch.net/insights/five-key-trends-driving-purchasing-decisions-in-healthcare-it/</ext-link></comment></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lenharo</surname><given-names>M</given-names> </name></person-group><article-title>Medicine&#x2019;s rapid adoption of AI has researchers concerned</article-title><source>Nature New Biol</source><year>2025</year><month>06</month><day>9</day><pub-id pub-id-type="doi">10.1038/d41586-025-01748-y</pub-id><pub-id pub-id-type="medline">40490519</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Henry</surname><given-names>T</given-names> </name></person-group><article-title>Physicians&#x2019; greatest use for AI? Cutting administrative burdens</article-title><source>American Medical Association</source><year>2025</year><month>03</month><day>20</day><access-date>2025-08-08</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.ama-assn.org/practice-management/digital-health/physicians-greatest-use-ai-cutting-administrative-burdens">https://www.ama-assn.org/practice-management/digital-health/physicians-greatest-use-ai-cutting-administrative-burdens</ext-link></comment></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Lohr</surname><given-names>S</given-names> </name></person-group><article-title>A.i. was coming for radiologists&#x2019; jobs. So far, they&#x2019;re just more efficient</article-title><source>The New York Times</source><year>2025</year><month>05</month><day>14</day><access-date>2025-05-16</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.nytimes.com/2025/05/14/technology/ai-jobs-radiologists-mayo-clinic.html">https://www.nytimes.com/2025/05/14/technology/ai-jobs-radiologists-mayo-clinic.html</ext-link></comment></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schuitmaker</surname><given-names>L</given-names> </name><name name-style="western"><surname>Drogt</surname><given-names>J</given-names> </name><name name-style="western"><surname>Benders</surname><given-names>M</given-names> </name><name name-style="western"><surname>Jongsma</surname><given-names>K</given-names> </name></person-group><article-title>Physicians&#x2019; required competencies in AI-assisted clinical settings: a systematic review</article-title><source>Br Med Bull</source><year>2025</year><month>01</month><day>16</day><volume>153</volume><issue>1</issue><fpage>ldae025</fpage><pub-id pub-id-type="doi">10.1093/bmb/ldae025</pub-id><pub-id pub-id-type="medline">39821209</pub-id></nlm-citation></ref></ref-list></back></article>