<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="review-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Educ</journal-id><journal-id journal-id-type="publisher-id">mededu</journal-id><journal-id journal-id-type="index">20</journal-id><journal-title>JMIR Medical Education</journal-title><abbrev-journal-title>JMIR Med Educ</abbrev-journal-title><issn pub-type="epub">2369-3762</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v12i1e88933</article-id><article-id pub-id-type="doi">10.2196/88933</article-id><article-categories><subj-group subj-group-type="heading"><subject>Review</subject></subj-group></article-categories><title-group><article-title>The Effectiveness of Artificial Intelligence in Undergraduate Health Professions Education: Systematic Review and Meta-Analysis of Randomized Controlled Trials</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Lai</surname><given-names>Nai Ming</given-names></name><degrees>MBBS, MRCP, MRCPCH</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lim</surname><given-names>Yin Sear</given-names></name><degrees>MBBS, MRCPCH</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Win</surname><given-names>Min Thein</given-names></name><degrees>MMedSc, MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Bhargava</surname><given-names>Prabal</given-names></name><degrees>MD, DNB (Ophthal)</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Thomas</surname><given-names>Paraidathathu</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Ong</surname><given-names>Qi Chwen</given-names></name><degrees>MBBS, MPH</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib></contrib-group><aff id="aff1"><institution>School of Medicine, Faculty of Health and Medical Sciences, Taylor's University</institution><addr-line>Subang Jaya</addr-line><addr-line>Selangor</addr-line><country>Malaysia</country></aff><aff id="aff2"><institution>Faculty of Health and Medical Sciences, Taylor's University</institution><addr-line>Subang Jaya</addr-line><addr-line>Selangor</addr-line><country>Malaysia</country></aff><aff id="aff3"><institution>Division of Cardiovascular Medicine, Radcliffe Department of Medicine, John Radcliffe Hospital, University of Oxford</institution><addr-line>Level 6, West Wing, John Radcliffe Hospital, Headington</addr-line><addr-line>Oxford</addr-line><addr-line>England</addr-line><country>United Kingdom</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Brini</surname><given-names>Stefano</given-names></name></contrib><contrib contrib-type="editor"><name name-style="western"><surname>Leung</surname><given-names>Tiffany</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Shilo</surname><given-names>Polina</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Hu</surname><given-names>Yihan</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Qi Chwen Ong, MBBS, MPH, Division of Cardiovascular Medicine, Radcliffe Department of Medicine, John Radcliffe Hospital, University of Oxford, Level 6, West Wing, John Radcliffe Hospital, Headington, Oxford, England, OX3 9DU, United Kingdom, 44 01865234657; <email>qichwen.ong@gtc.ox.ac.uk</email></corresp></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>5</day><month>5</month><year>2026</year></pub-date><volume>12</volume><elocation-id>e88933</elocation-id><history><date date-type="received"><day>04</day><month>12</month><year>2025</year></date><date date-type="rev-recd"><day>30</day><month>03</month><year>2026</year></date><date date-type="accepted"><day>30</day><month>03</month><year>2026</year></date></history><copyright-statement>&#x00A9; Nai Ming Lai, Yin Sear Lim, Min Thein Win, Prabal Bhargava, Paraidathathu Thomas, Qi Chwen Ong. Originally published in JMIR Medical Education (<ext-link ext-link-type="uri" xlink:href="https://mededu.jmir.org">https://mededu.jmir.org</ext-link>), 5.5.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Education, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://mededu.jmir.org/">https://mededu.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://mededu.jmir.org/2026/1/e88933"/><abstract><sec><title>Background</title><p>Health professions education faces increasing challenges from rising health care complexity, pedagogical shifts, and constrained curricular space, and rapidly expanding knowledge and technological advances. While artificial intelligence (AI) shows promise for transforming health professions education, evidence of its effectiveness remains unclear.</p></sec><sec><title>Objective</title><p>This study synthesized evidence from randomized controlled trials (RCTs) on the effectiveness of AI in undergraduate health professions education.</p></sec><sec sec-type="methods"><title>Methods</title><p>We included RCTs, randomized crossover trials, and cluster RCTs comparing AI against standard educational interventions at the undergraduate level. We excluded quasi-experimental studies and those without clear AI components. We searched PubMed, Cochrane, Embase, Educational Resources Information Center, and Web of Science up to January 26, 2026. Outcomes were categorized according by Kirkpatrick levels; risk of bias was assessed using the Risk Of Bias Instrument for Use in Systematic Reviews for Randomised Controlled Trials tool; random-effects meta-analysis was conducted in RevMan (Cochrane); and certainty of evidence was rated using the Grading of Recommendations, Assessment, Development, and Evaluation approach. AI interventions were subcategorized by technology type and educational functions, yielding 13 subcategories.</p></sec><sec sec-type="results"><title>Results</title><p>Of 39,783 records identified, 66 RCTs (N=4911 participants; 2020-2026) were included. Subcategorized analyses across 7 outcome domains yielded 48 comparisons. Most studies had high risk of bias, mainly due to poor allocation concealment and blinding, and certainty of evidence ranged from low to very low. Large language model (LLM)&#x2013;based personalized learning aids comprised the largest evidence base and showed positive effects for satisfaction (standardized mean difference [SMD] 0.93, 95% CI 0.40-1.46; 7 studies; 430 participants; <italic>I</italic>&#x00B2;=74%), confidence (SMD 0.91, 95% CI 0.54-1.29; 7 studies; 609 participants; <italic>I</italic>&#x00B2;=64%), and theoretical knowledge (SMD 0.53, 95% CI 0.13-0.94; 12 studies; 955 participants; <italic>I</italic>&#x00B2;=86%), all with very low certainty. Other AI subtypes, including LLM content generators, natural language processing (NLP) chatbots, and non-LLM adaptive learning platforms, showed generally favorable point estimates but substantial heterogeneity and wide CIs, often included no effect. Prediction intervals frequently crossed the null, indicating uncertainty across educational setting. No studies assessed Kirkpatrick levels 3 or 4.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>This review synthesized RCT evidence on AI in undergraduate health professions education by technology type and function, incorporating evidence certainty. Despite the large number of included studies, evidence remains insufficient to inform educational practice. Some AI interventions may improve some learning outcomes, but effects are inconsistent and not reliably reproducible. High risk of bias, heterogeneity, imprecision, and absence of higher-level outcomes limit conclusions. AI applications should therefore be used cautiously and on a trial basis.</p></sec><sec><title>Trial Registration</title><p>PROSPERO CRD42021243832; https://www.crd.york.ac.uk/PROSPERO/view/CRD42021243832</p></sec></abstract><kwd-group><kwd>health professions education</kwd><kwd>undergraduate</kwd><kwd>artificial intelligence</kwd><kwd>meta-analysis</kwd><kwd>educational resource</kwd><kwd>large languge model</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><sec id="s1-1"><title>Background</title><p>Health professions education is rapidly expanding, with an estimated growth rate of 13.3% per year in the global health education market from 2024, reaching US $264 billion by 2030, according to a market analysis report [<xref ref-type="bibr" rid="ref1">1</xref>]. This growth reflects rising health care complexity, expanding knowledge bases, increasing pedagogical demands, evolving learner expectations, and technological capabilities [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref3">3</xref>]. Traditional health professions education approaches struggle to keep pace with these changes because of challenges in incorporating innovations within the constraints of curricular space [<xref ref-type="bibr" rid="ref4">4</xref>] and resistance from educators and stakeholders [<xref ref-type="bibr" rid="ref5">5</xref>]. Educators face substantial pressure to deliver comprehensive curricular content while ensuring understanding and practical competency in a personalized, student-directed learning paradigm [<xref ref-type="bibr" rid="ref6">6</xref>]. The challenge is compounded by the lack of strong evidence linking the adoption of any educational approach to consistent improvements in a demonstrated chain of relevant outcomes, from learning perception and knowledge to real-life practice and eventual health outcomes [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref8">8</xref>]. The COVID-19 pandemic accelerated the adoption of digital technologies in health professions education but also highlighted their limitations [<xref ref-type="bibr" rid="ref9">9</xref>]. Artificial intelligence (AI), which has been applied in health professions education for over 2 decades [<xref ref-type="bibr" rid="ref10">10</xref>], is now seen as a potentially powerful option to address these challenges given its recent rapid advancements.</p><p>A widely accepted definition of AI is the &#x201C;science of making machines do things that would require intelligence if done by humans&#x201D; [<xref ref-type="bibr" rid="ref11">11</xref>]. AI in its many forms, including machine learning (ML), deep learning (DL), and artificial neural networks, has been used to analyze data patterns and make predictions with efficiency and consistency unmatched by humans [<xref ref-type="bibr" rid="ref12">12</xref>]. Generative AI, a form of DL, has rapidly gained traction across a wide range of fields, including education, following the introduction of ChatGPT (OpenAI), a large language model (LLM), in November 2022 [<xref ref-type="bibr" rid="ref13">13</xref>].</p><p>In health professions education, AI has been used in teaching-learning and assessment for over 2 decades [<xref ref-type="bibr" rid="ref10">10</xref>]. AI-powered systems for teaching-learning include intelligent tutoring systems [<xref ref-type="bibr" rid="ref14">14</xref>], adaptive learning platforms [<xref ref-type="bibr" rid="ref15">15</xref>], and surgical training simulators such as the virtual operative assistant [<xref ref-type="bibr" rid="ref16">16</xref>]. In assessment, AI has been used to automate student scoring and predict future performance in health disciplines [<xref ref-type="bibr" rid="ref17">17</xref>]. To this end, LLMs have emerged as particularly influential, supporting clinical learning note compilation and summarization, virtual patient simulation, and personalized mentoring while facilitating self-directed learning and writing assistance [<xref ref-type="bibr" rid="ref18">18</xref>]. While some studies have reported impressive performance of LLMs in undergraduate and postgraduate medical examinations [<xref ref-type="bibr" rid="ref19">19</xref>], others have shown considerable gaps between LLMs and human experts in clinical diagnostic abilities [<xref ref-type="bibr" rid="ref20">20</xref>]. The findings raise concerns about the appropriate use and limitations of LLMs in contemporary medical education [<xref ref-type="bibr" rid="ref21">21</xref>].</p></sec><sec id="s1-2"><title>Rationale</title><p>Despite extensive AI use in health professions education, fundamental questions about effectiveness remain unclear. While one recent meta-analysis examined early studies on the use of generative AI in medical education [<xref ref-type="bibr" rid="ref22">22</xref>], no prior review has comprehensively synthesized randomized controlled trial (RCT) evidence across all forms of AI, including ML, DL, and generative AI, specifically in undergraduate health professions education, which differs from postgraduate training in pedagogical approaches [<xref ref-type="bibr" rid="ref23">23</xref>]. Published reviews on AI in health professions education are mostly scoping or narrative reviews that are useful in providing guidance on the extent, facilitators, and barriers of AI. However, they do not provide quantitative effect estimates, which are essential for educators and researchers who seek to understand the magnitude of benefits or harm from AI applications, particularly for undergraduate education. The surge in AI publications following ChatGPT&#x2019;s November 2022 introduction [<xref ref-type="bibr" rid="ref13">13</xref>] has created a critical mass of RCT evidence that is rapidly expanding, enabling meaningful meta-analysis.</p></sec><sec id="s1-3"><title>Objectives</title><p>This systematic review and meta-analysis aimed to synthesize evidence on the effects of AI educational applications, compared with standard teaching methods, on learning outcomes among undergraduate health professions students.</p></sec></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Eligibility Criteria</title><p>We included RCTs, randomized crossover, and cluster-randomized trials that enrolled undergraduate students of health sciences, including medicine, pharmacy, nursing, and allied health. We accepted studies conducted in any health training setting, including clinical teaching that involved patient engagement, skills training, classroom learning in the form of problem-based, task-based, or case-based learning, written assignments, as well as any form of assessment across all periods of follow-up. We excluded studies that evaluated postgraduate or residency training to limit the scope of our review, given the sufficient number of studies published at the undergraduate level. We imposed no restrictions on publication year, language, or report format (abstract or full paper).</p><p>In terms of intervention, we accepted studies that explicitly reported the use of AI, which included ML, DL, or generative AI, chiefly LLMs, used for teaching and assessment purposes, either as the main or supplementary tool. We only included studies that evaluated robotics and virtual or augmented reality with a clear description of AI involvement, as these applications may or may not involve an operating AI component [<xref ref-type="bibr" rid="ref24">24</xref>]. We classified AI interventions as generative AI if they were explicitly described as using LLMs or other generative architectures capable of producing novel text, images, or dialogue in response to user input (eg, ChatGPT, GPT-4&#x2013;based tutors (OpenAI), and LLM-driven virtual patients). We classified all remaining AI interventions as nongenerative AI, including ML and DL applications that analyze or classify data without generating novel output (eg, imaging diagnostic aids, automated scoring systems, and rule-based chatbots). The comparison included conventional methods of teaching and assessment involving direct human face-to-face input, as well as the use of various teaching-learning technologies except AI.</p><p>The review outcomes included any form of learning gain, evaluated using various scales in accordance with the levels depicted in the Kirkpatrick Model of Evaluation, a framework developed by Kirkpatrick in 1959 [<xref ref-type="bibr" rid="ref25">25</xref>] and that has since been widely used in evaluating all forms of training. The list of outcomes is detailed under the subsequent subheading of Data Items. We included studies that fulfilled the criteria in terms of population and intervention, regardless of whether the study reported relevant outcomes or produced suitable outcome data for our meta-analysis.</p></sec><sec id="s2-2"><title>Information Sources</title><p>We searched MEDLINE (PubMed), the Cochrane Central Register of Controlled Trials (which included records from PubMed, Embase, CINAHL, and trial registers, including the World Health Organization (WHO) International Trials Registry Platform and ClinicalTrials.gov), Embase, Web of Science (Ovid), and Educational Resources Information Center for published studies and trial register records up to January 26, 2026. We searched the reference lists of relevant reviews for additional studies but did not identify any. We did not search conferences or other online sources or contact authors or personnel for additional studies.</p></sec><sec id="s2-3"><title>Search Strategies</title><p>We report the search strategies following the PRISMA-S (Preferred Reporting Items for Systematic Reviews and Meta-Analyses&#x2013;Search extension) guideline [<xref ref-type="bibr" rid="ref26">26</xref>]. The search strategies were developed by an author (NML) who has extensive systematic review experience, and were reviewed by another author (QCO) with systematic review experience (refer to Part 2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> for search strategies). We searched without applying any preset search filter or any language and publication type restriction.</p></sec><sec id="s2-4"><title>Data Collection Process</title><p>Two authors (NML and YSL) independently screened titles and abstracts for shortlisting and evaluated shortlisted papers in full text to determine eligibility, after deduplication using EndNote (version 19; Clarivate). One author (NML) used Claude Sonnet (Anthropic) [<xref ref-type="bibr" rid="ref27">27</xref>], an LLM application, to perform preliminary data extraction, including population, intervention, comparison, and outcomes, and verified accuracy against the full texts. We provided the LLM with structured prompts that included the overall context, aided by the background information (background and methodology of the current review) as well as a data extraction spreadsheet template with desired headings, assigning the LLM a role as an expert in systematic review methodology, and instructing the platform to read and extract stipulated data types in a step-by-step manner, pointing out the location of the data source to facilitate human verification (examples of prompts are available in Part 11 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>), and manually transcribed the verified data into an Excel (Microsoft) spreadsheet, while a second author (QCO) cross-checked the extracted data by NML against the full texts. We assessed multiple publications of the same study via the setting and characteristics as enumerated above and only selected the reference with the most complete description of the study as the primary reference. We resolved disagreements by discussion leading to a consensus, with referral to the third author (MTW) as required.</p></sec><sec id="s2-5"><title>Data Items</title><p>We extracted participant characteristics, including setting (field of health professions education, eg, medicine, nursing, and dentistry), specific subject or skills evaluated (eg, clinical skills, ophthalmology, radiograph interpretation, and denture mounting), the region where the study was conducted, intervention technologies (eg, LLM or non-LLM and specific technology used, including the main purpose of teaching and learning vs assessment), frequency and duration of the intervention, assessment and follow-up period, outcomes reported, and funding sources.</p><p>We categorized the outcomes, based on the Kirkpatrick Model of Training Evaluation into perceptionor satisfaction and self-efficacy or confidence (level 1), knowledge gain (theoretical knowledge, clinical skills, practical skills [including competence and task efficiency], and generic or personal skills) (level 2), behavioral change assessed from real-life practice (level 3), and improvement in outcomes at the level of the recipient, namely, patient health outcomes (level 4) [<xref ref-type="bibr" rid="ref28">28</xref>].</p></sec><sec id="s2-6"><title>Assessment of Missing Data</title><p>If we found a significant dropout rate (&#x003E;20%), we would judge the study at high risk of bias in terms of missing outcome data. We did not contact any author to request further information, as we did not consider any missing data to be critical for our meta-analysis.</p></sec><sec id="s2-7"><title>Study Risk-of-Bias Assessment</title><p>Two authors (NML and QCO) independently assessed risk of bias using the ROBUST-RCT (Risk Of Bias Instrument for Use in Systematic Reviews for Randomised Controlled Trials) tool, developed by Wang et al [<xref ref-type="bibr" rid="ref29">29</xref>], with adaptation in the domain of blinding from &#x201C;health providers&#x201D; to &#x201C;education providers.&#x201D; The tool consisted of 6 core domains (random sequence generation, allocation concealment, blinding of participants, blinding of educational providers, blinding of outcome assessors, and outcome data not included in the analysis), with 8 optional domains that we did not assess. Detailed guidance, including suggested rules and a decision table of the tool, is freely available in the additional material of the paper by Wang et al [<xref ref-type="bibr" rid="ref29">29</xref>].</p><p>Risk of bias was assessed separately for subjective and objective outcomes, particularly for items related to blinding of education providers. For subjective outcomes, we split blinding of providers (Item 4) into 2 subdomains. Item 4a (perception and satisfaction) was rated probably low risk, as anonymous self-reporting by adult learners is unlikely to be systematically influenced by unblinded providers in the absence of evidence of coercion or deliberate influence on ratings. Item 4b (confidence and self-perceived competence) was rated probably high risk by default, as unblinded providers can plausibly influence learners&#x2019; self-appraisals through differential feedback, encouragement, and validation beyond the educational content delivered; exceptions were made for programs that were largely self-directed or used highly standardized provider-learner interactions. For objective outcomes, Item 4 was similarly rated probably high risk by default, as differential input from unblinded providers can affect actual competence even when outcomes are objectively scored, with exceptions for standardized or self-directed programs.</p><p>In ROBUST-RCT, there is no explicit rule provided in assigning the overall risk-of-bias status. Consequently, we described the risk of bias per study per domain and evaluated the overall degree of concern regarding risk of bias in our certainty of evidence rating by judging the proportion of high-risk domains in the body of evidence, as detailed in Part 1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. For example, if the majority of the included studies for the outcome assessed had a single high-risk domain (with blinding as a whole considered as 1 domain for this purpose), we would consider the body of evidence to have serious concerns in terms of risk of bias, and downgrade the certainty of the evidence by 1 level due to study limitations. However, if the majority of the included studies had 2 or more high-risk domains, we would consider the body of evidence to have very serious concerns in terms of risk of bias and downgrade the certainty of the evidence by 2 levels due to study limitations.</p></sec><sec id="s2-8"><title>Effect Measures</title><p>For continuous outcomes, we used standardized mean difference (SMD) to pool the study results, as each included study reported its outcome data using different scales. We used Cohen benchmarks as a descriptive reference (SMD &#x003C;0.2 small, 0.2&#x2010;0.5 small to medium, &#x003E;0.5&#x2010;0.8 medium to large, and &#x003E;0.8 large) [<xref ref-type="bibr" rid="ref30">30</xref>], while recognizing that these thresholds were not developed for educational intervention contexts and should not be interpreted as definitive indices of educational importance [<xref ref-type="bibr" rid="ref31">31</xref>]. Accordingly, we contextualized effect sizes using a default minimally important difference of SMD 0.5, consistent with the empirically derived half-standard-deviation rule [<xref ref-type="bibr" rid="ref32">32</xref>], applying this threshold in our interpretation of synthesis results. For dichotomous outcomes, we reported the results using relative risk (RR).</p></sec><sec id="s2-9"><title>Synthesis Methods</title><p>Studies with suitable numerical outcome data, either reported in text, table or derivable from figures were eligible for synthesis; those presenting results only narratively were excluded. We derived missing SD by multiplying the SE with the square root of the sample size in the corresponding group. For studies that reported their results in median and IQR, we approximated the median as the mean and obtained an estimate of the SD by dividing the IQR by 1.35, as recommended by Hozo et al [<xref ref-type="bibr" rid="ref33">33</xref>]. Examples of studies from which we derived outcome data are available in Part 1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. For multiarm studies, we selected the relevant AI and control arms. In studies with multiple control groups, we selected the group representing the current standard educational intervention. We ensured each study appeared only once per analysis to avoid counting participants multiple times. For randomized crossover studies, we planned to include outcome data from the first period only, before the crossover, following one of the approaches suggested in the Cochrane Handbook [<xref ref-type="bibr" rid="ref34">34</xref>]. If first-period data were not available, we would accept the data as reported by the authors and conduct sensitivity analysis to assess the impact of these crossover studies on the pooled results.</p><p>We tabulated major components of the intervention and comparison in each study in the characteristics of included studies table (Part 5 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). We performed separate comparisons for each subcategory of AI application and divided the outcomes according to the levels of the Kirkpatrick Model of Training Evaluation [<xref ref-type="bibr" rid="ref25">25</xref>], with further division for different subdomains based on the characteristics of the outcomes assessed, for example, self-efficacy and attitude (level 1) and theoretical knowledge, clinical, and practical skills (level 2), following discussion among the review authors.</p><p>If studies reported multiple outcomes in the same domain, we exercised our judgment in selecting the most suitable outcomes for our meta-analysis following a discussion among the review authors. For example, we used a combined score in preference to scores for individual components, and if no combined scores were reported, we selected one component that was most relevant to the domain; and if there were multiple components of equal relevance, we manually derived the combined mean and SD following the formula recommended in Chapter 10 of the Cochrane Handbook [<xref ref-type="bibr" rid="ref35">35</xref>].</p><p>We presented our meta-analysis results graphically using forest plots. However, because of the large number of meta-analyses performed, we presented the synthesis results numerically in <xref ref-type="table" rid="table1">Table 1</xref> and kept all the analysis details, including the forest plots in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendices 1</xref> and <xref ref-type="supplementary-material" rid="app2">2</xref>.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>The full analysis results for each comparison between artificial intelligence (AI) subcategories and controlto with prediction intervals where applicable and certainty of evidence.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Outcome and specific comparison</td><td align="left" valign="bottom">Studies (participants)</td><td align="left" valign="bottom"><italic>I</italic>&#x00B2;</td><td align="left" valign="bottom">Results (point estimate with 95% CI)</td><td align="left" valign="bottom">95% prediction interval</td><td align="left" valign="bottom">Certainty of evidence (GRADE<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup>)</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="6">Kirkpatrick level 1: satisfaction or motivation</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup> content generator vs control</td><td align="left" valign="top">5 (509)</td><td align="left" valign="top">96%</td><td align="left" valign="top">SMD<sup><xref ref-type="table-fn" rid="table1fn3">c</xref></sup> 0.65 (&#x2212;0.73 to 2.04)</td><td align="left" valign="top">&#x2212;2.67 to 3.98</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM personalized learning aid vs control</td><td align="left" valign="top">7 (430)</td><td align="left" valign="top">74%</td><td align="left" valign="top">SMD 0.93 (0.40 to 1.46)</td><td align="left" valign="top">&#x2212;0.40 to 2.26</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM virtual patient vs control</td><td align="left" valign="top">3 (127)</td><td align="left" valign="top">64%</td><td align="left" valign="top">SMD 0.69 (&#x2212;0.83 to 2.21)</td><td align="left" valign="top">&#x2212;1.86 to 3.24</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM content generator + LLM virtual patient + LLM personalized learning aid vs control</td><td align="left" valign="top">1 (88)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD &#x2212;0.02 (&#x2212;0.44 to 0.40)</td><td align="left" valign="top">N/A<sup><xref ref-type="table-fn" rid="table1fn4">h</xref></sup></td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM-integrated curriculum vs control</td><td align="left" valign="top">1 (96)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 1.31 (0.87 to 1.76)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup>-moderated adaptive learning platform vs control</td><td align="left" valign="top">2 (143)</td><td align="left" valign="top">0%</td><td align="left" valign="top">SMD 0.55 (&#x2212;1.00 to 2.11)</td><td align="left" valign="top">&#x2212;1.00 to 2.11</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>NLP<sup><xref ref-type="table-fn" rid="table1fn6">f</xref></sup> rule-based chatbot vs control</td><td align="left" valign="top">2 (146)</td><td align="left" valign="top">74%</td><td align="left" valign="top">SMD 0.74 (&#x2212;3.65 to 5.13)</td><td align="left" valign="top">&#x2212;6.18 to 7.66</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>NLP rule-based chatbot + rule based virtual patient vs control</td><td align="left" valign="top">1 (61)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.17 (&#x2212;0.33 to 0.67)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>NLP rule-based virtual patient vs control</td><td align="left" valign="top">1 (79)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.75 (0.29 to 1.20)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Low</td></tr><tr><td align="left" valign="top" colspan="6">Kirkpatrick level 1: self-efficacy or confidence</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM personalized learning aid vs control</td><td align="left" valign="top">7 (609)</td><td align="left" valign="top">64%</td><td align="left" valign="top">SMD 0.91 (0.54 to 1.29)</td><td align="left" valign="top">0.05 to 1.77</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM virtual patient vs control</td><td align="left" valign="top">2 (100)</td><td align="left" valign="top">91%</td><td align="left" valign="top">SMD 1.36 (&#x2212;8.09 to 10.81)</td><td align="left" valign="top">&#x2212;14.49 to 17.20</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM-integrated curriculum vs control</td><td align="left" valign="top">1 (96)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 1.23 (0.79 to 1.67)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI procedure assistant vs control</td><td align="left" valign="top">1 (40)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.00 (&#x2212;0.62 to 0.62)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI moderated adaptive learning platform vs control</td><td align="left" valign="top">1 (40)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 2.45 (1.61 to 3.28)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>NLP rule-based chatbot vs control</td><td align="left" valign="top">2 (146)</td><td align="left" valign="top">43%</td><td align="left" valign="top">SMD 0.87 (&#x2212;2.11 to 3.86)</td><td align="left" valign="top">&#x2212;3.21 to 4.96</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI-VR<sup><xref ref-type="table-fn" rid="table1fn7">g</xref></sup> virtual doctor vs control</td><td align="left" valign="top">1 (64)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD &#x2212;0.68 (&#x2212;1.18 to&#x2212;0.17)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI procedure assistant + adaptive learning platform vs control</td><td align="left" valign="top">1 (20)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.55 (&#x2212;0.34 to 1.45)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI procedure assistant (non-LLM) vs control (proportion of participants confident in echocardiography view)</td><td align="left" valign="top">1 (43)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">RR 1.26 (0.45 to 3.50)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top" colspan="6">Kirkpatrick level 2: theoretical knowledge score</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM content generator vs control</td><td align="left" valign="top">3 (359)</td><td align="left" valign="top">93%</td><td align="left" valign="top">SMD 0.99 (&#x2212;1.04 to 3.01)</td><td align="left" valign="top">&#x2212;2.96 to 4.93</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM gamification tool vs control</td><td align="left" valign="top">1 (48)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.79 (0.20 to 1.38)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM personalized learning aid vs control</td><td align="left" valign="top">12 (955)</td><td align="left" valign="top">86%</td><td align="left" valign="top">SMD 0.53 (0.13 to 0.94)</td><td align="left" valign="top">&#x2212;0.81 to 1.88</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI moderated adaptive learning platform vs control</td><td align="left" valign="top">1 (40)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.68 (0.04 to 1.32)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>NLP rule-based chatbot vs control</td><td align="left" valign="top">3 (530)</td><td align="left" valign="top">97%</td><td align="left" valign="top">SMD 1.06 (&#x2212;2.19 to 4.32)</td><td align="left" valign="top">&#x2212;5.28 to 7.41</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI imaging diagnostic aid vs control</td><td align="left" valign="top">2 (69)</td><td align="left" valign="top">78%</td><td align="left" valign="top">SMD 1.26 (&#x2212;6.23 to 8.74)</td><td align="left" valign="top">&#x2212;10.73 to 13.24</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI-VR virtual doctor vs control</td><td align="left" valign="top">1 (64)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.67 (0.16 to 1.17)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Proportion with grade A or B: non-LLM AI gamification tool vs control</td><td align="left" valign="top">1 (73)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">RR 1.33 (1.01 to 1.74)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top" colspan="6">Kirkpatrick level 2: clinical skills</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM content generator vs control</td><td align="left" valign="top">2 (295)</td><td align="left" valign="top">97%</td><td align="left" valign="top">SMD 0.52 (&#x2212;8.66 to 9.69)</td><td align="left" valign="top">&#x2212;15.22 to 16.25</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM personalized learning aid vs control</td><td align="left" valign="top">9 (609)</td><td align="left" valign="top">83%</td><td align="left" valign="top">SMD 0.49 (0.00 to 0.97)</td><td align="left" valign="top">&#x2212;0.93 to 1.90</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM virtual patient vs control</td><td align="left" valign="top">1 (56)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 2.53 (1.82 to 3.25)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM content generator + LLM virtual patient + LLM personalized learning aid vs control</td><td align="left" valign="top">1 (88)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD &#x2212;0.11 (&#x2212;0.53 to 0.31)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM virtual patient + LLM personalized learning aid vs control</td><td align="left" valign="top">3 (124)</td><td align="left" valign="top">95%</td><td align="left" valign="top">SMD 1.82 (&#x2212;3.63 to 7.27)</td><td align="left" valign="top">&#x2212;8.61 to 12.24</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI imaging diagnostic aid vs control</td><td align="left" valign="top">4 (176)</td><td align="left" valign="top">62%</td><td align="left" valign="top">SMD 0.43 (&#x2212;0.41 to 1.27)</td><td align="left" valign="top">&#x2212;1.13 to 1.99</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI moderated adaptive learning platform vs control</td><td align="left" valign="top">2 (139)</td><td align="left" valign="top">98%</td><td align="left" valign="top">SMD 0.59 (&#x2212;19.98 to 21.17)</td><td align="left" valign="top">&#x2212;34.78 to 35.97</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>NLP rule-based virtual patient vs control</td><td align="left" valign="top">1 (79)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 2.03 (1.49 to 2.58)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Moderate</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>NLP rule-based chatbot + virtual patient vs control</td><td align="left" valign="top">1 (61)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.24 (&#x2212;0.26 to 0.74)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI-VR virtual doctor vs control</td><td align="left" valign="top">1 (64)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.21 (&#x2212;0.28 to 0.71)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI procedure assistant + AI-moderated adaptive learning platform vs control</td><td align="left" valign="top">1 (20)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.80 (&#x2212;0.12 to 1.72)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top" colspan="6">Kirkpatrick level 2: practical skills</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM personalized learning aid vs control</td><td align="left" valign="top">1 (187)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.67 (0.37 to 0.96)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI procedure assistant vs control</td><td align="left" valign="top">6 (305)</td><td align="left" valign="top">92%</td><td align="left" valign="top">SMD 0.18 (&#x2212;0.97 to 1.34)</td><td align="left" valign="top">&#x2212;2.77 to 3.13</td><td align="left" valign="top">Low</td></tr><tr><td align="left" valign="top" colspan="6">Kirkpatrick level 2: task efficiency (time taken to perform tasks)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM personalized learning aid vs control</td><td align="left" valign="top">2 (100)</td><td align="left" valign="top">47%</td><td align="left" valign="top">SMD &#x2212;0.15 (&#x2212;4.24 to 3.95)</td><td align="left" valign="top">&#x2212;5.99 to 5.69</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI imaging diagnostic aid vs control</td><td align="left" valign="top">1 (40)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 2.70 (1.82 to 3.58)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI procedure assistant vs control</td><td align="left" valign="top">2 (52)</td><td align="left" valign="top">88%</td><td align="left" valign="top">SMD &#x2212;1.26 (&#x2212;14.65 to 12.12)</td><td align="left" valign="top">&#x2212;23.49 to 20.96</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI procedure assistant + AI-moderated adaptive learning platform vs control</td><td align="left" valign="top">1 (20)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD &#x2212;0.87 (&#x2212;1.80 to 0.05)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top" colspan="6">Kirkpatrick level 2: generic or personal skills</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM personalized learning aid vs control</td><td align="left" valign="top">1 (101)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.45 (0.05 to 0.84)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM virtual patient vs control</td><td align="left" valign="top">1 (27)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.00 (&#x2212;1.06 to 1.06)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Very low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM-integrated curriculum vs control</td><td align="left" valign="top">1 (96)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.60 (0.19 to 1.01)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Moderate</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Non-LLM AI communication analysis vs control</td><td align="left" valign="top">1 (25)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 1.85 (0.88 to 2.81)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Low</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI-VR virtual doctor vs control</td><td align="left" valign="top">1 (64)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">SMD 0.31 (&#x2212;0.18 to 0.81)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">Low</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>GRADE: Grading of Recommendations, Assessment, Development, and Evaluation.</p></fn><fn id="table1fn2"><p><sup>b</sup>LLM: large language model.</p></fn><fn id="table1fn3"><p><sup>c</sup>SMD: standardized mean difference.</p></fn><fn id="table1fn4"><p><sup>d</sup>Not applicable.</p></fn><fn id="table1fn5"><p><sup>e</sup>AI: artificial intelligence.</p></fn><fn id="table1fn6"><p><sup>f</sup>NLP: natural language processing.</p></fn><fn id="table1fn7"><p><sup>g</sup>AI-VR: artificial intelligence&#x2013;virtual reality.</p></fn></table-wrap-foot></table-wrap><p>We performed random-effects meta-analysis via the inverse-variance method using Review Manager (RevMan) online software (Cochrane), with between-study variance estimated using the restricted maximum likelihood method. We presented all point estimates with their 95% CIs, generated using the Hartung-Knapp-Sidik-Jonkman method, as simulation studies suggest this approach provides more appropriate coverage probabilities, particularly when heterogeneity is substantial [<xref ref-type="bibr" rid="ref36">36</xref>,<xref ref-type="bibr" rid="ref37">37</xref>]. To identify and quantify heterogeneity, we visually inspected the forest plots and used the <italic>I</italic><sup>2</sup> statistic (detailed below in the &#x201C;Assessment of Heterogeneity&#x201D; subheading), with generation of a 95% prediction interval (PI) as an indication of the likely range of effects in individual studies [<xref ref-type="bibr" rid="ref35">35</xref>].</p></sec><sec id="s2-10"><title>Assessment of Heterogeneity</title><p>We evaluated heterogeneity statistically and explored plausible explanations in terms of educational characteristics of the studies [<xref ref-type="bibr" rid="ref38">38</xref>]. We used the <italic>I</italic><sup>2</sup> statistic with a cutoff of 50% to indicate substantial heterogeneity [<xref ref-type="bibr" rid="ref38">38</xref>]. If substantial heterogeneity was found, we explored educational characteristics via subgroup analyses in terms of the field or program evaluated, the region where the study was conducted, the category of AI intervention delivered (divided broadly into LLM and non-LLM), the major purpose of intervention (teaching-learning or assessment), and the frequency of intervention (single or multiple sessions across a period of time). As part of heterogeneity exploration, we generated a 95% PI, which indicated the expected range of true effects in similar individual studies [<xref ref-type="bibr" rid="ref39">39</xref>]. AI subcategories and outcome types were not part of the heterogeneity exploration, as we had already separated the analyses according to these subcategories, as well as grouped the outcomes according to Kirkpatrick levels.</p></sec><sec id="s2-11"><title>Sensitivity Analysis</title><p>We planned to conduct sensitivity analyses by restricting analyses to studies at low or probably low risk of bias for each domain where at least 5 studies were available in each risk stratum, consistent with Cochrane guidance [<xref ref-type="bibr" rid="ref35">35</xref>]. However, sensitivity analysis was only feasible based on 2 risk-of-bias domains and for only 1 analysis, as detailed below in the Results section.</p></sec><sec id="s2-12"><title>Publication and Reporting Bias Assessment</title><p>We created funnel plots to screen for small-study effects, a possible reason for publication bias, for outcomes in which there were &#x2265;10 studies. If a funnel plot showed significant asymmetry, we would downgrade the certainty of the evidence based on a strong suspicion of publication bias from small-study effects [<xref ref-type="bibr" rid="ref40">40</xref>]. However, in this review, the substantial heterogeneity observed across outcomes rendered funnel plot interpretation unreliable, especially when asymmetry is present [<xref ref-type="bibr" rid="ref41">41</xref>]. Nonetheless, the single funnel plot generated in our review did not show significant asymmetry. We did not specifically assess the risk of bias from missing results in a meta-analysis (ie, reporting biases), using recently introduced risk-of-bias tools, such as the RoB-ME tool developed by Page et al [<xref ref-type="bibr" rid="ref42">42</xref>].</p></sec><sec id="s2-13"><title>Certainty of Evidence Rating</title><p>Two authors (NML and QCO) independently assessed the certainty of the evidence for 7 major outcomes for each comparison, namely, self-efficacy or confidence, attitude or satisfaction, theoretical knowledge, clinical skills, practical skills, task efficiency, and generic or personal skills using the Grading of Recommendations, Assessment, Development, and Evaluation approach [<xref ref-type="bibr" rid="ref43">43</xref>]. We considered evidence from RCTs as high certainty to begin with, downgrading 1 level for serious (or 2 levels for very serious) limitations based upon 5 considerations: risk of bias, inconsistency across studies, indirectness of the evidence, imprecision of estimates, and publication bias (refer to Part 1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> for details). We used GRADEpro GDT (The GRADE Working Group) to create the summary of findings table (Part 3 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>).</p></sec><sec id="s2-14"><title>Ethical Considerations</title><p>This work is a systematic review and meta-analysis of previously published studies; therefore, no ethics approval or consent to participate was required. The review was registered in PROSPERO (CRD42021243832). We conducted the review following Cochrane methods [<xref ref-type="bibr" rid="ref38">38</xref>] and reported according to the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) 2020 guidelines [<xref ref-type="bibr" rid="ref44">44</xref>] (<xref ref-type="supplementary-material" rid="app3">Checklist 1</xref>), with particular reference to the PRISMA 2020 expanded checklist [<xref ref-type="bibr" rid="ref45">45</xref>] in the structure of our paper. Additional details on our methods, including changes from the protocol, are described in Part 1 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Study Selection</title><p>The initial search identified 39,783 records, with 35,785 records remaining after deduplication. The vast majority of the initial records (n=35,608) were clearly nonrelevant from an inspection of the title or abstract, and these were rejected outright. We shortlisted 177 papers that appeared relevant and assessed their full texts in depth, excluding an additional 73 papers due to mismatch in study design (n=31, 42.5% papers), population (not undergraduate students; n=14, 19.2% papers), or intervention (not AI; n=28, 38.4% papers) (refer to <xref ref-type="fig" rid="figure1">Figure 1</xref> for reasons for exclusion and Part 4 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> for the full list). We also identified 30 relevant ongoing studies. Finally, we included 74 records that described 66 studies in our analysis, as 3 studies had multiple records, including 3 with 2 records [<xref ref-type="bibr" rid="ref46">46</xref>-<xref ref-type="bibr" rid="ref48">48</xref>], 1 with 3 records [<xref ref-type="bibr" rid="ref49">49</xref>], and 1 with 4 records [<xref ref-type="bibr" rid="ref50">50</xref>]. The flow diagram of the studies from the initial search to meta-analysis is shown in <xref ref-type="fig" rid="figure1">Figure 1</xref>.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram depicting the process from initial screening to final inclusion. AI: artificial intelligence.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v12i1e88933_fig01.png"/></fig></sec><sec id="s3-2"><title>Study Characteristics</title><p>The characteristics of included studies are summarized as follows (refer to Part 5 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> for a detailed tabulation).</p></sec><sec id="s3-3"><title>Design and Setting</title><p>Out of 66 included studies (N=4911 participants) published between 2020 and 2026 [<xref ref-type="bibr" rid="ref46">46</xref>-<xref ref-type="bibr" rid="ref49">49</xref>,<xref ref-type="bibr" rid="ref51">51</xref>-<xref ref-type="bibr" rid="ref116">116</xref>], 58 (87.9%) were parallel RCTs, 7 (10.6%) were randomized crossover trials, and 1 (1.5%) was a cluster-randomized trial. Most studies were conducted in China (n=14, 21.2%), followed by Turkey (n=11, 16.7%), the United States (n=5, 7.6%), Canada (n=4, 6.1%), Germany and Taiwan (n=3 each, 4.5%), and Italy, South Korea, Hong Kong, Morocco, Singapore and the United Kingdom (n=2 each, 3%), with a single study conducted in each of Colombia, Denmark, France, India, Ireland, Israel, Japan, New Zealand, Norway, Pakistan, Spain, and Thailand, with 2 multicenter studies involving different countries [<xref ref-type="bibr" rid="ref66">66</xref>,<xref ref-type="bibr" rid="ref84">84</xref>]. The number of participants per study ranged from 16 [<xref ref-type="bibr" rid="ref47">47</xref>,<xref ref-type="bibr" rid="ref79">79</xref>] to 426 [<xref ref-type="bibr" rid="ref97">97</xref>], with 63 [<xref ref-type="bibr" rid="ref82">82</xref>,<xref ref-type="bibr" rid="ref102">102</xref>] as the median sample size.</p></sec><sec id="s3-4"><title>Population and Educational Field</title><p>Over half of the included studies enrolled students from medicine (n=35, 53%), followed by nursing (n=15, 22.7%), dentistry (n=9, 13.6%), physiotherapy, pharmacy, and health sciences (n=2 each, 3%), and optometry (n=1, 1.5%). Within each program, there was a wide range of disciplines assessed. In medicine, these included preclinical sciences (physiology, anatomy, pathology, pharmacology, biochemistry, immunology, microbiology, and histology), as well as clinical specialties such as internal medicine, cardiorespiratory medicine, gastroenterology, pulmonology, obstetrics and gynecology, ophthalmology, urology, radiology, neurosurgery, surgery, and emergency and critical care medicine; generic skills such as clinical reasoning, evidence-based medicine, and clinical decision-making were also represented. In nursing, subfields included surgical nursing, critical care nursing, sepsis care, clinical skills, communication skills, medical terminology, and patient education. In dentistry, areas covered included dental radiology, endodontics, restorative and operative skills, and prosthodontics. Physiotherapy studies addressed clinical reasoning and chronic low back pain rehabilitation, while health sciences studies covered infection control and chronic disease management. Specific practical skills assessed across fields included simulated brain tumor resection, cardiac ultrasonography, and dental access cavity preparation.</p></sec><sec id="s3-5"><title>Intervention</title><p>The studies used AI mainly as teaching-learning enhancement tools (n=54, 81.8%), as an assessment tool only (n=1, 1.5%) [<xref ref-type="bibr" rid="ref104">104</xref>], or both (11, 16.7%). LLM applications were used in 38 (57.6%) studies, primarily featuring ChatGPT (versions 3.5 and 4.0; Open AI), alongside Gemini (Google), Copilot (Microsoft), Perplexity (Perplexity AI, Inc), Claude (Anthropic), and specialized chatbots. The remaining 28 (42.4%) studies used non-LLM AI systems composed of specific algorithms such as convolutional neural networks, linear support vector machines, NLP rule-based architectures, AI-powered virtual doctor or diagnostic systems, or intelligent monitoring and assessment systems, without any explicit use of LLM.</p><p>The AI applications were further categorized according to their predominant functions as described in the study. Among LLM studies, the applications included personalized learning aids (n=20), clinical content generators (n=6), virtual patients (n=6), combined virtual patient and personalized learning aid (n=3), a multifunction LLM combining content generation, virtual patient simulation, and personalized feedback (n=1), a gamification tool (n=1), and 1 study described as a general &#x201C;LLM-integrated curriculum.&#x201D;</p><p>Among non-LLM studies, applications included AI procedure assistant (n=7), AI imaging diagnostic aid (n=7), NLP rule-based chatbot (n=4), and AI-moderated adaptive learning platform (n=4), as well as an artificial intelligence&#x2013;virtual reality (AI-VR) virtual doctor, AI communication analysis system, AI gamification tool, rule-based virtual patient, a combined NLP rule-based chatbot and virtual patient, and a combined procedure assistant with adaptive learning platform (n=1 each). These applications aimed to assist in achieving competencies from foundational knowledge to complex clinical decision-making. A working description of each AI subcategory is available in <xref ref-type="table" rid="table2">Table 2</xref>. This categorization recognizes potential overlap between subcategories, as some studies incorporate multiple AI functionalities or serve overlapping educational purposes.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Working description of artificial intelligence (AI) subcategories for the purpose of classification in this review.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">AI<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> subcategory</td><td align="left" valign="bottom">Description</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="2">LLM<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup>-based interventions</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM content generator</td><td align="left" valign="top">LLM-based applications generating educational materials (clinical vignettes, questions, explanatory text, case presentations, and learning guides).</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM gamification tool</td><td align="left" valign="top">LLM-powered applications incorporating game-based elements (escape rooms and interactive challenges) to enhance learning engagement.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM personalized learning aid</td><td align="left" valign="top">LLM-based conversational tutors providing on-demand responses, explanations, postsimulation debriefing, and individualized feedback.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM virtual patient</td><td align="left" valign="top">LLM applications simulating patient encounters for clinical history-taking and communication skills practice.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>LLM-integrated curriculum</td><td align="left" valign="top">Structured program with LLMs systematically embedded throughout curriculum delivery without specifying any component in learning.</td></tr><tr><td align="left" valign="top" colspan="2">Non-LLM AI interventions</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI imaging diagnostic aid</td><td align="left" valign="top">AI systems using CNNs<sup><xref ref-type="table-fn" rid="table2fn3">c</xref></sup> or machine learning to analyze medical images for diagnostic support, lesion detection, or educational annotation.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI gamification tool</td><td align="left" valign="top">AI-powered applications incorporating game design elements using specialized algorithms for interactive skill-based learning.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI procedure assistant</td><td align="left" valign="top">AI systems providing real-time guidance, quality assessment, or feedback during practical skills training using machine learning classifiers or computer vision.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI-moderated adaptive learning platform</td><td align="left" valign="top">Intelligent tutoring systems using algorithms (Deep Q-Networks and sentiment analysis) to personalize learning pathways and optimize content delivery.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI-VR<sup><xref ref-type="table-fn" rid="table2fn4">d</xref></sup> virtual doctor</td><td align="left" valign="top">AI-powered virtual physician characters in VR simulations generating dynamic clinical responses through AI algorithms.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Natural language processing (NLP) rule-based chatbot</td><td align="left" valign="top">NLP chatbots using structured decision trees and preprogrammed dialogue patterns to deliver content and provide feedback.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Rule-based virtual patient</td><td align="left" valign="top">Virtual patient systems using decision tree algorithms to simulate patient responses through predetermined pathways.</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>AI communication analysis</td><td align="left" valign="top">AI systems using computer vision or behavioral analysis to assess and provide feedback on learner communication skills.</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>AI: artificial intelligence.</p></fn><fn id="table2fn2"><p><sup>b</sup>LLM: large language model.</p></fn><fn id="table2fn3"><p><sup>c</sup>CNN: convolutional neural network.</p></fn><fn id="table2fn4"><p><sup>d</sup>VR: virtual reality. </p></fn></table-wrap-foot></table-wrap><p>Study durations varied considerably, from single sessions of under 3 hours (n=24, 36.4%), short-term from 1 day to 3 weeks (n=19, 28.8%), medium-term from 1 to 2 months (n=13, 19.7%), and long-term from 2 to 6 months (n=5, 7.6%), with 5 studies not specifying duration.</p></sec><sec id="s3-6"><title>Comparison</title><p>Over half of the studies (n=39, 59.1%) reported the control group as having received standard methods of learning as implemented in the existing curriculum, including student presentations, lectures, web-based tools, standard assessment procedures with existing technologies led by human tutors, and human-based simulation methods such as standardized patients, peer role-play, and real patient-based training. In 18 (27.3%) studies, authors reported that participants in the control group received the same educational content without AI tools. Four studies [<xref ref-type="bibr" rid="ref53">53</xref>,<xref ref-type="bibr" rid="ref72">72</xref>,<xref ref-type="bibr" rid="ref79">79</xref>,<xref ref-type="bibr" rid="ref100">100</xref>] included 2 control groups that used different sources of evidence: 1 from standard institutional resources and the other from external resources via online searches. In these studies, we selected the group that used standard institutional resources as the control group for our analysis.</p></sec><sec id="s3-7"><title>Outcome Assessment</title><p>According to the Kirkpatrick Model of Evaluation [<xref ref-type="bibr" rid="ref28">28</xref>], all studies measured outcomes either in level 1 (reaction) or level 2 (knowledge), and none reported outcomes in level 3 (real-life practice) or level 4 (health outcomes). The outcomes reported are summarized as follows:</p><list list-type="bullet"><list-item><p>Level 1: 36 (54.5%) studies reported attitude, perception, or satisfaction, and 21 (31.8%) studies reported self-efficacy or confidence.</p></list-item><list-item><p>Level 2: 26 (39.4%) studies reported theoretical knowledge, 30 (45.5%) studies reported clinical skills, 8 (12.1%) studies reported practical skills, 6 (9.1%) studies reported task efficiency, and 5 (7.6%) studies reported generic or personal skills.</p></list-item></list><p>Most outcomes were measured immediately after the intervention. Five studies [<xref ref-type="bibr" rid="ref56">56</xref>,<xref ref-type="bibr" rid="ref62">62</xref>,<xref ref-type="bibr" rid="ref67">67</xref>,<xref ref-type="bibr" rid="ref86">86</xref>,<xref ref-type="bibr" rid="ref90">90</xref>] in addition reported outcomes at intermediate time points, ranging from 1 week to 3 months after the intervention.</p></sec><sec id="s3-8"><title>Funding</title><p>Nine (13.6%) studies received funding from government-linked agencies, including national research councils and government ministries. Eighteen (27.2%) studies were funded by universities, hospitals, or dedicated research institutes. Three studies received contributions from commercial companies: 2 received in-kind contributions in the form of AI platform or device access [<xref ref-type="bibr" rid="ref55">55</xref>,<xref ref-type="bibr" rid="ref82">82</xref>], and 1 received direct financial sponsorship from a commercial AI company [<xref ref-type="bibr" rid="ref98">98</xref>]. Seventeen (25.8%) studies declared no funding, while 19 (28.8%) did not provide any funding statement.</p></sec><sec id="s3-9"><title>Risk of Bias in Studies</title><p>The full risk-of-bias assessment is shown in Part 6 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. There are significant methodological limitations across the included studies, particularly in allocation concealment and blinding. Random sequence generation was adequate in 44 (66.7%) studies, but allocation concealment was adequate in only 15 (22.7%) studies. Nonblinding of participants resulted in high or probably high risk of bias from differential participant expectations in 55 (83.3%) studies.</p><p>Blinding of education providers was similarly not achieved in all but 4 studies. The risk implications, however, differed by outcome type. For participants&#x2019; perception or satisfaction (assessed in 47 studies), nonblinding of providers was judged probably low risk in 40 (85.1%) studies, on the basis that anonymous self-reported evaluations of a teaching program by adult learners were unlikely to be materially influenced by unblinded providers. For self-efficacy or confidence (assessed in 24 studies), the majority (n=20, 83.3%) were judged probably high risk, as provider awareness of group allocation could plausibly influence student confidence through differential feedback, encouragement, and validation beyond the educational content itself; this concern did not apply to the 4 remaining studies, which used largely self-directed learning formats [<xref ref-type="bibr" rid="ref47">47</xref>,<xref ref-type="bibr" rid="ref68">68</xref>,<xref ref-type="bibr" rid="ref70">70</xref>,<xref ref-type="bibr" rid="ref80">80</xref>]. For objective outcomes, 40 (66.7%) studies were judged probably high risk and 20 (33.3%) probably low risk, the latter reflecting programs that were largely self-directed or sufficiently standardized to limit meaningful provider influence.</p><p>While none of the assessors for subjective, self-reported outcomes were blinded, blinding of outcome assessors was achieved in 29 out of 60 (48.3%) studies that evaluated objective outcomes, but almost all studies were judged to have low or probably low risk of bias because of clear documentation of objective assessment criteria, with only 2 studies [<xref ref-type="bibr" rid="ref110">110</xref>,<xref ref-type="bibr" rid="ref113">113</xref>] judged as probably high risk due to a lack of description of the scoring criteria despite reporting numerical checklists. In terms of missing outcome data, most studies (n=59, 89.4%) demonstrated low risk of bias.</p></sec><sec id="s3-10"><title>Results of Individual Studies</title><p>Overall, 60 studies (90.9%; n=4506 participants) contributed suitable data for meta-analysis. The large number of AI subcategories resulted in numerous comparisons, most comprising few studies with small sample sizes. The results of individual studies are presented in the forest plots in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>. None of the 7 randomized crossover trials provided separate data for the first period, and so we included data as reported in the paper where applicable.</p></sec><sec id="s3-11"><title>Results of Synthesis</title><p>A full display of the synthesis results is shown numerically in <xref ref-type="table" rid="table1">Table 1</xref> and graphically in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>. Substantial heterogeneity was present across almost all pooled estimates and was not adequately explained by subgroup analyses based on field and region of study, LLM vs non-LLM, teaching-learning vs assessment, and single vs multiple sessions (<xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>). Consequently, the certainty of evidence was rated low or very low for most comparisons (<xref ref-type="table" rid="table1">Table 1</xref>). Ninety-five percent PIs, reported for all comparisons with 2 or more studies, were wide throughout and almost universally included the null value, indicating highly uncertain effects of AI interventions in any new study setting.</p><p>The results are summarized below, organized by Kirkpatrick level.</p></sec><sec id="s3-12"><title>Kirkpatrick Level 1: Participant Perception or Satisfaction</title><p>Twenty-three (1679 participants) studies assessed participant perception or satisfaction across 9 AI subtype comparisons (<xref ref-type="table" rid="table1">Table 1</xref>).</p><p>Among LLM-based comparisons, LLM personalized learning aids appeared to show an important positive effect (7 studies, n=430; SMD 0.93, 95% CI 0.40 to 1.46; <italic>I</italic>&#x00B2;=74%; 95% PI &#x2212;0.40 to 2.26; <xref ref-type="fig" rid="figure2">Figure 2</xref> [<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref70">70</xref>,<xref ref-type="bibr" rid="ref80">80</xref>,<xref ref-type="bibr" rid="ref84">84</xref>,<xref ref-type="bibr" rid="ref113">113</xref>,<xref ref-type="bibr" rid="ref115">115</xref>,<xref ref-type="bibr" rid="ref116">116</xref>]), whereas LLM content generators (5 studies, n=509; SMD 0.65, 95% CI &#x2212;0.73 to 2.04; <italic>I</italic>&#x00B2;=96%; 95% PI &#x2212;2.67 to 3.98; <xref ref-type="fig" rid="figure3">Figure 3</xref> [<xref ref-type="bibr" rid="ref49">49</xref>,<xref ref-type="bibr" rid="ref51">51</xref>,<xref ref-type="bibr" rid="ref61">61</xref>,<xref ref-type="bibr" rid="ref62">62</xref>,<xref ref-type="bibr" rid="ref100">100</xref>]) and LLM virtual patients (3 studies, n=127; SMD 0.69, 95% CI &#x2212;0.83 to 2.21; <italic>I</italic>&#x00B2;=64%, 95% PI &#x2212;1.86 to 3.24) showed no clear difference, with substantial heterogeneity and wide PIs spanning the null. The remaining comparisons each involved 1 or 2 studies and were imprecise (<xref ref-type="table" rid="table2">Tables 1</xref> and <xref ref-type="table" rid="table1">2</xref>).</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Forest plot for perception or satisfaction (large language model [LLM] personalized learning aid vs control) [<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref70">70</xref>,<xref ref-type="bibr" rid="ref80">80</xref>,<xref ref-type="bibr" rid="ref84">84</xref>,<xref ref-type="bibr" rid="ref113">113</xref>,<xref ref-type="bibr" rid="ref115">115</xref>,<xref ref-type="bibr" rid="ref116">116</xref>].</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v12i1e88933_fig02.png"/></fig><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Forest plot for perception or satisfaction (large language model [LLM] content generator vs control) [<xref ref-type="bibr" rid="ref49">49</xref>,<xref ref-type="bibr" rid="ref51">51</xref>,<xref ref-type="bibr" rid="ref61">61</xref>,<xref ref-type="bibr" rid="ref62">62</xref>,<xref ref-type="bibr" rid="ref100">100</xref>].</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v12i1e88933_fig03.png"/></fig></sec><sec id="s3-13"><title>Self-Efficacy or Confidence</title><p>Sixteen studies (1115 participants) assessed self-efficacy or confidence across 9 AI subtype comparisons, and 1 additional study reported a dichotomous outcome (<xref ref-type="table" rid="table1">Table 1</xref>).</p><p>The largest comparison was LLM personalized learning aids, which appeared to show an important positive effect (7 studies, n=609; SMD 0.91, 95% CI 0.54 to 1.29; <italic>I</italic>&#x00B2;=64%, 95% PI 0.05 to 1.77; very low certainty; <xref ref-type="fig" rid="figure4">Figure 4</xref> [<xref ref-type="bibr" rid="ref53">53</xref>,<xref ref-type="bibr" rid="ref70">70</xref>,<xref ref-type="bibr" rid="ref80">80</xref>,<xref ref-type="bibr" rid="ref84">84</xref>,<xref ref-type="bibr" rid="ref113">113</xref>,<xref ref-type="bibr" rid="ref115">115</xref>,<xref ref-type="bibr" rid="ref116">116</xref>]). All other comparisons were based on 1 or 2 studies with no clear differences between groups. One notable single-study finding was a non-LLM AI-VR virtual doctor that showed reduced self-efficacy compared with control (SMD &#x2212;0.68, 95% CI &#x2212;1.18 to &#x2212;0.17; very low certainty). A separate single study reported the proportion of participants confident in echocardiography views using a non-LLM AI procedure assistant, with no clear difference (RR 1.26, 95% CI 0.45 to 3.50; very low certainty).</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Forest plot for self-efficacy or confidence (large language model [LLM] personalized learning aid vs control) [<xref ref-type="bibr" rid="ref53">53</xref>,<xref ref-type="bibr" rid="ref70">70</xref>,<xref ref-type="bibr" rid="ref80">80</xref>,<xref ref-type="bibr" rid="ref84">84</xref>,<xref ref-type="bibr" rid="ref113">113</xref>,<xref ref-type="bibr" rid="ref115">115</xref>,<xref ref-type="bibr" rid="ref116">116</xref>].</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v12i1e88933_fig04.png"/></fig></sec><sec id="s3-14"><title>Kirkpatrick Level 2: Theoretical Knowledge</title><p>Twenty-four studies (2126 participants) assessed theoretical knowledge across 7 AI subtype comparisons, and 1 additional study reported a dichotomous outcome (<xref ref-type="table" rid="table1">Table 1</xref>).</p><p>The most informative subtype comparison was LLM personalized learning aids, which appeared to show an important positive effect but with substantial heterogeneity (12 studies, n=955; SMD 0.53, 95% CI 0.13 to 0.94; <italic>I</italic>&#x00B2;=86%, 95% PI &#x2212;0.81 to 1.88; very low certainty; <xref ref-type="fig" rid="figure5">Figure 5</xref> [<xref ref-type="bibr" rid="ref53">53</xref>,<xref ref-type="bibr" rid="ref63">63</xref>,<xref ref-type="bibr" rid="ref67">67</xref>,<xref ref-type="bibr" rid="ref70">70</xref>-<xref ref-type="bibr" rid="ref72">72</xref>,<xref ref-type="bibr" rid="ref80">80</xref>,<xref ref-type="bibr" rid="ref82">82</xref>,<xref ref-type="bibr" rid="ref84">84</xref>,<xref ref-type="bibr" rid="ref86">86</xref>,<xref ref-type="bibr" rid="ref90">90</xref>,<xref ref-type="bibr" rid="ref113">113</xref>]; publication bias not suggested from funnel plot, Part 8 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). LLM content generators (3 studies, n=359; SMD 0.99, 95% CI &#x2212;1.04 to 3.01; <italic>I</italic>&#x00B2;=80%, 95% PI &#x2212;2.96 to 4.93) and NLP rule-based chatbots (3 studies, n=530; SMD 1.06, 95% CI &#x2212;2.19 to 4.32; <italic>I</italic>&#x00B2;=65%, 95% PI &#x2212;5.28 to 7.41) showed imprecise results with substantial heterogeneity and very wide PIs, all of very low certainty. Single-study comparisons of LLM gamification, non-LLM AI-moderated adaptive learning, and AI-VR virtual doctor each appeared to favor AI, while a non-LLM AI imaging diagnostic aid comparison (2 studies) was inconclusive (<xref ref-type="table" rid="table1">Table 1</xref>). One small study reported the proportion achieving grade A or B with a non-LLM AI gamification tool (RR 1.33, 95% CI 1.01 to 1.74; very low certainty).</p><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>Forest plot for theoretical knowledge (large language model [LLM] personalized learning aid vs control) [<xref ref-type="bibr" rid="ref53">53</xref>,<xref ref-type="bibr" rid="ref63">63</xref>,<xref ref-type="bibr" rid="ref67">67</xref>,<xref ref-type="bibr" rid="ref70">70</xref>-<xref ref-type="bibr" rid="ref72">72</xref>,<xref ref-type="bibr" rid="ref80">80</xref>,<xref ref-type="bibr" rid="ref82">82</xref>,<xref ref-type="bibr" rid="ref84">84</xref>,<xref ref-type="bibr" rid="ref86">86</xref>,<xref ref-type="bibr" rid="ref90">90</xref>,<xref ref-type="bibr" rid="ref113">113</xref>].</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v12i1e88933_fig05.png"/></fig></sec><sec id="s3-15"><title>Clinical Skills</title><p>Twenty-six studies (n=1711) assessed clinical skills across 11 AI subtype comparisons (<xref ref-type="table" rid="table1">Table 1</xref>). The largest comparisons were LLM personalized learning aids (9 studies, n=609; SMD 0.49, 95% CI 0.00 to 0.97; <italic>I</italic>&#x00B2;=83%, 95% PI &#x2212;0.93 to 1.90; very low certainty; <xref ref-type="fig" rid="figure6">Figure 6</xref> [<xref ref-type="bibr" rid="ref71">71</xref>,<xref ref-type="bibr" rid="ref81">81</xref>,<xref ref-type="bibr" rid="ref84">84</xref>,<xref ref-type="bibr" rid="ref86">86</xref>,<xref ref-type="bibr" rid="ref91">91</xref>,<xref ref-type="bibr" rid="ref113">113</xref>,<xref ref-type="bibr" rid="ref115">115</xref>,<xref ref-type="bibr" rid="ref116">116</xref>]) and non-LLM imaging diagnostic aid (4 studies, n=176; SMD 0.43, 95% CI &#x2212;0.41 to 1.27; <italic>I</italic>&#x00B2;=62%, 95% PI &#x2212;1.13 to 1.99; very low certainty; <xref ref-type="fig" rid="figure7">Figure 7</xref> [<xref ref-type="bibr" rid="ref56">56</xref>,<xref ref-type="bibr" rid="ref60">60</xref>,<xref ref-type="bibr" rid="ref85">85</xref>,<xref ref-type="bibr" rid="ref99">99</xref>]), which showed no clear differences between groups and substantial heterogeneity. A single study of a NLP rule-based virtual patient showed a large effect (SMD 2.03, 95% CI 1.49 to 2.58; moderate certainty), and a single LLM virtual patient study similarly showed a large effect (SMD 2.53, 95% CI 1.82 to 3.25; low certainty). The rest of the comparisons were inconclusive.</p><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>Forest plot for clinical skills (large language model [LLM] personalized learning aid vs control) [<xref ref-type="bibr" rid="ref71">71</xref>,<xref ref-type="bibr" rid="ref81">81</xref>,<xref ref-type="bibr" rid="ref84">84</xref>,<xref ref-type="bibr" rid="ref86">86</xref>,<xref ref-type="bibr" rid="ref91">91</xref>,<xref ref-type="bibr" rid="ref113">113</xref>,<xref ref-type="bibr" rid="ref115">115</xref>,<xref ref-type="bibr" rid="ref116">116</xref>].</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v12i1e88933_fig06.png"/></fig><fig position="float" id="figure7"><label>Figure 7.</label><caption><p>Forest plot for clinical skills (non&#x2013;large language model [LLM] imaging diagnostic aid vs control) [<xref ref-type="bibr" rid="ref56">56</xref>,<xref ref-type="bibr" rid="ref60">60</xref>,<xref ref-type="bibr" rid="ref85">85</xref>,<xref ref-type="bibr" rid="ref99">99</xref>].</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v12i1e88933_fig07.png"/></fig></sec><sec id="s3-16"><title>Practical Skills</title><p>Seven studies (492 participants) assessed practical skills (<xref ref-type="table" rid="table1">Table 1</xref>). One LLM personalized learning aid study (187 participants) appeared to show an important improvement (SMD 0.67, 95% CI 0.37 to 0.96; low certainty), whereas 6 studies of non-LLM AI procedure assistants (=305 participants) showed no clear difference but substantial heterogeneity (SMD 0.18, 95% CI &#x2212;0.97 to 1.34; <italic>I</italic>&#x00B2;=92%, 95% PI &#x2212;2.77 to 3.13; low certainty; <xref ref-type="fig" rid="figure8">Figure 8</xref> [<xref ref-type="bibr" rid="ref49">49</xref>,<xref ref-type="bibr" rid="ref50">50</xref>,<xref ref-type="bibr" rid="ref55">55</xref>,<xref ref-type="bibr" rid="ref58">58</xref>,<xref ref-type="bibr" rid="ref69">69</xref>,<xref ref-type="bibr" rid="ref105">105</xref>]).</p><fig position="float" id="figure8"><label>Figure 8.</label><caption><p>Forest plot for practical skills (non&#x2013;large language model [LLM] artificial intelligence [AI] procedure assistant vs control) [<xref ref-type="bibr" rid="ref49">49</xref>,<xref ref-type="bibr" rid="ref50">50</xref>,<xref ref-type="bibr" rid="ref55">55</xref>,<xref ref-type="bibr" rid="ref58">58</xref>,<xref ref-type="bibr" rid="ref69">69</xref>,<xref ref-type="bibr" rid="ref105">105</xref>].</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v12i1e88933_fig08.png"/></fig></sec><sec id="s3-17"><title>Task Efficiency</title><p>Six studies (212 participants) measured task efficiency, defined as time taken to complete tasks, across 4 subtype comparisons (<xref ref-type="table" rid="table1">Table 1</xref>). A single study of a non-LLM AI imaging diagnostic aid showed markedly increased time required to complete task (SMD 2.70, 95% CI 1.82 to 3.58), whereas other comparisons showed no clear differences, all with very low certainty.</p></sec><sec id="s3-18"><title>Generic or Personal Skills</title><p>Five studies (313 participants) assessed generic or personal skills across 5 single-study comparisons (<xref ref-type="table" rid="table1">Table 1</xref>). A single LLM-integrated curriculum study appeared to show an important improvement (SMD 0.60, 95% CI 0.19-1.01; moderate certainty), as did a single non-LLM AI communication analysis study (SMD 1.85, 95% CI 0.88 to 2.81; low certainty). A single LLM personalized learning aid study showed a small improvement (SMD 0.45, 95% CI 0.05 to 0.84; low certainty). The remaining comparisons showed no clear differences.</p></sec><sec id="s3-19"><title>Investigation of Heterogeneity</title><p>A detailed report on the results of our heterogeneity exploration is available in Part 7 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. In summary, there was a very high degree of heterogeneity across almost all outcomes (<italic>I</italic>&#x00B2;: 86%&#x2010;94%) with no consistent studies or group of studies identified as the chief contributors. Subgroup analyses (by field and region of study, LLM vs non-LLM, predominant function, and single vs multiple sessions), conducted for all outcomes with sufficient studies (<xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>), revealed no consistent explanation of heterogeneity, as none of these factors consistently moderated the direction or magnitude of the pooled estimates.</p></sec><sec id="s3-20"><title>Sensitivity Analysis</title><p>Sensitivity analysis was only feasible for the domains of random sequence generation and allocation concealment, and for the analysis of theoretical knowledge (comparing LLM personalized learning aid against control), as this was the only analysis with at least 5 studies in more than 1 risk stratum. In this analysis, after removing 7 studies with high risk of bias in allocation concealment, the pooled estimate of 5 remaining studies changed from showing a significant improvement in knowledge to no significant difference between groups. This might indicate that low-risk studies tended to show more conservative estimates, or reflected reduced statistical power from the smaller number of remaining studies. We could not perform sensitivity analysis to exclude crossover trials because in none of the analyses where these trials were included, there were sufficient studies in each stratum to enable a meaningful analysis.</p></sec><sec id="s3-21"><title>Certainty of Evidence</title><p>Certainty of evidence was low to very low for most outcomes, due mainly to study limitations (high risk of bias across multiple domains), inconsistency (high degree of heterogeneity with wide 95% PIs), and imprecision (wide 95% CIs crossing multiple effect thresholds; for details refer to Part 3 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>).</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>This review offers the first comprehensive RCT synthesis of AI in undergraduate health professions education, with granular subcategorization by technology type and educational function across diverse global settings. However, the wide diversity of AI subcategories produced numerous specific analyses, most containing few studies with imprecise, very-low-certainty estimates, leaving the body of evidence far from sufficient to inform educational practice, despite the overall large number of included RCTs.</p><p>Based predominantly on very low-certainty evidence, the findings suggest that certain AI intervention types may improve specific educational outcomes, but the pattern of benefit is inconsistent. LLM-based personalized learning aids, with comparatively the largest volume of evidence across outcomes, show a possible positive effect in perception, confidence, and knowledge, but the PIs of these outcomes were wide and mostly crossing the null, indicating substantial uncertainty about the likely effect in future educational contexts.</p><p>Across all other AI subcategories, point estimates were generally positive but CIs and PIs were wide and typically crossed the null, precluding firm conclusions. No studies assessed Kirkpatrick Level 3 (real-world practice change) or Level 4 (health outcomes), where evidence would be most compelling for educational policy.</p><p>The subcategorized analyses also clarify which comparisons remain informationally sparse. Several AI subtypes, particularly multicomponent interventions and novel tool combinations, contributed only single-study estimates, highlighting where future research is most needed.</p></sec><sec id="s4-2"><title>Comparison With Prior Work</title><p>Our findings build on and extend previous syntheses in this field. Before this work, approximately 50 reviews examined AI in health education (Part 9 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>), but the majority are scoping or narrative reviews, and most express cautious optimism without quantitative evidence from RCTs. Two meta-analyses evaluating LLMs in national licensing examinations reported a wide performance range and concluded that further model development and evaluation are needed before LLMs can be recommended as primary teaching tools [<xref ref-type="bibr" rid="ref117">117</xref>,<xref ref-type="bibr" rid="ref118">118</xref>]. A scoping review of AI across medical education similarly concluded that AI applications show promise but require rigorous evaluation [<xref ref-type="bibr" rid="ref10">10</xref>].</p><p>Most directly comparable to our work is a recent meta-analysis of 11 RCTs (786 participants) examining generative AI in medical education, which reported no clear difference in overall knowledge acquisition but improved practical skills and satisfaction [<xref ref-type="bibr" rid="ref22">22</xref>]. Our review extends this in several important respects: a more recent search (January 2026 vs January 2025), broader scope (all AI applications and health disciplines, not only generative AI in medicine), larger evidence base (60 RCTs, 4506 participants in meta-analysis), systematic subcategorization of AI intervention types, outcome classification using Kirkpatrick framework, risk of bias assessment using ROBUST-RCT, and, crucially, Grading of Recommendations, Assessment, Development, and Evaluation certainty-of-evidence ratings, absent from all previously published reviews of this topic. Without certainty ratings, positive point estimates risk being interpreted as practice-ready evidence; yet, in almost every comparison, certainty is low or very low, and PIs indicate that effects may not reliably reproduce across settings.</p></sec><sec id="s4-3"><title>Limitations</title><p>We acknowledge several limitations of the evidence gathered and the review process. In terms of the evidence gathered, despite comprehensive and up-to-date searches, several limitations require acknowledgment. First, even after subcategorization, substantial residual heterogeneity remained within several comparisons (<italic>I</italic>&#x00B2;: 64%&#x2010;98%), which was not adequately explained by the study characteristics reported. This heterogeneity likely reflects the inherently context-dependent nature of educational interventions, where learner characteristics, instructor expertise, institutional culture, implementation fidelity, and assessment approaches interact in ways that are rarely reported in sufficient detail [<xref ref-type="bibr" rid="ref119">119</xref>,<xref ref-type="bibr" rid="ref120">120</xref>]. Rather than invalidating the findings, this suggests the effectiveness of AI tools may be context-sensitive, shaped by local factors rarely reported in sufficient detail. The consistent direction of effect (most studies favoring AI) provides a preliminary signal of benefit, but the magnitude of any effect in a given setting remains uncertain.</p><p>Second, granular subcategorization, while educationally more informative, resulted in many comparisons based on only 1 or 2 studies, which are too imprecise to provide actionable guidance for practice.</p><p>Next, we included 7 crossover trials [<xref ref-type="bibr" rid="ref47">47</xref>,<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref53">53</xref>,<xref ref-type="bibr" rid="ref66">66</xref>,<xref ref-type="bibr" rid="ref74">74</xref>,<xref ref-type="bibr" rid="ref80">80</xref>,<xref ref-type="bibr" rid="ref81">81</xref>], which may have additional concerns in their study design that we have not adequately addressed, because the newly established ROBUST-RCT tool has not been expanded to cover crossover design [<xref ref-type="bibr" rid="ref29">29</xref>]. Additionally, none of these studies provided data separately for the first period, as desired. However, given that most included studies were already rated at high risk of bias for allocation concealment and blinding and the evidence certainty was very low for most outcomes, this limitation is unlikely to substantially alter the overall certainty assessments, but cannot be dismissed.</p><p>In terms of the review process, despite screening nearly 40,000 records and shortlisting 177 potentially eligible records, we may have missed studies evaluating AI-enabled tools (such as robotic or VR-based systems) not described as AI interventions in their reports. Our risk of bias assessment and decisions on which outcomes to include for meta-analysis, although transparent, involved some subjective judgment and might have been influenced by personal biases. Pooling continuous outcomes using SMD across heterogeneous instruments, particularly for subjective outcomes such as satisfaction and self-efficacy, is another limitation of this review; however, we incorporated an SMD of 0.5 as a default minimally important difference and PIs to improve interpretability in educational practice. Finally, our subcategorization of specific AI applications was made for the purpose of the current review, and our decision to assign one study to one subcategory or another, although largely reliant on the information provided by the study, involved subjective judgment. However, we hope this taxonomy provides a working framework for future syntheses as the evidence base grows.</p></sec><sec id="s4-4"><title>Conclusions and Implications</title><p>Early evidence from RCTs suggests that the effects of AI applications are neither uniform across AI types nor consistent across settings. Although certain AI intervention types used in undergraduate health professions education appeared to improve some educational outcomes, particularly satisfaction and theoretical knowledge with LLM-based personalized learning aids, the findings are highly heterogeneous with low- to very low&#x2013;certainty evidence and are far from sufficient in informing educational practice. No studies have yet assessed real-world practice change or health-related outcomes, which are the levels of evidence most relevant to educational policy. We recommend that AI applications in undergraduate health education continue to be used on a trial basis. Future RCTs should be adequately powered and use robust allocation concealment, such as central randomization or online systems with real-time allocation, with transparent documentation. Standardized outcome measures should be used to minimize performance and detection bias, with clear reporting of implementation details to support meaningful subgroup and heterogeneity analyses. Studies should also begin to assess longer-term Kirkpatrick Level 3 and 4 outcomes, which reflect real-world behavior change and patient health outcomes, as these would particularly strengthen the evidence base for educational guideline adoption.</p></sec></sec></body><back><ack><p>We used Claude Sonnet 4.5 (Anthropic), a large language model [<xref ref-type="bibr" rid="ref27">27</xref>], for preliminary data extraction including study characteristics in terms of population, intervention, comparison, outcomes, and methodologies, as well as final proofreading of the manuscript draft, and used Elicit AI (Ought) [<xref ref-type="bibr" rid="ref121">121</xref>] on a trial basis to identify relevant papers for our background and discussion. Examples of prompts for each purpose mentioned, as well as human verification and the decision process are available in Part 10 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>. However, the human authors performed final verification, finalized the extracted data, selected the relevant citations to be included in the write-up, and wrote this manuscript.</p></ack><notes><sec><title>Funding</title><p>No external financial support or grants were received for this work.</p></sec><sec><title>Data Availability</title><p>The datasets used and analyzed during this study, including the complete data extraction forms, risk-of-bias assessments, forest plots, and RevMan analysis files, are available from the corresponding author on reasonable request. The PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) 2020 checklist and detailed search strategies are provided in <xref ref-type="supplementary-material" rid="app3">Checklist 1</xref> and <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>, respectively.</p></sec></notes><fn-group><fn fn-type="con"><p>NML conceptualized the review, drafted the protocol including methodology, performed the search, extracted data with assistance from Claude Sonnet, performed risk-of-bias assessment, curated the data for meta-analysis and certainty-of-evidence rating, drafted the initial manuscript, and critically reviewed and revised the manuscript. QCO performed data extraction, checked data accuracy, performed risk-of-bias assessment and certainty-of-evidence rating, and critically reviewed and revised the manuscript. MTW and PB drafted part of the first draft, critically reviewed and revised the manuscript. YSL checked data accuracy, critically reviewed and revised the manuscript. PT critically reviewed and revised the manuscript.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">AI-VR</term><def><p>artificial intelligence&#x2013;virtual reality</p></def></def-item><def-item><term id="abb3">DL</term><def><p>deep learning</p></def></def-item><def-item><term id="abb4">LLM</term><def><p>large language model</p></def></def-item><def-item><term id="abb5">ML</term><def><p>machine learning</p></def></def-item><def-item><term id="abb6">PI</term><def><p>prediction interval</p></def></def-item><def-item><term id="abb7">PRISMA</term><def><p>Preferred Reporting Items for Systematic Reviews and Meta-Analyses</p></def></def-item><def-item><term id="abb8">PRISMA-S</term><def><p>Preferred Reporting Items for Systematic Reviews and Meta-Analyses&#x2013;Search extension</p></def></def-item><def-item><term id="abb9">RCT</term><def><p>randomized controlled trial</p></def></def-item><def-item><term id="abb10">ROBUST-RCT</term><def><p>Risk Of Bias Instrument for Use in Systematic Reviews for Randomised Controlled Trials</p></def></def-item><def-item><term id="abb11">RR</term><def><p>relative risk</p></def></def-item><def-item><term id="abb12">SMD</term><def><p>standardized mean difference</p></def></def-item><def-item><term id="abb13">WHO</term><def><p>World Health Organization</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="report"><article-title>Healthcare education market (2025-2033) size, share &#x0026; trends analysis report by providers (continuing medical education providers, educational platforms, learning management systems, universities &#x0026; academic centers), by application, by delivery mode, by end use, by region, and segment forecasts</article-title><year>2024</year><access-date>2025-08-18</access-date><publisher-name>Grand View Research</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://www.grandviewresearch.com/industry-analysis/healthcare-education-market-report">https://www.grandviewresearch.com/industry-analysis/healthcare-education-market-report</ext-link></comment></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Alotaibi</surname><given-names>N</given-names> </name><name name-style="western"><surname>Wilson</surname><given-names>CB</given-names> </name><name name-style="western"><surname>Traynor</surname><given-names>M</given-names> </name></person-group><article-title>Enhancing digital readiness and capability in healthcare: a systematic review of interventions, barriers, and facilitators</article-title><source>BMC Health Serv Res</source><year>2025</year><month>04</month><day>4</day><volume>25</volume><issue>1</issue><fpage>500</fpage><pub-id pub-id-type="doi">10.1186/s12913-025-12663-3</pub-id><pub-id pub-id-type="medline">40186200</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Grainger</surname><given-names>R</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Gladman</surname><given-names>T</given-names> </name></person-group><article-title>Learning technology in health professions education: realising an (un)imagined future</article-title><source>Med Educ</source><year>2024</year><month>01</month><volume>58</volume><issue>1</issue><fpage>36</fpage><lpage>46</lpage><pub-id pub-id-type="doi">10.1111/medu.15185</pub-id><pub-id pub-id-type="medline">37555302</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Slavin</surname><given-names>S</given-names> </name><name name-style="western"><surname>D&#x2019;Eon</surname><given-names>MF</given-names> </name></person-group><article-title>Overcrowded curriculum is an impediment to change (Part A)</article-title><source>Can Med Educ J</source><year>2021</year><month>09</month><volume>12</volume><issue>4</issue><fpage>1</fpage><lpage>6</lpage><pub-id pub-id-type="doi">10.36834/cmej.73532</pub-id><pub-id pub-id-type="medline">34567300</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Talwar</surname><given-names>S</given-names> </name><name name-style="western"><surname>Dhir</surname><given-names>A</given-names> </name><name name-style="western"><surname>Islam</surname><given-names>N</given-names> </name><name name-style="western"><surname>Kaur</surname><given-names>P</given-names> </name><name name-style="western"><surname>Almusharraf</surname><given-names>A</given-names> </name></person-group><article-title>Resistance of multiple stakeholders to e-health innovations: integration of fundamental insights and guiding research paths</article-title><source>J Bus Res</source><year>2023</year><month>11</month><volume>166</volume><fpage>114135</fpage><pub-id pub-id-type="doi">10.1016/j.jbusres.2023.114135</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Caverzagie</surname><given-names>KJ</given-names> </name><name name-style="western"><surname>Nousiainen</surname><given-names>MT</given-names> </name><name name-style="western"><surname>Ferguson</surname><given-names>PC</given-names> </name><etal/></person-group><article-title>Overarching challenges to the implementation of competency-based medical education</article-title><source>Med Teach</source><year>2017</year><month>06</month><volume>39</volume><issue>6</issue><fpage>588</fpage><lpage>593</lpage><pub-id pub-id-type="doi">10.1080/0142159X.2017.1315075</pub-id><pub-id pub-id-type="medline">28598747</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kumaravel</surname><given-names>B</given-names> </name><name name-style="western"><surname>Hearn</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Jahangiri</surname><given-names>L</given-names> </name><name name-style="western"><surname>Pollard</surname><given-names>R</given-names> </name><name name-style="western"><surname>Stocker</surname><given-names>CJ</given-names> </name><name name-style="western"><surname>Nunan</surname><given-names>D</given-names> </name></person-group><article-title>A systematic review and taxonomy of tools for evaluating evidence-based medicine teaching in medical education</article-title><source>Syst Rev</source><year>2020</year><month>04</month><day>24</day><volume>9</volume><issue>1</issue><fpage>91</fpage><pub-id pub-id-type="doi">10.1186/s13643-020-01311-y</pub-id><pub-id pub-id-type="medline">32331530</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McCluskey</surname><given-names>A</given-names> </name><name name-style="western"><surname>Lovarini</surname><given-names>M</given-names> </name></person-group><article-title>Providing education on evidence-based practice improved knowledge but did not change behaviour: a before and after study</article-title><source>BMC Med Educ</source><year>2005</year><month>12</month><day>19</day><volume>5</volume><fpage>40</fpage><pub-id pub-id-type="doi">10.1186/1472-6920-5-40</pub-id><pub-id pub-id-type="medline">16364181</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Salehi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Ballen</surname><given-names>CJ</given-names> </name><name name-style="western"><surname>Bolander Laksov</surname><given-names>K</given-names> </name><etal/></person-group><article-title>Global perspectives of the impact of the COVID-19 pandemic on learning science in higher education</article-title><source>PLoS One</source><year>2023</year><volume>18</volume><issue>12</issue><fpage>e0294821</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0294821</pub-id><pub-id pub-id-type="medline">38060473</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gordon</surname><given-names>M</given-names> </name><name name-style="western"><surname>Daniel</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ajiboye</surname><given-names>A</given-names> </name><etal/></person-group><article-title>A scoping review of artificial intelligence in medical education: BEME Guide No. 84</article-title><source>Med Teach</source><year>2024</year><month>04</month><volume>46</volume><issue>4</issue><fpage>446</fpage><lpage>470</lpage><pub-id pub-id-type="doi">10.1080/0142159X.2024.2314198</pub-id><pub-id pub-id-type="medline">38423127</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McCarthy</surname><given-names>J</given-names> </name><name name-style="western"><surname>Minsky</surname><given-names>ML</given-names> </name><name name-style="western"><surname>Rochester</surname><given-names>N</given-names> </name><name name-style="western"><surname>Shannon</surname><given-names>CE</given-names> </name></person-group><article-title>A proposal for the Dartmouth summer research project on artificial intelligence: August 31, 1955</article-title><source>AI Mag</source><year>1955</year><month>08</month><day>31</day><volume>27</volume><issue>4</issue><fpage>12</fpage><pub-id pub-id-type="doi">10.1609/aimag.v27i4.1904</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ali</surname><given-names>O</given-names> </name><name name-style="western"><surname>Abdelbaki</surname><given-names>W</given-names> </name><name name-style="western"><surname>Shrestha</surname><given-names>A</given-names> </name><name name-style="western"><surname>Elbasi</surname><given-names>E</given-names> </name><name name-style="western"><surname>Alryalat</surname><given-names>MAA</given-names> </name><name name-style="western"><surname>Dwivedi</surname><given-names>YK</given-names> </name></person-group><article-title>A systematic literature review of artificial intelligence in the healthcare sector: benefits, challenges, methodologies, and functionalities</article-title><source>J Innov Knowl</source><year>2023</year><month>01</month><volume>8</volume><issue>1</issue><fpage>100333</fpage><pub-id pub-id-type="doi">10.1016/j.jik.2023.100333</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="web"><article-title>Introducing ChatGPT</article-title><source>OpenAI</source><year>2022</year><access-date>2025-08-18</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://openai.com/blog/chatgpt">https://openai.com/blog/chatgpt</ext-link></comment></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mukherjee</surname><given-names>M</given-names> </name><name name-style="western"><surname>Le</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chow</surname><given-names>YW</given-names> </name></person-group><article-title>Generative AI-enhanced intelligent tutoring system for graduate cybersecurity programs</article-title><source>Future Internet</source><year>2025</year><volume>17</volume><issue>4</issue><fpage>154</fpage><pub-id pub-id-type="doi">10.3390/fi17040154</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Caudell</surname><given-names>TP</given-names> </name><name name-style="western"><surname>Summers</surname><given-names>KL</given-names> </name><name name-style="western"><surname>Holten</surname><given-names>J</given-names>  <suffix>4th</suffix></name><etal/></person-group><article-title>Virtual patient simulator for distributed collaborative medical education</article-title><source>Anat Rec B New Anat</source><year>2003</year><month>01</month><volume>270</volume><issue>1</issue><fpage>23</fpage><lpage>29</lpage><pub-id pub-id-type="doi">10.1002/ar.b.10007</pub-id><pub-id pub-id-type="medline">12526063</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mirchi</surname><given-names>N</given-names> </name><name name-style="western"><surname>Bissonnette</surname><given-names>V</given-names> </name><name name-style="western"><surname>Yilmaz</surname><given-names>R</given-names> </name><name name-style="western"><surname>Ledwos</surname><given-names>N</given-names> </name><name name-style="western"><surname>Winkler-Schwartz</surname><given-names>A</given-names> </name><name name-style="western"><surname>Del Maestro</surname><given-names>RF</given-names> </name></person-group><article-title>The Virtual Operative Assistant: an explainable artificial intelligence tool for simulation-based training in surgery and medicine</article-title><source>PLoS ONE</source><year>2020</year><volume>15</volume><issue>2</issue><fpage>e0229596</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0229596</pub-id><pub-id pub-id-type="medline">32106247</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mastour</surname><given-names>H</given-names> </name><name name-style="western"><surname>Dehghani</surname><given-names>T</given-names> </name><name name-style="western"><surname>Moradi</surname><given-names>E</given-names> </name><name name-style="western"><surname>Eslami</surname><given-names>S</given-names> </name></person-group><article-title>Explainable artificial intelligence for predicting medical students&#x2019; performance in comprehensive assessments</article-title><source>Sci Rep</source><year>2025</year><month>07</month><day>3</day><volume>15</volume><issue>1</issue><fpage>23752</fpage><pub-id pub-id-type="doi">10.1038/s41598-025-07460-1</pub-id><pub-id pub-id-type="medline">40610576</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Vrdoljak</surname><given-names>J</given-names> </name><name name-style="western"><surname>Boban</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Vilovi&#x0107;</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kumri&#x0107;</surname><given-names>M</given-names> </name><name name-style="western"><surname>Bo&#x017E;i&#x0107;</surname><given-names>J</given-names> </name></person-group><article-title>A review of large language models in medical education, clinical decision support, and healthcare administration</article-title><source>Healthcare (Basel)</source><year>2025</year><month>03</month><day>10</day><volume>13</volume><issue>6</issue><fpage>603</fpage><pub-id pub-id-type="doi">10.3390/healthcare13060603</pub-id><pub-id pub-id-type="medline">40150453</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Waldock</surname><given-names>WJ</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Guni</surname><given-names>A</given-names> </name><name name-style="western"><surname>Nabeel</surname><given-names>A</given-names> </name><name name-style="western"><surname>Darzi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ashrafian</surname><given-names>H</given-names> </name></person-group><article-title>The accuracy and capability of artificial intelligence solutions in health care examinations and certificates: systematic review and meta-analysis</article-title><source>J Med Internet Res</source><year>2024</year><month>11</month><day>5</day><volume>26</volume><fpage>e56532</fpage><pub-id pub-id-type="doi">10.2196/56532</pub-id><pub-id pub-id-type="medline">39499913</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shan</surname><given-names>G</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>X</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>C</given-names> </name><etal/></person-group><article-title>Comparing diagnostic accuracy of clinical professionals and large language models: systematic review and meta-analysis</article-title><source>JMIR Med Inform</source><year>2025</year><month>04</month><day>25</day><volume>13</volume><fpage>e64963</fpage><pub-id pub-id-type="doi">10.2196/64963</pub-id><pub-id pub-id-type="medline">40279517</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lucas</surname><given-names>HC</given-names> </name><name name-style="western"><surname>Upperman</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Robinson</surname><given-names>JR</given-names> </name></person-group><article-title>A systematic review of large language models and their implications in medical education</article-title><source>Med Educ</source><year>2024</year><month>11</month><volume>58</volume><issue>11</issue><fpage>1276</fpage><lpage>1285</lpage><pub-id pub-id-type="doi">10.1111/medu.15402</pub-id><pub-id pub-id-type="medline">38639098</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yin</surname><given-names>K</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Jiang</surname><given-names>X</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>D</given-names> </name></person-group><article-title>Effectiveness of generative artificial intelligence-based teaching versus traditional teaching methods in medical education: a meta-analysis of randomized controlled trials</article-title><source>BMC Med Educ</source><year>2025</year><volume>25</volume><issue>1</issue><fpage>1175</fpage><pub-id pub-id-type="doi">10.1186/s12909-025-07750-2</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Samarakoon</surname><given-names>L</given-names> </name><name name-style="western"><surname>Fernando</surname><given-names>T</given-names> </name><name name-style="western"><surname>Rodrigo</surname><given-names>C</given-names> </name><name name-style="western"><surname>Rajapakse</surname><given-names>S</given-names> </name></person-group><article-title>Learning styles and approaches to learning among medical undergraduates and postgraduates</article-title><source>BMC Med Educ</source><year>2013</year><month>12</month><volume>13</volume><issue>1</issue><pub-id pub-id-type="doi">10.1186/1472-6920-13-42</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Sharma</surname><given-names>AD</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Goyal</surname><given-names>LS</given-names> </name><name name-style="western"><surname>Rehman</surname><given-names>A</given-names> </name></person-group><article-title>Virtual reality: robotic improved surgical precision using AI techniques</article-title><source>Artificial Intelligence and Internet of Things</source><year>2021</year><access-date>2026-04-30</access-date><publisher-name>CRC Press</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://www.taylorfrancis.com/chapters/edit/10.1201/9781003097204-16/virtual-reality-aditi-sharma-aman-dureja">https://www.taylorfrancis.com/chapters/edit/10.1201/9781003097204-16/virtual-reality-aditi-sharma-aman-dureja</ext-link></comment></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Alliger</surname><given-names>GM</given-names> </name><name name-style="western"><surname>Janak</surname><given-names>EA</given-names> </name></person-group><article-title>Kirkpatrick&#x2019;s levels of training criteria: thirty years later</article-title><source>Pers Psychol</source><year>1989</year><month>06</month><volume>42</volume><issue>2</issue><fpage>331</fpage><lpage>342</lpage><pub-id pub-id-type="doi">10.1111/j.1744-6570.1989.tb00661.x</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rethlefsen</surname><given-names>ML</given-names> </name><name name-style="western"><surname>Kirtley</surname><given-names>S</given-names> </name><name name-style="western"><surname>Waffenschmidt</surname><given-names>S</given-names> </name><etal/></person-group><article-title>PRISMA-S: an extension to the PRISMA Statement for Reporting Literature Searches in Systematic Reviews</article-title><source>Syst Rev</source><year>2021</year><month>01</month><day>26</day><volume>10</volume><issue>1</issue><fpage>39</fpage><pub-id pub-id-type="doi">10.1186/s13643-020-01542-z</pub-id><pub-id pub-id-type="medline">33499930</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="web"><article-title>Homepage</article-title><source>Claude</source><year>2025</year><access-date>2025-08-21</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://claude.ai">https://claude.ai</ext-link></comment></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Johnston</surname><given-names>S</given-names> </name><name name-style="western"><surname>Coyer</surname><given-names>FM</given-names> </name><name name-style="western"><surname>Nash</surname><given-names>R</given-names> </name></person-group><article-title>Kirkpatrick&#x2019;s evaluation of simulation and debriefing in health care education: a systematic review</article-title><source>J Nurs Educ</source><year>2018</year><month>07</month><day>1</day><volume>57</volume><issue>7</issue><fpage>393</fpage><lpage>398</lpage><pub-id pub-id-type="doi">10.3928/01484834-20180618-03</pub-id><pub-id pub-id-type="medline">29958308</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Keitz</surname><given-names>S</given-names> </name><name name-style="western"><surname>Briel</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Development of ROBUST-RCT: risk of bias instrument for use in systematic reviews-for randomised controlled trials</article-title><source>BMJ</source><year>2025</year><month>03</month><day>25</day><volume>388</volume><fpage>e081199</fpage><pub-id pub-id-type="doi">10.1136/bmj-2024-081199</pub-id><pub-id pub-id-type="medline">40132800</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Cohen</surname><given-names>J</given-names> </name></person-group><source>Statistical Power Analysis for the Behavioral Sciences</source><year>1988</year><edition>2</edition><publisher-name>Routledge</publisher-name><pub-id pub-id-type="doi">10.4324/9780203771587</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kraft</surname><given-names>MA</given-names> </name></person-group><article-title>Interpreting effect sizes of education interventions</article-title><source>Educ Res</source><year>2020</year><month>05</month><volume>49</volume><issue>4</issue><fpage>241</fpage><lpage>253</lpage><pub-id pub-id-type="doi">10.3102/0013189X20912798</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Norman</surname><given-names>GR</given-names> </name><name name-style="western"><surname>Sloan</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Wyrwich</surname><given-names>KW</given-names> </name></person-group><article-title>Interpretation of changes in health-related quality of life: the remarkable universality of half a standard deviation</article-title><source>Med Care</source><year>2003</year><month>05</month><volume>41</volume><issue>5</issue><fpage>582</fpage><lpage>592</lpage><pub-id pub-id-type="doi">10.1097/01.MLR.0000062554.74615.4C</pub-id><pub-id pub-id-type="medline">12719681</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hozo</surname><given-names>SP</given-names> </name><name name-style="western"><surname>Djulbegovic</surname><given-names>B</given-names> </name><name name-style="western"><surname>Hozo</surname><given-names>I</given-names> </name></person-group><article-title>Estimating the mean and variance from the median, range, and the size of a sample</article-title><source>BMC Med Res Methodol</source><year>2005</year><month>04</month><day>20</day><volume>5</volume><issue>1</issue><fpage>13</fpage><pub-id pub-id-type="doi">10.1186/1471-2288-5-13</pub-id><pub-id pub-id-type="medline">15840177</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Higgins</surname><given-names>JP</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Higgins</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Chandler</surname><given-names>J</given-names> </name><name name-style="western"><surname>Cumpston</surname><given-names>M</given-names> </name></person-group><article-title>Including variants on randomized trials</article-title><source>Cochrane Handbook for Systematic Reviews of Interventions Version 65</source><year>2024</year><access-date>2026-04-30</access-date><publisher-name>Cochrane</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://www.cochrane.org/authors/handbooks-and-manuals/handbook/current/chapter-23">https://www.cochrane.org/authors/handbooks-and-manuals/handbook/current/chapter-23</ext-link></comment></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Altman</surname><given-names>DG</given-names> </name><name name-style="western"><surname>McKenzie</surname><given-names>JE</given-names> </name><name name-style="western"><surname>Veroniki</surname><given-names>AA</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Higgins</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Chandler</surname><given-names>J</given-names> </name><name name-style="western"><surname>Cumpston</surname><given-names>M</given-names> </name></person-group><article-title>Analysing data and undertaking meta-analyses</article-title><source>Cochrane Handbook for Systematic Reviews of Interventions</source><access-date>2026-04-30</access-date><publisher-name>Cochrane</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://www.cochrane.org/authors/handbooks-and-manuals/handbook/current/chapter-10">https://www.cochrane.org/authors/handbooks-and-manuals/handbook/current/chapter-10</ext-link></comment></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Langan</surname><given-names>D</given-names> </name><name name-style="western"><surname>Higgins</surname><given-names>JPT</given-names> </name><name name-style="western"><surname>Jackson</surname><given-names>D</given-names> </name><etal/></person-group><article-title>A comparison of heterogeneity variance estimators in simulated random-effects meta-analyses</article-title><source>Res Synth Methods</source><year>2019</year><month>03</month><volume>10</volume><issue>1</issue><fpage>83</fpage><lpage>98</lpage><pub-id pub-id-type="doi">10.1002/jrsm.1316</pub-id><pub-id pub-id-type="medline">30067315</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>IntHout</surname><given-names>J</given-names> </name><name name-style="western"><surname>Ioannidis</surname><given-names>JPA</given-names> </name><name name-style="western"><surname>Borm</surname><given-names>GF</given-names> </name></person-group><article-title>The Hartung-Knapp-Sidik-Jonkman method for random effects meta-analysis is straightforward and considerably outperforms the standard DerSimonian-Laird method</article-title><source>BMC Med Res Methodol</source><year>2014</year><month>02</month><day>18</day><volume>14</volume><issue>1</issue><fpage>25</fpage><pub-id pub-id-type="doi">10.1186/1471-2288-14-25</pub-id><pub-id pub-id-type="medline">24548571</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Higgins</surname><given-names>JPT</given-names> </name><name name-style="western"><surname>Thomas</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chandler</surname><given-names>J</given-names> </name><name name-style="western"><surname>Cumpston</surname><given-names>M</given-names> </name><name name-style="western"><surname>Li</surname><given-names>T</given-names> </name><name name-style="western"><surname>Page</surname><given-names>MJ</given-names> </name><etal/></person-group><source>Cochrane Handbook for Systematic Reviews of Interventions</source><year>2022</year><access-date>2026-04-30</access-date><publisher-name>The Cochrane Collaboration</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://training.cochrane.org/handbook">https://training.cochrane.org/handbook</ext-link></comment></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>IntHout</surname><given-names>J</given-names> </name><name name-style="western"><surname>Ioannidis</surname><given-names>JPA</given-names> </name><name name-style="western"><surname>Rovers</surname><given-names>MM</given-names> </name><name name-style="western"><surname>Goeman</surname><given-names>JJ</given-names> </name></person-group><article-title>Plea for routinely presenting prediction intervals in meta-analysis</article-title><source>BMJ Open</source><year>2016</year><month>07</month><day>12</day><volume>6</volume><issue>7</issue><fpage>e010247</fpage><pub-id pub-id-type="doi">10.1136/bmjopen-2015-010247</pub-id><pub-id pub-id-type="medline">27406637</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Sterne</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Egger</surname><given-names>M</given-names> </name><name name-style="western"><surname>Moher</surname><given-names>D</given-names> </name><name name-style="western"><surname>Boutron</surname><given-names>I</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Higgins</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Chandler</surname><given-names>J</given-names> </name><name name-style="western"><surname>Cumpston</surname><given-names>MS</given-names> </name></person-group><article-title>Addressing reporting biases</article-title><source>Cochrane Handbook for Systematic Reviews of Interventions</source><year>2017</year><access-date>2026-04-30</access-date><publisher-name>The Cochrane Collaboration</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://onlinelibrary.wiley.com/doi/10.1002/9780470712184.ch10">https://onlinelibrary.wiley.com/doi/10.1002/9780470712184.ch10</ext-link></comment></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sterne</surname><given-names>JAC</given-names> </name><name name-style="western"><surname>Sutton</surname><given-names>AJ</given-names> </name><name name-style="western"><surname>Ioannidis</surname><given-names>JPA</given-names> </name><etal/></person-group><article-title>Recommendations for examining and interpreting funnel plot asymmetry in meta-analyses of randomised controlled trials</article-title><source>BMJ</source><year>2011</year><month>07</month><day>22</day><volume>343</volume><issue>jul22 1</issue><fpage>d4002</fpage><pub-id pub-id-type="doi">10.1136/bmj.d4002</pub-id><pub-id pub-id-type="medline">21784880</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Page</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Sterne</surname><given-names>JAC</given-names> </name><name name-style="western"><surname>Boutron</surname><given-names>I</given-names> </name><etal/></person-group><article-title>ROB-ME: a tool for assessing risk of bias due to missing evidence in systematic reviews with meta-analysis</article-title><source>BMJ</source><year>2023</year><month>11</month><day>20</day><volume>383</volume><fpage>e076754</fpage><pub-id pub-id-type="doi">10.1136/bmj-2023-076754</pub-id><pub-id pub-id-type="medline">37984978</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="book"><person-group person-group-type="editor"><name name-style="western"><surname>Sch&#x00FC;nemann</surname><given-names>H</given-names> </name><name name-style="western"><surname>Bro&#x017C;ek</surname><given-names>J</given-names> </name><name name-style="western"><surname>Guyatt</surname><given-names>G</given-names> </name><name name-style="western"><surname>Oxman</surname><given-names>A</given-names> </name></person-group><source>Handbook for Grading the Quality of Evidence and the Strength of Recommendations Using the GRADE Approach</source><year>2013</year><access-date>2026-04-30</access-date><publisher-name>Group GW</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://gdt.gradepro.org/app/handbook/handbook.html">https://gdt.gradepro.org/app/handbook/handbook.html</ext-link></comment></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Page</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>McKenzie</surname><given-names>JE</given-names> </name><name name-style="western"><surname>Bossuyt</surname><given-names>PM</given-names> </name><etal/></person-group><article-title>The PRISMA 2020 statement: an updated guideline for reporting systematic reviews</article-title><source>Syst Rev</source><year>2021</year><month>03</month><day>29</day><volume>10</volume><issue>1</issue><fpage>89</fpage><pub-id pub-id-type="doi">10.1186/s13643-021-01626-4</pub-id><pub-id pub-id-type="medline">33781348</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Page</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>McKenzie</surname><given-names>JE</given-names> </name><name name-style="western"><surname>Bossuyt</surname><given-names>PM</given-names> </name><etal/></person-group><article-title>The PRISMA 2020 statement: an updated guideline for reporting systematic reviews</article-title><source>BMJ</source><year>2021</year><month>03</month><day>29</day><volume>372</volume><fpage>n71</fpage><pub-id pub-id-type="doi">10.1136/bmj.n71</pub-id><pub-id pub-id-type="medline">33782057</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liaw</surname><given-names>SY</given-names> </name><name name-style="western"><surname>Tan</surname><given-names>JZ</given-names> </name><name name-style="western"><surname>Bin Rusli</surname><given-names>KD</given-names> </name><etal/></person-group><article-title>Artificial intelligence versus human-controlled doctor in virtual reality simulation for sepsis team training: randomized controlled study</article-title><source>J Med Internet Res</source><year>2023</year><month>07</month><day>26</day><volume>25</volume><fpage>e47748</fpage><pub-id pub-id-type="doi">10.2196/47748</pub-id><pub-id pub-id-type="medline">37494112</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ting</surname><given-names>PW</given-names> </name><name name-style="western"><surname>Wolffsohn</surname><given-names>JS</given-names> </name></person-group><article-title>Artificial intelligence-driven patient history and symptoms combined with slit-lamp eye simulation for enhancing the clinical training of students</article-title><source>Clin Exp Optom</source><year>2026</year><month>04</month><volume>109</volume><issue>3</issue><fpage>434</fpage><lpage>443</lpage><pub-id pub-id-type="doi">10.1080/08164622.2025.2544809</pub-id><pub-id pub-id-type="medline">40975895</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Veras</surname><given-names>M</given-names> </name><name name-style="western"><surname>Dyer</surname><given-names>JO</given-names> </name><name name-style="western"><surname>Shannon</surname><given-names>H</given-names> </name><etal/></person-group><article-title>A mixed methods crossover randomized controlled trial exploring the experiences, perceptions, and usability of artificial intelligence (ChatGPT) in health sciences education</article-title><source>Digit Health</source><year>2024</year><volume>10</volume><fpage>20552076241298485</fpage><pub-id pub-id-type="doi">10.1177/20552076241298485</pub-id><pub-id pub-id-type="medline">39917060</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yilmaz</surname><given-names>R</given-names> </name><name name-style="western"><surname>Bakhaidar</surname><given-names>M</given-names> </name><name name-style="western"><surname>Alsayegh</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Real-time multifaceted artificial intelligence vs in-person instruction in teaching surgical technical skills: a randomized controlled trial</article-title><source>Sci Rep</source><year>2024</year><month>07</month><day>2</day><volume>14</volume><issue>1</issue><fpage>15130</fpage><pub-id pub-id-type="doi">10.1038/s41598-024-65716-8</pub-id><pub-id pub-id-type="medline">38956112</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fazlollahi</surname><given-names>AM</given-names> </name><name name-style="western"><surname>Bakhaidar</surname><given-names>M</given-names> </name><name name-style="western"><surname>Alsayegh</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Effect of artificial intelligence tutoring vs expert instruction on learning simulated surgical skills among medical students</article-title><source>JAMA Netw Open</source><year>2022</year><month>02</month><day>1</day><volume>5</volume><issue>2</issue><fpage>e2149008</fpage><pub-id pub-id-type="doi">10.1001/jamanetworkopen.2021.49008</pub-id><pub-id pub-id-type="medline">35191972</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Akutay</surname><given-names>S</given-names> </name><name name-style="western"><surname>Y&#x00FC;celer Ka&#x00E7;maz</surname><given-names>H</given-names> </name><name name-style="western"><surname>Kahraman</surname><given-names>H</given-names> </name></person-group><article-title>The effect of artificial intelligence supported case analysis on nursing students&#x2019; case management performance and satisfaction: a randomized controlled trial</article-title><source>Nurse Educ Pract</source><year>2024</year><month>10</month><volume>80</volume><fpage>104142</fpage><pub-id pub-id-type="doi">10.1016/j.nepr.2024.104142</pub-id><pub-id pub-id-type="medline">39299058</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ali</surname><given-names>M</given-names> </name><name name-style="western"><surname>Rehman</surname><given-names>S</given-names> </name><name name-style="western"><surname>Cheema</surname><given-names>E</given-names> </name></person-group><article-title>Impact of artificial intelligence on the academic performance and test anxiety of pharmacy students in objective structured clinical examination: a randomized controlled trial</article-title><source>Int J Clin Pharm</source><year>2025</year><month>08</month><volume>47</volume><issue>4</issue><fpage>1034</fpage><lpage>1041</lpage><pub-id pub-id-type="doi">10.1007/s11096-025-01876-5</pub-id><pub-id pub-id-type="medline">39903358</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Aneesh</surname><given-names>KV</given-names> </name><name name-style="western"><surname>Mohanan</surname><given-names>S</given-names> </name><name name-style="western"><surname>Jose</surname><given-names>S</given-names> </name><name name-style="western"><surname>Sajla</surname><given-names>K</given-names> </name><name name-style="western"><surname>Indulekha</surname><given-names>C</given-names> </name><name name-style="western"><surname>Sukumaran</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Effectiveness of generative AI versus traditional resources for self-directed learning in physiology among MBBS students: a comparative interventional study</article-title><source>Int J Med Public Health</source><year>2025</year><volume>15</volume><issue>3</issue><pub-id pub-id-type="doi">10.70034/ijmedph.2025.3.385</pub-id><pub-id pub-id-type="medline">40830476</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Arkan</surname><given-names>B</given-names> </name><name name-style="western"><surname>Dall&#x0131;</surname><given-names>&#x00D6;E</given-names> </name><name name-style="western"><surname>Varol</surname><given-names>B</given-names> </name></person-group><article-title>The impact of ChatGPT training in the nursing process on nursing students&#x2019; problem-solving skills, attitudes towards artificial intelligence, competency, and satisfaction levels: single-blind randomized controlled study</article-title><source>Nurse Educ Today</source><year>2025</year><month>09</month><volume>152</volume><fpage>106765</fpage><pub-id pub-id-type="doi">10.1016/j.nedt.2025.106765</pub-id><pub-id pub-id-type="medline">40334550</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Aronovitz</surname><given-names>N</given-names> </name><name name-style="western"><surname>Hazan</surname><given-names>I</given-names> </name><name name-style="western"><surname>Jedwab</surname><given-names>R</given-names> </name><etal/></person-group><article-title>The effect of real-time EF automatic tool on cardiac ultrasound performance among medical students</article-title><source>PLoS ONE</source><year>2024</year><volume>19</volume><issue>3</issue><fpage>e0299461</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0299461</pub-id><pub-id pub-id-type="medline">38547257</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ayan</surname><given-names>E</given-names> </name><name name-style="western"><surname>Bayraktar</surname><given-names>Y</given-names> </name><name name-style="western"><surname>&#x00C7;elik</surname><given-names>&#x00C7;</given-names> </name><name name-style="western"><surname>Ayhan</surname><given-names>B</given-names> </name></person-group><article-title>Dental student application of artificial intelligence technology in detecting proximal caries lesions</article-title><source>J Dent Educ</source><year>2024</year><month>04</month><volume>88</volume><issue>4</issue><fpage>490</fpage><lpage>500</lpage><pub-id pub-id-type="doi">10.1002/jdd.13437</pub-id><pub-id pub-id-type="medline">38200405</pub-id></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Br&#x00FC;gge</surname><given-names>E</given-names> </name><name name-style="western"><surname>Ricchizzi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Arenbeck</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Large language models improve clinical decision making of medical students through patient simulation and structured feedback: a randomized controlled trial</article-title><source>BMC Med Educ</source><year>2024</year><month>11</month><day>28</day><volume>24</volume><issue>1</issue><fpage>1391</fpage><pub-id pub-id-type="doi">10.1186/s12909-024-06399-7</pub-id><pub-id pub-id-type="medline">39609823</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Bliss</surname><given-names>L</given-names> </name><name name-style="western"><surname>Angelov</surname><given-names>N</given-names> </name><name name-style="western"><surname>Glick</surname><given-names>A</given-names> </name></person-group><article-title>Artificial intelligence-assisted full-mouth radiograph mounting in dental education</article-title><source>J Dent Educ</source><year>2024</year><month>07</month><volume>88</volume><issue>7</issue><fpage>933</fpage><lpage>939</lpage><pub-id pub-id-type="doi">10.1002/jdd.13524</pub-id><pub-id pub-id-type="medline">38545660</pub-id></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>Y</given-names> </name></person-group><article-title>Evaluation of the impact of AI-driven personalized learning platform on medical students&#x2019; learning performance</article-title><source>Front Med</source><year>2025</year><volume>12</volume><issue>1610012</issue><pub-id pub-id-type="doi">10.3389/fmed.2025.1610012</pub-id><pub-id pub-id-type="medline">02913790</pub-id></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Cheng</surname><given-names>CT</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>CC</given-names> </name><name name-style="western"><surname>Fu</surname><given-names>CY</given-names> </name><etal/></person-group><article-title>Artificial intelligence-based education assists medical students&#x2019; interpretation of hip fracture</article-title><source>Insights Imaging</source><year>2020</year><month>11</month><day>23</day><volume>11</volume><issue>1</issue><fpage>119</fpage><pub-id pub-id-type="doi">10.1186/s13244-020-00932-0</pub-id><pub-id pub-id-type="medline">33226480</pub-id></nlm-citation></ref><ref id="ref61"><label>61</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>&#x00C7;i&#x00E7;ek</surname><given-names>FE</given-names> </name><name name-style="western"><surname>&#x00DC;lker</surname><given-names>M</given-names> </name><name name-style="western"><surname>&#x00D6;zer</surname><given-names>M</given-names> </name><name name-style="western"><surname>K&#x0131;yak</surname><given-names>YS</given-names> </name></person-group><article-title>ChatGPT versus expert feedback on clinical reasoning questions and their effect on learning: a randomized controlled trial</article-title><source>Postgrad Med J</source><year>2025</year><month>04</month><day>22</day><volume>101</volume><issue>1195</issue><fpage>458</fpage><lpage>463</lpage><pub-id pub-id-type="doi">10.1093/postmj/qgae170</pub-id></nlm-citation></ref><ref id="ref62"><label>62</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Co&#x015F;kun</surname><given-names>&#x00D6;</given-names> </name><name name-style="western"><surname>K&#x0131;yak</surname><given-names>YS</given-names> </name><name name-style="western"><surname>Budako&#x011F;lu</surname><given-names>I&#x0130;</given-names> </name></person-group><article-title>ChatGPT to generate clinical vignettes for teaching and multiple-choice questions for assessment: a randomized controlled experiment</article-title><source>Med Teach</source><year>2025</year><month>02</month><volume>47</volume><issue>2</issue><fpage>268</fpage><lpage>274</lpage><pub-id pub-id-type="doi">10.1080/0142159X.2024.2327477</pub-id><pub-id pub-id-type="medline">38478902</pub-id></nlm-citation></ref><ref id="ref63"><label>63</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Digiacomo</surname><given-names>A</given-names> </name><name name-style="western"><surname>Orsini</surname><given-names>A</given-names> </name><name name-style="western"><surname>Cicchetti</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Chatgpt vs traditional pedagogy: a comparative study in urological learning</article-title><source>World J Urol</source><year>2025</year><month>05</month><day>8</day><volume>43</volume><issue>1</issue><fpage>286</fpage><pub-id pub-id-type="doi">10.1007/s00345-025-05654-w</pub-id><pub-id pub-id-type="medline">40338279</pub-id></nlm-citation></ref><ref id="ref64"><label>64</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fazlollahi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Bakhaidar</surname><given-names>M</given-names> </name><name name-style="western"><surname>Alsayegh</surname><given-names>A</given-names> </name><etal/></person-group><article-title>510 artificial intelligence tutoring compared with expert instruction in neurosurgical simulation training: a randomized controlled trial</article-title><source>Clin Neurosurg</source><year>2022</year><volume>68</volume><issue>Supplement_1</issue><fpage>128</fpage><lpage>129</lpage><pub-id pub-id-type="doi">10.1227/NEU.0000000000001880_510</pub-id></nlm-citation></ref><ref id="ref65"><label>65</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ferrer-Pe&#x00F1;a</surname><given-names>R</given-names> </name><name name-style="western"><surname>Di-Bonaventura</surname><given-names>S</given-names> </name><name name-style="western"><surname>P&#x00E9;rez-Gonz&#x00E1;lez</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ler&#x00ED;n-Calvo</surname><given-names>A</given-names> </name></person-group><article-title>Feasibility of a randomized controlled trial of large AI-based linguistic models for clinical reasoning training of physical therapy students: pilot randomized parallel-group study</article-title><source>JMIR Form Res</source><year>2025</year><month>07</month><day>23</day><volume>9</volume><fpage>e66126</fpage><pub-id pub-id-type="doi">10.2196/66126</pub-id><pub-id pub-id-type="medline">40702721</pub-id></nlm-citation></ref><ref id="ref66"><label>66</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fung</surname><given-names>TCJ</given-names> </name><name name-style="western"><surname>Chan</surname><given-names>SL</given-names> </name><name name-style="western"><surname>Lam</surname><given-names>CFM</given-names> </name><etal/></person-group><article-title>Effects of generative artificial intelligence (GenAI) patient simulation on perceived clinical competency among global nursing undergraduates: a cross-over randomised controlled trial</article-title><source>BMC Nurs</source><year>2025</year><month>07</month><day>17</day><volume>24</volume><issue>1</issue><fpage>934</fpage><pub-id pub-id-type="doi">10.1186/s12912-025-03492-0</pub-id><pub-id pub-id-type="medline">40676632</pub-id></nlm-citation></ref><ref id="ref67"><label>67</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gan</surname><given-names>W</given-names> </name><name name-style="western"><surname>Ouyang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Li</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Integrating ChatGPT in orthopedic education for medical undergraduates: randomized controlled trial</article-title><source>J Med Internet Res</source><year>2024</year><month>08</month><day>20</day><volume>26</volume><fpage>e57037</fpage><pub-id pub-id-type="doi">10.2196/57037</pub-id><pub-id pub-id-type="medline">39163598</pub-id></nlm-citation></ref><ref id="ref68"><label>68</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Han</surname><given-names>JW</given-names> </name><name name-style="western"><surname>Park</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>H</given-names> </name></person-group><article-title>Development and effects of a chatbot education program for self-directed learning in nursing students</article-title><source>BMC Med Educ</source><year>2025</year><month>06</month><day>2</day><volume>25</volume><issue>1</issue><fpage>825</fpage><pub-id pub-id-type="doi">10.1186/s12909-025-07316-2</pub-id><pub-id pub-id-type="medline">40457281</pub-id></nlm-citation></ref><ref id="ref69"><label>69</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>H&#x00F6;hne</surname><given-names>E</given-names> </name><name name-style="western"><surname>Bauer</surname><given-names>E</given-names> </name><name name-style="western"><surname>Bauer</surname><given-names>C</given-names> </name><etal/></person-group><article-title>A comparative bicentric study on ultrasound education for students: app- and AI-supported learning versus traditional hands-on instruction (AI-teach study)</article-title><source>Acad Radiol</source><year>2025</year><month>08</month><volume>32</volume><issue>8</issue><fpage>4930</fpage><lpage>4938</lpage><pub-id pub-id-type="doi">10.1016/j.acra.2025.04.024</pub-id><pub-id pub-id-type="medline">40300993</pub-id></nlm-citation></ref><ref id="ref70"><label>70</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wen</surname><given-names>C</given-names> </name><name name-style="western"><surname>Bai</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Exploring the application capability of ChatGPT as an instructor in skills education for dental medical students: randomized controlled trial</article-title><source>J Med Internet Res</source><year>2025</year><volume>27</volume><fpage>e68538</fpage><pub-id pub-id-type="doi">10.2196/68538</pub-id><pub-id pub-id-type="medline">02854722</pub-id></nlm-citation></ref><ref id="ref71"><label>71</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hui</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Zewu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Jiao</surname><given-names>H</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>C</given-names> </name></person-group><article-title>Application of ChatGPT-assisted problem-based learning teaching method in clinical medical education</article-title><source>BMC Med Educ</source><year>2025</year><volume>25</volume><issue>1</issue><fpage>CN</fpage><pub-id pub-id-type="doi">10.1186/s12909-024-06321-1</pub-id><pub-id pub-id-type="medline">02798528</pub-id></nlm-citation></ref><ref id="ref72"><label>72</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kalam</surname><given-names>KA</given-names> </name><name name-style="western"><surname>Masoud</surname><given-names>FD</given-names> </name><name name-style="western"><surname>Muntaser</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ranga</surname><given-names>R</given-names> </name><name name-style="western"><surname>Geng</surname><given-names>X</given-names> </name><name name-style="western"><surname>Goyal</surname><given-names>M</given-names> </name></person-group><article-title>ChatGPT as a learning tool for medical students: results from a randomized controlled trial</article-title><source>Cureus</source><year>2025</year><month>06</month><volume>17</volume><issue>6</issue><fpage>e85767</fpage><pub-id pub-id-type="doi">10.7759/cureus.85767</pub-id><pub-id pub-id-type="medline">40656250</pub-id></nlm-citation></ref><ref id="ref73"><label>73</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kestel</surname><given-names>S</given-names> </name><name name-style="western"><surname>&#x00C7;al&#x0131;k</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ku&#x015F;</surname><given-names>M</given-names> </name></person-group><article-title>The effect of chatbot-supported instruction on nursing students&#x2019; history-taking questioning skills and stress level: a randomized controlled study</article-title><source>J Prof Nurs</source><year>2025</year><volume>60</volume><fpage>93</fpage><lpage>100</lpage><pub-id pub-id-type="doi">10.1016/j.profnurs.2025.07.004</pub-id><pub-id pub-id-type="medline">40915772</pub-id></nlm-citation></ref><ref id="ref74"><label>74</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>J</given-names> </name><name name-style="western"><surname>Ouyang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Artificial intelligence-based online platform assists blood cell morphology learning: a mixed-methods sequential explanatory designed research</article-title><source>Med Teach</source><year>2023</year><month>06</month><volume>45</volume><issue>6</issue><fpage>596</fpage><lpage>603</lpage><pub-id pub-id-type="doi">10.1080/0142159X.2023.2190483</pub-id><pub-id pub-id-type="medline">36971649</pub-id></nlm-citation></ref><ref id="ref75"><label>75</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lyu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Dong</surname><given-names>L</given-names> </name><name name-style="western"><surname>Fan</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>Artificial intelligence-based graded training of pulmonary nodules for junior radiology residents and medical imaging students</article-title><source>BMC Med Educ</source><year>2024</year><volume>24</volume><issue>1</issue><fpage>CN</fpage><pub-id pub-id-type="doi">10.1186/s12909-024-05723-5</pub-id><pub-id pub-id-type="medline">02725020</pub-id></nlm-citation></ref><ref id="ref76"><label>76</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mahrous</surname><given-names>A</given-names> </name><name name-style="western"><surname>Botsko</surname><given-names>DL</given-names> </name><name name-style="western"><surname>Elgreatly</surname><given-names>A</given-names> </name><name name-style="western"><surname>Tsujimoto</surname><given-names>A</given-names> </name><name name-style="western"><surname>Qian</surname><given-names>F</given-names> </name><name name-style="western"><surname>Schneider</surname><given-names>GB</given-names> </name></person-group><article-title>The use of artificial intelligence and game-based learning in removable partial denture design: a comparative study</article-title><source>J Dent Educ</source><year>2023</year><month>08</month><volume>87</volume><issue>8</issue><fpage>1188</fpage><lpage>1199</lpage><pub-id pub-id-type="doi">10.1002/jdd.13225</pub-id><pub-id pub-id-type="medline">37186466</pub-id></nlm-citation></ref><ref id="ref77"><label>77</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McCarrick</surname><given-names>CA</given-names> </name><name name-style="western"><surname>McEntee</surname><given-names>PD</given-names> </name><name name-style="western"><surname>Boland</surname><given-names>PA</given-names> </name><etal/></person-group><article-title>A randomized controlled trial of a deep language learning model-based simulation tool for undergraduate medical students in surgery</article-title><source>J Surg Educ</source><year>2025</year><month>09</month><volume>82</volume><issue>9</issue><fpage>103629</fpage><pub-id pub-id-type="doi">10.1016/j.jsurg.2025.103629</pub-id><pub-id pub-id-type="medline">40729832</pub-id></nlm-citation></ref><ref id="ref78"><label>78</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Molu</surname><given-names>B</given-names> </name></person-group><article-title>Improving nursing students&#x2019; learning outcomes in neonatal resuscitation: a quasi-experimental study comparing ai-assisted care plan learning with traditional instruction</article-title><source>J Eval Clin Pract</source><year>2025</year><month>02</month><volume>31</volume><issue>1</issue><fpage>e14286</fpage><pub-id pub-id-type="doi">10.1111/jep.14286</pub-id><pub-id pub-id-type="medline">39733257</pub-id></nlm-citation></ref><ref id="ref79"><label>79</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Montagna</surname><given-names>M</given-names> </name><name name-style="western"><surname>Chiabrando</surname><given-names>F</given-names> </name><name name-style="western"><surname>De Lorenzo</surname><given-names>R</given-names> </name><name name-style="western"><surname>Rovere Querini</surname><given-names>P</given-names> </name><collab>Medical Students</collab></person-group><article-title>Impact of clinical decision support systems on medical students&#x2019; case-solving performance: comparison study with a focus group</article-title><source>JMIR Med Educ</source><year>2025</year><month>03</month><day>18</day><volume>11</volume><fpage>e55709</fpage><pub-id pub-id-type="doi">10.2196/55709</pub-id><pub-id pub-id-type="medline">40101183</pub-id></nlm-citation></ref><ref id="ref80"><label>80</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ng</surname><given-names>ISH</given-names> </name><name name-style="western"><surname>Siu</surname><given-names>A</given-names> </name><name name-style="western"><surname>Han</surname><given-names>CSJ</given-names> </name><etal/></person-group><article-title>Evaluating a custom chatbot in undergraduate medical education: randomised crossover mixed-methods evaluation of performance, utility, and perceptions</article-title><source>Behav Sci (Basel)</source><year>2025</year><month>09</month><day>19</day><volume>15</volume><issue>9</issue><fpage>1284</fpage><pub-id pub-id-type="doi">10.3390/bs15091284</pub-id><pub-id pub-id-type="medline">41009314</pub-id></nlm-citation></ref><ref id="ref81"><label>81</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nissen</surname><given-names>L</given-names> </name><name name-style="western"><surname>Rother</surname><given-names>JF</given-names> </name><name name-style="western"><surname>Heinemann</surname><given-names>M</given-names> </name><name name-style="western"><surname>Reimer</surname><given-names>LM</given-names> </name><name name-style="western"><surname>Jonas</surname><given-names>S</given-names> </name><name name-style="western"><surname>Raupach</surname><given-names>T</given-names> </name></person-group><article-title>A randomised cross-over trial assessing the impact of AI-generated individual feedback on written online assignments for medical students</article-title><source>Med Teach</source><year>2025</year><month>09</month><volume>47</volume><issue>9</issue><fpage>1544</fpage><lpage>1550</lpage><pub-id pub-id-type="doi">10.1080/0142159X.2025.2451870</pub-id><pub-id pub-id-type="medline">39831699</pub-id></nlm-citation></ref><ref id="ref82"><label>82</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ramsamooj</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ibrahim</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Gerriets</surname><given-names>VA</given-names> </name><name name-style="western"><surname>Cusick</surname><given-names>JK</given-names> </name><name name-style="western"><surname>Ramsamooj</surname><given-names>R</given-names> </name></person-group><article-title>Artificial intelligence versus traditional learning in a medical school setting</article-title><source>Cureus</source><year>2025</year><volume>17</volume><issue>6</issue><fpage>CN</fpage><pub-id pub-id-type="doi">10.7759/cureus.85262</pub-id><pub-id pub-id-type="medline">02881262</pub-id></nlm-citation></ref><ref id="ref83"><label>83</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Saat&#x00E7;i</surname><given-names>G</given-names> </name><name name-style="western"><surname>Korkut</surname><given-names>S</given-names> </name><name name-style="western"><surname>&#x00DC;nsal</surname><given-names>A</given-names> </name></person-group><article-title>The effect of the use of artificial intelligence in the preparation of patient education materials by nursing students on the understandability, actionability and quality of the material: a randomized controlled trial</article-title><source>Nurse Educ Pract</source><year>2024</year><month>11</month><volume>81</volume><fpage>104186</fpage><pub-id pub-id-type="doi">10.1016/j.nepr.2024.104186</pub-id><pub-id pub-id-type="medline">39520840</pub-id></nlm-citation></ref><ref id="ref84"><label>84</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sahin</surname><given-names>GE</given-names> </name><name name-style="western"><surname>Bayram</surname><given-names>GA</given-names> </name><name name-style="western"><surname>Sierra</surname><given-names>AS</given-names> </name><etal/></person-group><article-title>Effects of artificial intelligence based physiotherapy educational approach in developing clinical reasoning skills: a randomized controlled trial</article-title><source>BMC Med Educ</source><year>2025</year><month>10</month><day>9</day><volume>25</volume><issue>1</issue><fpage>1378</fpage><pub-id pub-id-type="doi">10.1186/s12909-025-07926-w</pub-id><pub-id pub-id-type="medline">41068907</pub-id></nlm-citation></ref><ref id="ref85"><label>85</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schropp</surname><given-names>L</given-names> </name><name name-style="western"><surname>S&#x00F8;rensen</surname><given-names>APS</given-names> </name><name name-style="western"><surname>Devlin</surname><given-names>H</given-names> </name><name name-style="western"><surname>Matzen</surname><given-names>LH</given-names> </name></person-group><article-title>Use of artificial intelligence software in dental education: a study on assisted proximal caries assessment in bitewing radiographs</article-title><source>Eur J Dent Educ</source><year>2024</year><month>05</month><volume>28</volume><issue>2</issue><fpage>490</fpage><lpage>496</lpage><pub-id pub-id-type="doi">10.1111/eje.12973</pub-id><pub-id pub-id-type="medline">37961027</pub-id></nlm-citation></ref><ref id="ref86"><label>86</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kejingyun</surname><given-names>S</given-names> </name><name name-style="western"><surname>Mingjun</surname><given-names>R</given-names> </name></person-group><article-title>Randomized controlled study on the impact of problem-based learning combined with large language models on critical thinking skills in nursing students</article-title><source>Nurse Educ</source><year>2025</year><volume>50</volume><issue>4</issue><fpage>216</fpage><lpage>220</lpage><pub-id pub-id-type="doi">10.1097/NNE.0000000000001879</pub-id><pub-id pub-id-type="medline">40261697</pub-id></nlm-citation></ref><ref id="ref87"><label>87</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shin</surname><given-names>H</given-names> </name><name name-style="western"><surname>De Gagne</surname><given-names>JC</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>SS</given-names> </name><name name-style="western"><surname>Hong</surname><given-names>M</given-names> </name></person-group><article-title>The impact of artificial intelligence-assisted learning on nursing students&#x2019; ethical decision-making and clinical reasoning in pediatric care</article-title><source>Comput Inform Nurs</source><year>2024</year><month>10</month><day>1</day><volume>42</volume><issue>10</issue><fpage>704</fpage><lpage>711</lpage><pub-id pub-id-type="doi">10.1097/CIN.0000000000001177</pub-id><pub-id pub-id-type="medline">39152099</pub-id></nlm-citation></ref><ref id="ref88"><label>88</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Song</surname><given-names>D</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>P</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Effects of generative artificial intelligence on higher-order thinking skills and artificial intelligence literacy in nursing undergraduates: a quasi-experimental study</article-title><source>Nurse Educ Pract</source><year>2025</year><month>10</month><volume>88</volume><fpage>104549</fpage><pub-id pub-id-type="doi">10.1016/j.nepr.2025.104549</pub-id><pub-id pub-id-type="medline">40915080</pub-id></nlm-citation></ref><ref id="ref89"><label>89</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Staples</surname><given-names>G</given-names> </name><name name-style="western"><surname>Webster</surname><given-names>C</given-names> </name><name name-style="western"><surname>Baraza</surname><given-names>W</given-names> </name></person-group><article-title>SP3.03 artificially intelligent simulated patients in undergraduate surgical teaching: a single blinded randomized controlled study</article-title><source>Br J Surg</source><year>2025</year><month>08</month><day>28</day><volume>112</volume><issue>Supplement_13</issue><fpage>CN</fpage><pub-id pub-id-type="doi">10.1093/bjs/znaf166.037</pub-id><pub-id pub-id-type="medline">02918785</pub-id></nlm-citation></ref><ref id="ref90"><label>90</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Svendsen</surname><given-names>K</given-names> </name><name name-style="western"><surname>Askar</surname><given-names>M</given-names> </name><name name-style="western"><surname>Umer</surname><given-names>D</given-names> </name><name name-style="western"><surname>Halvorsen</surname><given-names>KH</given-names> </name></person-group><article-title>Short-term learning effect of ChatGPT on pharmacy students&#x2019; learning</article-title><source>Explor Res Clin Soc Pharm</source><year>2024</year><month>09</month><volume>15</volume><fpage>100478</fpage><pub-id pub-id-type="doi">10.1016/j.rcsop.2024.100478</pub-id><pub-id pub-id-type="medline">39139501</pub-id></nlm-citation></ref><ref id="ref91"><label>91</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shalong</surname><given-names>W</given-names> </name><name name-style="western"><surname>Yi</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Bin</surname><given-names>Z</given-names> </name><etal/></person-group><article-title>Enhancing self-directed learning with custom GPT AI facilitation among medical students: a randomized controlled trial</article-title><source>Med Teach</source><year>2025</year><month>07</month><volume>47</volume><issue>7</issue><fpage>1126</fpage><lpage>1133</lpage><pub-id pub-id-type="doi">10.1080/0142159X.2024.2413023</pub-id><pub-id pub-id-type="medline">39425996</pub-id></nlm-citation></ref><ref id="ref92"><label>92</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>YF</given-names> </name><name name-style="western"><surname>Hsu</surname><given-names>MH</given-names> </name><name name-style="western"><surname>Yue-Feng Wang</surname><given-names>M</given-names> </name></person-group><article-title>Gamified mobile learning: an escape room chatbot to enhance medical terminology acquisition</article-title><source>Health Educ J</source><year>2025</year><month>03</month><volume>84</volume><issue>2</issue><fpage>174</fpage><lpage>188</lpage><pub-id pub-id-type="doi">10.1177/00178969241313442</pub-id></nlm-citation></ref><ref id="ref93"><label>93</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Fan</surname><given-names>TT</given-names> </name><name name-style="western"><surname>Li</surname><given-names>ML</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>NJ</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>XC</given-names> </name></person-group><article-title>Feasibility study of using GPT for history-taking training in medical education: a randomized clinical trial</article-title><source>BMC Med Educ</source><year>2025</year><volume>25</volume><issue>1</issue><fpage>CN</fpage><pub-id pub-id-type="doi">10.1186/s12909-025-07614-9</pub-id><pub-id pub-id-type="medline">02880440</pub-id></nlm-citation></ref><ref id="ref94"><label>94</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yeo</surname><given-names>CT</given-names> </name><name name-style="western"><surname>Ungi</surname><given-names>T</given-names> </name><name name-style="western"><surname>U-Thainual</surname><given-names>P</given-names> </name><name name-style="western"><surname>Lasso</surname><given-names>A</given-names> </name><name name-style="western"><surname>McGraw</surname><given-names>RC</given-names> </name><name name-style="western"><surname>Fichtinger</surname><given-names>G</given-names> </name></person-group><article-title>The effect of augmented reality training on percutaneous needle placement in spinal facet joint injections</article-title><source>IEEE Trans Biomed Eng</source><year>2011</year><month>07</month><volume>58</volume><issue>7</issue><fpage>2031</fpage><lpage>2037</lpage><pub-id pub-id-type="doi">10.1109/TBME.2011.2132131</pub-id><pub-id pub-id-type="medline">21435970</pub-id></nlm-citation></ref><ref id="ref95"><label>95</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gokkurt Yilmaz</surname><given-names>BN</given-names> </name><name name-style="western"><surname>Ozbey</surname><given-names>F</given-names> </name><name name-style="western"><surname>Yilmaz</surname><given-names>BE</given-names> </name></person-group><article-title>Effect of artificial intelligence-assisted personalized feedback on radiographic diagnostic performance of dental students: a controlled study</article-title><source>BMC Med Educ</source><year>2025</year><month>10</month><day>10</day><volume>25</volume><issue>1</issue><fpage>1403</fpage><pub-id pub-id-type="doi">10.1186/s12909-025-07875-4</pub-id><pub-id pub-id-type="medline">41074130</pub-id></nlm-citation></ref><ref id="ref96"><label>96</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zeng</surname><given-names>J</given-names> </name><name name-style="western"><surname>Sun</surname><given-names>K</given-names> </name><name name-style="western"><surname>Qin</surname><given-names>P</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>S</given-names> </name></person-group><article-title>Enhancing ophthalmology students&#x2019; awareness of retinitis pigmentosa: assessing the efficacy of ChatGPT in AI-assisted teaching of rare diseases-a quasi-experimental study</article-title><source>Front Med (Lausanne)</source><year>2025</year><volume>12</volume><fpage>1534294</fpage><pub-id pub-id-type="doi">10.3389/fmed.2025.1534294</pub-id><pub-id pub-id-type="medline">40171502</pub-id></nlm-citation></ref><ref id="ref97"><label>97</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Al Kahf</surname><given-names>S</given-names> </name><name name-style="western"><surname>Roux</surname><given-names>B</given-names> </name><name name-style="western"><surname>Clerc</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Chatbot-based serious games: a useful tool for training medical students? A randomized controlled trial</article-title><source>PLoS ONE</source><year>2023</year><volume>18</volume><issue>3</issue><fpage>e0278673</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0278673</pub-id><pub-id pub-id-type="medline">36913346</pub-id></nlm-citation></ref><ref id="ref98"><label>98</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Castano-Villegas</surname><given-names>N</given-names> </name><name name-style="western"><surname>Llano</surname><given-names>I</given-names> </name><name name-style="western"><surname>Villa</surname><given-names>MC</given-names> </name><name name-style="western"><surname>Zea</surname><given-names>J</given-names> </name><name name-style="western"><surname>Vel&#x00E1;squez</surname><given-names>L</given-names> </name></person-group><article-title>Clinical med students&#x2019; validation of arkangel AI are their responses any better when supported by the AI?</article-title><source>SSRN</source><comment>Preprint posted online on 2025</comment><pub-id pub-id-type="doi">10.2139/ssrn.5679000</pub-id></nlm-citation></ref><ref id="ref99"><label>99</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Dao</surname><given-names>L</given-names> </name><name name-style="western"><surname>Harmouch</surname><given-names>SS</given-names> </name><name name-style="western"><surname>Chin</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Effect of an artificial intelligence chest x-ray disease prediction system on the radiological education of medical students: a pilot study</article-title><source>medRxiv</source><comment>Preprint posted online on 2022</comment><pub-id pub-id-type="doi">10.1101/2022.07.11.22277278</pub-id></nlm-citation></ref><ref id="ref100"><label>100</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>D&#x00F6;ner</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ceyhan</surname><given-names>&#x00D6;</given-names> </name><name name-style="western"><surname>Ta&#x015F;ci</surname><given-names>S</given-names> </name></person-group><article-title>Effects of artificial intelligence-supported case analysis method on nursing students&#x2019; clinical competencies: mixed design research</article-title><source>Nurse Educ</source><year>2026</year><volume>51</volume><issue>3</issue><fpage>E199</fpage><lpage>E205</lpage><pub-id pub-id-type="doi">10.1097/NNE.0000000000002086</pub-id><pub-id pub-id-type="medline">41441838</pub-id></nlm-citation></ref><ref id="ref101"><label>101</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fang</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Reynaldi</surname><given-names>R</given-names> </name><name name-style="western"><surname>Araminta</surname><given-names>AS</given-names> </name><etal/></person-group><article-title>Artificial intelligence (AI)-driven dental education: exploring the role of chatbots in a clinical learning environment</article-title><source>J Prosthet Dent</source><year>2025</year><month>10</month><volume>134</volume><issue>4</issue><fpage>1296</fpage><lpage>1303</lpage><pub-id pub-id-type="doi">10.1016/j.prosdent.2024.03.038</pub-id><pub-id pub-id-type="medline">38644064</pub-id></nlm-citation></ref><ref id="ref102"><label>102</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>B bei</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>X yan</given-names> </name><name name-style="western"><surname>Luo</surname><given-names>Y cheng</given-names> </name><name name-style="western"><surname>Teng</surname><given-names>M miao</given-names> </name><name name-style="western"><surname>Weng</surname><given-names>X</given-names> </name></person-group><article-title>Implementation and evaluation of an optimized surgical clerkship teaching model utilizing ChatGPT</article-title><source>BMC Med Educ</source><year>2024</year><volume>24</volume><issue>1</issue><fpage>CN</fpage><pub-id pub-id-type="doi">10.1186/s12909-024-06575-9</pub-id><pub-id pub-id-type="medline">02794097</pub-id></nlm-citation></ref><ref id="ref103"><label>103</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jiang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Fu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Enhancing medical education with chatbots: a randomized controlled trial on standardized patients for colorectal cancer</article-title><source>BMC Med Educ</source><year>2024</year><month>12</month><day>20</day><volume>24</volume><issue>1</issue><fpage>1511</fpage><pub-id pub-id-type="doi">10.1186/s12909-024-06530-8</pub-id><pub-id pub-id-type="medline">39707245</pub-id></nlm-citation></ref><ref id="ref104"><label>104</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kobayashi</surname><given-names>M</given-names> </name><name name-style="western"><surname>Iwamoto</surname><given-names>M</given-names> </name><name name-style="western"><surname>Une</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kurazume</surname><given-names>R</given-names> </name><name name-style="western"><surname>Nakazawa</surname><given-names>A</given-names> </name><name name-style="western"><surname>Honda</surname><given-names>M</given-names> </name></person-group><article-title>Simulated communication skills training program effects using augmented reality with real&#x2010;time feedback: a randomized control study</article-title><source>Alzheimer&#x2019;s &#x0026; Dementia</source><year>2022</year><month>12</month><volume>18</volume><issue>S8</issue><comment><ext-link ext-link-type="uri" xlink:href="https://alz-journals.onlinelibrary.wiley.com/toc/15525279/18/S8">https://alz-journals.onlinelibrary.wiley.com/toc/15525279/18/S8</ext-link></comment><pub-id pub-id-type="doi">10.1002/alz.062055</pub-id></nlm-citation></ref><ref id="ref105"><label>105</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lau</surname><given-names>YH</given-names> </name><name name-style="western"><surname>Acharyya</surname><given-names>S</given-names> </name><name name-style="western"><surname>Wee</surname><given-names>CWL</given-names> </name><etal/></person-group><article-title>Effectiveness of traditional, artificial intelligence-assisted, and virtual reality training modalities for focused cardiac ultrasound skill acquisition: a randomised controlled study</article-title><source>Ultrasound J</source><year>2025</year><month>11</month><day>21</day><volume>17</volume><issue>1</issue><fpage>61</fpage><pub-id pub-id-type="doi">10.1186/s13089-025-00469-7</pub-id><pub-id pub-id-type="medline">41269495</pub-id></nlm-citation></ref><ref id="ref106"><label>106</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>HY</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>J</given-names> </name><name name-style="western"><surname>Choi</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Comparing AI chatbot simulation and peer role-play for OSCE preparation: a pilot randomized controlled trial</article-title><source>BMC Med Educ</source><year>2025</year><month>11</month><day>24</day><volume>25</volume><issue>1</issue><fpage>1755</fpage><pub-id pub-id-type="doi">10.1186/s12909-025-08308-y</pub-id><pub-id pub-id-type="medline">41286823</pub-id></nlm-citation></ref><ref id="ref107"><label>107</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Louie</surname><given-names>YLJ</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>CTL</given-names> </name><name name-style="western"><surname>Fong</surname><given-names>BYR</given-names> </name><name name-style="western"><surname>Yung</surname><given-names>KK</given-names> </name><name name-style="western"><surname>See</surname><given-names>C</given-names> </name><name name-style="western"><surname>Poon</surname><given-names>LCY</given-names> </name><etal/></person-group><article-title>AI vs conventional learning of clinical skills in O&#x0026;G - A comparison</article-title><source>BJOG</source><year>2025</year><volume>132</volume><issue>18&#x2010;9</issue><fpage>CN</fpage><pub-id pub-id-type="doi">10.1111/1471-0528.18215</pub-id><pub-id pub-id-type="medline">02920561</pub-id></nlm-citation></ref><ref id="ref108"><label>108</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Luo</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Bi</surname><given-names>S</given-names> </name><name name-style="western"><surname>Pang</surname><given-names>J</given-names> </name><etal/></person-group><article-title>A large language model digital patient system enhances ophthalmology history taking skills</article-title><source>NPJ Digit Med</source><year>2025</year><month>08</month><day>4</day><volume>8</volume><issue>1</issue><fpage>502</fpage><pub-id pub-id-type="doi">10.1038/s41746-025-01841-6</pub-id><pub-id pub-id-type="medline">40760042</pub-id></nlm-citation></ref><ref id="ref109"><label>109</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Meng</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>J</given-names> </name><name name-style="western"><surname>Lai</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>N</given-names> </name><name name-style="western"><surname>Ho</surname><given-names>CF</given-names> </name><name name-style="western"><surname>Tartari</surname><given-names>E</given-names> </name><etal/></person-group><article-title>AI-based augmented reality for hand hygiene: preliminary CRCT results in health sciences students</article-title><source>Antimicrob Resist Infect Control</source><year>2025</year><fpage>CN</fpage><pub-id pub-id-type="doi">10.1186/s13756-025-01610-w</pub-id><pub-id pub-id-type="medline">02923120</pub-id></nlm-citation></ref><ref id="ref110"><label>110</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Simsek-Cetinkaya</surname><given-names>S</given-names> </name><name name-style="western"><surname>Cakir</surname><given-names>SK</given-names> </name></person-group><article-title>Evaluation of the effectiveness of artificial intelligence assisted interactive screen-based simulation in breast self-examination: an innovative approach in nursing students</article-title><source>Nurse Educ Today</source><year>2023</year><month>08</month><volume>127</volume><fpage>105857</fpage><pub-id pub-id-type="doi">10.1016/j.nedt.2023.105857</pub-id><pub-id pub-id-type="medline">37253303</pub-id></nlm-citation></ref><ref id="ref111"><label>111</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Usta</surname><given-names>SN</given-names> </name><name name-style="western"><surname>Silva</surname><given-names>E</given-names> </name><name name-style="western"><surname>Tekkanat</surname><given-names>H</given-names> </name><name name-style="western"><surname>Keskin</surname><given-names>C</given-names> </name></person-group><article-title>Virtual reality haptic simulators and mobile applications-potential AI-enhanced tools for improving clinical endodontic training: a randomized controlled trial</article-title><source>J Endod</source><year>2026</year><month>03</month><volume>52</volume><issue>3</issue><fpage>337</fpage><lpage>344</lpage><pub-id pub-id-type="doi">10.1016/j.joen.2025.11.005</pub-id><pub-id pub-id-type="medline">41242655</pub-id></nlm-citation></ref><ref id="ref112"><label>112</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Vannaprathip</surname><given-names>N</given-names> </name><name name-style="western"><surname>Haddawy</surname><given-names>P</given-names> </name><name name-style="western"><surname>Schultheis</surname><given-names>H</given-names> </name><name name-style="western"><surname>Suebnukarn</surname><given-names>S</given-names> </name></person-group><article-title>SDMentor: a virtual reality-based intelligent tutoring system for surgical decision making in dentistry</article-title><source>Artif Intell Med</source><year>2025</year><month>04</month><volume>162</volume><fpage>103092</fpage><pub-id pub-id-type="doi">10.1016/j.artmed.2025.103092</pub-id><pub-id pub-id-type="medline">40015210</pub-id></nlm-citation></ref><ref id="ref113"><label>113</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>C</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>L</given-names> </name><name name-style="western"><surname>Han</surname><given-names>M</given-names> </name><name name-style="western"><surname>Li</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>N</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>C</given-names> </name></person-group><article-title>Application of ChatGPT-based blended medical teaching in clinical education of hepatobiliary surgery</article-title><source>Med Teach</source><year>2025</year><month>03</month><volume>47</volume><issue>3</issue><fpage>445</fpage><lpage>449</lpage><pub-id pub-id-type="doi">10.1080/0142159X.2024.2339412</pub-id><pub-id pub-id-type="medline">38614458</pub-id></nlm-citation></ref><ref id="ref114"><label>114</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wu</surname><given-names>D</given-names> </name><name name-style="western"><surname>Xiang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Artificial intelligence-tutoring problem-based learning in ophthalmology clerkship</article-title><source>Ann Transl Med</source><year>2020</year><month>06</month><volume>8</volume><issue>11</issue><fpage>700</fpage><pub-id pub-id-type="doi">10.21037/atm.2019.12.15</pub-id><pub-id pub-id-type="medline">32617320</pub-id></nlm-citation></ref><ref id="ref115"><label>115</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Benfatah</surname><given-names>M</given-names> </name><name name-style="western"><surname>Elazizi</surname><given-names>I</given-names> </name><name name-style="western"><surname>Lamiri</surname><given-names>A</given-names> </name><etal/></person-group><article-title>AI-assisted prebriefing to enhance simulation readiness in nursing education</article-title><source>Teach Learn Nurs</source><year>2026</year><month>01</month><volume>21</volume><issue>1</issue><fpage>e57</fpage><lpage>e63</lpage><pub-id pub-id-type="doi">10.1016/j.teln.2025.07.030</pub-id></nlm-citation></ref><ref id="ref116"><label>116</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Benfatah</surname><given-names>M</given-names> </name><name name-style="western"><surname>Marfak</surname><given-names>A</given-names> </name><name name-style="western"><surname>Saad</surname><given-names>E</given-names> </name><name name-style="western"><surname>Hilali</surname><given-names>A</given-names> </name><name name-style="western"><surname>Nejjari</surname><given-names>C</given-names> </name><name name-style="western"><surname>Youlyouz-Marfak</surname><given-names>I</given-names> </name></person-group><article-title>Assessing the efficacy of ChatGPT as a virtual patient in nursing simulation training: a study on nursing students&#x2019; experience</article-title><source>Teach Learn Nurs</source><year>2024</year><month>07</month><volume>19</volume><issue>3</issue><fpage>e486</fpage><lpage>e493</lpage><pub-id pub-id-type="doi">10.1016/j.teln.2024.02.005</pub-id></nlm-citation></ref><ref id="ref117"><label>117</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jin</surname><given-names>HK</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>HE</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>E</given-names> </name></person-group><article-title>Performance of ChatGPT-3.5 and GPT-4 in national licensing examinations for medicine, pharmacy, dentistry, and nursing: a systematic review and meta-analysis</article-title><source>BMC Med Educ</source><year>2024</year><month>09</month><day>16</day><volume>24</volume><issue>1</issue><fpage>1013</fpage><pub-id pub-id-type="doi">10.1186/s12909-024-05944-8</pub-id><pub-id pub-id-type="medline">39285377</pub-id></nlm-citation></ref><ref id="ref118"><label>118</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>M</given-names> </name><name name-style="western"><surname>Okuhara</surname><given-names>T</given-names> </name><name name-style="western"><surname>Chang</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Performance of ChatGPT across different versions in medical licensing examinations worldwide: systematic review and meta-analysis</article-title><source>J Med Internet Res</source><year>2024</year><month>07</month><day>25</day><volume>26</volume><fpage>e60807</fpage><pub-id pub-id-type="doi">10.2196/60807</pub-id><pub-id pub-id-type="medline">39052324</pub-id></nlm-citation></ref><ref id="ref119"><label>119</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Murray</surname><given-names>E</given-names> </name></person-group><article-title>Challenges in educational research</article-title><source>Med Educ</source><year>2002</year><month>02</month><volume>36</volume><issue>2</issue><fpage>110</fpage><lpage>112</lpage><pub-id pub-id-type="doi">10.1046/j.1365-2923.2002.01155.x</pub-id><pub-id pub-id-type="medline">11869436</pub-id></nlm-citation></ref><ref id="ref120"><label>120</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Riding</surname><given-names>R</given-names> </name></person-group><article-title>Individual differences and educational performance</article-title><source>Educ Psychol (Lond)</source><year>2005</year><month>12</month><volume>25</volume><issue>6</issue><fpage>659</fpage><lpage>672</lpage><pub-id pub-id-type="doi">10.1080/01443410500344712</pub-id></nlm-citation></ref><ref id="ref121"><label>121</label><nlm-citation citation-type="web"><article-title>The AI research assistant</article-title><source>Elicit</source><year>2025</year><access-date>2025-08-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://elicit.com">https://elicit.com</ext-link></comment></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Compilation of parts 1-11.</p><media xlink:href="mededu_v12i1e88933_app1.docx" xlink:title="DOCX File, 1084 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Full analyses with forest plots.</p><media xlink:href="mededu_v12i1e88933_app2.pdf" xlink:title="PDF File, 323 KB"/></supplementary-material><supplementary-material id="app3"><label>Checklist 1</label><p>PRISMA 2020 checklist.</p><media xlink:href="mededu_v12i1e88933_app3.docx" xlink:title="DOCX File, 30 KB"/></supplementary-material><supplementary-material id="app4"><label>Checklist</label><p>PRISMA-S checklist.</p><media xlink:href="mededu_v12i1e88933_app4.docx" xlink:title="DOCX File, 17 KB"/></supplementary-material></app-group></back></article>