<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Educ</journal-id><journal-id journal-id-type="publisher-id">mededu</journal-id><journal-id journal-id-type="index">20</journal-id><journal-title>JMIR Medical Education</journal-title><abbrev-journal-title>JMIR Med Educ</abbrev-journal-title><issn pub-type="epub">2369-3762</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v10i1e57451</article-id><article-id pub-id-type="doi">10.2196/57451</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Performance of GPT-3.5 and GPT-4 on the Korean Pharmacist Licensing Examination: Comparison Study</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Jin</surname><given-names>Hye Kyung</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Kim</surname><given-names>EunYoung</given-names></name><degrees>BCPS, PharmD, PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref></contrib></contrib-group><aff id="aff1"><institution>Research Institute of Pharmaceutical Sciences, College of Pharmacy, Chung-Ang University</institution>, <addr-line>Seoul</addr-line>, <country>Republic of Korea</country></aff><aff id="aff2"><institution>Data Science, Evidence-Based and Clinical Research Laboratory, Department of Health, Social, and Clinical Pharmacy, College of Pharmacy, Chung-Ang University</institution>, <addr-line>Seoul</addr-line>, <country>Republic of Korea</country></aff><aff id="aff3"><institution>Division of Licensing of Medicines and Regulatory Science, The Graduate School of Pharmaceutical Management and Regulatory Science Policy, The Graduate School of Pharmaceutical Regulatory Sciences, Chung-Ang University</institution>, <addr-line>84 Heukseok-Ro, Dongjak-gu</addr-line><addr-line>Seoul</addr-line>, <country>Republic of Korea</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Lesselroth</surname><given-names>Blake</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Ashraf</surname><given-names>Amir Reza</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Chatzimina</surname><given-names>Maria</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Song</surname><given-names>Sanghoun</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to EunYoung Kim, BCPS, PharmD, PhD, Division of Licensing of Medicines and Regulatory Science, The Graduate School of Pharmaceutical Management and Regulatory Science Policy, The Graduate School of Pharmaceutical Regulatory Sciences, Chung-Ang University, 84 Heukseok-Ro, Dongjak-gu, Seoul, 06974, Republic of Korea, 82 2-820-5791, 82 2-816-7338; <email>eykimjcb777@cau.ac.kr</email></corresp></author-notes><pub-date pub-type="collection"><year>2024</year></pub-date><pub-date pub-type="epub"><day>4</day><month>12</month><year>2024</year></pub-date><volume>10</volume><elocation-id>e57451</elocation-id><history><date date-type="received"><day>17</day><month>02</month><year>2024</year></date><date date-type="rev-recd"><day>28</day><month>08</month><year>2024</year></date><date date-type="accepted"><day>09</day><month>10</month><year>2024</year></date></history><copyright-statement>&#x00A9;Hye Kyung Jin, EunYoung Kim. Originally published in JMIR Medical Education (<ext-link ext-link-type="uri" xlink:href="https://mededu.jmir.org">https://mededu.jmir.org</ext-link>), 4.12.2024. </copyright-statement><copyright-year>2024</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Education, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://mededu.jmir.org/">https://mededu.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://mededu.jmir.org/2024/1/e57451"/><abstract><sec><title>Background</title><p>ChatGPT, a recently developed artificial intelligence chatbot and a notable large language model, has demonstrated improved performance on medical field examinations. However, there is currently little research on its efficacy in languages other than English or in pharmacy-related examinations.</p></sec><sec><title>Objective</title><p>This study aimed to evaluate the performance of GPT models on the Korean Pharmacist Licensing Examination (KPLE).</p></sec><sec sec-type="methods"><title>Methods</title><p>We evaluated the percentage of correct answers provided by 2 different versions of ChatGPT (GPT-3.5 and GPT-4) for all multiple-choice single-answer KPLE questions, excluding image-based questions. In total, 320, 317, and 323 questions from the 2021, 2022, and 2023 KPLEs, respectively, were included in the final analysis, which consisted of 4 units: Biopharmacy, Industrial Pharmacy, Clinical and Practical Pharmacy, and Medical Health Legislation.</p></sec><sec sec-type="results"><title>Results</title><p>The 3-year average percentage of correct answers was 86.5% (830/960) for GPT-4 and 60.7% (583/960) for GPT-3.5. GPT model accuracy was highest in Biopharmacy (GPT-3.5 77/96, 80.2% in 2022; GPT-4 87/90, 96.7% in 2021) and lowest in Medical Health Legislation (GPT-3.5 8/20, 40% in 2022; GPT-4 12/20, 60% in 2022). Additionally, when comparing the performance of artificial intelligence with that of human participants, pharmacy students outperformed GPT-3.5 but not GPT-4.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>In the last 3 years, GPT models have performed very close to or exceeded the passing threshold for the KPLE. This study demonstrates the potential of large language models in the pharmacy domain; however, extensive research is needed to evaluate their reliability and ensure their secure application in pharmacy contexts due to several inherent challenges. Addressing these limitations could make GPT models more effective auxiliary tools for pharmacy education.</p></sec></abstract><kwd-group><kwd>GPT-3.5</kwd><kwd>GPT-4</kwd><kwd>Korean</kwd><kwd>Korean Pharmacist Licensing Examination</kwd><kwd>KPLE</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Recently, artificial intelligence (AI) based on large language models (LLMs) has shown promise in various fields and industries [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref3">3</xref>]. On November 30, 2022, ChatGPT (GPT-3.5), an AI language model trained using deep-learning algorithms, was released by OpenAI [<xref ref-type="bibr" rid="ref4">4</xref>]. Since its release, ChatGPT has become a popular topic, showing promise in optimizing performance on various examinations, such as the US Certified Public Accountant examination [<xref ref-type="bibr" rid="ref5">5</xref>] and those in MBA and law school programs [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>]. ChatGPT has also demonstrated potential efficacy in the health care field, such as in optimizing clinical workflows and supporting clinical decisions and diagnoses [<xref ref-type="bibr" rid="ref8">8</xref>-<xref ref-type="bibr" rid="ref11">11</xref>]. Furthermore, it has performed adequately in medical education, with demonstrated effectiveness on 6 different national medical licensing examinations, including those in Italy, France, Spain, the United States, India, and the United Kingdom [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref13">13</xref>]. These findings indicate the potential of ChatGPT as an innovative method for medical education and as a study resource, with efficient and accurate responses [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref15">15</xref>].</p><p>ChatGPT&#x2019;s demonstrated ability to perform well on medical and licensing examinations suggests that this technology could also be applicable to other health care&#x2013;related examinations. In Japan, ChatGPT&#x2019;s performance on the National Nurse Examinations from 2019 to 2023 showed an average accuracy over 5 years of 75.1% for basic knowledge questions and 64.5% for general questions, with the passing criteria being 80% and approximately 60%, respectively [<xref ref-type="bibr" rid="ref16">16</xref>]. In addition, ChatGPT achieved response accuracy rates between 54.1%-63.8% across 1510 questions on Taiwan&#x2019;s registered nurse license examination [<xref ref-type="bibr" rid="ref17">17</xref>]. Similarly, for dentistry questions via the Swiss Federal Licensing Examination in Dental Medicine, it showed an average accuracy rate of 63.3% [<xref ref-type="bibr" rid="ref18">18</xref>].</p><p>Compared with its predecessor GPT-3.5, the proficiency of GPT-4 in responding to the United States Medical Licensing Examination (USMLE) questions showed an accuracy of 90.7% across the entire USMLE, which surpassed the passing threshold of approximately 60% accuracy [<xref ref-type="bibr" rid="ref19">19</xref>]. Furthermore, while GPT-3.5 scored 42.8% on the National Medical Licensing Examination in Japan, GPT-4 achieved a score of 81.5%, surpassing the passing threshold of 72% [<xref ref-type="bibr" rid="ref20">20</xref>]. However, despite their high performance in medical education, the utility of these programs has yet to be extensively studied in the context of pharmacy education. In particular, research on their performance on national pharmacist licensing examinations is limited, with Nisar et al [<xref ref-type="bibr" rid="ref21">21</xref>] being one of the few relevant studies. Their study demonstrated that ChatGPT can achieve satisfactory accuracy and relevance when responding to pharmacology textbook queries related to pharmacokinetics, clinical applications, adverse effects, and drug interactions, suggesting that the application of LLMs in the pharmacy domain is increasingly viable. However, the study also highlighted the need for further improvements in ChatGPT&#x2019;s performance when addressing more intricate and complex questions. This indicates that while promising, the technology still requires refinement for broader clinical applications. Given that pharmacists are responsible for providing comprehensive drug information and suggesting personalized treatment plans for patients, the performance of AI technologies like ChatGPT that can assist in these tasks must be evaluated.</p><p>This study aims to investigate the accuracy of GPT-3.5 and GPT-4 on the Korean Pharmacist Licensing Examinations (KPLEs) conducted from 2021 to 2023. Responses from the KPLEs were used to conduct a comparative analysis of ChatGPT&#x2019;s performance across various units.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>ChatGPT Models</title><p>In this study, we assessed the performance of 2 versions of an AI model: GPT-3.5 and GPT-4 (a newer, paid version available through the ChatGPT+ platform). We accessed these models through the online interface provided on OpenAI&#x2019;s website rather than the application programming interface.</p></sec><sec id="s2-2"><title>KPLE Datasets</title><p>We used the original questions from the 72nd, 73rd, and 74th KPLEs, respectively held in 2021, 2022, and 2023. These examinations are conducted annually and comprise 350 questions classified into 4 units: Biopharmacy (100 questions), which includes biochemistry, molecular biology, microbiology, immunology, pharmacology, preventive pharmacy, and pathophysiology; Industrial Pharmacy (90 questions), covering areas such as physical pharmacy, synthetic chemistry, medicinal chemistry, pharmaceutical analysis, pharmaceutics, pharmacognosy, and herbal medicine; Clinical and Practical Pharmacy, divided into part I (77 questions) focusing on pharmacotherapy and part II (63 questions) covering pharmacy practice, pharmaceutical manufacturing, pharmaceutical quality control, and pharmacy administration and management; and Medical Health Legislation (20 questions), which includes Pharmaceutical Affairs Act, Narcotics Control Act, National Health Promotion Act, Framework Act on Health and Medical Services, National Health Insurance Act, and Regional Public Health Act. Each question is worth 1 point. The passing criteria for the KPLE requires a minimum score of 40% or higher for each subject and a total score of 60% or higher, equating to at least 210 points across all subjects.</p></sec><sec id="s2-3"><title>Procedures</title><p>The questions from the KPLEs, along with their multiple-choice responses, were utilized in their Korean format, in conjunction with the official national examination guidelines. Image-based questions that ChatGPT could not recognize were excluded when calculating accuracy. Specifically, 30, 33, and 27 questions from 2021, 2022, and 2023 were excluded from the final analysis. To elicit diverse responses, we provided specific instructions via a prompt stating, &#x201C;Only one best option can be selected.&#x201D; The determination of &#x201C;correct&#x201D; responses to the inquiries posed to GPT-3.5 and GPT-4 was grounded in the Korea Health Personnel Licensing Examination Institute&#x2019;s (KHPLEI) database and accessed from their official website [<xref ref-type="bibr" rid="ref22">22</xref>]. Only responses that strictly adhered to the question instructions were considered &#x201C;correct.&#x201D; Across all 3 years, the most image-based questions that were excluded came from Unit 2, Industrial Pharmacy, ranging from 20 (22.2%) to 24 (26.7%) of the 90 questions. Additionally, for non&#x2013;image-based questions included in tables, we redescribed the information in a manner similar to how a test administrator or proctor reads examination questions for visually impaired students (ie, those who are blind or have low vision). Responses that were ambiguous, contained clear errors, or for which ChatGPT generated multiple answers with incorrect options instead of a single response were considered inaccurate. We provided no additional content or hints in our study, aiming to simulate a real examination scenario. However, we followed up by asking, &#x201C;Do you have confidence in this?&#x201D; to assess the model&#x2019;s stability and consistency and to prompt ChatGPT&#x2019;s potential re-evaluation of its initial response. However, these data were not included in the results and were kept for reference purposes only. If the model&#x2019;s answer changed, it could suggest a degree of doubt in its initial response. Monitoring how frequently and under what conditions the model adjusts its answers provides important information about its capacity for self-correction, which is crucial for learning and decision-making. We collected and evaluated responses from GPT-3.5 and GPT-4 on December 17 and 18, 2023, according to these criteria for correctness.</p></sec><sec id="s2-4"><title>Data Analysis</title><p>We employed standard descriptive statistics including numbers, proportions, and averages for each dataset. A Fisher exact test was utilized to compare the rates of correct responses. Statistical analyses were performed using SPSS software, version 29 (IBM Corp). All tests were 2-tailed, and a <italic>P</italic> value of less than .05 was considered statistically significant.</p></sec><sec id="s2-5"><title>Ethics Approval</title><p>Ethical approval was not required for this study because it involved the analysis of data from a publicly available database. The test questions and answers used in this study were initially developed and copyrighted by the KHPLEI and are available for academic research purposes. The KHPLEI holds all copyrights pertaining to the examination content and ensured that this research complied with these copyrights without violation.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>ChatGPT&#x2019;s Performance</title><p>Out of a total of 350 questions each, 320, 317, and 323 questions from the 2021, 2022, and 2023 KPLEs, respectively, were included in the final analysis. The 3-year average percentage of correct answers from GPT-3.5 and GPT-4 were 60.7% (583/960) and 86.5% (830/960), respectively. When analyzing accuracy rates by unit, the GPT models showed their most notable performance over the 3-year period in Biopharmacy. Specifically, GPT-4 achieved a 96.7% (87/90) accuracy rate in 2021, while GPT-3.5 recorded a lower but still impressive accuracy rate of 80.2% (77/96) in 2022. In contrast, the accuracy rates were lowest in Medical Health Legislation out of the 4 units. The lowest scores were observed in 2022, with accuracy rates of 40% (8/20) and 60% (12/20) for GPT-3.5 and GPT-4, respectively.</p><p>Beyond our numerical analysis, we found that GPT-4 provided more comprehensive and accurate explanations for its responses compared to its predecessors.</p><p>In the 2021 KPLE, 320 (91.4%) of the 350 questions were suitable for analysis in both GPT-3.5 and GPT-4, excluding 30 image-based questions. Of these, GPT-3.5 correctly answered 200 questions, resulting in an accuracy rate of 62.5%. In contrast, GPT-4 answered 275 questions correctly, achieving an accuracy rate of 85.9%. <xref ref-type="table" rid="table1">Table 1</xref> presents the detailed scores for each unit of the 2021 KPLE. Regarding specific question types, GPT-4 notably surpassed GPT-3.5 in achieving higher rates of correct responses across all sections, with statistically significant differences (all <italic>P</italic>&#x003C;.05) observed in all comparisons except for Unit 4. The highest accuracy rates were in Biopharmacy (GPT-3.5 66/90, 73.3%; GPT-4 87/90, 96.7%), while the lowest were in Medical Health Legislation (GPT-3.5 10/20, 50%; GPT-4 13/20, 65%).</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Comparison of GPT-3.5&#x2019;s and GPT-4&#x2019;s performances on the 2021 Korean Pharmacist Licensing Examination.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Question category</td><td align="left" valign="bottom">All questions, n</td><td align="left" valign="bottom">Student correct response rate, n (%)</td><td align="left" valign="bottom">Questions answerable by GPT, n</td><td align="left" valign="bottom">GPT-3.5 correct response rate, n (%)</td><td align="left" valign="bottom">GPT-4 correct response rate, n (%)</td><td align="left" valign="bottom">Passing criteria, %</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Total</td><td align="left" valign="top">350</td><td align="left" valign="top">246 (70.3)</td><td align="left" valign="top">320</td><td align="left" valign="top">200 (62.5)</td><td align="left" valign="top">275 (85.9)</td><td align="left" valign="top">&#x2265;60</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Unit 1: Biopharmacy<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup></td><td align="left" valign="top">100</td><td align="left" valign="top">70.3 (70.3)</td><td align="left" valign="top">90</td><td align="left" valign="top">66 (73.3)</td><td align="left" valign="top">87 (96.7)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Unit 2: Industrial Pharmacy<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup></td><td align="left" valign="top">90</td><td align="left" valign="top">60.2 (66.9)</td><td align="left" valign="top">70</td><td align="left" valign="top">39 (55.7)</td><td align="left" valign="top">55 (78.6)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">.007</td></tr><tr><td align="left" valign="top">Unit 3: Clinical and Practical Pharmacy I and II<sup><xref ref-type="table-fn" rid="table1fn3">c</xref></sup></td><td align="left" valign="top">140</td><td align="left" valign="top">100.2 (71.6)</td><td align="left" valign="top">140</td><td align="left" valign="top">85 (60.7)</td><td align="left" valign="top">120 (85.7)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Unit 4: Medical Health Legislation<sup><xref ref-type="table-fn" rid="table1fn4">d</xref></sup></td><td align="left" valign="top">20</td><td align="left" valign="top">15.3 (76.5)</td><td align="left" valign="top">20</td><td align="left" valign="top">10 (50)</td><td align="left" valign="top">13 (65)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">.52</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>Biochemistry, molecular biology, microbiology, immunology, pharmacology, preventive pharmacy, and pathophysiology.</p></fn><fn id="table1fn2"><p><sup>b</sup>Physical pharmacy, synthetic chemistry, medicinal chemistry, pharmaceutical analysis, pharmaceutics, pharmacognosy, and herbal medicine.</p></fn><fn id="table1fn3"><p><sup>c</sup>Pharmacotherapy, pharmacy practice, pharmaceutical manufacturing, pharmaceutical quality control, and pharmacy administration and management.</p></fn><fn id="table1fn4"><p><sup>d</sup>Pharmaceutical Affairs Act, Narcotics Control Act, National Health Promotion Act, Framework Act on Health and Medical Services, National Health Insurance Act, and Regional Public Health Act.</p></fn></table-wrap-foot></table-wrap><p>Out of 350 questions on the 2022 KPLE, 317 (90.6%) were analyzed by both GPT-3.5 and GPT-4, and the correct response rates are shown in <xref ref-type="table" rid="table2">Table 2</xref>. Neither version could process 33 image-based questions. GPT-3.5 correctly answered 188 of the 317 questions, resulting in a 59.3% accuracy rate, falling short of the required 60% standard for passing. In contrast, GPT-4 answered 273 of the 317 questions correctly, achieving an accuracy rate of 86.1%. Regarding question types, the highest accuracy rates were observed in Biopharmacy, at 80.2% (77/96) and 95.8% (92/96) for GPT-3.5 and GPT-4, respectively. In contrast, the lowest accuracy rates were recorded for Medical Health Legislation, with GPT-3.5 and GPT-4 achieving 40% (8/20) and 60% (12/20), respectively. Notably, GPT-4 showed a considerable 40.9% increase in accuracy for Industrial Pharmacy questions compared with GPT-3.5 (<italic>P</italic>&#x003C;.001).</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Comparison of GPT-3.5&#x2019;s and GPT-4&#x2019;s performances on the 2022 Korean Pharmacist Licensing Examination.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Question category</td><td align="left" valign="bottom">All questions, n</td><td align="left" valign="bottom">Student correct response rate, n (%)</td><td align="left" valign="bottom">Questions answerable by GPT, n</td><td align="left" valign="bottom">GPT-3.5 correct response rate, n (%)</td><td align="left" valign="bottom">GPT-4 correct response rate, n (%)</td><td align="left" valign="bottom">Passing criteria, %</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Total</td><td align="left" valign="top">350</td><td align="left" valign="top">248 (70.9)</td><td align="left" valign="top">317</td><td align="left" valign="top">188 (59.3)</td><td align="left" valign="top">273 (86.1)</td><td align="left" valign="top">&#x2265;60</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Unit 1: Biopharmacy<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup></td><td align="left" valign="top">100</td><td align="left" valign="top">67 (67)</td><td align="left" valign="top">96</td><td align="left" valign="top">77 (80.2)</td><td align="left" valign="top">92 (95.8)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">.001</td></tr><tr><td align="left" valign="top">Unit 2: Industrial Pharmacy<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup></td><td align="left" valign="top">90</td><td align="left" valign="top">62.8 (69.8)</td><td align="left" valign="top">66</td><td align="left" valign="top">32 (48.5)</td><td align="left" valign="top">59 (89.4)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Unit 3: Clinical and Practical Pharmacy I and II<sup><xref ref-type="table-fn" rid="table2fn3">c</xref></sup></td><td align="left" valign="top">140</td><td align="left" valign="top">101.7 (72.6)</td><td align="left" valign="top">135</td><td align="left" valign="top">71 (52.6)</td><td align="left" valign="top">110 (81.5)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Unit 4: Medical Health Legislation<sup><xref ref-type="table-fn" rid="table2fn4">d</xref></sup></td><td align="left" valign="top">20</td><td align="left" valign="top">16.5 (82.5)</td><td align="left" valign="top">20</td><td align="left" valign="top">8 (40)</td><td align="left" valign="top">12 (60)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">.34</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>Biochemistry, molecular biology, microbiology, immunology, pharmacology, preventive pharmacy, and pathophysiology.</p></fn><fn id="table2fn2"><p><sup>b</sup>Physical pharmacy, synthetic chemistry, medicinal chemistry, pharmaceutical analysis, pharmaceutics, pharmacognosy, and herbal medicine.</p></fn><fn id="table2fn3"><p><sup>c</sup>Pharmacotherapy, pharmacy practice, pharmaceutical manufacturing, pharmaceutical quality control, and pharmacy administration and management.</p></fn><fn id="table2fn4"><p><sup>d</sup> Pharmaceutical Affairs Act, Narcotics Control Act, National Health Promotion Act, Framework Act on Health and Medical Services, National Health Insurance Act, and Regional Public Health Act.</p></fn></table-wrap-foot></table-wrap><p>As shown in <xref ref-type="table" rid="table3">Table 3</xref>, 323 out of 350 (92.3%) questions on the 2023 KPLE were answered by both GPT-3.5 and GPT-4, with 27 image-based questions that could not be processed. GPT-4 substantially exceeded GPT-3.5 in correct response rates across all question types, with statistically significant differences (<italic>P</italic>&#x003C;.001) in all categories except for Unit 4. The highest accuracy rates were in Biopharmacy (GPT-3.5 72/99, 72.7%; GPT-4 93/99, 93.9%), whereas the lowest rates were in Industrial Pharmacy for GPT-3.5 (31/66, 47%) and Medical Health Legislation for GPT-4 (14/20, 70%).</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Comparison of GPT-3.5&#x2019;s and GPT-4&#x2019;s performances on the 2023 Korean Pharmacist Licensing Examination.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Question category</td><td align="left" valign="bottom">All questions, n</td><td align="left" valign="bottom">Student correct response rate, n (%)</td><td align="left" valign="bottom">Questions answerable by GPT, n</td><td align="left" valign="bottom">GPT-3.5 correct response rate, n (%)</td><td align="left" valign="bottom">GPT-4 correct response rate, n (%)</td><td align="left" valign="bottom">Passing criteria, %</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Total</td><td align="left" valign="top">350</td><td align="left" valign="top">257.3 (73.5)</td><td align="left" valign="top">323</td><td align="left" valign="top">195 (60.4)</td><td align="left" valign="top">282 (87.3)</td><td align="left" valign="top">&#x2265;60</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Unit 1: Biopharmacy<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td><td align="left" valign="top">100</td><td align="left" valign="top">74.7 (74.7)</td><td align="left" valign="top">99</td><td align="left" valign="top">72 (72.7)</td><td align="left" valign="top">93 (93.9)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Unit 2: Industrial Pharmacy<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">90</td><td align="left" valign="top">63.8 (70.9)</td><td align="left" valign="top">66</td><td align="left" valign="top">31 (47)</td><td align="left" valign="top">53 (80.3)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Unit 3: Clinical and Practical Pharmacy I and II<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="top">140</td><td align="left" valign="top">103 (73.6)</td><td align="left" valign="top">138</td><td align="left" valign="top">81 (58.7)</td><td align="left" valign="top">122 (88.4)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Unit 4: Medical Health Legislation<sup><xref ref-type="table-fn" rid="table3fn4">d</xref></sup></td><td align="left" valign="top">20</td><td align="left" valign="top">15.7 (78.5)</td><td align="left" valign="top">20</td><td align="left" valign="top">11 (55)</td><td align="left" valign="top">14 (70)</td><td align="left" valign="top">&#x2265;40</td><td align="left" valign="top">.51</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>Biochemistry, molecular biology, microbiology, immunology, pharmacology, preventive pharmacy, and pathophysiology.</p></fn><fn id="table3fn2"><p><sup>b</sup>Physical pharmacy, synthetic chemistry, medicinal chemistry, pharmaceutical analysis, pharmaceutics, pharmacognosy, and herbal medicine.</p></fn><fn id="table3fn3"><p><sup>c</sup>Pharmacotherapy, pharmacy practice, pharmaceutical manufacturing, pharmaceutical quality control, and pharmacy administration and management.</p></fn><fn id="table3fn4"><p><sup>d</sup>Pharmaceutical Affairs Act, Narcotics Control Act, National Health Promotion Act, Framework Act on Health and Medical Services, National Health Insurance Act, and Regional Public Health Act.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-2"><title>GPT Models Versus Humans</title><p>KPLE participants totaled 1920, 1993, and 2014 in 2021, 2022, and 2023, respectively. For the 2021 examination, 59.7% (1147/1920) of the participants were female, with the largest age group being those in their 20s (1205/1920, 62.8%). This trend continued in 2022 and 2023, with women comprising 59% (1175/1993 in 2022; 1188/2014 in 2023) of the total participants in both years, and those in their 20s comprising 58.6% (1167/1993) and 59.2% (1193/2014), respectively. All participants were either graduates or expected graduates of a pharmacy school, as this level of education is required to qualify for the examination [<xref ref-type="bibr" rid="ref23">23</xref>-<xref ref-type="bibr" rid="ref26">26</xref>].</p><p>When comparing AI and human performance, pharmacy students correctly answered an average of 70.3%, 70.9%, and 73.5% of the same questions as those used in our study, in 2021, 2022, and 2023, respectively, with an average passing rate of 92.3%. GPT-3.5 showed lower accuracy rates than human participants, with scores of 62.5% (200/320), 59.3% (188/317), and 60.4% (195/323), respectively, whereas GPT-4 demonstrated higher accuracy rates of 85.9% (275/320), 86.1% (273/317), and 87.3% (282/323), respectively (<xref ref-type="fig" rid="figure1">Figure 1</xref>).</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Performance of GPT-3.5, GPT-4, and pharmacy students on the Korean Pharmacist Licensing Examination.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v10i1e57451_fig01.png"/></fig></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>This study evaluated the performance of both GPT-3.5 and GPT-4 on the KPLE over 3 recent years. GPT-4 consistently achieved scores above the passing level, whereas GPT-3.5 did not reach the 60% passing threshold in 1 of the 3 years, instead scoring very close to this criteria at 59.3% (188/317), indicating the limitations of GPT-3.5 in answering KPLE questions accurately. In contrast, GPT-4 exhibited significantly improved correct response rates across all units compared to GPT-3.5, surpassing the passing threshold with an overall accuracy of 86.5% (830/960). In evaluations of the performance of AI compared with that of human participants, pharmacy students&#x2019; accuracy rates were higher than those of GPT-3.5 but lower than those of GPT-4. This aligns with a recent study comparing medical students&#x2019; knowledge and interpretation-based responses with those of GPT-3.5, wherein GPT-3.5 achieved an accuracy rate of 60.8% (48/79). This was lower than the students&#x2019; overall performance, with an average accuracy rate of approximately 90.8% (71.8/79) [<xref ref-type="bibr" rid="ref27">27</xref>].</p><p>The findings of GPT-3.5&#x2019;s underperformance in comparison to GPT-4 are consistent with those of a previous study that investigated the performance of GPT-3 and GPT-4 in the North American Pharmacist Licensure Examination [<xref ref-type="bibr" rid="ref28">28</xref>] as well as in other health professional examinations [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref30">30</xref>]. In addition, similar studies have shown that GPT-3.5 failed to pass such examinations, such as the Taiwanese Pharmacist Licensing Examination [<xref ref-type="bibr" rid="ref31">31</xref>]. This could be attributed to GPT-3.5&#x2019;s training data cutoff in September 2021, rendering it outdated in terms of recent pharmacy practice advancements, research, and guideline updates. Moreover, disparity in performance might be attributed to differences in language and culture, as well as variations in the content of the examinations [<xref ref-type="bibr" rid="ref32">32</xref>]. Both GPT models have acquired substantial information on health care policies in English-speaking nations because of their comprehensive English datasets. Similarly, previous studies have reported higher accuracy rates for questions in English when compared to other languages [<xref ref-type="bibr" rid="ref31">31</xref>,<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref34">34</xref>]. Consequently, these findings imply that GPT models might benefit from supplementary training data in languages besides English, to expand their knowledge and enhance their performance on language-specific pharmacy examinations. These findings suggest that LLMs possess great capabilities in addressing KPLE questions, and with further refinement, these models can be reasonably expected to provide even higher levels of accuracy. This aligns with recent research in medical natural language processing, which has increasingly emphasized the importance of domain- and language-specific modeling [<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref36">36</xref>]. Models specialized in specific languages, such as Korean, and the medical domain have shown superior performance in tasks like processing medical documents compared to general models, highlighting the need for the continued development of AI models tailored to specific domains and languages [<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref36">36</xref>].</p><p>Notably, among the 4 units, the accuracy rate was highest for Biopharmacy questions among the specific question datasets used in our study, followed by Clinical and Practical Pharmacy. This suggests that GPT-4 could potentially support pharmacists within clinical settings. However, this may not accurately reflect the complexity of real-world clinical settings, which involve a variety of patient cases and collaborations with other health care professionals. Additionally, the KPLE does not involve communication in the pharmacist-patient relationship, interpersonal skills, or empathy. Brin et al [<xref ref-type="bibr" rid="ref37">37</xref>] found that GPT-4 exhibited superior performance in soft skills such as empathy, ethics, and judgment. This finding indicates AI&#x2019;s potential to address complex ethical challenges, show empathy, and effectively support patient care and family interactions. However, that study had a limited question pool of only 80 multiple-choice questions drawn from 2 sources, potentially leading to selection bias. Consequently, the questions might not accurately reflect the full scope of the actual USMLE content or cover all the soft skills that are vital in medical practice. Furthermore, ChatGPT models scored the lowest in Medical Health Legislation, which may be attributable to the differences in policies and laws between Korea and the United States, as the AI is likely more familiar with legal frameworks in English-speaking countries.</p><p>Additionally, GPT models are susceptible to a phenomenon termed &#x201C;hallucination,&#x201D; where they generate scientifically inaccurate information that seems plausible to individuals lacking expertise [<xref ref-type="bibr" rid="ref34">34</xref>]. For example, in our study, GPT-4 generated an incorrect drug interaction between 2 medications that are not actually known to interact. In another example, GPT-3.5 suggested an incorrect dosage for a common medication, which could lead to potential harm if used in a clinical setting. Moreover, GPT-3.5 showed lower concordance and higher self-contradiction compared to GPT-4. These examples highlight the importance of critically evaluating AI-generated information, especially in the context of clinical environment. Depending solely on generated content carries risks; therefore, those receiving the output should have the professional pharmaceutical knowledge necessary to assess its accuracy. These limitations must be recognized and addressed to comprehensively evaluate GPT-4&#x2019;s practical applicability in classrooms and clinical practice.</p><p>Previous studies suggested that GPT models have the potential to become useful tools in the field of medical education because of their ability to generate appropriate and precise information in response to well-defined inputs [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref38">38</xref>,<xref ref-type="bibr" rid="ref39">39</xref>]. Due to their ease of access and rapid information generation capabilities, AI-based LLMs, like ChatGPT, are poised to serve as valuable educational aids. Our evaluation findings align with these studies and indicate that GPT models could facilitate a &#x201C;self-directed&#x201D; learning approach, helping students enhance their knowledge and reasoning skills. However, it is essential to validate the information provided by GPT models because they lack standard references for the retrieved information. Hence, with careful implementation, clear protocols, and oversight by health care professionals, AI-driven chatbots show significant potential to transform clinical pharmacology and drug information services. Specific guidelines, such as ensuring chatbot use is supervised by licensed professionals and continuously updated with the latest medical knowledge, are crucial for their effective and safe application.</p></sec><sec id="s4-2"><title>Limitations</title><p>Our study has some limitations. First, it is essential to recognize that the GPT models employed in this study may not reflect the latest models available. Therefore, caution should be exercised when implementing these findings in practical clinical settings, since depending on up-to-date references is essential. Specifically, the findings represent the GPT models&#x2019; capabilities up to December 18, 2023, and variations in results may occur in the future, given the expected rapid improvement in the capabilities of ChatGPT versions through user feedback and deep learning. Additional updates are expected in the future, and it is crucial to consistently assess them. Second, image-based test questions were excluded because the AI models could not support them. Moreover, the AI chatbots were unable to interpret the information presented in tables; therefore, we manually entered this information. Third, the findings of this study are specific to the datasets and conditions of the KPLE. However, the adjustments noted above were randomly distributed across various subjects and not biased toward any specific topics. Furthermore, as both GPT-3.5 and GPT-4 were evaluated under the same conditions, these exclusions likely had minimal impact on the final performance comparison. However, the generalizability of these results to other professional environments or licensing examinations may still be limited. Further research is needed to assess the performance of AI chatbots in different contexts and with various datasets, as it remains possible that correct answers may not be obtained under different conditions. Finally, it should be noted that this investigation focused exclusively on GPT-3.5 and GPT-4. In the future, it is worth considering the possibility of implementing other LLMs, such as New Bing and Bard, in the pharmaceutical field.</p></sec><sec id="s4-3"><title>Conclusions</title><p>Our study demonstrates the potential of ChatGPT (ie, GPT-3.5 and GPT-4) to assist in pharmaceutical knowledge comprehension within the Korean context. GPT-4 exhibited expert-level performance and consistently passed the KPLE, while GPT-3.5 fell short of the passing criteria in 1 instance. Pharmacy students outperformed GPT-3.5 but scored lower than GPT-4. Although GPT-4 outperformed students, educators and students should not rely solely on chatbots for learning, as AI tools may produce misleading or inaccurate information. Therefore, it is imperative to conduct thorough testing and validation before successfully implementing AI and to examine the feasibility of all GPT versions in real-world clinical contexts. Future research should incorporate more extensive and diverse question sets and ethical scenarios to provide a more accurate representation of pharmacy practice.</p></sec></sec></body><back><ack><p>This research was supported by the Basic Science Research Program through the National Research Foundation of Korea funded by the Ministry of Education (grant number 2021R1A6A1A03044296).</p></ack><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">KHPLEI</term><def><p>Korea Health Personnel Licensing Examination Institute</p></def></def-item><def-item><term id="abb3">KPLE</term><def><p>Korean Pharmacist Licensing Examination</p></def></def-item><def-item><term id="abb4">LLM</term><def><p>large language model</p></def></def-item><def-item><term id="abb5">USMLE</term><def><p>United States Medical Licensing Examination</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Baidoo-anu</surname><given-names>D</given-names> </name><name name-style="western"><surname>Owusu Ansah</surname><given-names>L</given-names> </name></person-group><article-title>Education in the era of generative artificial intelligence (AI): understanding the potential benefits of ChatGPT in promoting teaching and learning</article-title><source>J AI</source><year>2023</year><month>12</month><day>31</day><volume>7</volume><issue>1</issue><fpage>52</fpage><lpage>62</lpage><pub-id pub-id-type="doi">10.61969/jai.1337500</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fui-Hoon Nah</surname><given-names>F</given-names> </name><name name-style="western"><surname>Zheng</surname><given-names>R</given-names> </name><name name-style="western"><surname>Cai</surname><given-names>J</given-names> </name><name name-style="western"><surname>Siau</surname><given-names>K</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>L</given-names> </name></person-group><article-title>Generative AI and ChatGPT: applications, challenges, and AI-human collaboration</article-title><source>J Inf Technol Case Appl Res</source><year>2023</year><month>07</month><day>21</day><volume>25</volume><issue>3</issue><fpage>277</fpage><lpage>304</lpage><pub-id pub-id-type="doi">10.1080/15228053.2023.2233814</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Miller</surname><given-names>DD</given-names> </name><name name-style="western"><surname>Brown</surname><given-names>EW</given-names> </name></person-group><article-title>Artificial intelligence in medical practice: the question to the answer?</article-title><source>Am J Med</source><year>2018</year><month>02</month><volume>131</volume><issue>2</issue><fpage>129</fpage><lpage>133</lpage><pub-id pub-id-type="doi">10.1016/j.amjmed.2017.10.035</pub-id><pub-id pub-id-type="medline">29126825</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Heacock</surname><given-names>L</given-names> </name><name name-style="western"><surname>Elias</surname><given-names>J</given-names> </name><etal/></person-group><article-title>ChatGPT and other large language models are double-edged swords</article-title><source>Radiology</source><year>2023</year><month>04</month><volume>307</volume><issue>2</issue><fpage>e230163</fpage><pub-id pub-id-type="doi">10.1148/radiol.230163</pub-id><pub-id pub-id-type="medline">36700838</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Bommarito</surname><given-names>J</given-names> </name><name name-style="western"><surname>Bommarito</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Katz</surname><given-names>JAM</given-names> </name><name name-style="western"><surname>Katz</surname><given-names>DM</given-names> </name></person-group><article-title>GPT as knowledge worker: a zero-shot evaluation of (AI)CPA capabilities</article-title><source>SSRN J</source><year>2023</year><month>01</month><day>11</day><pub-id pub-id-type="doi">10.2139/ssrn.4322372</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="web"><person-group person-group-type="author"><name name-style="western"><surname>Terwiesch</surname><given-names>C</given-names> </name></person-group><article-title>Would ChatGPT get a Wharton MBA?</article-title><source>Mack Institute for Innovation Management</source><year>2023</year><month>01</month><day>17</day><access-date>2024-01-10</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://mackinstitute.wharton.upenn.edu/2023/would-chat-gpt3-get-a-wharton-mba-new-white-paper-by-christian-terwiesch/">https://mackinstitute.wharton.upenn.edu/2023/would-chat-gpt3-get-a-wharton-mba-new-white-paper-by-christian-terwiesch/</ext-link></comment></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Choi</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Hickman</surname><given-names>KE</given-names> </name><name name-style="western"><surname>Monahan</surname><given-names>A</given-names> </name><name name-style="western"><surname>Schwarcz</surname><given-names>DB</given-names> </name></person-group><article-title>ChatGPT goes to law school</article-title><source>J Legal Educ</source><year>2022</year><volume>71</volume><issue>3</issue><fpage>387</fpage><lpage>400</lpage><pub-id pub-id-type="doi">10.2139/ssrn.4335905</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rao</surname><given-names>A</given-names> </name><name name-style="western"><surname>Pang</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Assessing the utility of ChatGPT throughout the entire clinical workflow: development and usability study</article-title><source>J Med Internet Res</source><year>2023</year><month>08</month><day>22</day><volume>25</volume><fpage>e48659</fpage><pub-id pub-id-type="doi">10.2196/48659</pub-id><pub-id pub-id-type="medline">37606976</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fraser</surname><given-names>H</given-names> </name><name name-style="western"><surname>Crossland</surname><given-names>D</given-names> </name><name name-style="western"><surname>Bacher</surname><given-names>I</given-names> </name><name name-style="western"><surname>Ranney</surname><given-names>M</given-names> </name><name name-style="western"><surname>Madsen</surname><given-names>T</given-names> </name><name name-style="western"><surname>Hilliard</surname><given-names>R</given-names> </name></person-group><article-title>Comparison of diagnostic and triage accuracy of Ada Health and WebMD symptom checkers, ChatGPT, and physicians for patients in an emergency department: clinical data analysis study</article-title><source>JMIR Mhealth Uhealth</source><year>2023</year><month>10</month><day>3</day><volume>11</volume><fpage>e49995</fpage><pub-id pub-id-type="doi">10.2196/49995</pub-id><pub-id pub-id-type="medline">37788063</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jeblick</surname><given-names>K</given-names> </name><name name-style="western"><surname>Schachtner</surname><given-names>B</given-names> </name><name name-style="western"><surname>Dexl</surname><given-names>J</given-names> </name><etal/></person-group><article-title>ChatGPT makes medicine easy to swallow: an exploratory case study on simplified radiology reports</article-title><source>Eur Radiol</source><year>2024</year><month>05</month><volume>34</volume><issue>5</issue><fpage>2817</fpage><lpage>2825</lpage><pub-id pub-id-type="doi">10.1007/s00330-023-10213-1</pub-id><pub-id pub-id-type="medline">37794249</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dave</surname><given-names>T</given-names> </name><name name-style="western"><surname>Athaluri</surname><given-names>SA</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>S</given-names> </name></person-group><article-title>ChatGPT in medicine: an overview of its applications, advantages, limitations, future prospects, and ethical considerations</article-title><source>Front Artif Intell</source><year>2023</year><month>05</month><day>4</day><volume>6</volume><fpage>1169595</fpage><pub-id pub-id-type="doi">10.3389/frai.2023.1169595</pub-id><pub-id pub-id-type="medline">37215063</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kung</surname><given-names>TH</given-names> </name><name name-style="western"><surname>Cheatham</surname><given-names>M</given-names> </name><name name-style="western"><surname>Medenilla</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models</article-title><source>PLOS Dig Health</source><year>2023</year><month>02</month><day>9</day><volume>2</volume><issue>2</issue><fpage>e0000198</fpage><pub-id pub-id-type="doi">10.1371/journal.pdig.0000198</pub-id><pub-id pub-id-type="medline">36812645</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Alfertshofer</surname><given-names>M</given-names> </name><name name-style="western"><surname>Hoch</surname><given-names>CC</given-names> </name><name name-style="western"><surname>Funk</surname><given-names>PF</given-names> </name><etal/></person-group><article-title>Sailing the Seven Seas: a multinational comparison of ChatGPT&#x2019;s performance on medical licensing examinations</article-title><source>Ann Biomed Eng</source><year>2024</year><month>06</month><volume>52</volume><issue>6</issue><fpage>1542</fpage><lpage>1545</lpage><pub-id pub-id-type="doi">10.1007/s10439-023-03338-3</pub-id><pub-id pub-id-type="medline">37553555</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gilson</surname><given-names>A</given-names> </name><name name-style="western"><surname>Safranek</surname><given-names>CW</given-names> </name><name name-style="western"><surname>Huang</surname><given-names>T</given-names> </name><etal/></person-group><article-title>How does ChatGPT perform on the United States Medical Licensing Examination (USMLE)? The implications of large language models for medical education and knowledge assessment</article-title><source>JMIR Med Educ</source><year>2023</year><month>02</month><day>8</day><volume>9</volume><fpage>e45312</fpage><pub-id pub-id-type="doi">10.2196/45312</pub-id><pub-id pub-id-type="medline">36753318</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>P</given-names> </name><name name-style="western"><surname>Bubeck</surname><given-names>S</given-names> </name><name name-style="western"><surname>Petro</surname><given-names>J</given-names> </name></person-group><article-title>Benefits, limits, and risks of GPT-4 as an AI chatbot for medicine</article-title><source>N Engl J Med</source><year>2023</year><month>03</month><day>29</day><volume>388</volume><issue>13</issue><fpage>1233</fpage><lpage>1239</lpage><pub-id pub-id-type="doi">10.1056/NEJMsr2214184</pub-id><pub-id pub-id-type="medline">36988602</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Taira</surname><given-names>K</given-names> </name><name name-style="western"><surname>Itaya</surname><given-names>T</given-names> </name><name name-style="western"><surname>Hanada</surname><given-names>A</given-names> </name></person-group><article-title>Performance of the large language model ChatGPT on the National Nurse Examinations in Japan: evaluation study</article-title><source>JMIR Nurs</source><year>2023</year><month>06</month><day>27</day><volume>6</volume><fpage>e47305</fpage><pub-id pub-id-type="doi">10.2196/47305</pub-id><pub-id pub-id-type="medline">37368470</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huang</surname><given-names>H</given-names> </name></person-group><article-title>Performance of ChatGPT on registered nurse license exam in Taiwan: a descriptive study</article-title><source>Healthcare (Basel)</source><year>2023</year><month>10</month><day>30</day><volume>11</volume><issue>21</issue><fpage>2855</fpage><pub-id pub-id-type="doi">10.3390/healthcare11212855</pub-id><pub-id pub-id-type="medline">37958000</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fuchs</surname><given-names>A</given-names> </name><name name-style="western"><surname>Trachsel</surname><given-names>T</given-names> </name><name name-style="western"><surname>Weiger</surname><given-names>R</given-names> </name><name name-style="western"><surname>Eggmann</surname><given-names>F</given-names> </name></person-group><article-title>ChatGPT&#x2019;s performance in dentistry and allergy-immunology assessments: a comparative study</article-title><source>Swiss Dent J</source><year>2023</year><month>10</month><day>6</day><volume>134</volume><issue>5</issue><pub-id pub-id-type="medline">37799027</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Yang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Yao</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Tasmin</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Performance of multimodal GPT-4V on USMLE with image: potential for imaging diagnostic support with explanations</article-title><source>medRxiv</source><comment>Preprint posted online on  Nov 15, 2023</comment><pub-id pub-id-type="doi">10.1101/2023.10.26.23297629</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yanagita</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Yokokawa</surname><given-names>D</given-names> </name><name name-style="western"><surname>Uchida</surname><given-names>S</given-names> </name><name name-style="western"><surname>Tawara</surname><given-names>J</given-names> </name><name name-style="western"><surname>Ikusaka</surname><given-names>M</given-names> </name></person-group><article-title>Accuracy of ChatGPT on medical questions in the National Medical Licensing Examination in Japan: evaluation study</article-title><source>JMIR Form Res</source><year>2023</year><month>10</month><day>13</day><volume>7</volume><fpage>e48023</fpage><pub-id pub-id-type="doi">10.2196/48023</pub-id><pub-id pub-id-type="medline">37831496</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Nisar</surname><given-names>S</given-names> </name><name name-style="western"><surname>Aslam</surname><given-names>MS</given-names> </name></person-group><article-title>Is ChatGPT a good tool for T&#x0026;CM students in studying pharmacology?</article-title><source>SSRN</source><comment>Preprint posted online on  Jan 17, 2023</comment><pub-id pub-id-type="doi">10.2139/ssrn.4324310</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="web"><article-title>Official released questions from the Korean Pharmacist Licensing Examination. &#xAD6D;&#xAC00;&#xC2DC;&#xD5D8; &#xAE30;&#xCD9C;&#xBB38;&#xC81C;</article-title><source>Korea Health Personnel Licensing Examination Institute. &#xD55C;&#xAD6D;&#xBCF4;&#xAC74;&#xC758;&#xB8CC;&#xC778;&#xAD6D;&#xAC00;&#xC2DC;&#xD5D8;&#xC6D0;</source><access-date>2024-01-12</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.kuksiwon.or.kr/CollectOfQuestions/brd/m_116/list.do">https://www.kuksiwon.or.kr/CollectOfQuestions/brd/m_116/list.do</ext-link></comment></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="web"><article-title>Statistics of the 72nd Korean Pharmacist Licensing Examination in 2021. 2021&#xB144;&#xB3C4; &#xC81C;72&#xD68C; &#xC57D;&#xC0AC; &#xAD6D;&#xAC00;&#xC2DC;&#xD5D8; &#xBD84;&#xC11D;&#xACB0;&#xACFC;</article-title><source>Korea Health Personnel Licensing Examination Institute &#xD55C;&#xAD6D;&#xBCF4;&#xAC74;&#xC758;&#xB8CC;&#xC778;&#xAD6D;&#xAC00;&#xC2DC;&#xD5D8;&#xC6D0;</source><access-date>2024-01-12</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.kuksiwon.or.kr/analysis/brd/m_91/view.do?seq=254&#x0026;srchFr=&#x0026;srchTo=&#x0026;srchWord=%C2%A0%EC%95%BD%EC%82%AC%C2%A0%EA%B5%AD%EA%B0%80%EC%8B%9C%ED%97%98&#x0026;srchTp=0&#x0026;itm_seq_1=0&#x0026;itm_seq_2=0&#x0026;multi_itm_seq=0&#x0026;company_cd=&#x0026;company_nm=&#x0026;etc1=">https://www.kuksiwon.or.kr/analysis/brd/m_91/view.do?seq=254&#x0026;srchFr=&#x0026;srchTo=&#x0026;srchWord=%C2%A0%EC%95%BD%EC%82%AC%C2%A0%EA%B5%AD%EA%B0%80%EC%8B%9C%ED%97%98&#x0026;srchTp=0&#x0026;itm_seq_1=0&#x0026;itm_seq_2=0&#x0026;multi_itm_seq=0&#x0026;company_cd=&#x0026;company_nm=&#x0026;etc1=</ext-link></comment></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="web"><article-title>Statistics of the 73rd Korean Pharmacist Licensing Examination in 2022. 2022&#xB144;&#xB3C4; &#xC81C;73&#xD68C; &#xC57D;&#xC0AC; &#xAD6D;&#xAC00;&#xC2DC;&#xD5D8; &#xBD84;&#xC11D;&#xACB0;&#xACFC;</article-title><source>Korea Health Personnel Licensing Examination Institute. &#xD55C;&#xAD6D;&#xBCF4;&#xAC74;&#xC758;&#xB8CC;&#xC778;&#xAD6D;&#xAC00;&#xC2DC;&#xD5D8;&#xC6D0;</source><access-date>2024-01-12</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.kuksiwon.or.kr/analysis/brd/m_91/view.do?seq=292&#x0026;srchFr=&#x0026;srchTo=&#x0026;srchWord=%C2%A0%EC%95%BD%EC%82%AC%C2%A0%EA%B5%AD%EA%B0%80%EC%8B%9C%ED%97%98&#x0026;srchTp=0&#x0026;itm_seq_1=0&#x0026;itm_seq_2=0&#x0026;multi_itm_seq=0&#x0026;company_cd=&#x0026;company_nm=&#x0026;etc1=">https://www.kuksiwon.or.kr/analysis/brd/m_91/view.do?seq=292&#x0026;srchFr=&#x0026;srchTo=&#x0026;srchWord=%C2%A0%EC%95%BD%EC%82%AC%C2%A0%EA%B5%AD%EA%B0%80%EC%8B%9C%ED%97%98&#x0026;srchTp=0&#x0026;itm_seq_1=0&#x0026;itm_seq_2=0&#x0026;multi_itm_seq=0&#x0026;company_cd=&#x0026;company_nm=&#x0026;etc1=</ext-link></comment></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="web"><article-title>Statistics of the 74th Korean Pharmacist Licensing Examination in 2023. 2023&#xB144;&#xB3C4; &#xC81C;74&#xD68C; &#xC57D;&#xC0AC; &#xAD6D;&#xAC00;&#xC2DC;&#xD5D8; &#xBD84;&#xC11D;&#xACB0;&#xACFC;</article-title><source>Korea Health Personnel Licensing Examination Institute. &#xD55C;&#xAD6D;&#xBCF4;&#xAC74;&#xC758;&#xB8CC;&#xC778;&#xAD6D;&#xAC00;&#xC2DC;&#xD5D8;&#xC6D0;</source><access-date>2024-01-12</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.kuksiwon.or.kr/analysis/brd/m_91/view.do?seq=333&#x0026;srchFr=&#x0026;srchTo=&#x0026;srchWord=%C2%A0%EC%95%BD%EC%82%AC%C2%A0%EA%B5%AD%EA%B0%80%EC%8B%9C%ED%97%98&#x0026;srchTp=0&#x0026;itm_seq_1=0&#x0026;itm_seq_2=0&#x0026;multi_itm_seq=0&#x0026;company_cd=&#x0026;company_nm=&#x0026;etc1=">https://www.kuksiwon.or.kr/analysis/brd/m_91/view.do?seq=333&#x0026;srchFr=&#x0026;srchTo=&#x0026;srchWord=%C2%A0%EC%95%BD%EC%82%AC%C2%A0%EA%B5%AD%EA%B0%80%EC%8B%9C%ED%97%98&#x0026;srchTp=0&#x0026;itm_seq_1=0&#x0026;itm_seq_2=0&#x0026;multi_itm_seq=0&#x0026;company_cd=&#x0026;company_nm=&#x0026;etc1=</ext-link></comment></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="web"><article-title>Korea Health Personnel Licensing Examination Institute: statistics of candidates for the KPLE. &#xD55C;&#xAD6D;&#xBCF4;&#xAC74;&#xC758;&#xB8CC;&#xC778;&#xAD6D;&#xAC00;&#xC2DC;&#xD5D8;&#xC6D0;: &#xC57D;&#xC0AC;&#xAD6D;&#xAC00;&#xACE0;&#xC2DC; &#xC751;&#xC2DC;&#xC790; &#xD604;&#xD669;</article-title><source>Data.go.kr. &#xACF5;&#xACF5;&#xB370;&#xC774;&#xD130;&#xD3EC;&#xD138;</source><access-date>2024-01-12</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.data.go.kr/tcs/dss/selectDataSetList.do?keyword=%EC%95%BD%EC%82%AC+%EA%B5%AD%EA%B0%80%EA%B3%A0%EC%8B%9C&#x0026;brm=&#x0026;svcType=&#x0026;recmSe=N&#x0026;conditionType=init&#x0026;extsn=&#x0026;kwrdArray=">https://www.data.go.kr/tcs/dss/selectDataSetList.do?keyword=%EC%95%BD%EC%82%AC+%EA%B5%AD%EA%B0%80%EA%B3%A0%EC%8B%9C&#x0026;brm=&#x0026;svcType=&#x0026;recmSe=N&#x0026;conditionType=init&#x0026;extsn=&#x0026;kwrdArray=</ext-link></comment></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Huh</surname><given-names>S</given-names> </name></person-group><article-title>Are ChatGPT&#x2019;s knowledge and interpretation ability comparable to those of medical students in Korea for taking a parasitology examination?: a descriptive study</article-title><source>J Educ Eval Health Prof</source><year>2023</year><volume>20</volume><fpage>1</fpage><pub-id pub-id-type="doi">10.3352/jeehp.2023.20.1</pub-id><pub-id pub-id-type="medline">36627845</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Angel</surname><given-names>M</given-names> </name><name name-style="western"><surname>Patel</surname><given-names>A</given-names> </name><name name-style="western"><surname>Alachkar</surname><given-names>A</given-names> </name><name name-style="western"><surname>Baldi</surname><given-names>P</given-names> </name></person-group><article-title>Clinical knowledge and reasoning abilities of AI large language models in pharmacy: a comparative study on the NAPLEX exam</article-title><source>bioRxiv</source><comment>Preprint posted online on  Jun 12, 2023</comment><pub-id pub-id-type="doi">10.1101/2023.06.07.544055</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Flores-Cohaila</surname><given-names>JA</given-names> </name><name name-style="western"><surname>Garc&#x00ED;a-Vicente</surname><given-names>A</given-names> </name><name name-style="western"><surname>Vizcarra-Jim&#x00E9;nez</surname><given-names>SF</given-names> </name><etal/></person-group><article-title>Performance of ChatGPT on the Peruvian National Licensing Medical Examination: cross-sectional study</article-title><source>JMIR Med Educ</source><year>2023</year><month>09</month><day>28</day><volume>9</volume><fpage>e48039</fpage><pub-id pub-id-type="doi">10.2196/48039</pub-id><pub-id pub-id-type="medline">37768724</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Knoedler</surname><given-names>L</given-names> </name><name name-style="western"><surname>Alfertshofer</surname><given-names>M</given-names> </name><name name-style="western"><surname>Knoedler</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Pure wisdom or Potemkin villages? A comparison of ChatGPT 3.5 and ChatGPT 4 on USMLE Step 3 style questions: quantitative analysis</article-title><source>JMIR Med Educ</source><year>2024</year><month>01</month><day>5</day><volume>10</volume><fpage>e51148</fpage><pub-id pub-id-type="doi">10.2196/51148</pub-id><pub-id pub-id-type="medline">38180782</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>YM</given-names> </name><name name-style="western"><surname>Shen</surname><given-names>HW</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>TJ</given-names> </name></person-group><article-title>Performance of ChatGPT on the pharmacist licensing examination in Taiwan</article-title><source>J Chin Med Assoc</source><year>2023</year><month>07</month><day>1</day><volume>86</volume><issue>7</issue><fpage>653</fpage><lpage>658</lpage><pub-id pub-id-type="doi">10.1097/JCMA.0000000000000942</pub-id><pub-id pub-id-type="medline">37227901</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Seghier</surname><given-names>ML</given-names> </name></person-group><article-title>ChatGPT: not all languages are equal</article-title><source>Nature</source><year>2023</year><month>03</month><volume>615</volume><issue>7951</issue><fpage>216</fpage><pub-id pub-id-type="doi">10.1038/d41586-023-00680-3</pub-id><pub-id pub-id-type="medline">36882613</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tong</surname><given-names>W</given-names> </name><name name-style="western"><surname>Guan</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Artificial intelligence in global health equity: an evaluation and discussion on the application of ChatGPT, in the Chinese National Medical Licensing Examination</article-title><source>Front Med (Lausanne)</source><year>2023</year><month>10</month><day>19</day><volume>10</volume><fpage>1237432</fpage><pub-id pub-id-type="doi">10.3389/fmed.2023.1237432</pub-id><pub-id pub-id-type="medline">38020160</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><collab>OpenAI</collab><name name-style="western"><surname>Achiam</surname><given-names>J</given-names> </name><name name-style="western"><surname>Adler</surname><given-names>S</given-names> </name><etal/></person-group><article-title>GPT-4 technical report</article-title><source>arXiv</source><comment>Preprint posted online on  Mar 15, 2023</comment><pub-id pub-id-type="doi">10.48550/arXiv.2303.08774</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>JM</given-names> </name><etal/></person-group><article-title>A pre-trained BERT for Korean medical natural language processing</article-title><source>Sci Rep</source><year>2022</year><month>08</month><day>16</day><volume>12</volume><issue>1</issue><fpage>13847</fpage><pub-id pub-id-type="doi">10.1038/s41598-022-17806-8</pub-id><pub-id pub-id-type="medline">35974113</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>YM</given-names> </name><name name-style="western"><surname>Song</surname><given-names>S</given-names> </name><name name-style="western"><surname>Joo</surname><given-names>HJ</given-names> </name></person-group><article-title>Predicting medical specialty from text based on a domain-specific pre-trained BERT</article-title><source>Int J Med Inform</source><year>2023</year><month>02</month><volume>170</volume><fpage>104956</fpage><pub-id pub-id-type="doi">10.1016/j.ijmedinf.2022.104956</pub-id><pub-id pub-id-type="medline">36512987</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Brin</surname><given-names>D</given-names> </name><name name-style="western"><surname>Sorin</surname><given-names>V</given-names> </name><name name-style="western"><surname>Vaid</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Comparing ChatGPT and GPT-4 performance in USMLE soft skill assessments</article-title><source>Sci Rep</source><year>2023</year><month>10</month><day>1</day><volume>13</volume><issue>1</issue><fpage>16492</fpage><pub-id pub-id-type="doi">10.1038/s41598-023-43436-9</pub-id><pub-id pub-id-type="medline">37779171</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>O&#x2019;Connor</surname><given-names>S</given-names> </name></person-group><article-title>Open artificial intelligence platforms in nursing education: tools for academic progress or abuse?</article-title><source>Nurse Educ Pract</source><year>2023</year><month>01</month><volume>66</volume><fpage>103537</fpage><pub-id pub-id-type="doi">10.1016/j.nepr.2022.103537</pub-id><pub-id pub-id-type="medline">36549229</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Abd-Alrazaq</surname><given-names>A</given-names> </name><name name-style="western"><surname>AlSaad</surname><given-names>R</given-names> </name><name name-style="western"><surname>Alhuwail</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Large language models in medical education: opportunities, challenges, and future directions</article-title><source>JMIR Med Educ</source><year>2023</year><month>06</month><day>1</day><volume>9</volume><fpage>e48291</fpage><pub-id pub-id-type="doi">10.2196/48291</pub-id><pub-id pub-id-type="medline">37261894</pub-id></nlm-citation></ref></ref-list></back></article>