Abstract
Background: With the increasing application of large language models like ChatGPT in various industries, its potential in the medical domain, especially in standardized examinations, has become a focal point of research.
Objective: The aim of this study is to assess the clinical performance of ChatGPT, focusing on its accuracy and reliability in the Chinese National Medical Licensing Examination (CNMLE).
Methods: The CNMLE 2022 question set, consisting of 500 single-answer multiple choices questions, were reclassified into 15 medical subspecialties. Each question was tested 8 to 12 times in Chinese on the OpenAI platform from April 24 to May 15, 2023. Three key factors were considered: the version of GPT-3.5 and 4.0, the prompt’s designation of system roles tailored to medical subspecialties, and repetition for coherence. A passing accuracy threshold was established as 60%. The χ2 tests and κ values were employed to evaluate the model’s accuracy and consistency.
Results: GPT-4.0 achieved a passing accuracy of 72.7%, which was significantly higher than that of GPT-3.5 (54%; P<.001). The variability rate of repeated responses from GPT-4.0 was lower than that of GPT-3.5 (9% vs 19.5%; P<.001). However, both models showed relatively good response coherence, with κ values of 0.778 and 0.610, respectively. System roles numerically increased accuracy for both GPT-4.0 (0.3%‐3.7%) and GPT-3.5 (1.3%‐4.5%), and reduced variability by 1.7% and 1.8%, respectively (P>.05). In subgroup analysis, ChatGPT achieved comparable accuracy among different question types (P>.05). GPT-4.0 surpassed the accuracy threshold in 14 of 15 subspecialties, while GPT-3.5 did so in 7 of 15 on the first response.
Conclusions: GPT-4.0 passed the CNMLE and outperformed GPT-3.5 in key areas such as accuracy, consistency, and medical subspecialty expertise. Adding a system role insignificantly enhanced the model’s reliability and answer coherence. GPT-4.0 showed promising potential in medical education and clinical practice, meriting further study.
doi:10.2196/52784
Keywords
Introduction
ChatGPT, a general large language model (LLM) developed by OpenAI, has gained substantial attention since its launch on November 30, 2022. Known for its advanced natural language processing capabilities, ChatGPT has the potential to make significant impacts on many industries, including medical education. Its performance in medicine was first tested at or near the passing threshold of the United States Medical Licensing Examination (USMLE) [
, ]. While ChatGPT’s accuracy varies across languages [ ], it has been tested on a series of medical exams like the Japanese National Medical Licensing Examination in languages including English [ ], Chinese [ ], Dutch [ ], Japanese [ ], and Korean [ ]. The research scope related to ChatGPT has expanded to medical education in fields like nuclear medicine [ ], neurosurgery [ ], ophthalmology [ ], general chemistry, nursing[ ], life support [ ], dentology [ ], and radiation oncology physics [ ]. Overall, while ChatGPT demonstrates heterogeneous capabilities, it shows promising potential in these medical specialties.Several factors might influence ChatGPT’s performance. First, the updated version of ChatGPT, GPT-4, understands and generates natural language in more complex and nuanced scenarios, leading to more accurate responses [
], which is important in analyzing complex clinical case questions [ ]. Thus, GPT-4 conclusively demonstrated significantly better performance than GPT-3.5, as evidenced by various official medical exams [ ]. Besides the model version, ChatGPT allows users to guide its behavior by adding prompts that describe its system role. These system roles influence the direction of ChatGPT’s answers and may affect its reliability. However, the impact of these system roles on ChatGPT’s performance in medical field has not yet been investigated. As a professional chatbot tool, ChatGPT uses sampling to predict the next token with varying distribution probabilities, ensuring responses are varied and natural in real-world applications. Zhu et al [ ] have found that composite answers derived from repeated questioning can enhance the accuracy of ChatGPT. Typically, 2 or 3 repeated responses are necessary to ensure response stability [ - ].Currently, the peer-reviewed research still lacks highlights on the strength of ChatGPT when it comes to the Chinese National Medical Licensing Examination (CNMLE). This study aimed to evaluate the performance of ChatGPT in answering CNMLE questions in the clinical setting of China, with consideration of the version of ChatGPT and system role.
Methods
The CNMLE 2022 Question Data
As an industry admission examination, passing the CNMLE means that a medical practitioner meets the minimum medical competencies. The written part of the examination, which emphasizes medical knowledge and clinical decision-making skills, is created and supervised by the Chinese National Medical Examination Center (NMEC). In 2021, the CNMLE transitioned from the traditional paper-based format to a computer-based examination. Each candidate is presented with 600 questions, arranged in a slightly varied order, from the exam year’s question data set. According to OpenAI’s introduction, ChatGPT’s responses are based on information available up to September 2021. Thus, we selected the CNMLE 2022 questions, which were purchased from a web-based bookstore [
], for our evaluation. This choice ensured that the questions had not been previously encountered and trained by the model. The publisher has confirmed that these released questions are the original ones from the examination.The CNMLE 2022 covered 600 single-answer multiple-choice questions, which were evenly divided into 4 units [
]. Each unit had 4 specific question types: A1, the single-sentence optimal choice questions; A2, case summary optimal choice questions; A3/A4, case group optimal choice questions; and B1, standard combination questions. Detailed explanations of each question type was conveyed to ChatGPT via a structured prompt prior to inquiry (see in ). The CNMLE 2022 questions did not involve table or image-based questions. Therefore, ChatGPT, despite lacking multimodal capabilities, was still suited to effectively complete the test.According to the introduction of the Chinese NMEC [
], each examination unit always addresses specific medical subspecialties. Unit 1 covers medical knowledge, policies, regulations, and preventive medicine; unit 2 mainly pertains to the cardiovascular, urinary, musculoskeletal, and endocrine systems; unit 3 involves the digestive, respiratory, and associated systems; unit 4 focuses on obstetrics and gynecology, pediatrics, and neurological or psychiatric domains. However, such distribution is not absolute. Therefore, 2 clinicians independently reclassified the 600 questions into 15 medical subspecialties, resolving discrepancies through discussion. The κ value for the result of their classifications was 0.935. The Sankey diagram of the 3 question classifications, medical subspecialties, units, and types is shown in .Instructions Before Testing Part
Before manually inputting questions, ChatGPT was informed about an upcoming series of queries. ChatGPT needed to identify the most plausible response from the available options and explain the reasoning behind its selection. The question types determined the relevant lead-in prompts provided. For the A1 and A2 question types, each input question was deemed independent, rendering any interquestion relationships irrelevant. In contrast, A3/A4 question types implied that multiple questions within a single clinical case shared a connection. However, individual clinical cases were treated as discrete entities, eliminating the need to consider relationships between them. For the B1 question type, 5 shared options were given. ChatGPT needed to identify the correct answers for subsequent questions. Chaining was used in A3/A4 and B1 question types to ensure that multiple questions within a single clinical case in A3/A4 shared the same context, and multiple questions in B1 shared the same options. The number of questions inputted at one time depended on the text’s length, such as 5‐8 questions for A1/A2 types. If necessary, ChatGPT was forced to disregard prior conversational content and commence a fresh chat.
Temperature
The temperature parameter in ChatGPT influences the randomness of the model’s responses. A higher temperature yields more varied and creative answers. In our study, we did not manually adjust the temperature; instead, we used the default setting on the OpenAI platform, commonly at 0.7, to simulate real-world user interactions on the front end. This balance between typical user habits and diverse thought processes was intentional. The default relatively high temperature was expected to enable ChatGPT to generate more creative reasoning processes while still arriving at accurate answers.
Testing Strategy
All the CNMLE 2022 questions were tested in Chinese according to the following 2 factors:
- ChatGPT model selection. Both GPT-3.5 (version from March 23) and GPT-4.0 (version from May 3) were rigorously evaluated on the OpenAI platform from April 24 to May 15, 2023, to ascertain any evolution in the model’s capability in the medical domain.
- System role. This refers to the specific identity or role, such as “gastroenterology specialist,” assigned to ChatGPT to determine if relevant knowledge is applied more accurately. Questions were evaluated both with and without assigning a system role related to the 15 specific clinical subspecialties. This system role was designated by providing a tailored system prompt before the testing instructions, aiming to guide ChatGPT’s approach and align it with specialist viewpoints in the relevant medical field.
Testing Process
Considering the evaluation of the ChatGPT model, system role, and response coherence, each question was tested 8‐12 times. The prompts included those specific to question types, the assignment of system role, and the use of chaining. Slight modifications in these prompts were adopted to avoid potential systematic errors introduced by rigid wording. For example, the prompt “Assume you are a gastroenterology specialist” might vary as “Assume you are highly proficient in gastroenterology.” For coherence evaluation, each question was presented again to ChatGPT. If the regenerated response matched the initial answer, the process was halted. However, if the 2 responses differed, the question was posed once more to ChatGPT.
Response Determination
The first and second responses from ChatGPT were directly assessed against the given standard answers for accuracy. For the final response (referred to as joint response), if 2 of the 3 answers were consistent, this was taken as the conclusive answer and evaluated against the standard. However, if the 3 responses were all distinct, it was automatically marked as incorrect without any further comparison to the standard answer.
The first response was more applicable to assessing whether ChatGPT could pass the CNMLE in the same situation as a student examinee. In contrast, the joint response represented an overall accuracy (the proportion of questions answered correctly at least twice) [
], which was more suitable for demonstrating the potential of ChatGPT in medical education.According to the announcement from the CNMLE Committee of the National Health Commission of China, the passing score for licensed physicians is 360 points, which means an accuracy rate of 60% or above is considered a pass.
Statistical Analysis
Data were collected and managed using Excel software. The statistical analyses were conducted with SPSS (version 26.0.0; IBM Corp). A χ2 test was used to compare the accuracy of CNMLE question responses between different testing strategies and subgroups of question types. Variability was calculated by the number of consistently correct or wrong answers in 2 repeated responses divided by the total number of questions (600). Additionally, the κ statistic was used to evaluate answer consistency. A difference was considered statistically significant when P<.05.
Ethical Considerations
This study collected information that was already published in the bookstore and did not involve human subjects; therefore, approval by the Institutional Review Board of Henan Provincial People’s Hospital was not required.
Results
Accuracy and System Role Assignment
In model comparison, GPT-3.5 achieved an initial accuracy of 54% (324/600) and did not meet the exam criteria. Conversely, GPT-4.0 achieved a passing accuracy of 72.7% (436/600), which was significantly higher than GPT-3.5 (P<.001). Similarly, with a designated system role, GPT-4.0 still exhibited higher accuracy than GPT-3.5 (73% vs 55.3%; P<.001).
Upon system role assignment, both GPT-3.5 and GPT-4.0 showed a slight increase in accuracy compared to when no role was assigned; specifically, 55.3% (332/500) from 54% (324/600) for GPT-3.5 (P>.05) and 73% (438/600) from 72.7% (436/600) for GPT-4.0 (P>.05).
The upper comparisons for the second and joint responses paralleled the initial results, as shown in
.Accuracy | GPT-3.5, n (%) | GPT-4.0, n (%) | P value | GPT-3.5 + SR | , n (%)GPT-4.0 + SR, n (%) | P value |
IR | 324 (54.0) | 436 (72.7) | <.001 | 332 (55.3) | 438 (73.0) | <.001 |
2R | 303 (50.5) | 426 (71.0) | <.001 | 310 (51.7) | 448 (74.7) | <.001 |
JR | 302 (50.3) | 435 (72.5) | <.001 | 329 (54.8) | 437 (72.8) | <.001 |
aSR: system role.
bIR: initial response.
c2R: second response.
dJR: joint response.
Variability of Responses
The GPT-3.5 model exhibited a variability rate of 19.5% (117/600), which decreased to 17.7% (106/600) upon the designation of a system role. The variability rate for GPT-4.0 was observed at 9% (54/600), and further reduced to 7.3% after a system role was assigned. These results indicated a smaller response variability for GPT-4.0 compared to GPT-3.5, and specifying system roles also decreased the variability rates. Both models showed relatively high coherence between the initial and second response, with κ values of 0.778 and 0.610. Detailed information for repeated response can be seen in
.Accuracy for Subgroups
For GPT-4.0, when accounting for system role and repeated responses, there was a statistically significant difference in accuracy across the different units for the CNMLE test, with accuracy ranging from 62% (93/150) to 84% (126/150; P range from<.001 to .01). However, when grouped by question type, the accuracy ranged from 69.4% (145/209) to 83.1% (59/71) without statistical difference (P>.28).
In contrast, for GPT-3.5, only the initial response with system role designation showed a statistical difference in accuracy (P=.04) for question type subgroups. In other groupings by unit or question type, as well as in subsequent responses, the accuracy remained without significant variations (P>.14; see
).Accuracy for initial and joint responses of GPT-3.5/4.0 classified by 15 medical subspecialties is shown in
. In multiple testing strategies, GPT-4.0 outperformed GPT-3.5 in accuracy for 14 distinct clinical subspecialty questions, consistently surpassing the 60% passing threshold.Model strategy | Unit 1 (n=150), n (%) | Unit 2 (n=150), n (%) | Unit 3 (n=150), n (%) | Unit 4 (n=150), n (%) | P value | A1 (n=220), n (%) | A2 (n=209), n (%) | A3/A4 (n=100), n (%) | B1 (n=71), n (%) | P value |
GPT3.5: IR | 82 (54.7) | 83 (55.3) | 88 (58.7) | 71 (47.3) | .25 | 122 (55.5) | 109 (52.2) | 59 (59.0) | 34 (47.9) | .47 |
GPT3.5: 2R | 71 (47.3) | 77 (51.3) | 85 (56.7) | 70 (46.7) | .28 | 115 (52.3) | 103 (49.3) | 57 (57.0) | 28 (39.4) | .14 |
GPT3.5: JR | 72 (48.0) | 75 (50.0) | 86 (57.3) | 69 (46.0) | .22 | 114 (51.8) | 101 (48.3) | 57 (57.0) | 30 (42.3) | .24 |
GPT3.5: IR+ SR | 85 (56.7) | 84 (56.0) | 91 (60.7) | 72 (48.0) | .16 | 129 (58.6) | 113 (54.1) | 61 (61.0) | 29 (40.8) | .04 |
GPT3.5: 2R + SR | 83 (55.3) | 74 (49.3) | 82 (54.7) | 71 (47.3) | .42 | 121 (55.0) | 102 (48.8) | 57 (57.0) | 30 (42.3) | .15 |
GPT3.5: JR+ SR | 84 (56.0) | 80 (53.3) | 91 (60.7) | 74 (49.3) | .25 | 126 (57.3) | 110 (52.6) | 61 (61.0) | 32 (45.1) | .16 |
GPT4.0: IR | 102 (68.0) | 118 (78.7) | 119 (79.3) | 97 (64.7) | .006 | 154 (70.0) | 152 (72.7) | 79 (79.0) | 51 (71.8) | .42 |
GPT4.0: 2R | 100 (66.7) | 112 (74.7) | 119 (79.3) | 95 (63.3) | .009 | 155 (70.5) | 145 (69.4) | 76 (76.0) | 50 (70.4) | .68 |
GPT4.0: JR | 104 (69.3) | 114 (76.0) | 121 (80.7) | 96 (64.0) | .007 | 157 (71.4) | 146 (69.9) | 79 (79.0) | 53 (74.6) | .37 |
GPT4.0: IR+ SR | 103 (68.7) | 116 (77.3) | 126 (84.0) | 93 (62.0) | <.001 | 157 (71.4) | 151 (72.2) | 72 (72.0) | 58 (81.7) | .37 |
GPT4.0: 2R + SR | 104 (69.3) | 117 (78.0) | 124 (82.7) | 103 (68.7) | .01 | 159 (72.3) | 153 (73.2) | 77 (77.0) | 59 (83.1) | .28 |
GPT4.0: JR+ SR | 101 (67.3) | 115 (76.7) | 124 (82.7) | 97 (64.7) | .001 | 156 (70.9) | 151 (72.2) | 73 (73.0) | 57 (80.3) | .47 |
aIR: initial response.
b2R: second response.
cJR: joint response.
dSR: system role.
Discussion
Overview
The CNMLE syllabus outlines the essential knowledge and competencies that physicians need for diagnostic and therapeutic procedures. Acquiring these competencies typically demands that a medical student invest several years in both theoretical education and practical skill development. The application of ChatGPT in medical examinations, particularly within the CNMLE framework, offers a pioneering approach to gauge the potential of LLMs in clinical diagnosis and treatment planning. This study comprehensively assessed ChatGPT’s performance in addressing CNMLE questions, focusing on model evolution and system role designation, which has not yet been fully investigated.
Model Evolution and Performance
In our study, GPT-4.0 consistently outperformed GPT-3.5 in accuracy and reliably met the passing criteria set by the CNMLE Committee. Despite GPT-3.5 achieving an accuracy rate of over 50%, it failed to pass the examination. A noncomparison study using GPT-3.5 to test CNMLE 2020‐2022 achieved an accuracy of (36.5%‐47%) [
]. The lower accuracy might be attributed to the fact that the testing was conducted before February, shortly after the release of GPT-3.5. The better performance of GPT-4.0 compared with GPT-3.5 was also reported by Wang et al [ ]. However, it is noteworthy that their assessment was based on a limited sample of 100 questions, rather than a full set of 600 questions. The small sample might have contributed to the overall favorable results (GPT-4.0: 84%; GPT-3.5: 56%). Therefore, our findings might provide a more representative comparison of the real-world performance of GPT-4.0 and 3.5 on the CNMLE.Other research on evaluating ChatGPT’s accuracy on national medical licensing examinations included assessments of the USMLE [
, ] and the Japanese National Medical Licensing Examination [ ]. The conclusions were similar to ours: while GPT-3.5 was often at or near the passing threshold, GPT-4.0 passed relevant exams and had higher testing accuracy compared to GPT-3.5. This trend was not only limited to national medical licensing examinations but also applied to other medical-related examinations. However, the specific accuracy varied across models, possibly due to differences in study countries, testing time, exam content, and other variables. A comprehensive review of existing published and non–peer-reviewed research findings is available in .Study | Country | Test model | Examination | Data sample, n | Passing threshold | Accuracy (%) |
Gilson et al [ | ]United States | GPT-3.5 | The United States Medical Licensing Examination Step 1 and Step 2 exams | 87‐102 | 60% | GPT-3.5: 44.0‐64.4 |
Kung et al [ | ]United States | GPT-3.5 | The United States Medical Licensing Exam | 376 | 60% | At or near 60% |
Guerra et al [ | ]United States | GPT-4.0 and 3.5 | Congress of Neurological Surgeons Self-Assessment Neurosurgery Exam | 591 | — | GPT-4.0: 76.6; GPT-3.5: 60.2 |
Takagi et al [ | ]Japan | GPT-4.0 and 3.5 | Japanese National Medical Licensing Examination (2023) | 254 | GPT-4.0: Pass; GPT-3.5: Failed | GPT-4.0: 79.9; GPT-3.5: 50.8 |
Wang et al [ | ]China | GPT-4.0 and 3.5 | The Chinese National Medical Licensing Examination | 100 | — | GPT-4.0: 84; GPT-3.5: 56 |
Cai et al [ | ]United States | GPT-4.0 and 3.5 | Ophthalmology Board-Style Questions | 250 | — | GPT-4.0: 71.6; GPT-3.5: 58.8 |
Oh et al [ | ]Korea | GPT-4.0 and 3.5 | Korean General Surgery Board Exams | 280 | — | GPT-4.0: 76.4; GPT-3.5: 46.8 |
Skalidis et al [ | ]Switzerland | GPT-3.5 | The European Exam in Core Cardiology | 488 | Pass | GPT-3.5: 58.8 |
Saad et al [ | ]United Kingdom | GPT-4.0 | The Orthopaedic FRCS Orth Part A exam | 240 | Failed | GPT-4.0: 67.5 |
Weng et al [ | ]China | GPT-3.5 | Taiwan’s 2022 Family Medicine Board Exam | 125 | Failed | GPT-3.5: 41.6 |
Kumah-Crystal et al [ | ]United States | GPT-3.5 | The Clinical Informatics Board Examination | 254 | 60%, Pass | GPT-3.5: 74 |
Mihalache et al [ | ]Canada | GPT-4.0 | OphthoQuestions practice question bank for board certification examination | 125 | — | GPT-4.0: 84 |
Ali et al [ | ]United States | GPT-4.0 and 3.5 | Self-Assessment Neurosurgery Examination Indications Examination | 149 | — | GPT-4.0: 82.6; GPT-3.5: 62.4 |
Oztermeliet al [ | ]Turkey | GPT-3.5 | Turkey Medical Specialty Exams | 1177 | — | GPT-3.5: 54.3‐70.9 |
Fijaoko et al [ | ]United States | GPT-3.5 | American Heart Association Basic Life Support and Advanced Cardiovascular Life Support exams | 126 | 84%, Failed | GPT-3.5: 64‐68.4 |
Su et al [ | ]China (Taiwan) | GPT-3.5 | Taiwan\'s 2022 Nursing Licensing Exam | 400 | Pass | GPT-3.5: 80.75% |
Lewandowski et al [ | ]Poland | GPT-4.0 and 3.5 | The Dermatology Specialty Certificate Examinations | 120 × 3 | GPT-4 Pass | GPT-4.0: >70% better than GPT-3.5 |
Kung et al [ | ]United States | GPT-4.0 and 3.5 | Orthopaedic In-Training Examination (2020‐2022) | 360 | GPT-4.0: >PGY | -5 level; GPT-3.5: PGY-1 levelGPT-4.0: 73.6; GPT-3.5: 54.3 |
Gencer and Aydin [ | ]Turkey | GPT-4.0 and 3.5 | Turkish-language thoracic surgery exam | 105 | Surpass students’ scores | GPT-4.0: 93.3; GPT-3.5: 90.5 |
aNot available.
bPGY: postgraduate year.
System Role for Accuracy
While it was expected that introducing system role tailored for clinical subspecialties would enhance the reliability of ChatGPT’s medical responses, this effect had not been systematically studied. Our research addressed this gap. Our findings revealed slight but noteworthy improvements in accuracy for both GPT-3.5 (1.3%‐4.5%) and GPT-4.0 (0.3%‐3.7%), although these gains were not statistically significant. This might imply that ChatGPT’s inherent abilities are already robust enough to discern and address the medical inquiries without narrowing down its response scope.
Response Variability
As an LLM, ChatGPT naturally exhibits variability in responses when the temperature hyperparameter is not zero. In this study, we adopted the default temperature of 0.7 to simulate real-world use conditions on the front end. Our results showed relatively high coherence between the initial and second responses for both GPT-4.0 and GPT-3.5. Therefore, the relatively high temperature of 0.7 is feasible and recommended when testing ChatGPT’s performance on the CNMLE. Furthermore, our results highlighted that both model evolution and system roles contribute to ChatGPT’s variability in scenarios such as the Chinese Medical Licensing Exams. This variability can be valuable for medical education, as ChatGPT not only provides answers to questions but also includes the rationale and references for its choices, which allows students to easily follow and comprehend [
]. Repeatedly submitting questions allows groups or individuals to engage with the explanatory content generated by ChatGPT, which is particularly beneficial for open-ended case scenario discussions [ ].Subgroup and Multispecialty Analysis
Our subgroup analysis revealed that ChatGPT demonstrated consistent accuracy across different types of questions. This indicated that ChatGPT was capable of understanding and analyzing complex medical cases and scenarios (A2, A3/A4 questions), which can be challenging even for humans, and making correct decisions. This decision-making ability was equally proficient when addressing more straightforward, common-sense questions that did not require reasoning (A1, B1 questions).
In comparisons among unit subgroups representing different subspecialties, significant performance variations were observed in GPT-4.0 across CNMLE units. GPT-4.0 exhibited higher accuracy in units 2‐3, which predominantly featured questions from subspecialties such as cardiovascular, urinary, digestive, and respiratory systems. This was further corroborated by our multispecialty analysis results. GPT-4.0 achieved an accuracy rate of over 75% for these 4 subspecialties, surpassing its overall accuracy rate of 72.7%. Given that these 4 subspecialties accounted for a substantial proportion (34.5%) of all 15 subspecialties, such a disparity might have been advantageous. However, this disparity disappeared upon the introduction of system roles as prompts, with the overall accuracy of GPT-4.0 increasing to 78.6%. This might suggest that the appropriate use of system roles could compensate for individual subspecialty question accuracy, thereby enhancing the overall accuracy of ChatGPT.
Furthermore, we used CNMLE questions, divided into 15 medical subspecialties, to comprehensively assess the medical expertise of ChatGPT models. This approach provided a robust framework for evaluating model proficiency across a variety of medical fields. Notably, GPT-4.0 surpassed the 60% passing threshold in 14 of the 15 distinct clinical subspecialties, in contrast to GPT-3.5, which only passed in 7 out of 15 subspecialties. This highlighted the superiority of GPT-4.0 and its potential in medical applications.
Generalizability of Findings
Previous studies [
] often excluded table and image-based questions when evaluating ChatGPT’s performance in medical exams. This approach limited the generalizability of these findings due to ChatGPT’s lack of multimodal data processing. In contrast, our study, focusing on the CNMLE’s multiple-choice format, which almost exclusively consists of nongraphical and nontabular questions, offers greater generalizability in real exam settings. Zhu et al [ ] suggested that ChatGPT, as a chatbot, had advantages in responding to open-ended questions, corresponding more closely with real-world scenarios where users sought medical support knowledge from ChatGPT. The potential of ChatGPT in exams with open-ended questions merits further exploration.Limitations
First, this study assessed ChatGPT’s ability to answer questions from the Chinese version of the CNMLE. As ChatGPT is mainly trained on English data, Chinese questions could have underestimated its capabilities. Second, the CNMLE questions were multiple-choice, introducing the chance factor in selecting correct answers. Limited by paper length, we did not evaluate the logic behind ChatGPT’s choices, although this aspect is critical and merits deeper investigation. Third, real-world medical questions often have open-ended, multiple, or uncertain answers. Therefore, the CNMLE may not represent the full scope of challenges ChatGPT might face in clinical settings. Consequently, GPT-4.0’s success on the CNMLE may only indicate its partial competence in clinical decision-making. Future studies should broaden the range of question types to better assess ChatGPT’s medical performance. Despite these limitations, we believe this study provided valuable insights into ChatGPT’s capabilities in medicine.
Conclusions
This study comprehensively evaluated the performance of GPT-4.0 and GPT-3.5 in the context of the CNMLE. Our findings indicated that GPT-4.0 not only met the CNMLE passing criteria but also significantly outperformed GPT-3.5 in key areas such as accuracy, consistency, and medical subspecialty expertise. Furthermore, the implementation of system roles served as a pivotal factor in enhancing the model’s reliability and answer coherence. These results collectively underscored GPT-4.0’s promising potential as a valuable tool for medical professionals, educators, and students, warranting further research and application in the medical field.
Acknowledgments
This research was supported by Medical Science and Technology Tackling Plan of Henan Province (LHGJ20210078).
Authors' Contributions
SM and BL conceived the study and share the corresponding author. SM, QG, and WC collected all relevant data and assisted in results interpretation. SM designed the study, carried out data analysis, and drafted the manuscript. BL participated in the design and reviewed the manuscript. All authors contributed to the article and approved the submitted version.
Conflicts of Interest
None declared.
Question type explanations conveyed to ChatGPT via structured prompts.
DOCX File, 12 KBDetail information for repeated responses and their κ value under the ChatGPT default temperature of 0.7.
DOCX File, 16 KBReferences
- Gilson A, Safranek CW, Huang T, et al. How does ChatGPT perform on the United States Medical Licensing Examination (USMLE)? The implications of large language models for medical education and knowledge assessment. JMIR Med Educ. Feb 8, 2023;9:e45312. [CrossRef] [Medline]
- Kung TH, Cheatham M, Medenilla A, et al. Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. PLOS Digit Health. Feb 2023;2(2):e0000198. [CrossRef] [Medline]
- Seghier ML. ChatGPT: not all languages are equal. Nature. Mar 2023;615(7951):216. [CrossRef] [Medline]
- Fijačko N, Gosak L, Štiglic G, Picard CT, John Douma M. Can ChatGPT pass the life support exams without entering the American Heart Association course? Resuscitation. Apr 2023;185:109732. [CrossRef] [Medline]
- Weng TL, Wang YM, Chang S, Chen TJ, Hwang SJ. ChatGPT failed Taiwan’s Family Medicine Board Exam. J Chin Med Assoc. Aug 1, 2023;86(8):762-766. [CrossRef] [Medline]
- Morreel S, Mathysen D, Verhoeven V. Aye, AI! ChatGPT passes multiple-choice family medicine exam. Med Teach. Jun 3, 2023;45(6):665-666. [CrossRef] [Medline]
- Takagi S, Watari T, Erabi A, Sakaguchi K. Performance of GPT-3.5 and GPT-4 on the Japanese Medical Licensing Examination: comparison study. JMIR Med Educ. Jun 29, 2023;9:e48002. [CrossRef] [Medline]
- Oh N, Choi GS, Lee WY. ChatGPT goes to the operating room: evaluating GPT-4 performance and its potential in surgical education and training in the era of large language models. Ann Surg Treat Res. May 2023;104(5):269-273. [CrossRef] [Medline]
- Currie G, Barry K. ChatGPT in nuclear medicine education. J Nucl Med Technol. Sep 2023;51(3):247-254. [CrossRef] [Medline]
- Ali R, Tang OY, Connolly ID, et al. Performance of ChatGPT and GPT-4 on neurosurgery written board examinations. Neurosurgery. Dec 1, 2023;93(6):1353-1365. [CrossRef] [Medline]
- Antaki F, Touma S, Milad D, El-Khoury J, Duval R. Evaluating the performance of ChatGPT in ophthalmology: an analysis of its successes and shortcomings. Ophthalmol Sci. Dec 2023;3(4):100324. [CrossRef] [Medline]
- Su M, Lin L, Lin L, Chen Y. Assessing question characteristic influences on ChatGPT's performance and response-explanation consistency: Insights from Taiwan's Nursing Licensing Exam. Int J Nurs Stud. May 2024;153:104717. [CrossRef] [Medline]
- Ali K, Barhom N, Tamimi F, Duggal M. ChatGPT-a double-edged sword for healthcare education? Implications for assessments of dental students. Eur J Dent Educ. Feb 2024;28(1):206-211. [CrossRef] [Medline]
- Holmes J, Liu Z, Zhang L, et al. Evaluating large language models on a highly-specialized topic, radiation oncology physics. Front Oncol. Jul 2023;13:1219326. [CrossRef] [Medline]
- GPT-4. OpenAI. URL: https://openai.com/research/gpt-4/ [Accessed 2023-11-21]
- Yang R, Tan TF, Lu W, Thirunavukarasu AJ, Ting DSW, Liu N. Large language models in health care: development, applications, and challenges. Health Care Sci. Aug 2023;2(4):255-263. [CrossRef] [Medline]
- Zhu L, Mou W, Yang T, Chen R. ChatGPT can pass the AHA exams: open-ended questions outperform multiple-choice format. Resuscitation. Jul 2023;188:109783. [CrossRef] [Medline]
- Sarraju A, Bruemmer D, Van Iterson E, Cho L, Rodriguez F, Laffin L. Appropriateness of cardiovascular disease prevention recommendations obtained from a popular online chat-based artificial intelligence model. JAMA. Mar 14, 2023;329(10):842-844. [CrossRef] [Medline]
- Zhu L, Mou W, Chen R. Can the ChatGPT and other large language models with internet-connected database solve the questions and concerns of patient with prostate cancer and help democratize medical knowledge? J Transl Med. Apr 19, 2023;21(1):269. [CrossRef] [Medline]
- Strong E, DiGiammarino A, Weng Y, et al. Chatbot vs medical student performance on free-response clinical reasoning examinations. JAMA Intern Med. Sep 1, 2023;183(9):1028-1030. [CrossRef] [Medline]
- National Clinical Practitioner Qualification Exam: past years’ real exam papers and detailed solutions [Article in Chinese]. JD. 2022. URL: https://item.jd.com/30821733544.html/ [Accessed 2023-04-20]
- Introduction of medical licensing examination. The Chinese National Medical Examination Center. URL: https://www1.nmec.org.cn/Pages/ArticleInfo-13-10706.html/ [Accessed 2023-11-21]
- Wang X, Gong Z, Wang G, et al. ChatGPT performs on the Chinese National Medical Licensing Examination. J Med Syst. Aug 15, 2023;47(1):86. [CrossRef] [Medline]
- Wang H, Wu W, Dou Z, He L, Yang L. Performance and exploration of ChatGPT in medical examination, records and education in Chinese: pave the way for medical AI. Int J Med Inform. Sep 2023;177:105173. [CrossRef] [Medline]
- Guerra GA, Hofmann H, Sobhani S, et al. GPT-4 artificial intelligence model outperforms ChatGPT, medical students, and neurosurgery residents on neurosurgery written board-like questions. World Neurosurg. Nov 2023;179:e160-e165. [CrossRef] [Medline]
- Cai LZ, Shaheen A, Jin A, et al. Performance of generative large language models on ophthalmology board–style questions. Am J Ophthalmol. Oct 2023;254:141-149. [CrossRef] [Medline]
- Skalidis I, Cagnina A, Luangphiphat W, et al. ChatGPT takes on the European exam in core cardiology: an artificial intelligence success story? Eur Heart J Digit Health. May 2023;4(3):279-281. [CrossRef] [Medline]
- Saad A, Iyengar KP, Kurisunkal V, Botchu R. Assessing ChatGPT’s ability to pass the FRCS orthopaedic part A exam: a critical analysis. Surgeon. Oct 2023;21(5):263-266. [CrossRef] [Medline]
- Kumah-Crystal Y, Mankowitz S, Embi P, Lehmann CU. ChatGPT and the clinical informatics board examination: the end of unproctored maintenance of certification? J Am Med Inform Assoc. Aug 18, 2023;30(9):1558-1560. [CrossRef] [Medline]
- Mihalache A, Huang RS, Popovic MM, Muni RH. Performance of an upgraded artificial intelligence chatbot for ophthalmic knowledge assessment. JAMA Ophthalmol. Aug 1, 2023;141(8):798-800. [CrossRef] [Medline]
- Ali R, Tang OY, Connolly ID, et al. Performance of ChatGPT, GPT-4, and Google Bard on a neurosurgery oral boards preparation question bank. Neurosurgery. Nov 1, 2023;93(5):1090-1098. [CrossRef] [Medline]
- Oztermeli AD, Oztermeli A. ChatGPT performance in the medical specialty exam: an observational study. Medicine (Baltimore). Aug 11, 2023;102(32):e34673. [CrossRef] [Medline]
- Lewandowski M, Łukowicz P, Świetlik D, Barańska-Rybak W. An original study of ChatGPT-3.5 and ChatGPT-4 dermatological knowledge level based on the Specialty Certificate Examination in Dermatology. Clin Exp Dermatol. Jun 25, 2024;49(7):686-691. [CrossRef] [Medline]
- Kung JE, Marshall C, Gauthier C, Gonzalez TA, Jackson JB. Evaluating ChatGPT performance on the orthopaedic in-training examination. JB JS Open Access. Sep 8, 2023;8(3):e23.00056. [CrossRef] [Medline]
- Gencer A, Aydin S. Can ChatGPT pass the thoracic surgery exam? Am J Med Sci. Oct 2023;366(4):291-295. [CrossRef] [Medline]
Abbreviations
CNMLE: Chinese National Medical Licensing Examination |
LLM: large language model |
NMEC: National Medical Examination Center |
USMLE: United States Medical Licensing Examination |
Edited by Blake Lesselroth; submitted 15.09.23; peer-reviewed by Andrew Mihalache, Rui Yang; final revised version received 20.05.24; accepted 20.06.24; published 13.08.24.
Copyright© Shuai Ming, Qingge Guo, Wenjun Cheng, Bo Lei. Originally published in JMIR Medical Education (https://mededu.jmir.org), 13.8.2024.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Education, is properly cited. The complete bibliographic information, a link to the original publication on https://mededu.jmir.org/, as well as this copyright and license information must be included.