TY - JOUR AU - Tseng, Liang-Wei AU - Lu, Yi-Chin AU - Tseng, Liang-Chi AU - Chen, Yu-Chun AU - Chen, Hsing-Yu PY - 2025 DA - 2025/3/19 TI - Performance of ChatGPT-4 on Taiwanese Traditional Chinese Medicine Licensing Examinations: Cross-Sectional Study JO - JMIR Med Educ SP - e58897 VL - 11 KW - artificial intelligence KW - AI language understanding tools KW - ChatGPT KW - natural language processing KW - machine learning KW - Chinese medicine license exam KW - Chinese medical licensing examination KW - medical education KW - traditional Chinese medicine KW - large language model AB - Background: The integration of artificial intelligence (AI), notably ChatGPT, into medical education, has shown promising results in various medical fields. Nevertheless, its efficacy in traditional Chinese medicine (TCM) examinations remains understudied. Objective: This study aims to (1) assess the performance of ChatGPT on the TCM licensing examination in Taiwan and (2) evaluate the model’s explainability in answering TCM-related questions to determine its suitability as a TCM learning tool. Methods: We used the GPT-4 model to respond to 480 questions from the 2022 TCM licensing examination. This study compared the performance of the model against that of licensed TCM doctors using 2 approaches, namely direct answer selection and provision of explanations before answer selection. The accuracy and consistency of AI-generated responses were analyzed. Moreover, a breakdown of question characteristics was performed based on the cognitive level, depth of knowledge, types of questions, vignette style, and polarity of questions. Results: ChatGPT achieved an overall accuracy of 43.9%, which was lower than that of 2 human participants (70% and 78.4%). The analysis did not reveal a significant correlation between the accuracy of the model and the characteristics of the questions. An in-depth examination indicated that errors predominantly resulted from a misunderstanding of TCM concepts (55.3%), emphasizing the limitations of the model with regard to its TCM knowledge base and reasoning capability. Conclusions: Although ChatGPT shows promise as an educational tool, its current performance on TCM licensing examinations is lacking. This highlights the need for enhancing AI models with specialized TCM training and suggests a cautious approach to utilizing AI for TCM education. Future research should focus on model improvement and the development of tailored educational applications to support TCM learning. SN - 2369-3762 UR - https://mededu.jmir.org/2025/1/e58897 UR - https://doi.org/10.2196/58897 DO - 10.2196/58897 ID - info:doi/10.2196/58897 ER -