Published on in Vol 12 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/66381, first published .
Perceptions and Attitudes of Medical Students Toward the Integration of Large Language Models in Medical Education: Cross-Sectional Survey in China

Perceptions and Attitudes of Medical Students Toward the Integration of Large Language Models in Medical Education: Cross-Sectional Survey in China

Perceptions and Attitudes of Medical Students Toward the Integration of Large Language Models in Medical Education: Cross-Sectional Survey in China

1Department of Urology, Xiangya Hospital, Central South University, No.87 Xiangya Road, Kaifu District, Changsha, China

2Department of Neurology, The Second Xiangya Hospital, Central South University, Changsha, China

3Department of Urology, The Third Xiangya Hospital, Central South University, Changsha, China

4Department of Education, Xiangya Hospital, Central South University, Changsha, China

5Department of Epidemiology and Health Statistics, Xiangya School of Public Health, Central South University, Changsha, China

Corresponding Author:

Cheng Zhao, MD


Background: Although artificial intelligence (AI) is being rapidly integrated into medical education, insights from medical students, particularly in the Chinese context, remain limited.

Objective: This study was designed to explore Chinese medical students’ perceptions of and attitudes toward the integration of AI into medical education, as well as the factors that may influence their perspectives. The findings of this research offer valuable insights to assist medical educators in the effective implementation of these innovative educational approaches.

Methods: On the basis of the estimated total number of clinical medical students at the target institutions, the sample size was calculated to be 379. A web-based questionnaire survey was designed to investigate the acceptance level of medical students toward the application of AI. The questionnaire consisted of 14 questions across 4 dimensions, which included demographic characteristics, perceptions of AI application, willingness, and concerns. Each dimension contained 3 to 4 questions. Descriptive statistics were used to tabulate the frequency of each variable. Chi-square tests and multiple regression analyses were conducted to measure the influencing factors.

Results: A total of 566 cross-sectional online surveys were distributed from December 2023 to January 2024 through a snowball sampling method. Finally, 490 medical students from various local tertiary medical centers were involved. Overall, a majority of the participants showed a positive attitude toward future learning and the usage of AI, manifested as totally willing to acquire relevant knowledge (222/490, 45.3%), totally willing to use AI tools (230/490, 46.9%), and totally desiring that schools would offer AI-related courses (230/490, 46.9%). However, there is still a large proportion (392/490, 80.0%) of participants who expressed concerns regarding ethical issues. The findings also indicated that gender and educational level were significantly correlated with the AI application. Specifically, regression analysis indicated that male participants were more inclined to acquire AI information through social media (odds ratio 0.458, 95% CI 0.33‐0.67; P<.001) and that male or graduate-level participants were more likely to use AI for academic writing purposes (odds ratio 0.476, 95% CI 0.38‐0.82; P=.001 for male; odds ratio 1.552, 95% CI 1.32‐1.77; P=.009 for graduate students, respectively).

Conclusions: Our findings indicate that a general awareness of AI’s role in medical education is evident among students. However, subgroup-specific differences must be taken into account, particularly when designing and optimizing educational strategies integrated with AI. This consideration is critical to ensuring that such tools align with the diverse learning needs of distinct student groups.

JMIR Med Educ 2026;12:e66381

doi:10.2196/66381

Keywords



Artificial intelligence (AI) is increasingly used across various sectors, such as finance, health care, and education [1-4]. In the field of medical education, AI holds the potential to enhance teaching and learning by improving students’ critical thinking and communication skills [5]. However, while numerous studies have examined the views of medical educators, the perspectives of medical students themselves have received less attention. As modern medical education emphasizes collaboration between educators and learners, a thorough understanding of students’ attitudes is crucial for designing and implementing effective teaching strategies [6,7].

As China undergoes a broad transformation driven by AI, the distribution of educational resources in medicine remains uneven [8]. Currently, only a limited number of medical schools have established specialized AI courses, primarily located in economically developed regions [9]. This regional disparity in educational resources has led to widespread misconceptions about AI among medical students, which in turn may amplify ethical and privacy-related risks [10]. While Chinese policymakers acknowledge the need for regulation in this rapidly evolving field, most existing policies remain high-level guidelines, providing limited practical direction for specific sectors such as medical education [11].

To address the aforementioned issues, a cross-sectional study was conducted. The primary goal of this study was to (1) investigate the perceptions, attitudes, and willingness of medical students in China toward the integration of large language models (LLMs) into medical education; (2) identify key influencing factors associated with these perceptions and attitudes; and (3) provide empirical evidence to inform the design of targeted AI-related educational strategies for Chinese medical institutions and discuss the regional disparities while using this novel technology. No a priori hypotheses were formulated, given the exploratory nature of investigating region-specific perspectives on LLM integration in medical education.


Study Population

A web-based questionnaire survey was conducted in the central south region of China, and all participants were recruited from 2 regional tertiary hospitals in Changsha and 3 tertiary hospitals affiliated with Central South University. The inclusion criteria were as follows: (1) full-time students majoring in clinical medicine, (2) each participant could respond only once per account, (3) the total response time for completing the questionnaire was longer than 60 seconds, and (4) participants were recruited between December 2023 and January 2024. The exclusion criteria included (1) inability to complete the questionnaire in Chinese, (2) a history of cognitive impairment or mental illness, and (3) no prior clinical practice experience. Specifically, the assessment of cognitive impairment or mental illness was conducted via a self-reported question at the start of the questionnaire: “Have you ever been diagnosed with a cognitive disorder or mental illness by a professional physician?” Participants who answered “yes” were automatically excluded from the survey to ensure data quality and protect their well-being. This screening method was conducted in compliance with the Declaration of Helsinki [12]. Finally, we used the CHERRIES (Checklist for Reporting Results of Internet E-Surveys) to guide our report [13].

Sample Size Calculation

The sample size for this cross-sectional survey was calculated to estimate the proportion of clinical medical students holding positive attitudes toward AI applications in medical education. The following formula for finite populations was used:

Formula: n=[N×z²×p(1−p)]/[(N−1)×e²+z²×p(1−p)]

Parameters: N=total population size (estimated at 2000 clinical students across target hospitals), z=1.96 (for a 95% confidence level), P=.50 (conservative proportion assuming maximum variability), and e=0.05 (desired margin of error)

Calculation:

n=[N×z²×p(1-p)]/[(N-1)×e²+z²×p(1-p)]

=[2000×(1.96)²×0.5×.5]/[(1999×0.0025)+(3.8416×0.25)]

=[2000×3.8416×.25]/[4.9975+0.9604]

=1920.8/5.9579 ≈ 322

Nonresponse adjustment (15% attrition):

n_adjusted=322/0.85 ≈ 379

A minimum of 379 participants was required.

Questionnaire Design

The questionnaire was developed based on a comprehensive review of the existing literature on related topics and was modified according to local perceptions in China [14,15]. As most of the participants included in our study were preclinical medical students, the questionnaire items were designed to avoid esoteric or ambiguous research-related or clinical questions as much as possible. Before the formal investigation, a pilot test of the questionnaires was conducted in this study. Three experts (LW, JW, and WY) were invited to conduct a systematic content validity evaluation. Subsequently, 20 students were recruited from different clinical disciplines, maintaining a male-to-female ratio of 1:1. The participants included both undergraduate and postgraduate students. These medical students completed the questionnaire to identify any encountered issues. On the basis of their feedback, the questionnaire was further refined. It should be noted that the responses from the pilot test were excluded from the final data analysis. The final questionnaire, consisting of 14 questions, was published on the WJX platform [16]. It addressed four dimensions: demographic characteristics, perceptions, willingness, and concerns. Each dimension encompassed 3 to 4 questions, and depending on the research target, question formats were either single or multiple choices.

Distribution of Questionnaire

We distributed the online questionnaire through “WeChat,” (Tencent Holdings Limited) a widely used real-time messaging application in China. The team members contacted educators from various hospitals through WeChat, informed them of the purpose of the survey, and then forwarded the questionnaire to them in WeChat groups of students from different grades. Medical students who agreed to participate scanned the QR code to anonymously answer the questionnaire online. To further enhance the reach and minimize potential bias, we used a snowball sampling method by requesting initial respondents to forward the survey link to other medical students in their networks [17].

Ethical Considerations

Informed consent and instructions were displayed at specific locations in the questionnaire. Participants were required to complete an informed consent form, which stated, “You acknowledge and agree that the data collected may be used for research purpose,” as a prerequisite for proceeding with the survey. Participants could freely modify their selected options before final submission and retract their questionnaire at any time after delivery.

Given the nature of the study (anonymous, noninterventional, and without collection of personal or sensitive data), it was exempt from formal institutional review board review and approval, in accordance with the ethical guidelines of the Medical Ethics Committee of Xiangya Hospital, Central South University. The study was conducted in full compliance with relevant ethical principles, ensuring the protection of participants’ privacy and rights throughout the research process. As the survey was voluntary and designed to take no more than 10 minutes to complete, no financial or nonfinancial compensation was provided to participants.

Data Analysis

After questionnaire collection, data cleaning was conducted. The questionnaires were screened based on the following predetermined criteria: (1) responses with excessive missing data, accounting for two-thirds of the total questions; (2) responses exhibiting uniform or highly regular answer selections; and (3) questionnaires completed in less than 60 seconds. These time thresholds were established as reasonable response times based on the pilot test conducted previously.

We used SPSS Statistics (version 26; IBM Corporation) for data analysis. First, descriptive analysis was performed to summarize the frequency distribution characteristics of the variables. Subsequently, subgroup analysis using chi-square tests was conducted to explore the associations between different demographic characteristics. Finally, regression analysis was applied to control for confounding factors and further investigate the associations between factors such as gender, age, and education level with the study objectives. If P<.05, differences were considered statistically significant. The reliability of the items from the willingness and concerns dimensions was tested using the Cronbach α coefficient analysis. The results of the Cronbach α were considered to have an acceptable reliability, as the generally accepted rule is that α values of 0.6 to 0.7 indicate an acceptable level of reliability, and 0.8 or greater is a very good level [18].


Demographic Characteristics

A total of 566 questionnaires were distributed online, and 490 valid responses were collected (a response rate of 86.6%). The demographic characteristics of the participants are presented in Table 1. Overall, the gender distribution of participants was balanced (248/490, 50.7%, men vs 242/490, 49.3%, women), with the majority concentrated in the age range of 20 to 25 years (304/490, 62.0%), and only 32 individuals aged 35 years and older (32/490, 6.5%). The majority of participants were undergraduate students (258/490, 52.7%), followed by master’s candidates (134/490, 27.3%), PhD candidates (74/490, 15.1%), and postdoctoral fellows (16/490, 3.3%). To clarify the Chinese medical education model, the following definitions apply to participant degree programs: (1) undergraduate students, that is, students pursuing a 5-year Bachelor of Medicine degree with a primary focus on foundational medical courses and who are not concurrently enrolled in other degree programs; (2) graduate students, including master’s (3-y program) and PhD (3‐ to 4-y program) students, who specialize in clinical subfields or medical research and are required to complete clinical rotations and publish academic papers; and (3) postdoctoral fellows, who are researchers who have completed a PhD and engage in advanced clinical or research work at hospitals; while not “students” in the traditional sense, they were included because they continue to participate in medical education training programs and represent the upper end of the medical learner spectrum. This classification aligns with national medical education standards in China [19].

Table 1. Demographic characteristics of study (N=490).
CharacteristicsValues (%)
Sex
Male248 (50.7)
Female242 (49.3)
Age (y)
<208 (1.6)
21‐25304 (62.0)
26‐30100 (20.4)
31‐3546 (9.4)
>3532 (6.5)
Educational level
Junior college8 (1.6)
Undergraduate degree258 (52.7)
Master’s degree134 (27.3)
PhD74 (15.1)
Postdoctorate16 (3.3)

The specific frequency distribution is as follows: The age range of participants is from 18 to 38 years, and the median age is 23.1 (18.7-31.6) years. The extreme values were defined as ages less than 20 (8/490, 1.6%) or greater than 35 (32/490, 6.5%) years. Specifically, the proportion of participants aged less than 20 years was (8/490, 1.6%), all of whom were undergraduate students in their first or second year. Most participants (304/490, 62.0%) were aged 21 to 25 years, corresponding to undergraduate fourth- or fifth-year students or first-year postgraduates, followed by those aged 26 to 30 years (100/490, 20.4%), who were second- or third-year postgraduates, and those aged older than 30 years (78/490, 15.9%), who were PhD or postdoctoral fellows. This distribution is consistent with the age structure of medical students in China, where most enter postgraduate studies after 23 years of age [19].

Perceptions of AI Application

This set of questions primarily focused on participants’ perceptions of the application of LLMs. The description of participants’ perceptions of LLMs is presented in Tables 2-4. Subsequently, we conducted chi-square tests to investigate the correlation between other factors and perceptions. First, male participants were more likely to have been exposed to information related to LLMs before the conduction of the survey (158/248, 63.7% vs 132/242, 54.5% for women; χ²1=5.4; P=.02), as well as educational level also contributed to this issue (undergraduates: 162/258, 62.7%; master’s students: 80/134, 59.7%; PhD candidates: 30/74, 40.5%; χ²1=19.854; P=.001). Second, gender (180/248, 72.5% vs 130/242, 53.7% for women; χ²1=18.7; P<.001), age (21‐25 y: 142/304, 46.7%; 26‐30 y: 60/100, 60%; and 31‐35 y: 38/46, 82%; χ²1=26.4; P<.001), and educational level (undergraduates: 170/258, 65.8%; master’s students: 68/134, 50.7%; PhD candidates: 56/74, 75.6%; χ²1=38.7; P<.001) were all correlated with the way that participants primarily acquired LLM-related information. A clear trend was observed that as education level increased, participants were more likely to access LLM information via academic journals or friends chatting and to recognize the value of LLM for research. Third, gender (206/248, 83.0% vs 168/242, 69.4% for women; χ²1=12.6; P<.001), age (21‐25 y: 222/304, 73.0%; 26‐30 y: 84/100, 84.0%; 31‐35 y: 40/46, 87.0%; χ²1=25.5; P<.001), and educational level (undergraduates: 186/258, 72.1%; master’s students: 110/134, 82.1%; PhD candidates: 60/74, 81.1%; χ²1=22.5; P<.001) also influenced participants’ views on the application methods of LLMs in the medical field, especially in assisting in academic writing and literature translation. Finally, age (21‐25 y: 232/304, 76.3%; 26‐30 y: 84/100, 84%; 31‐35 y: 42/46, 91.3%; χ²1=25.7; P<.001) and educational level (undergraduates: 198/258, 76.7%; master’s students: 110/134, 82.1%; PhD candidates: 70/74, 94.6%; χ²1=25.0: P<.001) were also correlated with the challenges of LLMs in medical applications, specifically concerns about privacy protection. These trends indicated that older participants expressed greater concern about privacy risks compared to younger participants, which may reflect 2 key aspects: first, a higher demand for AI applications in the academic domain among more experienced learners, and second, a stronger awareness of data protection among this group.

Table 2. Chi-squared test between gender and categorical variables.
VariableGender, nPearson χ² (df)P valuea
MaleFemale
Have you ever been exposed to information related to large language models?5.4 (1).02
Yes158132
No90110
In what ways have you primarily acquired information about large language models?18.7 (1)<.001
Social media
Yes180130
No68112
Friend chatting
Yes118138
No130104
What do you think large language models can be used for in the field of medicine?12.6 (1)<.001
Assistance in writing academic papers
Yes206168
No4274
Translation of literature
Yes186160
No6282

aP<.05 was considered statistically significant.

Table 3. Chi-squared test between age and categorical variables.
VariableAge (y), nPearson χ² (df)P valuea
<2021‐2526‐3031‐35>35
In what ways have you primarily acquired information about large language models?26.4 (1)<.001
Friend chatting
Yes2142603814
No616240818
What do you think large language models can be used for in the field of medicine?25.5 (1)<.001
Assistance in writing academic papers
Yes2222844026
No6821666
Translation of literature
Yes0208803622
No896201010
What challenges do you believe large language models encounter in their applications within the field of medicine?25.7 (1)<.001
Privacy protection
Yes2232844230
No6721642

aP<.05 was considered statistically significant.

Table 4. Chi-squared test between education level and categorical variables.
VariableEducation level, nPearson χ² (df)P valuea
Junior collegeUndergraduate degreeMaster’s degreePhDPostdoctorate
Have you ever been exposed to information related to large language models?19.9 (1).001
Yes616280304
No296544412
In what ways have you primarily acquired information about large language models?  38.7 (1)<.001
Academic journal
Yes2102643612
No615670384
Social media
Yes4170685612
No48866184
Lectures
Yes08618146
No81721166010
Friend chatting
Yes2108943814
No615040362
What do you think large language models can be used for in the field of medicine?22.6 (1)<.001
Summarize medical research papers
Yes62161247214
No2421022
Assistance in writing academic papers
Yes21861106016
No67224140
Translation of literature
Yes21721025614
No68632182
What challenges do you believe large language models encounter in their applications within the field of medicine?25.0 (1)<.001
Privacy protection
Yes4198110708
No4602448

aP<.05 was considered statistically significant.

Willingness

The 3 questions in this dimension examined the participants’ attitudes regarding their future studies and the application of or training courses related to LLMs. The responses were recorded using a 5-point Likert scale, ranging from 1 (totally agree) to 5 (totally disagree). Overall, all the participants demonstrated a positive willingness to accept and engage with AI-related learning and training. Specifically, the proportion of participants who chose “completely agree” accounted for nearly half of the total number in all 3 questions. Additionally, the median scores were all relatively low (with lower scores indicating a higher degree of acceptance by the participants). Specific data are present in Figure 1.

Figure 1. Descriptive statistical analysis of the willingness domain. Numbers represent the percentage of participants in each category.

Concerns

We also assessed potential concerns about the integration of LLMs into medical education. First, the majority of participants expressed concerns about ethical issues, with more than half of them choosing to totally agree. However, in the subsequent questions related to medical issues, the number of participants choosing to totally agree largely decreased compared to the ethical questions. Moreover, regarding the issue of whether LLMs would lead to the unemployment of physicians, the vast majority of participants did not express any significant concern, with the median scores being relatively high. Specific data are shown in Figure 2.

Figure 2. Descriptive statistical analysis of the concerns domain. Numbers represent the percentage of participants in each category.

Regression Analysis

After adjusting for gender, age, and education, we observed that both gender and educational level had a significant influence on participants’ awareness and attitudes. Specifically, male participants were more inclined to acquire relevant information through social media (odds ratio [OR] 0.458, 95% CI 0.33‐0.67; P<.001). Furthermore, after adjusting for other variables, male participants (OR 0.476, 95% CI 0.38‐0.82; P=.001) and graduate students (OR 1.552, 95% CI 1.32‐1.77; P=.009) both showed a strong inclination toward using AI for academic writing. Data are presented in Table 5.

Table 5. Binary logistic regression between the categorical variables and different options.
CharacteristicsSocial mediaAssistance in writing academic papers
P valueORa (95% CI)P valueOR (95% CI)
Sex<.001b.001b
MaleRef.cRef.
Female0.458 (0.33‐0.67)0.476 (0.38‐0.82)
Education level.81.009b
Undergraduate degreeRef.Ref.
Graduate degree1.033 (0.95‐1.21)1.552 (1.32‐1.77)

aOR: odds ratio.

bP<.05 was considered statistically significant.

cRef.: reference.


Principal Findings

To summarize the principal findings of this study: Among 490 valid responses from Chinese medical students in Hunan, (1) participants generally held a positive attitude toward LLM integration into medical education, with 45.3% expressing full agreement with learning AI-related knowledge and 46.9% willing to use AI tools; (2) gender and educational level significantly influenced LLM-related perceptions—men were more likely to access LLM information via social media and use LLMs for academic writing, while postgraduates compared with undergraduates showed stronger recognition of the LLM’s academic value; (3) ethical concerns were the most prominent worry, followed by privacy risks, but few participants feared LLM-induced physician unemployment; and (4) most students acquired LLM information through informal channels (social media or friends chatting), with only a few relying on formal medical school courses (academic journals or lectures).

Implications of Findings

Importantly, this study does not claim that students possess “substantial foundational knowledge of AI”—instead, it reports students’ self-reported awareness and use of AI. No conclusions can be drawn about students’ actual mastery of AI foundational knowledge because of the self-reported data limitations. However, for AI course design, the study’s findings highlight specific gaps rather than definitive knowledge levels: (1) informal information channels, such as social media, may lead to an incomplete understanding of AI principles—thus, courses should include foundational modules on AI basics; (2) high ethical concerns suggest a need for modules on AI ethics (eg, case studies of AI-related academic misconduct and guidelines for patient data protection); and (3) the preference for research applications indicates that courses should integrate AI tools for thesis writing, literature review, and data analysis (aligned with Chinese medical education requirements). Future research should use objective assessments to identify specific knowledge gaps and inform the targeted curriculum design.

Comparison to the Literature

This study’s findings align with some conventional knowledge in medical education while revealing region-specific differences.

Findings aligning with existing literature include the following. First, informal AI learning channels: consistent with one study [20], our results show that most Chinese medical students (63.2% and 52.2%) access AI information via nonformal channels (social media or friends chatting) rather than school courses, which aligns with global trends [20,21]. Specifically, 1 study has demonstrated that, on the premise of enhancing transparency and interpretability, leveraging social media as a channel helps expand the audience of medical research and provides opportunities for the early intervention of diseases [22]. However, another study also indicates that social media are associated with risks of data misuse, and its fundamental flaw lies in the lack of procedural standards and extremely low transparency. As an emerging field, AI will further amplify the risks when students acquire relevant information through social media platforms [23]. Second, gender differences in AI attitudes: our findings indicated that male participants were more likely to use AI (68.2% vs 51.2% of female participants), whereas female participants expressed greater concerns about risks (72.3% vs 54.8% of male participants). This result is consistent with surveys from other studies [24-26], which suggest cross-cultural consistency in gender-related AI perceptions. Third, low formal AI education coverage: our data (25.3% of students with formal AI courses) align with a Chinese survey reporting that regional disparities of AI-related instruction exist among medical schools in China, confirming the existence of a national education gap [9].

Novel or region-specific findings are as follows. First, preference for research over clinical AI use: unlike Western studies [27-29], where medical students prioritize AI for clinical tasks, our participants favored research applications (76.3% vs 32.7% for clinical use). This difference could be explained by China’s medical education requirements, which mandate research experience and paper publication for postgraduates [19]. Second, low fear of AI-induced unemployment: only 12.7% of our participants worried about physician unemployment due to AI. This may reflect China’s large health care workforce gap [30], where AI is viewed as a complement rather than a replacement for physicians. Notably, all participants in our study were recruited from large tertiary hospitals located in provincial capitals; therefore, this conclusion may be subject to selection bias. Third, high ethical concerns focused on academic misconduct: while Western studies highlight clinical ethics [31,32], our participants prioritized academic ethics more likely due to strict academic integrity policies in Chinese medical schools [33].

At the policy level, the value of policy guidance in the field of AI applications has gained growing recognition. These policy frameworks not only establish effective norms for AI implementation in practical scenarios but also play a pivotal role in mitigating risks associated with AI deployment [34]. China’s regulatory environment is characterized by centralized management and alignment with national strategic priorities, emphasizing that technological development should serve the overall interests of society, safeguard national security, and uphold social stability and moral norms [11]. This top-down approach facilitates coordinated deployment of AI in health care and education, ensuring consistency with long-term national goals. For example, the “Implementation Opinions on Promoting and Regulating the Application and Development of “AI+Healthcare” (issued by the National Health Commission and 4 other ministries in November 2025) sets targets for 2027 and 2030, explicitly supporting AI application in clinical scenarios [35]. Moreover, the “Reference Guidelines for AI Application Scenarios in the Health Industry” (issued by the National Health Commission in November 2024) defines 84 AI application directions in health care, including medical image analysis and intelligent drug research—providing clear guidance for designing AI education modules aligned with clinical needs [36]. These policies’ emphasis on “application-oriented AI development” aligns with our finding that medical students prioritize AI use for academic and clinical tasks, highlighting the need for education to bridge the gap between policy goals and student practice. In contrast, the regulatory environment in Western countries adopts a more decentralized and issue-specific focus. For example, Europe prioritizes robust protection of personal privacy and data sovereignty, as exemplified by the General Data Protection Regulation and the EU AI Act [37], which classifies medical AI as “high risk” and mandates strict transparency and safety requirements. The United States emphasizes fostering technological innovation and market competition, often relying on industry self-regulation supplemented by targeted federal guidelines [38]. These measures exemplify China’s proactive efforts toward digital transformation in the health care and education sectors, with policy acting as a catalyst for cross-sector collaboration.

Notably, disparities in resource allocation between different regions may give rise to structural differences in the AI literacy levels of medical students [39]. This gap is unlikely to be alleviated in the short term. In Hunan Province, as a relatively underdeveloped province located in the central-southern area of China, there are relatively fewer high-quality AI medical resources compared to economically developed coastal provinces. Although the “Artificial Intelligence+Teacher Team Construction Project” initiated by Hunan Province may promote the balanced development of education within the province [40], this policy is unlikely to be promoted to other provinces. Moreover, the concept that this policy views AI as a tool rather than a decision-making entity may inhibit students’ critical thinking about the ethical risks of AI, thereby increasing the potential risks of its usage. Therefore, as the primary cohort of future users of AI-related applications, their insights are critical to facilitating the effective and contextually relevant adoption of this transformative technology, particularly in aligning its use with the unique demands of medical practice and education. These findings are not only regionally relevant but also have a significant impact on the digital transformation of medical education nationwide. However, when applying these findings to other regions, adjustments must be made based on the differences in the local policy and medical ecosystems.

Limitations

Notably, this study has several limitations. First, cross-sectional studies, although useful for identifying correlations between variables such as the use of LLMs and learning outcomes, are limited in their ability to establish causality. Second, while the sample is representative of the targeted region, its regional specificity may preclude full generalization to the broader population of Chinese medical students. Third, the relatively small sample size of the questionnaire survey may render certain conclusions vulnerable to random variation; expanding the sample size could potentially alter the robustness of these results. Furthermore, selection bias may exist, as medical students with limited interest in the medical applications of LLMs and those who are currently experiencing emotional distress were likely less motivated to participate in the survey. This nonresponse bias may have led to an overrepresentation of positive attitudes toward the use of LLMs in medical contexts. Finally, the study lacks data from two key stakeholders in the medical education ecosystem: practicing physicians and patients. Their perspectives on AI applications in medical practice and education would serve as a valuable complement to the current findings.

Conclusions

In summary, while the momentum toward AI integration in medicine is undeniable, its healthy development relies on coordinated efforts to advance technology responsibly, implement proactive regulation, and prioritize the diverse perspectives of those most impacted by its implementation. This approach will not only address the limitations identified in this study but also lay the groundwork for an AI-enabled medical ecosystem that is equitable, safe, and aligned with the core goals of health care and medical education.

Funding

This work was supported by the Education and Teaching Reform Project of Central South University (2023jy102 to WY; 2023jy077 to CZ); the Degree & Postgraduate Education Reform Project of Central South University (2023JGB010 to WY; 2024JGB141 to CZ); the Degree & Postgraduate Education Reform Project of Hunan Province (2023JGYB019 to WY); and the Natural Science Foundation of Hunan Province (2024JJ5572 to CZ).

Data Availability

The minimal dataset necessary to replicate the study findings is available from the corresponding author upon reasonable request.

Authors' Contributions

WY collected the data and drafted the manuscript. LW and JW performed the literature review and questionnaire design. RY conducted statistical analysis. HPN performed language polishing. CZ conceptualized the study and revised the manuscript.

Conflicts of Interest

None declared.

  1. Cao Y. Using AI and big data analytics to support entrepreneurial decisions in the digital economy. Sci Rep. Oct 22, 2025;15(1):36933. [CrossRef] [Medline]
  2. Roy P, Ghose B, Singh PK, Tyagi PK, Vasudevan A. Artificial intelligence and finance: a bibliometric review on the trends, influences, and research directions. F1000Res. 2025;14:122. [CrossRef] [Medline]
  3. Reddy S. Generative AI in healthcare: an implementation science informed translational path on application, integration and governance. Implement Sci. Mar 15, 2024;19(1):27. [CrossRef] [Medline]
  4. Boscardin CK, Gin B, Golde PB, Hauer KE. ChatGPT and generative artificial intelligence for medical education: potential impact and opportunity. Acad Med. Jan 1, 2024;99(1):22-27. [CrossRef] [Medline]
  5. Dave M, Patel N. Artificial intelligence in healthcare and education. Br Dent J. May 2023;234(10):761-764. [CrossRef] [Medline]
  6. Westerdahl F, Carlson E, Wennick A, Borglin G. Bachelor nursing students´ and their educators´ experiences of teaching strategies targeting critical thinking: A scoping review. Nurse Educ Pract. Aug 2022;63:103409. [CrossRef] [Medline]
  7. Alamrani MH, Alammar KA, Alqahtani SS, Salem OA. Comparing the effects of simulation-based and traditional teaching methods on the critical thinking abilities and self-confidence of nursing students. J Nurs Res. Jun 2018;26(3):152-157. [CrossRef] [Medline]
  8. Tong W, Zhang X, Zeng H, Pan J, Gong C, Zhang H. Reforming China’s secondary vocational medical education: adapting to the challenges and opportunities of the AI era. JMIR Med Educ. Aug 15, 2024;10:e48594. [CrossRef] [Medline]
  9. Hu N, Jiang XQ, Wang YD, et al. Status and perceptions of ChatGPT utilization among medical students: a survey-based study. BMC Med Educ. Jun 4, 2025;25(1):831. [CrossRef] [Medline]
  10. Lee H. Assessment of digital capabilities by 9 countries in the alliance for healthy cities using AI: cross-sectional analysis. JMIR Form Res. Feb 7, 2025;9:e62935. [CrossRef] [Medline]
  11. He J, Baxter SL, Xu J, Xu J, Zhou X, Zhang K. The practical implementation of artificial intelligence technologies in medicine. Nat Med. Jan 2019;25(1):30-36. [CrossRef] [Medline]
  12. World Medical Association. World Medical Association Declaration of Helsinki: ethical principles for medical research involving human subjects. JAMA. Nov 27, 2013;310(20):2191-2194. [CrossRef] [Medline]
  13. Eysenbach G. Improving the quality of Web surveys: the Checklist for Reporting Results of Internet E-Surveys (CHERRIES). J Med Internet Res. Sep 29, 2004;6(3):e34. [CrossRef] [Medline]
  14. Tangadulrat P, Sono S, Tangtrakulwanich B. Using ChatGPT for clinical practice and medical education: cross-sectional survey of medical students’ and physicians’ perceptions. JMIR Med Educ. Dec 22, 2023;9:e50658. [CrossRef] [Medline]
  15. Alkhaaldi SMI, Kassab CH, Dimassi Z, et al. Medical student experiences and perceptions of ChatGPT and artificial intelligence: cross-sectional study. JMIR Med Educ. Dec 22, 2023;9:e51302. [CrossRef] [Medline]
  16. Wenjuanxing. 2025. URL: https://www.wjx.cn/ [Accessed 2025-12-31]
  17. Sarangi PK, Panda BB, P S, Pattanayak D, Panda S, Mondal H. Exploring radiology postgraduate students’ engagement with large language models for educational purposes: a study of knowledge, attitudes, and practices. Indian J Radiol Imaging. Jan 2025;35(1):35-42. [CrossRef] [Medline]
  18. Tao W, Yang J, Qu X. Utilization of, perceptions on, and intention to use AI chatbots among medical students in China: national cross-sectional study. JMIR Med Educ. Oct 28, 2024;10:e57132. [CrossRef] [Medline]
  19. Tie H, Luo Y, Chen D. Thinkings on the reform of medical education system in China. Med Educ Online. Dec 31, 2024;29(1):2302677. [CrossRef] [Medline]
  20. Ravšelj D, Keržič D, Tomaževič N, et al. Higher education students’ perceptions of ChatGPT: a global study of early reactions. PLoS ONE. 2025;20(2):e0315011. [CrossRef] [Medline]
  21. Gupta S, Kattapuram TM, Patel TY. Social media’s role in the perception of radiologists and artificial intelligence. Clin Imaging. Dec 2020;68:158-160. [CrossRef] [Medline]
  22. Hameed S, Nauman M, Akhtar N, Fayyaz MAB, Nawaz R. Explainable AI-driven depression detection from social media using natural language processing and black box machine learning models. Front Artif Intell. 2025;8:1627078. [CrossRef] [Medline]
  23. Kostygina G, Kim Y, Seeskin Z, LeClere F, Emery S. Disclosure standards for social media and generative artificial intelligence research: toward transparency and replicability. Soc Media Soc. 2023;9(4):20563051231216947. [CrossRef] [Medline]
  24. McDermott H, Choudhury N, Lewin-Runacres M, Aemn I, Moss E. Gender differences in understanding and acceptance of robot-assisted surgery. J Robot Surg. Feb 2020;14(1):227-232. [CrossRef] [Medline]
  25. Shalaby R, Vuong W, Hrabok M, et al. Gender differences in satisfaction with a text messaging program (Text4Hope) and anticipated receptivity to technology-based health support during the COVID-19 pandemic: cross-sectional survey study. JMIR Mhealth Uhealth. Apr 15, 2021;9(4):e24184. [CrossRef] [Medline]
  26. Dutta B, Peng MH, Sun SL. Modeling the adoption of personal health record (PHR) among individual: the effect of health-care technology self-efficacy and gender concern. Libyan J Med. Dec 2018;13(1):1500349. [CrossRef] [Medline]
  27. Karimov Z, Allahverdiyev I, Agayarov OY, Demir D, Almuradova E. ChatGPT vs UpToDate: comparative study of usefulness and reliability of Chatbot in common clinical presentations of otorhinolaryngology-head and neck surgery. Eur Arch Otorhinolaryngol. Apr 2024;281(4):2145-2151. [CrossRef] [Medline]
  28. Uche-Anya E, Anyane-Yeboa A, Berzin TM, Ghassemi M, May FP. Artificial intelligence in gastroenterology and hepatology: how to advance clinical practice while ensuring health equity. Gut. Sep 2022;71(9):1909-1915. [CrossRef] [Medline]
  29. Shickel B, Ozrazgat-Baslanti T, Bihorac A. Artificial intelligence at the bedside for the prevention, detection, and management of acute kidney injury. Curr Opin Nephrol Hypertens. Nov 1, 2025;34(6):483-490. [CrossRef] [Medline]
  30. Dong E, Sun X, Xu T, et al. Measuring the inequalities in healthcare resource in facility and workforce: a longitudinal study in China. Front Public Health. 2023;11:1074417. [CrossRef] [Medline]
  31. Badawy W, Zinhom H, Shaban M. Navigating ethical considerations in the use of artificial intelligence for patient care: a systematic review. Int Nurs Rev. Sep 2025;72(3):e13059. [CrossRef] [Medline]
  32. Kleebayoon A, Wiwanitkit V. Ethics for artificial intelligence use in clinical pharmacology. Indian J Pharmacol. May 1, 2024;56(3):224-225. [CrossRef] [Medline]
  33. Jin Y, Wei J, Zhang J, et al. A study on the knowledge, attitude, and practice of research integrity among medical professionals in Ningxia, China. BMC Med Educ. Nov 23, 2024;24(1):1355. [CrossRef] [Medline]
  34. Preiksaitis C, Rose C. Opportunities, challenges, and future directions of generative artificial intelligence in medical education: scoping review. JMIR Med Educ. Oct 20, 2023;9:e48785. [CrossRef] [Medline]
  35. Implementation opinions on promoting and regulating the application and development of “AI + healthcare”. National Health Commission of the People’s Republic of China. 2025. URL: https://www.gov.cn/zhengce/202511/content_7047015.htm [Accessed 2025-12-31]
  36. National Health Commission of the People’s Republic of China reference guidelines for AI application scenarios in the health industry. CNR. 2024. URL: http://www.cnr.cn/newscenter/native/gd/20251110/t20251110_527425853.shtml [Accessed 2025-12-31]
  37. Data protection under GDPR. European Union. URL: https://europa.eu/youreurope/business/dealing-with-customers/data-protection/data-protection-gdpr/ [Accessed 2025-12-31]
  38. Artificial intelligence program: research on AI/ML-based medical devices. US Food and Drug Administration. URL: https:/​/www.​fda.gov/​medical-devices/​medical-device-regulatory-science-research-programs-conducted-osel/​artificial-intelligence-program-research-aiml-based-medical-devices [Accessed 2025-12-31]
  39. Gumilar KE, Indraprasta BR, Hsu YC, et al. Disparities in medical recommendations from AI-based chatbots across different countries/regions. Sci Rep. Jul 24, 2024;14(1):17052. [CrossRef] [Medline]
  40. Changsha city, hunan province, is taking multiple measures to promote the development of digital teachers. Ministry of Education of the People’s Republic of China. 2025. URL: http://www.moe.gov.cn/jyb_xwfb/s6192/s222/moe_1750/202509/t20250916_1413597.html [Accessed 2025-12-31]


AI: artificial intelligence
CHERRIES: Checklist for Reporting Results of Internet E-Surveys
LLM: large language model
OR: odds ratio


Edited by Blake Lesselroth; submitted 11.Sep.2024; peer-reviewed by Omar Ali Aboshady, Omer Allela, Pradosh Kumar Sarangi; accepted 03.Dec.2025; published 09.Mar.2026.

Copyright

© Cheng Zhao, Weiqian Yan, Long Wang, Jing Wu, Herve Pasteur Ndikuriyo, Renhe Yu. Originally published in JMIR Medical Education (https://mededu.jmir.org), 9.Mar.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Education, is properly cited. The complete bibliographic information, a link to the original publication on https://mededu.jmir.org/, as well as this copyright and license information must be included.