Original Paper
Abstract
Background: Artificial intelligence (AI) has the potential to revolutionize the way medicine is learned, taught, and practiced, and medical education must prepare learners for these inevitable changes. Academic medicine has, however, been slow to embrace recent AI advances. Since its launch in November 2022, ChatGPT has emerged as a fast and user-friendly large language model that can assist health care professionals, medical educators, students, trainees, and patients. While many studies focus on the technology’s capabilities, potential, and risks, there is a gap in studying the perspective of end users.
Objective: The aim of this study was to gauge the experiences and perspectives of graduating medical students on ChatGPT and AI in their training and future careers.
Methods: A cross-sectional web-based survey of recently graduated medical students was conducted in an international academic medical center between May 5, 2023, and June 13, 2023. Descriptive statistics were used to tabulate variable frequencies.
Results: Of 325 applicants to the residency programs, 265 completed the survey (an 81.5% response rate). The vast majority of respondents denied using ChatGPT in medical school, with 20.4% (n=54) using it to help complete written assessments and only 9.4% using the technology in their clinical work (n=25). More students planned to use it during residency, primarily for exploring new medical topics and research (n=168, 63.4%) and exam preparation (n=151, 57%). Male students were significantly more likely to believe that AI will improve diagnostic accuracy (n=47, 51.7% vs n=69, 39.7%; P=.001), reduce medical error (n=53, 58.2% vs n=71, 40.8%; P=.002), and improve patient care (n=60, 65.9% vs n=95, 54.6%; P=.007). Previous experience with AI was significantly associated with positive AI perception in terms of improving patient care, decreasing medical errors and misdiagnoses, and increasing the accuracy of diagnoses (P=.001, P<.001, P=.008, respectively).
Conclusions: The surveyed medical students had minimal formal and informal experience with AI tools and limited perceptions of the potential uses of AI in health care but had overall positive views of ChatGPT and AI and were optimistic about the future of AI in medical education and health care. Structured curricula and formal policies and guidelines are needed to adequately prepare medical learners for the forthcoming integration of AI in medicine.
doi:10.2196/51302
Keywords
Introduction
Innovation drives health care and health professional education forward. Yet medical education has historically been slow to embrace major change. For example, despite the availability of digital infrastructure and multiple online resources, many medical schools continue to rely on traditional lectures and hands-on experiential learning and have not incorporated the “flipped classroom” model or virtual reality simulations into the curriculum [
]. In recent years, health systems have been challenged by large-scale disruptions, with significant and wide-sweeping impacts on medical education. The COVID-19 pandemic forced an abrupt leap into the virtual learning environment, expediting the widespread use of technology-enhanced learning [ ]. Concomitantly, the pandemic contributed to increased awareness of social and health disparities, spurring the implementation of diversity initiatives and social determinants of health curricula in medical schools and residency programs worldwide [ ]. We are currently on the precipice of another transformational shift in health care and medical education. Artificial intelligence (AI) and AI-based large language models (LLMs), such as ChatGPT, have the potential to revolutionize the way medicine is learned, taught, and practiced.Since its launch in November 2022, ChatGPT has emerged as a fast and user-friendly LLM that can assist health care professionals, medical educators, students/trainees, and patients [
- ]. It is capable of amalgamating and processing large amounts of data and has received passing scores equivalent to a third-year medical student on steps 1 and 2 of the United States Medical Licensing Exam (USMLE) [ ]. Moreover, ChatGPT can be used as a testing tool, providing learners with logical explanations for incorrect responses and allowing them to gain knowledge [ ]. ChatGPT can provide medical students and residents with personalized learning experiences in a safe setting tailored to their learning styles and needs and supported with immediate feedback [ ]. Students and trainees can also have access to readily synthesized evidence-based information that they can use in academic writing [ ] and clinical care and decision-making. This can contribute to better training and, ultimately, improved patient care [ , ].The use of generative AI in medical education is not without controversy. Legal and ethical concerns include bias, copyright and privacy infringements, and overreliance on the technology with potential dehumanization in the learning process [
- ]. ChatGPT has also been found to provide incorrect or fabricated data or “hallucinate,” whereby its generated responses may appear plausible and convincing but are inaccurate or illogical [ - ].The literature on AI and ChatGPT is growing rapidly and is primarily focused on this technology as a transformative innovation and its capabilities, possibilities, and risks. Many studies discuss ChatGPT’s potential to significantly impact teaching and learning, but there is no consensus on how to incorporate it into the medical curriculum. Arguably, transformative innovation goes beyond policy and curricular changes; it disrupts the status quo and challenges the medical education community to question previously held beliefs and practices [
]. As the adoption of technology into medical education progresses, it becomes important to understand medical students’ and residents’ perceptions, concerns, and expectations. This understanding can identify gaps in their knowledge and skills to help educators and policymakers design and implement effective educational interventions tailored to student needs [ ]. Therefore, we conducted a study of recently graduated medical students in an international academic medical center to gauge their experiences and perspectives on the uses of ChatGPT and AI in their medical training and on their future careers in medicine.Methods
Ethical Considerations
We conducted a cross-sectional web-based survey of medical students in the United Arab Emirates. The Sheikh Khalifa Medical City Institutional Review Board approved this study with a waiver of informed consent (RS-804). We used the Checklist for Reporting Results of Internet E-Surveys (CHERRIES) to guide our reporting [
] ( ).Setting and Participants
Participants included medical student applicants to all residency training programs in an academic medical center in the United Arab Emirates. There are currently 2 models of undergraduate medical education in the United Arab Emirates, whereby most medical schools have undergraduate entry (following high school) and a 6-year curriculum and 1 school has postgraduate entry (after a bachelor’s degree) and a 4-year curriculum. Application to residency training is open to graduates from medical schools worldwide. Graduate medical education in the United Arab Emirates is competency based and models the US training structure, with similar resident roles and responsibilities [
].Study Development
The open survey instrument was developed after a comprehensive review of the literature on ChatGPT and AI in health care and medical education and iteratively revised by a panel of 5 medical educators and bioinformatics specialists. The Formsite (Vroman Systems, Inc) survey tool was used. Questions were in English and aimed to understand the students’ formal and informal experiences with ChatGPT and AI in medical school, expectations of using ChatGPT and AI in residency training, and overall perceptions of the impact of AI and LLMs on health care and their professional careers. The instrument was pilot-tested on 15 medical students for length and clarity with only minor changes made based on their comments. These responses were not included in the data analysis. The final version consisted of 42 questions divided into 4 sections (
). Each survey question only allowed for 1 response, which could be freely changed until survey completion and submission. Following basic demographic questions, participants were asked about prior experiences with LLMs and digitally enhanced education, anticipated use of LLMs in residency training, and overall perceptions of AI technology.Data Collection
Between May 5, 2023, and June 13, 2023, an administrator who was not involved in the residency recruitment process invited all medical student applicants on site visits of the hospital and its training programs to scan a QR code that directed them to the web-based survey. Once scanned, the survey could be completed at any time. The first page of the survey provided the description and purpose of the study and explained that it was anonymous and confidential. Participation was voluntary and no incentives were offered. No IP addresses were collected. Consent to participate in the study was indicated by the completion and submission of the survey.
Data Analysis
Data were analyzed using R (version 4.2.2; R Foundation for Statistical Computing). Descriptive statistics were used to tabulate the frequency of the variables. Subgroup analysis was performed to determine the correlation between the demographics and the different variables, and significance was assessed using the chi-square test. Regression analysis was used to determine the association between age, gender, and previous experience with positive perception while controlling confounding variables. P<.05 indicated a significant difference between the variables.
Results
Of 325 applicants to the residency programs, 265 completed the survey (for an 81.5% response rate). The demographic characteristics of the participants are represented in
. The majority of participants (n=174, 65.7%) were women, which is consistent with the gender distribution in the region’s medical schools and residency programs [ ]. Respondents trained in medical schools in multiple countries, but most participants graduated from local medical schools (n=187, 70.6%) and were applying to different medical specialties.Respondents reported minimal incorporation of advanced technology into their medical school curricula (
). The vast majority of respondents also denied using ChatGPT in medical school, with only 20.4% (n=54) using the technology to help complete written assessments and less than 10% (n=25, 9.4%) using ChatGPT to help write patient notes ( ). Despite their limited experience with ChatGPT, more students planned to use it during residency, primarily for exploring new medical topics and research (n=168, 63.4%) and for exam preparation (n=151, 57%). Less than half intended to use ChatGPT to write case reports (n=122, 46%) or research papers (n=127, 47.9%), and fewer than a third of respondents anticipated using ChatGPT for clinical purposes, such as writing patient notes (n=79, 29.8%) or assisting in decision-making (n=73, 27.5%) ( ).Respondents expressed interest in using newer versions of ChatGPT (n=159, 60%) and believed that it would improve their learning (n=141, 53.2%). However, they were more ambivalent about its utility in career progression, with many students expressing uncertainty about AI’s impact on their future opportunities (n=108, 40.8%) and job options (n=85, 32.1%), whereas only 78 participants (n=29.4%) agreed that ChatGPT would expand career opportunities (
). Most respondents were optimistic about AI’s potential and agreed (n=188, 70.9%) that AI will have a major impact on health care during their careers by improving patient care (n=155, 58.5%), though less than half believed that it would improve diagnostic accuracy (n=116, 43.8%) or reduce medical errors (n=124, 46.8%). Although few students frankly disagreed with the positive impact of AI on clinical care, many responses were neutral ( ).Concerning the ethical implications of AI, the students believed that AI could decrease humanism in medicine (n=168, 63.4%) and reduce patient trust in physicians (n=157, 59.2%) (
). The majority (n=163, 61.5%) agreed that medical schools and residency training programs should develop policies to regulate the use of ChatGPT and AI by trainees. Moreover, the vast majority (n=165, 62.3%) recognized that ChatGPT’s answers required verification. When asked if their peers use ChatGPT ethically, 24.2% (n=64) of respondents disagreed and 54.7% (n=145) were unsure ( ).Gender differences in responses were noted. When compared to the female students, the male students were significantly more likely to believe that AI will improve diagnostic accuracy (n=47, 51.7% vs n=69, 39.7%; P=.001), reduce medical errors (n=53, 58.2% vs n=71, 40.8%; P=.002), and improve patient care (n=60, 65.9% vs n=95, 54.6%; P=.007).
After adjusting for gender, we found that there was no significant association between age and perceptions of AI in health care. As for previous ChatGPT experiences, after adjusting for age and gender, prior experience with ChatGPT in medical school was positively correlated with beliefs that AI will improve patient care, decrease medical error and misdiagnosis, and increase the accuracy of diagnoses (P=.001, P<.001, and P=.008, respectively).
Characteristic | Participants, n (%) | ||
Gender | |||
Female | 174 (65.7) | ||
Male | 91 (34.3) | ||
Age (years) | |||
20-24 | 138 (52.1) | ||
25-30 | 110 (41.5) | ||
31-35 | 17 (6.4) | ||
Geographic region of medical school | |||
Africa | 6 (2.3) | ||
Asia | 26 (9.8) | ||
Europe | 12 (4.5) | ||
Middle East or North Africa | 34 (12.8) | ||
United Arab Emirates | 187 (70.6) | ||
Residency choice | |||
Anesthesia | 4 (1.5) | ||
Dermatology | 17 (6.4) | ||
Emergency medicine | 23 (8.7) | ||
Family medicine | 13 (4.9) | ||
Internal medicine or subspecialties | 94 (35.5) | ||
Obstetrics/gynecology | 10 (3.8) | ||
Pediatrics | 40 (15.1) | ||
Psychiatry | 9 (3.4) | ||
Radiology | 16 (6) | ||
Surgery or surgical specialties | 37 (14) | ||
Undeclared | 2 (0.8) |
Survey questions | Participants, n (%) | |
Advanced technology or AI | ||
Digital anatomy | 163 (61.5) | |
High fidelity simulation | 96 (36.2) | |
Virtual dissection | 92 (34.7) | |
AI-generated cases for simulation | 71 (26.8) | |
Computational pathology | 70 (26.4) | |
Tasks for which ChatGPT was used | ||
Complete written assignments | 54 (20.4) | |
Write case reports | 44 (16.6) | |
Write research papers | 42 (15.8) | |
Study or exam preparation | 40 (15.1) | |
Generate case scenarios | 40 (15.1) | |
Suggest research topics or questions | 39 (14.7) | |
Generate questions to test oneself | 35 (13.2) | |
Write patient notes | 25 (9.4) |
Survey questions | Participants, n (%) |
Explore new medical topics or research | 168 (63.4) |
Study or exam preparation | 151 (57) |
Write research papers | 127 (47.9) |
Write case reports | 122 (46) |
Answer medical questions | 115 (43.4) |
Write patient notes | 79 (29.8) |
Clinical decision-making | 73 (27.5) |
Discussion
Principal Findings
In this cross-sectional study of 265 applicants to United Arab Emirates residency programs, most participants had minimal experience with ChatGPT in medical school but had positive perceptions of the technology and planned to use it during residency. Men and students with prior experience with the technology were significantly more likely to have positive views of the technology.
AI Technology in Medical Education
Over the past decade, the use of AI and LLMs in health care has grown substantially in many areas. AI algorithms can provide clinical decision-making support and assist physicians in analyzing medical images, identifying high-risk patients, and recognizing potential drug interactions [
]. LLMs also have the potential to ease the burden of medical documentation by producing first drafts of patient progress notes, result notifications, and medical summaries, thereby saving valuable time that can be spent on patient interactions and tasks involving more advanced knowledge [ ]. As the adoption of AI and LLMs continues to grow in health care, medical education must prepare learners for these developments. Integrating AI early in the medical curriculum will enable future physicians not only to be proficient users of AI tools, but to also take a leading role in steering, evaluating, and overseeing the technology to ensure its appropriate and ethical integration into training and clinical practice [ ].Our study adds to the AI literature by providing the perspectives of medical students—the end users of the technology. Our survey of medical trainees in a developed country in the Middle East shows minimal formal and informal experience with AI tools and limited perceptions of the potential uses of AI in health care. This appears to be a global problem. Our findings are consistent with other studies that show an inconsistent and fragmented approach to teaching AI and digital technology [
]. In a survey of US medical school students, 91.2% either denied or were unsure about their access to AI resources [ ]. Further, a review of AI in undergraduate medical education found substantial variability and limited consensus on how to inculcate AI into the curriculum [ ].In our study, survey respondents displayed positive attitudes about the future of AI in education and medicine. However, less than half of the students anticipated integrating ChatGPT and AI tools into their studying or scientific writing. Previous experience was correlated with positive familiarity and perception, which is also supported by the literature [
]. LLMs such as ChatGPT can provide personalized learning with immediate, individualized feedback that can help trainees identify areas of weakness and improve performance [ ]. These tools can also synthesize concepts from varied resources and can provide feedback on language and writing style [ ]. For non–native English speaking students, this can defray the time and financial burdens of English language editing services and promote diversity and equity in the scientific publication landscape. Without proper guidance, our students may be missing out on these important opportunities.AI Technology in Medical Practice
Notably, only a small fraction of our study participants anticipated integrating AI technology into clinical practice; there was substantial ambiguity about the technology’s potential to improve decision-making. A recent Saudi study also showed that health care workers were highly interested in using AI tools for medical research (69.5%) but less so for decision-making (39.5%) or patient care (44.7%) [
]. Research has shown that LLMs can improve efficiency by completing routine tasks, such as writing discharge summaries or patient instructions [ , ]. Cascella et al [ ] documented ChatGPT’s ability to create a medical note for a patient in the intensive care unit and correctly categorize treatments and test samples. This aligns with the observations made by Javaid et al [ ] that ChatGPT can assist health care professionals with clerical tasks, including report creation and medical record transcription, which can streamline the clinical workflow and free up physician time to focus on patient care. The authors also observed that ChatGPT can be trained to match data from tests, laboratories, vital signs, and symptoms, and then provide recommendations [ ]. Machine learning algorithms have also demonstrated the ability to improve diagnostic precision. In one study, an AI system was more successful than radiologists in interpreting medical images and predicting breast cancer [ ]. Another study showed that AI had an accuracy rate of 71.7% in clinical decision-making [ ]. Similarly, Liu et al [ ] found that AI-generated suggestions could complement clinical decision-support alerts and assist experts in formulating their own recommendations.Survey participants worried that AI would decrease the humanism in medical care. While some authors have argued that AI cannot provide the depth of awareness that human health care professionals have of the intricacies of medical care and the emotional and social circumstances of their patients [
], one study showed that in an online forum, chatbots generated high-quality responses to patient queries that were consistently rated to be more empathetic than physician responses [ ].It is notable that male survey participants had significantly more optimistic views of AI applications in health care than their female colleagues. There are several possible explanations for the observed gender differences in our study. Prior studies have shown that AI can perpetuate racial and gender biases and stereotypes [
]. Other research suggests AI might disproportionately benefit men in some domains, thereby widening the educational gender gap [ ]. In addition, there are potential sociocultural factors, where due to traditional gender roles and societal expectations in the Middle East, men may have been exposed to AI and technology in a way that fosters more positive attitudes and leads to greater comfort and familiarity with these tools [ ]. This can extend into the workforce, where women make up only 30% of those employed in the AI sector according to the 2023 World Economic Forum Global Gender Gap Report [ ]. In response to these challenges, feminist AI has emerged as an approach to ensure that digital technologies are developed and used in ways that are equitable and inclusive [ ]. While it is possible that these concerns contributed to the more tempered enthusiasm among female respondents, further research is needed to fully understand the underlying reasons.Furthermore, we found no significant differences between age groups, likely because our cohort is young and of similar ages. In studies, younger, better educated, and more experienced individuals adopted AI technologies more readily [
]. We are reassured that the medical student respondents expressed some skepticism about the ethical impact of AI. Innovations like AI can have unintended consequences. Trainees can be encouraged to explore these tools under supervision and should be forewarned about potential issues of accuracy, reliability, bias, privacy, and academic integrity [ , ]. It is important for medical educators to reconcile the potential benefits and drawbacks of this disruptive innovation. To do this, the medical education community must develop core competencies in AI, as well as embed AI technology into clinical curricula and practice coupled with clear regulations on its use. Medical students and residents will also need AI ethics training to guide the responsible and equitable use of these technologies [ ]. Some medical educators have already started this process. Suggested AI-related clinical competencies for health care professionals include basic knowledge of AI, social and ethical implications of AI, AI-enhanced clinical encounters that integrate diverse sources of information in creating patient-centered care plans, evidence-based evaluation of AI-based tools, and workflow analysis for AI-based tools [ ]. Developing faculty expertise is an important first step in this process [ ].Limitations
Our study has several limitations. Given the recent launch of ChatGPT, survey respondents had limited experience with this tool during medical school and their clinical rotations. Also, although respondents were from multiple medical schools in several countries, data collection was conducted at 1 hospital, limiting generalizability. Only graduating medical students were surveyed; understanding the experiences and perceptions of all medical trainees and teaching faculty is important. Finally, the cross-sectional design provides a snapshot view and does not capture long-term trends and changes in attitudes or use habits over time.
Conclusions
ChatGPT and AI technology as a whole have the potential to revolutionize medical education and clinical practice. Our study shows that despite limited experience and some ethics concerns, medical students were overall positive and optimistic about the future of AI in medical education and health care but unclear about its role in their own training and careers. Structured curricula and formal policies and guidelines are needed to adequately prepare medical learners for the forthcoming integration of AI in medicine.
Acknowledgments
SMIA, CHK, ZD, MAF, LOA, and HI contributed to conceptualization. All authors contributed to methodology. MAF collected the data. CAH, LOA, and HI contributed to formal analysis and investigation. SMIA, CHK, and LOA drafted the manuscript. HI and ZD revised the manuscript. All authors read and approved the final version.
Conflicts of Interest
None declared.
Checklist for Reporting Results of Internet E-Surveys (CHERRIES).
DOCX File , 19 KBQuestionnaire.
DOCX File , 49 KBReferences
- Chen CH, Mullen AJ. COVID-19 can catalyze the modernization of medical education. JMIR Med Educ. Jun 12, 2020;6(1):e19725. [FREE Full text] [CrossRef] [Medline]
- Rajab MH, Gazal AM, Alkattan K. Challenges to online medical education during the COVID-19 pandemic. Cureus. Jul 02, 2020;12(7):e8966. [FREE Full text] [CrossRef] [Medline]
- Frenk J, Chen LC, Chandran L, Groff EOH, King R, Meleis A, et al. Challenges and opportunities for educating health professionals after the COVID-19 pandemic. Lancet. Oct 29, 2022;400(10362):1539-1556. [FREE Full text] [CrossRef] [Medline]
- Subramani M, Jaleel I, Krishna Mohan S. Evaluating the performance of ChatGPT in medical physiology university examination of phase I MBBS. Adv Physiol Educ. Jun 01, 2023;47(2):270-271. [FREE Full text] [CrossRef] [Medline]
- Khan RA, Jawaid M, Khan AR, Sajjad M. ChatGPT - Reshaping medical education and clinical management. Pak J Med Sci. 2023;39(2):605-607. [FREE Full text] [CrossRef] [Medline]
- Lee H. The rise of ChatGPT: Exploring its potential in medical education. Anat Sci Educ. Mar 14, 2023:1-6. [CrossRef] [Medline]
- Gilson A, Safranek CW, Huang T, Socrates V, Chi L, Taylor RA, et al. How does ChatGPT perform on the United States Medical Licensing Examination? the implications of large language models for medical education and knowledge assessment. JMIR Med Educ. Feb 08, 2023;9:e45312. [FREE Full text] [CrossRef] [Medline]
- Kung TH, Cheatham M, Medenilla A, Sillos C, De Leon L, Elepaño C, et al. Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. PLOS Digit Health. Feb 2023;2(2):e0000198. [FREE Full text] [CrossRef] [Medline]
- Li R, Kumar A, Chen JH. How chatbots and large language model artificial intelligence systems will reshape modern medicine: fountain of creativity or Pandora's box? JAMA Intern Med. Jun 01, 2023;183(6):596-597. [CrossRef] [Medline]
- Liu J, Wang C, Liu S. Utility of ChatGPT in clinical practice. J Med Internet Res. Jun 28, 2023;25:e48568. [FREE Full text] [CrossRef] [Medline]
- How AI and ChatGPT in healthcare elevating the game? Kellton. URL: https://www.kellton.com/kellton-tech-blog/how-ai-and-chatgpt-in-healthcare-elevating-the-game [accessed 2023-11-10]
- Katznelson G, Gerke S. The need for health AI ethics in medical school education. Adv Health Sci Educ Theory Pract. Oct 2021;26(4):1447-1458. [CrossRef] [Medline]
- Mohammad B, Supti T, Alzubaidi M, Shah H, Alam T, Shah Z, et al. The pros and cons of using ChatGPT in medical education: a scoping review. Stud Health Technol Inform. Jun 29, 2023;305:644-647. [CrossRef] [Medline]
- Khairatun Hisan U, Miftahul Amri M. ChatGPT and Medical Education: A Double-Edged Sword. J Pedagog Educ Sci. Mar 11, 2023;2(01):71-89. [CrossRef]
- Bhattacharyya M, Miller VM, Bhattacharyya D, Miller LE. High rates of fabricated and inaccurate references in ChatGPT-generated medical content. Cureus. May 2023;15(5):e39238. [FREE Full text] [CrossRef] [Medline]
- Cascella M, Montomoli J, Bellini V, Bignami E. Evaluating the feasibility of ChatGPT in healthcare: an analysis of multiple clinical and research scenarios. J Med Syst. Mar 04, 2023;47(1):33. [FREE Full text] [CrossRef] [Medline]
- Eysenbach G. The role of ChatGPT, generative language models, and artificial intelligence in medical education: a conversation with ChatGPT and a call for papers. JMIR Med Educ. Mar 06, 2023;9:e46885. [FREE Full text] [CrossRef] [Medline]
- Alkaissi H, McFarlane SI. Artificial hallucinations in ChatGPT: implications in scientific writing. Cureus. Feb 2023;15(2):e35179. [FREE Full text] [CrossRef] [Medline]
- Sukhera J, Fung C, Kulasegaram K. Disruption and dissonance: exploring constructive tensions within research in medical education. Acad Med. Nov 01, 2021;96(11S):S1-S5. [CrossRef] [Medline]
- McLean M, Gibbs TJ. Learner-centred medical education: improved learning or increased stress? Educ Health (Abingdon). Dec 2009;22(3):287. [Medline]
- Eysenbach G. Improving the quality of web surveys: the Checklist for Reporting Results of Internet E-Surveys (CHERRIES). J Med Internet Res. Sep 29, 2004;6(3):e34. [FREE Full text] [CrossRef] [Medline]
- Ibrahim H, Tatari H, Holmboe ES. The transition to competency-based pediatric training in the United Arab Emirates. BMC Med Educ. 2015;15:65. [CrossRef] [Medline]
- Mohamed NA, Abdulhadi NN, Al-Maniri AA, Al-Lawati NR, Al-Qasmi AM. The trend of feminization of doctors' workforce in Oman: is it a phenomenon that could rouse the health system? Hum Resour Health. Apr 27, 2018;16(1):19. [FREE Full text] [CrossRef] [Medline]
- Dorr DA, Adams L, Embí P. Harnessing the promise of artificial intelligence responsibly. JAMA. Apr 25, 2023;329(16):1347-1348. [CrossRef] [Medline]
- Brender TD. Medicine in the era of artificial intelligence: hey chatbot, write me an H & P. JAMA Intern Med. Jun 01, 2023;183(6):507-508. [CrossRef] [Medline]
- Lee J, Wu AS, Li D, Kulasegaram KM. Artificial intelligence in undergraduate medical education: a scoping review. Acad Med. Nov 01, 2021;96(11S):S62-S70. [CrossRef] [Medline]
- McCoy LG, Nagaraj S, Morgado F, Harish V, Das S, Celi LA. What do medical students actually need to know about artificial intelligence? NPJ Digit Med. 2020;3:86. [FREE Full text] [CrossRef] [Medline]
- Liu DS, Sawyer J, Luna A, Aoun J, Wang J, Boachie L, et al. Perceptions of US medical students on artificial intelligence in medicine: mixed methods survey study. JMIR Med Educ. Oct 21, 2022;8(4):e38325. [FREE Full text] [CrossRef] [Medline]
- Elkhodr M, Gide E, Wu R, Darwish O. ICT students' perceptions towards ChatGPT: An experimental reflective lab analysis. STEM Educ. 2023;3(2):70-88. [CrossRef]
- Hooda M, Rana C, Dahiya O, Shet JP, Singh BK. Integrating LA and EDM for improving students success in higher education using FCN algorithm. Math Probl Eng. Mar 29, 2022;2022:1-12. [CrossRef]
- Kim S. Using ChatGPT for language editing in scientific articles. Maxillofac Plast Reconstr Surg. Mar 08, 2023;45(1):13. [FREE Full text] [CrossRef] [Medline]
- Temsah M, Aljamaan F, Malki KH, Alhasan K, Altamimi I, Aljarbou R, et al. ChatGPT and the future of digital health: a study on healthcare workers' perceptions and expectations. Healthcare (Basel). Jun 21, 2023;11(13):1812. [FREE Full text] [CrossRef] [Medline]
- Patel SB, Lam K. ChatGPT: the future of discharge summaries? Lancet Digit Health. Mar 2023;5(3):e107-e108. [FREE Full text] [CrossRef] [Medline]
- Duffourc M, Gerke S. Generative AI in health care and liability risks for physicians and safety concerns for patients. JAMA. Jul 25, 2023;330(4):313-314. [CrossRef] [Medline]
- Javaid M, Haleem A, Singh RP. ChatGPT for healthcare services: An emerging stage for an innovative perspective. BenchCouncil Transactions on Benchmarks, Standards and Evaluations. Feb 2023;3(1):100105. [CrossRef]
- McKinney SM, Sieniek M, Godbole V, Godwin J, Antropova N, Ashrafian H, et al. International evaluation of an AI system for breast cancer screening. Nature. Jan 2020;577(7788):89-94. [CrossRef] [Medline]
- Rao A, Pang M, Kim J, Kamineni M, Lie W, Prasad AK, et al. Assessing the utility of ChatGPT throughout the entire clinical workflow. medrXiv. Preprint posted online February 26, 2023. [FREE Full text] [CrossRef] [Medline]
- Fear K, Gleber C. Shaping the future of older adult care: ChatGPT, advanced AI, and the transformation of clinical practice. JMIR Aging. Sep 13, 2023;6:e51776. [CrossRef] [Medline]
- Ayers JW, Poliak A, Dredze M, Leas EC, Zhu Z, Kelley JB, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. Jun 01, 2023;183(6):589-596. [CrossRef] [Medline]
- Shrestha S, Das S. Exploring gender biases in ML and AI academic research through systematic literature review. Front Artif Intell. 2022;5:976838. [FREE Full text] [CrossRef] [Medline]
- Özdemir V. Digital Is political: why we need a feminist conceptual lens on determinants of digital health. OMICS. Apr 2021;25(4):249-254. [CrossRef] [Medline]
- Gibert K, Valls A. Building a territorial working group to reduce gender gap in the field of artificial intelligence. Applied Sciences. Mar 18, 2022;12(6):3129. [CrossRef]
- Global gender gap report 2023. World Economic Forum. URL: https://www.weforum.org/publications/global-gender-gap-report-2023/ [accessed 2023-11-10]
- The impact of artificial intelligence on the future of workforces in the European Union and the United States of America. The White House. URL: https://www.whitehouse.gov/cea/written-materials/2022/12/05/the-impact-of-artificial-intelligence/ [accessed 2023-11-10]
- Kitamura FC. ChatGPT Is shaping the future of medical writing but still requires human judgment. Radiology. Apr 2023;307(2):e230171. [CrossRef] [Medline]
- Paranjape K, Schinkel M, Nannan Panday R, Car J, Nanayakkara P. Introducing artificial intelligence training in medical education. JMIR Med Educ. Dec 03, 2019;5(2):e16048. [FREE Full text] [CrossRef] [Medline]
- Russell RG, Lovett Novak L, Patel M, Garvey KV, Craig KJT, Jackson GP, et al. Competencies for the use of artificial intelligence-based tools by health care professionals. Acad Med. Mar 01, 2023;98(3):348-356. [CrossRef] [Medline]
Abbreviations
AI: artificial intelligence |
CHERRIES: Checklist for Reporting Results of Internet E-Surveys |
LLM: large language model |
Edited by K Venkatesh; submitted 27.07.23; peer-reviewed by E Vashishtha, A DiGiammarino; comments to author 20.10.23; revised version received 10.11.23; accepted 11.12.23; published 22.12.23.
Copyright©Saif M I Alkhaaldi, Carl H Kassab, Zakia Dimassi, Leen Oyoun Alsoud, Maha Al Fahim, Cynthia Al Hageh, Halah Ibrahim. Originally published in JMIR Medical Education (https://mededu.jmir.org), 22.12.2023.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Education, is properly cited. The complete bibliographic information, a link to the original publication on https://mededu.jmir.org/, as well as this copyright and license information must be included.