Abstract
This paper proposes a framework for leveraging large language models (LLMs) to generate misconceptions as a tool for collaborative learning in health care education. While misconceptions—particularly those generated by AI—are often viewed as detrimental to learning, we present an alternative perspective: that LLM-generated misconceptions, when addressed through structured peer discussion, can promote conceptual change and critical thinking. The paper outlines use cases across health care disciplines, including both clinical and basic science contexts, and a practical 10-step guidance for educators to implement the framework. It also highlights the need for medium- to long-term research to evaluate the impact of LLM-supported learning on student outcomes. This framework may support health care educators globally in integrating emerging AI technologies into their teaching, regardless of the disciplinary focus.
JMIR Med Educ 2026;12:e81875doi:10.2196/81875
Keywords
Introduction
In an age where the availability of information is the norm and not the exception, misinformation from a range of information sources—including social media—can dominate the information landscape. Potential misinformation exists in the medical world, given the increasing rates of knowledge dissemination and the presence of potentially misleading information sources.
Medical practitioners and health professionals are not exempt from its effects, with the consequences having wide-reaching implications. An example of a high-profile case where physicians generated harm due to misinformation includes the physician-authored Great Barrington Declaration during the COVID-19 pandemic, which may have potentially increased resistance to lockdown policies in the United States and Europe []. Other examples include a Houston-based physician tweeting in 2022 to her account of more than 50,000 followers that there were no medically valid reasons for vaccination at the time [], or physician-promoted anti-diet movements supporting the consumption of discretionary foods []. Medical training must therefore explicitly cultivate the capacity to detect and dismantle misinformation.
To address misconceptions stemming from misinformation, examples of current approaches by medical educators include manual assessment of student answers in examinations [] and in-class polling [] to identify misconceptions for use in misconception-based learning. However, these methods assume that (1) various student cohorts over time share similar misconceptions, without potential idiosyncratic variability in thought patterns; (2) educators have a breadth of experience with multiple cohorts; and (3) educators have the time capacity to generate multiple misconceptions in-depth. These assumptions may not hold true in practice, ultimately limiting the scalability of traditional misconception-based learning to educator experience.
The advent of ChatGPT and other large language models (LLMs) fueled by artificial intelligence (AI) presents opportunities to augment misconception-based learning. LLMs have demonstrated potential to personalize learning experiences, offer instant feedback and enhance student engagement []. Within misconception-based learning, LLMs offer the potential to both assist educators in generating misconceptions and students in identifying inaccurate thought patterns through feedback. However, limited research exists on LLM-assisted misconception-based learning in health care education. Therefore, key questions arise: (1) how do LLMs fit within the process of misconception-based learning; and (2) how can LLMs be applied in the context of misconception-based learning in health care education? To address the research gap, this paper serves as a theoretical foundation and a conceptual framework with potential cases to apply LLMs in health care education.
Theoretical Premise of Misconception-Based Learning
Misinformation, and later entrenched misconceptions, can result from cognitive biases [] and the lack of critical thinking []. It is important to note that scholarly debate exists on the use of the word misconception, particularly surrounding its definition and appropriateness of use []. Here misconceptions are referred to in the lens of the Piagetian constructivist theory, which sees misconceptions as a part of active constructions of students’ mental models []. Misconceptions, in this context, are not deficiencies in knowledge but rather a gap between the student’s perceptions and scientifically accurate information. To be more effective, educators could elucidate and challenge misconceptions early in the process of knowledge acquisition.
One of the foundations of health care professions education lies in conceptual acquisition. Conceptual acquisition could range from logic-derived processes (ie, mechanics of inspiration and breathing in physiology) to memory-based internalization (ie, eukaryotic transcription factors in molecular biology). Misconceptions would thus be the students’ prior beliefs and constructed mental models that contradict accepted scientific concepts []. In health care professions, not addressing misconceptions originating from education may have ramifications on clinical reasoning and patient safety [].
presents a simplified model of the process surrounding scientific conceptual acquisition. The assumed educational goal is the possession of student-learned knowledge supported by factual accuracy (output). The student is first given an input of target knowledge either through didactic or autodidactic methods. Given that the target knowledge input in health sciences can be highly conceptual, which has little to no attribution to sensorimotor information [], the student is likely to begin the process of mental abstraction that includes simplification and consolidation of knowledge (labeled (A) in ). Conceptual acquisition could occur in a mechanistic process, with the assembly of smaller and cognitively plausible concepts as a bootstrap for developing larger concepts [].
Ideally, the iterative process of conceptual acquisition is factually accurate, with the student utilizing prior learned knowledge to acquire new concepts, a core component of critical thinking (labeled (B) in ). However, a student is likely to develop misconceptions due to a continuously evolving, plausible intuitive interpretation of the concept according to lived experiences (denoted by the dotted arrow labeled (C) in ) []. Misconceptions can interfere or inhibit with the further process of mental abstraction, given that the student is likely to form a false belief that the concept learned is accurate, resulting in a persistent misconception.
One of the ways to manage misconceptions arising from the process of scientific conceptual acquisition is to pre-emptively expose students to misconceptions in the classroom and systematically debunk such misconceptions. A method to which misconception-based learning can be applied in practice is using refutation text. Refutation text is a misconception-centric pedagogical method first rooted in theories of conceptual accommodation in science []. Conceptual accommodation is the internalization of new concepts upon facing limitations in utilizing their existing conceptions to solve problems []. Conceptual accommodation, therefore, is the goal of refutation text and misconception-based learning.
Three components work together to form the refutation text pedagogical method: (1) stating common but inaccurate knowledge held by the student; (2) explicit indication of the incorrect aspects of said knowledge; and (3) stating factually correct information with supporting explanation and applications []. The advantages of refutation text over traditional text or didactic instruction include explicit mention of assumed prior incorrect beliefs, and increased propensity for conceptual acquisition even for students with low prior topic knowledge []. Given that students may find it challenging to state their own misconceptions held, the educator would be the primary source of providing the refutation text. Misconception exposure and the scientific method have already been in practice within the National University of Singapore common curriculum [].

Integrating LLMs: An LLM-Augmented Misconception-Based Collaborative Learning Approach
There are two potential areas for LLM augmentation: educator-led or student-derived inputs of misconceptions into LLMs (denoted by the parts labeled (D) and (E), respectively, in ). In the educator-led case, LLMs can be used to simulate common student misconceptions. Educators can input a defined set of assessment-relevant concepts and prompt the model to generate plausible errors, providing a basis for targeted instruction. Techniques in prompt engineering include few-shot prompting [], which is the provision of additional examples of desired tasks by the user in follow-up prompts. Given that LLMs predict appropriate outputs based on patterns in training data, providing structured follow-up prompts such as through few-shot prompting can help refine the scope and relevance of these generated misconceptions. For more complex cognitive processes, such as clinical reasoning, advanced strategies like chain-of-thought prompting—which simulates step-by-step heuristic thinking—can be employed to trace and identify potential misconceptions within the reasoning pathways of clinicians-in-training [].
In the student-derived case, LLMs can drive deeper engagement between students and the course material. Prior student examination answers, or reference texts for the concepts, could be supplied to an LLM along with a pre-defined prompt to role-play as a junior trainee. The students could then begin a Socratic-style conversation with the LLM, identifying misconceptions based on the LLM’s questions. Afterwards, the students could collaboratively correct the misconception and explain the correct understanding to the LLM. Such dialogic engagement could encourage metacognitive awareness, sustain curiosity due to the personalized nature of responses, and foster deeper conceptual understanding.
LLM-generated misconceptions complement student misconceptions revealed through assessment responses. Importantly, however, LLM-generated misconceptions may not be universally applicable across all student populations or educational contexts. As such, the lecturer’s role includes adapting LLM-generated content to suit specific curricular needs, positioning the LLM as a quasi-teaching assistant in the generation of discussion prompts.
Overall, this framework in outlines the approach of LLM-augmented, misconception-based collaborative learning (CL). In addition to saving educator time, reducing reliance on educator experience, and identifying idiosyncratic student misconceptions, LLMs can also drive deeper engagement between students and the learning material. Educators could then shift their focus toward fostering higher order thinking skills aligned with the upper levels of Bloom’s Taxonomy, though the extent of this shift will depend on the course’s learning objectives and institutional priorities. Additionally, student-led presentation on evaluating misconceptions generated by AI can integrate into existing pedagogical approaches on the flipped classroom, encouraging student autonomy. Ultimately, the flexibility in this framework’s extent of implementation potentially allows for a variety of use cases.
The Utility of CL in LLM-Assisted Misconception-Based Learning
Refuting misconceptions generated by LLMs could be achieved in CL settings, such as tutorials or workshops. CL emphasizes consensus building and collective problem-solving within a peer group, in contrast to individualistic or competitive approaches. By shifting learners from passive to active modes of engagement, CL promotes critical thinking, deeper conceptual understanding, and the development of social support networks [].
In the context of addressing misconceptions, CL provides a structured environment for students to identify, articulate, and refute conceptual errors. A three-step approach can be used to scaffold this process: (1) students first work individually to pinpoint and justify any conceptual flaws within a statement; (2) small groups are then allocated time to discuss and contrast these ideas with factually accurate knowledge, refining their reasoning through peer interaction; and (3) groups present their consensus and rationale to the larger class, facilitating the synthesis and reinforcement of accurate understanding. After the group presentation, educators could identify strengths and weaknesses in the student groups’ arguments. Additionally, LLMs could provide supplemental feedback alongside educators’ comments on the group’s quality of arguments. CL can thus have positive feedback on student-learned knowledge and conceptual retention (denoted by the part labeled (F) in ).
CL can present a risk of unmoderated reinforcement of misconceptions in LLM-supported environments through distractor information. Distractor information generated by LLMs adds unnecessary extraneous cognitive load to a student’s working memory. This generates a quasi-redundancy effect that involves processing unnecessary or misleading information, possibly reducing learning efficiency [].
However, the presence of CL here can be a crucial first line “guardrail” to the accuracy of knowledge. Through peer discussion, learners can identify and correct hallucination errors, making CL an essential rather than peripheral component of the learning process. CL has shown to be resource efficient in digitally enabled classroom settings when addressing misconceptions []. Educators could then review the understanding of the student group, acting as a second “guardrail” to address problematic misconceptions. Educators could further explore initiatives to scaffold understanding through diagnosis of student understanding and reinforcing understanding within CL groups with learning checkpoints [].
Operational Considerations and Obstacles for Implementation
A key challenge in applying LLMs to misconception-based CL lies in data quality. LLMs learn from data patterns, and their performance depends critically on the quality and reliability of training materials []. Low-quality or inappropriate data can result in hallucinations or the generation of misconceptions that misrepresent core concepts. To mitigate this, domain-specific LLM architectures could be trained to draw exclusively from vetted, purpose-built educational resources such as notes, textbooks, or slides. A cost-efficient alternative for developing domain-specific LLMs is to input relevant educational resources into general-purpose models, such as GPT-4o, prior to prompt generation—though this approach may yield less domain-specific or contextually relevant outputs.
Providing high-quality data for LLM training is especially vital in health care education, where complex competencies such as clinical reasoning and clinical history taking have significant patient safety implications. In these contexts, custom LLM pipelines that utilize high-quality data—including pre-existing manual diagnostic frameworks currently in use by clinicians—are required. However, given present concerns on data security and ethical use, such data must be rigorously vetted for content, suitability, and compliance with privacy regulations before it is used to support learning and assessment. Privacy breaches in AI-assisted learning environments could compromise patient confidentiality and institutional trust in patient data security. Any adoption of LLMs in medical curricula must therefore follow the same rigor as clinical data governance.
Additionally, given that there is an absence of gold standards for clinical reasoning pipelines, it may be challenging in practice to develop an AI model specifically for clinical reasoning. Such pipelines of reasoning could be obtained from experienced clinicians for database construction, which would lead to greater quality of input data for LLMs to generate misconceptions for use in education.
While LLMs can generate customized feedback, pitfalls of LLMs in education include its potential to mislead students with false answers under the pretense of conversational and easy-to-understand responses []. However, a learning opportunity exists amidst the discourse, that is, how might we, as educators (1) utilize such misconceptions as a teaching opportunity, and (2) demonstrate the value of critical thinking and source evaluation in the process? A shift in the paradigm of conceptual acquisition can be achieved by seeing LLMs not as a source of harm, but as a pedagogical opportunity. This is pertinent in the health care education context, as hallucinated misinformation generated by LLMs may reinforce clinically unsafe assumptions if students are not trained to critically verify references.
Recent advances in ChatGPT-4o functionality include a deep research mode, which attempts to provide source references for the content it generates. However, because many publications remain behind paywalls and inaccessible to LLMs, the tool may still produce inaccurate citations or “hallucinated” references despite this feature. This limitation presents an opportunity for educators to help students critically evaluate the credibility of retrieved materials, distinguishing peer-reviewed from non-peer-reviewed sources and identifying instances of fabricated references.
Use Cases for LLM-Assisted Misconception Generation
Educators seeking to use LLMs such as ChatGPT for generating misconceptions can employ role-play within a single prompt, followed by targeted clarification prompts to leverage the few-shot learning capabilities of LLMs. illustrates example use cases across disciplines including pharmacy, nursing, medicine, molecular biology, and microbiology, highlighting the potential application of this approach to both the practice-oriented and basic science domains of health care education. only displays the initial prompt and excludes any subsequent prompts for elaboration based on the individual educator’s expertise and use needs.
| Discipline | Concept(s) | Prompt | Learning goal | ChatGPT-4o output |
| Pharmacy | Hepatic extraction ratio | I am an educator within the pharmacy department of my university. I would like to teach the concept of hepatic extraction ratio. Come up with likely misconceptions that students may possess. |
|
|
| Nursing | High-flow oxygen therapy | I am a nurse educator in my university’s school of medicine. I am trying to convey that not all patients benefit from high-flow oxygen therapy, and that post-operation patients should be given fluid intake. Come up with likely misconceptions that students may have about this concept. |
|
|
| Medicine | Dengue immunity and waning antibody titres | I am a clinician educator in my university hospital. I am delivering a lecture to a group of medical students. I would like to convey the concept of waning immunity and the fact that having dengue once does not guarantee long-term immunity. Provide some common student misconceptions. |
|
|
| Molecular Biology | DNA replication | I am teaching an introductory biology class to freshmen in Pharmacy. I am trying to convey the concept that DNA replication begins at multiple points of origin and DNA Polymerase uses dNTPs for DNA replication. Come up with likely misconceptions that students may have about these two concepts. |
|
|
| Microbiology | Viral spillover effects in humans, and viral evolution | I am an educator within my university’s microbiology and immunology department. I am delivering a lecture on SARS-CoV-2 spillover effects and viral evolution. Generate a list of plausible student misconceptions that I can debunk and explain. |
|
|
aSpO2: oxygen saturation.
bCOPD: chronic obstructive pulmonary disease.
cARDS: acute respiratory distress syndrome.
dNPO: nil per os (nothing by mouth).
edNTP: deoxynucleotide triphosphate.
Of interest are the inherent assumptions made by ChatGPT about a student’s prior knowledge in the subject. For example, in the molecular biology misconception on DNA replication (), ChatGPT assumes that students in an introductory molecular biology class can contrast DNA replication between prokaryotes and eukaryotes, despite the prompt only indicating misconceptions on multiple points of DNA replication. Such an assumption may not hold true in an average cohort. This may indicate a form of metacognitive hindsight bias intrinsic within GPT-related architectures. Educators could then include clarification prompts on the existing knowledge base of the target audience and modify AI-generated misconceptions to suit teaching needs.
In addition, educators can draw students’ attention to instances of hasty generalization within LLM-generated misconceptions such as in . For example, absolute qualifiers such as “all” (nursing prompt), “guarantees” (medical prompt), or “always” (microbiology prompt) imply universal causation with no exceptions. By explicitly highlighting these terms, educators can introduce counterexamples that challenge such overgeneralizations, providing a method for debunking misconceptions.
Finally, educators with prior experience of running past editions of the course in previous years could provide students’ examination answers as an input for LLM analysis, after anonymizing student answer scripts. This could clarify the scope and customize the misconceptions generated to a specific educator’s need. However, given excess specific input, the LLM runs a risk of factoring in irrelevant information, potentially leading to erroneous outputs. Educators could potentially mitigate this by limiting excessive irrelevant information and vetting the input contents to prevent overloading of the LLM’s context window.
A 10-Step Guide for Implementing LLM-Augmented Misconception-Based CL
For educators who intend on implementing the framework in in practice, the following is a suggested 10-step guide with the concept of serotype-specific immunity in used as an example.
Educator-Led Steps
Step 1: Setup
Input lecture slides and/or reference textbook chapters into the LLM. Include specifics of your role (eg, clinician educator) and audience (eg, medical students) in the prompt.
In this step, educators should input lecture slides and/or reference textbook chapters into the LLM. Include specifics of your role (eg, clinician educator) and audience (eg, medical students) in the prompt.
Step 2: Review
Educators review the misconceptions provided. LLMs such as ChatGPT-4o may occasionally provide misconceptions based on content outside the scope of coverage.
Step 3: Tailor
Educators tailor the subsequent prompts to vary the length of misconceptions provided by the LLM. For example, if the concepts covered pertain to the nature of serotype-specific immunity and its relationship to long-term immunity such as antibody-dependent enhancement, a prompt instructing the LLM to integrate these concepts into a single misconception statement may be necessary. The integration of multiple related concepts into one misconception can test students’ attention to detail.
Student-Derived Steps
Step 4: Setup
Upload anonymized student examination answers and lecture notes into an institutional pre-trained LLM chatbot.
Step 5: Role-play
Prompt the chatbot to: (1) assume the role of a junior medical student, (2) engage in a Socratic-style conversational dialogue, (3) ask questions individually and reply to responses with a follow-up question, and (4) provide some misconceptions within its train of thought based on the uploaded student examination answers.
Alternatively, if institutional chatbots are not available, ChatGPT-4o may be used. Only input lecture notes and/or reference textbook chapters and not student examination answers would be used in this case, as sending past student examination answers to current students may potentially infringe upon data privacy.
Step 6: Group
Depending on the cohort size and availability of educators, split the students into groups of 3‐5 to discuss the generated misconceptions for CL purposes. CL should preferably be implemented after students have gone through the content at least once, either through pre-reading or classroom instruction.
Step 7: Present
Student groups should conduct short presentations to identify and explain the LLM-generated misconceptions.
Step 8: Converse
Instruct the students to access the prompted institutional chatbot or ChatGPT-4o and initiate the conversation.
Step 9: Collaborate
Allow the students to work collaboratively in groups to address the questions posed by the LLM and “teach” it the concepts. Students could download the chat history or transcripts after the conclusion of the “lesson.”
Step 10: Assess
Students’ misconceptions could be assessed afterwards by reviewing their responses within the downloaded chat transcripts.
Limitations and Directions for AI-Assisted Pedagogy in Health Care Education Studies
It should be noted that only 2 quantitative studies [,] have been conducted on misconception-based reasoning and the wider misconception-based learning literature within the context of health care education. More than half of these studies are focused on the fields of Newtonian physics, genetically modified organisms, and evolution []. While misconception-based learning was determined to be robust across disciplines independent of specific text design [], generalizing misconception-based learning’s efficacy from other fields to health care education may not be fully applicable. More studies are required to examine the impact of this pedagogical method on student outcomes in health care education.
A common methodological critique of this pedagogical approach is its vulnerability to pre-test priming effects. In typical designs, students complete a pre-test before exposure to the instructional intervention and a post-test afterward. However, evidence suggests that notifying students about an impending test may activate prior knowledge and trigger conceptual change, making it challenging to distinguish between the effects of pre-test priming and those of LLM-assisted, misconception-based CL. To address this, researchers could track individual student responses and apply statistical controls—such as paired t tests or ANCOVA analyses—to adjust for pretest influences and better isolate the intervention’s true impact.
CL applied in LLM-generated misconception generation assumes that there is a form of homogeneity in conceptual understanding within and between groups. In a group comprised only of weaker students, false explanations of the LLM-generated misconceptions may be negatively reinforced instead. While ensuring the distribution of students with varying backgrounds and degrees of conceptual understanding within groups may be ideal to fully capitalize on the benefits of CL, such methods are difficult to implement in practice and are a key limitation of this pedagogical method.
Given the limited number of studies present to examine AI-assisted pedagogy in health care education, multiple avenues for follow-up research studies exist. Few studies have assessed cognitive attitudes and student perspectives on CL using AI []. Importantly, the critical nature of such studies stems from the increasingly interconnected working environment. Studying cognitive attitudes on CL using AI from students’ perspectives could better associate AI technology implementation in the classroom with student outcomes.
Additionally, there has been a lack of medium- to long-term research on student outcomes using AI []. The prevalence of AI technologies and its impact on the educational setting have far-reaching implications for changing pedagogy and student learning. However, since the rise of LLMs to prominence in 2023, studies have largely been cross-sectional, capturing only a snapshot of AI’s influence at a specific point in its rapid development.
Cross-sectional studies are likely only able to capture a snapshot of AI’s impact on student learning at a specific point of time in AI’s development. Such studies often overlook how a student’s learning trajectory evolves alongside advances in AI itself. For example, an educational experience supported by GPT-3 would likely differ substantially from one shaped by GPT-4o, which introduced capabilities such as image generation. Longitudinal, mixed-methods studies are needed to track these evolving dynamics, providing deeper insights into both shifts in student perceptions and longer-term learning outcomes as AI technologies continue to evolve. Such an approach could also assist to identify key factors that support or hinder the long-term implementation of AI in teaching practice.
Yet, it should be noted that the fast-evolving nature of LLMs—which at the time of writing include Gemini 2.5 and the upcoming GPT-5 by OpenAI—could impact the reproducibility of generated misconceptions. With larger token sizes and greater architecture complexity, prompts provided to newer models may not produce the same outputs. Future educational research involving LLMs could ensure the reporting of the specific model version used, sampling conditions, and prompts used for an effort towards greater reproducibility. While the stochastic nature of LLMs carries a risk of low reproducibility of outputs with the same prompt, the role of educators to mitigate such a risk could be to verify and vet the LLM-generated outputs. Thus, LLM-generated misconceptions should serve to complement curriculum design, and the educator should not be removed from the process.
Additionally, greater data privacy risk comes with the evolving dynamics of new LLM models. The extent to which newer models retain data inputs by users through larger context windows are not well understood. Therefore, such risks could be mitigated by educators through anonymization and the removal of original file metadata. On the end of educational technology developers, methods to protect data security such as adversarial attack testing or encryption specific to LLM modalities could be explored [].
Finally, explicit indications of the misconception and subsequent explanations by educators may not always work if the misconception is entrenched in the students’ mental models. Greater emphasis needs to be placed on understanding the root of the falsehood, perhaps through analysis of past lived experiences of students that led to this network of beliefs, to facilitate effective conceptual change. Addressing misconceptions in education would therefore likely not be a one-size-fits-all solution. Supplementing refutation-based learning with other educational approaches such as concept maps and audiovisual representation may improve coherence of concept representation [].
Conclusion – LLM’s Potential for Augmenting Misconception-Based Learning in Health Care Education
Misconceptions, which could stem from naive misinformation, have implications on critical thinking and student learning. Yet, when leveraged intentionally—for example, through LLM-augmented, misconception-based CL—they can catalyze transformative shifts in teaching practice. This approach not only deepens students’ understanding of subject matter but also provides a valuable teaching opportunity for exploring both the benefits and limitations of LLMs.
Additionally, the framework serves as a springboard for future empirical research, generating several potentially testable hypotheses. For example, does pre-bunking with LLM-generated misconceptions elicit stronger propensity for conceptual accommodation as compared to educator-generated refutation text? Does CL around LLM-generated misconceptions improve reasoning accuracy and quality compared to individual learning? Does LLM-generated misconceptions change classroom dynamics for CL? These are important questions for identifying and testing the value of LLMs in misconception-based CL.
However, the true value of LLMs in misconception-based learning exists not to replace educator involvement and expertise. Rather, its purpose is to augment student learning with personalized feedback, synthesize “edge-case” concepts across topics, and reduce the time taken for educators by supplementing misconceptions. The educator’s role is still important in context-setting, the facilitation of CL, and critical review of the AI output. Ultimately, the proposed framework positions LLMs not as an endpoint for teaching and learning, but to foster self-directed, critical, and reflective thinking.
Acknowledgments
We used the generative AI tool ChatGPT by OpenAI to generate examples of use cases in . Additionally, ChatGPT was used to refine the grammatical structure of the main text to enhance clarity and readability. We would like to thank Dr Ronald Boyd Anderson (National University of Singapore) for his kind comments to improve the paper on the concepts surrounding LLMs and reproducibility.
Funding
This work was funded by the National University of Singapore Startup Grant (NUHSRO/2023/018/Startup/10).
Data Availability
No original data were generated or collected for the purpose of this paper.
Authors' Contributions
Conceptualization: BCJC (lead), JHC (supporting), CWT (supporting)
Investigation: BCJC
Methodology: BCJC
Evidence curation: BCJC
Writing – Original draft: BCJC
Writing – Review and editing: BCJC (lead), JHC (supporting), CWT (supporting), SS (supporting)
Supervision: JHC (equal), CWT (equal), SS (supporting)
Funding acquisition: CWT
Conflicts of Interest
None declared.
References
- Tandar CE, Lin JC, Stanford FC. Combating medical misinformation and rebuilding trust in the USA. Lancet Digit Health. Nov 2024;6(11):e773-e774. [CrossRef] [Medline]
- Rubin R. When physicians spread unscientific information about COVID-19. JAMA. Mar 8, 2022;327(10):904. [CrossRef]
- Denniss E, Lindberg R. Social media and the spread of misinformation: infectious and a threat to public health. Health Promot Int. Mar 5, 2025;40(2):daaf023. [CrossRef] [Medline]
- Olde Bekkink M, Donders A, Kooloos JG, de Waal RMW, Ruiter DJ. Uncovering students’ misconceptions by assessment of their written questions. BMC Med Educ. Aug 24, 2016;16(1):221. [CrossRef] [Medline]
- Bordes SJ, Manyevitch R, Huntley JD, Li Y, Murray IVJ. Medical student misconceptions in cardiovascular physiology. Adv Physiol Educ. Jun 1, 2021;45(2):241-249. [CrossRef] [Medline]
- Al-Zahrani AM, Alasmari TM. Exploring the impact of artificial intelligence on higher education: the dynamics of ethical, social, and educational implications. Humanit Soc Sci Commun. Dec 1, 2024;11(1). [CrossRef]
- Britt MA, Rouet JF, Blaum D, Millis K. A reasoned approach to dealing with fake news. Policy Insights Behav Brain Sci. Mar 2019;6(1):94-101. [CrossRef]
- List JA, Ramirez LM, Seither J, Unda J, Vallejo BH. Critical thinking and misinformation vulnerability: experimental evidence from Colombia. PNAS Nexus. Oct 2024;3(10):pgae361. [CrossRef] [Medline]
- Leonard MJ, Kalinowski ST, Andrews TC. Misconceptions yesterday, today, and tomorrow. LSE. Jun 2014;13(2):179-186. [CrossRef]
- Rowell JA, Dawson CJ, Lyndon H. Changing misconceptions: a challenge to science educators. Int J Sci Educ. Apr 1990;12(2):167-175. [CrossRef]
- Boshuizen HPA, Marambe KN. Misconceptions in medicine, their origin and development in education and working life. Int J Educ Res. 2020;100:101536. [CrossRef]
- Södervik I, Mikkilä-Erdmann M, Chi MTH. Conceptual change challenges in medicine during professional development. Int J Educ Res. 2019;98:159-170. [CrossRef]
- Yee E. Abstraction and concepts: when, how, where, what and why? Lang Cogn Neurosci. Nov 26, 2019;34(10):1257-1265. [CrossRef]
- diSessa AA. Toward an epistemology of physics. Cogn Instr. Apr 1993;10(2-3):105-225. [CrossRef]
- Vosniadou S, Ioannides C. From conceptual development to science education: a psychological point of view. Int J Sci Educ. Dec 1998;20(10):1213-1230. [CrossRef]
- Posner GJ, Strike KA, Hewson PW, Gertzog WA. Accommodation of a scientific conception: toward a theory of conceptual change. Sci Educ. Apr 1982;66(2):211-227. [CrossRef]
- Nussbaum J, Novick S. Alternative frameworks, conceptual conflict and accommodation: toward a principled teaching strategy. Instr Sci. Dec 1982;11(3):183-200. [CrossRef]
- Danielson RW, Jacobson NG, Patall EA, et al. The effectiveness of refutation text in confronting scientific misconceptions: a meta-analysis. Educ Psychol. Jan 2, 2025;60(1):23-47. [CrossRef]
- Hynd CR. Refutational texts and the change process. Int J Educ Res. Jan 2001;35(7-8):699-714. [CrossRef]
- National University of Singapore, Faculty of Science. Stay curious, question everything: ‘Citizen scientists’ explore how and why science works. 2024. URL: https://www.science.nus.edu.sg/blog/2024/04/innovative-educators-assoc-prof-ryan-bettens/ [Accessed 2025-07-09]
- Chen B, Zhang Z, Langrené N, Zhu S. Unleashing the potential of prompt engineering for large language models. Patterns (N Y). Jun 13, 2025;6(6):101260. [CrossRef] [Medline]
- Laal M, Ghodsi SM. Benefits of collaborative learning. Procedia - Social and Behavioral Sciences. 2012;31:486-490. [CrossRef]
- Sweller J. Cognition in education. In: Ross BH, Mestre JP, editors. Psychol Learn Motiv. Elsevier; 2011:313. URL: https://shop.elsevier.com/books/cognition-in-education/mestre/978-0-12-387691-1 [Accessed 2026-01-05] ISBN: 9780123876911
- Jansen M, Kim Cavdar I. Adapting team-based learning for medical education: a case study with scalable and resource-efficient implementation. MedSciEduc. 35(2):883-892. [CrossRef]
- van de Pol J, Mercer N, Volman M. Scaffolding student understanding in small-group work: students’ uptake of teacher support in subsequent small-group interaction. Journal of the Learning Sciences. Mar 15, 2019;28(2):206-239. [CrossRef]
- Cheah BCJ, Vicente CR, Chan KR. Machine learning and artificial intelligence for infectious disease surveillance, diagnosis, and prognosis. Viruses. Jun 23, 2025;17(7):882. [CrossRef] [Medline]
- Li R, Wu T. Delving into the practical applications and pitfalls of large language models in medical education: narrative review. Adv Med Educ Pract. 2025;16:625-636. [CrossRef] [Medline]
- Holford D, Mäki KO, Karlsson LC, Lewandowsky S, Gould VC, Soveri A. A randomized controlled trial of empathetic refutational learning with health care professionals. BMC Public Health. Feb 12, 2025;25(1):583. [CrossRef] [Medline]
- Versteeg M, van Loon MH, Wijnen-Meijer M, Steendijk P. Refuting misconceptions in medical physiology. BMC Med Educ. Aug 5, 2020;20(1):250. [CrossRef] [Medline]
- Zengilowski A, Schuetze BA, Nash BL, Schallert DL. A critical review of the refutation text literature: methodological confounds, theoretical problems, and possible solutions. Educ Psychol. Jul 3, 2021;56(3):175-195. [CrossRef]
- Schroeder NL, Kucera AC. Refutation text facilitates learning: a meta-analysis of between-subjects experiments. Educ Psychol Rev. 2022;34(2):957-987. [CrossRef] [Medline]
- Yu JH, Chauhan D, Iqbal RA, Yeoh E. Mapping academic perspectives on AI in education: trends, challenges, and sentiments in educational research (2018–2024). Education Tech Research Dev. Feb 2025;73(1):199-227. [CrossRef]
- Adiguzel T, Kaya MH, Cansu FK. Revolutionizing education with AI: exploring the transformative potential of ChatGPT. CONT ED TECHNOLOGY. 2023;15(3):ep429. [CrossRef]
- Yan B, Li K, Xu M, et al. On protecting the data privacy of large language models (llms): a survey. 2024. Presented at: 2024 International Conference on Meta Computing (ICMC). [CrossRef]
Abbreviations
| AI: artificial intelligence |
| CL: collaborative learning |
| LLM: large language model |
| MBC: misconception-based collaborative |
Edited by Sian Tsuei; submitted 09.Aug.2025; peer-reviewed by Anas Alhur, Jason Wen Yau Lee, Mohd Nihra Haruzuan Mohamad SaidnAff; final revised version received 05.Nov.2025; accepted 06.Nov.2025; published 16.Jan.2026.
Copyright© Brandon C J Cheah, Shefaly Shorey, Jun Hong Ch'ng, Chee Wah Tan. Originally published in JMIR Medical Education (https://mededu.jmir.org), 16.Jan.2026.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Education, is properly cited. The complete bibliographic information, a link to the original publication on https://mededu.jmir.org/, as well as this copyright and license information must be included.

