<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JME</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Med Educ</journal-id>
      <journal-title>JMIR Medical Education</journal-title>
      <issn pub-type="epub">2369-3762</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v12i1e82988</article-id>
      <article-id pub-id-type="pmid">41911020</article-id>
      <article-id pub-id-type="doi">10.2196/82988</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Susceptibility of Assessment Types to AI-Generated Content in Digital Health and Health Information Management Education: Quasi-Experimental Pilot Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Leung</surname>
            <given-names>Tiffany</given-names>
          </name>
        </contrib>
        <contrib contrib-type="editor">
          <name>
            <surname>Eriksen</surname>
            <given-names>Jeppe</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Li</surname>
            <given-names>Joan</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Azizi</surname>
            <given-names>Amirabbas</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Wani</surname>
            <given-names>Tafheem Ahmad</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>School of Psychology and Public Health</institution>
            <institution>La Trobe University</institution>
            <addr-line>Health Sciences 2</addr-line>
            <addr-line>Plenty Road, Bundoora</addr-line>
            <addr-line>Melbourne, Victoria, 3086</addr-line>
            <country>Australia</country>
            <phone>61 451906170</phone>
            <email>t.wani@latrobe.edu.au</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-9884-0563</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Liem</surname>
            <given-names>Michael</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-0568-3942</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Prasad</surname>
            <given-names>Natasha</given-names>
          </name>
          <degrees>MHIM</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-9395-414X</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Robinson</surname>
            <given-names>Kerin</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-9037-6022</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Nexhip</surname>
            <given-names>Abbey</given-names>
          </name>
          <degrees>BHSc (Hons), BHSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-7000-4985</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author">
          <name name-style="western">
            <surname>Tassos</surname>
            <given-names>Melanie</given-names>
          </name>
          <degrees>MAppSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-2634-8892</ext-link>
        </contrib>
        <contrib id="contrib7" contrib-type="author">
          <name name-style="western">
            <surname>Gjorgioski</surname>
            <given-names>Stephanie</given-names>
          </name>
          <degrees>BHSc (Hons), BHIM</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-1350-1416</ext-link>
        </contrib>
        <contrib id="contrib8" contrib-type="author">
          <name name-style="western">
            <surname>Khan</surname>
            <given-names>Urooj Raza</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-5910-0004</ext-link>
        </contrib>
        <contrib id="contrib9" contrib-type="author">
          <name name-style="western">
            <surname>Boyd</surname>
            <given-names>James</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-8925-8811</ext-link>
        </contrib>
        <contrib id="contrib10" contrib-type="author">
          <name name-style="western">
            <surname>Riley</surname>
            <given-names>Merilyn</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4230-7062</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>School of Psychology and Public Health</institution>
        <institution>La Trobe University</institution>
        <addr-line>Melbourne, Victoria</addr-line>
        <country>Australia</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Tafheem Ahmad Wani <email>t.wani@latrobe.edu.au</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2026</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>30</day>
        <month>3</month>
        <year>2026</year>
      </pub-date>
      <volume>12</volume>
      <elocation-id>e82988</elocation-id>
      <history>
        <date date-type="received">
          <day>26</day>
          <month>8</month>
          <year>2025</year>
        </date>
        <date date-type="rev-request">
          <day>27</day>
          <month>10</month>
          <year>2025</year>
        </date>
        <date date-type="rev-recd">
          <day>21</day>
          <month>1</month>
          <year>2026</year>
        </date>
        <date date-type="accepted">
          <day>31</day>
          <month>1</month>
          <year>2026</year>
        </date>
      </history>
      <copyright-statement>©Tafheem Ahmad Wani, Michael Liem, Natasha Prasad, Kerin Robinson, Abbey Nexhip, Melanie Tassos, Stephanie Gjorgioski, Urooj Raza Khan, James Boyd, Merilyn Riley. Originally published in JMIR Medical Education (https://mededu.jmir.org), 30.03.2026.</copyright-statement>
      <copyright-year>2026</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Education, is properly cited. The complete bibliographic information, a link to the original publication on https://mededu.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://mededu.jmir.org/2026/1/e82988" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Generative artificial intelligence (AI) tools, such as ChatGPT, are increasingly used in higher education and have raised significant concerns about assessment validity and academic integrity. In Digital Health and Health Information Management (DIGHIM) programs, assessments are designed to evaluate a mix of technical skills, contextual reasoning, and professional judgment that underpin medical and health practice. Understanding how generative AI performs across different assessment types is, therefore, critical to identifying which formats are most susceptible to AI-generated content and how assessments may be redesigned to remain authentic and educationally meaningful.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This study aimed to evaluate ChatGPT’s performance across diverse assessment types in DIGHIM education by examining how task complexity influences AI-generated output quality, and develop recommendations for ethical and effective AI integration in assessments.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>A pilot quasi-experimental design compared ChatGPT-generated responses with deidentified student submissions across 5 assessment types: digital health solution design, business case analysis, reflective assessment, SQL health database programming, and a health classification quiz. For each task, multiple AI submissions were produced using different prompting strategies, including rubric integration and the use of ChatGPT (GPT-4 and o1 Preview model). Blinded academic markers evaluated all AI-generated submissions and previously submitted deidentified student assessments against standard rubrics, and descriptive statistics were used to compare performance.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>ChatGPT’s performance varied considerably across assessment types. It achieved its highest accuracy scores in objective, rule-based tasks such as multiple-choice quiz items in health classification (mean 88%, SD 0%) and produced well-structured, coherent responses for reflective assessments (mean 69%, SD 12.8%), though these often lacked personalization and nuanced industry context. In descriptive analytical tasks, such as digital health business cases and solution designs, ChatGPT produced logically structured work with reasonable use of evidence but failed to provide deep contextualization, domain-specific insights, or visual elements expected in DIGHIM practice. Technical assessments revealed the greatest limitations: SQL programming tasks averaged 42% (SD 17.2%) with persistent schema errors, incomplete queries, and weak interpretation of health data outputs, while scenario-based clinical coding scored just 7% (SD 0%), reflecting a lack of precision in applying <italic>ICD-10-AM</italic> (<italic>International Classification of Diseases, Tenth Revision, Australian Modification</italic>) rules and coding conventions. Structured prompting and rubric integration improved results, particularly in descriptive and reflective tasks (up to 80%), but the advanced o1 Preview model did not consistently outperform earlier versions.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>While ChatGPT performs well in structured, rule-based, and reflective tasks, it remains limited in technical accuracy, contextual reasoning, and applied DIGHIM competencies. To support academic integrity and workforce readiness, assessment design should prioritize critical thinking, ethical reasoning, and scenario-based problem-solving aligned with health care practice. Using AI as a tool for critique and refinement, rather than a substitute for student work, may help educators prepare learners for responsible AI use in medical and health professional education.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>academic integrity</kwd>
        <kwd>assessment design</kwd>
        <kwd>ChatGPT performance</kwd>
        <kwd>digital health education</kwd>
        <kwd>generative artificial intelligence</kwd>
        <kwd>health information management</kwd>
        <kwd>quasi-experimental study</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>Artificial intelligence (AI) is a transformative technology that enables systems to generate outputs, such as content, recommendations, or decisions, based on human-defined objectives [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. A key branch, generative AI (GenAI), produces human-like media (text, images, videos, and code) in response to prompts [<xref ref-type="bibr" rid="ref3">3</xref>]. At the forefront are large language models (LLMs) such as ChatGPT (OpenAI), Microsoft Copilot, and Gemini (Google), which use deep learning and vast datasets to understand context, generate coherent responses, and adapt tone and style [<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref8">8</xref>]. These models improve continuously through reinforcement learning with human feedback [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref11">11</xref>].</p>
        <p>In higher education, LLMs have rapidly gained attention among academics and students [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref12">12</xref>-<xref ref-type="bibr" rid="ref14">14</xref>], supporting assessment design, automated marking, and curriculum development [<xref ref-type="bibr" rid="ref14">14</xref>-<xref ref-type="bibr" rid="ref17">17</xref>], improving feedback timeliness and reducing grading errors [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>]. They also help students refine language, generate ideas, and start research inquiries [<xref ref-type="bibr" rid="ref20">20</xref>]. Overall, recent evidence in medical education highlights the rapid diversification of GenAI applications across learning resources, instructional methods, and assessment practices, offering tangible opportunities in areas such as documentation support, simulation-based training, and personalized learning [<xref ref-type="bibr" rid="ref21">21</xref>-<xref ref-type="bibr" rid="ref23">23</xref>].</p>
        <p>Despite the benefits, the integration of GenAI technologies into higher education can present significant ethical challenges [<xref ref-type="bibr" rid="ref24">24</xref>]. These include ChatGPT’s use of inaccurate content (including fictitious reference material) [<xref ref-type="bibr" rid="ref25">25</xref>], concerns regarding human teacher replacement [<xref ref-type="bibr" rid="ref11">11</xref>], and negative repercussions on students’ critical thinking and problem-solving skills [<xref ref-type="bibr" rid="ref26">26</xref>]. Most significantly, the use of ChatGPT can pose a threat to academic integrity and ethics [<xref ref-type="bibr" rid="ref27">27</xref>]. A recent 2024 scoping review has highlighted that traditional assessment methods do not operate effectively in GenAI-facilitated learning environments, prompting the need for innovative and refocused assessment designs that foster career-driven competencies and lifelong learning skills [<xref ref-type="bibr" rid="ref28">28</xref>]. The ethical framework surrounding students’ use of AI in educational assessments primarily involves transparency, accountability, and reliability [<xref ref-type="bibr" rid="ref29">29</xref>]. Transparency requires that students acknowledge the use of AI appropriately in their submissions. Students must also be accountable for the content they submit, as “outputs of AI tools can include biased, inaccurate, or incorrect content that users should be aware of” [<xref ref-type="bibr" rid="ref29">29</xref>]. In some instances, GenAI can fabricate information, which raises the question of the reliability of the outputs [<xref ref-type="bibr" rid="ref7">7</xref>].</p>
        <p>The accessibility of AI tools and the difficulty in detecting AI-generated content [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref16">16</xref>] may tempt students to engage in academic misconduct. This raises concerns about assessment fairness, equity, and the credibility of academic credentials earned through digital platforms. Educators, therefore, face challenges in maintaining the integrity of assessments and ensuring educational quality [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref30">30</xref>]. Recent studies also report learners’ desire for clear institutional guidance on appropriate AI use, aligning with calls for explicit policy and training [<xref ref-type="bibr" rid="ref31">31</xref>,<xref ref-type="bibr" rid="ref32">32</xref>].</p>
        <p>La Trobe University, Australia, offers a wide range of health programs. The Digital Health and Health Information Management (DIGHIM) programs provide formal education in disciplines that support contemporary health care and medical practice, aiming to develop students’ technical, professional, and interpersonal skills [<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref34">34</xref>]. Typical enrolments include undergraduate and master’s-level students from clinical or quasi-clinical backgrounds such as nursing, allied health, health sciences, health administration, biomedical sciences, and related fields, as well as graduates seeking roles at the intersection of health care delivery, health data, and digital systems. The curriculum is aligned with medical education priorities, including clinical documentation, health classification, health data governance, digital health (DH) system design, and the ethical use of health information, all of which directly support contemporary medical practice and health service delivery. The DIGHIM programs emphasize discipline-specific and generic competencies, which are inclusive of problem-solving, critical thinking, and ethical decision-making that are typically assessed through academic essays, reports, multiple-choice quizzes, presentations, case studies, programming exercises, and practical simulations [<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref36">36</xref>]. The proliferation of AI-generated content threatens the validity and reliability of assessments in most of these areas, potentially compromising the effectiveness of evaluating students’ comprehension and competence [<xref ref-type="bibr" rid="ref16">16</xref>]. In this context, the broader medical-education literature now differentiates by task type and competency domain: from automated short-answer scoring and case generation to simulation and reflective work, strengthening the rationale for DIGHIM-specific analysis of assessment susceptibility and redesign [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref37">37</xref>-<xref ref-type="bibr" rid="ref39">39</xref>].</p>
        <p>Previous research on the impact of GenAI tools in higher education has primarily focused on specific types of assessments, such as multiple-choice questions or essay evaluations, mainly within the medical [<xref ref-type="bibr" rid="ref40">40</xref>-<xref ref-type="bibr" rid="ref42">42</xref>] and business fields [<xref ref-type="bibr" rid="ref4">4</xref>]. For example, Chaudhry et al [<xref ref-type="bibr" rid="ref16">16</xref>] undertook an evaluation of AI-generated assessments in a Bachelor of Business Administration, including case analysis, empirical study report, self-reflection, group work, and calculation-based assessments. These findings do not necessarily translate well to the more specialized fields of DH and health information management (HIM), each of which requires a distinct combination of technical, analytical, and decision-making skills. DIGHIM programs, for instance, demand expertise in areas such as health classification, epidemiology and biostatistics, DH solution design, health data governance, and interoperability [<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref34">34</xref>]. These specialist areas of study necessitate a rigorous and diverse set of assessment methods, incorporating scenario-based problem-solving, data interpretation, and system implementation tasks that test both theoretical knowledge and practical application. Given the interdisciplinary nature of DH, which aims to bridge clinical practice, data science, and IT, there is a need for a separate evaluation of how GenAI interacts with these unique assessments [<xref ref-type="bibr" rid="ref43">43</xref>]. This study responds to that gap by examining assessment type susceptibility in DIGHIM and aligning implications with contemporary redesign strategies and governance recommendations emerging from the 2024-2025 literature [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref31">31</xref>,<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref44">44</xref>].</p>
      </sec>
      <sec>
        <title>Aims</title>
        <p>This study examined how GenAI, specifically ChatGPT, performed across different DIGHIM assessment types and how its outputs compared with student work, with the aim of identifying assessment formats most susceptible to AI-generated content. It further sought to inform assessment redesign by identifying approaches that promoted authenticity, higher-order reasoning, and responsible AI use, thereby strengthening academic integrity through design rather than detection.</p>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Study Design</title>
        <p>A quasi-experimental design was piloted to evaluate ChatGPT’s performance on a range of assessment types, comparing its results with submitted assessments of past students. This exploratory approach was intended to provide preliminary insights and inform the design of larger-scale studies. While similar methods have been applied in ChatGPT response evaluation within the education sector, this study pilots the approach in the context of DIGHIM, where it has not yet been examined [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref41">41</xref>].</p>
      </sec>
      <sec>
        <title>Assessment Materials and Markers (Participants)</title>
        <p>Assessments from 5 DIGHIM subjects were purposively selected because of specific assessment characteristics, to maximize diversity and breadth in the type and nature of assessments chosen (<xref ref-type="table" rid="table1">Table 1</xref>). Past students’ deidentified assessments, originally submitted for academic credit, were reevaluated for research purposes as a benchmark against which the ChatGPT-generated responses were compared. Markers, blinded to the origin of each assessment, were responsible for evaluating both the ChatGPT-generated assessments and the past students’ assessments. Markers were academic staff, both full-time and casual, with postgraduate qualifications and advanced professional and teaching experience in DH or HIM relevant to the specific assessments they were assigned, and with demonstrated expertise in evaluating student work using standardized marking rubrics.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Summary of assessment types/sample chosen for the study.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="170"/>
            <col width="90"/>
            <col width="90"/>
            <col width="390"/>
            <col width="140"/>
            <col width="120"/>
            <thead>
              <tr valign="top">
                <td>Assessment type</td>
                <td>Level</td>
                <td>Discipline</td>
                <td>Assessment description</td>
                <td>Experimental (AI<sup>a</sup>) group assessments marked, n</td>
                <td>Control (student) group assessments marked, n</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Digital health solution design</td>
                <td>Masters</td>
                <td>Digital health</td>
                <td>Propose and justify a structured approach to addressing a specific health care challenge through a digital health solution, incorporating design principles, implementation frameworks, and the consideration of barriers and enablers.</td>
                <td>4</td>
                <td>3</td>
              </tr>
              <tr valign="top">
                <td>Digital health business case proposal report</td>
                <td>Masters</td>
                <td>Digital health</td>
                <td>Develop a proposal identifying gaps in a digital health system, analyzing barriers, and recommending evidence-based, innovative solutions.</td>
                <td>4</td>
                <td>3</td>
              </tr>
              <tr valign="top">
                <td>Health information management reflective assessment</td>
                <td>Fourth-year undergraduate</td>
                <td>Health information management</td>
                <td>Reflection on health information management students’ professional practice (work-integrated learning) placement experience and learning</td>
                <td>4</td>
                <td>3</td>
              </tr>
              <tr valign="top">
                <td>Health database SQL programming</td>
                <td>Second-year undergraduate</td>
                <td>Health information management</td>
                <td>Use of SQL to query and analyze a health database, generating reports, and extracting relevant data</td>
                <td>4</td>
                <td>3</td>
              </tr>
              <tr valign="top">
                <td>HIM<sup>c</sup> health classification online quiz examination</td>
                <td>First-year undergraduate</td>
                <td>Health information management</td>
                <td>Online examination comprising long, short, and objective (MCQ<sup>b</sup>, true/false, and 1-word answer) questions, assessing knowledge of health classification systems (<italic>ICD-10-AM</italic><sup>d</sup>) and clinical coding principles</td>
                <td>2</td>
                <td>3</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>AI: artificial intelligence.</p>
            </fn>
            <fn id="table1fn2">
              <p><sup>b</sup>MCQ: multiple-choice question.</p>
            </fn>
            <fn id="table1fn3">
              <p><sup>c</sup>HIM: health information management.</p>
            </fn>
            <fn id="table1fn4">
              <p><sup>d</sup><italic>ICD-10-AM</italic>: International Classification of Diseases, Tenth Revision, Australian Modification.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Sample</title>
        <sec>
          <title>Control Group</title>
          <p>Subject coordinators selected a sample of student assessments (3 per each assessment type) completed between July 1, 2023, and June 30, 2024, to form the control group. To ensure representativeness, assessments were chosen across 3 grade bands: high performance (&#62;80%), medium performance (70%-79%), and low performance (50%-59%). This resulted in the inclusion of a total of 15 student assessments across the 5 subjects and 5 assessment types. All assessments were deidentified prior to analysis to maintain student anonymity.</p>
        </sec>
        <sec>
          <title>Experimental Group</title>
          <p>ChatGPT was used to complete the same assessment tasks as were selected for the control group. A total of 2-4 AI-generated assessments were created for each assessment type (<xref ref-type="table" rid="table1">Table 1</xref>; n=18) to reflect variations in how a typical student might approach the task. The criteria for these variations and their alignment with student input patterns are detailed below in the Experiment section.</p>
          <p>In total, 33 assessments, across the 5 assessment types, were selected for the study: 18 were AI-generated (experimental group), and 15 were past students’ assessments (control group). The 5 assessment types include both undergraduate and postgraduate assessments, reflecting core educational activities that develop competencies directly relevant to medical education, including clinical documentation, health classification, health data analysis, DH system design, and reflective practice in health care settings.</p>
        </sec>
      </sec>
      <sec>
        <title>Experiment</title>
        <sec>
          <title>Overview</title>
          <p>An independent research assistant (RA) followed a structured process, established by the researchers, to gather ChatGPT assessment responses for the chosen sample of assessments (<xref rid="figure1" ref-type="fig">Figure 1</xref>).</p>
          <fig id="figure1" position="float">
            <label>Figure 1</label>
            <caption>
              <p>Overview of methodology.</p>
            </caption>
            <graphic xlink:href="mededu_v12i1e82988_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Generating ChatGPT Responses</title>
          <p>To ensure reliability, completeness, and alignment with academic expectations, multiple submission versions (ie, basic answer generation [V1], section-based refinement [V2], rubric-based refinement [V3], and rubric-based refinement using o1 Preview model [V4]) of assessment instructions were developed for each assessment type for submission to ChatGPT and its versions (<xref ref-type="table" rid="table2">Table 2</xref>). Following the preliminary testing phase described in the Testing and Review subsection below, a refined and detailed protocol was created to guide the research assistant in generating all 4 structured versions (V1-V4; <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>).</p>
          <p>It was expected that full-instruction (V1) and step-by-step approaches (V2 and V3) would allow for a nuanced analysis of AI-generated outputs, with V2 and V3 specifically designed to assess how incremental guidance impacted response quality. Furthermore, inclusion of rubrics (V3) was expected to further improve contextual alignment. The inclusion of o1 Preview (V4) used the o1 model, which was designed for advanced reasoning, and provided insights into GenAI’s evolving capabilities. For SQL tasks, submission versions (V2 and V3) included the database schema and rubric to support contextual accuracy; however, the underlying health dataset itself was not provided to ChatGPT. In contrast, for objective-style questions in the health classification assessment, only 2 submission versions were needed due to the structured format of the questions and the straightforward nature of the required responses. It is important to note that in all versions, ChatGPT was provided with the same assessment instructions and marking rubrics that were originally given to students, ensuring comparability between human and AI-generated work.</p>
          <table-wrap position="float" id="table2">
            <label>Table 2</label>
            <caption>
              <p>Description of artificial intelligence–generated submission versions and prompting strategies.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="30"/>
              <col width="180"/>
              <col width="790"/>
              <thead>
                <tr valign="top">
                  <td colspan="2">Assessment type and submission version</td>
                  <td>Description</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td colspan="3">
                    <bold>Subjective and analytical assessments (reflective assessment, business case proposal/case study, and digital health solution)</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>V1</td>
                  <td>Full assessment instructions entered into ChatGPT using the GPT-4 model for a single-step response generation.</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>V2</td>
                  <td>Assessment broken into parts as per instructions; ChatGPT using the GPT-4 model generated each part separately and compiled.</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>V3</td>
                  <td>Same as V2, with the rubric provided to guide response generation for better alignment with criteria.</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>V4<break/>  <break/>  </td>
                  <td>Same as V3, but generated using the o1 Preview model to evaluate differences in performance.</td>
                </tr>
                <tr valign="top">
                  <td colspan="3">
                    <bold>SQL/programming assessment</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>V1</td>
                  <td>Query instructions provided without a schema for response generation using GPT-4.</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>V2</td>
                  <td>Query instructions with the schema provided to enhance contextual accuracy using GPT-4.</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>V3</td>
                  <td>Schema and rubric provided to ensure responses aligned with evaluation criteria using GPT-4.</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>V4</td>
                  <td>Same as V3, but generated using the o1 Preview model for comparative analysis.</td>
                </tr>
                <tr valign="top">
                  <td colspan="3">
                    <bold>HIM health classification/clinical coding online quiz examination (scenario-based and objective questions)</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>V1</td>
                  <td>Objective questions generated using GPT-4.</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>V2</td>
                  <td>Objective questions generated using the o1 Preview model to compare differences in outputs.</td>
                </tr>
              </tbody>
            </table>
          </table-wrap>
        </sec>
        <sec>
          <title>Blinded Review Preparation</title>
          <p>The research assistant formatted ChatGPT-generated responses to resemble student submissions. These AI-generated responses were then mixed with real, deidentified student assessments to maintain blinding. To ensure an unbiased review process, all assessments were anonymized and assigned coded identifiers.</p>
        </sec>
        <sec>
          <title>Assessment Allocation</title>
          <p>Based on their expertise in the subject, each subject coordinator suggested potential markers to the RA to assess and mark both ChatGPT responses and student assessments. The RA contacted potential markers individually via email to seek their participation and to shortlist 1 marker per assessment type. Markers were provided with participant information consent forms. To maintain impartiality, the nominated markers did not include the academics who originally marked the student assessments. Assessments were randomly assigned to the markers by the RA, acting as an independent coordinator to ensure a fair distribution and prevent any conflicts of interest.</p>
        </sec>
        <sec>
          <title>Marking Process</title>
          <p>Markers followed a standardized marking rubric to evaluate the assessments. To ensure consistency in grading, clear instructions were provided by the principal investigator and communicated by the independent RA. Additionally, support was available throughout the process to address any questions or uncertainties. In addition to numerical scores, markers provided written feedback aligned with the rubric criteria, along with overall free-form comments to justify their grading and offer insights into the quality of the work. To further ensure reliability, if a discrepancy of more than 10 marks was identified in any student assessment (between the original mark and remark), all assessments within that assessment type were independently reviewed by a second subject expert. Any differences were then discussed, and a consensus was reached to finalize the grades.</p>
        </sec>
        <sec>
          <title>Data Tracking and Management</title>
          <p>The RA maintained a Microsoft Excel spreadsheet to track the marking process, including details of assessments sent to staff members and their feedback.</p>
        </sec>
        <sec>
          <title>Testing and Review</title>
          <p>As part of the experimental process, an initial round of testing was undertaken to refine the methodology and confirm the feasibility of the planned research. This involved a smaller sample of assessments across various types (<xref ref-type="table" rid="table2">Table 2</xref>) to trial the generation of AI responses, the collection and deidentification of student work, and the blinded review process (see the sections “Blinded Review Preparation,” “Assessment Allocation,” “Marking Process,” and “Data Tracking and Management” above). Insights from this stage were used to make necessary adjustments prior to the main experiment. The assessments used during this testing phase were excluded from the final data analysis. During preliminary testing, we determined that generating multiple outputs from the same input was not suitable for this study’s design. When repeated prompts were issued within the same ChatGPT account, model-state carryover effects created dependency between responses due to latent conversational memory, thereby compromising independence across outputs. Conversely, when the same inputs were generated using different accounts, the resulting outputs were highly similar in structure, sequencing, and wording. Such similarity posed a risk to the blinded marking process, as a single marker evaluating near-identical responses could readily identify them as AI-generated, introducing potential confirmation bias. For these reasons, we adopted a structured versioning approach (V1-V4) instead of generating multiple outputs for identical inputs.</p>
        </sec>
      </sec>
      <sec>
        <title>Synthesis and Analysis</title>
        <p>Marks and qualitative feedback provided by academic markers were compiled for each assessment type. Descriptive statistics (means, percentages, and score ranges) were calculated across all student grade bands (high, medium, and low) to support the blinded marking process and ensure natural variation within the control group. While ChatGPT outputs were compared against all remarked student assessments, the primary analytical focus was on the highest-graded submissions, as these offered a stable and meaningful benchmark for assessing whether AI-generated work could approximate high-quality human performance. Performance was examined both within individual assessment types and across assessment categories.</p>
        <p>Comparative trend analysis was undertaken to evaluate progression across ChatGPT versions (V1-V4), focusing on improvements associated with structured prompting, schema inclusion, and rubric integration. In addition, rubric-level synthesis was conducted to identify recurring strengths and weaknesses.</p>
        <p>Feedback comments for both student and AI-generated submissions were reviewed and compared, with particular focus on top-performing student work. This comparison helped identify where the evaluative feedback on AI outputs differed most noticeably from that of human-authored work.</p>
        <p>Results were further synthesized into cross-assessment comparisons, with performance grouped into broader task categories (objective, reflective, descriptive analytical, scenario-based analytical, communication/referencing, and programming) to identify where AI demonstrated relative strengths versus persistent limitations.</p>
      </sec>
      <sec>
        <title>GenAI Tools Used</title>
        <p>The study used GPT-4 (free version/limited) and o1 Preview (pro/paid version) for the experimentation. The researchers opted out of ChatGPT’s AI training model using an available feature, ensuring that the submitted assessment guidelines and instructions were not used to further train the AI. All necessary privacy settings were enabled to maintain data confidentiality throughout the research process.</p>
      </sec>
      <sec>
        <title>Ethical Considerations</title>
        <p>The research study was approved by La Trobe University’s Human Research Ethics Committee (application number HEC24286). No student assessment submissions, student-generated content, or identifiable educational data were entered into any GenAI tools as part of this study. Only publicly available assessment instructions and marking rubrics were provided to ChatGPT to generate AI responses. All student work used for comparison was deidentified, stored securely, and assessed offline by academic markers.</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Performance by Assessment Type</title>
        <sec>
          <title>Overview</title>
          <p>This section provides a detailed breakdown of ChatGPT’s performance across each assessment type, with comparisons to the highest-graded student assessments. Comprehensive data analysis and full qualitative feedback for all assessments are provided in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>. This appendix includes comparisons with all types of student submissions. Results are presented by assessment type and task category in the same sequence as outlined in the Methods section, enabling direct comparison between study design and outcomes.</p>
        </sec>
        <sec>
          <title>Assessment Type 1: DH Solution Design</title>
          <p>This assessment required students to design and justify a DH solution using implementation frameworks and a critical analysis of barriers and enablers, assessed against rubric criteria for problem framing, contextual relevance, and professional communication.</p>
          <p>AI-generated responses showed progressive improvements with structured prompting and rubric integration, most notably in GPT-4 (V3), which scored 69% compared to 56% for GPT-4 (V1; <xref ref-type="table" rid="table3">Table 3</xref>). Overall, even with these refinements, AI outputs lacked the depth, specificity, and contextual alignment observed in student submissions. Notably, the use of the more advanced o1 model (V4) did not improve performance over GPT-4 (V3), scoring only 63%, and continued to exhibit generic recommendations and limited tailoring to the scenario. In contrast, the top-performing student submission (81%) demonstrated nuanced justification, visual clarity, and strong alignment with project objectives.</p>
          <table-wrap position="float" id="table3">
            <label>Table 3</label>
            <caption>
              <p>Comparison of outcomes from ChatGPT submission versions for digital health solution assessment.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="230"/>
              <col width="90"/>
              <col width="350"/>
              <col width="330"/>
              <thead>
                <tr valign="top">
                  <td>GPT model and submission version</td>
                  <td>Score (%)</td>
                  <td>Key positive points noted by the marker</td>
                  <td>Key points for improvement noted by the marker</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>GPT-4 (V1<sup>a</sup>)</td>
                  <td>56</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Design stages justification incomplete</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Vague barrier discussion</p>
                      </list-item>
                      <list-item>
                        <p>Unclear language</p>
                      </list-item>
                      <list-item>
                        <p>Misplaced citations</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>GPT-4 (V2<sup>b</sup>)</td>
                  <td>60</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Improved clarity and referencing (than V1)</p>
                      </list-item>
                      <list-item>
                        <p>Better articulation of barriers</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Generic roadmap activities</p>
                      </list-item>
                      <list-item>
                        <p>Limited alignment with goals</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>GPT-4 (V3<sup>c</sup>)</td>
                  <td>69</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Design approach supported with clear evidence</p>
                      </list-item>
                      <list-item>
                        <p>Better understanding of challenges</p>
                      </list-item>
                      <list-item>
                        <p>Strong structure</p>
                      </list-item>
                      <list-item>
                        <p>Clearer communication</p>
                      </list-item>
                      <list-item>
                        <p>Improved referencing</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Activities remained broad</p>
                      </list-item>
                      <list-item>
                        <p>Lacked visual aids and specificity</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>o1 Preview (V4<sup>d</sup>)</td>
                  <td>63</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Consistent framework discussion</p>
                      </list-item>
                      <list-item>
                        <p>Professional tone</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Roadmap remained generic</p>
                      </list-item>
                      <list-item>
                        <p>Limited contextual alignment</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table3fn1">
                <p><sup>a</sup>V1: basic answer generation.</p>
              </fn>
              <fn id="table3fn2">
                <p><sup>b</sup>V2: section-based refinement.</p>
              </fn>
              <fn id="table3fn3">
                <p><sup>c</sup>V3: rubric-based refinement.</p>
              </fn>
              <fn id="table3fn4">
                <p><sup>d</sup>V4: rubric-based refinement using o1 Preview model.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Assessment Type 2: DH Business Case Proposal Report</title>
          <p>This assessment required students to develop a DH business proposal identifying system gaps and recommending innovative technologies, with emphasis on evidence-based justification, analysis of implementation barriers, and professional communication demonstrating health care value and impact.</p>
          <p>ChatGPT-generated submissions yielded similar scores across versions (<xref ref-type="table" rid="table4">Table 4</xref>). GPT-4 (V2) scored the highest (68%), showing improved referencing and clearer structure. However, despite incremental refinements, all AI outputs were critiqued for their generic content, lack of tailored insights, and absence of visual elements such as patient journey maps or system flow diagrams. Notably, the more advanced o1 Preview model (V4) did not improve performance (score: 64%) and continued to present directive rather than strategic framing. In contrast, the top-performing student submission (score: 82%) demonstrated personalized problem-solution alignment, stronger analytical depth, and enhanced clarity through well-integrated visual aids.</p>
          <table-wrap position="float" id="table4">
            <label>Table 4</label>
            <caption>
              <p>Comparison of outcomes from ChatGPT submission versions (V1-V4)—digital health business proposal/case study assessment.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="150"/>
              <col width="110"/>
              <col width="340"/>
              <col width="400"/>
              <thead>
                <tr valign="top">
                  <td>GPT model and submission version</td>
                  <td>Score (%)</td>
                  <td>Key positive points noted by the marker</td>
                  <td>Key points for improvement noted by the marker</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>GPT-4 (V1<sup>a</sup>)</td>
                  <td>62</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Basic identification of system gaps</p>
                      </list-item>
                      <list-item>
                        <p>logical structure</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Weak justification of solutions</p>
                      </list-item>
                      <list-item>
                        <p>insufficient references</p>
                      </list-item>
                      <list-item>
                        <p>vague insights</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>GPT-4 (V2<sup>b</sup>)</td>
                  <td>68</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Clearer articulation of barriers</p>
                      </list-item>
                      <list-item>
                        <p>improved referencing and layout</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Lacked personalization</p>
                      </list-item>
                      <list-item>
                        <p>limited strategic alignment with health care needs</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>GPT-4 (V3<sup>c</sup>)</td>
                  <td>67</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Stronger evidence-based rationale</p>
                      </list-item>
                      <list-item>
                        <p>professional tone</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Solutions not well integrated with patient needs</p>
                      </list-item>
                      <list-item>
                        <p>absence of visual aids such as journey maps</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>o1 Preview (V4<sup>d</sup>)</td>
                  <td>64</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Consistently identified issues and proposed digital interventions</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Lacked strategic framing</p>
                      </list-item>
                      <list-item>
                        <p>weak linkage between problems and proposed actions</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table4fn1">
                <p><sup>a</sup>V1: basic answer generation.</p>
              </fn>
              <fn id="table4fn2">
                <p><sup>b</sup>V2: section-based refinement.</p>
              </fn>
              <fn id="table4fn3">
                <p><sup>c</sup>V3: rubric-based refinement.</p>
              </fn>
              <fn id="table4fn4">
                <p><sup>d</sup>V4: rubric-based refinement using o1 Preview model.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Assessment Type 3: HIM-Reflective Assessment</title>
          <p>This assessment required students to critically reflect on their HIM placement, using empirical evidence to evaluate professional competencies, identify strengths and weaknesses, set actionable development goals, and demonstrate professional communication.</p>
          <p>ChatGPT-generated responses showed a wide range in performance (<xref ref-type="table" rid="table5">Table 5</xref>). GPT-4 (V3) achieved the highest AI score (80%), presenting a well-structured report with detailed reflections and systematic use of empirical evidence, even surpassing the top student marked (70%). However, it exceeded the word limit and lacked visual enhancements, while some goals remained broad and misaligned with competencies expected in the health industry. Earlier versions showed key weaknesses: V1 (53%) was concise but underdeveloped and lacked depth in reflection and linkage to placement experience, and V2 (70%) improved on structure and relevance but included fabricated references and overly clinical goals. The most advanced version, o1 Preview (V4), scored only 55%, meeting word limits but offering shallow insights and repetitive, vague reflections.</p>
          <p>While ChatGPT demonstrated fluency, structure, and effective use of reflective frameworks, it fell short in personalization, contextual relevance, and depth, elements that distinguished the top student submission, which provided clear placement-specific insights, nuanced analysis, and realistic strategies for professional development. This also aided in evaluating whether AI could produce realistic reflective narratives and to inform how such assessments might be strengthened in the future.</p>
          <table-wrap position="float" id="table5">
            <label>Table 5</label>
            <caption>
              <p>Comparison of outcomes from ChatGPT submission versions (V1-V4)—HIM reflective assessment.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="230"/>
              <col width="90"/>
              <col width="350"/>
              <col width="330"/>
              <thead>
                <tr valign="top">
                  <td>GPT model and submission version</td>
                  <td>Score (%)</td>
                  <td>Key positive points noted by the marker</td>
                  <td>Key points for improvement noted by the marker</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>GPT-4 (V1<sup>a</sup>)</td>
                  <td>53</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Structured format</p>
                      </list-item>
                      <list-item>
                        <p>Identified gaps in reflective practice</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Omitted details</p>
                      </list-item>
                      <list-item>
                        <p>Weak placement linkage</p>
                      </list-item>
                      <list-item>
                        <p>Lacked depth and clarity</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>GPT-4 (V2<sup>b</sup>)</td>
                  <td>70</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Strong organization</p>
                      </list-item>
                      <list-item>
                        <p>Good competency linkage, visual structure</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Overly clinical focus</p>
                      </list-item>
                      <list-item>
                        <p>Exceeded word count</p>
                      </list-item>
                      <list-item>
                        <p>Included fabricated reference</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>GPT-4 (V3<sup>c</sup>)</td>
                  <td>80</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Comprehensive, well-referenced</p>
                      </list-item>
                      <list-item>
                        <p>Detailed placement reflection</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Wordy</p>
                      </list-item>
                      <list-item>
                        <p>Lacked visuals</p>
                      </list-item>
                      <list-item>
                        <p>Some goals too broad for HIM context</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>o1 Preview (V4<sup>d</sup>)</td>
                  <td>55</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Concise and grammatically sound</p>
                      </list-item>
                      <list-item>
                        <p>Some relevant evidence</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Shallow reflection</p>
                      </list-item>
                      <list-item>
                        <p>Repetitive content</p>
                      </list-item>
                      <list-item>
                        <p>Weak critical analysis</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table5fn1">
                <p><sup>a</sup>V1: basic answer generation.</p>
              </fn>
              <fn id="table5fn2">
                <p><sup>b</sup>V2: section-based refinement.</p>
              </fn>
              <fn id="table5fn3">
                <p><sup>c</sup>V3: rubric-based refinement.</p>
              </fn>
              <fn id="table5fn4">
                <p><sup>d</sup>V4: rubric-based refinement using o1 Preview model.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Assessment Type 4: Health Database SQL Programming</title>
          <p>This assessment evaluated students’ ability to assess identification of health information needs, write accurate SQL queries to analyze health data, extract relevant information, and interpret results to support health care decision‑making.</p>
          <p>AI-generated responses demonstrated only modest improvement across iterations, with schema and rubric integration resulting in higher scores (<xref ref-type="table" rid="table6">Table 6</xref>). GPT-4 (V1), generated without schema input, scored the lowest (17%), with all queries based on incorrect table and column names. Schema inclusion in V2 improved structural accuracy (47%), but errors in data grouping and missing outputs persisted. V3 (49%) refined query logic slightly, though misspellings and omission of key fields remained problematic. The most advanced model, o1 Preview (V4), achieved the highest AI score (56%), correctly executing about half of the queries and demonstrating improved adherence to SQL conventions, but still suffered from inaccuracies in age group breakdowns, incorrect counts, and missing contextual details such as diagnosis descriptions.</p>
          <p>In contrast, the top student submission (100%) showed precise application of SQL logic, full query completion, and context-aware interpretation of health trends. It demonstrated superior attention to schema structure, data accuracy, and professional communication in presenting findings, areas where ChatGPT consistently underperformed.</p>
          <table-wrap position="float" id="table6">
            <label>Table 6</label>
            <caption>
              <p>Comparison of outcomes from ChatGPT submission versions (V1-V4)—health database SQL programming.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="230"/>
              <col width="90"/>
              <col width="350"/>
              <col width="330"/>
              <thead>
                <tr valign="top">
                  <td>GPT model and submission version</td>
                  <td>Score (%)</td>
                  <td>Key positive points noted by the marker</td>
                  <td>Key points for improvement noted by the marker</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>GPT-4 (V1<sup>a</sup>)</td>
                  <td>17</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Correct use of SQL structure</p>
                      </list-item>
                      <list-item>
                        <p>Basic understanding of syntax</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>All table/column names incorrect</p>
                      </list-item>
                      <list-item>
                        <p>No valid outputs</p>
                      </list-item>
                      <list-item>
                        <p>Missing queries</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>GPT-4 (V2<sup>b</sup>)</td>
                  <td>47</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Some correct queries</p>
                      </list-item>
                      <list-item>
                        <p>Schema improved the structural logic</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Wrong grouping</p>
                      </list-item>
                      <list-item>
                        <p>Spelling errors</p>
                      </list-item>
                      <list-item>
                        <p>Partial outputs</p>
                      </list-item>
                      <list-item>
                        <p>Inconsistent results</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>GPT-4 (V3<sup>c</sup>)</td>
                  <td>49</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Better rubric alignment</p>
                      </list-item>
                      <list-item>
                        <p>Clearer structure</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Inaccurate grouping</p>
                      </list-item>
                      <list-item>
                        <p>Incomplete queries</p>
                      </list-item>
                      <list-item>
                        <p>Diagnosis data omitted</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>o1 Preview (V4<sup>d</sup>)</td>
                  <td>56</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Improved accuracy</p>
                      </list-item>
                      <list-item>
                        <p>Half the queries returned correct results</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Ongoing errors in counts and terminology</p>
                      </list-item>
                      <list-item>
                        <p>Limited interpretation</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table6fn1">
                <p><sup>a</sup>V1: basic answer generation.</p>
              </fn>
              <fn id="table6fn2">
                <p><sup>b</sup>V2: section-based refinement.</p>
              </fn>
              <fn id="table6fn3">
                <p><sup>c</sup>V3: rubric-based refinement.</p>
              </fn>
              <fn id="table6fn4">
                <p><sup>d</sup>V4: rubric-based refinement using o1 Preview model.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Assessment Type 5: HIM Health Classification/Clinical Coding Online Quiz Examination</title>
          <p>This quiz assessed students’ application of <italic>ICD-10-AM</italic> (<italic>International Classification of Diseases, Tenth Revision, Australian Modification</italic>) health classification standards through scenario-based coding, line coding, and objective questions.</p>
          <p>AI responses (GPT-4 [V1] and o1 Preview [V2]) performed best in the objective section (88%) of the quiz but scored only 7% in scenario-based coding tasks, with a 32% overall score, compared to 87% scored by the top student (<xref ref-type="table" rid="table7">Table 7</xref>). Key issues included incorrect block numbers, inaccurate sequencing, and vague or missing code justifications. In contrast, the top student demonstrated precision and contextual understanding aligned with national coding standards. These results reinforce that while AI handles structured recall effectively, it lacks the nuanced reasoning required for applied clinical coding.</p>
          <table-wrap position="float" id="table7">
            <label>Table 7</label>
            <caption>
              <p>Comparison of outcomes from ChatGPT submission versions (V1-V4)—health classification quiz. Only 2 versions were tested, as rubric- and breakdown-based versions were not applicable to this quiz format.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="230"/>
              <col width="90"/>
              <col width="340"/>
              <col width="340"/>
              <thead>
                <tr valign="top">
                  <td>GPT model and submission version</td>
                  <td>Score (%)</td>
                  <td>Key positive points noted by the marker</td>
                  <td>Key points for improvement noted by the marker</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>GPT-4 (V1<sup>a</sup>)</td>
                  <td>32.4</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Strong performance in the objective section (14/16 correct)</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Incorrect block numbers</p>
                      </list-item>
                      <list-item>
                        <p>Poor sequencing</p>
                      </list-item>
                      <list-item>
                        <p>Inaccurate tabular usage</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
                <tr valign="top">
                  <td>o1 Preview (V2<sup>b</sup>)</td>
                  <td>32.4</td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Consistent objective accuracy</p>
                      </list-item>
                      <list-item>
                        <p>Good understanding of conventions</p>
                      </list-item>
                    </list>
                  </td>
                  <td>
                    <list list-type="bullet">
                      <list-item>
                        <p>Weak in scenario coding</p>
                      </list-item>
                      <list-item>
                        <p>Missing procedural codes</p>
                      </list-item>
                      <list-item>
                        <p>Vague codes/justifications</p>
                      </list-item>
                    </list>
                  </td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table7fn1">
                <p><sup>a</sup>V1: basic answer generation.</p>
              </fn>
              <fn id="table7fn2">
                <p><sup>b</sup>V2: section-based refinement using o1 Preview model.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
      </sec>
      <sec>
        <title>Cumulative Performance Comparison Across Assessment Types and Tasks</title>
        <sec>
          <title>Comparative Analysis Across Assessment Types</title>
          <p>An overall performance comparison across different versions of ChatGPT responses, based on average scores, revealed noticeable variations in performance across assessments (<xref rid="figure2" ref-type="fig">Figure 2</xref>).</p>
          <p>Descriptive and reflective assessments consistently achieved higher average scores in ChatGPT-generated responses, compared to technical assessments such as SQL programming and health classification quiz, which received substantially lower average scores. Lower averages in these tasks reflected the complexity of accurately interpreting technical prompts, adhering to database structures, and executing precise scripts. Progression from V1 to V4 in technical tasks demonstrated the importance of context, schema inclusion, and rubrics, which helped improve performance over iterations. However, even the best-performing versions did not match the precision and depth required for these tasks, underlining limitations in handling computational assessments.</p>
          <p>Overall, V3 achieved the highest average scores across assessment types, benefiting from rubric-driven structuring and better alignment with task criteria. V2 showed notable success in descriptive tasks due to its structured approach, while V4 demonstrated improvements in technical assessments through contextual and schema support. In contrast, V1 consistently underperformed across most assessments, reflecting its reliance on single-step, unguided generation, which often lacked depth, accuracy, and alignment with task expectations.</p>
          <fig id="figure2" position="float">
            <label>Figure 2</label>
            <caption>
              <p>Overall performance comparison across assessment types. V1: basic answer generation; V2: section-based refinement; V3: rubric-based refinement; V4: rubric-based refinement using o1 Preview model.</p>
            </caption>
            <graphic xlink:href="mededu_v12i1e82988_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Comparative Analysis of ChatGPT Versions Across Task Types</title>
          <p>Evaluating ChatGPT’s performance across assessment task types, ChatGPT performed best on objective tasks, showing strong accuracy in factual, rule-based questions. Reflective tasks followed, with later versions demonstrating improved reasoning. Descriptive analytical tasks were next in terms of average performance, though gains plateaued in newer versions. Communication and referencing tasks also showed similar performance with a clearer structure in later iterations. Programming tasks improved over time but continued to face challenges in accuracy and contextual understanding. The lowest performance was in complex scenario-based health classification tasks, reflecting current limitations in contextual and analytical reasoning (<xref rid="figure3" ref-type="fig">Figure 3</xref>).</p>
          <fig id="figure3" position="float">
            <label>Figure 3</label>
            <caption>
              <p>Overall performance comparison of ChatGPT versions across assessment task types/rubric criteria. V1: basic answer generation; V2: section-based refinement; V3: rubric-based refinement; V4: rubric-based refinement using o1 Preview model.</p>
            </caption>
            <graphic xlink:href="mededu_v12i1e82988_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Overview</title>
        <p>This research aimed to pilot the evaluation of ChatGPT’s performance across diverse DH and HIM assessment tasks, with a focus on promoting academic integrity. This analysis not only revealed the specific contexts in which ChatGPT excelled or fell short but also suggested broader implications for designing assessments that foster critical thinking, ethical practice, and real-world applicability. By situating ChatGPT’s performance within the study’s original goals, we were able to gain clearer insights into the potential and the limitations of GenAI in academic and professional environments. However, because this pilot examined only the final outputs generated by ChatGPT, it does not evaluate AI’s role as a formative learning partner, such as when students iteratively critique, refine, or codevelop AI-generated drafts.</p>
      </sec>
      <sec>
        <title>Principal Findings and Comparison With Prior Literature</title>
        <p>The DH case study and solution design assessments revealed clear limitations in ChatGPT’s capacity to address complex, context-specific problems. While AI-generated submissions consistently identified foundational frameworks and produced well-structured outputs, they lacked the depth and contextualization required to develop effective, patient-specific solutions. Weak alignment between identified gaps and proposed interventions reflects broader challenges in maintaining contextual coherence. This is consistent with prior findings across disciplines that report similar shortcomings despite fluent language, structural coherence, and high grammatical accuracy [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref44">44</xref>]. The absence of visual artifacts, such as patient journey maps or flowcharts, further constrained the communication of complex ideas, underscoring limitations in GenAI’s ability to produce meaningful visual representations [<xref ref-type="bibr" rid="ref45">45</xref>]. These findings reinforce the essential role of human expertise in refining and contextualizing solutions. For DIGHIM assessments, tasks should prioritize deep contextual engagement, patient-specific scenarios, and dynamic problem-solving. Requiring the integration of real-world data, graphical visualizations, and detailed case-based justification can better assess applied competence while reducing overreliance on AI tools. This approach aligns with constructivist learning theory, which emphasizes learning through authentic engagement rather than surface-level reproduction [<xref ref-type="bibr" rid="ref46">46</xref>,<xref ref-type="bibr" rid="ref47">47</xref>], and with higher-order cognitive processes in Bloom taxonomy that current GenAI systems cannot reliably replicate [<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref49">49</xref>].</p>
        <p>Reflective assessments demonstrated ChatGPT’s strongest performance, with later versions, particularly GPT-4 (V3), achieving the highest score at 80% and surpassing student submissions in structure, coherence, and adherence to reflective frameworks. Similar patterns have been reported in dental education, where ChatGPT performed strongly in structured-portfolio tasks [<xref ref-type="bibr" rid="ref17">17</xref>]. However, despite polished language and systematic organization, AI-generated reflections often lacked genuine depth, contextual nuance, and personalized insight, which are central to HIM professional development. These results indicate a need to redesign reflective assessments to preserve authenticity. This aligns with social cognitive theory, which emphasizes authentic self-reflection, mastery experiences, and the development of self-efficacy as central to professional learning, all of which are elements that GenAI cannot meaningfully replicate [<xref ref-type="bibr" rid="ref50">50</xref>]. Emphasizing personalized insights, unique contextual connections, and critical evaluation of lived experiences, alongside dynamic scenarios, peer interaction, or real-time contextual observations, can reduce AI dependence while reinforcing academic integrity and professional skill development [<xref ref-type="bibr" rid="ref51">51</xref>,<xref ref-type="bibr" rid="ref52">52</xref>].</p>
        <p>Objective assessments, including multiple-choice and true or false questions, showed consistently high AI performance with minimal error rates. This reflects known strengths of AI in structured, rule-based tasks requiring factual recall and logical reasoning and aligns with prior evaluations of GenAI performance on objective assessments [<xref ref-type="bibr" rid="ref53">53</xref>-<xref ref-type="bibr" rid="ref56">56</xref>]. However, this raises concerns regarding the validity of such tasks in measuring individual learning, particularly in online contexts. To maintain assessment integrity, objective questions should be supplemented with higher-order components such as justification, explanation, or reasoning. Embedding real-world context and multistep problem-solving can further enhance authenticity and alignment with learning outcomes while limiting AI dominance [<xref ref-type="bibr" rid="ref57">57</xref>].</p>
        <p>Clinical coding scenario-based assessments revealed a clear performance gap between AI and students, particularly in tasks requiring nuanced contextual interpretation and application of coding rules. While AI performed comparably on objective components, it struggled with complex scenarios, resulting in significantly lower scores than high-performing students. This aligns with US-based findings showing that ChatGPT performs adequately for simple, single-diagnosis coding tasks but struggles with complex patient data [<xref ref-type="bibr" rid="ref58">58</xref>,<xref ref-type="bibr" rid="ref59">59</xref>]. Soroush et al [<xref ref-type="bibr" rid="ref59">59</xref>] reported that GPT-4 achieved only a 34% exact match rate for <italic>ICD-10-CM</italic> (<italic>International Classification of Diseases, Tenth Revision, Clinical Modification</italic>)codes and concluded that LLMs lack a complete internal representation of medical coding rules, rendering them unsuitable for clinical coding tasks. These findings underscore the need for assessments that emphasize contextual reasoning and real-world coding application, reinforcing the importance of developing independent clinical coding proficiency.</p>
        <p>SQL programming assessments highlighted further limitations in ChatGPT’s ability to perform technically precise, context-dependent tasks. Although later versions showed incremental improvement, persistent issues included incorrect table and column usage, overreliance on prompts, limited understanding of health data conventions, and failure to execute complex multistep queries. Errors such as misspellings and the absence of real-time testing and debugging capabilities further undermined output reliability. These findings contrast with prior research reporting strong SQL generation performance by LLMs [<xref ref-type="bibr" rid="ref60">60</xref>], suggesting that health database contexts demand deeper domain knowledge and contextual awareness. Previous work has emphasized the importance of iterative testing and contextual understanding in SQL code generation [<xref ref-type="bibr" rid="ref61">61</xref>]. For educators, this highlights the value of assessments that extend beyond syntax to include debugging, real-world application, and analytical interpretation of health data, thereby fostering job-ready competencies that AI cannot fully replicate.</p>
        <p>In communication, writing, and referencing tasks, ChatGPT demonstrated strong performance in structure, clarity, and adherence to academic conventions, particularly in later versions such as V3. Nonetheless, limitations in depth, contextual relevance, repetition, and critical engagement persisted, especially in longer writing tasks. Similar findings have been reported in business education, where AI-generated responses for extended academic writing scored between 40% and 77% due to insufficient depth and analysis despite high grammatical quality [<xref ref-type="bibr" rid="ref16">16</xref>]. Issues such as hallucinated references and inconsistent citation formatting further undermined credibility, consistent with prior research [<xref ref-type="bibr" rid="ref62">62</xref>,<xref ref-type="bibr" rid="ref63">63</xref>]. These results indicate that while AI can support foundational academic writing, it struggles with nuanced argumentation and original critical thought. Assessment designs should therefore require critical engagement with sources, precise referencing, and personalized application of concepts to ensure authentic student learning.</p>
        <p>Chain prompting, involving iterative refinement of prompts, significantly improved AI performance, particularly in descriptive and reflective tasks, with later versions benefiting from rubric-aligned guidance. Comparable improvements have been observed in pharmacy education, where structured prompting enhanced performance in knowledge recall tasks [<xref ref-type="bibr" rid="ref56">56</xref>]. This demonstrates the effectiveness of multistep prompting in optimizing AI outputs for well-defined assessment formats.</p>
        <p>Overall, o1 Preview (V4) underperformed in descriptive tasks relative to V3, although performance was comparable or improved in simpler objective assessments. This suggests that newer AI versions do not necessarily yield better outcomes for complex, context-driven tasks. While earlier studies reported GPT-4 superiority in medical licensing examinations [<xref ref-type="bibr" rid="ref64">64</xref>] and basic sciences [<xref ref-type="bibr" rid="ref65">65</xref>], such improvements were not consistently observed in this study. There was no clear evidence that o1 Preview enhanced performance in tasks requiring critical thinking or reflection. These findings highlight the importance of assessment designs that prioritize originality, contextual understanding, and higher-order cognition over tasks that can be easily completed using AI assistance.</p>
      </sec>
      <sec>
        <title>Strengths and Limitations</title>
        <p>The findings from this pilot study highlight the importance of designing assessments that align with the evolving roles in HIM and DH while fostering the ethical and practical use of GenAI. Health information managers play a crucial role in health data management, analytics, and health information and communication technology [<xref ref-type="bibr" rid="ref33">33</xref>], and as DH increasingly integrates disciplines such as clinical care, data science, and IT [<xref ref-type="bibr" rid="ref66">66</xref>], it is essential for educational assessments to reinforce interdisciplinary skills that align with the evolving demands of these domains [<xref ref-type="bibr" rid="ref67">67</xref>,<xref ref-type="bibr" rid="ref68">68</xref>]. Recent evidence highlights that traditional assessment methods in higher education are increasingly ineffective in GenAI-facilitated learning environments, necessitating redesigned assessments that promote critical thinking, creativity, and lifelong learning skills [<xref ref-type="bibr" rid="ref28">28</xref>]. For instance, practical redesign approaches, such as “flipped assessment,” can be implemented, where students cocreate and then critically validate AI-generated items under educator supervision, improving confidence while maintaining expert oversight [<xref ref-type="bibr" rid="ref31">31</xref>]. This type of redesign aligns with this study’s recommendation for tasks that encourage students to critique, refine, and contextualize AI outputs rather than merely generate them.</p>
        <p>Assessments should emphasize tasks that require contextual understanding, technical proficiency, and decision-making. Furthermore, these assessments should integrate AI as a tool to augment, rather than a substitute for critical thinking. For example, students could refine AI-generated clinical coding or SQL outputs to align with standards-based frameworks, critically evaluate AI-generated business proposals for depth and contextual relevance, or integrate visualizations such as patient journey maps into health information reports. These approaches not only build essential competencies such as critical analysis, programming, and communication but also encourage students to navigate AI’s limitations and responsibly incorporate its outputs into their work, promoting the ethical use of GenAI to support, rather than replace, student efforts.</p>
        <p>In DIGHIM education, AI use in assessments should be carefully structured to align with professional competencies while maintaining academic integrity. Some tasks should avoid reliance on AI, particularly those requiring personal reflections, ethical decision-making, and critical thinking in complex DH scenarios. Other tasks can adopt AI for efficiency, such as assisting in summarizing health policies, generating structured reports, or organizing data-driven insights. Finally, AI-generated content can be adapted in assessments that require students to critically refine, validate, and contextualize AI outputs, such as evaluating AI-generated clinical coding recommendations, refining SQL queries for health databases, or assessing the applicability of AI-generated DH solutions within industry frameworks. Additionally, educators and students must ensure privacy by refraining from supplying GenAI tools with copyrighted or sensitive health data and should receive structured training on the responsible and ethical use of AI in DH and HIM to support professional readiness.</p>
        <p>Taken together, current 2024-2025 evidence converges on a dual imperative: redesign assessments to cultivate critique, verification, and reflection skills, and establish clear governance mechanisms, including policy guidance, training programs, and disclosure norms, proportionate to the evolving capabilities and risks of GenAI systems [<xref ref-type="bibr" rid="ref31">31</xref>,<xref ref-type="bibr" rid="ref32">32</xref>,<xref ref-type="bibr" rid="ref38">38</xref>].</p>
        <p>These approaches prepare students for real-world applications of GenAI in HIM and DH roles, where such tools may be used to assist with technical tasks such as data analysis, report generation, and strategic planning. By designing assessments that both test and build skills in these areas while fostering responsible AI use, educators can equip graduates to navigate the interdisciplinary and technology-driven landscape of modern DH care effectively.</p>
        <p><xref ref-type="table" rid="table8">Table 8</xref> provides a summary of important recommendations.</p>
        <p>This study had several limitations. First, as a pilot study, the findings are preliminary and intended to provide directional insights rather than definitive conclusions; however, the use of multiple assessment types and structured AI submission versions was intended to strengthen internal comparison. Larger-scale studies are needed to confirm these patterns. Second, the quasi-experimental design lacked random assignment, which may limit generalizability. This was partially mitigated by comparing AI-generated submissions with authentic, previously graded student work under blinded conditions. Future studies could strengthen external validity through multiinstitutional or randomized designs. Third, assessments were marked by discipline-specific academic staff, and individual grading variability may have influenced results despite mitigation strategies, including standardized rubrics, blinded review, and second-marker consensus when remarked scores differed by more than 10 marks. Future work could further enhance reliability by incorporating multiple independent markers per assessment. Fourth, the number of assessment samples within each category was limited, which may reduce the robustness of task-specific conclusions. This was addressed by selecting assessments across multiple levels and disciplines, though larger samples are required in future studies. Fifth, the study could not fully replicate real-time student interactions with GenAI, such as iterative prompting or blended human-AI workflows. The structured versioning approach (V1-V4) was used to approximate common usage patterns, but longitudinal studies are needed to capture authentic use over time. Sixth, while all student submissions underwent formal academic integrity checks and no concerns were identified, it was not possible to determine with absolute certainty whether any student work involved AI assistance, which may have influenced comparisons. Finally, the study did not systematically assess the detectability of AI-generated content. Although markers occasionally inferred AI authorship, these judgments were informal and sometimes inaccurate, consistent with prior reports of false positives [<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref69">69</xref>] and highlighting the unreliability of human detection methods.</p>
        <p>Future research should explicitly investigate the accuracy of markers in distinguishing AI- from student-generated work and consider how assessment design can emphasize originality, critical thinking, and personalized insights to reduce reliance on detection and strengthen academic integrity. Another area that future studies could explore is expanding the scope beyond ChatGPT to include other advanced LLMs (eg, Gemini, Claude, and AI-assisted coding tools such as GitHub Copilot), to allow broader comparison of capabilities and provide a more comprehensive understanding of how different AI systems perform across assessment types in DIGHIM.</p>
        <table-wrap position="float" id="table8">
          <label>Table 8</label>
          <caption>
            <p>Recommendations for ethical assessment design in Digital Health and Health Information Management in the artificial intelligence (AI) era.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="150"/>
            <col width="340"/>
            <col width="510"/>
            <thead>
              <tr valign="top">
                <td>Theme</td>
                <td>Key findings</td>
                <td>Recommendations</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Contextual and higher-order assessment design</td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>AI struggled with context-specific, patient-centered, and judgement-based tasks.</p>
                    </list-item>
                    <list-item>
                      <p>Underperformance in descriptive, analytical, and coding scenario questions.</p>
                    </list-item>
                  </list>
                </td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>Use assessments requiring deep contextualization, scenario complexity, and justification of decisions.</p>
                    </list-item>
                    <list-item>
                      <p>Emphasize ethical reasoning, critique, and higher-order application rather than factual recall.</p>
                    </list-item>
                  </list>
                </td>
              </tr>
              <tr valign="top">
                <td>Reflective and authentic assessments</td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>AI produced a strong structure but fabricated plausible experiences.</p>
                    </list-item>
                    <list-item>
                      <p>Risk to authenticity in reflective tasks.</p>
                    </list-item>
                  </list>
                </td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>Require personalized and context-verified reflections (eg, workplace events, supervisor-signed logs).</p>
                    </list-item>
                    <list-item>
                      <p>Use scenario-based reflections, peer engagement, or real-time journaling to ensure authenticity.</p>
                    </list-item>
                  </list>
                </td>
              </tr>
              <tr valign="top">
                <td>Objective and procedural tasks</td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>AI performed strongly on MCQs<sup>a</sup>, true/false, and structured objective items.</p>
                    </list-item>
                    <list-item>
                      <p>Risk of undermining originality.</p>
                    </list-item>
                  </list>
                </td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>Add reasoning, explanation, or linked vignettes to objective tasks.</p>
                    </list-item>
                    <list-item>
                      <p>Combine correctness with justification to assess understanding.</p>
                    </list-item>
                  </list>
                </td>
              </tr>
              <tr valign="top">
                <td>Technical and programming tasks (SQL/coding)</td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>AI struggled with schema use, debugging, and domain conventions.</p>
                    </list-item>
                    <list-item>
                      <p>Frequent inaccuracies despite rubric/schema input.</p>
                    </list-item>
                  </list>
                </td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>Use live, hands-on coding tasks requiring testing, debugging, and scenario-based logic.</p>
                    </list-item>
                    <list-item>
                      <p>Assess understanding of health-data structures and domain-specific conventions.</p>
                    </list-item>
                  </list>
                </td>
              </tr>
              <tr valign="top">
                <td>Academic writing and referencing</td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>Strong structure but limited depth and contextual argumentation</p>
                    </list-item>
                  </list>
                </td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>Require deeper analysis, original reasoning, and evidence validation</p>
                    </list-item>
                    <list-item>
                      <p>Include tasks where students examine or verify AI-generated claims.</p>
                    </list-item>
                  </list>
                </td>
              </tr>
              <tr valign="top">
                <td>Ethics, equity, privacy, and governance</td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>Prompting skill differences create inequities</p>
                    </list-item>
                    <list-item>
                      <p>AI detection unreliable</p>
                    </list-item>
                    <list-item>
                      <p>Privacy and rubric transparency concerns.</p>
                    </list-item>
                  </list>
                </td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>Provide standardized AI literacy training</p>
                    </list-item>
                    <list-item>
                      <p>Avoid overreliance on detection tools; design for originality and personal insight</p>
                    </list-item>
                    <list-item>
                      <p>Train students not to enter sensitive data into AI tools.</p>
                    </list-item>
                    <list-item>
                      <p>Use rubrics that guide learning without enabling “AI gaming.”</p>
                    </list-item>
                    <list-item>
                      <p>Develop institutional policies for ethical AI use.</p>
                    </list-item>
                  </list>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table8fn1">
              <p><sup>a</sup>MCQ: multiple-choice question.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>This pilot study has highlighted the potential and limitations of ChatGPT in HIM and DH assessments. While GenAI demonstrates strengths in structured tasks and foundational content generation, it struggles with contextualization, depth, and the technical precision required for assessments requiring independent critical thinking. These findings emphasize the need for carefully designed assessments that integrate AI ethically and prioritize tasks that require human judgment and contextual understanding to ensure meaningful learning outcomes and academic integrity.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Protocol for research assistant.</p>
        <media xlink:href="mededu_v12i1e82988_app1.docx" xlink:title="DOCX File , 356 KB"/>
      </supplementary-material>
      <supplementary-material id="app2">
        <label>Multimedia Appendix 2</label>
        <p>Detailed performance analysis by assessment type.</p>
        <media xlink:href="mededu_v12i1e82988_app2.xlsx" xlink:title="XLSX File  (Microsoft Excel File), 84 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AI</term>
          <def>
            <p>artificial intelligence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">DH</term>
          <def>
            <p>digital health</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">DIGHIM</term>
          <def>
            <p>Digital Health and Health Information Management</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">GenAI</term>
          <def>
            <p>generative artificial intelligence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">HIM</term>
          <def>
            <p>health information management</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">ICD-10-AM</term>
          <def>
            <p>International Classification of Diseases, Tenth Revision, Australian Modification</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">ICD-10-CM</term>
          <def>
            <p>International Classification of Diseases, Tenth Revision, Clinical Modification</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">LLM</term>
          <def>
            <p>large language model</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">RA</term>
          <def>
            <p>research assistant</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">V1</term>
          <def>
            <p>basic answer generation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">V2</term>
          <def>
            <p>section-based refinement</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">V3</term>
          <def>
            <p>rubric-based refinement</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">V4</term>
          <def>
            <p>rubric-based refinement using o1 Preview model</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>The authors sincerely thank Sheikh Hanan Bashir, Saja Sammour, and Kay Bonello for their valuable contributions to this study. The authors also acknowledge the staff of the Digital Health and Health Information Management cluster, the participating markers, and the academic leadership of the School of Psychology and Public Health at La Trobe University for their support.</p>
      <p>Generative artificial intelligence tools such as Microsoft Copilot were used for limited language editing and refinement of the manuscript. All content was reviewed, verified, and finalized by the authors, who take full responsibility for the accuracy, originality, and integrity of the work.</p>
    </ack>
    <notes>
      <sec>
        <title>Funding</title>
        <p>The research was funded through La Trobe University’s internal artificial intelligence in teaching grant scheme. The funder had no role in the study design; data collection, analysis, or interpretation; manuscript preparation; or the decision to submit the manuscript for publication.</p>
      </sec>
    </notes>
    <notes>
      <sec>
        <title>Data Availability</title>
        <p>Summary data and detailed analyses supporting this study are provided in the manuscript and <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>. Access to underlying deidentified student assessment data is restricted due to ethical and privacy considerations; however, reasonable requests for additional deidentified data may be considered by the corresponding author, subject to institutional ethics approval.</p>
      </sec>
    </notes>
    <fn-group>
      <fn fn-type="con">
        <p>Conceptualization: TAW, ML, NP, MR</p>
        <p>Data curation: TAW, ML, NP, AN, MR</p>
        <p>Formal analysis: TAW, KR, AN, MT, SG, URK, JB, MR</p>
        <p>Investigation: TAW, ML, NP, KR, AN, MT, SG</p>
        <p>Methodology: TAW, ML, NP, MR</p>
        <p>Validation: TAW, ML, MT, SG, URK, JB, MR</p>
        <p>Project administration: TAW, ML, NP, MR</p>
        <p>Resources: TAW, ML, NP, MR</p>
        <p>Supervision: TAW, KR, MR</p>
        <p>Visualization: TAW</p>
        <p>Writing—original draft: TAW</p>
        <p>Writing—review &#38; editing: ML, NP, KR, AN, MT, SG, URK, JB, MR</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Makridakis</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>The forthcoming artificial intelligence (AI) revolution: its impact on society and firms</article-title>
          <source>Futures</source>
          <year>2017</year>
          <volume>90</volume>
          <fpage>46</fpage>
          <lpage>60</lpage>
          <pub-id pub-id-type="doi">10.1016/j.futures.2017.03.006</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="web">
          <article-title>ISO/IEC 22989:2022(en), Information technology -- Artificial intelligence concepts and terminology</article-title>
          <source>ISO</source>
          <year>2022</year>
          <access-date>2025-05-16</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.iso.org/obp/ui/#iso:std:iso-iec:22989:ed-1:v1:en">https://www.iso.org/obp/ui/#iso:std:iso-iec:22989:ed-1:v1:en</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Marr</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>The 4 types Of generative AI transforming our world</article-title>
          <source>Forbes</source>
          <access-date>2025-05-16</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.forbes.com/sites/bernardmarr/2024/04/29/the-4-types-of-generative-ai-transforming-our-world/">https://www.forbes.com/sites/bernardmarr/2024/04/29/the-4-types-of-generative-ai-transforming-our-world/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hasanein</surname>
              <given-names>AM</given-names>
            </name>
            <name name-style="western">
              <surname>Sobaih</surname>
              <given-names>AEE</given-names>
            </name>
          </person-group>
          <article-title>Drivers and consequences of ChatGPT use in higher education: key stakeholder perspectives</article-title>
          <source>Eur J Investig Health Psychol Educ</source>
          <year>2023</year>
          <volume>13</volume>
          <issue>11</issue>
          <fpage>2599</fpage>
          <lpage>2614</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=ejihpe13110181"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/ejihpe13110181</pub-id>
          <pub-id pub-id-type="medline">37998071</pub-id>
          <pub-id pub-id-type="pii">ejihpe13110181</pub-id>
          <pub-id pub-id-type="pmcid">PMC10670526</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kasneci</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Sessler</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Küchemann</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Bannert</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Dementieva</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Fischer</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Gasser</surname>
              <given-names>U</given-names>
            </name>
            <name name-style="western">
              <surname>Groh</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Günnemann</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hüllermeier</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Krusche</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kutyniok</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Michaeli</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Nerdel</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Pfeffer</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Poquet</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Sailer</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Schmidt</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Seidel</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Stadler</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Weller</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kuhn</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kasneci</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT for good? On opportunities and challenges of large language models for education</article-title>
          <source>Learn Individ Differ</source>
          <year>2023</year>
          <volume>103</volume>
          <fpage>102274</fpage>
          <pub-id pub-id-type="doi">10.1016/j.lindif.2023.102274</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nazi</surname>
              <given-names>ZA</given-names>
            </name>
            <name name-style="western">
              <surname>Peng</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>Large language models in healthcare and medical domain: a review</article-title>
          <source>Informatics</source>
          <year>2024</year>
          <volume>11</volume>
          <issue>3</issue>
          <fpage>57</fpage>
          <pub-id pub-id-type="doi">10.3390/informatics11030057</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gao</surname>
              <given-names>CA</given-names>
            </name>
            <name name-style="western">
              <surname>Howard</surname>
              <given-names>FM</given-names>
            </name>
            <name name-style="western">
              <surname>Markov</surname>
              <given-names>NS</given-names>
            </name>
            <name name-style="western">
              <surname>Dyer</surname>
              <given-names>EC</given-names>
            </name>
            <name name-style="western">
              <surname>Ramesh</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Pearson</surname>
              <given-names>AT</given-names>
            </name>
          </person-group>
          <article-title>Comparing scientific abstracts generated by ChatGPT to real abstracts with detectors and blinded human reviewers</article-title>
          <source>NPJ Digit Med</source>
          <year>2023</year>
          <volume>6</volume>
          <issue>1</issue>
          <fpage>75</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-023-00819-6"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-023-00819-6</pub-id>
          <pub-id pub-id-type="medline">37100871</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-023-00819-6</pub-id>
          <pub-id pub-id-type="pmcid">PMC10133283</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lund</surname>
              <given-names>BD</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Mannuru</surname>
              <given-names>NR</given-names>
            </name>
            <name name-style="western">
              <surname>Nie</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Shimray</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT and a new academic reality: artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing</article-title>
          <source>Asso Info Science Tech</source>
          <year>2023</year>
          <volume>74</volume>
          <issue>5</issue>
          <fpage>570</fpage>
          <lpage>581</lpage>
          <pub-id pub-id-type="doi">10.1002/asi.24750</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>Why and how to embrace AI such as ChatGPT in your academic life</article-title>
          <source>R Soc Open Sci</source>
          <year>2023</year>
          <volume>10</volume>
          <issue>8</issue>
          <fpage>230658</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/37621662"/>
          </comment>
          <pub-id pub-id-type="doi">10.1098/rsos.230658</pub-id>
          <pub-id pub-id-type="medline">37621662</pub-id>
          <pub-id pub-id-type="pii">rsos230658</pub-id>
          <pub-id pub-id-type="pmcid">PMC10445029</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Deng</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>The benefits and challenges of ChatGPT: an overview</article-title>
          <source>FCIS</source>
          <year>2023</year>
          <volume>2</volume>
          <issue>2</issue>
          <fpage>81</fpage>
          <lpage>83</lpage>
          <pub-id pub-id-type="doi">10.54097/fcis.v2i2.4465</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pradana</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Elisa</surname>
              <given-names>HP</given-names>
            </name>
            <name name-style="western">
              <surname>Syarifuddin</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Discussing ChatGPT in education: a literature review and bibliometric analysis</article-title>
          <source>Cogent Educ</source>
          <year>2023</year>
          <volume>10</volume>
          <issue>2</issue>
          <fpage>2243134</fpage>
          <pub-id pub-id-type="doi">10.1080/2331186x.2023.2243134</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Baig</surname>
              <given-names>MI</given-names>
            </name>
            <name name-style="western">
              <surname>Yadegaridehkordi</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT in the higher education: a systematic literature review and research challenges</article-title>
          <source>Int J Educ Res</source>
          <year>2024</year>
          <volume>127</volume>
          <fpage>102411</fpage>
          <pub-id pub-id-type="doi">10.1016/j.ijer.2024.102411</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rasul</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Nair</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Kalendra</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Robin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Santini</surname>
              <given-names>FDO</given-names>
            </name>
            <name name-style="western">
              <surname>Ladeira</surname>
              <given-names>WJ</given-names>
            </name>
          </person-group>
          <article-title>The role of ChatGPT in higher education: benefits, challenges, and future research directions</article-title>
          <source>JALT</source>
          <year>2023</year>
          <volume>6</volume>
          <issue>1</issue>
          <fpage>41</fpage>
          <lpage>56</lpage>
          <pub-id pub-id-type="doi">10.37074/jalt.2023.6.1.29</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rawas</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT: empowering lifelong learning in the digital age of higher education</article-title>
          <source>Educ Inf Technol</source>
          <year>2023</year>
          <volume>29</volume>
          <issue>6</issue>
          <fpage>6895</fpage>
          <lpage>6908</lpage>
          <pub-id pub-id-type="doi">10.1007/s10639-023-12114-8</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nikolic</surname>
              <given-names>S</given-names>
            </name>
            <collab>Daniel</collab>
            <name name-style="western">
              <surname>Haque</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Belkina</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Hassan</surname>
              <given-names>GM</given-names>
            </name>
            <name name-style="western">
              <surname>Grundy</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lyden</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Neal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Sandison</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT versus engineering education assessment: a multidisciplinary and multi-institutional benchmarking and analysis of this generative artificial intelligence tool to investigate assessment integrity</article-title>
          <source>Eur J Eng Educ</source>
          <year>2023</year>
          <volume>48</volume>
          <issue>4</issue>
          <fpage>559</fpage>
          <lpage>614</lpage>
          <pub-id pub-id-type="doi">10.1080/03043797.2023.2213169</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chaudhry</surname>
              <given-names>IS</given-names>
            </name>
            <name name-style="western">
              <surname>Sarwary</surname>
              <given-names>SAM</given-names>
            </name>
            <name name-style="western">
              <surname>El Refae</surname>
              <given-names>GA</given-names>
            </name>
            <name name-style="western">
              <surname>Chabchoub</surname>
              <given-names>H</given-names>
            </name>
          </person-group>
          <article-title>Time to revisit existing student’s performance evaluation approach in higher education sector in a new era of ChatGPT — a case study</article-title>
          <source>Cogent Educ</source>
          <year>2023</year>
          <volume>10</volume>
          <issue>1</issue>
          <fpage>2210461</fpage>
          <pub-id pub-id-type="doi">10.1080/2331186x.2023.2210461</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ali</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Barhom</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Tamimi</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Duggal</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT-A double-edged sword for healthcare education? Implications for assessments of dental students</article-title>
          <source>Eur J Dent Educ</source>
          <year>2024</year>
          <volume>28</volume>
          <issue>1</issue>
          <fpage>206</fpage>
          <lpage>211</lpage>
          <pub-id pub-id-type="doi">10.1111/eje.12937</pub-id>
          <pub-id pub-id-type="medline">37550893</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hadi Mogavi</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Deng</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Juho Kim</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>D. Kwon</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Hosny Saleh Metwally</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Tlili</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bassanelli</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Bucchiarone</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Gujar</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Nacke</surname>
              <given-names>LE</given-names>
            </name>
            <name name-style="western">
              <surname>Hui</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT in education: a blessing or a curse? A qualitative study exploring early adopters’ utilization and perceptions</article-title>
          <source>Comput Hum Behav Artif Hum</source>
          <year>2024</year>
          <volume>2</volume>
          <issue>1</issue>
          <fpage>100027</fpage>
          <pub-id pub-id-type="doi">10.1016/j.chbah.2023.100027</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sallam</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Salim</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Barakat</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Al-Tammemi</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT applications in medical, dental, pharmacy, and public health education: a descriptive study highlighting the advantages and limitations</article-title>
          <source>Narra J</source>
          <year>2023</year>
          <volume>3</volume>
          <issue>1</issue>
          <fpage>e103</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/38450035"/>
          </comment>
          <pub-id pub-id-type="doi">10.52225/narra.v3i1.103</pub-id>
          <pub-id pub-id-type="medline">38450035</pub-id>
          <pub-id pub-id-type="pii">NarraJ-3-e103</pub-id>
          <pub-id pub-id-type="pmcid">PMC10914078</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Akiba</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Fraboni</surname>
              <given-names>MC</given-names>
            </name>
          </person-group>
          <article-title>AI-supported academic advising: exploring ChatGPT’s current state and future potential toward student empowerment</article-title>
          <source>Educ Sci</source>
          <year>2023</year>
          <volume>13</volume>
          <issue>9</issue>
          <fpage>885</fpage>
          <pub-id pub-id-type="doi">10.3390/educsci13090885</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Luo</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Ye</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zhong</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Zhao</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Applications, challenges, and prospects of generative artificial intelligence empowering medical education: scoping review</article-title>
          <source>JMIR Med Educ</source>
          <year>2025</year>
          <volume>11</volume>
          <fpage>e71125</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mededu.jmir.org/2025//e71125/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/71125</pub-id>
          <pub-id pub-id-type="medline">41128430</pub-id>
          <pub-id pub-id-type="pii">v11i1e71125</pub-id>
          <pub-id pub-id-type="pmcid">PMC12547994</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Janumpally</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Nanua</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ngo</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Youens</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Generative artificial intelligence in graduate medical education</article-title>
          <source>Front Med (Lausanne)</source>
          <year>2024</year>
          <volume>11</volume>
          <fpage>1525604</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.3389/fmed.2024.1525604"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/fmed.2024.1525604</pub-id>
          <pub-id pub-id-type="medline">39867924</pub-id>
          <pub-id pub-id-type="pmcid">PMC11758457</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Weightman</surname>
              <given-names>MJ</given-names>
            </name>
            <name name-style="western">
              <surname>Chur-Hansen</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Clark</surname>
              <given-names>SR</given-names>
            </name>
          </person-group>
          <article-title>AI in psychiatric education and training from 2016 to 2024: scoping review of trends</article-title>
          <source>JMIR Med Educ</source>
          <year>2025</year>
          <volume>11</volume>
          <fpage>e81517</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mededu.jmir.org/2025//e81517/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/81517</pub-id>
          <pub-id pub-id-type="medline">41474738</pub-id>
          <pub-id pub-id-type="pii">v11i1e81517</pub-id>
          <pub-id pub-id-type="pmcid">PMC12755346</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Farhud</surname>
              <given-names>DD</given-names>
            </name>
            <name name-style="western">
              <surname>Zokaei</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Ethical issues of artificial intelligence in medicine and healthcare</article-title>
          <source>Iran J Public Health</source>
          <year>2021</year>
          <volume>50</volume>
          <issue>11</issue>
          <fpage>i</fpage>
          <lpage>v</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/35223619"/>
          </comment>
          <pub-id pub-id-type="doi">10.18502/ijph.v50i11.7600</pub-id>
          <pub-id pub-id-type="medline">35223619</pub-id>
          <pub-id pub-id-type="pii">IJPH-50-i</pub-id>
          <pub-id pub-id-type="pmcid">PMC8826344</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rahman</surname>
              <given-names>MM</given-names>
            </name>
            <name name-style="western">
              <surname>Watanobe</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT for education and research: opportunities, threats, and strategies</article-title>
          <source>Appl Sci</source>
          <year>2023</year>
          <volume>13</volume>
          <issue>9</issue>
          <fpage>5783</fpage>
          <pub-id pub-id-type="doi">10.3390/app13095783</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Neumann</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Rauschenberger</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Schön</surname>
              <given-names>EM</given-names>
            </name>
          </person-group>
          <article-title>“We need to talk about ChatGPT”: the future of AI and higher education</article-title>
          <year>2023</year>
          <conf-name>2023 IEEE/ACM 5th International Workshop on Software Engineering Education for the Next Generation (SEENG)</conf-name>
          <conf-date>2023 May 16</conf-date>
          <conf-loc>Melbourne, Australia</conf-loc>
          <publisher-name>IEEE</publisher-name>
          <pub-id pub-id-type="doi">10.1109/seeng59157.2023.00010</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>AlAfnan</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Dishari</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Jovic</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Lomidze</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT as an educational tool: opportunities, challenges, and recommendations for communication, business writing, and composition courses</article-title>
          <source>JAIT</source>
          <year>2023</year>
          <volume>3</volume>
          <fpage>60</fpage>
          <lpage>68</lpage>
          <pub-id pub-id-type="doi">10.37965/jait.2023.0184</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Weng</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>XIA</surname>
              <given-names>Q</given-names>
            </name>
            <name name-style="western">
              <surname>Gu</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Rajaram</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Chiu</surname>
              <given-names>TKF</given-names>
            </name>
          </person-group>
          <article-title>Assessment and learning outcomes for generative AI in higher education: a scoping review on current research status and trends</article-title>
          <source>AJET</source>
          <year>2024</year>
          <volume>40</volume>
          <issue>6</issue>
          <fpage>37</fpage>
          <lpage>55</lpage>
          <pub-id pub-id-type="doi">10.14742/ajet.9540</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Foltynek</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Bjelobaba</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Glendinning</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Khan</surname>
              <given-names>ZR</given-names>
            </name>
            <name name-style="western">
              <surname>Santos</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Pavletic</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Kravjar</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>ENAI recommendations on the ethical use of artificial intelligence in education</article-title>
          <source>Int J Educ Integr</source>
          <year>2023</year>
          <volume>19</volume>
          <issue>1</issue>
          <fpage>1</fpage>
          <lpage>4</lpage>
          <pub-id pub-id-type="doi">10.1007/s40979-023-00133-4</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Elkhatat</surname>
              <given-names>AM</given-names>
            </name>
            <name name-style="western">
              <surname>Elsaid</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Almeer</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Evaluating the efficacy of AI content detection tools in differentiating between human and AI-generated text</article-title>
          <source>Int J Educ Integr</source>
          <year>2023</year>
          <volume>19</volume>
          <issue>1</issue>
          <fpage>1</fpage>
          <lpage>16</lpage>
          <pub-id pub-id-type="doi">10.1007/s40979-023-00140-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Birks</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Gray</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Darling-Pomranz</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Using artificial intelligence to provide a 'flipped assessment' approach to medical education learning opportunities</article-title>
          <source>Med Teach</source>
          <year>2025</year>
          <volume>47</volume>
          <issue>8</issue>
          <fpage>1377</fpage>
          <lpage>1384</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.tandfonline.com/doi/10.1080/0142159X.2024.2434101?url_ver=Z39.88-2003&#38;rfr_id=ori:rid:crossref.org&#38;rfr_dat=cr_pub  0pubmed"/>
          </comment>
          <pub-id pub-id-type="doi">10.1080/0142159X.2024.2434101</pub-id>
          <pub-id pub-id-type="medline">39616548</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ichikawa</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Olsen</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Vinod</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Glenn</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Hanna</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Lund</surname>
              <given-names>GC</given-names>
            </name>
            <name name-style="western">
              <surname>Pierce-Talsma</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Generative artificial intelligence in medical education-policies and training at us osteopathic medical schools: descriptive cross-sectional survey</article-title>
          <source>JMIR Med Educ</source>
          <year>2025</year>
          <volume>11</volume>
          <fpage>e58766</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mededu.jmir.org/2025//e58766/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/58766</pub-id>
          <pub-id pub-id-type="medline">39934984</pub-id>
          <pub-id pub-id-type="pii">v11i1e58766</pub-id>
          <pub-id pub-id-type="pmcid">PMC11835596</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gjorgioski</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Riley</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Prasad</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Tassos</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Nexhip</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Richardson</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Robinson</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Workforce survey of Australian health information management graduates, 2017-2021: a 5-year follow-on study</article-title>
          <source>Health Inf Manag</source>
          <year>2025</year>
          <volume>54</volume>
          <issue>1</issue>
          <fpage>43</fpage>
          <lpage>54</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://journals.sagepub.com/doi/10.1177/18333583231197936?url_ver=Z39.88-2003&#38;rfr_id=ori:rid:crossref.org&#38;rfr_dat=cr_pub  0pubmed"/>
          </comment>
          <pub-id pub-id-type="doi">10.1177/18333583231197936</pub-id>
          <pub-id pub-id-type="medline">37753774</pub-id>
          <pub-id pub-id-type="pmcid">PMC11705756</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Riley</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Robinson</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Prasad</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Gleeson</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Barker</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Wollersheim</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Price</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Workforce survey of Australian graduate health information managers: employability, employment, and knowledge and skills used in the workplace</article-title>
          <source>Health Inf Manag</source>
          <year>2020</year>
          <volume>49</volume>
          <issue>2-3</issue>
          <fpage>88</fpage>
          <lpage>98</lpage>
          <pub-id pub-id-type="doi">10.1177/1833358319839296</pub-id>
          <pub-id pub-id-type="medline">31006266</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Temsah</surname>
              <given-names>MH</given-names>
            </name>
            <name name-style="western">
              <surname>Aljamaan</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Malki</surname>
              <given-names>KH</given-names>
            </name>
            <name name-style="western">
              <surname>Alhasan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Altamimi</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Aljarbou</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Bazuhair</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Alsubaihin</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Abdulmajeed</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Alshahrani</surname>
              <given-names>FS</given-names>
            </name>
            <name name-style="western">
              <surname>Temsah</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Alshahrani</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Al-Eyadhy</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Alkhateeb</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Saddik</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Halwani</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Jamal</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Al-Tawfiq</surname>
              <given-names>JA</given-names>
            </name>
            <name name-style="western">
              <surname>Al-Eyadhy</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT and the future of digital health: a study on healthcare workers' perceptions and expectations</article-title>
          <source>Healthcare (Basel)</source>
          <year>2023</year>
          <volume>11</volume>
          <issue>13</issue>
          <fpage>1812</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=healthcare11131812"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/healthcare11131812</pub-id>
          <pub-id pub-id-type="medline">37444647</pub-id>
          <pub-id pub-id-type="pii">healthcare11131812</pub-id>
          <pub-id pub-id-type="pmcid">PMC10340744</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tudor Car</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Kyaw</surname>
              <given-names>BM</given-names>
            </name>
            <name name-style="western">
              <surname>Nannan Panday</surname>
              <given-names>RS</given-names>
            </name>
            <name name-style="western">
              <surname>van der Kleij</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Chavannes</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Majeed</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Car</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Digital health training programs for medical students: scoping review</article-title>
          <source>JMIR Med Educ</source>
          <year>2021</year>
          <volume>7</volume>
          <issue>3</issue>
          <fpage>e28275</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mededu.jmir.org/2021/3/e28275/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/28275</pub-id>
          <pub-id pub-id-type="medline">34287206</pub-id>
          <pub-id pub-id-type="pii">v7i3e28275</pub-id>
          <pub-id pub-id-type="pmcid">PMC8339984</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Fang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Fu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Ling</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yan</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Jiang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Cao</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Abudukeremu</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Performance of ChatGPT on the Chinese postgraduate examination for clinical medicine: survey study</article-title>
          <source>JMIR Med Educ</source>
          <year>2024</year>
          <volume>10</volume>
          <fpage>e48514</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mededu.jmir.org/2024//e48514/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/48514</pub-id>
          <pub-id pub-id-type="medline">38335017</pub-id>
          <pub-id pub-id-type="pii">v10i1e48514</pub-id>
          <pub-id pub-id-type="pmcid">PMC10891494</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>QY</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Ong</surname>
              <given-names>CW</given-names>
            </name>
            <name name-style="western">
              <surname>Ho</surname>
              <given-names>CSH</given-names>
            </name>
          </person-group>
          <article-title>The role of generative artificial intelligence in psychiatric education- a scoping review</article-title>
          <source>BMC Med Educ</source>
          <year>2025</year>
          <volume>25</volume>
          <issue>1</issue>
          <fpage>438</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://bmcmededuc.biomedcentral.com/articles/10.1186/s12909-025-07026-9"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s12909-025-07026-9</pub-id>
          <pub-id pub-id-type="medline">40133891</pub-id>
          <pub-id pub-id-type="pii">10.1186/s12909-025-07026-9</pub-id>
          <pub-id pub-id-type="pmcid">PMC11938615</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cherif</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Moussa</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Missaoui</surname>
              <given-names>AM</given-names>
            </name>
            <name name-style="western">
              <surname>Salouage</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Mokaddem</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Dhahri</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Appraisal of ChatGPT's aptitude for medical education: comparative analysis with third-year medical students in a pulmonology examination</article-title>
          <source>JMIR Med Educ</source>
          <year>2024</year>
          <volume>10</volume>
          <fpage>e52818</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mededu.jmir.org/2024//e52818/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/52818</pub-id>
          <pub-id pub-id-type="medline">39042876</pub-id>
          <pub-id pub-id-type="pii">v10i1e52818</pub-id>
          <pub-id pub-id-type="pmcid">PMC11303904</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Eysenbach</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>The role of ChatGPT, generative language models, and artificial intelligence in medical education: a conversation with ChatGPT and a call for papers</article-title>
          <source>JMIR Med Educ</source>
          <year>2023</year>
          <volume>9</volume>
          <fpage>e46885</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mededu.jmir.org/2023//e46885/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/46885</pub-id>
          <pub-id pub-id-type="medline">36863937</pub-id>
          <pub-id pub-id-type="pii">v9i1e46885</pub-id>
          <pub-id pub-id-type="pmcid">PMC10028514</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Meo</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Al-Masri</surname>
              <given-names>AA</given-names>
            </name>
            <name name-style="western">
              <surname>Alotaibi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Meo</surname>
              <given-names>MZS</given-names>
            </name>
            <name name-style="western">
              <surname>Meo</surname>
              <given-names>MOS</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT knowledge evaluation in basic and clinical medical sciences: multiple choice question examination-based performance</article-title>
          <source>Healthcare (Basel)</source>
          <year>2023</year>
          <volume>11</volume>
          <issue>14</issue>
          <fpage>2046</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.mdpi.com/resolver?pii=healthcare11142046"/>
          </comment>
          <pub-id pub-id-type="doi">10.3390/healthcare11142046</pub-id>
          <pub-id pub-id-type="medline">37510487</pub-id>
          <pub-id pub-id-type="pii">healthcare11142046</pub-id>
          <pub-id pub-id-type="pmcid">PMC10379728</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Benito</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Isla-Jover</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>González-Castro</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Fernández Esparcia</surname>
              <given-names>PJ</given-names>
            </name>
            <name name-style="western">
              <surname>Carpio</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Blay-Simón</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Gutiérrez-Bedia</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Lapastora</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Carratalá</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Carazo-Casas</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>GPT-4o and openAI o1 performance on the 2024 Spanish competitive medical specialty access examination: cross-sectional quantitative evaluation study</article-title>
          <source>JMIR Med Educ</source>
          <year>2026</year>
          <volume>12</volume>
          <fpage>e75452</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mededu.jmir.org/2026//e75452/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/75452</pub-id>
          <pub-id pub-id-type="medline">41525685</pub-id>
          <pub-id pub-id-type="pii">v12i1e75452</pub-id>
          <pub-id pub-id-type="pmcid">PMC12795474</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wani</surname>
              <given-names>TA</given-names>
            </name>
            <name name-style="western">
              <surname>Liem</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Boyd</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wijesooriya</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Khan</surname>
              <given-names>UR</given-names>
            </name>
          </person-group>
          <article-title>Digital health education in Australian universities: trends, gaps, and future directions</article-title>
          <source>Int J Med Inform</source>
          <year>2026</year>
          <volume>205</volume>
          <fpage>106105</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1386-5056(25)00322-3"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.ijmedinf.2025.106105</pub-id>
          <pub-id pub-id-type="medline">40916275</pub-id>
          <pub-id pub-id-type="pii">S1386-5056(25)00322-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kolade</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Owoseni</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Egbetokun</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Is AI changing learning and assessment as we know it? Evidence from a ChatGPT experiment and a conceptual framework</article-title>
          <source>Heliyon</source>
          <year>2024</year>
          <volume>10</volume>
          <issue>4</issue>
          <fpage>e25953</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S2405-8440(24)01984-4"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.heliyon.2024.e25953</pub-id>
          <pub-id pub-id-type="medline">38379960</pub-id>
          <pub-id pub-id-type="pii">S2405-8440(24)01984-4</pub-id>
          <pub-id pub-id-type="pmcid">PMC10877295</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Moin</surname>
              <given-names>KA</given-names>
            </name>
            <name name-style="western">
              <surname>Nasir</surname>
              <given-names>AA</given-names>
            </name>
            <name name-style="western">
              <surname>Petroff</surname>
              <given-names>DJ</given-names>
            </name>
            <name name-style="western">
              <surname>Loveless</surname>
              <given-names>BA</given-names>
            </name>
            <name name-style="western">
              <surname>Moshirfar</surname>
              <given-names>OA</given-names>
            </name>
            <name name-style="western">
              <surname>Hoopes</surname>
              <given-names>PC</given-names>
            </name>
            <name name-style="western">
              <surname>Moshirfar</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Assessment of generative artificial intelligence (AI) models in creating medical illustrations for various corneal transplant procedures</article-title>
          <source>Cureus</source>
          <year>2024</year>
          <volume>16</volume>
          <issue>8</issue>
          <fpage>e67833</fpage>
          <pub-id pub-id-type="doi">10.7759/cureus.67833</pub-id>
          <pub-id pub-id-type="medline">39328681</pub-id>
          <pub-id pub-id-type="pmcid">PMC11424388</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vygotsky</surname>
              <given-names>LS</given-names>
            </name>
          </person-group>
          <source>Mind in Society: The Development of Higher Psychological Processes</source>
          <year>1978</year>
          <publisher-loc>Cambridge, Massachusetts</publisher-loc>
          <publisher-name>Harvard University Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tran</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Balasooriya</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Semmler</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Rhee</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Generative artificial intelligence: the 'more knowledgeable other' in a social constructivist framework of medical education</article-title>
          <source>NPJ Digit Med</source>
          <year>2025</year>
          <volume>8</volume>
          <issue>1</issue>
          <fpage>430</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-025-01823-8"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-025-01823-8</pub-id>
          <pub-id pub-id-type="medline">40646156</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-025-01823-8</pub-id>
          <pub-id pub-id-type="pmcid">PMC12254308</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Anderson</surname>
              <given-names>LW</given-names>
            </name>
            <name name-style="western">
              <surname>Krathwohl</surname>
              <given-names>DR</given-names>
            </name>
          </person-group>
          <article-title>A taxonomy for learning, teaching, and assessing: a revision of Bloom's taxonomy of educational objectives: complete edition</article-title>
          <source>EDUQ.info</source>
          <year>2001</year>
          <access-date>2025-11-14</access-date>
          <publisher-name>Addison Wesley Longman</publisher-name>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://eduq.info/xmlui/handle/11515/18824">https://eduq.info/xmlui/handle/11515/18824</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref49">
        <label>49</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bloom</surname>
              <given-names>BS</given-names>
            </name>
            <name name-style="western">
              <surname>Engelhart</surname>
              <given-names>MD</given-names>
            </name>
            <name name-style="western">
              <surname>Furst</surname>
              <given-names>EJ</given-names>
            </name>
            <name name-style="western">
              <surname>Hill</surname>
              <given-names>WH</given-names>
            </name>
            <name name-style="western">
              <surname>Krathwohl</surname>
              <given-names>DR</given-names>
            </name>
          </person-group>
          <source>Taxonomy of Educational Objectives</source>
          <year>1964</year>
          <publisher-loc>New York</publisher-loc>
          <publisher-name>Longmans, Green</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref50">
        <label>50</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bandura</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <source>Social Foundations of Thought and Action</source>
          <year>1986</year>
          <publisher-loc>Englewood Cliffs NJ</publisher-loc>
          <publisher-name>Prentice Hall</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref51">
        <label>51</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Powell</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Forsyth</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Generative AI and the implications for authentic assessment</article-title>
          <source>Using Gener AI Eff High Educ</source>
          <year>2024</year>
          <publisher-loc>New York</publisher-loc>
          <publisher-name>Routledge</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref52">
        <label>52</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Salinas-Navarro</surname>
              <given-names>DE</given-names>
            </name>
            <name name-style="western">
              <surname>Vilalta-Perdomo</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Michel-Villarreal</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Montesinos</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Designing experiential learning activities with generative artificial intelligence tools for authentic assessment</article-title>
          <source>ITSE</source>
          <year>2024</year>
          <volume>21</volume>
          <issue>4</issue>
          <fpage>708</fpage>
          <lpage>734</lpage>
          <pub-id pub-id-type="doi">10.1108/itse-12-2023-0236</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref53">
        <label>53</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hersh</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Fultz Hollis</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>Results and implications for generative AI in a large introductory biomedical and health informatics course</article-title>
          <source>NPJ Digit Med</source>
          <year>2024</year>
          <volume>7</volume>
          <issue>1</issue>
          <fpage>247</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-024-01251-0"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-024-01251-0</pub-id>
          <pub-id pub-id-type="medline">39271955</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41746-024-01251-0</pub-id>
          <pub-id pub-id-type="pmcid">PMC11399285</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref54">
        <label>54</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Morjaria</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Burns</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Bracken</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Ngo</surname>
              <given-names>QN</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Levinson</surname>
              <given-names>AJ</given-names>
            </name>
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Thompson</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Sibbald</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Examining the threat of ChatGPT to the validity of short answer assessments in an undergraduate medical program</article-title>
          <source>J Med Educ Curric Dev</source>
          <year>2023</year>
          <volume>10</volume>
          <fpage>23821205231204178</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://journals.sagepub.com/doi/10.1177/23821205231204178?url_ver=Z39.88-2003&#38;rfr_id=ori:rid:crossref.org&#38;rfr_dat=cr_pub  0pubmed"/>
          </comment>
          <pub-id pub-id-type="doi">10.1177/23821205231204178</pub-id>
          <pub-id pub-id-type="medline">37780034</pub-id>
          <pub-id pub-id-type="pii">10.1177_23821205231204178</pub-id>
          <pub-id pub-id-type="pmcid">PMC10540597</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref55">
        <label>55</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Newton</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Xiromeriti</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>ChatGPT performance on multiple choice question examinations in higher education. A pragmatic scoping review</article-title>
          <source>Assess Eval High Educ</source>
          <year>2024</year>
          <volume>49</volume>
          <issue>6</issue>
          <fpage>781</fpage>
          <lpage>798</lpage>
          <pub-id pub-id-type="doi">10.1080/02602938.2023.2299059</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref56">
        <label>56</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Hu</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Most</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Hawkins</surname>
              <given-names>WA</given-names>
            </name>
            <name name-style="western">
              <surname>Murray</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Smith</surname>
              <given-names>SE</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Sikora</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Evaluating accuracy and reproducibility of large language model performance on critical care assessments in pharmacy education</article-title>
          <source>Front Artif Intell</source>
          <year>2024</year>
          <volume>7</volume>
          <fpage>1514896</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.3389/frai.2024.1514896"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/frai.2024.1514896</pub-id>
          <pub-id pub-id-type="medline">39850846</pub-id>
          <pub-id pub-id-type="pmcid">PMC11754395</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref57">
        <label>57</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nguyen Thanh</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Vo</surname>
              <given-names>DTH</given-names>
            </name>
            <name name-style="western">
              <surname>Nguyen Nhat</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Pham</surname>
              <given-names>TTT</given-names>
            </name>
            <name name-style="western">
              <surname>Thai Trung</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Ha Xuan</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Race with the machines: assessing the capability of generative AI in solving authentic assessments</article-title>
          <source>AJET</source>
          <year>2023</year>
          <volume>39</volume>
          <issue>5</issue>
          <fpage>59</fpage>
          <lpage>81</lpage>
          <pub-id pub-id-type="doi">10.14742/ajet.8902</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref58">
        <label>58</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Abdelgadir</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Thongprayoon</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Miao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Suppadungsuk</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Pham</surname>
              <given-names>JH</given-names>
            </name>
            <name name-style="western">
              <surname>Mao</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Craici</surname>
              <given-names>IM</given-names>
            </name>
            <name name-style="western">
              <surname>Cheungpasitporn</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>AI integration in nephrology: evaluating ChatGPT for accurate ICD-10 documentation and coding</article-title>
          <source>Front Artif Intell</source>
          <year>2024</year>
          <volume>7</volume>
          <fpage>1457586</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.3389/frai.2024.1457586"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/frai.2024.1457586</pub-id>
          <pub-id pub-id-type="medline">39286549</pub-id>
          <pub-id pub-id-type="pmcid">PMC11402808</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref59">
        <label>59</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Soroush</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Glicksberg</surname>
              <given-names>BS</given-names>
            </name>
            <name name-style="western">
              <surname>Zimlichman</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Barash</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Freeman</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Charney</surname>
              <given-names>AW</given-names>
            </name>
            <name name-style="western">
              <surname>Nadkarni</surname>
              <given-names>GN</given-names>
            </name>
            <name name-style="western">
              <surname>Klang</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Large language models are poor medical coders — benchmarking of medical code querying</article-title>
          <source>NEJM AI</source>
          <year>2024</year>
          <volume>1</volume>
          <issue>5</issue>
          <fpage>AIdbp2300040</fpage>
          <pub-id pub-id-type="doi">10.1056/aidbp2300040</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref60">
        <label>60</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pornphol</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Chittayasothorn</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Verification of relational database languages codes generated by ChatGPT</article-title>
          <year>2024</year>
          <conf-name>ASSE '23: Proceedings of the 2023 4th Asia Service Sciences and Software Engineering Conference</conf-name>
          <conf-date>2023 October 27 - 29</conf-date>
          <conf-loc>Aizu-Wakamatsu City Japan</conf-loc>
          <publisher-name>Association for Computing Machinery</publisher-name>
          <fpage>17</fpage>
          <lpage>22</lpage>
          <pub-id pub-id-type="doi">10.1145/3634814.3634817</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref61">
        <label>61</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Carr</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Shawon</surname>
              <given-names>FR</given-names>
            </name>
            <name name-style="western">
              <surname>Jamil</surname>
              <given-names>HM</given-names>
            </name>
          </person-group>
          <article-title>An experiment on leveraging ChatGPT for online teaching and assessment of database students</article-title>
          <year>2023</year>
          <conf-name>2023 IEEE Int Conf Teach Assess Learn Eng TALE</conf-name>
          <conf-date>2023 November 27 – 1 December</conf-date>
          <conf-loc>Auckland, New Zealand</conf-loc>
          <fpage>1</fpage>
          <lpage>8</lpage>
          <pub-id pub-id-type="doi">10.1109/tale56641.2023.10398239</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref62">
        <label>62</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Walters</surname>
              <given-names>WH</given-names>
            </name>
            <name name-style="western">
              <surname>Wilder</surname>
              <given-names>EI</given-names>
            </name>
          </person-group>
          <article-title>Fabrication and errors in the bibliographic citations generated by ChatGPT</article-title>
          <source>Sci Rep</source>
          <year>2023</year>
          <volume>13</volume>
          <issue>1</issue>
          <fpage>14045</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-023-41032-5"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-023-41032-5</pub-id>
          <pub-id pub-id-type="medline">37679503</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-023-41032-5</pub-id>
          <pub-id pub-id-type="pmcid">PMC10484980</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref63">
        <label>63</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Williams</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>Comparison of generative AI performance on undergraduate and postgraduate written assessments in the biomedical sciences</article-title>
          <source>Int J Educ Technol High Educ</source>
          <year>2024</year>
          <volume>21</volume>
          <issue>1</issue>
          <fpage>52</fpage>
          <pub-id pub-id-type="doi">10.1186/s41239-024-00485-y</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref64">
        <label>64</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Okuhara</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Chang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Shirabe</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Nishiie</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Okada</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Kiuchi</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Performance of ChatGPT across different versions in medical licensing examinations worldwide: systematic review and meta-analysis</article-title>
          <source>J Med Internet Res</source>
          <year>2024</year>
          <volume>26</volume>
          <fpage>e60807</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2024//e60807/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/60807</pub-id>
          <pub-id pub-id-type="medline">39052324</pub-id>
          <pub-id pub-id-type="pii">v26i1e60807</pub-id>
          <pub-id pub-id-type="pmcid">PMC11310649</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref65">
        <label>65</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>'In awe': scientists impressed by latest ChatGPT model o1</article-title>
          <source>Nature</source>
          <year>2024</year>
          <volume>634</volume>
          <issue>8033</issue>
          <fpage>275</fpage>
          <lpage>276</lpage>
          <pub-id pub-id-type="doi">10.1038/d41586-024-03169-9</pub-id>
          <pub-id pub-id-type="medline">39354139</pub-id>
          <pub-id pub-id-type="pii">10.1038/d41586-024-03169-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref66">
        <label>66</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Meehan</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Health informatics workforce in the digital health ecosystem</article-title>
          <source>Stud Health Technol Inform</source>
          <year>2024</year>
          <volume>310</volume>
          <fpage>1226</fpage>
          <lpage>1230</lpage>
          <pub-id pub-id-type="doi">10.3233/SHTI231160</pub-id>
          <pub-id pub-id-type="medline">38270010</pub-id>
          <pub-id pub-id-type="pii">SHTI231160</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref67">
        <label>67</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Butler-Henderson</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Gray</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Arabi</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Roles and responsibilities of the global specialist digital health workforce: analysis of global census data</article-title>
          <source>JMIR Med Educ</source>
          <year>2024</year>
          <volume>10</volume>
          <fpage>e54137</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mededu.jmir.org/2024//e54137/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/54137</pub-id>
          <pub-id pub-id-type="medline">39118468</pub-id>
          <pub-id pub-id-type="pii">v10i1e54137</pub-id>
          <pub-id pub-id-type="pmcid">PMC11327619</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref68">
        <label>68</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Keep</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Janssen</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>McGregor</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Brunner</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Baysari</surname>
              <given-names>MT</given-names>
            </name>
            <name name-style="western">
              <surname>Quinn</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Shaw</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Mapping eHealth education: review of eHealth content in health and medical degrees at a metropolitan tertiary institute in Australia</article-title>
          <source>JMIR Med Educ</source>
          <year>2021</year>
          <volume>7</volume>
          <issue>3</issue>
          <fpage>e16440</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mededu.jmir.org/2021/3/e16440/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/16440</pub-id>
          <pub-id pub-id-type="medline">34420920</pub-id>
          <pub-id pub-id-type="pii">v7i3e16440</pub-id>
          <pub-id pub-id-type="pmcid">PMC8414287</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref69">
        <label>69</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Khalil</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Er</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Will ChatGPT get you caught? Rethinking of plagiarism detection</article-title>
          <year>2023</year>
          <conf-name>10th International Conference, LCT 2023, Held as Part of the 25th HCI International Conference, HCII 2023</conf-name>
          <conf-date>2023  July 23–28</conf-date>
          <conf-loc>Copenhagen, Denmark</conf-loc>
          <publisher-loc>Cham</publisher-loc>
          <publisher-name>Springer Nature Switzerland</publisher-name>
          <fpage>475</fpage>
          <lpage>487</lpage>
          <pub-id pub-id-type="doi">10.35542/osf.io/fnh48</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
