<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Med Educ</journal-id><journal-id journal-id-type="publisher-id">mededu</journal-id><journal-id journal-id-type="index">20</journal-id><journal-title>JMIR Medical Education</journal-title><abbrev-journal-title>JMIR Med Educ</abbrev-journal-title><issn pub-type="epub">2369-3762</issn></journal-meta><article-meta><article-id pub-id-type="publisher-id">58126</article-id><article-id pub-id-type="doi">10.2196/58126</article-id><title-group><article-title>Use of Multiple-Choice Items in Summative Examinations: Questionnaire Survey Among German Undergraduate Dental Training Programs</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>R&#x00F6;ssler</surname><given-names>Lena</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Herrmann</surname><given-names>Manfred</given-names></name><degrees>Dr rer nat</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Wiegand</surname><given-names>Annette</given-names></name><degrees>Prof Dr med dent</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Kanzow</surname><given-names>Philipp</given-names></name><degrees>MSc, Dr rer medic, PD Dr med dent</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Preventive Dentistry, Periodontology and Cariology, University Medical Center G&#x00F6;ttingen</institution>, <addr-line>G&#x00F6;ttingen</addr-line>, <country>Germany</country></aff><aff id="aff2"><institution>Study Deanery, University Medical Center G&#x00F6;ttingen</institution>, <addr-line>G&#x00F6;ttingen</addr-line>, <country>Germany</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Lesselroth</surname><given-names>Blake</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Feofanova</surname><given-names>Elena</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Melzow</surname><given-names>Florentina</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Scholz</surname><given-names>Konstantin</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Philipp Kanzow, MSc, Dr rer medic, PD Dr med dent<email>philipp.kanzow@med.uni-goettingen.de</email></corresp></author-notes><pub-date pub-type="collection"><year>2024</year></pub-date><pub-date pub-type="epub"><day>27</day><month>6</month><year>2024</year></pub-date><volume>10</volume><elocation-id>e58126</elocation-id><history><date date-type="received"><day>07</day><month>03</month><year>2024</year></date><date date-type="rev-recd"><day>15</day><month>04</month><year>2024</year></date><date date-type="accepted"><day>07</day><month>05</month><year>2024</year></date></history><copyright-statement>&#x00A9; Lena R&#x00F6;ssler, Manfred Herrmann, Annette Wiegand, Philipp Kanzow. Originally published in JMIR Medical Education (<ext-link ext-link-type="uri" xlink:href="https://mededu.jmir.org">https://mededu.jmir.org</ext-link>), 27.6.2024. </copyright-statement><copyright-year>2024</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Education, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://mededu.jmir.org/">https://mededu.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://mededu.jmir.org/2024/1/e58126"/><abstract><sec><title>Background</title><p>Multiple-choice examinations are frequently used in German dental schools. However, details regarding the used item types and applied scoring methods are lacking.</p></sec><sec><title>Objective</title><p>This study aims to gain insight into the current use of multiple-choice items (ie, questions) in summative examinations in German undergraduate dental training programs.</p></sec><sec sec-type="methods"><title>Methods</title><p>A paper-based 10-item questionnaire regarding the used assessment methods, multiple-choice item types, and applied scoring methods was designed. The pilot-tested questionnaire was mailed to the deans of studies and to the heads of the Department of Operative/Restorative Dentistry at all 30 dental schools in Germany in February 2023. Statistical analysis was performed using the Fisher exact test (<italic>P</italic>&#x003C;.05).</p></sec><sec sec-type="results"><title>Results</title><p>The response rate amounted to 90% (27/30 dental schools). All respondent dental schools used multiple-choice examinations for summative assessments. Examinations were delivered electronically by 70% (19/27) of the dental schools. Almost all dental schools used single-choice Type A items (24/27, 89%), which accounted for the largest number of items in approximately half of the dental schools (13/27, 48%). Further item types (eg, conventional multiple-select items, Multiple-True-False, and Pick-N) were only used by fewer dental schools (&#x2264;67%, up to 18 out of 27 dental schools). For the multiple-select item types, the applied scoring methods varied considerably (ie, awarding [intermediate] partial credit and requirements for partial credit). Dental schools with the possibility of electronic examinations used multiple-select items slightly more often (14/19, 74% vs 4/8, 50%). However, this difference was statistically not significant (<italic>P</italic>=.38). Dental schools used items either individually or as key feature problems consisting of a clinical case scenario followed by a number of items focusing on critical treatment steps (15/27, 56%). Not a single school used alternative testing methods (eg, answer-until-correct). A formal item review process was established at about half of the dental schools (15/27, 56%).</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>Summative assessment methods among German dental schools vary widely. Especially, a large variability regarding the use and scoring of multiple-select multiple-choice items was found.</p></sec></abstract><kwd-group><kwd>alternate-choice</kwd><kwd>assessment</kwd><kwd>best-answer</kwd><kwd>dental</kwd><kwd>dental schools</kwd><kwd>dental training</kwd><kwd>education</kwd><kwd>educational assessment</kwd><kwd>educational measurement</kwd><kwd>examination</kwd><kwd>German</kwd><kwd>Germany</kwd><kwd>k of n</kwd><kwd>Kprim</kwd><kwd>K&#x2019;</kwd><kwd>medical education</kwd><kwd>medical student</kwd><kwd>MTF</kwd><kwd>Multiple-True-False</kwd><kwd>multiple choice</kwd><kwd>multiple-select</kwd><kwd>Pick-N</kwd><kwd>scoring</kwd><kwd>scoring system</kwd><kwd>single choice</kwd><kwd>single response</kwd><kwd>test</kwd><kwd>testing</kwd><kwd>true/false</kwd><kwd>true-false</kwd><kwd>Type A</kwd><kwd>Type K</kwd><kwd>Type K&#x2019;</kwd><kwd>Type R</kwd><kwd>Type X</kwd><kwd>undergraduate</kwd><kwd>undergraduate curriculum</kwd><kwd>undergraduate education</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Summative examinations of theoretical knowledge are an integral part of university degree programs. As they are intended to assess examinees&#x2019; ability regarding predefined learning objectives, they should reflect examinees&#x2019; true knowledge as closely as possible. To assess examinees objectively and efficiently, multiple-choice examinations were described as early as 1916 [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. To date, these types of examinations have been expanded by further item types, and multiple-choice examinations are frequently used within higher education including but not limited to dental training programs [<xref ref-type="bibr" rid="ref3">3</xref>-<xref ref-type="bibr" rid="ref5">5</xref>]. Multiple-choice items (ie, questions) can be subdivided into single-choice items (eg, Type A, Type K, Type R, and alternate-choice) and multiple-select items (eg, Pick-N and Multiple-True-False [Type K&#x2019;]) [<xref ref-type="bibr" rid="ref6">6</xref>]. While dichotomous scoring (ie, 1 full credit point is awarded if examinees mark the correct answer option or statements, otherwise no credit is awarded) is most commonly proposed for single-choice items [<xref ref-type="bibr" rid="ref7">7</xref>], scoring methods for multiple-select items are more heterogeneous: Besides dichotomous scoring, further scoring methods resulting in (intermediate) partial credit or even negative points (ie, malus points) have been described [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>].</p><p>Besides paper-based examinations, examinations are nowadays frequently delivered electronically. While electronic examinations are well perceived by examinees [<xref ref-type="bibr" rid="ref10">10</xref>], comprehensive studies regarding their effectiveness are still lacking [<xref ref-type="bibr" rid="ref11">11</xref>]. However, the use of different examination software (eg, UCAN&#x2019;s [Umbrella Consortium for Assessment Networks] CAMPUS examination software) might improve the ease of multiple-choice examinations, accelerate the evaluation of examinations and item analysis, and allow for more complex scoring algorithms. Despite the benefits associated with electronic examinations, the availability of hardware and software at the level of individual institutions might limit its use.</p><p>In Germany, the revised undergraduate dental curriculum consists of 10 semesters and includes preclinical training (4 semesters), training using simulators or phantom heads (2 semesters), and clinical training (4 semesters). Following the state examinations after each part (ie, after the fourth, sixth, and 10th semester), students receive their license (&#x201C;Approbation&#x201D;) to practice dentistry. Besides practical skills, theoretical knowledge is taught within the undergraduate dental curriculum, and students&#x2019; ability is often assessed using written multiple-choice examinations. However, such examinations are not standardized among German dental schools. While general recommendations exist for their design and evaluation [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref13">13</xref>], details such as suitable item types and applied scoring methods are often defined in local examination guidelines at the level of individual dental schools. However, these details might impact examinees&#x2019; scoring results [<xref ref-type="bibr" rid="ref5">5</xref>]. To the best of our knowledge, a comprehensive overview regarding the used item types and applied scoring methods at German dental schools does not exist.</p><p>Therefore, this study aimed to gain insight into the current use of multiple-choice items in summative examinations in German undergraduate dental training programs. The null hypothesis is that the use of digital examinations does not impact the use of more complex (ie, multiple-select) multiple-choice items.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Ethical Considerations</title><p>The study was designed as a prospective investigation. In preparation for the investigation, the websites of all German dental schools were screened (n=30), and the names of the heads of the Department of Operative/Restorative Dentistry and the deans of studies were noted for later procedures.</p><p>The study was performed after approval by the local ethics committee of the University Medical Center G&#x00F6;ttingen (approval number 22/1/23). Participation in this study was voluntary, and participants gave their informed consent for the anonymous evaluation of the provided answers by returning the questionnaires. Participants did not receive any incentives or compensation.</p></sec><sec id="s2-2"><title>Questionnaire</title><p>A paper-based questionnaire, consisting of 10 items about the construction and evaluation of summative examinations, was jointly designed by the authors and pilot-tested in the University Medical Center G&#x00F6;ttingen (<xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). Both closed and open-ended items were used. The opening questions related to different examination types used for the summative assessment of theoretical knowledge, and whether or not electronic examinations were being used. Additionally, it was asked whether the examination items undergo a formal review process and if so, the participants had the chance to give a brief description of this procedure. The more specific questions related to the types of multiple-choice items used and asked for the relative percentage to which these items were being used. Furthermore, the participants were asked to describe the applied scoring methods for each of the item types used. Finally, participants were provided with a text field open for comments and their contact details (ie, if required for further clarification) and were asked to supply a copy of their local examination guidelines or program regulations.</p><p>Following the evaluation of the pilot survey among 5 dentists at the University Medical Center G&#x00F6;ttingen, the questionnaire was slightly modified for clarification, printed, and mailed to (1) the heads of the Department of Operative/Restorative Dentistry and to (2) the deans of studies on February 1, 2023. The wording was slightly adjusted for each recipient: (1) &#x201C;used in your department&#x201D; versus (2) &#x201C;permitted at your dental school&#x201D;. Mailings included a personalized cover letter, an overview illustrating different multiple-choice item types (<xref ref-type="fig" rid="figure1">Figure 1</xref>), and a stamped return envelope. The survey was closed after 12 weeks. Nonresponders were reminded once 6 weeks after the initial distribution of the questionnaires.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Exemplary presentation of the most commonly used multiple-choice item types referenced in the questionnaire. Round marking boxes represent 1 answer option to be selected (1 out of x), while square marking boxes imply that multiple answer options or statements (x out of X) can be chosen.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mededu_v10i1e58126_fig01.png"/></fig></sec><sec id="s2-3"><title>Statistical Analysis</title><p>First, data were manually transferred into a digital chart using a piloted spreadsheet containing columns for each item of the questionnaire. This step was independently performed by 2 authors (LR and PK). In case of disagreement, data were repeatedly extracted from the returned questionnaires.</p><p>In case of disagreement between the heads of the Department of Operative/Restorative Dentistry and the deans of studies, results were based on the responses from the heads of the Department of Operative/Restorative Dentistry. For further clarification, responses were cross-validated with the supplied or publicly available examination guidelines and program regulations. If required, respondents were contacted for further clarification if they had agreed to do so previously.</p><p>Second, statistical analysis was performed using the software SPSS Statistics (Macintosh version 29.0.0.0; IBM Corp). The effect of delivering digital examinations on the use of multiple-select items was assessed using the Fisher exact test. The level of significance was set at .05.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Overview</title><p>In total, responses from 27 dental schools were received yielding a response rate of 90% (27/30 dental schools). More specifically, 25 Departments of Operative/Restorative Dentistry and 17 deans of studies replied. All dental schools responded that they use written multiple-choice examinations for the assessment of examinees&#x2019; theoretical knowledge. Therefore, subsequent results are based on the number of respondent dental schools.</p></sec><sec id="s3-2"><title>Multiple-Choice Items Used</title><p>The most commonly used multiple-choice item types at German dental schools were single-choice Type A or Type A<sub>negative</sub> items with 3 to 6 answer options (24/27, 89%). Pick-N items (ie, the number of answer options to be selected is known to examinees) were reported to contain between 3 and 26 answer options and were used by 67% (18/27) of dental schools. Type K items were reported to contain between 3 and 6 statements and were used by 52% (14/27) of the dental schools. Multiple-True-False (also known under further names such as Kprim, Type K&#x2019;, or Type X) and conventional multiple-select items (ie, the number of answer options to be selected is unknown to examinees) were reported to contain between 4 and 6 statements or answer options and were both used by 44% (12/27) of the dental schools. The use of further item types is shown in <xref ref-type="table" rid="table1">Table 1</xref>.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Different multiple-choice item types for the assessment of theoretical knowledge at the respondent dental schools (N=27).</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Item type</td><td align="left" valign="bottom">Dental schools, n (%)</td></tr></thead><tbody><tr><td align="left" valign="top">Type A</td><td align="left" valign="top">24 (89)</td></tr><tr><td align="left" valign="top">Pick-N</td><td align="left" valign="top">18 (67)</td></tr><tr><td align="left" valign="top">Type K</td><td align="left" valign="top">14 (52)</td></tr><tr><td align="left" valign="top">Conventional multiple-select</td><td align="left" valign="top">12 (44)</td></tr><tr><td align="left" valign="top">Multiple-True-False (Type K&#x2019;)</td><td align="left" valign="top">12 (44)</td></tr><tr><td align="left" valign="top">Type R</td><td align="left" valign="top">6 (22)</td></tr><tr><td align="left" valign="top">Alternate-choice</td><td align="left" valign="top">4 (15)</td></tr></tbody></table></table-wrap></sec><sec id="s3-3"><title>Examination Setting</title><p>Key feature problems consisting of a clinical case scenario followed by a number of items focusing on critical treatment steps were used by approximately half of the dental schools (15/27, 56%). Not a single school used alternative testing methods (eg, answer-until-correct). Also, a formal item review process prior to the delivery of the examination was only established at about half of the dental schools (15/27, 56%).</p></sec><sec id="s3-4"><title>Delivery of Examinations</title><p>The percentage of dental schools that deliver examinations electronically amounted to 70% (19/27). However, the software used by the dental schools differed: a dedicated examination software (ie, UCAN&#x2019;s CAMPUS or tEXAM, Q-Exam [IQUL GmbH]) was used by 8 dental schools, while learning management systems such as Moodle (Moodle Pty Ltd), ILIAS (ILIAS open source e-Learning e.V.), or OpenOLAT (frentix GmbH) were used by 7 dental schools for the purpose of examination delivery. The remaining 4 dental schools did not provide any information regarding the examination software they used.</p><p>Dental schools with the possibility of electronic examinations used multiple-select items slightly more often (14/19, 74% vs 4/8, 50%). However, this difference was statistically not significant (<italic>P</italic>=.38).</p></sec><sec id="s3-5"><title>Applied Scoring Methods</title><p>All dental schools scored single-choice items (ie, Type A, Type A<sub>negative</sub>, Type K, Type R, and alternate-choice) dichotomously (ie, 1 full credit point is awarded if examinees mark the correct answer option or statements, otherwise no credit is awarded).</p><p>Scoring of multiple-select items was more heterogeneous and no single scoring method that was commonly used was identified: some dental schools used scoring algorithms resulting in partial (ie, 0.5 credit points) or intermediate partial credit (ie, 1/n partial credit for each correct response) besides dichotomous scoring on multiple-select items. However, scoring methods resulting in negative points (ie, malus points) were not used at any location.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>The aim of this study was to gain insight into summative assessment methods that involve the use of multiple-choice items and are used at German dental schools. The purpose of summative assessment is to evaluate examinees&#x2019; knowledge at the end of a course by comparing their scores to a predefined standard (ie, cutoff score) [<xref ref-type="bibr" rid="ref14">14</xref>]. Our results demonstrate that all respondent dental schools use multiple-choice examinations for summative assessment of theoretical knowledge. Besides individual items, approximately half of the dental schools also use key feature problems.</p><p>Single-choice Type A items are the most popular item types used at German dental schools. These items are used by almost every respondent dental school and often account for the largest number of items at the respective dental schools. This might be explained by the demand for ease of scoring (ie, dichotomous scoring, no partially correct responses).</p><p>Multiple-select item types such as Pick-N or Multiple-True-False are used by fewer dental schools. For these item types, the applied scoring methods vary considerably: Some dental schools award partial or even intermediate partial credit for partially correct responses while others do not. However, the exact cutoff levels and scoring methods for partial credit differed. For example, Partial Scoring 50% (PS<sub>50</sub>) was used by some dental schools for Pick-N items: In these cases, 1 full credit point is awarded if all answer options are marked correctly, and 0.5 credit points are awarded if at least half of the true answer options are marked, otherwise no credit is awarded [<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref15">15</xref>]. Furthermore, a similar scoring method named Half-point Scoring was used by some dental schools for Multiple-True-False and conventional multiple-select items: 1 full credit point is awarded if all statements or answer options are marked correctly, 0.5 credit points are awarded if the response to 1 statement or answer option is incorrect, otherwise no credit is awarded [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref16">16</xref>]. In addition, some dental schools awarded intermediate partial credit on multiple-select items: In the case of Partial Scoring 1/n (PS<sub>1/n</sub>), 1/n partial credit was awarded for each correct response [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>]. Some dental schools also subtracted 1/n partial credit for each incorrect response (Blasberg-Method) [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref17">17</xref>].</p><p>As a result, the scoring of multiple-select items at different German dental schools can be considered very heterogeneous. This is not surprising, as a vast number of different scoring methods for multiple-select items have been described in the literature [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>]. As stated previously, it is not possible to suggest a single versatile scoring method. Different requirements as defined in dental schools&#x2019; local examination guidelines (eg, fixed pass-mark and fixed proportion of true answer options) impact the scoring method to be selected. Regarding jurisdictional requirements, scoring methods resulting in negative points (ie, malus points) must not be used in Germany [<xref ref-type="bibr" rid="ref13">13</xref>]. Consequently, not a single dental school uses scoring methods resulting in malus points. However, almost half of the dental schools do not use a formal item review process. A formal review process is recommended prior to the delivery of the examinations and might further improve the quality and overall validity of the examinations.</p><p>In addition, 70% (19/27) of all dental schools stated to deliver examinations electronically. While the electronic delivery of examinations allows for automatic scoring and more complex scoring methods (ie, within the context of multiple-select items), no statistically significant relation between the type of delivery (paper-based vs electronic) and the use of multiple-select item types was found. Therefore, our results fail to reject the null hypothesis. This might be explained by the software used for the delivery and scoring of electronic examinations: it was found that dental schools use learning management systems such as Moodle, ILIAS, or OpenOLAT besides dedicated examination software such as UCAN&#x2019;s CAMPUS, UCAN&#x2019;s tEXAM, or Q-Exam for the delivery and scoring of summative assessments. This is of relevance, as learning management systems usually support fewer item types and scoring methods than dedicated examination software [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>]. As a result, electronic delivery of examinations does not necessarily result in an increased use of multiple-select items.</p><p>Interestingly, not a single dental school used alternative testing methods that deviate from the standard setting during examinations (ie, examinees mark the answer options or statements they believe to be correct or true but receive no immediate feedback regarding correctly or incorrectly marked answer options or statements). Within multiple-choice examinations, alternative testing methods such as confidence weighting scoring (ie, examinees are requested to indicate the degree of confidence in their marking) [<xref ref-type="bibr" rid="ref18">18</xref>], elimination scoring (ie, examinees are instructed to mark the incorrect instead of correct answer options) [<xref ref-type="bibr" rid="ref19">19</xref>], or answer-until-correct [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref21">21</xref>] have been described in the literature. Within the answer-until-correct method, examinees receive immediate feedback and examinees may correct their marking on previously incorrectly marked items, thereby still receiving partial credit. However, the benefit of such testing methods within the field of dental education is questionable. Dental school examinees are becoming future dentists. While treating patients, dentists are required to make informed choices and dentists might not always have a second chance without potentially harming their patients. In addition, such alternative testing methods benefit from the electronic delivery of examinations and set even higher requirements for the used examination software.</p></sec><sec id="s4-2"><title>Strengths and Limitations</title><p>To the best of our knowledge, this is the first study to systematically assess the use and scoring of multiple-choice item types in summative examinations among German dental schools. A number of strengths are present. First, a pretested questionnaire was used. Second, our questionnaire survey study yielded a high response rate of 90% (27/30 dental schools). Third, our results might be considered representative of the current use of multiple-choice items in summative examinations among German dental schools.</p><p>Nevertheless, limitations are also present. First, our questionnaire focused on multiple-choice items; therefore, the use of other assessment types (eg, objective structured clinical examinations, oral examinations) remains unknown. Second, the number of dental schools in Germany is limited. Thereby, results from the Fisher exact test might be underpowered despite the high response rate. Furthermore, this study could not control for potential confounders (eg, location, number of students per dental school) due to the overall low number of dental schools. Third, transferability and generalizability to other educational settings might be limited due to different jurisdictional requirements or the overall lower importance of written examinations.</p></sec><sec id="s4-3"><title>Future Directions</title><p>New dental licensing regulations (&#x201C;Approbationsordnung&#x201D;) have been in effect since 2021, which restructured the undergraduate dental curriculum in Germany. For the first time, a nationwide written board examination with single-choice items takes place at the end of all undergraduate dental programs (ie, after the 10th semester) [<xref ref-type="bibr" rid="ref22">22</xref>]. Therefore, multiple-choice examinations in general and especially single-choice Type A items will remain a popular format for summative examinations among German undergraduate dental programs. Ideally, examinees already become familiar with single-choice Type A items during their studies. Therefore, all dental schools should use single-choice Type A items to adequately prepare their students for the final board examination.</p><p>Nevertheless, additional examinations (eg, objective structured clinical or practical examinations) are required to test examinees&#x2019; practical skills [<xref ref-type="bibr" rid="ref3">3</xref>]. Regardless of the used item type, multiple-choice examinations are not suitable to assess the higher levels Miller&#x2019;s Pyramid of clinical competence (ie, does and shows how) [<xref ref-type="bibr" rid="ref23">23</xref>].</p></sec><sec id="s4-4"><title>Conclusion</title><p>While students from almost all dental schools can be expected to be familiar with single-choice Type A items, techniques for the summative assessment of theoretical knowledge differ widely among German dental schools. Especially, a large variability regarding the use and scoring of multiple-select multiple-choice items was found. In addition, implementing a formal item review process might further improve the quality and overall validity of the examinations.</p></sec></sec></body><back><ack><p>The authors acknowledge support from the Open Access Publication Funds of G&#x00F6;ttingen University. The funder had no role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. Generative artificial intelligence was not used in any portion of this manuscript.</p></ack><notes><sec><title>Data Availability</title><p>All data generated during or analyzed during this study are included in this published article.</p></sec></notes><fn-group><fn fn-type="con"><p>LR, MH, AW, and PK contributed to the study&#x2019;s conception and designed the questionnaire. LR and PK transferred the data. PK performed statistical analyses. All authors interpreted the data, critically revised the manuscript, and approved the final version of the manuscript.</p></fn><fn fn-type="conflict"><p>PK is an associate editor of <italic>JMIR Medical Education</italic> at the time of this publication. Other authors have no competing interests to declare.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">PS<sub>1/n</sub></term><def><p>Partial Scoring 1/n</p></def></def-item><def-item><term id="abb2">PS<sub>50</sub></term><def><p>Partial Scoring 50%</p></def></def-item><def-item><term id="abb3">UCAN</term><def><p>Umbrella Consortium for Assessment Networks</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kelly</surname><given-names>FJ</given-names> </name></person-group><article-title>The Kansas silent reading tests</article-title><source>J Educ Psychol</source><year>1916</year><month>02</month><volume>7</volume><issue>2</issue><fpage>63</fpage><lpage>80</lpage><pub-id pub-id-type="doi">10.1037/h0073542</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ruch</surname><given-names>GM</given-names> </name><name name-style="western"><surname>Stoddard</surname><given-names>GD</given-names> </name></person-group><article-title>Comparative reliabilities of five types of objective examinations</article-title><source>J Educ Psychol</source><year>1925</year><month>03</month><volume>16</volume><issue>2</issue><fpage>89</fpage><lpage>103</lpage><pub-id pub-id-type="doi">10.1037/h0072894</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gerhard-Szep</surname><given-names>S</given-names> </name><name name-style="western"><surname>G&#x00FC;ntsch</surname><given-names>A</given-names> </name><name name-style="western"><surname>Pospiech</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Assessment formats in dental medicine: an overview</article-title><source>GMS J Med Educ</source><year>2016</year><month>08</month><volume>33</volume><issue>4</issue><fpage>Doc65</fpage><pub-id pub-id-type="doi">10.3205/zma001064</pub-id><pub-id pub-id-type="medline">27579365</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kanzow</surname><given-names>P</given-names> </name><name name-style="western"><surname>Schuelper</surname><given-names>N</given-names> </name><name name-style="western"><surname>Witt</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Effect of different scoring approaches upon credit assignment when using Multiple True-False items in dental undergraduate examinations</article-title><source>Eur J Dent Educ</source><year>2018</year><month>11</month><volume>22</volume><issue>4</issue><fpage>e669</fpage><lpage>e678</lpage><pub-id pub-id-type="doi">10.1111/eje.12372</pub-id><pub-id pub-id-type="medline">29934980</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kanzow</surname><given-names>P</given-names> </name><name name-style="western"><surname>Schmidt</surname><given-names>D</given-names> </name><name name-style="western"><surname>Herrmann</surname><given-names>M</given-names> </name><name name-style="western"><surname>Wassmann</surname><given-names>T</given-names> </name><name name-style="western"><surname>Wiegand</surname><given-names>A</given-names> </name><name name-style="western"><surname>Raupach</surname><given-names>T</given-names> </name></person-group><article-title>Use of multiple-select multiple-choice items in a dental undergraduate curriculum: retrospective study involving the application of different scoring methods</article-title><source>JMIR Med Educ</source><year>2023</year><month>03</month><day>27</day><volume>9</volume><fpage>e43792</fpage><pub-id pub-id-type="doi">10.2196/43792</pub-id><pub-id pub-id-type="medline">36841970</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Krebs</surname><given-names>R</given-names> </name></person-group><source>Pr&#x00FC;fen mit Multiple Choice: Kompetent planen, entwickeln, durchf&#x00FC;hren und auswerten [Testing With Multiple Choice: Plan, Develop, Implement, and Evaluate Competently]</source><year>2019</year><publisher-name>Hogrefe</publisher-name><pub-id pub-id-type="other">9783456859026</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kanzow</surname><given-names>AF</given-names> </name><name name-style="western"><surname>Schmidt</surname><given-names>D</given-names> </name><name name-style="western"><surname>Kanzow</surname><given-names>P</given-names> </name></person-group><article-title>Scoring single-response multiple-choice items: scoping review and comparison of different scoring methods</article-title><source>JMIR Med Educ</source><year>2023</year><month>05</month><day>19</day><volume>9</volume><fpage>e44084</fpage><pub-id pub-id-type="doi">10.2196/44084</pub-id><pub-id pub-id-type="medline">37001510</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schmidt</surname><given-names>D</given-names> </name><name name-style="western"><surname>Raupach</surname><given-names>T</given-names> </name><name name-style="western"><surname>Wiegand</surname><given-names>A</given-names> </name><name name-style="western"><surname>Herrmann</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kanzow</surname><given-names>P</given-names> </name></person-group><article-title>Relation between examinees&#x2019; true knowledge and examination scores: systematic review and exemplary calculations on Multiple-True-False items</article-title><source>Educ Res Rev</source><year>2021</year><month>11</month><volume>34</volume><fpage>100409</fpage><pub-id pub-id-type="doi">10.1016/j.edurev.2021.100409</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schmidt</surname><given-names>D</given-names> </name><name name-style="western"><surname>Raupach</surname><given-names>T</given-names> </name><name name-style="western"><surname>Wiegand</surname><given-names>A</given-names> </name><name name-style="western"><surname>Herrmann</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kanzow</surname><given-names>P</given-names> </name></person-group><article-title>Relation between examinees&#x2019; true knowledge and examination scores: systematic review and exemplary calculations on Pick-N items</article-title><source>Educ Res Rev</source><year>2022</year><month>11</month><volume>37</volume><fpage>100483</fpage><pub-id pub-id-type="doi">10.1016/j.edurev.2022.100483</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nardi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ranieri</surname><given-names>M</given-names> </name></person-group><article-title>Comparing paper&#x2010;based and electronic multiple&#x2010;choice examinations with personal devices: impact on students&#x2019; performance, self&#x2010;efficacy and satisfaction</article-title><source>Brit J Educ Tech</source><year>2019</year><month>05</month><volume>50</volume><issue>3</issue><fpage>1495</fpage><lpage>1506</lpage><pub-id pub-id-type="doi">10.1111/bjet.12644</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Way</surname><given-names>WD</given-names> </name><name name-style="western"><surname>Davis</surname><given-names>LL</given-names> </name><name name-style="western"><surname>Keng</surname><given-names>L</given-names> </name><name name-style="western"><surname>Strain-Seymour</surname><given-names>E</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Drasgow</surname><given-names>F</given-names> </name></person-group><article-title>Increasing the accessibility of assessments through technology</article-title><source>Technology and Testing: Improving Educational and Psychological Measurement</source><year>2015</year><publisher-name>Routledge</publisher-name><fpage>217</fpage><lpage>234</lpage><pub-id pub-id-type="other">9781315871493</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>J&#x00FC;nger</surname><given-names>J</given-names> </name><name name-style="western"><surname>Just</surname><given-names>I</given-names> </name></person-group><article-title>Recommendations of the German Society for Medical Education and the German Association of Medical Faculties regarding university-specific assessments during the study of human, dental and veterinary medicine</article-title><source>GMS Z Med Ausbild</source><year>2014</year><month>08</month><volume>31</volume><issue>3</issue><fpage>Doc34</fpage><pub-id pub-id-type="doi">10.3205/zma000926</pub-id><pub-id pub-id-type="medline">25228936</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kubinger</surname><given-names>KD</given-names> </name></person-group><article-title>Gutachten zur Erstellung &#x201E;gerichtsfester&#x201C; Multiple-Choice-Pr&#x00FC;fungsaufgaben</article-title><source>Psychol Rundschau</source><year>2014</year><month>07</month><volume>65</volume><issue>3</issue><fpage>169</fpage><lpage>178</lpage><pub-id pub-id-type="doi">10.1026/0033-3042/a000218</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Scriven</surname><given-names>M</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Tyler</surname><given-names>RW</given-names> </name><name name-style="western"><surname>Gagn&#x00E9;</surname><given-names>RM</given-names> </name><name name-style="western"><surname>Scriven</surname><given-names>M</given-names> </name></person-group><article-title>The methodology of evaluation</article-title><source>Perspectives of Curriculum Evaluation</source><year>1967</year><publisher-name>Rand McNally</publisher-name><pub-id pub-id-type="other">9780528618109</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bauer</surname><given-names>D</given-names> </name><name name-style="western"><surname>Holzer</surname><given-names>M</given-names> </name><name name-style="western"><surname>Kopp</surname><given-names>V</given-names> </name><name name-style="western"><surname>Fischer</surname><given-names>MR</given-names> </name></person-group><article-title>Pick-N multiple choice-exams: a comparison of scoring algorithms</article-title><source>Adv Health Sci Educ Theory Pract</source><year>2011</year><month>05</month><volume>16</volume><issue>2</issue><fpage>211</fpage><lpage>221</lpage><pub-id pub-id-type="doi">10.1007/s10459-010-9256-1</pub-id><pub-id pub-id-type="medline">21038082</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Vorkauf</surname><given-names>H</given-names> </name></person-group><article-title>Teilpunktbewertung bei K&#x2019;-Items [Partial credit scoring of Multiple-True-False items]</article-title><source>Jahresbericht 1986 der Gruppe Medizinalpr&#x00FC;fungen und der Gruppe Statistik und EDV</source><year>1987</year><publisher-name>Institut f&#x00FC;r Ausbildungs- und Examensforschung, Medizinische Fakult&#x00E4;t der Universit&#x00E4;t Bern</publisher-name><fpage>44</fpage><lpage>48</lpage></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Blasberg</surname><given-names>R</given-names> </name><name name-style="western"><surname>G&#x00FC;ngerich</surname><given-names>U</given-names> </name><name name-style="western"><surname>M&#x00FC;ller-Esterl</surname><given-names>W</given-names> </name><name name-style="western"><surname>Neumann</surname><given-names>D</given-names> </name><name name-style="western"><surname>Schappel</surname><given-names>S</given-names> </name></person-group><article-title>Erfahrungen mit dem Fragentyp &#x201E;k aus n&#x201C; in Multiple-Choice-Klausuren [Experiences with item type &#x201C;k from n&#x201D; in multiple-choice-tests]</article-title><source>Med Ausbild</source><year>2001</year><volume>18</volume><issue>S1</issue><fpage>73</fpage><lpage>76</lpage></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hevner</surname><given-names>K</given-names> </name></person-group><article-title>A method of correcting for guessing in true-false tests and empirical evidence in support of it</article-title><source>J Soc Psychol</source><year>1932</year><month>08</month><volume>3</volume><issue>3</issue><fpage>359</fpage><lpage>362</lpage><pub-id pub-id-type="doi">10.1080/00224545.1932.9919159</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Collet</surname><given-names>LS</given-names> </name></person-group><article-title>Elimination scoring: an empirical evaluation</article-title><source>J Educ Meas</source><year>1971</year><month>09</month><volume>8</volume><issue>3</issue><fpage>209</fpage><lpage>214</lpage><pub-id pub-id-type="doi">10.1111/j.1745-3984.1971.tb00927.x</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gilman</surname><given-names>DA</given-names> </name><name name-style="western"><surname>Ferry</surname><given-names>P</given-names> </name></person-group><article-title>Increasing test reliability through self-scoring procedures</article-title><source>J Educ Meas</source><year>1972</year><month>09</month><volume>9</volume><issue>3</issue><fpage>205</fpage><lpage>207</lpage><pub-id pub-id-type="doi">10.1111/j.1745-3984.1972.tb00953.x</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hanna</surname><given-names>GS</given-names> </name></person-group><article-title>Incremental reliability and validity of multiple-choice tests with an answer-until-correct procedure</article-title><source>J Educ Meas</source><year>1975</year><month>09</month><volume>12</volume><issue>3</issue><fpage>175</fpage><lpage>178</lpage><pub-id pub-id-type="doi">10.1111/j.1745-3984.1975.tb01019.x</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Schmitz</surname><given-names>UJ</given-names> </name><name name-style="western"><surname>Daubl&#x00E4;nder</surname><given-names>M</given-names> </name></person-group><article-title>&#x00DC;bertragung papierbasierter Multiple Choice Aufgaben in die digitale Welt &#x2013; ein Weg zur Verbesserung der Pr&#x00FC;fungsqualit&#x00E4;t? [Transferring paper-based multiple-choice items to the digital world &#x2013; a way to improve examination quality?] [Abstract]</article-title><source>Jahrestagung der Gesellschaft f&#x00FC;r Medizinische Ausbildung (GMA)</source><year>2022</year><publisher-name>German Medical Science GMS Publishing House</publisher-name></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Miller</surname><given-names>GE</given-names> </name></person-group><article-title>The assessment of clinical skills/competence/performance</article-title><source>Acad Med</source><year>1990</year><month>09</month><volume>65</volume><issue>9 Suppl</issue><fpage>S63</fpage><lpage>S67</lpage><pub-id pub-id-type="doi">10.1097/00001888-199009000-00045</pub-id><pub-id pub-id-type="medline">2400509</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Authors&#x2019; translation of the used questionnaire, which was originally distributed in German.</p><media xlink:href="mededu_v10i1e58126_app1.pdf" xlink:title="PDF File, 247 KB"/></supplementary-material></app-group></back></article>