Published on in Vol 11 (2025)

This is a member publication of University of Birmingham (Jisc)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/52332, first published .
Global Disparities in Simulation-Based Learning Performance: Serial Cross-Sectional Mixed Methods Study

Global Disparities in Simulation-Based Learning Performance: Serial Cross-Sectional Mixed Methods Study

Global Disparities in Simulation-Based Learning Performance: Serial Cross-Sectional Mixed Methods Study

1Department of Surgery, Dayanand Medical College, Punjab, India

2Department of Applied Health Sciences, College of Medical and Dental Sciences, University of Birmingham, Edgbaston, Birmingham, United Kingdom

3College of Medical and Dental Sciences, University of Birmingham, Birmingham, United Kingdom

4Northwick Park Hospital, London North West University Healthcare NHS Trust, London, United Kingdom

5Queen Elizabeth Hospital, University Hospitals Birmingham NHS Foundation Trust, Birmingham, United Kingdom

*these authors contributed equally

Corresponding Author:

Punith Kempegowda, PhD


Background: Simulated programs provide health care professionals (HCPs) with a learning opportunity to develop clinical competencies and improve patient outcomes in a safe and controlled environment. While the benefits of simulation training are well established, there is a paucity of research assessing its differential impact, if any. SIMBA (Simulation via Instant Messaging for Bedside Application) provides simulation-based learning through WhatsApp and Zoom (Zoom Video Communications, Inc) to increase HCPs’ confidence in managing various medical conditions.

Objectives: This study aims to explore whether there are differences in the clinical performance of HCPs participating in SIMBA sessions based on gender, country of work, and training grade.

Methods: This study assessed participants in 17 SIMBA sessions from May 2020 to June 2022. WhatsApp chats containing participants’ approach to the simulated scenarios were graded using an adapted version of the Global Rating Scale consisting of 6 domains: eliciting history; physical examination; investigations, diagnostic tests, and imaging; interpretation of investigations and imaging; clinical judgment; and management and follow-up or discharge plan. These domains were rated using a Likert-type scale of 1 (not done) to 5 (excellent) prior to the session based on expert inputs. All WhatsApp transcripts were evaluated against the scale postsimulation session. Unadjusted and adjusted means and 95% CIs of the scores for the 6 performance variables were calculated using multiple linear regression models. The P value for heterogeneity between the mean performance scores was calculated using likelihood ratio tests by using an analysis of variance.

Results: A total of 289 participants across 49 countries who completed pre-SIMBA and post-SIMBA surveys in the 17 simulation sessions were included in the analysis. Participants from high-income countries scored higher in all categories of the Global Rating Scale (GRS) except the physical examination and interpretation score. Junior-grade participants scored significantly higher in history taking (junior=4.2, middle=3.7, and senior=3.7; P=.003) and physical examination (junior=4.0, middle=3.7, and senior=3.5; P=.068), but this was not significantly different. There were no statistically significant differences in GRS scores between male and female participants.

Conclusions: The significant differences in clinical performance scores between low- and middle-income countries and high-income countries highlight the need for better medical education resources to bridge existing gaps in health care globally. The decrease in some clinical competency scores following career progression could be addressed by simulation-based training to maintain the same quality of history taking and physical examination skills. These outcomes, including no gendered differences in simulation-based learning, hold profound implications for tailoring medical education strategies, fostering equitable training, and elevating patient care standards on a global scale. The need for targeted interventions and capacity-building efforts via context-specific training and tailored approaches to health care education is emphasized.

JMIR Med Educ 2025;11:e52332

doi:10.2196/52332

Keywords



Understanding the multifaceted global health care disparities is crucial for ensuring access to high-quality medical education [1]. The most common barriers to high-quality medical education are due to financial, resource, and accessibility inequities [2]. There is a shortage in the training resources to recruit and retain health care professionals (HCPs) in low- and middle-income countries (LMICs), leading to disparities in clinical learning opportunities compared with HCPs in high-income countries (HICs) [3]. This was further exacerbated following the COVID-19 pandemic [4,5].

E-learning and distance virtual simulation sessions can provide case-based learning and support the training of medical professionals in LMICs [6]. Specifically, initiatives using accessible technology platforms to develop high-quality educational programs can enhance postgraduate health care training [7,8]. Simulated programs provide HCPs with a learning opportunity to develop clinical competencies and improve patient outcomes in a safe and controlled environment [9,10]. However, there is an unequal distribution of simulation-based learning across LMICs, due to misconceptions regarding the cost and availability of programs [11]. Also, while the benefits of simulation training are well established, there is a paucity of research studying the differences, if any, in the clinical performance of HCPs from HICs and LMICs in simulation-based case scenarios. Understanding the similarities and differences in clinical performance can help identify areas for improvement in health care education and resource allocation in LMICs.

SIMBA (Simulation via Instant Messaging for Bedside Application) is a simulation-based learning model that uses WhatsApp and Zoom (Zoom Video Communications, Inc) platforms to improve HCPs’ confidence in managing various medical conditions [12,13]. The model is built on the principles of Kolb’s experiential learning theory and simulation gaming theory [14]. Kolb’s experiential learning theory emphasizes the importance of learning through reflection on doing, where learners actively engage in a cycle of concrete experience, reflective observation, abstract conceptualization, and active experimentation. This framework guided the design of our SIMBA sessions, ensuring that participants were exposed to real-world scenarios and encouraged to reflect on their actions and apply new knowledge in future simulations. Simulation gaming theory further shaped our curriculum by integrating gamelike elements to create an immersive and interactive learning environment. Using realistic clinical cases, role-playing, and feedback loops in our sessions aligns with the theory’s emphasis on learning through interaction and problem-solving within a simulated context. This approach fosters engagement and enhances skill development, making the learning process dynamic and participatory. It helped deliver and sustain good quality medical education during the COVID-19 pandemic [15]. It also provided alternative work experiences for premedical students [16] and a platform for medical students and junior doctors to improve teamwork and leadership skills [17]. Our previous work described how SIMBA provided equitable and high-quality postgraduate medical education to health care physicians in the LMICs and HICs [18].

As health care systems strive to provide safe and effective medical services, understanding the nuances of clinical competence among different training grades becomes a pivotal aspect of medical education and professional development. Exploring the clinical performance variations across various stages of HCP training offers valuable insights into how educational strategies and experiential learning impact the acquisition and application of clinical competencies. The need for such exploration is underscored by the potential ramifications these findings may have on patient care. An HCP’s ability to accurately diagnose, interpret diagnostic tests, make informed clinical judgments, and formulate appropriate management plans can vary across different training grades. These variations can impact patient safety, the quality of health care services, and the overall effectiveness of medical teams [19-21]. In addition, as health care systems increasingly seek to bridge global disparities in medical education and patient outcomes, it is imperative to identify how training grade–related differences influence clinical performance across diverse settings.

In this study, we explore the differences in the performance and outcomes of HCPs participating in SIMBA sessions based on gender, country of work, and training grade. This study addresses this critical knowledge gap by examining the disparities among training grades within a single context and how the geographic context of HICs and LMICs influences these disparities. The inclusion of gender as a parameter is aligned with the aspiration to tailor medical education to the diverse needs of HCPs, while also acknowledging the broader societal implications of gender-sensitive health care delivery. The study’s findings hold the potential to influence medical training practices, curriculum development, and professional mentorship in ways that are more attuned to the nuances of gender-specific experiences and perspectives.


Conducting SIMBA Sessions

The preparation and delivery of SIMBA sessions, including detailed flowcharts and examples, have been described in detail in our previous publication [12]. Each SIMBA session featured 4-6 clinical case scenarios representing a range of medical presentations commonly encountered in secondary care. These sessions were promoted via social media, junior doctor bulletins, and supporting organizations’ websites.

The case scenarios were adapted from real-life cases with all patient-identifiable information removed. They included comprehensive details such as presenting complaints, medical history, examination findings, clinical observations, investigation results (eg, blood tests and imaging), differential diagnoses, management strategies, and follow-up plans. Experts chairing the sessions rigorously reviewed and approved the case transcripts to ensure scientific accuracy and alignment with current medical guidelines.

Participation in the sessions was voluntary and free of charge. Participants registered in advance by providing their email addresses and WhatsApp numbers. Emails were used for presession communication, while WhatsApp facilitated real-time interaction during the simulations, providing access to all necessary session details and resources.

Before the simulation, participants completed a presimulation survey, which included informed consent, sociodemographic data collection, and self-assessment of their confidence in managing various clinical scenarios. This survey established a baseline measure of their confidence levels, covering both the session scenarios and the similar medical presentations.

During the simulation, HCPs from HICs and LMICs participated concurrently via WhatsApp. Moderators guided participants through the scenarios, encouraging them to interact as they would during real-life patient encounters. WhatsApp was also used to share key resources, such as links to surveys, investigations, and meeting invitations.

After completing all simulated cases, participants joined a Zoom meeting led by an expert. This interactive session provided an opportunity to discuss the cases, ask questions, and gain additional insights. Experts facilitated peer-to-peer discussions in the Zoom chat, fostering collaborative learning and ensuring a thorough exploration of each case.

After the simulation, participants completed a postsimulation survey, which included the confidence-rating questions from the presurvey to assess changes in confidence levels. The survey also collected feedback on their experience, highlighting successful aspects of the session and identifying areas for improvement.

WhatsApp interactions were recorded and later analyzed. Participants’ performance was objectively assessed using a GRS [22], validated by experts prior to the session. The scale evaluated 6 domains: eliciting history, physical examination, investigations and diagnostic tests, interpretation of findings, clinical judgment, and management and follow-up plans. Each domain was scored on a Likert-type scale from 1 (not done) to 5 (excellent) [23], with total scores ranging from 6 to 30. This assessment measured participants’ knowledge, accuracy, and thoroughness in handling the cases.

In addition, moderators provided personalized feedback using Pendleton’s feedback model [24,25]. Feedback highlighted participants’ strengths and offered constructive guidance for improvement, aiming to enhance their skills and confidence in managing similar scenarios in the future.

Data Collection

This study assessed participants in 17 SIMBA sessions from May 2020 to June 2022. The topics simulated included a wide range of medical specialties, as seen in Figure 1. All participants were invited to complete both the pre-SIMBA and post-SIMBA surveys voluntarily. The pre-SIMBA survey, distributed just before the session began, gathered basic sociodemographic information and self-reported confidence levels in managing simulated cases. The post-SIMBA survey, shared immediately after the expert case discussion, included similar questions on self-reported confidence in managing these cases. It also sought participants’ feedback on their experience in the session. Each survey was administered at a single time point rather than as part of a test-retest design.

In addition, following each SIMBA session, WhatsApp chats containing the participants’ approach to the simulated scenarios were graded using an adapted version of the GRS [22], which is reviewed by experts to confirm their appropriateness for the simulated case. Our assessment tool consisted of 6 domains (eliciting history; physical examination; investigations, diagnostic tests, and imaging; interpretation of investigations and imaging; clinical judgment; and management and follow-up or discharge plan), which are rated using a Likert-type scale of 1 (not done) to 5 (excellent) [23]. Moderators calculated the participant’s score between a minimum of 6 and a maximum of 30, measuring the participant’s knowledge, accuracy, and completeness in approaching the simulated case and assessing the patient. In addition to a numerical score, moderators provided participants with written feedback based on Pendleton’s feedback model [24,25]. Pendleton’s model is a structured approach to feedback that aims to create a positive and constructive learning environment. The model emphasizes a balanced feedback process by first highlighting the learner’s strengths, then identifying areas for improvement, and finally engaging the learner in self-reflection. This structured feedback process encourages active learning and helps participants gain a deeper understanding of their performance, making it an effective tool for clinical education. GRS scores and feedback were emailed to participants 24‐48 hours after the session. In the case of any queries, participants were provided with the contact details of the SIMBA team. To ensure uniformity and fairness across scoring, all moderators attended training sessions where they were taught how to use our assessment tool by experienced moderators. Any uncertainties that arised during the scoring process were resolved by discussion with experienced moderators.

Figure 1. Timeline of SIMBA (Simulation via Instant Messaging for Bedside Application) sessions held from 2020 to 2022. PCOS: polycystic ovary syndrome.

Statistical Analysis

Participants who completed both pre-SIMBA and post-SIMBA surveys were included in the analysis. This pre-post test design with presurvey and postsurvey specifically focused on assessing changes over time rather than the consistency of responses. Participants self-reported their country of residence during surveys, and the data (countries) were grouped into HICs and LMICs according to their country of residence based on the 2022 World Bank report [26]. Participants were classified as junior, middle, and senior grades based on their work experience. Junior-, middle-, and senior-grade HCPs had 0‐2 years, 2‐5 years, and >5 years of work experience, respectively.

Unadjusted and adjusted means and 95% CIs of the scores for the 6 performance variables (history taking, physical examination, investigation, interpretation, judgment, and management) by economic classification (HIC or LMIC), level of training (junior, middle, or senior grade), and gender (male or female) were calculated using multiple linear regression models, adjusting for gender, country, training, and number of WhatsApp messages (continuous), as appropriate. The mean (95% CI) difference of the pairwise comparisons between each performance score was calculated using paired t tests. The P value for heterogeneity between the mean performance scores was calculated using likelihood ratio tests by using an ANOVA. Two-sided P values of <.05 were considered statistically significant. We did not adjust the P value for multiple comparisons. All statistical analyses were performed using the Stata statistical software package (version 27.0; StataCorp).

Ethical Considerations

SIMBA was approved as an educational modality for postgraduate medical education by the Health Education West Midlands Diabetes and Endocrinology Specialist Training Committee. The study was approved by the Science, Technology, Engineering and Mathematics Ethics Committee at the University of Birmingham (ERN_2023‐0495 and ERN_2023‐0544). The study complied with all relevant ethical guidelines and institutional regulations throughout its design, implementation, analysis, and dissemination. Informed consent was taken from each participant and participation was voluntary. Participants were provided with an information sheet outlining the purpose of the study, the estimated duration of their involvement, the nature of the data collected, the identity of the investigators, and the storage and retention details. All data were anonymized at the point of analysis and stored on a secure, restricted access online drive. No identifiable patient data were used. No financial or material compensation was offered for participation.


Overview

A total of 289 participants completed both pre-SIMBA and post-SIMBA surveys in the 17 simulation sessions. A total of 281 participants provided their country of residence. These participants worked in 49 countries across 5 continents, of which 36.7% (18/49) were classified as HICs, whereas 63.2% (31/49) were LMICs. Most participants worked in HICs (HICs: 191/281, 67.9%; LMICs: 90/281, 32.0%). Overall, the mean (SD) scores for the cohort were history taking: 3.6 (SD 3.9), physical examination: 3.5 (SD 3.7), investigation: 3.3 (SD 3.5), interpretation: 2.6 (SD 2.9), judgment: 3.1 (SD 3.3), and management: 2.6 (SD 2.8) (Table 1).

The values with the footnotes are the mean (95% CIs) for that particular score. Other values and corresponding P values are the mean difference (95% CIs) for the pairwise comparisons between scores taken from a paired t test for the 264 participants who had a value for all 6 outcome scores. The difference is for the performance measure on the column compared with the performance measure on the corresponding row.

Table 1. Mean and 95% CIs difference between various performance measures using the global rating score for those participating in Simulation via Instant Messaging for Bedside Application (SIMBA).
HistoryP valuePhysical examinationP valueInvestigationP valueInterpretationP valueJudgmentP valueManagement
History3.7 (3.6‐3.9)a
Physical examination0.1 (0.0‐0.2).023.6 (3.5‐3.7)a
Investigation0.3 (0.2‐0.4)<.010.2 (0.1‐0.3)<.013.4 (3.3‐3.5)a
Interpretation1.0 (0.8‐1.2)<.010.9 (0.7‐1.1)<.010.7 (0.6‐0.8)<.012.7 (2.6‐2.9)a
Judgment0.5 (0.4‐0.6)<.010.4 (0.3‐0.5)<.010.2 (0.1‐0.3)<.01−0.5 (−0.6 to 0.3)<.013.2 (3.1‐3.3)a
Management1.0 (0.9‐1.2)<.010.9 (0.8‐1.1)<.010.7 (0.6‐0.8)<.010.0 (−0.1 to 0.2).520.5 (0.4‐0.6)<.012.7 (2.6‐2.8)a

aThese values are the mean (95% CIs) for that particular score.

Comparing the Performance Between Participants From HIC and LMIC

After adjusting for sex, training, and number of WhatsApp messages, there were statistically significant differences in performance that were identified in the categories of history taking (HIC vs LMIC: 3.8 [3.7‐3.9] vs 3.5 [3.3‐3.7]; P<.01), investigations (3.6 [3.5‐3.7] vs 3.1 [2.9‐3.2]; P<.01), clinical judgment (3.4 [3.3‐3.5] vs 2.9 [2.7‐3.1]; P<.01), and management (2.9 [2.7‐3.0] vs 2.3 [2.1‐2.5]; P<.01) (Table 2 and Multimedia Appendix 1). A slight difference, although not statistically significant, was also found in physical examination (3.7 [3.5‐3.8] vs 3.5 [3.2‐3.7]; P=.13) and interpretation (2.8 [2.6‐3.0] vs 2.6 [2.3‐2.9]; P=.23).

Table 2. Mean and 95% CIs of various performance measures using the global rating score for low- or middle- and high-income countries for those participating in the SIMBA (Simulation via Instant Messaging for Bedside Application).
Performance measureUnadjusted and adjustedLow and middle incomeHigh incomeP valuea
History takingUnadjusted3.5 (3.2‐3.7)3.8 (3.7‐4.0)<.01
Adjusteda3.5 (3.3‐3.7)3.8 (3.7‐3.9)<.01
Physical examinationUnadjusted3.4 (3.2‐3.6)3.7 (3.5‐3.8).05
Adjusteda3.5 (3.2‐3.7)3.7 (3.5‐3.8).13
InvestigationUnadjusted3.1 (2.9‐3.2)3.6 (3.5‐3.7)<.01
Adjusteda3.1 (2.9‐3.2)3.6 (3.5‐3.7)<.01
InterpretationUnadjusted2.6 (2.3‐2.9)2.8 (2.6‐2.9).44
Adjusteda2.6 (2.3‐2.9)2.8 (2.6‐3.0).23
JudgmentUnadjusted2.9 (2.7‐3.1)3.4 (3.3‐3.5)<.01
Adjusteda2.9 (2.7‐3.1)3.4 (3.3‐3.5)<.01

aAdjusting for country, level of training, and the number of WhatsApp messages

Comparing the Performance Between Participants Based on Their Training Grade

A total of 213 participants provided their training grade in pre-SIMBA and post-SIMBA surveys, of which we classified 28.2% (60/213) as junior grade, 55.9% (119/213) as middle grade, and 16% (34/213) as senior grade. Junior-grade participants scored significantly higher in history taking (junior vs middle vs senior: 4.2 [4.0‐4.5] vs 3.7 [3.5‐3.9] vs 3.7 [3.4‐4.0]; P<.01). Otherwise, there was no significant difference across the training grade for the rest of the domains (physical examination: 4.0 [3.7‐4.2] vs 3.7 [3.5‐3.9] vs 3.5 [3.1‐3.8]; P=.07; investigation: 3.5 [3.3‐3.7] vs 3.4 [3.3‐3.6] vs 3.6 [3.3‐3.8]; P=.53; clinical interpretation: 2.6 [2.3‐2.9] vs 2.5 [2.3‐2.8] vs 3.1 [2.7‐3.6]; P=.08; clinical judgment: 3.4 [3.1‐3.6] vs 3.3 [3.1‐3.4] vs 3.4 [3.1‐3.7]; P=.63; and management: 2.8 [2.5‐3.0] vs 2.7 [2.6‐2.9] vs 2.8 [2.5‐3.1]; P=.94) (Table 3 and Multimedia Appendix 2).

Table 3. Mean and 95% CIs of various performance measures using the global rating score by level of training for those participating in SIMBA (Simulation via Instant Messaging for Bedside Application).
Performance measureUnadjusted and adjustedJunior gradeMiddle gradeSenior gradeP valuea
History takingUnadjusted4.2 (4.0‐4.5)3.7 (3.6‐3.9)3.6 (3.3‐3.9)<.01
Adjusteda4.2 (4.0‐4.5)3.7 (3.5‐3.9)3.7 (3.4‐4.0)<.01
Physical examinationUnadjusted3.9 (3.7‐4.2)3.7 (3.5‐3.9)3.5 (3.1‐3.8).09
Adjusteda4.0 (3.7‐4.2)3.7 (3.5‐3.9)3.5 (3.1‐3.8).07
InvestigationUnadjusted3.5 (3.3‐3.7)3.5 (3.3‐3.6)3.5 (3.2‐3.8).97
Adjusteda3.5 (3.3‐3.7)3.4 (3.3‐3.6)3.6 (3.3‐3.8).53
InterpretationUnadjusted2.6 (2.3‐3.0)2.5 (2.3‐2.8)3.1 (2.7‐3.5).06
Adjusteda2.6 (2.3‐2.9)2.5 (2.3‐2.8)3.1 (2.7‐3.6).08
JudgmentUnadjusted3.3 (3.1‐3.6)3.3 (3.2‐3.5)3.4 (3.1‐3.7).96
Adjusteda3.4 (3.1‐3.6)3.3 (3.1‐3.4)3.4 (3.1‐3.7).63
ManagementUnadjusted2.7 (2.5‐3.0)2.8 (2.6‐3.0)2.7 (2.4‐3.0).87
Adjusteda2.8 (2.5‐3.0)2.7 (2.6‐2.9)2.8 (2.5‐3.1).94

aAdjusting for sex, country, and the number of WhatsApp messages.

Comparing the Performance Between Participants Based on Their Gender

A total of 199 participants provided their gender identity in pre-SIMBA and post-SIMBA surveys (male: 83/199, 41.7%; female: 116/199, 58.3%). There were no statistically significant differences in GRS scores between male and female participants (male vs female; history taking: 3.7 [3.5‐3.9] vs 3.8 [3.6‐3.9]; P=.53; physical examination: 3.7 [3.5‐3.9] vs 3.6 [3.4‐3.7]; P=.30; investigation: 3.4 [3.3‐3.6] vs 3.4 [3.3‐3.6]; P=.93; clinical interpretation: 2.9 [2.7‐3.2] vs 2.7 [2.5‐2.9]; P=.12; clinical judgment: 3.3 [3.1‐3.5] vs 3.3 [3.1‐3.4]; P=.75; and management: 2.8 [2.6‐3.0] vs 2.7 [2.5‐2.8]; P=.42) (Table 4).

Table 4. Mean and 95% CIs of various performance measures using the global rating score by number of WhatsApp messages for those participating in SIMBA (Simulation via Instant Messaging for Bedside Application).
Performance measureUnadjusted and adjustedMenWomenP valuea
History takingUnadjusted3.7 (3.4‐3.9)3.8 (3.6‐4.0).52
Adjusteda3.7 (3.5‐3.9)3.8 (3.6‐3.9).53
Physical examinationUnadjusted3.7 (3.5‐3.9)3.6 (3.4‐3.8).35
Adjusteda3.7 (3.5‐3.9)3.6 (3.4‐3.7).30
InvestigationUnadjusted3.4 (3.3‐3.6)3.4 (3.3‐3.6).89
Adjusteda3.4 (3.3‐3.6)3.4 (3.3‐3.6).93
InterpretationUnadjusted3.0 (2.7‐3.2)2.7 (2.4‐2.9).09
Adjusteda2.9 (2.7‐3.2)2.7 (2.5‐2.9).12
JudgmentUnadjusted3.3 (3.1‐3.5)3.3 (3.1‐3.4).79
Adjusteda3.3 (3.1‐3.5)3.3 (3.1‐3.4).75
ManagementUnadjusted2.8 (2.5‐3.0)2.7 (2.5‐2.8).51
Adjusteda2.8 (2.6‐3.0)2.7 (2.5‐2.8).42

aAdjusting for country, level of training, and the number of WhatsApp messages.


Principal Findings

Overall, participants scored higher in history taking and physical examination skills, while their scores were lower in interpretation and management skills. This information is valuable for fine-tuning future simulation programs to focus on improving interpretation and management abilities.

GRS is a validated tool for assessing clinical competencies across 6 critical domains: eliciting history, physical examination, investigations and diagnostic tests, interpretation of findings, clinical judgment, and management and follow-up plans. Although the GRS was not used to evaluate participants’ abilities before the curriculum, it served as an objective measure during the simulation to assess performance and track growth. Combined with presimulation and postsimulation surveys, it allowed us to analyze changes in confidence levels and skill application.

The Pendleton model was adopted to provide structured feedback for the GRS output. This approach starts with positive reinforcement by highlighting what participants did well and then constructive guidance on areas for improvement. This approach ensures supportive feedback and facilitates skill enhancement.

Demographic data were collected during the presimulation survey to contextualize participants’ backgrounds and identify trends in performance or confidence changes across demographic groups. While these characteristics are static, integrating them into a pre-post framework enabled us to correlate them with dynamic variables such as confidence and skill development.

The significant differences in clinical performance scores between LMICs and HICs highlight the need for better medical education resources to bridge existing gaps in health care across the globe. A recent analysis by the General Medical Council has identified a substantial attainment gap between international medical graduates from LMIC and candidates from the United Kingdom working in the National Health Service [27]. Subsequent recommendations advise the need for tailored approaches to tackle inequality and ensure consistent quality of medical education to trainees. Free and accessible programs, such as SIMBA, can facilitate this.

The observed differences in clinical performance between HCPs from HICs and LMICs can be attributed to various factors. HCPs from HICs often benefit from better infrastructure, advanced technology, and comprehensive training programs. In contrast, LMIC professionals face limited access to resources, inadequate training opportunities, and contextual factors that may impact their clinical performance. The comparative analysis highlights the need for context-specific training and tailored approaches to health care education in LMICs. Of note, a commonly identified issue with technology-enhanced simulation is the cost of simulators and limited understanding of the use and limitations of artificial intelligence or embedded computer algorithms [28,29]. SIMBA bypasses these hurdles as it has minimal costs, and the sessions can be conducted virtually.

Whilst all training grades performed similarly across investigation, clinical judgment, and management domains, significant differences were noted in history taking and physical examination scores. The decrease in these scores following career progression suggests the need for simulation-based training for senior trainees to maintain the high quality of history taking and physical examination skills. These insights hold implications for refining medical education approaches and optimizing patient care delivery. In an era of global health care parity, comprehending training grade–related performance disparities is instrumental in fostering contextually apt and universally equitable medical training standards. A study validating the use of smartphone accelerometers in neurosurgical simulations found higher acceleration scores among junior doctors than among senior registrars and consultants, suggesting improved technical performance in junior HCPs [30]. However, our findings also contrast with findings from a study of the American Board of Internal Medicine patient satisfaction questionnaire, which found a systematic improvement in the quality of consultations in directly observed clinical skills in senior attending physicians compared with junior residents [31]. The decline in history and examination skills may stem from increased workload, encouraging reliance on diagnostic investigations over thorough assessments. Deskilling—reduced use of certain skills in practice—also contributes to a decline in history and examination skills. The potential reason for contrast with findings from various national and international educational boards likely arises from differences in settings; formal assessments occur in controlled environments, encouraging conscious performance. In contrast, SIMBA sessions reflect real-life scenarios without direct observation, emphasizing authentic application of skills, which may reveal different trends.

Recommendations by Health Education England highlight the growing role of simulation-based education in core medical training, including procedural and emergency presentation learning sessions [32]. Its widespread use and implementation of a clear framework across all National Health Service trusts may allow physicians to further develop their skills and knowledge in line with up-to-date guidelines throughout their careers.

We did not find a significant difference in GRS between male and female participants. This is consistent with previous findings, which showed no significant gender-related differences in simulation assessment scores among emergency medicine residents assessed against 2 simulation cases related to the emergency medicine curriculum [33]. Similarly, simulation-based learning has been shown to eliminate gender differences in a virtual surgical training course in endoscopic proficiency [34]. However, while gender differences in the simulation were investigated, the translation of simulated skill to physical performance in the endoscopy suite was not investigated. Consequently, further prospective studies are required to assess the impact of gender on clinical performance. Previous studies have further demonstrated the role of simulation-based education in developing interpersonal communication skills and teamwork among medical students and HCPs [35,36]. Therefore, simulation-based learning may provide objective and standardized scenarios to help remove gender-based differences in clinical outcome scores.

While mobile simulation-based learning appears to be widely accessible, reliable web access and cost of access continue to pose a barrier to virtual learning in many LMICs [37]. This is a particular issue in remote areas of the world, which may be subject to unscheduled power cuts [38]. Although communication platforms such as WhatsApp serve as low-cost initiatives for delivering medical education, emphasis must be placed on data security and confidentiality [39]. Consequently, standardized guidelines concerning the use of instant messaging in health care education are required to ensure its sustainability [40].

Limitations

We did not assess the change in perception toward virtual education over time. This may have influenced the findings of experience and difficulties that the participants encountered with web-based education from both LMICs and HICs. Future longitudinal studies must determine whether subsequent clinical decision-making is improved in real-life scenarios. Brain drain, that is, migration of doctors usually from LMICs to HICs for better opportunities and other varied reasons, is another important factor that needs to be considered as this phenomenon may have skewed the results. As SIMBA sessions are currently conducted only in English, it is a potential language barrier. As all the data collected were self-reported rather than objectively verified or tested, it is an inherent limitation of the study.

Regarding the physical examination, participants requested specific types of examinations, such as general physical examination, abdominal examination, or limb examination, and the corresponding findings were provided to them. The virtual clinical physical examination was not performed by the participants themselves, which is a limitation of our study. As we collected minimal identifying data, we do not have further information on the career breaks taken by a doctor for any reason that may have led to inaccurate classification as a junior, middle, or senior-grade doctor. Moreover, given the growing role of virtual learning in medical education, future work should focus on user perception of the feasibility of remote simulation-based training in LMICs.

Conclusions

Future simulation programs should enhance interpretation and management skills by incorporating customized content, feedback mechanisms, and real-world scenarios for HCPs. The significant differences in clinical performance scores between LMICs and HICs highlight the need for better medical education resources to bridge existing gaps in health care globally. The decrease in some clinical competency scores following career progression could be addressed by simulation-based training to maintain the same quality of history taking and physical examination skills. These outcomes, including no gendered differences in simulation-based learning, hold profound implications for tailoring medical education strategies, fostering equitable training, and elevating patient care standards on a global scale. The need for targeted interventions and capacity-building efforts via context-specific training and tailored approaches to health care education is emphasized.

Acknowledgments

The authors thank all the health care professionals who participated in this study. They thank all the moderators, session chairs, and endorsing organizations for their contribution and continued support. During the preparation of this work the authors used ChatGPT to improve language and readability. After using this tool or service, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.

Data Availability

Data will be made available upon request.

Authors' Contributions

KM and HB are the joint first authors, having contributed to all aspects of the study. FC analyzed and interpreted the data analysis. PK conceptualized and supervised the delivery of all study aspects and critically reviewed the manuscript. EW, VS, and DZ contributed substantially to delivering the SIMBA sessions, developing the study and writing the manuscript. SIMBA and CoMICs team contributed toward all stages of this manuscript. All authors contributed substantially to drafting and approving the final draft of the manuscript. The final version has been reviewed and approved by all the authors.

The SIMBA and CoMICs Team members are Carina Pan, Pavithra Sakthivel, Anisah Ali, Isabel Allison, Tamzin Ogiliev, Haaziq Sheikh, Sung Yat Ng, Zahra Olateju, Maiar Elhariry, Eka Melson, Sangamithra Ravi, Abby Radcliffe, Rachel Nirmal, Aditya Swaminathan, Shams Ali Baig, Dwi Delson, Soon Chee Yap, Vardhan Venkatesh, and Fazna Rahman.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Global Rating Scale score of HIC versus LMIC participants (statistically significant results in history taking, investigations, clinical judgment, and management). HIC: high-income country; LMIC: low- and middle-income country.

PNG File, 30 KB

Multimedia Appendix 2

Global Rating Scale score of participants by training grade (statistically significant results in history taking).

PNG File, 37 KB

  1. Morton S, Pencheon D, Squires N. Sustainable development goals (SDGs), and their implementation: a national global framework for health, development and equity needs a systems approach at every level. Br Med Bull. Dec 1, 2017;124(1):81-90. [CrossRef] [Medline]
  2. Shah R, Ahluwalia S. The challenges of understanding differential attainment in postgraduate medical education. Br J Gen Pract. Sep 2019;69(686):426-427. [CrossRef] [Medline]
  3. Kinfu Y, Dal Poz MR, Mercer H, Evans DB. The health worker shortage in Africa: are enough physicians and nurses being trained? Bull World Health Organ. Mar 2009;87(3):225-230. [CrossRef] [Medline]
  4. Papapanou M, Routsi E, Tsamakis K, et al. Medical education challenges and innovations during COVID-19 pandemic. Postgrad Med J. May 2022;98(1159):321-327. [CrossRef] [Medline]
  5. Alsoufi A, Alsuyihili A, Msherghi A, et al. Impact of the COVID-19 pandemic on medical education: medical students’ knowledge, attitudes, and practices regarding electronic learning. PLoS ONE. 2020;15(11):e0242905. [CrossRef] [Medline]
  6. Frehywot S, Vovides Y, Talib Z, et al. E-learning in medical education in resource constrained low- and middle-income countries. Hum Resour Health. Feb 4, 2013;11:4. [CrossRef] [Medline]
  7. Ahmady S, Kallestrup P, Sadoughi MM, et al. Distance learning strategies in medical education during COVID-19: a systematic review. J Educ Health Promot. 2021;10:421. [CrossRef] [Medline]
  8. Stojan J, Haas M, Thammasitboon S, et al. Online learning developments in undergraduate medical education in response to the COVID-19 pandemic: a BEME systematic review: BEME Guide No. 69. Med Teach. Feb 2022;44(2):109-129. [CrossRef] [Medline]
  9. Al-Elq AH. Simulation-based medical teaching and learning. J Family Community Med. Jan 2010;17(1):35-40. [CrossRef] [Medline]
  10. Melson E, Chen W, Zhou D, et al. Adaptation and use of media in an innovative simulation-based clinician training programme. BMJ Simul Technol Enhanc Learn. 2021;7(6):650-652. [CrossRef] [Medline]
  11. Rosen KR. The history of medical simulation. J Crit Care. Jun 2008;23(2):157-166. [CrossRef] [Medline]
  12. Melson E, Davitadze M, Aftab M, et al. Simulation via instant messaging—Birmingham advance (SIMBA) model helped improve clinicians’ confidence to manage cases in diabetes and endocrinology. BMC Med Educ. Aug 18, 2020;20(1):274. [CrossRef] [Medline]
  13. Wallett L, Chen W, Thomas L, et al. Developing a simulation-based learning model for acute medical education during COVID-19 pandemic with Simulation via Instant Messaging - Birmingham Advance (SIMBA). BMJ Open Qual. Apr 2022;11(2):e001565. [CrossRef] [Medline]
  14. Davitadze M, Ooi E, Ng CY, et al. SIMBA: using Kolb’s learning theory in simulation-based learning to improve participants’ confidence. BMC Med Educ. Feb 22, 2022;22(1):116. [CrossRef] [Medline]
  15. Morgan G, Melson E, Davitadze M, et al. Utility of Simulation via Instant Messaging—Birmingham Advance (SIMBA) in medical education during COVID-19 pandemic. J R Coll Physicians Edinb. Jun 2021;51(2):168-172. [CrossRef] [Medline]
  16. Evans N, Davitadze M, Narendran A, et al. SIMBA as an alternative and/or an adjunct to pre-medical work experience during the COVID-19 pandemic. Future Healthc J. Mar 2021;8(1):e142-e145. [CrossRef] [Medline]
  17. Ng CY, Allison I, Ooi E, Davitadze M, Melson E, Kempegowda P. Medical students’ and junior doctors’ leadership and teamwork skills improved after involvement with Simulation via Instant Messaging—Birmingham Advance (SIMBA). BMJ Lead. Sep 2022;6(3):233-236. [CrossRef] [Medline]
  18. Malhotra K, Ali A, Soran V, et al. Levelling the learning ground for healthcare professionals across the world through SIMBA: a mixed-methods study. BMJ Open. Jul 10, 2023;13(7):e069109. [CrossRef] [Medline]
  19. Adams LV, Wagner CM, Nutt CT, Binagwaho A. The future of global health education: training for equity in global health. BMC Med Educ. Nov 21, 2016;16(1):296. [CrossRef] [Medline]
  20. Williams JS, Walker RJ, Egede LE. Achieving equity in an evolving healthcare system: opportunities and challenges. Am J Med Sci. Jan 2016;351(1):33-43. [CrossRef] [Medline]
  21. Kruk ME, Gage AD, Arsenault C, et al. High-quality health systems in the Sustainable Development Goals era: time for a revolution. Lancet Glob Health. Nov 2018;6(11):e1196-e1252. [CrossRef] [Medline]
  22. Gerard JM, Kessler DO, Braun C, Mehta R, Scalzo AJ, Auerbach M. Validation of global rating scale and checklist instruments for the infant lumbar puncture procedure. Simul Healthc. Jun 2013;8(3):148-154. [CrossRef] [Medline]
  23. Joshi A, Kale S, Chandel S, Pal DK. Likert scale: explored and explained. BJAST. 2015;7(4):396-403. [CrossRef]
  24. Burgess A, van Diggele C, Roberts C, Mellis C. Feedback in the clinical setting. BMC Med Educ. Dec 3, 2020;20(Suppl 2):460. [CrossRef] [Medline]
  25. Pendleton D. The Consultation: An Approach to Learning and Teaching: No6. Oxford University Press; 1984.
  26. World bank country and lending groups. World Bank Data Help Desk. Mar 2022. URL: https://datahelpdesk.worldbank.org/knowledgebase/articles/906519 [Accessed 2022-09-04]
  27. Tackling disadvantage in medical education. Health Education England. Mar 2023. URL: https:/​/www.​gmc-uk.org/​-/​media/​documents/​96887270_tackling-disadvantage-in-medical-education-020323.​pdf [Accessed 2023-04-24]
  28. Kansal R, Bawa A, Bansal A, et al. Differences in knowledge and perspectives on the usage of artificial intelligence among doctors and medical students of a developing country: a cross-sectional study. Cureus. Jan 2022;14(1):e21434. [CrossRef] [Medline]
  29. Barteit S, Guzek D, Jahn A, Bärnighausen T, Jorge MM, Neuhann F. Evaluation of e-learning for medical education in low- and middle-income countries: a systematic review. Comput Educ. Feb 2020;145(103726):103726. [CrossRef] [Medline]
  30. Jensen Ang WJ, Hopkins ME, Partridge R, et al. Validating the use of smartphone-based accelerometers for performance assessment in a simulated neurosurgical task. Neurosurgery. Mar 2014;10 Suppl 1(57–64):57-64. [CrossRef] [Medline]
  31. Boudreau D, Tamblyn R, Dufresne L. Evaluation of consultative skills in respiratory medicine using a structured medical consultation. Am J Respir Crit Care Med. Nov 1994;150(5 Pt 1):1298-1304. [CrossRef] [Medline]
  32. Enhancing UK Core Medical Training through simulation-based education: an evidence-based approach. Joint JRCPTB/HEE Expert Group. Oct 4, 2018. URL: https://www.jrcptb.org.uk/sites/default/files/HEE_Report_FINAL.pdf [Accessed 2023-08-23]
  33. Siegelman JN, Lall M, Lee L, Moran TP, Wallenstein J, Shah B. Gender bias in simulation-based assessments of emergency medicine residents. J Grad Med Educ. Aug 2018;10(4):411-415. [CrossRef] [Medline]
  34. Ritter EM, Lineberry M, Hashimoto DA, et al. Simulation-based mastery learning significantly reduces gender differences on the fundamentals of endoscopic surgery performance exam. Surg Endosc. Dec 2018;32(12):5006-5011. [CrossRef] [Medline]
  35. Blackmore A, Kasfiki EV, Purva M. Simulation-based education to improve communication skills: a systematic review and identification of current best practice. BMJ Simul Technol Enhanc Learn. 2018;4(4):159-164. [CrossRef] [Medline]
  36. Amsalem D, Gothelf D, Soul O, Dorman A, Ziv A, Gross R. Single-day simulation-based training improves communication and psychiatric skills of medical students. Front Psychiatry. 2020;11:221. [CrossRef] [Medline]
  37. Nyemike Simeon A, Abdulmujeeb Babatunde A, Lukman Abiodun N, Omogbadegun Olu R, Ido Emem A. Uptake, barriers, and determinants of e-learning among university students in selected low income countries in Sub-Saharan Africa amidst the COVID-19 disruption: an online survey. Adv Med Educ Pract. 2022;13(609–17):609-617. [CrossRef] [Medline]
  38. Ukrani RD, Shaikh AN, Martins RS, Fatima SS, Naseem HA, Baig MA. Low-cost peer-taught virtual research workshops for medical students in Pakistan: a creative, scalable, and sustainable solution for student research. BMC Med Educ. Nov 1, 2021;21(1):557. [CrossRef] [Medline]
  39. Mars M, Morris C, Scott RE. WhatsApp guidelines—what guidelines? A literature review. J Telemed Telecare. Oct 2019;25(9):524-529. [CrossRef] [Medline]
  40. Luna D, Almerares A, Mayan JC 3rd, González Bernaldo de Quirós F, Otero C. Health informatics in developing countries: going beyond pilot practices to sustainable implementations: a review of the current challenges. Healthc Inform Res. Jan 2014;20(1):3-10. [CrossRef] [Medline]


HCPs: health care professionals
HICs: high-income countries
LMICs: low- and middle-income countries
SIMBA: Simulation via Instant Messaging for Bedside Application


Edited by Blake Lesselroth; submitted 31.08.23; peer-reviewed by Cedric Manlhiot, Eren Ogut, Faisal Ismail, Meri Davitadze; final revised version received 09.12.24; accepted 25.02.25; published 11.08.25.

Copyright

© Kashish Malhotra, Harshin Balakrishnan, Emily Warmington, Vina Soran, Francesca Crowe, Dengyi Zhou, SIMBA AND CoMICs Team, Punith Kempegowda. Originally published in JMIR Medical Education (https://mededu.jmir.org), 11.8.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Education, is properly cited. The complete bibliographic information, a link to the original publication on https://mededu.jmir.org/, as well as this copyright and license information must be included.