Introduction

This blog is about medical education in the US and around the world. My interest is in education research and the process of medical education.



The lawyers have asked that I add a disclaimer that makes it clear that these are my personal opinions and do not represent any position of any University that I am affiliated with including the American University of the Caribbean, the University of Kansas, the KU School of Medicine, Florida International University, or the FIU School of Medicine. Nor does any of this represent any position of the Northeast Georgia Medical Center or Northeast Georgia Health System.



Wednesday, September 3, 2025

Residency Selection in the Age of Artificial Intelligence: Promise and Peril

 Residency Selection in the Age of Artificial Intelligence: Promise and Peril

Residency selection has always been a high-stakes, high-stress process. For applicants, it can feel like condensing years of study and service into a few fleeting data points. For programs, it is like drinking from a firehose—thousands of applications to sift through with limited time and resources, and an abiding fear of missing the “right fit.” In recent years, the pressures have only grown: more applications, Step 1 shifting to pass/fail, and increased calls for holistic review in the name of equity and mission alignment. 

Into this crucible comes artificial intelligence (AI). Advocates promise that AI can tame the flood of applications, find overlooked gems, and help restore a measure of balance to an overloaded system. Critics worry that it will encode and amplify existing biases, creating new blind spots behind the sheen of algorithmic authority. A set of recent papers provide a window into this crossroads, with one central question: will AI be a tool for fairness, or just a faster way of making the same mistakes?

What We Know About Interviews   

Before even considering AI, it helps to step back and look at the traditional residency interview process. Lin and colleagues (1) recently published a systematic review of evidence-based practices for interviewing residency applicants in Journal of Graduate Medical Education. Their review of nearly four decades of research is sobering: most studies are low to moderate quality, and many of our cherished traditions—long unstructured interviews, interviewer “gut feelings”—have little evidence behind them. What does work? Structure helps. The multiple mini interview (MMI) shows validity and reliability. Interviewer training improves consistency. Applicants prefer shorter, one-on-one conversations, and they value time with current residents. Even virtual interviews, despite mixed reviews, save money and broaden access. 

In other words, structure beats vibe. If interviews are going to continue as a central part of residency selection, they need to be thoughtfully designed and consistently delivered.

The Scoping Review: AI Arrives 

The most important new contribution to this debate is Sumner and colleagues’ scoping review in JGME (2). They examined the small but growing literature on AI in residency application review. Of the twelve studies they found, three-quarters focused on predicting interview offers or rank list positions using machine learning.  Three articles used natural language processing (NLP) to review and analyze letters of recommendation. 

The results are promising but fragmented. Some models could replicate or even predict program decisions with decent accuracy. Others showed how NLP might highlight subtle patterns in narrative data, such as differences in the language of recommendation letters. But strikingly, only a quarter of the studies explicitly modeled bias. Most acknowledged it as a limitation but stopped short of systematically addressing it. The authors conclude that AI in residency recruitment is here, but it is underdeveloped, under-regulated, and under-evaluated. Without common standards for reporting accuracy, fairness, and transparency, we risk building shiny black boxes that give an illusion of precision while quietly perpetuating inequity.

Early Prototypes in Action

Several studies give us a glimpse of what AI might look like in practice. Burk-Rafel and colleagues at NYU (3) developed a machine learning–based decision support tool, trained on over 8,000 applications across three years of internal medicine interview cycles. The training data included 61 features, including demographics, time since graduation, medical school location, USMLE scores or status, awards (AOA), and publications among many others. Their model achieved an area under the ROC of 0.95 and even performed well (0.94) without USMLE scores. Interestingly, when deployed prospectively, it identified twenty applicants for interview who had been overlooked by human reviewers, many of whom later proved strong candidates. Here, AI wasn’t replacing judgment but augmenting it, catching “diamonds in the rough” that busy faculty reviewers had missed.

Rees and Ryder’s work (4) published in Teaching and Learning in Medicine took a different angle, building machine learning algorithm Random Forest models to predict ranked applicants and matriculants in internal medicine. Their models could predict with high accuracy (area under ROC 0.925) who would be ranked, but struggled to predict who would ultimately matriculate (area under ROC 0.597). The lesson: AI may be able to mimic program decisions, but it is far less certain whether those decisions correlate with outcomes that matter—like performance, retention, or alignment with mission.

Finally, Hassan and colleagues in the Journal of Surgical Education (5) directly compared AI with manual selection of surgical residency applicants. Their findings were provocative: the two applicant lists (AI selected vs PD selected) only had an overlap of 7.4%. AI was able to identify high-performing applicants with efficiency comparable to traditional manual selection, but there were significant differences. The AI selected applicants who were more frequently white/Hispanic (p<0.001), more US medical graduates (p=0.027), younger (p=0.024), and had more publications (p<0.001). This raises questions about both list generation processes. There are questions transparency and acceptance by faculty. Programs faculty trust their own collective wisdom, but will they trust an machine learning process that highlights candidates they initially passed over?

Where AI Could Help

Taken together, these studies suggest that AI could help in several ways:

- Managing volume: AI tools can quickly sort thousands of applications, highlighting candidates who meet baseline thresholds or who might otherwise be filtered out by crude metrics.
- Surfacing hidden talent: By integrating many data points, AI may identify applicants overlooked because of a single weak metric, such as a lower Step score or an atypical background.
- Standardizing review: Algorithms can enforce consistency, reducing the idiosyncrasies of individual reviewers.
- Exposing bias: When designed well, AI can make explicit the patterns of selection, shining light on where programs may unintentionally disadvantage certain groups.

Where AI Could Harm

But the risks are equally real:

- Amplifying bias: Models trained on past decisions will replicate the biases of those decisions. If a program historically favored certain schools or demographics, the algorithm will “learn” to do the same.
- False precision: High AUROC scores may mask the reality that models are only as good as their training data. Predicting interviews is not the same as predicting good residents.
- Transparency and trust: Faculty may resist adopting tools they don’t understand, and applicants may lose faith in a process that feels automated and impersonal.
- Gaming the system: When applicants learn which features are weighted, they may tailor applications to exploit those cues—turning AI from a tool for fairness into just another hoop to jump through.

Broad Reflections: The Future of Recruitment

What emerges from these studies is less a roadmap and more a set of crossroads. Residency recruitment is under enormous pressure. AI offers tantalizing relief, but also real danger.

For programs, the key is humility and intentionality. AI should never completely replace human judgment, but it can augment it. Program directors can use AI to help manage scale, to catch outliers, and to audit their own biases. But the human values—commitment to service, value in diversity, and the mission of training compassionate physicians—cannot be delegated to an algorithm.

For applicants, transparency matters most. A process already viewed as opaque will only grow more fraught if decisions are seen as coming from a black box. Clear communication about how AI is being used, and ongoing study of its impact on residency selection is essential. 

For the medical education community, the moment calls for leadership. We need reporting standards for AI models, fairness audits, and shared best practices. Otherwise, each program will reinvent the wheel—and the mistakes.

Residency recruitment has always been an imperfect science, equal parts art and data. AI does not change that. What it does offer is a new lens—a powerful, potentially distorting one. Our task is not to embrace it blindly nor to reject it out of fear, but to use it wisely, always remembering that behind every application is a human being hoping for a chance to serve.

References

(1) Lin JC, Hu DJ, Scott IU, Greenberg PB. Evidence-based practices for interviewing graduate medical education applicants: A systematic review. J Grad Med Educ. 2024; 16 (2): 151-165.

(2) Sumner MD, Howell TC, Soto AL, et al. The use of artificial intelligence in residency application evaluation: A scoping review. J Grad Med Educ. 2025; 17 (3): 308-319.

(3) Burk-Rafel J, Reinstein I, Feng J, et al. Development and validation of a machine learning–based decision support tool for residency applicant screening and review. Acad Med. 2021; 96 (11S): S54-S61.

(4) Rees CA, Ryder HF. Machine learning for the prediction of ranked applicants and matriculants to an internal medicine residency program. Teach Learn Med. 2022; 35 (3): 277-286.

(5) Hassan S, et al. Artificial intelligence compared to manual selection of prospective surgical residents. J Surg Educ. 2025; 82 (1): 103308.

Monday, August 25, 2025

The Future of Simulation in Medical Education: From Novelty to Necessity

 The Future of Simulation in Medical Education: from Novelty to Necessity

Medical education has always wrestled with the challenge of teaching complex, high-stakes skills in an environment where mistakes can carry real consequences. Historically, students learned at the bedside, often relying on apprenticeship models where experience came in unpredictable bursts. While this “see one, do one, teach one” tradition had its strengths, it also left gaps. Simulation-based training (SBT) emerged to fill those gaps, and it is no longer a niche tool—it is a core component of medical education. A recent article describes simulation-based research and innovation. The authors suggest that the next decade will transform simulation from a supplemental experience into a foundational pillar of how we prepare physicians.

Why Simulation Matters

Simulation provides a safe space where learners can make mistakes, reflect, and try again—without putting patients at risk. Elendu and colleagues’ 2024 review (1) highlights several key benefits: learners gain clinical competence more quickly, retain knowledge longer, and demonstrate improved patient safety outcomes. Equally important, simulation supports deliberate practice, structured feedback, and team-based scenarios that mirror the realities of modern healthcare. In an era where patient safety is paramount and medical knowledge is expanding faster than ever, the controlled environment of simulation offers a vital buffer between the classroom and the clinic. 

Emerging Technologies Driving Change

The next wave of simulation training will be shaped by technology. In an article posted by Education Management Solutions (2), artificial intelligence (AI) is poised to revolutionize how scenarios are created and adapted. Instead of static, one-size-fits-all cases, AI can generate patient interactions tailored to a learner’s level, performance, and even biases. Imagine a resident who consistently misses subtle diagnostic cues being repeatedly exposed to cases that hone that specific skill. Adaptive learning, powered by AI, promises to accelerate mastery and personalize education in ways we’ve only begun to imagine.

Another major trend is the improvement in simulation technology such as high-fidelity mannequins (Sim Man and Harvey), virtual endoscopy and ultrasound simulators, and surgical simulators. Virtual Reality and Augmented Reality have moved from gaming into the world of education. (3) VR headsets are smaller, more affordable, and more accessible. For medical schools committed to widening access to education and reducing disparities, portability is a game-changer.
These tools allow learners to step into highly realistic, immersive scenarios. VR can recreate the chaos of a mass casualty event or the precision of an operating room, while AR overlays digital information onto the real world—imagine seeing a patient with anatomy labeled in real time. The potential for engagement and realism is enormous. Still, VR/AR must avoid becoming flashy gimmicks. Their power lies in creating experiences that are both immersive and educationally sound, rooted in clear learning objectives.

Feeling is Believing: the Role of Haptics 

Simulation has long been strong in visual and auditory fidelity, but haptics—the sense of touch—has lagged behind. That is changing. New advances in haptic feedback allow learners to “feel” the resistance of tissue during a procedure, the snap of a joint during reduction, or the subtle give of a vessel wall during cannulation. For skill-based specialties like surgery, obstetrics, and emergency medicine, this tactile realism can shorten the learning curve and increase confidence before performing procedures on patients. A recent systematic review in the Journal of Surgical Education (4) identified the challenge with surgical simulation. Feedback from the surgical instrument which is typical for minimally invasive techniques such as laparoscopy is easier to simulate than the feel of soft tissues in the body. The review identified nine studies of haptics but there is much inconsistency in the evidence.

Competency Tracking

Perhaps one of the most exciting—and potentially controversial—advances is the integration of data analytics into simulation. Systems are emerging that can measure everything from the angle of a needle insertion to the response time in a code scenario. These metrics can provide real-time feedback and generate longitudinal reports of a learner’s progress. For competency-based medical education (CBME), which emphasizes outcomes over time served, such analytics could provide the objective measures we have long struggled to capture. Of course, this raises important questions about how such data are used in assessment, promotion, and even remediation. Transparency and fairness will be critical if analytics are to fulfill their promise without creating new inequities.

Challenges Ahead  

Despite its promise, simulation faces hurdles. Costs are significant—high-fidelity mannequins, VR systems, and haptic devices are expensive, and simulation centers require space, staff, and upkeep. Faculty development is another challenge: effective simulation requires skilled facilitators who can guide debriefings, not just operate the technology. Finally, while simulation improves competence, translating those skills into clinical performance is not automatic. More research, like that synthesized by Elendu et al., is needed to understand how best to integrate simulation into curricula to maximize transfer to patient care. 

Implications for Medical Education

For medical schools (and residency training programs), the message is clear: simulation is not optional. Schools that fail to invest in simulation risk graduating physicians less prepared for the realities of modern healthcare. The most forward-thinking institutions will not only build simulation centers but also embed simulation across the curriculum—from preclinical years through residency. This requires leadership willing to make strategic investments and faculty committed to weaving simulation into teaching, assessment, and remediation. It also requires attention to equity, ensuring that students across campuses and resource levels have access to the same opportunities.

Looking Forward

As simulation matures, its role will expand beyond technical training. It will increasingly serve as a platform for teaching professionalism, interprofessional teamwork, cultural humility, and even resilience. The “hidden curriculum” of medicine—the values, habits, and attitudes we pass on—can be intentionally addressed in simulated spaces. AI-driven avatars may even help address bias, exposing learners to diverse patient populations in ways that are not possible in traditional settings.

In short, the future of simulation is bright. What began as a supplemental tool is becoming the backbone of modern medical education. The convergence education and technology is creating a learning ecosystem that is safer, smarter, and more responsive to individual learners. The challenge for medical educators is not whether to adopt simulation, but how to do so thoughtfully, equitably, and in ways that truly enhance patient care.

 

References

(1)   Elendu C, Amaechi DC, Okatta AU, et al. The impact of simulation-based training in medical education: A review. Medicine  2024; 103 (27): e38813. doi: 10.1097/MD.0000000000038813. PMID: 38968472; PMCID: PMC11224887.

(2)   https://ems-works.com/blog/content/7-future-trends-in-healthcare-simulation-training/

(3)   Dhar E, Upadhyay U, Huang Y, Uddin M, Manias G, Kyriazis D, Wajid U, AlShawaf H, Syed Abdul S. A scoping review to assess the effects of virtual reality in medical education and clinical care. Digit Health. 2023; 9: 20552076231158022. doi: 10.1177/20552076231158022. PMID: 36865772; PMCID: PMC9972057.

(4)   Rangarajan K, Davis H, Pucher PH.  Systematic Review of Virtual Haptics in Surgical Simulation: A Valid Educational Tool? J of Surgical Education 2020; 77 (2); 337-347.  https://doi.org/10.1016/j.jsurg.2019.09.006