back to top
Thursday, 4 September, 2025
HomeArtificial Intelligence (AI)AI threat of 'de-skilling' for medical trainees

AI threat of 'de-skilling' for medical trainees

Medical learners who use powerful large language models (LLMs) as education aids, run the risk of being “de-skilled”, where over-reliance on AI erodes fundamental clinical reasoning skills, “mis-skilled”, where trainees adopt AI-generated errors, or “never-skilled”, the failure to develop essential competencies in the first place.

In a recent review in New England Journal of Medicine, researchers elucidated the challenges of supervising such early-career medical learners, and proposed a structured educational framework called “diagnosis, evidence, feedback, teaching and recommendation for AI engagement (DEFT-AI)” to counter these potential AI demerits by scaffolding critical thinking.

News-Medical.net reports that the review also introduces the “cyborg” and “centaur” models of human-AI collaboration, urging clinicians to adopt an adaptive practice where they learn to engage critically with AI-generated outputs rather than unquestioningly trusting them.

Recognise the ‘leap of faith’

The paper defines AI interactions as moments where clinicians receive outputs they cannot fully retrace, prompting a critical pause to assess trustworthiness before acting.

Recent advancements in AI, particularly in computation and LLMs, are progressing at an astonishing rate. LLMs such as OpenAI’s ChatGPT and Google’s Gemini are increasingly being used in medical learning, raising both opportunities and risks for clinical reasoning.

A growing body of literature suggests that AI tools are fundamentally reshaping medical learning and practice.

However, integrating AI into clinical practice presents unprecedented opportunities and significant risks for medical education. While rapid access to information and the ability to consolidate vast swaths of data into easily accessible summaries may become integral in future medical education and practice, LLMs are known to simulate human-like reasoning to create an “appearance of agency” – an effect where systems simulate reasoning and outputs appear to show agency when none actually exists.

This can be extremely dangerous for inexperienced medical trainees.

Medical educators thus face a novel and urgent challenge – guiding and supervising trainees who might be more proficient at leveraging AI than the educators themselves, creating an “inversion of expertise” where teachers become learners too.

The present review highlights three specific hurdles (“de-skilling”, “never-skilling”, and “mis-skilling”) that must be overcome before AI can cement its role in ensuring a safer and healthier future.

About the review

Model co-management of uncertainty: educators should openly acknowledge their own AI learning curve alongside trainees, turning discomfort into shared inquiry and explicit teaching moments.

This review aims to address the urgent and critical need by conducting a comprehensive examination of the scientific literature that explores the challenges and opportunities presented by AI in medical education.

It collates and synthesises the outcomes of more than 70 prior publications across existing educational theory, cognitive science, and emerging research on human-AI interaction and uses these insights to develop novel conceptual frameworks for the clinical supervision of AI:

Diagnosis, evidence, feedback, teaching (DEFT), and recommendation for AI engagement (DEFT-AI) – an adapted framework for promoting critical thinking during educational conversations around AI use; Cyborg vs centaur models: a new typology to describe two distinct modalities of human-AI collaboration. These models are designed to help teachers and learners adapt their use of AI to the specific clinical task and associated risk.

Review findings

Ask “How do you think the AI got there?”: DEFT-AI conversations should probe the learner’s understanding of the AI tool’s limitations and potential biases, not just its output.

The review identifies and addresses several cognitive traps imposed on medical education by today’s AI age. “Cognitive offloading”, the process of over-relying on AI for complex tasks like clinical reasoning, is highlighted for its link to “automation bias”, subsequent over-reliance on the AI’s output, and a failure to catch its mistakes.

Alarmingly, cognitive offloading and automation bias are not just theoretical concerns. A study found that more than a third of advanced medical students failed to identify erroneous LLM answers to clinical scenarios. Another study reported a significant negative correlation between the frequent use of AI tools and critical thinking abilities, mediated by increased offloading, and this effect was especially pronounced among younger participants.

The review recommends addressing these concerns by developing and adopting the DEFT-AI framework, a structured approach for educators as a response to a trainee’s dependence on AI. It proposes leveraging a critical conversation that moves beyond the AI’s answer to probe the learner’s reasoning.

Key questions include: “What prompts did you use?”, “How did you verify the AI-generated output?”, and “How did the AI’s suggestion influence or change your diagnostic approach?”

Educators are also encouraged to teach evidence-based appraisal of AI outputs using Sackett’s framework (ask, acquire, appraise, apply, assess) and effective prompt engineering techniques, such as chain-of-thought reasoning.

The review further stresses that supervision must distinguish between evaluating the AI tool itself and evaluating its specific output. For example, institutional scorecards or model leader-boards may be used to judge tools, while evidence-based medicine appraisal steps should be applied to each individual output.

Use chain-of-thought prompting: asking the AI to explain its reasoning step-by-step (“think out loud”) improves its accuracy and allows clinicians to better evaluate its logic.

Finally, the review presents the “cyborg” and “centaur” modes of clinician-AI interaction. In centaur mode, tasks are strategically divided so that the clinician delegates low-risk, well-defined tasks (such as summarising data or drafting communications) to the AI, while retaining complete control over high-stakes clinical judgment and decision-making.

This mode is recommended when addressing complex or uncertain cases.

In contrast, the cyborg mode assumes that the clinician and AI co-construct a solution to the task at hand. This mode is efficient for low-risk, routine tasks but carries a higher risk of automation bias if not used with ongoing reflective oversight and justification.

The review also warns that performance heterogeneity and bias in LLMs can exacerbate health inequities. AI systems may underperform for certain populations, and uncritical adoption could widen disparities rather than close them.

Conclusions

The present review concludes that while the integration of AI into medicine and medical education is inevitable (and largely beneficial), its successful and safe adoption is not. It highlights that medical education must proactively address the risks of de-skilling, never-skilling, and mis-skilling by fundamentally changing how clinical reasoning is taught, particularly against the backdrop of AI.

Critical thinking remains foundational for “adaptive practice” – the ability to shift between efficient routines and innovative problem-solving when faced with the unpredictability of AI.

In summary, this review demonstrates that the ultimate goal is not to create doctors who are dependent on AI, but to cultivate clinicians who can skilfully and safely leverage it as a powerful tool to augment, but not replace, their own expertise through a “verify and trust” paradigm.

 

News-Medical.net article – How can medical trainees use AI without losing critical thinking skills? (Open access)

 

See more from MedicalBrief archives:

 

OECD: How artificial intelligence could change the future of health

 

The challenges of rapidly evolving AI in healthcare

 

Growing role for AI in everyday medical interactions

MedicalBrief — our free weekly e-newsletter

We'd appreciate as much information as possible, however only an email address is required.