Faculty AI Fluency Is Falling Behind Trainees, and CME Is Not Ready
Earlier coverage of ai oversight and its implications for CME providers.
Oncologists are selecting guideline-anchored AI tools over general LLMs for accuracy and safety, creating a targeted training gap for CME. Faculty development is shifting toward explicit clinician-educator identity as a
Oncologists are already selecting guideline-anchored AI tools over general large language models, citing gains in accuracy, clarity, and safety. The clearest signal this week is oncology-led, with a separate faculty-development thread pointing to the same provider problem: education has to help clinicians operate inside new roles and tools, not merely understand them in the abstract.
The sharpest clinician signal came from an oncology thread comparing guideline-anchored retrieval-augmented generation with broader LLM approaches. The author’s summary was blunt: “NCCN-anchored RAG > baseline GPT-4 > broad literature LLMs for accuracy, clarity, and safety.” The same post framed the issue as source quality, not model novelty: AI decision support is only as good as the evidence it can retrieve and apply (source).
That matters because many CME AI offerings still risk treating “AI literacy” as a single category. This week’s signal argues for a narrower skill set: knowing when a curated, guideline-restricted system is preferable to a general model; how to inspect the evidence boundary; and when to override an output that appears plausible but exceeds the tool’s source base. We saw a related pattern in an earlier brief on barriers to AI adoption in practice, but the task has become more concrete: not just whether clinicians trust AI, but what kind of AI they are being asked to trust.
A society-adjacent oncology discussion made the workflow side explicit, describing a gap between exciting AI abstracts and tools that patient-facing physicians can actually use in everyday practice. The speaker pointed to information overload in cancer care and the value of AI for decision support and guideline-based practice, while still emphasizing the translation gap from technology to clinic (source).
For CME teams, the implication is to move from “what is an LLM?” to “which tool is safe enough for this decision, under which evidence constraints, and how would you know?” A useful activity here is not a demo; it is a case where learners compare outputs from a general model and a guideline-anchored tool, identify the evidence boundary, and decide whether to accept, question, or reject the recommendation.
A separate, more institutional signal came from a faculty-development podcast on educator identity in health professions education. The discussion described clinicians who enter teaching roles with little formal education training, then begin to see clinician and educator identities as more connected after structured HPE participation (source).
This is a single institutional source, so it should not be treated as broad consensus. But it is useful because it names an outcome CME providers often leave implicit. Faculty development commonly measures whether participants learned feedback models, questioning strategies, or learning science concepts. The conversation suggests another measurable endpoint: whether participants become more intentional about acting as educators during routine clinical work.
That distinction matters operationally. If educator identity is only a byproduct, programs may stop at short skills sessions. If it is an intended outcome, design changes: participants need reflection on how they see their role, community with peers making the same transition, and follow-up evidence that teaching behaviors changed in clinical settings.
The concrete question for CME teams is whether faculty-development curricula are built to help clinicians practice being educators, not just collect teaching techniques.
The useful shift is specificity. AI education is moving from general comfort with new tools toward disciplined choices about source curation, guardrails, and clinical use. Faculty development is making a similar move from generic teaching improvement toward the role transition clinicians have to make when they become educators. For CME providers, the shared lesson is simple: the next layer of learning design should train clinicians for the exact decision they must make in practice, whether that decision is trusting a tool or stepping fully into a teaching role.
Clinicians on X explicitly compared RAG-restricted outputs to ChatGPT, citing clarity and safety gains from curated guidelines.
"AI decision support in Oncology is only as good as its evidence source. NCCN-anchored RAG > baseline GPT-4 > broad literature LLMs for accuracy, clarity, and safety. #ASCOGI2026 #AIinOncology Congrats to Connor & our other co-authors, more to follow!! 🚀 Partnerships @EvidenceOpen + @nccn | @ASCO +@googlecloud bring much better results"
Show captured excerptCollapse excerptEarlier coverage of ai oversight and its implications for CME providers.
Earlier coverage of ai oversight and its implications for CME providers.
Earlier coverage of learning design and its implications for CME providers.
ChatCME surfaces the questions clinicians actually ask — so you can build activities that close real knowledge gaps.
Request a demoVideo discussion reinforced the need for expert-curated sources and accompanying AI literacy training.
Faculty participants described increased intentionality in feedback and learner community-building once identities merged.
Open source