Generative AI Moves Into Clinical Teaching Workflows, Raising Immediate Privacy and Trust Questions
Earlier coverage of ai oversight and its implications for CME providers.
Clinician discussion highlighted two credibility tests: AI tools need validation checkpoints, and patient contributors need clearer funding and role rules.
AI chatbots outperformed physicians on empathy metrics in patient messaging this week, yet produced hallucinated outputs in 12% of cases and non-concordant recommendations one-third of the time. The examples were oncology-led in places, but the provider issue is broader: CME teams are being asked to add new voices and tools without weakening trust.
Clinician conversation split the AI story in two. In a lung cancer conference thread, clinicians discussed chatbots producing more empathetic patient-message responses and possible uses in screening and trial matching, while one reply described using GPT-4 to make patient messages warmer after a quick accuracy review (source). A separate physician thread framed AI adoption as an advantage for early users, tied to model performance in medical training contexts (source).
The caution came from the same week’s oncology education discussion: a chatbot comparison against cancer treatment guidelines was described as producing hallucinated outputs in 12% of cases and non-concordant recommendations in about one-third of recommendations (source). That combination matters. Empathy performance may make AI feel safer than it is, because the output can sound polished even when the recommendation needs verification.
This advances an earlier brief on generative AI entering clinical teaching workflows: the conversation has moved from general privacy and trust concerns to concrete control points. For CME teams, the question is no longer whether AI belongs in education. It is which tasks can be AI-assisted, which outputs require guideline validation, and how clearly learners are told when AI shaped the activity.
The second trust issue was patient participation. A provider-owned CME podcast described patient contributors as authors, discussants, and program participants who bring questions clinicians may not surface on their own (source). That is a legitimate learning-design point: patient experience can change what clinicians notice, especially around treatment burden, communication, and what matters at the point of care.
But the evidence base here needs a caveat. One major signal came from a CME-provider podcast, so it reflects a supplier-side view of education design. The counter-signal came from a commentator critique of pharma-funded patient advocacy, not broad independent clinician consensus. Still, the policy implication is concrete. The commentator opened with the premise, “The first thing we have to acknowledge is patients are the most important voice at the table of all of biomedicine,” then argued that advocacy organizations can represent a selective and funded slice of patient experience rather than the full range of patients (source).
A separate COI discussion around patents, expertise, and disclosure shows why this is not solved by listing conflicts alone (source). CME teams need to distinguish roles: lived-experience speaker, patient reviewer, co-author, planning committee member, moderator, or guideline-adjacent leader. Each role carries a different independence threshold. The concrete question is whether your patient-contributor policy checks funding sources, representation breadth, and role limits before the program is built—not after the disclosure slide is finalized.
Two systems many CME teams already have—AI experimentation and patient engagement—are becoming trust-governance issues. The week’s change is not that AI is suddenly ready, or that patient voice is suddenly risky. It is that both now need explicit rules before they are scaled. For AI, a fluent answer should trigger verification, not relief. For patient contributors, authenticity should include funding transparency, sampling breadth, and clear limits on decision-making roles. The providers who move fastest here will not be the ones with the most tools or the most voices; they will be the ones whose review process can explain why those tools and voices can be trusted.
Independent clinicians highlight AI chatbots outperforming physicians on empathy scores in patient message responses while noting hallucination risks.
"Coming off the heels of an incredible #SAGES2024 is a brand new @NEJM #AI article comparing GPT-3.5 & GPT-4 performance against resident performance on in-service training exams. Note the difference between the two models… Early adopters of AI tools will have a major advantage!"
Show captured excerptCollapse excerptDiscussion of over-reliance dangers and need for guardrails when AI assists with faculty selection or trial matching.
Earlier coverage of ai oversight and its implications for CME providers.
Earlier coverage of ai oversight and its implications for CME providers.
Earlier coverage of commercial disclosure trust and its implications for CME providers.
ChatCME surfaces the questions clinicians actually ask — so you can build activities that close real knowledge gaps.
Request a demo"Hilarious that AI chatbot was more empathetic than physicians when responding to messages from patients?! Several possibilities of AI in oncology including lung cancer screening, clinical trial matching, and more. Enlightening talk by Dr. @PatelOncology @TLCconference"
Show captured excerptCollapse excerptOncology-focused analysis of AI for literature review and trial matching alongside accuracy and hallucination concerns.
Open sourceCME-provider discussion of patients contributing directly to programs and the practical questions they raise that experts overlook.
Open sourceIndependent commentator analysis of pharma funding routes to patient advocacy groups and resulting influence risks.
Open sourceClinician thread on need for stricter COI policies when patients or advocates hold guideline or education leadership roles.
"When COIs and patents are the entirety of your post, it’s an entirely reasonable assumption that they are the focus of your concern. Not exactly a stretch. 🤷🏻♂️"
Show captured excerptCollapse excerpt