Forty-two percent of AI-generated emergency department summaries contained hallucinated information. Forty-seven percent omitted clinically relevant details. These findings are published. Ambient AI documentation tools are entering health systems under the banner of efficiency, reduced burden, and throughput optimization. Both realities now coexist. Large language models generate portions of clinical documentation through probabilistic prediction. Clinicians review, edit, and sign. At first glance, this appears incremental. It is not. Authorship has shifted. Prediction optimizes for statistical likelihood. Clinical documentation is meant to preserve clinical intent. That divergence is structural. A medical record is not administrative residue. It is clinical memory, operational infrastructure, and legal evidence. It governs care continuity, reimbursement integrity, audit defensibility, and malpractice exposure. When narrative construction changes, governance must change with it. WHAT THE EVIDENCE ESTABLISHES A UCSF-led evaluation of large language model generated emergency department summaries found: Only 33 percent were entirely error free. 42 percent contained hallucinated information. 47 percent omitted clinically relevant details. Error concentration was highest in the Plan section, the portion directing follow-up and treatment decisions. Research published in npj Digital Medicine identified hallucinations in 1.47 percent of sentences and omissions in 3.45 percent. Of the hallucinations observed, 44 percent were categorized as major and capable of influencing diagnosis or management. WHY OMISSION IS THE STRUCTURAL RISK Absence carries consequence. Unrecorded symptoms are treated as absent. Undocumented uncertainty reads as confidence. Missing qualifiers alter downstream reasoning. Documentation functions as shared interpretive infrastructure. When contextual detail thins, care pathways adjust accordingly. Nearly half of summaries in the UCSF study omitted clinically relevant information. At scale, modest omission rates compound into systemic narrative drift. Drift rarely declares itself. It accumulates through normalization. As documentation transitions from authored reasoning to AI-assisted review, cognitive posture shifts. Constructing a note requires synthesis. Validating a draft requires plausibility assessment. Those are not equivalent disciplines. Early outputs may appear accurate. Vigilance softens. Familiar phrasing spreads. Narrative variance narrows despite clinical complexity. Erosion proceeds incrementally. ACCOUNTABILITY HAS NOT SHIFTED Signature standards governing medical documentation remain unchanged. A clinician's signature signifies completeness and accuracy. Courts and regulators evaluate the signed record rather than the generative workflow behind it. AI introduces probabilistic drafting upstream, while existing liability standards for signed documentation remain unchanged. Content retained in the record, whether fabricated or incomplete, is attributed to the signer. Legal expectations have not evolved in parallel with generative capability. This is not a frontline issue. Executive leadership defines the safeguards that sit around this shift. EFFICIENCY AND INTEGRITY SIGNAL DIFFERENT THINGS Documentation burden is substantial. Reducing after-hours charting carries legitimate value. Efficiency deserves attention. Speed is easily quantified. Narrative integrity is not. Completion time can improve while nuance erodes. High-performing systems monitor operational efficiency alongside interpretive fidelity. If fidelity remains unmeasured, efficiency becomes the dominant signal. Dominant signals shape behavior. WHEN REVIEW REPLACES AUTHORSHIP Professional engagement evolves when documentation moves from authored narrative to AI-assisted drafting. Synthesis and validation require different forms of attention. Observable indicators of drift include: Repetitive phrasing across encounters. Erosion of contextual qualifiers. Structurally complete yet clinically thinned plans. Reduced narrative variance across diverse cases. Individually, these signals appear minor. Collectively, they reflect attenuation. Organizational culture responds to reinforced priorities. Sustained emphasis on velocity over fidelity recalibrates norms. Consistent reinforcement of interpretive integrity stabilizes them. Individual vigilance is insufficient. System design must reinforce engagement. Leadership communication must elevate fidelity alongside efficiency. Governance structures must institutionalize narrative review. GOVERNANCE MATURITY DETERMINES OUTCOME AI documentation adoption is accelerating. Healthcare systems are navigating rapid technological expansion alongside ongoing regulatory, legal, and public accountability pressures. Documentation AI sits at that intersection. Deployment is not the strategic question. Governance maturity is. Mature oversight frameworks incorporate: Defined variance thresholds for hallucination and omission. Routine transcript-to-note sampling. Section-specific monitoring for high-impact domains such as the Plan. Transparent traceability from transcript to finalized note. Explicit executive ownership of documentation fidelity. LEADERSHIP CHECKPOINT Organizations implementing AI documentation should examine: What omission rate is acceptable? How frequently are transcript-to-note comparisons conducted? Which executive role owns documentation fidelity? Are interpretive drift indicators tracked alongside throughput? What vulnerabilities would surface under external audit? These are governance determinations. Efficiency gains are visible. Fidelity erosion is not. Design influences drift. Monitoring strengthens resilience. Stewardship determines whether innovation reinforces or weakens trust.

Keep reading