Freddie Seba headshot

AI Keynote Speaker & Ethics Speaker & Executive Workshops

Dr. Freddie Seba

Issue #59: Filtering Is Governance Infrastructure Before Drift

Operational Safety in Healthcare AI, Slop Markets, AI Literacy Risk, and the New Trust Surface

AI Ethics & Governance for Leaders, Boards & Trustees

By Dr. Freddie Seba

© 2026 Freddie Seba. All rights reserved.

This Week’s Governance Lesson

AI capability scales fast.

But governance maturity only shows up when institutions build filtering + operational safety infrastructure—before drift forces a reactive response.

The institutional failure mode is now predictable:

If your institution adopts AI without filtering and operational safety, you do not get transformation.

You get variance → noise → drift → trust decay.

So the board question becomes:

Do we have filtering infrastructure and an operational safety case… or are we hoping smart tools will self-govern at scale?

Podcast Update

New episode drops midweek (Episode 8): Enabling Responsible AI in Higher Education — Making the Right Thing Easy.

Guest: Christopher Brooks, PhD | Professor, Computer Science & Engineering | @University of San Francisco

Why it belongs in this issue: filtering governance fails when responsible behavior is hard. Enablement governance makes the right thing easy—so filters and safety controls actually get used.

Available on @Spotify, @YouTube, @Apple Podcasts, and the @Substack app.

Executive Reflection: The Signal Is Not More AI. It is More Variance.

The signal this week is not that AI is generating more.

It is that AI is generating more variance—more drafts, more summaries, more decisions, more content, more answers, more almost-right.

That creates a governance shift boards can’t ignore:

When output triples, filtering becomes the value layer.

And in high-stakes domains like healthcare and education, filtering becomes a safety layer, not a productivity feature.

One of the cleanest metaphors comes from the book market signal: with LLM diffusion, new book releases tripled, average quality fell, but the upper tier improved—meaning the net impact depends on whether readers can reliably find the good work. (NBER)

That is not just publishing.

That is governance.

Because scale increases variance, and that variance becomes institutional risk unless you build systems that separate signal from noise.

What We are Seeing: Signals

1) Write Access to the World is the governance threshold—not model intelligence.

A second signal is ideological but useful: the sovereign AI / Web 4.0 narrative—AI that reads, writes, and owns without humans. Regardless of the timeline, governance leaders should treat this as a diagnostic.

The moment you grant systems real-world write access—posting, purchasing, messaging, changing records, deploying code—governance becomes operational safety.

AI governance translation: Stop debating capability. Start governing permission surfaces:

  • What can the system read?
  • What can it write?
  • What can it transact?
  • What can it change that becomes hard to undo?

Board move: Classify write access as a high-risk control tier requiring:

  • Explicit decision rights
  • Staged approvals
  • Monitoring
  • Rollback plans
  • Fast pause authority

2) Clinical AI governance is shifting from accuracy to operational safety

A crisp signal from @npjDigitalMedicine: a new framing for defining operational safety in clinical AI systems—not just does it work, but does it remain safe in deployment conditions. (Nature)

AI governance translation:

Boards should stop accepting validated models as a proxy for safety. Require an operational safety case, including:

  • Intended-use boundaries + prohibited uses
  • Known failure modes
  • Workflow integration risks
  • Monitoring + drift detection
  • Human factors (overreliance, alert fatigue, trust erosion)
  • Incident pathways (pause, notify, remediate)

3) Public trust is now evidence-based—and it points to governance, not AI vibes

A major trust signal from @Coalition for Health AI (CHAI): adoption is high, comfort is low, and trust is conditional on accountability. (Chai)

  • 75% report using AI; only 13% feel very comfortable (Chai)
  • 51% say AI makes them trust healthcare less; 12% say it increases trust (Chai)
  • 80%+ say trust would increase with clear accountability measures (Chai)
  • Disclosure matters—but disclosure alone does not reliably build trust (Chai)

AI governance translation:

Trust is not a downstream PR outcome. Trust is a design constraint—and the public is signaling they want enforceable accountability, not marketing.

Board move: Treat trust as an oversight deliverable:

  • Named accountable owners
  • Visible escalation pathways
  • Auditable monitoring
  • Multi-layer governance (internal + independent + regulatory-ready)

4) The build-from-within health system strategy is becoming a governance advantage

A high-signal deployment note from arXiv describes a health system building an internal, model-agnostic LLM capability, integrated into EHR workflows via ChatEHR, with continuous monitoring and explicit value assessment. (arXiv)

Key operational reality: In real-world use, summary generation had measurable error rates (hallucinations and inaccuracies), and the institution needed new monitoring methods beyond benchmark evaluations. (arXiv)

AI governance translation:

This is governance maturity: internal platform, measurement, monitoring, and value accounting.

Board move: Require a “value + risk accounting” framework:

  • Which tasks get automated first
  • What error rates are tolerable (by task)
  • What gets escalated
  • What savings count—and what hidden costs must be captured (burden shift, trust impact, downstream harm)

5) Health misinformation is no longer an edge case—it’s a predictable failure mode

Two complementary safety signals:

  • A large-scale study of LLM responses to patient medical questions found high rates of unsafe answers across models and conditions—showing why helpfulness without safety controls poses a clinical risk. (Nature)
  • This aligns with what operational safety demands: safety must be defined in terms of real prompts, not just ideal tasks. (Nature)

Once outputs are patient-facing, disclosure, monitoring, and escalation are safety controls—not optional communications.

Board move: For patient-facing or patient-influencing systems, require:

  • Misinformation red-teaming
  • Do not answer/redirect protocols
  • Audit trails for transformations
  • Complaint-to-remediation pipelines

6) AI literacy is becoming a governance control, because human behavior is the bottleneck

Here’s the governance trap: the riskiest failure modes are often human—hidden use, overtrust, and miscalibrated reliance.

Regulators are now wrestling with the realities of implementation. In the EU Digital Omnibus on AI draft, stakeholders argue that a one-size-fits-all AI literacy obligation can be ineffective and burdensome, pushing toward guidance/support models—while still recognizing literacy as crucial for informed deployment.

The same draft also reflects implementation capacity constraints, including proposals to postpone certain high-risk obligation timelines to allow standards/guidance to catch up.

AI governance translation:

AI literacy is not a training initiative. It is a control—because the same tool can be safe or unsafe depending on user behavior, context, and incentives.

Board move: Require:

  • Measurement beyond self-report (logs, anonymized telemetry, workflow sampling)
  • Role-based literacy (what clinicians, staff, faculty, and leaders must know)
  • Standards that prohibit magic framing were imposed, documented, and escalated.

7) Benchmarks are accelerating—but governance needs real-world validity filters

Math benchmark escalation is another governance mirror: leaderboard wins come at the expense of deployment constraints, eval leakage risks, or task mismatch.

An @IEEE Spectrum analysis highlights how AI math benchmarks keep evolving to keep pace with rapid model gains—illustrating why governance can’t rely on benchmark proof as a safety proxy. (IEEE Spectrum)

AI governance translation:

Benchmark gains are not operational safety. A benchmark is a claim—filtering is the verification system.

Board move: Require evaluation that matches the deployment reality:

  • In-domain performance
  • Real workflow constraints
  • Real failure costs
  • Incident-ready monitoring

8) Data quality failures are governance failures—especially in healthcare

A technical signal with direct governance implications: sparse EHRs and unknown missingness can distort clinical predictions unless institutions explicitly govern how missingness is handled. A new method proposal frames missingness imputation as a core performance variable. (arXiv)

Separately, a Nature portfolio paper points toward fairness-aware design for clinical language models—another reminder that fairness does not appear automatically; it emerges when institutions engineer and govern for it. (Nature)

AI governance translation:

In healthcare, “data reality” is safety reality. If your data pipeline is brittle, your governance is brittle.

Board move: Demand explicit ownership of:

  • Data provenance and missingness policy
  • Drift monitoring
  • Equity impact measurement
  • Update and rollback authority

9) The Slop Market is the clearest governance mirror: output up, average down, top-tier up.

A working paper from @NBER finds that after LLM diffusion, releases increased sharply, average quality declined, but the top tier improved—and consumer benefits depend on filtering and discovery. (NBER)

AI governance translation:

Your institution is now a marketplace of AI outputs. If you do not build filters—standards, provenance, review pathways, escalation—your AI adoption becomes slop adoption.

Board move: Require every AI deployment to declare its filtering stack:

  • Pre-use constraints (inputs, sources, tool boundaries)
  • In-process checks (retrieval, guardrails, structured outputs)
  • Post-use review (spot checks, audits, incident learning)

The Seba Framework: The 12 Ps of Responsible AI Oversight ©

How I translate signals into board-ready governance.

Purpose — mission alignment vs. cost extraction

Problems — decision-relevant framing (not metric chasing)

Profits — who benefits vs. who bears risk

People — patients/students/workforce; lived impacts

Planet — compute, infrastructure, scale costs

Process — lifecycle monitoring + incident learning

Policy — risk-specific rules (health, youth, education, employment)

Protections — vulnerable populations + escalation paths

Privacy — enforceable limits on data use/training

Provenance — traceability of data, models, vendors, outputs

Preparedness — board competence + governance cadence

Product Ownership — institutions own outcomes once AI acts

Board-Ready Next Step: A One-Page Filtering & Operational Safety Sheet

If you only do one thing this quarter:

Require a Filtering & Operational Safety Sheet for every AI use case. One page. Audit-ready. Updated post-deployment.

It should state:

  1. Output category (what does it produce?)
  • Summaries
  • Recommendations
  • Clinical interpretations
  • Patient-facing explanations
  • Operational decisions
  • External-facing content
  1. Write-access surface (what can it change?)
  • None (read-only)
  • Drafts only
  • Sends with approval
  • Executes actions
  • Modifies records
  • Deploys code/triggers workflows
  • Financial transactions
  1. Filtering stack (3 layers)
  • Pre-output filters: allowed inputs, source rules, retrieval constraints
  • In-output structure: required format, citations, confidence markers, refusal rules
  • Post-output review: sampling, peer review, clinician sign-off, audit logs
  1. Operational safety case
  • Intended use + prohibited use
  • Known failure modes (misinfo, bias, overreliance, drift)
  • Human factors risks
  • Mitigations + accountable owners
  1. Quality + safety metrics (measured, not promised)
  • Task-specific hallucination/inaccuracy estimates
  • Complaint rate
  • Override rate
  • Near-miss incidents
  • Equity signals (who is harmed, who is excluded)
  1. Monitoring + escalation
  • Who can pause
  • Who must be notified
  • What counts as an incident
  • Response timelines
  • Remediation + learning loop
  1. Value assessment
  • Cost savings
  • Time savings
  • Revenue impacts (if any)
  • And the unpriced outcomes: quality, trust, clinician burden, patient harm risk

This sheet turns ‘we’re using AI’ into we are governing AI.

About the Author

Freddie Seba is a researcher and practitioner—and an #AIKeynoteSpeaker and #AIEthicsSpeaker—focused on AI ethics and AI governance for leaders across higher education, healthcare, and financial services.

He holds an MBA (@Yale University), an MA in International Policy Studies (@Stanford University), and an EdD in Organization and Leadership (@University of San Francisco), with a dissertation on AI ethics and governance defended in Fall 2025.

He writes AI Ethics & Governance for Leaders, Boards & Trustees and hosts AI Governance with Dr. Freddie Seba, translating practitioner signals into board-ready oversight: decision rights, risk tiering, vendor accountability, monitoring, and incident preparedness.

Speaking, Briefings, and Board Workshops

Freddie Seba delivers a keynote on AI governance, risk, trust infrastructure, and institutional legitimacy. He also delivers AI executive workshops and board briefings with practical takeaways—accountability, transparency, safety, responsible adoption, escalation design, and governance maturity:

Inventory → Tiering → Controls → Dashboards → Incident Drills

To book: freddieseba.com and @LinkedIn.

Gratitude

Grateful for the communities that keep this work grounded:

@University of San Francisco • @AMIA Informatics • @Stanford HAI • @Coalition for Health AI

Transparency + Disclaimer

Educational content only. This newsletter does not constitute legal, medical, clinical, insurance, financial, or professional advice.

Drafted and refined with AI-assisted tools for synthesis and clarity. Final editorial control and responsibility remain with the author.

© 2026 Freddie Seba. All rights reserved.

#AIGovernance #ResponsibleAI #BoardOversight #HealthcareAI #OperationalSafety #AILeadership #TrustInfrastructure #RiskManagement #AIandTrust #HigherEd

SOURCES

Leave a Reply

Your email address will not be published. Required fields are marked *