CHAI 2025: Healthcare AI Stops Being Theoretical
CHAI 2025: Healthcare AI Stops Being Theoretical
A patient receives HIV test results on their phone. The terminology confuses them. â4th generation testâ instead of the familiar finger stick. They message their provider through the portal and wait.
One minute. Two minutes. Frustrated, they click the AI chat option.
Fifteen minutes of detailed conversation follows. The AI explains window periods, test accuracy, reviews testing history. When the human provider finally responds 45 minutes later with âYeah, thatâs our new test. Donât worry about it,â the patient replies: âThanks, I already got my answer from the AI.â
This real-world interaction, shared at the Coalition for Health AIâs 2025 conference at Stanford, reveals where healthcare stands today. Demand overwhelms capacity. Technology capabilities exceed human bandwidth. Patients choose AI responses over delayed human ones.
David Entwistle, Stanford Healthcareâs CEO, framed the moment precisely: âSix years ago, if you asked healthcare CEOs what it would take to achieve 20% productivity improvement, theyâd say impossible. Today with AI, we might actually get there.â
The Procurement Conversation Flipped
Walk the Stanford halls during CHAIâs conference and youâd overhear health system executives comparing deployment strategies. Startup founders shared outcome metrics, not projected ones. Clinicians debated workflow integration details, not whether AI belonged in healthcare at all.
The numbers tell part of the story. Sixty-six percent of physicians now use AI tools. Major health systems report thousands of daily patient interactions through AI agents. Procurement conversations shifted from âWhat is AI?â to âHow quickly can we deploy this?â Michael Blumenthal from Hiro noted this transformation took six years.
Six years from healthcare IT conference curiosity to production deployment discussions. That timeline matters for anyone building a career in clinical informatics or healthcare technology leadership.
NIH Commits to AI Infrastructure
The room fell silent when Dr. Christopher Muller took the stage. Heâs Deputy Principal Director of the NIH, the organization funding 87% of global biomedical research. When they announce strategic direction, the field listens.
The first comprehensive NIH AI strategic plan represents more than another funding announcement. It signals fundamental reimagining of medical discovery processes.
âThe kind of science thatâs going to move the fields forward is no longer one scientist sitting in their lab by themselves doing experiments,â Muller explained. The volume of data generated now exceeds any individualâs or small teamâs ability to make connections.
Executive orders positioning AI as essential for âhuman flourishing, economic competitiveness and national securityâ back the strategic plan. This isnât peripheral investment. Itâs infrastructure building for AI-powered discovery.
What makes NIHâs approach significant: their focus on reliability and reproducibility. âWe have to be able to feed reliable data into them, otherwise what theyâre generating is not going to be reliable,â Muller stressed. Quality over quantity. A mature approach acknowledging both transformative potential and current limitations.
The Trust Revolution
Perhaps the most surprising theme of the conference was the enthusiasm for governanceânot typically a word that generates excitement in Silicon Valley. Yet presenter after presenter emphasized that governance frameworks arenât barriers to innovation; theyâre enablers of it.
Barry Stein from Hartford Healthcare captured this perfectly with his Formula One analogy: âDrivers can only go around the track at 100 miles an hour because theyâve got good brakes.â Hartford has built six core capabilities over a decade to safely accelerate AI adoption, treating each new AI tool like âa new drug or angioplasty balloonâ that requires rigorous evaluation before clinical use.
This governance-first approach is paying dividends. CHAI announced five new Quality Assurance Resource Providers joining Beekeeper AIâAlignment AI, Signal 1, Lens AI, Click, and Ferrumâcreating an ecosystem for independent AI validation. The first 25 solutions to enroll in CHAIâs new model-card registry will receive free independent verification, turning transparency into a market incentive. The message was clear: in healthcare, âmove fast and break thingsâ isnât just inappropriateâitâs potentially lethal.
The startups get it too. In a revealing panel discussion, Sam Varma from Healthvana admitted something you rarely hear in Silicon Valley: âWe are actually really excited about governance.â Why? Because in a market where anyone can âtake ChatGPT, scrub off the letters and write âmy cool health appâ on top,â governance frameworks help serious companies differentiate themselves from what he called âconvincing fake products.â
The Race for Real-World Impact
Stanfordâs Dr. Nigam Shah presented sobering data about the current state of medical AI research. His teamâs analysis found that 95% of healthcare AI papers used no actual electronic health record data. The top âhealthcare taskâ being evaluated? Taking medical licensing exams. âPretty sorry state of affairs,â he concluded.
This disconnect between research and reality motivated Shahâs team to create Med-HELM, a comprehensive evaluation framework for medical language models. With 35 benchmarks across 121 clinical tasks, itâs the most ambitious attempt yet to measure what actually matters in healthcare AI.
The results are illuminating. While AI excels at generating clinical documentation, it struggles with administrative tasks like billing and codingâironically, some of the most pressing pain points for health systems. The framework costs $11,700 to run completely, but Shah is giving it away free through CHAI, hoping to crowdsource the thousands of private datasets needed to truly validate these models.
Agents Take the Stage
If there was a breakout star of the conference, it was agentic AIâautonomous systems that can take actions on behalf of users. The demonstrations were impressive: Hiroâs AI agent handling complex appointment rescheduling while checking multiple systems and applying business logic. BrainHi serving half the population of Puerto Rico with AI receptionists. Healthvana showing those remarkable patient engagement statistics.
But what struck attendees wasnât just the technologyâit was the scale of deployment. These arenât pilots anymore. Intermountain Healthâs Mona Bassett captured the moment perfectly: âWe have too many ideas now.â The challenge has shifted from finding AI use cases to managing the flood of possibilities.
Recognizing this new reality, CHAI announced the formation of an Agentic AI Working Group launching this summer to draft safety guidelines for semi-autonomous care systems. The group will tackle thorny questions about autonomy boundaries, human oversight requirements, and fail-safe mechanisms.
The top five use cases have crystallized around scheduling, prescription refills, general FAQs, call routing, and billing inquiries. Not the sexiest applications, perhaps, but they address real pain points that affect millions of interactions daily. As Emmanuel Okwuendo from BrainHi noted, these systems can help patients find specialty care âin minutes rather than weeks or months.â
The Ambient Revolution Grows Up
The ambient documentation market has matured remarkably. What began as simple transcription services has evolved into sophisticated clinical partners. The panel featuring leaders from Suki, Nuance, Abridge, and Ambience revealed an industry moving beyond competing on accuracy to differentiating on workflow integration and user experience.
âWeâre very early,â cautioned Ed Lee from Nuance. âIf youâre talking about truly ambient technology using our five senses to help clinicians, all weâve done is take speech data and EHR data and put it together into essentially a billing artifact.â The vision extends far beyond current capabilitiesâreal-time decision support, multi-modal inputs, predictive documentation.
Yet even todayâs âbasicâ capabilities are transforming clinical practice. As one panelist noted, these tools donât just save timeââthey make peopleâs work lives more sustainable and let them go home and recharge.â
The Foundation Model Dilemma
How do you choose between OpenAI, Anthropic, Google, and others when new models drop weekly? The foundation model panel tackled this head-on, revealing strategies that would surprise those expecting brand loyalty.
âTest every model empirically for each use case,â advised Zubair from Anthropic, adding that âmodels cannot forecast in advance how they might perform.â The cost variations are staggeringâpanelists reported 10 to 1,000x differences for similar performance depending on the use case.
Stanfordâs approach, shared by Kathleen, was pragmatic: âThe model itself beyond a certain threshold for performance doesnât matter that much. It has to be good enough and easy for us to access.â Theyâve standardized on Azure OpenAI not because itâs always best, but because it enables rapid experimentation.
The consensus? Healthcare organizations will operate in a multi-model world, optimizing different models for different tasks. The key is building the organizational capability to rapidly test and deploy, not picking a winner.
The Global Expansion
CHAIâs announcement of Singapore as its first international chapter signals recognition that healthcare AI canât be solved by any one country. As Brian Anderson, CHAIâs leader, explained, theyâre not building âhigh-level international standardsâ but rather taking technical frameworks and applying them to specific communities with âappropriate adherence to local regulations, priorities, and values.â The Singapore chapter will adapt CHAI frameworks to APEC regulations starting in Q3 2025.
The Q1 2026 global summit co-hosted with Spain and the OECD represents the next phaseâmoving from American-centric standards to truly global frameworks. This matters because healthcareâs challenges are universal, even if the solutions need localization.
Closer to home, CHAI is partnering with the National Association of Community Health Centers to create governance guides specifically tailored to resource-constrained environments. These playbooks will help safety-net clinics adopt AI responsibly without requiring enterprise-level resources.
What This Means for Healthcare Leaders
After three days of presentations, panels, and corridor conversations, several imperatives emerged for healthcare executives:
Start with problems, not technology. Barry Stein mentioned this principle six times in his presentation, and for good reason. The successful implementations all began with clearly defined pain points, not exciting technology looking for applications.
Governance is your competitive advantage. In a world where base AI capabilities are commoditizing, your ability to safely and reliably deploy AI becomes the differentiator. Build those muscles now.
Think portfolio, not projects. The leaders are managing AI as a portfolio of initiatives, not one-off projects. This requires new organizational capabilities and governance structures.
Invest in education at scale. Hartfordâs partnership with MIT to educate everyone from ward clerks to executives recognizes a fundamental truth: AI transformation requires widespread literacy, not just expert knowledge.
Prepare for the talent war. As one lunch conversation revealed, âThose really passionate about healthcare tend to outlastâ others in this space. Build teams that combine technical expertise with healthcare commitment.
Workforce Education at Scale
The AI transformation demands more than just technical deploymentâit requires comprehensive workforce readiness. Hartford Healthcareâs partnership with MIT has become the model, training everyone from ward clerks to cardiac surgeons on AI fundamentals. This isnât about creating AI specialists; itâs about ensuring every healthcare worker understands how to work alongside AI systems.
CHAI is scaling this approach nationally. A new nursing-specific curriculum developed with the American Nurses Association addresses the unique challenges nurses face as AI enters bedside care. Specialty societies in cardiology, radiology, and pathology are creating discipline-specific guideline tracks, recognizing that AI implementation varies dramatically by clinical context.
Most importantly, CHAI announced a patient-facing literacy program with the National Health Council. This initiative lets citizens understand what it means when an AI system participates in their careâincluding how to read a model card before consenting to AI-assisted treatment. As one panelist emphasized, âEducation, education, educationâwhich is governance.â
Solving the Alignment Problem
Perhaps the most philosophical yet practical challenge emerged from Mercy Healthâs announcement of their Value-Alignment Task Force, developed in partnership with faith-based bioethicists. The question theyâre tackling is deceptively simple: When AI systems disagree with human clinicians, who decides what constitutes the ârightâ answer?
Moderator Patrick VN posed the uncomfortable reality: âWe can have five specialists look at the same case and get five different opinions. Now weâre adding AI as a sixth voice. But unlike human disagreement, AI operates at scale. When itâs wrong, itâs wrong thousands of times.â
The task force is developing frameworks for encoding institutional values into AI systemsânot just clinical accuracy, but ethical priorities around end-of-life care, resource allocation, and patient autonomy. Itâs an acknowledgment that healthcare AI isnât just a technical challenge; itâs fundamentally about values, and different communities may reasonably want their AI systems to reflect different priorities.
Building a Career in Healthcare AI Leadership
The conversations at CHAI revealed something striking about professional opportunities. Six years ago, âclinical informaticistâ meant someone who managed EHR implementations. Today it means architecting AI governance frameworks, evaluating foundation models, designing agentic systems, and translating between clinical reality and technical possibility.
Every presentation underscored demand for professionals who bridge medicine and technology. Hartford Healthcareâs decade-long build of AI governance capabilities required hiring specialists who understand both HIPAA compliance and machine learning validation. Stanfordâs multi-model strategy needs people who can empirically test GPT-4 against Claude across clinical use cases. BrainHiâs deployment serving half of Puerto Rico required clinical informaticists who could design conversation flows that work in Spanish and English while maintaining medical accuracy.
The skill gaps are widening faster than traditional academic programs can address. Dr. Nigam Shah noted that 95% of medical AI research uses no actual EHR data because researchers lack access and expertise to work with real clinical systems. Meanwhile, health systems need people who can evaluate vendor claims, design pilot studies, analyze outcomes, and scale what works. Thatâs not a computer science problem or a clinical problemâitâs both simultaneously.
What makes this career inflection point different from previous health IT waves: AI deployment happens at physician-touching workflow levels. You canât design ambient documentation pilots without understanding clinical conversation patterns. You canât evaluate agentic AI for scheduling without knowing how appointment types, insurance verification, and provider preferences interact. Technology decisions increasingly require clinical judgment, and clinical decisions increasingly require technology expertise.
Michael Blumenthalâs observation about the procurement conversation shift applies equally to careers. Six years took healthcare from âWhat is AI?â to âHow quickly can we deploy?â That same acceleration is creating opportunities for professionals who positioned themselves at the intersection early. The clinical informaticists who understand both NIHâs research infrastructure plans and Hartfordâs governance frameworks will architect the systems that define the next decade of healthcare.
For physicians considering career pivots, nurses exploring informatics roles, or healthcare administrators building new skillsets, the message from CHAI was consistent: learn governance before chasing capabilities. The technical models will change every six months. The frameworks for safely evaluating, deploying, and monitoring healthcare AI will compound over decades.
The View from 2030
During the foundation model panel, moderator Sue Cho asked panelists to envision healthcare in 2030. The responses were telling. Graham from Kaiser worried that despite AI adoption, âthe staffing shortage will get way worseâ due to demographics. Others saw administrative roles transformed while clinical positions evolved rather than disappeared.
But perhaps the most provocative vision came from Grahamâs twenty-year outlook: âWill we even need health systems?â Itâs a question that would have seemed absurd just five years ago. Today, with AI agents handling routine interactions and patients increasingly comfortable with AI-mediated care, it deserves serious consideration.
The Road Ahead
As attendees departed Stanford, the mood was distinctly different from previous years. The question has shifted from âifâ to âhow fast.â The experimenters have become implementers. The governance skeptics have become framework advocates.
Yet challenges remain formidable. The pace mismatch between healthcareâs cautious culture and AIâs exponential advancement creates constant tension. Shadow AI usage threatens patient privacy. The workforce fears displacement even as shortages worsen. And underneath it all runs the fundamental question: Who decides what constitutes âgoodâ medical AI when even specialists disagree?
Whatâs clear is that healthcare AI has reached escape velocity. The combination of technological capability, economic pressure, and demonstrated outcomes has created irreversible momentum. As one attendee observed during lunch, âItâs nice to be up here with a bunch of people just talking innovation like itâs a regular day in the office.â
That normalizationâfrom exotic experiment to operational necessityâmay be CHAI 2025âs most important signal. Healthcare AI isnât coming. Itâs here. The only question now is how quickly and responsibly we can scale it to meet healthcareâs mounting challenges.
For organizations like OrbDoc, focused on bringing AIâs benefits to rural hospitals, the opportunity has never been clearer. As the market matures, success will come not from the most sophisticated AI, but from the most thoughtful implementation. In healthcareâs AI revolution, the winners will be those who remember that behind every algorithm is a human needing care.
The plane is indeed being built while flying, as Brian Anderson memorably put it. But after CHAI 2025, at least we know where weâre headed. And for the first time, we have the navigation tools to get there safely.
Ready to implement healthcare AI thoughtfully? Learn how AI medical scribes work or request a demo to see OrbDocâs approach to safe, effective clinical AI.