What AI Can Do – And What It Can’t

by Grace Chen

For decades, the promise of digital health has followed a predictable, often frustrating cycle: a wave of innovation arrives with grand promises of transformation, only to deliver incremental improvements that often leave clinicians more burdened than they were before. From the early days of electronic health records to the rollout of telehealth, the technology frequently failed to account for the chaotic, human reality of the bedside.

But according to John Halamka, president of the Mayo Clinic Platform, we have reached a tipping point where the trajectory has fundamentally shifted. For Halamka, a veteran of health IT who has advised multiple presidential administrations, the current era of artificial intelligence is not just another iteration of digital health. It is a categorical revolution driven by a “perfect storm” of available compute power, sufficient multimodal data, and clearly defined clinical use cases.

The goal is no longer simply to digitize a paper process—which often resulted in clinicians spending half their shifts staring at keyboards—but to use AI to return the focus to the patient. By understanding what AI can do in healthcare and, more importantly, where its boundaries lie, the medical community is attempting to build a framework that prioritizes safety over hype.

Moving from Documentation to Dialogue

One of the most immediate impacts of AI is the reduction of the “administrative tax” on providers. In previous iterations of health IT, such as the “meaningful use” era following the 2009 HITECH Act, clinicians were required to enter vast amounts of data—sometimes over 140 elements per visit—to satisfy regulatory and quality measures. This created a physical and emotional barrier between the doctor and the patient.

Moving from Documentation to Dialogue
Dialogue One

Ambient AI is now being deployed to dismantle that barrier. By using ambient listening, AI can capture the natural dialogue between a provider and a patient, “automagically” populating the medical record. At Mayo Clinic, the objective for nursing staff is a future where a nurse does not need to touch a keyboard during an entire shift, allowing them to work at the top of their license by focusing on empathy and active listening.

Beyond documentation, AI is proving its worth in high-stakes diagnostics. In cardiology, for example, algorithms can now analyze ECG data gathered from consumer-grade devices in a patient’s home. This capability can potentially spare patients from expensive, invasive procedures by identifying conduction defects or confirming heart health through non-invasive, algorithm-driven screening.

The Guardrails: Data Cards and Model Cards

Because the stakes in medicine are life and death, the deployment of AI requires a level of rigor similar to pharmaceutical trials. The danger is not just a “hallucination” in the text, but “data drift”—where an algorithm trained on one population fails when applied to another.

To combat this, Mayo Clinic utilizes a system of “data cards” and “model cards.” A data card explicitly defines the phenotype, genotype, and environment of the training set. This prevents the mistake of applying an algorithm trained on a specific demographic—such as Scandinavian Lutherans in Minnesota—to a vastly different population in rural Georgia without proper validation.

Model cards further track how the AI performs across different races, ethnicities, ages, and zip codes. This is paired with a six-rank risk stratification process. An algorithm that suggests a patient walk more steps carries a low risk of harm; an algorithm that automatically injects insulin carries a critical risk. The level of qualification and surveillance required is scaled according to that risk.

AI Capability Clinical Constraint / Risk Mitigation Strategy
Ambient Charting Accuracy of captured dialogue Clinician review and sign-off
Predictive Diagnostics Data drift across populations Data and Model Cards
Triage & Referrals False positives/negatives Human-in-the-loop validation
Autonomous Action Catastrophic system failure Strict risk stratification

Where AI Falls Short

Despite the progress, AI cannot yet replace the nuanced judgment of a physician. Currently, most clinical AI acts as a “smart consultant” rather than an autonomous decision-maker. Whether in radiology or digital pathology, the AI highlights areas of concern, but a human remains “nearby” to make the final call.

John Halamka, M.D. – President : Mayo Clinic Platform – A very unique opportunity

There is also a significant cultural and legal hurdle. In many specialties, such as radiology, clinicians express frustration with AI not because it is inaccurate, but because of the legal burden of disagreement. If an AI has a 95% positive predictive value, the time and documentation required for a doctor to legally justify disagreeing with the AI can outweigh the benefit of the tool.

the risk of “agentic AI”—systems that can take action autonomously—introduces severe cybersecurity concerns. Without rigorous security stacks, an autonomous agent could potentially be hijacked by a lousy actor to disrupt entire healthcare delivery systems.

Solving the Workforce Crisis

The urgency to integrate AI is not merely about efficiency; it is a necessity driven by global demographics. In industrialized nations, particularly Japan and South Korea, birth rates have fallen well below replacement levels while lifespans continue to increase. This creates a widening gap between the demand for care and the number of available clinicians.

AI is seen as the primary tool to close this gap by extending the capabilities of mid-level providers, such as nurse practitioners and physician assistants. By providing these professionals with high-level decision support, health systems can deliver specialized care to more patients in more regions, reducing the reliance on a dwindling number of specialists.

To ensure these tools are safe and standardized, the Coalition for Health AI (CHAI) is working to establish a “community standard of care.” Much like malpractice is judged by whether a doctor deviated from accepted standards, AI governance requires a shared definition of what constitutes a “good enough” and safe algorithm.

The final hurdle is educational. Many medical schools still rely on a curriculum based on memorization rather than data science. To safely navigate this revolution, the next generation of physicians must be trained not just as clinicians, but as AI interpreters who can critically assess whether a tool’s output is credible or a hallucination.

Disclaimer: This article is for informational purposes only and does not constitute medical advice. Always seek the advice of your physician or other qualified health provider with any questions you may have regarding a medical condition.

The next major milestone in AI governance will be the continued development of community standards through organizations like CHAI, as the industry moves toward a more formalized “post-market surveillance” model for medical algorithms.

Do you think AI will eventually replace the need for certain medical specialties, or will it always require a human in the loop? Share your thoughts in the comments.

You may also like

Leave a Comment