In health care, the promise of speed is powerful—but it’s trust that makes real progress possible. Artificial intelligence is already unlocking new capabilities in care delivery, from streamlining workflows to spotting patterns clinicians can’t always see. But as adoption accelerates, one challenge remains: ensuring AI technologies are trusted by the people who rely on them every day.
This finding is central to the 2025 Philips Future Health Index (FHI) report—our 10th annual survey of health-care leaders, providers, and patients worldwide. And this year, the message is clear: Trust is key to using AI to deliver better care for more people.
Across 16 countries, we interviewed nearly 2,000 health-care professionals and over 16,000 patients. In the U.S., where health-care AI adoption is growing, clinicians are optimistic—63% believe AI can improve patient outcomes. But only 48% of U.S. patients feel the same, reflecting a broader trust gap we observed across markets.
This trust divide has consequences. If clinicians lack confidence, innovation stalls. If patients are hesitant, adoption slows. In both cases, the real cost is time—time that today’s strained health systems cannot afford to lose.
Reclaiming time for patient care
As a physician, I’ve seen the real-world impact of delayed care—and I’m not alone. According to the 2025 FHI, 83% of U.S. health-care professionals say they lose clinical time due to incomplete or inaccessible patient data, with two in five (44%) losing 45 minutes or more per shift.
These aren’t abstract challenges—they’re daily barriers to delivering quality care. And they’re addressable. But only if we accelerate the adoption of technologies that improve both the care experience and patient outcomes.
That’s where AI comes in—as a system-level catalyst that enhances care delivery. The health-care professionals we surveyed recognize that potential: 85% believe AI can help reduce administrative burden and overtime, while 74% say it could elevate staff skills and improve patient access.
From detecting subtle patterns in imaging to streamlining documentation, AI can free up what matters most: the time, attention and energy providers need to care for their patients. But to scale those benefits, we must first bridge the trust gap in health-care AI.
Defining trustworthy AI in health care
So what does trustworthy AI actually look like? It’s transparent, not opaque. It integrates seamlessly into clinical workflows instead of disrupting them. Most of all, it keeps people—providers and patients—at the center.
According to the 2025 FHI, doctors are the information source that would make patients feel most comfortable about the use of AI in their health care, with 79% saying so. That’s a powerful insight—and a reminder that earning patient trust starts by earning clinician trust first. When providers feel confident using AI, they can explain it clearly, use it responsibly, and help their patients feel comfortable with it too.
Technology should augment human expertise—not replace it. That’s why AI solutions should be co-developed with clinicians and health systems to ensure they’re designed for real-world use.
Delivering better care for more people
For health-care leaders weighing where to invest limited resources, the path forward isn’t always clear. But inaction comes at a cost. Delaying AI adoption doesn’t just slow innovation—it risks widening gaps in care and further straining a workforce already under pressure.
To close the trust gap and unlock AI’s full potential, three priorities must guide the way forward:
Human-led design. Health-care AI should be shaped by the people who use it—providers and patients alike. That means designing solutions that fit naturally into existing workflows, amplify clinical expertise, and meaningfully improve care delivery.
Proven performance—in real-world settings. Trust starts with results. AI must be safe, effective, and unbiased—designed with clinicians, tested across patient groups, and backed by clear evidence. That’s how we build confidence and ensure AI improves care for all.
Cross-sector collaboration. Trust in health-care AI can’t be built in silos. It takes close collaboration with clinicians, tech partners, policymakers, and patients to design solutions that truly work. We also need clear, consistent guidelines to move forward confidently and bring trusted AI to the point of care.
The future of health care isn’t just about delivering care faster. It’s about delivering better care for more people—care that’s more connected, more personal, and built on trust. The need is real, and the time is now. Let’s build that future together.
The opinions expressed in Fortune.com commentary pieces are solely the views of their authors and do not necessarily reflect the opinions and beliefs of Fortune.
This story was originally featured on Fortune.com
Recent Comments