AI in Health Care Could Save Lives and Money—but Change Won’t Happen Overnight

Rapid diagnosis is one of the big promises of artificial intelligence for use in health care

Turgay Ayer, PhD/MSc

Turgay Ayer, PhD/MSc, is the Virginia C. and Joseph C. Mello Endowed Chair and a professor in the H. Milton Stewart School of Industrial and Systems Engineering at the Georgia Institute of Technology.

ViewFull Profile
Learn about ourEditorial Policies.
Published:Jul 14, 2025
|4 min read
Register for free to listen to this article
Listen with Speechify
0:00
4:00

Imagine walking into your doctor’s office feeling sick—and rather than flipping through pages of your medical history or running tests that take days, your doctor instantly pulls together data from your health records, genetic profile, and wearable devices to help decipher what’s wrong.

This kind of rapid diagnosis is one of the big promises of artificial intelligence for use in health care. Proponents of the technology say that over the coming decades, AI has the potential to save hundreds of thousands, even millions of lives.

What’s more, a 2023 study found that if the health care industry significantly increased its use of AI, up to US$360 billion annually could be saved.

But though artificial intelligence has become nearly ubiquitous, from smartphones to chatbots to self-driving cars, its impact on health care so far has been relatively low.

A 2024 American Medical Association survey found that 66 percent of US physicians had used AI tools in some capacity, up from 38 percent in 2023. But most of it was for administrative or low-risk support. And although 43 percent of US healthcare organizations had added or expanded AI use in 2024, many implementations are still exploratory, particularly when it comes to medical decisions and diagnoses.

I’m a professor and researcher who studies AI and health care analytics. I’ll try to explain why AI’s growth will be gradual, and how technical limitations and ethical concerns stand in the way of AI’s widespread adoption by the medical industry.

Inaccurate diagnoses, racial bias

Artificial intelligence excels at finding patterns in large sets of data. In medicine, these patterns could signal early signs of disease that a human physician might overlook—or indicate the best treatment option, based on how other patients with similar symptoms and backgrounds responded. Ultimately, this will lead to faster, more accurate diagnoses and more personalized care.

AI can also help hospitals run more efficiently by analyzing workflows, predicting staffing needs, and scheduling surgeries so that precious resources, such as operating rooms, are used most effectively. By streamlining tasks that take hours of human effort, AI can let health care professionals focus more on direct patient care.

But for all its power, AI can make mistakes. Although these systems are trained on data from real patients, they can struggle when encountering something unusual, or when data doesn’t perfectly match the patient in front of them.

As a result, AI doesn’t always give an accurate diagnosis. This problem is called algorithmic drift—when AI systems perform well in controlled settings but lose accuracy in real-world situations.

Racial and ethnic bias is another issue. If data includes bias because it doesn’t include enough patients of certain racial or ethnic groups, then AI might give inaccurate recommendations for them, leading to misdiagnoses. Some evidence suggests this has already happened.

Data-sharing concerns, unrealistic expectations

Health care systems are labyrinthian in their complexity. The prospect of integrating artificial intelligence into existing workflows is daunting; introducing a new technology like AI disrupts daily routines. Staff will need extra training to use AI tools effectively. Many hospitals, clinics, and doctor’s offices simply don’t have the time, personnel, money, or will to implement AI.

Also, many cutting-edge AI systems operate as opaque “black boxes.” They churn out recommendations, but even its developers might struggle to fully explain how. This opacity clashes with the needs of medicine, where decisions demand justification.

But developers are often reluctant to disclose their proprietary algorithms or data sources, both to protect intellectual property and because the complexity can be hard to distill. The lack of transparency feeds skepticism among practitioners, which then slows regulatory approval and erodes trust in AI outputs. Many experts argue that transparency is not just an ethical nicety but a practical necessity for adoption in health care settings.

There are also privacy concerns; data sharing could threaten patient confidentiality. To train algorithms or make predictions, medical AI systems often require huge amounts of patient data. If not handled properly, AI could expose sensitive health information, whether through data breaches or unintended use of patient records.

For instance, a clinician using a cloud-based AI assistant to draft a note must ensure no unauthorized party can access that patient’s data. US regulations such as the HIPAA law impose strict rules on health data sharing, which means AI developers need robust safeguards.

Privacy concerns also extend to patients’ trust: If people fear their medical data might be misused by an algorithm, they may be less forthcoming or even refuse AI-guided care.

The grand promise of AI is a formidable barrier in itself. Expectations are tremendous. AI is often portrayed as a magical solution that can diagnose any disease and revolutionize the health care industry overnight. Unrealistic assumptions like that often lead to disappointment. AI may not immediately deliver on its promises.

Finally, developing an AI system that works well involves a lot of trial and error. AI systems must go through rigorous testing to make certain they’re safe and effective. This takes years, and even after a system is approved, adjustments may be needed as it encounters new types of data and real-world situations.

Incremental change

Today, hospitals are rapidly adopting AI scribes that listen during patient visits and automatically draft clinical notes, reducing paperwork and letting physicians spend more time with patients. Surveys show over 20 percent of physicians now use AI for writing progress notes or discharge summaries. AI is also becoming a quiet force in administrative work. Hospitals deploy AI chatbots to handle appointment scheduling, triage common patient questions and translate languages in real time.

Clinical uses of AI exist but are more limited. At some hospitals, AI is a second eye for radiologists looking for early signs of disease. But physicians are still reluctant to hand decisions over to machines; only about 12 percent of them currently rely on AI for diagnostic help.

Suffice to say that health care’s transition to AI will be incremental. Emerging technologies need time to mature, and the short-term needs of health care still outweigh long-term gains. In the meantime, AI’s potential to treat millions and save trillions awaits.

This article is republished from The Conversation under a Creative Commons license. Read the original article.


Turgay Ayer, PhD/MSc

Turgay Ayer is the Virginia C. and Joseph C. Mello Endowed Chair and a professor in the H. Milton Stewart School of Industrial and Systems Engineering at the Georgia Institute of Technology, where he also serves as the director of Business Intelligence and Healthcare Analytics at the Center for Health and Humanitarian Systems. In addition, Dr. Ayer serves as a senior scientist at the Centers for Disease Control and Prevention (CDC) and holds a courtesy appointment at Emory University School of Medicine.

Dr. Ayer’s research focuses on healthcare analytics, AI and generative AI in medicine, socially responsible operations, and practice-focused research. He has published over 80 peer-reviewed articles in top management, engineering, medical, and health policy journals. His work has also been widely covered by major media outlets such as The Wall Street Journal, The Washington Post, US News, NPR, and Fox News. In addition to his research, Ayer actively teaches courses on AI and healthcare to engineering, management, and medical students, fostering interdisciplinary education in these rapidly evolving fields.

Ayer's research has made a significant policy impact, notably changing blood collection operations for the American Red Cross nationwide, influencing lung cancer drug pricing and coverage decisions in the U.S. and U.K., optimizing distribution strategies for scarce hepatitis C treatment drugs in correctional health systems, and improving the allocation of medical surplus products to the developing world by a major medical surplus recovery organization.

In recognition of his contributions, Ayer has received numerous best paper awards and has been named a Bonder Scholar and an Edelman Laureate. His research has been supported by over $3 million in grant funding from various institutions, including the National Science Foundation.

Ayer is actively involved with INFORMS and has been a long-term member for over 15 years. He currently serves as a Department Editor for one of the flagship journals of INFORMS, Operations Research, in the area of Real-World OR Innovations. Additionally, he is a past president of the INFORMS Health Applications Society, the second-largest society within INFORMS.


Tags:

Health CareDiagnosticsArtificial Intelligence
Top Image:
Though AI has become nearly ubiquitous, from smartphones to chatbots to self-driving cars, its impact on health care so far has been relatively low.