The Challenges of AI in Healthcare
Navigating Complexities to Unlock AI’s Full Potential in Healthcare
Vikram Narayan
Nov 1, 2024
With all the talk about AI in healthcare, there are still many challenges. The technology is still immature, Ethical concerns, complexity in integration and governance, and costs of maintaining such systems are cropping up. In this article we review some of the major challenges of applying AI to healthcare.
To begin with, let’s start with a fundamental question.
What is AI?
While AI today has become synonymous with ChatGPT-like technologies, the real definition of AI is far more expansive. The following diagram helps most of my workshop participants when thinking about how the various pieces of AI actually fit together.
AI is an all-encompassing term that includes a variety of techniques and algorithms that help machines simulate human intelligence (including playing chess or solving Sudoku puzzles). Inside this broad umbrella called ‘AI’ we’ve got ML (Machine Learning) which is the idea of being able to learn from data rather than hard-coded rules. Within the umbrella of machine learning, we’ve got Deep Learning which is a kind of machine learning technique which uses massive amounts of data and many layers of ‘neurons’ to learn from and process that data. Then, we’ve got Generative AI (or Gen AI) which, having learned from data, attempts to produce new data - be it images or text. And finally, we’ve got LLM’s or Large Language Models - the technology that powers ChatGPT.
So when two people use the word AI, they may be talking about two very different things!
However, today, when most people mention the word, AI, they are referring to technologies like Gemini, Claude or ChatGPT - these seemingly super-intelligent beings capable of doing many things that humans can do and which promise to solve all of humanity's problems and this class of AI will be the focus of our article.
Now, this brings up the next question:
How Does ChatGPT Actually Work?
For all of the magical things that ChatGPT does, we need to bear in mind that it’s just a guessing machine which guesses the next set of words (or ‘tokens) based on prior words (or tokens). And this idea is best explained with the following example:
Here, the large language model is evaluating which word to produce after producing the words “Twinkle, twinkle, little …”. The model has access to various alternative words with varying probabilities and finally chooses “star” based on its probability of being the right word.
So Large Language Models are simply “stochastic parrots” that can guess the next word really, really well!
And given this context, we are in a great place to examine the challenges that we face when attempting to integrate AI into healthcare.
Challenge 1 - Model Hallucination
When ChatGPT first came out, there were a number of instances where naive users were caught flat-footed when attempting to use the technology. You’ve probably seen headlines where ChatGPT invented new legal cases when it first came out in 2023.
This ‘model hallucination’ occurs because the model does not know where its knowledge ends and where ignorance begins and because it’s simply predicting tokens based on previous tokens. And while the problem of hallucination is going down dramatically, it still exists and we can never be 100% sure if a large language model will spout an inaccuracy. For example, there have been instances of hallucinations in generating medical summaries.
Challenge 2 - Blackbox Decision Making
If you ask a human why they made a particular decision or said something, they’ll often be able to give you a rational explanation. If you ask a large language model why it makes a decision, it gives you a rational-sounding explanation and we can never be too sure as to why it made a particular choice. A lot of work is being done to mitigate these challenges with an idea called ‘explainable AI’ but we’re still far away from a model being able to explain clearly why it’s doing or saying something.
Challenge 3 - Inconsistent Behavior
AI models are complex creations. They have billions (or even trillions) of ‘knobs’ that can be turned and tuned. And which of these knobs light up can vary based on the inputs. Ask a question in a different way and you may get a different answer. This inconsistency, while great for creative tasks, is not well-suited for tasks that require consistent outputs.
Challenge 4 - PHI Safety and HIPAA Compliance
An additional consideration in the healthcare space is complying with HIPAA requirements and protecting PHI (Protected Health Information). The challenging part of this is that PHI will need to be sent to a large language model provider like Open AI or AWS Bedrock for most meaningful applications. And even with a BAA (Business Associate Agreement), it’s still a challenge to ensure ongoing compliance over the long run - especially if there are multiple vendors. The recent incident involving a Microsoft healthcare chatbot potentially exposing patient data is particularly instructive.
Challenge 5 - The Uncanny Valley
Humans appreciate it when AI becomes human-like. But only up to a point. If the AI becomes too ‘human’ we tend to feel repelled. This design principle - referred to as “the uncanny valley” - is likely to play out in the healthcare space as patients reject AI caregivers and AI admin staff trying too hard to sound and act human. The refrain “let me talk to a real human” is likely to grow louder as health systems try to save costs by cramming AI nurses down the throats of unwilling patients.
Overcoming Challenges for AI in Healthcare
Apply AI to the Right Problems:
Differentiate Mission-Critical from Life-Critical Tasks: To safely deploy AI, healthcare organizations should categorize tasks as mission-critical (e.g., efficiency improvements) or life-critical (e.g., diagnosis or treatment recommendations). For life-critical tasks, AI should act in a supportive role rather than making decisions independently to minimize patient risk.
Use AI for Non-Sensitive, High-Impact Tasks: AI can significantly streamline non-patient-facing areas like claims processing, appointment scheduling, and administrative workload reduction, where errors don't directly affect patient outcomes.
Embed Domain Knowledge Checks: In AI tools like chatbots, integrating knowledge verified by healthcare professionals can help avoid dangerous recommendations, providing safe support without compromising quality.
Establish Comprehensive AI Governance:
Create a Cross-Functional AI Governance Committee: A governance committee should include stakeholders from IT, clinical staff, legal, compliance, and executive leadership to address technical, ethical, and operational concerns. This group can act as a checkpoint for new AI initiatives.
Monitor Bias, Security, and Transparency: AI governance must address data ownership, systemic bias, and transparency to ensure AI outputs are fair, accurate, and explainable. For example, setting guidelines for data usage and periodic bias testing can help prevent discrimination.
Regular Audits and Accountability Measures: Introduce audits that regularly assess AI performance, checking for data integrity, accuracy, and regulatory compliance. Assign accountability to specific roles to ensure any deviations are addressed promptly.
Enhance Explainability and Transparency in AI:
Invest in Explainable AI (XAI) Solutions: Developing explainable AI methods allows clinicians to better understand AI-driven insights, fostering trust and effective integration with clinical workflows. Techniques like local interpretable model-agnostic explanations (LIME) can break down complex decisions into understandable formats.
Transparency in AI Training Data: Clearly document and communicate the sources, limitations, and biases of training data to help clinicians understand the context and potential gaps in AI models. This can also guide users on how best to leverage AI recommendations.
Provide User Training and Decision-Making Protocols: For models that are less transparent, organizations can train users to identify potential AI limitations and establish protocols for decision-making, where AI is consulted as one part of a broader decision-making process.
Safeguard Patient Privacy and Compliance with PHI Standards:
Implement Encrypted Data Pipelines: PHI should be protected by strong encryption both in transit and at rest. If using cloud-based models, organizations should rely on secure APIs and encryption standards to minimize data breaches.
Select Vendors with HIPAA-Compliance: Choose AI providers with clear HIPAA compliance documentation and comprehensive data protection measures, ensuring they adhere to Business Associate Agreements (BAAs) and other regulatory mandates.
Introduce a ‘Minimum Necessary’ Standard: For AI models, only the necessary amount of PHI should be used to achieve specific healthcare outcomes. Limiting data exposure not only mitigates risk but also simplifies compliance by reducing sensitive data in the system.
Address the Uncanny Valley in AI-Patient Interactions:
Set Boundaries for AI’s Role in Patient Care: Define roles for AI that align with patient comfort levels, such as supporting tasks without pretending to be fully human. For instance, AI assistants can provide information without mimicking emotional responses, which might make patients uncomfortable.
Conduct Patient-Centric Design Research: Gather patient feedback to better understand how AI’s tone, appearance, and behavior can be adjusted to avoid the uncanny valley effect. This data can help tailor AI interactions to be helpful and non-intrusive.
Design AI with Clear Human-Support Options: Ensure patients can easily reach human support when desired, promoting transparency and a patient-centered approach to healthcare that integrates AI effectively without displacing human connection.
Conclusion
While AI offers significant potential to transform healthcare, realizing that potential requires thoughtful, responsible integration. By targeting non-sensitive tasks, establishing robust governance, enhancing transparency, and safeguarding patient privacy, healthcare organizations can mitigate risks and maximize AI’s benefits.