• Thread Author
Artificial intelligence has begun to fundamentally reshape the contours of healthcare, offering the tantalizing promise of smarter diagnostics, more efficient workflows, and democratized medical expertise. Yet, as the technology gallops ahead, the scaffolding of laws, norms, and ethics lags behind, struggling to keep pace with the speed and scale of innovation. This is the reality explored in a compelling Microsoft Research podcast series, which convenes leaders at the intersection of technology, medicine, and bioethics to discuss both the breakthroughs and challenges of governing AI in health.

The Dawn of Generative AI in Medicine​

When ChatGPT was launched in November 2022, followed quickly by GPT-4 in the spring of 2023, it marked a seismic shift for healthcare AI. Peter Lee, a senior executive at Microsoft Research and co-author of “The AI Revolution in Medicine,” describes this moment as a "revolution in cheap expertise," opening the floodgates to advanced large language models accessible through any smartphone. Unlike electronic health records or telemedicine—previous digital disruptions that remained siloed in scope—generative AI has the potential to touch every corner of healthcare, from frontline clinicians to underserved rural patients.
Laura Adams, senior advisor at the National Academy of Medicine (NAM) and leader of its Artificial Intelligence Code of Conduct initiative, highlights four defining characteristics of generative AI’s impact:
  • Extraordinary speed of development and improvement
  • Scalable solutions, especially for tasks like ambient clinical listening and medical scribing
  • Ubiquity, with universal access via smartphones
  • The democratization of expertise, offering high-level knowledge at low cost
But Adams also underscores the sobering reality: technology reflects the data it’s trained on—data that is often riddled with historical bias and inequity. The same tools democratizing expertise also risk deepening the digital divide between well-resourced urban hospitals and underfunded rural or inner-city providers.

AI as Savior, AI as Source of Harm: The Need for Guardrails​

The tension is palpable: the optimism for transformative improvement is balanced by the recognition of new and old risks. Adams, reflecting on the spread of electronic health records, notes that poor design and lack of provider input led to unintended consequences—fatigue and time away from patient care. There’s a lesson here: innovation must go hand in hand with ethical frameworks, thoughtful governance, and a focus on real-world impact.
To that end, Adams and the NAM embarked on a nationwide—indeed, increasingly international—effort to craft an AI Code of Conduct. The project’s core ambitions are clear:
  • Protect and advance human health and connection: In a world where technology becomes omnipresent, the primary goal must remain patient welfare and the preservation of meaningful human relationships.
  • Ensure equity in distribution of risks and benefits: It's not enough for AI to "work"—it must work for everyone, not just the privileged.
  • Embed monitoring and openness into the lifecycle: Post-implementation feedback and transparency are essential for safe, effective AI deployment.
These and other principles—distilled down from extensive analysis of over sixty existing frameworks—are designed to be both accessible and adaptable, functioning as a touchstone rather than a rigid prescription.

Beyond Technical Specifications: Ethics and Inclusivity​

What sets the NAM’s initiative apart is its insistence on inclusivity—not only by involving clinicians, technologists, and ethicists in guidelines development, but also by including patient advocates as true co-designers. This moves the discussion beyond box-ticking compliance to a more participatory model of ethics, where those directly impacted by AI have real agency in shaping it.
This inclusive approach is echoed by Vardit Ravitsky, President and CEO of The Hastings Center for Bioethics, who emphasizes the need for bioethics to act not as a brake on AI innovation but as a facilitator for responsible, context-sensitive adoption. Ravitsky observes that many concerns—such as patient autonomy, privacy, trust, and informed consent—are old dilemmas in a powerful new guise.
Yet, as AI moves from reading medical images (potentially outperforming radiologists in certain settings) to generating empathetic patient notes or providing decision support, new nuances emerge. A recent study cited by Ravitsky found that patients expressed even greater satisfaction with AI-generated responses than from human clinicians—unless they were told the message was written by AI, at which point satisfaction dropped. This conundrum reveals a delicate balance between transparency (disclosure of AI use) and the potential positive impact on patient experiences.

The Perils of Fragmentation and the Case for Governance Interoperability​

A recurring theme throughout these expert conversations is the fragmentation of regulatory oversight and ethical norms. Adams compares the current patchwork of AI guidelines to the once balkanized world of health information exchange in the US, where different privacy laws and data standards made it difficult—and expensive—to transfer patient data across state lines. Without a similar push for governance interoperability, Adams warns, healthcare may squander the potential of AI while exposing vulnerable populations to greater risk.
To address this, the NAM code of conduct seeks common ground among divergent frameworks, while allowing for context-specific adaptation. This process is inherently iterative and dynamic—a crucial recognition in a field moving as fast as AI in health.

Unpacking the Issue of Bias: Spotlight on Clinical AI​

Perhaps the thorniest challenge facing AI in health is bias. Dr. Roxana Daneshjou, assistant professor at Stanford and an expert on fairness in medical AI, has shown in multiple landmark studies that algorithms trained on unrepresentative data can systematically underperform—and sometimes cause harm—for underrepresented groups. Her research on dermatological image recognition, for instance, demonstrated that computer vision models fared significantly worse on brown and black skin but that this bias could be reduced by diversifying training data.
Daneshjou’s later work, focusing on race-based medicine in large language models, exposed how even the most advanced AI (including GPT-4 and Claude) can unwittingly repeat harmful medical myths—such as the incorrect use of race in estimating kidney function. Importantly, her team also pioneered the idea of using ensembles of AI models to audit one another for bias, and publicly released datasets to spur community-wide progress.
Yet, bias isn’t solely a technical flaw to be debugged. As Daneshjou notes, even the “fairest” AI, when combined with human clinicians operating within structurally biased systems, can perpetuate inequities. Her call to action is twofold: improve both the models and the systems in which they are deployed, working toward synergy rather than replacement.

Regulation in an Era of Generative AI: Old Tools, New Realities​

Adequately regulating AI in healthcare is proving to be a wicked problem. Traditional approaches—certifying tools as “safe” before deployment—are insufficient when models can change, adapt, and interact with real-world data in unpredictable ways. Adams points to the emergence of regulatory sandboxes and provocative new proposals—such as whether to regulate generative AI like a licensed clinician, subjecting it to ongoing assessment, reporting, and “retraining” as medical knowledge evolves.
Another underappreciated facet is the proliferation of AI technologies directly in the hands of consumers. As patients gain agency and autonomy using these tools, outdated regulatory paradigms—built around professional oversight—may need to be completely rethought. Both Adams and Ravitsky emphasize the inevitability of patients leveraging AI and the need to channel this “revolution” to reinforce, rather than erode, the patient-provider relationship.

From Black Boxes to Trustworthy Systems: Transparency and the Limits of Informed Consent​

Perhaps the greatest philosophical and practical debate centers around trust. If we are comfortable with “black box” elements in medical practice—a surgeon’s skill, a doctor’s mood, the unmeasured variation between clinicians—should AI be singled out for special scrutiny? Ravitsky warns of “AI exceptionalism,” where technology is held to an unrealistically high standard simply because it is new or unfamiliar. Yet the consequences of error—particularly at scale—remain profound, justifying the continued emphasis on transparency, explainability, and informed consent.
Still, tests of patient acceptance are evolving. While some find comfort in knowing a real human wrote their care instructions, others are reassured—even impressed—by the thoroughness and empathy of AI-generated messages. As AI becomes more integrated and less visible, the ethical imperative may shift from blanket disclosure to contextually appropriate transparency—tailored to the material impact on care and the evolving expectations of society.

Can AI Replace the Clinician? The Future of Medical Expertise​

Fears that AI would make radiologists or dermatologists obsolete have been widespread—yet premature. Daneshjou notes the profound complexity of clinical reasoning, which involves not just image interpretation but tactile diagnosis, biopsies, treatment planning, and, crucially, the ability to manage ambiguity and disagreement. Even the best AI can handle well-bounded, structured tasks, but human judgment remains essential for the many cases that defy clear-cut rules.
Instead, the emerging vision is not replacement, but augmentation: a “human-AI system” in which each compensates for the other’s weaknesses. The challenge is to ensure that this synergy closes, rather than exacerbates, gaps in access, quality, and safety.

Environmental and Societal Trade-offs: The Hidden Costs​

One surprising lesson from the NAM code of conduct development was the importance of environmental considerations. As Adams recounts, even leaders focused primarily on human health were forced to confront the outsized carbon footprint of advanced AI models—a reminder that innovation always comes with trade-offs, some of which may only become clear in hindsight.

Final Analysis: Strengths, Risks, and the Road Forward​

The integration of generative AI into healthcare is perhaps the most consequential technological shift in medicine since the widespread adoption of digital records. Its potential strengths are formidable:
  • Reduction in clinician burnout via automation of documentation and administrative tasks
  • Improved diagnostic accuracy in imaging, language, and multimodal data
  • New pathways for patient engagement, information access, and self-advocacy
  • The democratization of high-level expertise, potentially redressing geographic and economic inequities
Yet the risks are just as real:
  • Amplification of existing biases, resulting in systemic harm to already underserved populations
  • Novel privacy, security, and consent dilemmas, especially as AI moves closer to the patient
  • Potential for eroded trust if technologies are implemented without transparency or recourse
  • The proliferation of fragmented, incompatible regulatory frameworks, which could stifle both safety and innovation
The outlook, therefore, is neither utopian nor dystopian, but balanced and contingent. The principles articulated by the National Academy of Medicine—a commitment to human health, equity, participatory governance, iterative oversight, and environmental stewardship—offer a solid foundation, but will only succeed if adopted in spirit as well as in letter.

Conclusion​

As AI becomes ever more deeply woven into the fabric of medicine, the twin imperatives of safety and access must guide its evolution. The stories of Adams, Ravitsky, and Daneshjou show that genuine progress depends less on dazzling technology than on humble, inclusive dialogue—anchored in evidence, open to critique, and informed by lived experience. The future of AI in healthcare will be shaped not simply by what is technically possible, but by the collective will to harness its power for all, while never losing sight of the human at its heart.

Source: Microsoft Laws, norms, and ethics for AI in health