Healthcare
Critical risk sector
AI in diagnostics, triage, and clinical decision support is subject to medical device rules, privacy laws, and patient safety standards.
Overview
Healthcare AI covers imaging, diagnostic support, triage, clinical documentation, and patient-facing assistants. Regulators look at safety, effectiveness, bias, and clinical validation.
What this means for you
AI may help read your scan, triage your symptoms, or recommend care. You have privacy rights under HIPAA (US), GDPR (EU), and similar laws, and providers should disclose AI use in important decisions.
Relevant laws & frameworks
-
EU AI Act
The world's first comprehensive horizontal AI law, imposing risk-based obligations across the EU.
-
Colorado AI Act
First comprehensive US state AI law targeting consequential decisions; effective in 2026.
-
NIST AI RMF
Voluntary US framework for managing AI risks across the life cycle (Govern, Map, Measure, Manage).
-
CCPA/CPRA ADMT
California is issuing regulations on automated decision-making technology under the CCPA/CPRA.
Business examples
-
Healthcare AI — FDA AI/ML device list
The FDA publishes a regularly updated list of AI/ML-enabled medical devices authorized in the US.