Federal health technology regulators on Wednesday finalized new rules to force software vendors to disclose how artificial intelligence tools are trained, developed, and tested — a move to protect patients against biased and harmful decisions about their care.
The rules are aimed at placing guardrails around a new generation of AI models gaining rapid adoption in hospitals and clinics around the country. These tools are meant to help predict health risks and emergent medical problems, but little is publicly known about their effectiveness, reliability, or fairness.
Starting in 2025, electronic health record vendors who develop or supply these tools, which increasingly use a type of AI known as machine learning, will be required to disclose more technical information to clinical users about their performance and testing, as well as the steps taken to manage potential risks.
To submit a correction request, please visit our Contact Us page.
STAT encourages you to share your voice. We welcome your commentary, criticism, and expertise on our subscriber-only platform, STAT+ Connect