Health care companies are racing to incorporate generative AI tools into their product pipelines and IT systems after the technology displayed an ability to perform many tasks faster, cheaper — and sometimes better — than humans.
But the rush to harness the power of so-called large language models, which are trained on vast troves of data, is outpacing efforts to assess their value. AI experts are still trying to understand, and explain, how and why they work better than prior systems, and what blind spots might undermine their usefulness in medicine.
It remains unclear, for example, how well these models will perform, and what privacy and ethical quandaries will arise, when they’re exposed to new types of data, such as genetic sequences, CT scans, and electronic health records. Even knowing exactly how much data must be fed into a model to achieve peak performance on a given task is still largely guesswork.
This article is exclusive to STAT+ subscribers
Unlock this article — and get additional analysis of the technologies disrupting health care — by subscribing to STAT+.
Already have an account? Log in
Already have an account? Log in
To submit a correction request, please visit our Contact Us page.
STAT encourages you to share your voice. We welcome your commentary, criticism, and expertise on our subscriber-only platform, STAT+ Connect