Inside Precision Medicine October 15, 2024
Anita Chakraverty

Most large language models (LLMs) used in health care settings, such as AI chatbots, are being assessed in a fragmented and inconsistent way that does not encompass real patient information, researchers have warned.

A new study, in the journal JAMA, found that just one in 20 evaluation studies included real patient data that reflected the complexities of clinical practice.

Most instead focused on the accuracy of answering medical examinations, which resulted in limited attention on considerations such as fairness, bias, toxicity and deployment.

The investigators said that testing these AI programs on hypothetical medical questions has previously been likened to certifying a car for road worthiness using a multiple-choice questionnaire.

“Real patient care data encompasses the complexities of clinical practice,...

Today's Sponsors

LEK
ZeOmega

Today's Sponsor

LEK

 
Topics: AI (Artificial Intelligence), Provider, Survey / Study, Technology, Trends
Generative AI Is Helping To Clear Up Brain Fog
Getting started with AI agents (part 1): Capturing processes, roles and connections
Unlocking The Genetic Code: AI Reveals New Insights Into Psychiatric Disorders
5 questions for the Abundance Institute's Neil Chilson
AI agents are unlike any technology ever

Share This Article