MedCity News July 22, 2024
Arundhati Parmar

Anthropic’s Claude3-Opus performed better than GPT-4, but both fell short of humans on a test of objective medical knowledge. The study was conducted by a firm developing LLMs specifically for healthcare that claim to be incorporating peer-reviewed sources of information.

A new study that pitted six humans, OpenAI’s GPT-4 and Anthropic’s Claude3-Opus to evaluate which of them can answer medical questions most accurately found that flesh and blood still beat out artificial intelligence.

Both the LLMs answered roughly a third of questions incorrectly though GPT-4 performed worse than Claude3-Opus. The survey questionnaire were based on objective medical knowledge drawn from a Knowledge Graph created by another AI firm – Israel-based Kahun. The company created their proprietary Knowledge Graph with a...

Today's Sponsors

LEK
ZeOmega

Today's Sponsor

LEK

 
Topics: AI (Artificial Intelligence), Physician, Provider, Survey / Study, Technology, Trends
Expert Insights on How Utilization Management Drives Physician Burnout
Shaping the Future of Cardiology: Key Takeaways From AHA 2024
109 hospitals receiving new Medicare-backed residency slots
STAT+: UnitedHealth pays its own physician groups considerably more than others, driving up consumer costs and its profits
AI Robot Scanner as Good as Rheumatologists at Assessing RA

Share This Article