MedPage Today April 1, 2024
— GPT-4 earned higher clinical reasoning scores than residents and attendings
When evaluating simulated clinical cases, Open AI’s GPT-4 chatbot outperformed physicians in clinical reasoning, a cross-sectional study showed.
Median R-IDEA scores — an assessment of clinical reasoning — were 10 for the chatbot, 9 for attending physicians, and 8 for internal medicine residents, Adam Rodman, MD, of Beth Israel Deaconess Medical Center in Boston, and colleagues reported in a research letter in JAMA Internal Medicine.
In logistic regression analysis, GPT-4 had the highest estimated probability of achieving high R-IDEA scores (0.99, 95% CI 0.98-1.00) followed by attendings (0.76, 95% CI 0.51-1.00) and residents (0.56, 95% CI 0.23-0.90), with the chatbot being significantly higher than both attendings (P=0.002) and residents...