Medical Economics October 29, 2024
Key Takeaways
- GPT-4 did not significantly improve physicians’ diagnostic reasoning compared to conventional resources like UpToDate and Google.
- The study involved 50 U.S.-licensed physicians across family, internal, and emergency medicine.
- GPT-4 independently outperformed both physician groups, indicating potential for enhanced physician-AI collaboration.
- A bi-coastal AI evaluation network, ARiSE, has been established to further assess generative AI outputs in healthcare.
A recent study compared the diagnostic performance of physicians referencing AI to those limited to conventional resources.
Researchers from the University of Minnesota Medical School, Stanford University, Beth Israel Deaconess Medical Center and the University of Virginia analyzed the efficacy of GPT-4, an artificial intelligence (AI) large language model (LLM) system, as a diagnostic tool to assist physicians’ diagnoses....