VentureBeat March 28, 2024
Michael Nuñez

A new study from Google’s DeepMind research unit has found that an artificial intelligence system can outperform human fact-checkers when evaluating the accuracy of information generated by large language models.

The paper, titled “Long-form factuality in large language models” and published on the pre-print server arXiv, introduces a method called Search-Augmented Factuality Evaluator (SAFE). SAFE uses a large language model to break down generated text into individual facts, and then uses Google Search results to determine the accuracy of each claim.

“SAFE utilizes an LLM to break down a long-form response into a set of individual facts and to evaluate the accuracy of each fact using a multi-step reasoning process comprising sending search queries to Google Search and determining whether...

Today's Sponsors

LEK
ZeOmega

Today's Sponsor

LEK

 
Topics: AI (Artificial Intelligence), Survey / Study, Technology, Trends
Mayo, AdventHealth, Emory: 6 Big Tech health system partnerships
Apple’s OpenGL and iOS 18: Shaping the future of health care technology
Altman handpicked for Homeland Security's AI safety board
Royal Papworth CIO presents six actions for trusts’ AI journey
Hospitals Are Using AI To Help Manage Patient Messages to Physicians

Share This Article