MIT Technology Review April 25, 2024
Will Douglas Heaven

In many high-stakes situations, large language models are not worth the risk. Knowing which outputs to throw out might fix that.

Large language models are famous for their ability to make things up—in fact, it’s what they’re best at. But their inability to tell fact from fiction has left many businesses wondering if using them is worth the risk.

A new tool created by Cleanlab, an AI startup spun out of a quantum computing lab at MIT, is designed to give high-stakes users a clearer sense of how trustworthy these models really are. Called the Trustworthy Language Model, it gives any output generated by a large language model a score between 0 and 1, according to its reliability. This lets...

Today's Sponsors

LEK
ZeOmega

Today's Sponsor

LEK

 
Topics: AI (Artificial Intelligence), Technology
How to prepare your workforce to think like AI pros
Warren Buffett warns of AI risks
UK sets out position on regulating AI as a medical device
AI: A powerful tool for improving health care efficiency and safety
Enhancing Global Mental Health Care With Digital Tools and AI for Scalable Interventions

Share This Article