VentureBeat May 7, 2023
By Ivan Smetannikov, Serokell

Large language models (LLMs) are one of the hottest innovations today. With companies like OpenAI and Microsoft working on releasing new impressive NLP systems, no one can deny the importance of having access to large amounts of quality data that can’t be undermined.

However, according to recent research done by Epoch, we might soon need more data for training AI models. The team has investigated the amount of high-quality data available on the internet. (“High quality” indicated resources like Wikipedia, as opposed to low-quality data, such as social media posts.)

The analysis shows that high-quality data will be exhausted soon, likely before 2026. While the sources for low-quality data will be exhausted only decades later, it’s clear that the...

Today's Sponsors

Venturous
Got healthcare questions? Just ask Transcarent

Today's Sponsor

Venturous

 
Topics: AI (Artificial Intelligence), Big Data, Technology
Apple to open AI server factory in Texas as part of $500 billion U.S. investment
Nvidia Invests Further Into Healthcare And Releases The Largest Biology Foundation Model With The Arc Institute
Forget ChatGPT: Why Agentic AI Is The Next Big Retail Disruption
NHS trust group joins European network for responsible use of AI
How UCHealth monitors 22K hospital beds with AI

Share This Article