VentureBeat December 18, 2024
Ben Dickson

The AI research community continues to find new ways to improve large language models (LLMs), the latest being a new architecture introduced by scientists at Meta and the University of Washington.

Their technique, Byte latent transformer (BLT), could be the next important paradigm for making LLMs more versatile and scalable.

BLT solves one of the longstanding problems of LLMs that operate at byte level as opposed to tokens. BLT can open the way for new models that can process raw data, are robust to changes and don’t rely on fixed vocabularies.

Tokens vs bytes

Most LLMs are trained based on a static set of tokens, predefined groups of byte sequences.

During inference, a tokenizer breaks the input sequence down into...

Today's Sponsors

Venturous
Got healthcare questions? Just ask Transcarent

Today's Sponsor

Venturous

 
Topics: AI (Artificial Intelligence), Technology
OpenAI Launches New Tools for Building AI Agents
AI And Platforms Challenge A Key Management Theory. Again.
Norm Ai Raises $48 Million to Develop Regulatory AI Agents
The Dangers Of Automated Governance: Can AI Lead A Nation?
The Rise Of The One-Person Unicorn: How AI Agents Are Redefining Entrepreneurship

Share This Article