VentureBeat December 18, 2024
Ben Dickson

The AI research community continues to find new ways to improve large language models (LLMs), the latest being a new architecture introduced by scientists at Meta and the University of Washington.

Their technique, Byte latent transformer (BLT), could be the next important paradigm for making LLMs more versatile and scalable.

BLT solves one of the longstanding problems of LLMs that operate at byte level as opposed to tokens. BLT can open the way for new models that can process raw data, are robust to changes and don’t rely on fixed vocabularies.

Tokens vs bytes

Most LLMs are trained based on a static set of tokens, predefined groups of byte sequences.

During inference, a tokenizer breaks the input sequence down into...

Today's Sponsors

LEK
ZeOmega

Today's Sponsor

LEK

 
Topics: AI (Artificial Intelligence), Technology
Johns Hopkins Medicine inks AI deal with Abridge
Congress' AI report leaves some tech-watchers on edge
Only 20% of AI devices for children used pediatric data to train: 3 notes
Top Decentralized AI Projects Of 2025 Amid OpenAI Copyright Concerns
SCAN Group: It's Time for a Healthtech Moonshot

Share This Article