Meta and Stanford Researchers Propose Fast Byte Latent Transformer That Reduces Inference Memory Bandwidth by Over 50% Without Tokenization
A workforce of researchers from Meta, Stanford University, and the University of Washington have launched three new strategies that considerably speed up era within the Byte Latent Transformer (BLT) — a language mannequin structure that operates straight on uncooked bytes as an alternative of tokens.
Byte-Level Models Are Slow at Inference
To perceive what this new analysis solves, it is advisable to perceive the tradeoff on the heart of byte-level language modeling.
Most language fashions at present work on tokens — chunks of textual content produced by subword tokenizers like byte-pair encoding (BPE). A token usually represents a number of characters or perhaps a entire phrase. While that is environment friendly, tokenization comes with recognized downsides: sensitivity to enter noise, poor dealing with of multilingual textual content, weak character-level understanding, and fragility on structured inputs like code and numbers.
Byte-level fashions sidestep all of this by working straight on uncooked bytes — the lowest-level illustration of textual content. The Byte Latent Transformer (BLT) was a significant step ahead: it matched the efficiency of tokenization-based fashions at scale by grouping bytes dynamically into variable-length patches utilizing an entropy-based segmentation technique. High-entropy (harder-to-predict) areas get shorter patches; extra predictable spans get longer ones. The bulk of computation runs over latent token representations, not uncooked bytes — utilizing three elements: an area encoder, a big international Transformer, and an area decoder — with a mean patch measurement of 4 bytes and a most of 8.
The remaining drawback is inference velocity. Even with BLT’s hierarchical design, the native decoder nonetheless generates one byte at a time autoregressively. Since a typical subword token corresponds to a number of bytes, BLT wants a number of decoder ahead passes to supply the identical quantity of textual content {that a} token-level mannequin produces in a single step. In trendy LLM serving, the bottleneck is commonly not compute however reminiscence bandwidth — repeatedly loading mannequin weights and key-value caches from reminiscence. More decoder ahead passes means extra reminiscence hundreds, which straight interprets to slower era.

Three Methods, One Goal: Fewer Forward Passes
The analysis workforce introduces three strategies that scale back this bottleneck, every buying and selling velocity towards era high quality otherwise.
BLT Diffusion (BLT-D)
It is the core contribution and the quickest variant. The key thought is to interchange autoregressive byte-by-byte decoding with block-wise discrete diffusion within the native decoder.
During coaching, the decoder receives two inputs: a clear byte sequence (the unique textual content) and a corrupted sequence of fixed-length byte blocks. For every block, a steady diffusion timestep t is sampled from U(0,1), and every byte within the block is independently changed with a [MASK] token with chance t. This means the diploma of masking varies per coaching instance — a decrease t leaves most bytes seen; the next t masks most of them. The block measurement B (set to 4, 8, or 16 bytes in experiments) usually extends past BLT’s common patch measurement of 4 bytes, instructing the decoder to foretell bytes additional into the longer term than it usually would. The complete coaching loss combines the usual autoregressive next-byte prediction loss on the clear sequence and a masked-byte prediction loss on the corrupted blocks — conceptually much like how masked language modeling in BERT works, however utilized on the byte stage inside BLT’s hierarchical structure.
At inference, BLT-D initializes a block of [MASK] positions and iteratively unmasks a number of byte positions per decoder step utilizing certainly one of two methods: confidence-based unmasking (unmask positions whose predicted chance exceeds a threshold α) or entropy-bounded (EB) sampling (choose the biggest subset of positions whose cumulative entropy stays under a threshold γ). Both methods generate a number of bytes per ahead move moderately than one. The encoder and international mannequin — BLT’s costly elements — are invoked as soon as per block moderately than as soon as per patch, additional lowering complete mannequin calls. BLT-D additionally helps KV caching, benefiting from any strategies that scale back KV-cache reminiscence footprint.
At 3B parameters, BLT-D-4 (block measurement 4) practically matches BLT’s process scores whereas requiring lower than half the reminiscence bandwidth. BLT-D-16 (block measurement 16) achieves an 87–92% discount in estimated memory-bandwidth price in comparison with BLT, making it the quickest configuration evaluated — although with decrease move@1 scores on coding benchmarks (HumanEval, MBPP).
BLT Self-Speculation (BLT-S)
It takes a unique route, drawing on speculative decoding — a method the place an affordable draft mannequin proposes tokens and a bigger mannequin verifies them in parallel. What makes BLT-S uncommon is that it requires no separate draft mannequin and no architectural adjustments or extra coaching. It repurposes BLT’s present light-weight native decoder because the drafter.
In normal BLT inference, the decoder stops producing every time the entropy-based patcher determines {that a} new patch boundary has been reached — usually each 4 bytes. BLT-S as an alternative lets the decoder autoregressively generate as much as a set window measurement ok (8 or 16 bytes in experiments) no matter entropy spikes, conditioning on the final accessible latent token. After producing a draft of ok bytes, the complete mannequin re-encodes the candidate sequence by means of the encoder, international mannequin, and decoder and produces next-byte predictions. Drafted bytes are accepted as much as the primary mismatch; the primary mismatched byte is changed with the verified prediction.
Under grasping decoding, this process ensures that verified outputs are an identical to straightforward autoregressive BLT decoding — no high quality loss. BLT-S will increase decoder ahead passes barely however considerably reduces encoder and international mannequin calls. At 3B parameters with ok=16, BLT-S might obtain as much as 77% memory-bandwidth discount with no loss in process efficiency.
BLT Diffusion+Verification (BLT-DV)
It sits within the center. Because BLT-D is skilled with each a diffusion goal and a normal next-byte prediction goal, the identical mannequin weights can run autoregressively utilizing causal decoder masks — no separate mannequin and no extra coaching wanted. BLT-DV exploits this: diffusion drafts a block of bytes first, then a single autoregressive ahead move verifies the draft, accepting bytes as much as the primary mismatch. Empirically, one-step diffusion mixed with verification yielded the quickest BLT-DV configuration. While one-step diffusion alone usually results in fast degradation in era high quality, the verification step successfully prevents this. At 3B parameters, BLT-DV might obtain as much as 81% memory-bandwidth discount in comparison with BLT.
Understanding the Numbers
All fashions have been skilled on the BLT-1T dataset (1 trillion tokens from public sources together with a subset of Datacomp-LM), with 1B-parameter fashions skilled for 240,000 steps and 3B-parameter fashions for 480,000 steps. Evaluation lined 4 era duties: French-to-English and German-to-English translation utilizing the FLORES-101 benchmark (4-shot, SentencePiece BLEU) and two coding benchmarks — HumanEval (0-shot, move@1) and MBPP (3-shot, move@1).
Beyond era duties, the analysis workforce additionally evaluates BLT-D on 5 likelihood-based benchmarks: ARC-Easy, ARC-Challenge, PIQA, HellaSwag, and MMLU. Since BLT-D is skilled with a next-byte prediction goal alongside the diffusion goal, it might compute autoregressive likelihoods by making use of a causal masks to the decoder — the identical mechanism BLT-DV’s verification step depends on. The outcomes present BLT-D variants obtain scores approaching BLT’s baseline on all 5 benchmarks, confirming that integrating block diffusion doesn’t compromise the mannequin’s autoregressive reasoning functionality.
Efficiency is reported by way of three proxy metrics: decoder community perform evaluations (NFEs), encoder/international mannequin NFEs, and an estimated memory-bandwidth determine in gigabytes derived from parameter counts and forward-pass counts underneath 16-bit precision. The analysis workforce is specific that these are proxy metrics — changing NFE reductions into precise wall-clock enhancements requires a extremely optimized inference implementation, which the analysis workforce flags as a very powerful path for future work.
Translation duties profit most from BLT-D throughout all block sizes. Coding duties present extra sensitivity to dam measurement: BLT-D-16 gives the biggest effectivity good points however exhibits significant rating drops on HumanEval and MBPP. A notable extra discovering comes from the era range evaluation: when utilizing entropy-bounded sampling with top-p sampling at inference, extra decoder NFEs correlate with larger type-token ratio (a measure of lexical range). This means the effectivity–range tradeoff is tunable at inference time with none retraining.

Key Takeaways
- BLT-D introduces block-wise discrete diffusion into BLT’s native decoder, coaching with a mixed next-byte prediction and masked-byte prediction loss to generate a number of bytes per ahead move as an alternative of separately
- BLT-S makes use of BLT’s personal light-weight decoder as a speculative drafter — no separate mannequin, no architectural adjustments, no extra coaching — and produces output an identical to straightforward BLT underneath grasping decoding
- BLT-DV combines diffusion drafting with an autoregressive verification step utilizing the identical BLT-D mannequin weights, recovering high quality misplaced in diffusion-only decoding with out further coaching
- All strategies might obtain an estimated memory-bandwidth price over 50% decrease than BLT on era duties; BLT-D-16 might attain 87–92% discount
- BLT-D’s autoregressive functionality stays strong on likelihood-based benchmarks (ARC-Easy, ARC-Challenge, PIQA, HellaSwag, MMLU), and its era range is tunable at inference time by way of entropy-bounded sampling thresholds
Check out the Paper. Also, be at liberty to observe us on Twitter and don’t overlook to affix our 150k+ ML SubReddit and Subscribe to our Newsletter. Wait! are you on telegram? now you can join us on telegram as well.
Need to associate with us for selling your GitHub Repo OR Hugging Face Page OR Product Release OR Webinar and many others.? Connect with us
The put up Meta and Stanford Researchers Propose Fast Byte Latent Transformer That Reduces Inference Memory Bandwidth by Over 50% Without Tokenization appeared first on MarkTechPost.
