#44285
Mentioned in 1 episodes

Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention

Book • 2024
This paper presents a novel approach to scaling Transformer-based large language models to handle infinitely long inputs with bounded memory and computation.

The Infini-attention mechanism incorporates compressive memory into the vanilla attention mechanism, enabling efficient processing of long contexts.

The approach is demonstrated on various language modeling benchmarks.

Mentioned by

Mentioned in 1 episodes

Mentioned as a Google paper on efficient infinite context transformers.
51 snips
#162 - Udio Song AI, TPU v5, Mixtral 8x22, Mixture-of-Depths, Musicians sign open letter

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app