

#44285
Mentioned in 1 episodes
Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention
Book • 2024
This paper presents a novel approach to scaling Transformer-based large language models to handle infinitely long inputs with bounded memory and computation.
The Infini-attention mechanism incorporates compressive memory into the vanilla attention mechanism, enabling efficient processing of long contexts.
The approach is demonstrated on various language modeling benchmarks.
The Infini-attention mechanism incorporates compressive memory into the vanilla attention mechanism, enabling efficient processing of long contexts.
The approach is demonstrated on various language modeling benchmarks.
Mentioned by
Mentioned in 1 episodes
Mentioned as a Google paper on efficient infinite context transformers.

51 snips
#162 - Udio Song AI, TPU v5, Mixtral 8x22, Mixture-of-Depths, Musicians sign open letter