Neural Search Talks — Zeta Alpha cover image

Transformer Memory as a Differentiable Search Index: memorizing thousands of random doc ids works!?

Neural Search Talks — Zeta Alpha

00:00

Using Direct Indexing to Get the First 32 Tokens of a Document

The simplest thing to do here is just to take the first, they call them L tokens. So you have a document, you say, I'll take the first 32 tokens. This is the common setting. And then you want to store these 32 tokens in the index in a way that associates them with the right doc ID. But that's, they call direct indexing. The other approach kind of feels like trying to go beyond this issue of only looking at the first 32 token. They come up with different chunks of the document and try to store them. That sounds like a pretty short context in modern transformers and stuff. Do you have any idea why they would only do that?

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app