Deep Papers cover image

Lost in the Middle: How Language Models Use Long Contexts

Deep Papers

00:00

How to Retrieve Context From Your Input

Mummels are relatively robust, but when you go above their trained token maximum, they start to struggle. Performance on multi-document question is actually lower than its performance with predicting without any documents. There's going to be this trade-off between adding more information and increasing the complexity of these models.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app