Deep Papers

Lost in the Middle: How Language Models Use Long Contexts

Jul 26, 2023
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Performance Drops In The Middle

  • Language models use long contexts unevenly, favoring the beginning and end of inputs.
  • Performance drops significantly when relevant information sits in the middle of long contexts.
INSIGHT

More Context Isn't Always Better

  • Increasing input context length (more documents or history) does not guarantee better outputs.
  • As context length grows, model accuracy steadily declines, especially for middle-positioned information.
ADVICE

Return Fewer, More Relevant Documents

  • When using retrieval for QA, prioritize returning fewer documents and ensure the single relevant document appears near the start or end.
  • Limit retrieved documents to reduce noise and raise the chance the model finds the correct answer.
Get the Snipd Podcast app to discover more snips from this episode
Get the app