

#35044
Mentioned in 1 episodes
Task Contamination: Language Models May Not Be Few-Shot Anymore
Book • 2024
This paper investigates the impact of task contamination on the performance of large language models (LLMs) in zero-shot and few-shot learning scenarios.
It highlights that LLMs often perform better on datasets released before their training data creation date, indicating task contamination.
The study employs methods like training data inspection and membership inference to detect contamination.
It highlights that LLMs often perform better on datasets released before their training data creation date, indicating task contamination.
The study employs methods like training data inspection and membership inference to detect contamination.
Mentioned by
Mentioned in 1 episodes
Mentioned in a discussion about language model evaluation and task contamination.

#149 - Reflecting on 2023, Midjourney v6, Anthropic Revenue, Unified-IO 2, NY Times sues OpenAI