The Nonlinear Library

AF - Your LLM Judge may be biased by Rachel Freedman

Mar 29, 2024
Rachel Freedman, an AI safety researcher, discusses the bias present in LLM judges used by researchers. She details experiments and strategies to mitigate biases, including adjusting labeling systems and validating judgments against human ones. The podcast covers analyzing biases in the LAMA2 model and exploring techniques like few-shot prompting to reduce biases in language models.
Ask episode
Chapters
Transcript
Episode notes