AI Safety Fundamentals: Alignment

Challenges in Evaluating AI Systems

Apr 7, 2024
Exploring challenges in evaluating AI systems, the podcast delves into limitations of current evaluation suites and offers policy recommendations. Topics include pitfalls of using MM LU metric, difficulties in measuring social biases, hurdles in Bias Benchmark Task, complexities of Big Bench framework, and methodologies for red teaming in security evaluations.
Ask episode
Chapters
Transcript
Episode notes