Machine Learning Street Talk (MLST)

Reasoning, Robustness, and Human Feedback in AI - Max Bartolo (Cohere)

348 snips
Mar 18, 2025
Max Bartolo, a researcher at Cohere, dives into the world of machine learning, focusing on model reasoning and robustness. He highlights the DynaBench platform's role in dynamic benchmarking and the complex challenges of evaluating AI performance. The conversation reveals the limitations of human feedback in training AI and the surprising reliance on distributed knowledge. Bartolo discusses the impact of adversarial examples on model reliability and emphasizes the need for tailored approaches to enhance AI systems, ensuring they align with human values.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

Distributed Reasoning vs. Factual Retrieval

  • Models may rely on distributed procedural knowledge for reasoning tasks, unlike factual retrieval.
  • Influence functions reveal this by showing broader influence patterns across training documents.
ANECDOTE

Reasoning vs. Control Questions

  • Questions like "What is the slope of a line defined by points (2,2) and (3,3)?" require reasoning.
  • Control questions, like "The slope is 1, what is the slope?", don't require reasoning, highlighting the difference.
INSIGHT

Reasoning Implies Robustness

  • Reasoning and robustness are interconnected; true reasoning implies consistent results.
  • Inconsistent performance questions whether a model truly reasons or just matches patterns.
Get the Snipd Podcast app to discover more snips from this episode
Get the app