Machine Learning Street Talk (MLST) cover image

Machine Learning Street Talk (MLST)

Reasoning, Robustness, and Human Feedback in AI - Max Bartolo (Cohere)

Mar 18, 2025
Max Bartolo, a researcher at Cohere, dives into the world of machine learning, focusing on model reasoning and robustness. He highlights the DynaBench platform's role in dynamic benchmarking and the complex challenges of evaluating AI performance. The conversation reveals the limitations of human feedback in training AI and the surprising reliance on distributed knowledge. Bartolo discusses the impact of adversarial examples on model reliability and emphasizes the need for tailored approaches to enhance AI systems, ensuring they align with human values.
01:23:11

Episode guests

Podcast summary created with Snipd AI

Quick takeaways

  • AI models must demonstrate unwavering consistency and reliability across similar tasks to foster trust and effectiveness in reasoning.
  • Human feedback significantly influences AI model training, yet reliance solely on it can compromise factual accuracy and model performance.

Deep dives

The Consistency and Robustness of AI Models

Consistency is a crucial expectation for AI models, paralleling the reliability one would demand from basic tools like calculators. If a model, similar to a calculator, produces inconsistent or incorrect answers when presented with variations of a question, it undermines trust in its functionality. For AI to be effective in reasoning tasks, it must demonstrate unwavering reliability across similar queries, indicating a more profound understanding rather than merely matching recognized patterns. This insistence on consistency sets a higher standard for AI performance, especially on simple tasks that should be effortlessly executed.

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner