Machine Learning Street Talk (MLST)

Prof. Melanie Mitchell 2.0 - AI Benchmarks are Broken!

15 snips
Sep 10, 2023
Prof. Melanie Mitchell, Davis Professor of Complexity at the Santa Fe Institute, dives into the murky waters of AI understanding. She argues that current benchmarks are inadequate, as machines often replicate human tasks without true comprehension. Mitchell highlights the limitations of large language models, noting their lack of common sense despite impressive statistical capabilities. She emphasizes the need for evolving evaluation methods and suggests a deeper, context-specific look at intelligence, advocating for more rigorous testing to reflect genuine understanding.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

Nature of Intelligence

  • Intelligence is ill-defined and multidimensional, not a simple yes/no.
  • AI systems might possess intelligence in specific ways or degrees, rather than absolute intelligence.
ANECDOTE

Grounded Understanding

  • Dilip George argues a university professor understands vectors better than AI.
  • This is because their knowledge is grounded in real-world situations, not just abstract concepts.
INSIGHT

Understanding in LLMs

  • Key question: Is AI understanding a category error, mistaking token associations for real-world connections?
  • Or do large language models create concept-based mental models like humans, and does scaling improve them?
Get the Snipd Podcast app to discover more snips from this episode
Get the app