Machine Learning Street Talk (MLST)

How Do AI Models Actually Think? - Laura Ruis

218 snips
Jan 20, 2025
Laura Ruis, a PhD student at University College London and researcher at Cohere, discusses her groundbreaking work on reasoning capabilities of large language models. She delves into whether these models rely on fact retrieval or procedural knowledge. The conversation highlights the influence of pre-training data on AI behavior and examines the complexities in defining intelligence. Ruis also explores the philosophical implications of AI agency and creativity, raising questions about how AI models mimic human reasoning and the potential risks they pose.
Ask episode
AI Snips
Chapters
Books
Transcript
Episode notes
INSIGHT

Reasoning vs. Retrieval

  • LLMs seem to perform reasoning by synthesizing knowledge from many documents.
  • This differs from fact retrieval, which relies on specific documents.
INSIGHT

Influence Functions

  • Influence functions reveal how pre-training data affects LLM reasoning steps.
  • They estimate how model parameters change if specific data points are removed during training.
INSIGHT

Code's Influence

  • Code heavily influences LLM reasoning processes, both positively and negatively.
  • This is surprising, as code differs from natural language data.
Get the Snipd Podcast app to discover more snips from this episode
Get the app