Machine Learning Street Talk (MLST)

Making deep learning perform real algorithms with Category Theory (Andrew Dudzik, Petar Velichkovich, Taco Cohen, Bruno Gavranović, Paul Lessard)

72 snips
Dec 22, 2025
This discussion features Andrew Dudzik, a mathematician specializing in category theory; Taco Cohen, a researcher in geometric deep learning; and Petar Veličković, an expert in graph neural networks. They delve into why LLMs struggle with basic math by highlighting their pattern recognition flaws. The conversation proposes category theory as a framework to transition AI from trial-and-error towards a scientific approach. They explore concepts like equivariance, compositional structures, and the potential for unifying diverse machine learning perspectives.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

LLMs Don’t Truly Perform Addition

  • Large language models learn patterns, not algorithmic addition, so simple digit changes break them.
  • Andrew Dudzik shows LLMs fail carrying operations when a single digit disrupts learned patterns.
ADVICE

Internalize Basic Computation

  • Equip models with internal computation instead of relying solely on external tools to avoid repeated calls.
  • Andrew Dudzik argues internalizing basic computation yields efficiency and stability over tool chaining.
INSIGHT

Equivariance Cuts Data Needs

  • Geometric deep learning uses equivariance to exploit symmetries like translations and permutations.
  • Taco Cohen explains this reduces required data massively and underpins transformers' token permutation invariance.
Get the Snipd Podcast app to discover more snips from this episode
Get the app