Machine Learning Street Talk (MLST) cover image

Machine Learning Street Talk (MLST)

Latest episodes

undefined
16 snips
Mar 8, 2025 • 1h 1min

Transformers Need Glasses! - Federico Barbero

Federico Barbero, a lead author at DeepMind/Oxford, dives into the quirks of transformers and why large language models falter at tasks like counting. He reveals fascinating architectural bottlenecks that affect their performance. By drawing parallels with graph neural networks, he sheds light on the softmax function's role in limiting decision-making clarity. But not all hope is lost! Federico shares innovative 'glasses' to enhance transformer performance, including input tweaks and structural modifications to boost their clarity and efficiency.
undefined
91 snips
Mar 1, 2025 • 1h 38min

Sakana AI - Chris Lu, Robert Tjarko Lange, Cong Lu

Chris Lu, a recent Oxford DPhil graduate specializing in meta-learning, and Robert Tjarko Lange, a TU Berlin PhD candidate focused on evolutionary algorithms, join forces to discuss innovative approaches to AI. They explore how language models can automate algorithm discovery and enhance training processes. The conversation dives into the interplay of human creativity and AI, addressing challenges like infinite regress in loss functions and the implications of evolutionary optimization. Together, they envision a future where AI systems co-create alongside researchers.
undefined
39 snips
Feb 19, 2025 • 51min

Clement Bonnet - Can Latent Program Networks Solve Abstract Reasoning?

Clement Bonnet, a researcher specializing in abstract reasoning, shares his cutting-edge approach to the ARC challenge using latent program networks. He contrasts his method of embedding programs in latent spaces with traditional neural networks, highlighting their struggles with tasks requiring genuine understanding. The discussion dives into the importance of induction versus transduction in machine learning, explores innovative training techniques, and examines the creative limitations of large language models, advocating for a balance between human cognition and AI capabilities.
undefined
82 snips
Feb 18, 2025 • 54min

Prof. Jakob Foerster - ImageNet Moment for Reinforcement Learning?

Jakob Foerster, a prominent AI researcher at Oxford University and Meta, joins to discuss the future of AI. He emphasizes the shift from mimicking human behavior to developing intelligent agents that can learn independently. The conversation delves into the importance of open-source AI for responsible innovation and addresses challenges such as AI scaling and goal misalignment. They also explore advancements in deep reinforcement learning, the significance of creativity, and the need for democratization in AI to foster collaboration and mitigate risks.
undefined
44 snips
Feb 12, 2025 • 1h 9min

Daniel Franzen & Jan Disselhoff - ARC Prize 2024 winners

Daniel Franzen and Jan Disselhoff, the winners of the ARC Prize 2024, dive into their innovative approaches with large language models. They discuss achieving a surprising 53.5% accuracy using novel techniques like depth-first search for token selection and test-time training. Their insights into model training complexities, ethical considerations, and the balance between performance and accuracy provide a fascinating look at cutting-edge AI research. Additionally, they share the importance of rapid innovation under competitive pressures and the challenges faced in algorithm development.
undefined
38 snips
Feb 12, 2025 • 1h 7min

Sepp Hochreiter - LSTM: The Comeback Story?

Sepp Hochreiter, the mastermind behind LSTM networks and founder of NXAI, shares insights from his journey in AI. He discusses the potential of XLSTM for robotics and industrial simulation. Hochreiter critiques Large Language Models' shortcomings in true reasoning and creativity. He emphasizes the need for hybrid approaches that integrate symbolic reasoning with neural networks. His reflections on the evolution of neural architectures reveal the exciting advancements in memory management and processing efficiency, hinting at a transformative future for AI.
undefined
113 snips
Feb 8, 2025 • 1h 18min

Want to Understand Neural Networks? Think Elastic Origami! - Prof. Randall Balestriero

Professor Randall Balestriero, an expert in machine learning, dives deep into neural network geometry and spline theory. He introduces the captivating concept of 'grokking', explaining how prolonged training can enhance adversarial robustness. The discussion also highlights the significance of representing data through splines to improve model design and performance. Additionally, Balestriero explores the geometric implications for large language models in toxicity detection, and delves into the challenges of reconstruction learning and the intricacies of representation in neural networks.
undefined
120 snips
Jan 25, 2025 • 1h 21min

Nicholas Carlini (Google DeepMind)

Nicholas Carlini, a research scientist at Google DeepMind specializing in AI security, delves into compelling insights about the vulnerabilities in machine learning systems. He discusses the unexpected chess-playing prowess of large language models and the broader implications of emergent behaviors. Carlini emphasizes the necessity for robust security designs to combat potential model attacks and the ethical considerations surrounding AI-generated code. He also highlights how language models can significantly enhance programming productivity, urging users to remain skeptical of their limitations.
undefined
28 snips
Jan 23, 2025 • 1h 32min

Subbarao Kambhampati - Do o1 models search?

In this engaging discussion, Professor Subbarao Kambhampati, an expert in AI reasoning systems, dives into OpenAI's O1 model. He explains how it employs reinforcement learning akin to AlphaGo and introduces the concept of 'fractal intelligence,' where models exhibit unpredictable performance. The conversation contrasts single-model approaches with hybrid systems like Google’s, and addresses the balance between AI as an intelligence amplifier versus an autonomous decision-maker, shedding light on the computational costs associated with advanced reasoning systems.
undefined
180 snips
Jan 20, 2025 • 1h 18min

How Do AI Models Actually Think? - Laura Ruis

Laura Ruis, a PhD student at University College London and researcher at Cohere, discusses her groundbreaking work on reasoning capabilities of large language models. She delves into whether these models rely on fact retrieval or procedural knowledge. The conversation highlights the influence of pre-training data on AI behavior and examines the complexities in defining intelligence. Ruis also explores the philosophical implications of AI agency and creativity, raising questions about how AI models mimic human reasoning and the potential risks they pose.

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode