
Machine Learning Street Talk (MLST)
Welcome! We engage in fascinating discussions with pre-eminent figures in the AI field. Our flagship show covers current affairs in AI, cognitive science, neuroscience and philosophy of mind with in-depth analysis. Our approach is unrivalled in terms of scope and rigour – we believe in intellectual diversity in AI, and we touch on all of the main ideas in the field with the hype surgically removed. MLST is run by Tim Scarfe, Ph.D (https://www.linkedin.com/in/ecsquizor/) and features regular appearances from MIT Doctor of Philosophy Keith Duggar (https://www.linkedin.com/in/dr-keith-duggar/).
Latest episodes

4 snips
Jun 6, 2020 • 1h 52min
OpenAI GPT-3: Language Models are Few-Shot Learners
Yannic Kilcher, a YouTube AI savant, and Connor Shorten, a machine learning contributor, dive into the revolutionary GPT-3 language model. They discuss its jaw-dropping 175 billion parameters and how it performs various NLP tasks with zero fine-tuning. The duo unpacks the differences between autoregressive models like GPT-3 and BERT, as well as the complexities of reasoning versus memorization in language models. Additionally, they tackle the implications of AI bias, the significance of transformer architecture, and the future of generative AI.

Jun 3, 2020 • 1h 13min
Jordan Edwards: ML Engineering and DevOps on AzureML
Jordan Edwards, Principal Program Manager for AzureML at Microsoft, dives into the world of ML DevOps and the challenges of deploying machine learning models. He discusses how to bridge the gap between science and engineering, emphasizing model governance and testing. Jordan shares insights from the recent Microsoft Build conference, highlighting innovations like FairLearn and GPT-3. He also introduces his maturity model for ML DevOps and explores the complexities of collaboration in machine learning workflows, making for a thought-provoking conversation.

13 snips
Jun 2, 2020 • 2h 29min
One Shot and Metric Learning - Quadruplet Loss (Machine Learning Dojo)
Join Eric Craeymeersch, a software engineer and innovation director with a focus on machine learning and computer vision, as he dives into the fascinating world of metric learning and one shot learning. Discover the revolutionary shift toward quadruplet loss over triplet loss and its implications for more efficient clustering and classification. Eric discusses the intricacies of Siamese networks, hard mining strategies, and the importance of experimentation in data science, sharing valuable insights that can propel your understanding of cutting-edge machine learning techniques.

May 25, 2020 • 1h 38min
Harri Valpola: System 2 AI and Planning in Model-Based Reinforcement Learning
Harri Valpola, the CEO and Founder of Curious AI, specializes in optimizing industrial processes through advanced AI. In this discussion, he dives into the fascinating world of System 1 and System 2 thinking in AI, illustrating the balance between instinctive and reflective reasoning. Valpola shares insights from his recent research on model-based reinforcement learning, emphasizing the challenges of real-world applications like water treatment. He also highlights innovative approaches using denoising autoencoders to improve planning in uncertain environments.

May 22, 2020 • 2h 34min
ICLR 2020: Yoshua Bengio and the Nature of Consciousness
Yoshua Bengio, a pioneer in deep learning and Professor at the University of Montreal, dives into the intriguing intersection of AI and consciousness. He discusses the role of attention in conscious processing and explores System 1 and System 2 thinking as outlined by Daniel Kahneman. Bengio raises profound questions about the nature of intelligence and self-awareness in machines. He also addresses the implications of sparse factor graphs and the philosophical dimensions of consciousness, offering fresh insights into how these concepts can enhance AI models.

32 snips
May 19, 2020 • 2h 12min
ICLR 2020: Yann LeCun and Energy-Based Models
Yann LeCun, a pioneer in machine learning and AI, discusses the latest in self-supervised learning and energy-based models (EBMs). He compares how humans and machines learn concepts, advocating for methods that mimic human cognition. The conversation dives into EBMs' applications in optimizing labels and addresses challenges in traditional models. LeCun also explores the potential of self-supervised learning techniques for enhancing AI capabilities, such as in natural language processing and image recognition.

May 19, 2020 • 1h 27min
The Lottery Ticket Hypothesis with Jonathan Frankle
Jonathan Frankle, author of The Lottery Ticket Hypothesis, shares his insights on Sparse Neural Networks and their pruning techniques. He delves into the implications of the lottery ticket hypothesis for improving neural network efficiency and discusses innovative strategies like linear mode connectivity. Frankle also explores the intersection of AI technology and policy, emphasizing the importance of ethical decision-making in AI development. Listeners will appreciate his journey in deep learning research and the challenges faced in academia.

May 19, 2020 • 1h 40min
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
The conversation dives into the fascinating world of Large-scale Transfer Learning in NLP. Key highlights include the innovative T5 model's impact and the importance of dataset size and fine-tuning strategies. The trio also explores embodied cognition and meta-learning, pondering the very nature of intelligence. They discuss the evolution of transformers and the intricacies of training paradigms, all while navigating the challenges of benchmarking and chatbot systems. This lively discussion is packed with insights into advancing AI technologies and their real-world applications.

May 2, 2020 • 1h 15min
CURL: Contrastive Unsupervised Representations for Reinforcement Learning
Aravind Srinivas, a technical staff member at OpenAI and PhD candidate at Berkeley, dives deep into the revolutionary CURL paper he co-authored. This approach leverages contrastive unsupervised learning to enhance data efficiency in reinforcement learning, nearly matching performance with traditional methods. The conversation covers the pivotal role of pixel inputs for robotic control, challenges in sample efficiency, and the evolving dynamics between unsupervised and supervised learning. Srinivas' insights shed light on the future of machine learning.

15 snips
Apr 24, 2020 • 1h 13min
Exploring Open-Ended Algorithms: POET
Mathew Salvaris is a research scientist specializing in computer vision. He dives into the revolutionary concept of open-ended algorithms, likening their evolution to natural selection. These AI-generating algorithms autonomously create their own learning pathways, presenting increasingly complex challenges. The conversation explores how these algorithms can lead to innovative solutions beyond traditional methods, fostering adaptability and improved performance. Excitingly, Salvaris also discusses the potential implications for future AI development and the collaborative relationship between humans and machines.