

Interconnects
Nathan Lambert
Audio essays about the latest developments in AI and interviews with leading scientists in the field. Breaking the hype, understanding what's under the hood, and telling stories. www.interconnects.ai
Episodes
Mentioned books

19 snips
Feb 12, 2025 • 14min
Deep Research, information vs. insight, and the nature of science
Explore the fascinating intersection of deep research and AI's role in transforming science. The discussion contrasts mere information gathering with the pursuit of genuine insights. Hear about how pioneering tools like AlphaFold are reshaping scientific practices. The podcast emphasizes the need for scientists to adapt to this AI-driven landscape. Finally, it delves into the implications of embracing AI within scientific paradigms, promoting collaboration and innovation in research strategies.

6 snips
Feb 5, 2025 • 16min
Making the U.S. the home for open-source AI
Explore the evolving landscape of open-source AI and its ideological debates. The discussion highlights the challenges of building a sustainable ecosystem amid competitive pressures from major players, particularly in the U.S. and China. Discover the significance of DeepSeek, which reshapes narratives surrounding open versus closed AI models. Delve into the vision of a future where AI is more accessible, safer, and collaboratively built by a broader community, pushing back against the dominance of super-rich companies.

27 snips
Jan 28, 2025 • 12min
Why reasoning models will generalize
Explore the fascinating evolution of reasoning models in AI, highlighting their potential to generalize beyond traditional domains like programming and math. Discover how chain of thought reasoning enhances performance, allowing models to manage complexity more effectively. The discussion touches on advancements in training methodologies and the future capabilities expected by 2025. The differences in reasoning between human intelligence and language models provide intriguing insights into how information is processed and stored.

11 snips
Jan 22, 2025 • 1h 13min
Interviewing OLMo 2 leads: Open secrets of training language models
Luca Soldaini, the Data lead for the OLMo project at AI2, joins the discussion to unveil the intricacies of training language models. He shares tales of overcoming challenges in pretraining efficiency and the quest for stability, especially after a significant 70B model attempt. The conversation dives into the strategic decisions behind building effective language modeling teams, the intricate balance of deep versus wide network architectures, and the importance of community-driven advancements in AI.

16 snips
Jan 21, 2025 • 20min
DeepSeek R1's recipe to replicate o1 and the future of reasoning LMs
Discover the latest in AI with the launch of a groundbreaking reasoning language model, R1, featuring a unique four-stage reinforcement learning approach. The discussion dives into how this innovation could disrupt the market with competitive pricing and open-source implications. The conversation also touches on advancements in reasoning models and the fine-tuning processes that enhance their capabilities, hinting at exciting developments for researchers and companies alike.

Jan 15, 2025 • 10min
Let me use my local LMs on Meta Ray-Bans
Exploring the intersection of AI and wearable tech, the conversation highlights how new gadgets like Meta's Ray-Bans are transforming our interaction with technology. The role of local language models emerges as essential for enhanced privacy and efficiency. Insights reveal that while the current devices feel a step behind, their potential impact is likened to the revolutionary iPod and iPad. The excitement surrounding these innovations echoes the initial buzz of AirPods, signaling a shift in how we perceive and utilize AI in daily life.

10 snips
Jan 9, 2025 • 17min
(Voiceover) DeepSeek V3 and the actual cost of training frontier AI models
Discover the groundbreaking innovations behind DeepSeek V3 and its impressive learning efficiency. The discussion dives into the complex financial aspects of training frontier AI models, shedding light on the true costs involved. Get insights into how these advancements could shape the future of AI development and the importance of transparency in computational resources. It's a fascinating look at technology's evolution and its implications for the industry.

18 snips
Jan 8, 2025 • 54min
The state of post-training in 2025
Explore the exciting advancements in post-training for language models as experts discuss reinforced learning from human feedback and preference tuning. Gain insights into the complexities of these techniques and the challenges of data acquisition and metric evaluation. The conversation highlights a promising future for open recipes and knowledge in the field by 2025. It's an optimistic take as the scientific community continues to push the boundaries of understanding and effective training methods.

25 snips
Jan 2, 2025 • 16min
Quick recap on the state of reasoning
The discussion dives into the intriguing intersection of reasoning, inference, and post-training in AI. It challenges the myth that language models lack reasoning capabilities, emphasizing their potential to manipulate tokens to draw conclusions. The speaker highlights advancements in reinforcement learning and how they enhance model performance. Future developments in Reasoning Language Models (RLMs) are also a hot topic, suggesting a shift in understanding AI capabilities is on the horizon.

12 snips
Dec 31, 2024 • 6min
(Voiceover) 2024 Interconnects year in review
The podcast reflects on the significant milestones in AI for 2024, featuring the launch of OpenAI's O1 model. It delves into the advancements in reinforcement learning and the impact of open-source policies. The discussion highlights how these developments have shaped the AI landscape, offering insights into future trends and growth in the field.