Interconnects

Nathan Lambert
undefined
19 snips
Feb 12, 2025 • 14min

Deep Research, information vs. insight, and the nature of science

Explore the fascinating intersection of deep research and AI's role in transforming science. The discussion contrasts mere information gathering with the pursuit of genuine insights. Hear about how pioneering tools like AlphaFold are reshaping scientific practices. The podcast emphasizes the need for scientists to adapt to this AI-driven landscape. Finally, it delves into the implications of embracing AI within scientific paradigms, promoting collaboration and innovation in research strategies.
undefined
6 snips
Feb 5, 2025 • 16min

Making the U.S. the home for open-source AI

Explore the evolving landscape of open-source AI and its ideological debates. The discussion highlights the challenges of building a sustainable ecosystem amid competitive pressures from major players, particularly in the U.S. and China. Discover the significance of DeepSeek, which reshapes narratives surrounding open versus closed AI models. Delve into the vision of a future where AI is more accessible, safer, and collaboratively built by a broader community, pushing back against the dominance of super-rich companies.
undefined
27 snips
Jan 28, 2025 • 12min

Why reasoning models will generalize

Explore the fascinating evolution of reasoning models in AI, highlighting their potential to generalize beyond traditional domains like programming and math. Discover how chain of thought reasoning enhances performance, allowing models to manage complexity more effectively. The discussion touches on advancements in training methodologies and the future capabilities expected by 2025. The differences in reasoning between human intelligence and language models provide intriguing insights into how information is processed and stored.
undefined
11 snips
Jan 22, 2025 • 1h 13min

Interviewing OLMo 2 leads: Open secrets of training language models

Luca Soldaini, the Data lead for the OLMo project at AI2, joins the discussion to unveil the intricacies of training language models. He shares tales of overcoming challenges in pretraining efficiency and the quest for stability, especially after a significant 70B model attempt. The conversation dives into the strategic decisions behind building effective language modeling teams, the intricate balance of deep versus wide network architectures, and the importance of community-driven advancements in AI.
undefined
16 snips
Jan 21, 2025 • 20min

DeepSeek R1's recipe to replicate o1 and the future of reasoning LMs

Discover the latest in AI with the launch of a groundbreaking reasoning language model, R1, featuring a unique four-stage reinforcement learning approach. The discussion dives into how this innovation could disrupt the market with competitive pricing and open-source implications. The conversation also touches on advancements in reasoning models and the fine-tuning processes that enhance their capabilities, hinting at exciting developments for researchers and companies alike.
undefined
Jan 15, 2025 • 10min

Let me use my local LMs on Meta Ray-Bans

Exploring the intersection of AI and wearable tech, the conversation highlights how new gadgets like Meta's Ray-Bans are transforming our interaction with technology. The role of local language models emerges as essential for enhanced privacy and efficiency. Insights reveal that while the current devices feel a step behind, their potential impact is likened to the revolutionary iPod and iPad. The excitement surrounding these innovations echoes the initial buzz of AirPods, signaling a shift in how we perceive and utilize AI in daily life.
undefined
10 snips
Jan 9, 2025 • 17min

(Voiceover) DeepSeek V3 and the actual cost of training frontier AI models

Discover the groundbreaking innovations behind DeepSeek V3 and its impressive learning efficiency. The discussion dives into the complex financial aspects of training frontier AI models, shedding light on the true costs involved. Get insights into how these advancements could shape the future of AI development and the importance of transparency in computational resources. It's a fascinating look at technology's evolution and its implications for the industry.
undefined
18 snips
Jan 8, 2025 • 54min

The state of post-training in 2025

Explore the exciting advancements in post-training for language models as experts discuss reinforced learning from human feedback and preference tuning. Gain insights into the complexities of these techniques and the challenges of data acquisition and metric evaluation. The conversation highlights a promising future for open recipes and knowledge in the field by 2025. It's an optimistic take as the scientific community continues to push the boundaries of understanding and effective training methods.
undefined
25 snips
Jan 2, 2025 • 16min

Quick recap on the state of reasoning

The discussion dives into the intriguing intersection of reasoning, inference, and post-training in AI. It challenges the myth that language models lack reasoning capabilities, emphasizing their potential to manipulate tokens to draw conclusions. The speaker highlights advancements in reinforcement learning and how they enhance model performance. Future developments in Reasoning Language Models (RLMs) are also a hot topic, suggesting a shift in understanding AI capabilities is on the horizon.
undefined
12 snips
Dec 31, 2024 • 6min

(Voiceover) 2024 Interconnects year in review

The podcast reflects on the significant milestones in AI for 2024, featuring the launch of OpenAI's O1 model. It delves into the advancements in reinforcement learning and the impact of open-source policies. The discussion highlights how these developments have shaped the AI landscape, offering insights into future trends and growth in the field.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app