Interconnects

Nathan Lambert
undefined
Sep 4, 2024 • 11min

OLMoE and the hidden simplicity in training better foundation models

Dive into the innovations behind OLMoE, a cutting-edge language model that excels among its peers. Explore the challenges of training complexity and organizational hurdles. Discover the secret sauce of compounding improvements that leads to better models. This conversation unpacks not just the tech, but the strategic thinking driving advancements in AI.
undefined
Aug 28, 2024 • 8min

On the current definitions of open-source AI and the state of the data commons

The discussion dives deep into the evolving definitions of open-source AI. It highlights the challenges faced by the data commons and the necessity for better documentation. Concerns about the implications of mandating fully released data are raised. Frustration with existing definitions is palpable, as examples are urgently needed to clarify the landscape. The dialogue emphasizes the balance between accessibility and regulation in the AI realm.
undefined
Aug 16, 2024 • 9min

Nous Hermes 3 and exploiting underspecified evaluations

The discussion kicks off with the launch of a new model, questioning what defines a 'frontier model.' Notable comparisons are drawn with LAMA 3.1 and the importance of transparent evaluation metrics emerges. The conversation elaborates on valuable lessons learned from the training process of Hermes 3. The broader implications for technology policy are also highlighted, emphasizing the need for integrity in AI evaluations.
undefined
Aug 8, 2024 • 1h 2min

Interviewing Ross Taylor on LLM reasoning, Llama fine-tuning, Galactica, agents

Ross Taylor, a former LLM lead at Meta AI and co-founder of Papers with Code, dives into the cutting-edge world of language models. He shares insights on the Galactica project, its ambitions, and the ethical complexities involved. The conversation explores the potential of language models in scientific discovery, the evolution of reasoning within AI, and innovative training methodologies. Taylor emphasizes the significance of collaboration in advancing AI research while highlighting the latest developments in model alignment and user experience.
undefined
Aug 7, 2024 • 10min

A recipe for frontier model post-training

The discussion dives into the latest advancements in reinforcement learning from human feedback, focusing on the Llama 3.1 model. Key players like Apple, Meta, and Nvidia emphasize the importance of synthetic data and iterative training. Data quality emerges as a pivotal theme, with agreements on new standards in model training. The episode showcases how companies are adapting to this evolving landscape, highlighting a shift towards refined approaches that include rigorous filtering and human preference data.
undefined
Aug 1, 2024 • 1h 4min

Interviewing Sebastian Raschka on the state of open LLMs, Llama 3.1, and AI education

Sebastian Raschka, a staff research engineer at Lightning AI and AI educator, dives into the dynamic landscape of open language models. He discusses the evolution of Llama 3.1 and its implications for AI research. Sebastian shares insights from his experience as an Arxiv moderator, shedding light on the challenges of navigating academic papers. The conversation also covers advancements in model training techniques, the importance of ethics in AI, and how open access enhances AI education. Tune in for a fascinating look at the future of AI and language models!
undefined
Jul 31, 2024 • 8min

GPT-4o-mini changed ChatBotArena

Uncover the transformation in the Chatbot Arena brought about by GPT-4o-mini. Delve into the fascinating world of model evaluations, exploring the strengths and weaknesses of the platform. Discover insights from recent performances of Llama 3 and the impact of community feedback on AI understanding. Hear about the intriguing partial solutions being developed and the roadmap ahead in the evolving landscape of language models.
undefined
Jul 23, 2024 • 15min

Llama 3.1 405b, Meta's AI strategy, and the new open frontier model ecosystem

Discussing Meta's AI strategy in the open-source AI ecosystem, comparing it to the Unix stack. Analyzing Zuckerberg's vision for open-source AI and the implications of the Llama 3.1 license. Exploring different futures for regulating frontier models in the AI economy.
undefined
Jul 17, 2024 • 14min

SB 1047, AI regulation, and unlikely allies for open models

A podcast discusses the open-source community's opposition to SB 1047 and its potential impact on AI regulation. They delve into the challenges of regulating AI developers, the emergence of unlikely allies for open models, and ponder on what should be regulated in the AI landscape today.
undefined
Jul 3, 2024 • 7min

Switched to Claude 3.5

Speculations on the role of RLHF, transitioning to Claude 3.5 for enhanced performance, product priorities, and the peak of RLHF discussed. AI generated audio with Python and 11Labs.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app