Interconnects

Nathan Lambert
undefined
Sep 27, 2024 • 14min

[Article Voiceover] Llama 3.2 Vision and Molmo: Foundations for the multimodal open-source ecosystem

Dive into the fascinating world of open-source AI with a detailed look at Llama 3.2 Vision and Molmo. Explore how multimodal models enhance capabilities by integrating visual inputs with text. Discover the architectural differences and performance comparisons among leading models. The discussion delves into current challenges, the future of generative AI, and what makes the open-source movement vital for developers. Tune in for insights that bridge technology and creativity in the evolving landscape of AI!
undefined
Sep 17, 2024 • 19min

[Article Voiceover] Reverse engineering OpenAI's o1

Dive into the future of AI with OpenAI's groundbreaking O1 reasoning system. Explore its innovative training methods and real-time inference capabilities. Delve into the challenges of scaling reinforcement learning models and the complexities of balancing human preferences with computational needs. Discover the evolution of language models and their potential to become more integrated tools in our lives. It's an exciting look at cutting-edge developments in artificial intelligence!
undefined
Sep 11, 2024 • 12min

Futures of the data foundry business model

The discussion dives into the competitive dynamics of data foundries, contrasting synthetic and human-annotated data for AI training. It explores the implications of advancing reinforcement learning with human feedback. A key focus is the future of data foundries as AI dependence escalates, highlighting potential growth vectors and the associated risks. The conversation also touches on how companies like Nvidia could dominate profits in the evolving data market. Expect insights that provoke thought about the future of AI and data sourcing!
undefined
Sep 10, 2024 • 6min

A post-training approach to AI regulation with Model Specs

Discover the pivotal role of model specifications in AI regulation. The discussion delves into current regulatory trends, emphasizing transparency and responsible AI use. It highlights the importance of documenting intentions behind computational models, fostering connections among stakeholders. The hosts explore how clear specifications can mitigate risks and anticipate future developments, paving the way for ongoing dialogue in the rapidly evolving AI landscape.
undefined
Sep 5, 2024 • 11min

OpenAI's Strawberry, LM self-talk, inference scaling laws, and spending more on inference

Discover the fascinating advancements in AI with OpenAI's Strawberry method, designed to enhance reasoning in language models. The discussion reveals the importance of inference spending and structural changes shaping future AI products. Dive into the complexities of scaling inference, where reinforcement learning and reward models play a pivotal role. Understand why optimizing inference time is crucial and explore promising avenues for further research in this rapidly evolving field.
undefined
Sep 4, 2024 • 11min

OLMoE and the hidden simplicity in training better foundation models

Dive into the innovations behind OLMoE, a cutting-edge language model that excels among its peers. Explore the challenges of training complexity and organizational hurdles. Discover the secret sauce of compounding improvements that leads to better models. This conversation unpacks not just the tech, but the strategic thinking driving advancements in AI.
undefined
Aug 28, 2024 • 8min

On the current definitions of open-source AI and the state of the data commons

The discussion dives deep into the evolving definitions of open-source AI. It highlights the challenges faced by the data commons and the necessity for better documentation. Concerns about the implications of mandating fully released data are raised. Frustration with existing definitions is palpable, as examples are urgently needed to clarify the landscape. The dialogue emphasizes the balance between accessibility and regulation in the AI realm.
undefined
Aug 16, 2024 • 9min

Nous Hermes 3 and exploiting underspecified evaluations

The discussion kicks off with the launch of a new model, questioning what defines a 'frontier model.' Notable comparisons are drawn with LAMA 3.1 and the importance of transparent evaluation metrics emerges. The conversation elaborates on valuable lessons learned from the training process of Hermes 3. The broader implications for technology policy are also highlighted, emphasizing the need for integrity in AI evaluations.
undefined
Aug 8, 2024 • 1h 2min

Interviewing Ross Taylor on LLM reasoning, Llama fine-tuning, Galactica, agents

Ross Taylor, a former LLM lead at Meta AI and co-founder of Papers with Code, dives into the cutting-edge world of language models. He shares insights on the Galactica project, its ambitions, and the ethical complexities involved. The conversation explores the potential of language models in scientific discovery, the evolution of reasoning within AI, and innovative training methodologies. Taylor emphasizes the significance of collaboration in advancing AI research while highlighting the latest developments in model alignment and user experience.
undefined
Aug 7, 2024 • 10min

A recipe for frontier model post-training

The discussion dives into the latest advancements in reinforcement learning from human feedback, focusing on the Llama 3.1 model. Key players like Apple, Meta, and Nvidia emphasize the importance of synthetic data and iterative training. Data quality emerges as a pivotal theme, with agreements on new standards in model training. The episode showcases how companies are adapting to this evolving landscape, highlighting a shift towards refined approaches that include rigorous filtering and human preference data.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app