

Interconnects
Nathan Lambert
Audio essays about the latest developments in AI and interviews with leading scientists in the field. Breaking the hype, understanding what's under the hood, and telling stories. www.interconnects.ai
Episodes
Mentioned books

May 1, 2024 • 13min
RLHF: A thin line between useful and lobotomized
Exploring the mechanisms for making models chattier, the chattiness paradox, and the next steps for RLHF research in AI generated audio with Python. Delving into the impact of style on model evaluation and improvement, advancements in training language models, and exploring preference alignment in data sets. Discussing biases in GPT-4, alternative models like alpaca and vacuna, and the importance of data in AI research.

Apr 30, 2024 • 10min
Phi 3 and Arctic: Outlier LMs are hints
The podcast discusses the future of AI industry with outlier LLMs, focusing on Phi 3 and Arctic models. It explores trends in open mixture of expert models and the impact of synthetic data on small models. The episode also touches on the training techniques of the new models, including Arctic's sparse mixture of experts architecture and Phi 3's evolution in the LLM space.

Apr 24, 2024 • 11min
AGI is what you want it to be
Exploring the ambiguous definition of AGI and its impact on AI discourse, focusing on the role of reinforcement learning and the challenges of defining genuine intelligence. Delving into various tests for AGI, such as the Modern Turing Test, and the debate around surpassing human abilities. Reflecting on the balance between compute power, energy efficiency, and the future of AGI in AI model training.

Apr 21, 2024 • 15min
Llama 3: Scaling open LLMs to AGI
Explore the challenges and advancements in scaling open large language models towards achieving AGI, including Meta's release of the LAMA3 model. Analysis of training models at Meta and comparisons with other hyperscalers. Discuss training and fine-tuning strategies for long context behavior. Exploration of pricing calculation methods, preference data, and AI platform features. Future prospects of the open LLM ecosystem through the upcoming llama 3 400b model.

Apr 17, 2024 • 8min
Stop "reinventing" everything to "solve" alignment
Delve into integrating non-computing science into reinforcement learning for AI alignment. Explore social choice theory for diverse human feedback. Discover OLMo 1.7 7B model with good benchmarks and open design. Unveil insights on pluralistic alignment in AI systems for inclusivity.

Apr 15, 2024 • 7min
The end of the "best open LLM"
Exploring the performance trade-offs of open LLM models like Mistrel 8 and DBRX. Discussion on the compute efficiency, growth trends, and optimizing resources for large language models. Debate on non-commercial licenses and comparisons between open and closed LLMs.

Apr 3, 2024 • 9min
Why we disagree on what open-source AI should be
The podcast discusses the varied perspectives on open-source AI, including accelerationists, scientists promoting transparency, inclusion and fighting power concentration, and freedom advocates. It explores the complexities of defining 'openness' in the tech industry and emphasizes the importance of diverse viewpoints in shaping open-source projects.

Mar 29, 2024 • 17min
DBRX: The new best open LLM and Databricks' ML strategy
Exploring Databricks' new model DBRX, surpassing Mixtral and Llama 2 in performance. Discussion on AI generated audio, Python, and 11Labs. Details on open LLM and AI strategy, playing with DBRX Instruct. Digging into the narrative and strategic ML implementation. Exploring model capabilities, author's AI newsletter, and system dynamics.

Mar 21, 2024 • 13min
Evaluations: Trust, performance, and price (bonus, announcing RewardBench)
Exploring the shift towards trust and performance-focused evaluations, the rising costs of evaluation tools, and the introduction of RewardBench for evaluating reward models. Discussing the challenges in evaluating different AI models, the need for standardized frameworks, and incremental upgrades in evaluation systems.

Mar 13, 2024 • 11min
Model commoditization and product moats
Exploring challenges in the AI industry with the rise of GPT4 class models and the development of unique language models. Insights on building moats despite commoditization, opportunities in the Open, and the usefulness of LLMs. Discussion on upcoming developments and the competitive dynamics between companies in the LLM space.


