Interconnects

Nathan Lambert
undefined
Jan 24, 2024 • 10min

Local LLMs, some facts some fiction

The podcast discusses the benefits of local LLMs, strategies to optimize latency, and the integration of LLMs into consumer devices. It explores the role of local models in machine learning for personalization and optimization for inference. The influence of ML labs and their larger ambitions on the future is also discussed, highlighting Alama's popularity and Meta's build-out plans and open-source strategy.
undefined
Jan 17, 2024 • 8min

Multimodal blogging: My AI tools to expand your audience

A podcast discusses multimodal blogging, AI tools for content creation, and expanding audience reach. The speaker shares their workflow in building a suite of tools for bloggers and explores the use of AI tools like Passport, audio conditioning, and voice cloning. They also discuss future advancements in text to video models and automation in research talks and video creation.
undefined
Jan 10, 2024 • 16min

Multimodal LM roundup: Unified IO 2, inputs and outputs, Gemini, LLaVA-RLHF, and RLHF questions

This podcast discusses recent developments in the multimodal space, including the Unified IO 2 model, collecting preference data for images, LLaVA-RLHF experiments, and challenges in multimodal RLHF. They explore the architecture and challenges of multimodal models, the potential of GPT for V in multimodal RLHF, and the use of RLHF technique in multimodal models. They also discuss the importance of clearer terminology and the adoption of synthetic data in this context.
undefined
Jan 5, 2024 • 14min

Where 2024’s “open GPT4” can’t match OpenAI’s

And why the comparisons don't really matter. Repeated patterns in the race for reproducing ChatGPT, another year of evaluation crises, and people who will take awesome news too far.This is AI generated audio with Python and 11LabsSource code: https://github.com/natolambert/interconnects-toolsOriginal post: https://www.interconnects.ai/p/open-gpt4-limitations00:00 Where 2024's "open GPT4" can't match OpenAI's03:19 Models vs. products04:51 RLHF progress: Revisiting Llama 2's release and potential in 202408:30 Smaller scale open RLHF10:33 Opportunities12:24 Commentary This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit www.interconnects.ai/subscribe
undefined
Dec 21, 2023 • 36min

Interviewing Tri Dao and Michael Poli of Together AI on the future of LLM architectures

Tri Dao, an incoming professor at Princeton and Chief Scientist at Together AI, joins Michael Poli, a Stanford PhD graduate and research scientist at Together AI. They dive into why traditional attention mechanisms may not scale effectively and introduce innovative models like Striped Hyena and Mamba. The duo discusses hardware optimization for these architectures and predicts exciting developments in AI for 2024, challenging the dominance of current transformer models. Their insights reflect a transformative wave in machine learning.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app