The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) cover image

OLMo: Everything You Need to Train an Open Source LLM with Akshita Bhagia - #674

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

00:00

Challenges in Training Large-Scale Language Models

This chapter explores the complexities faced when training large-scale machine learning models, particularly in reproducing existing results and the impact of architectural decisions like weight tying. It highlights the significance of collaborative research, the importance of evaluation methods like perplexity, and the innovative Paloma project that provides a comprehensive benchmarking framework. Additionally, the discussion emphasizes the balance between open sourcing models and maintaining security, arguing that transparency fosters a healthier AI development ecosystem.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app