ThursdAI - The top AI news from the past week

📅 ThursdAI - Sep 5 - 👑 Reflection 70B beats Claude 3.5, Anthropic Enterprise 500K context, 100% OSS MoE from AllenAI, 1000 agents world sim, Replit agent is the new Cursor? and more AI news

5 snips
Sep 6, 2024
Matt Shumer and Sahil, co-founders of Glaive.ai, share insights on their new Reflection-LLama 70B model, which surpasses Claude 3.5 in performance. Niklas, a leading author of OLMoE, discusses the significance of open-source AI in advancing model training techniques. Eugene, known as pico_creator, reveals RWKV's deployment to 1.5 billion devices. The conversation highlights major developments such as Anthropic's expansive context window and Elon Musk's massive AI training cluster, showcasing the rapid evolution of AI technologies and their potential impact.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ANECDOTE

RWKV Deployed on Billions

  • RWKV.CPP, an open source AI model, was found deployed to 1.5 billion Windows 10 and 11 devices.
  • Microsoft uses RWKV for their Windows Local Recall and Co-Pilot features to reduce resource usage and enable multilingual support.
INSIGHT

OLMoE Fully Open Source Release

  • Allen Institute for AI released OLMoE, a fully open source Mixture of Experts model with weights, data, and training code.
  • The release includes detailed logs and 244 intermediate checkpoints to facilitate research and reproducibility.
INSIGHT

The Efficiency of MOE Models

  • Mixture of Experts (MOE) models can reduce computation by activating only a subset of parameters, improving efficiency.
  • MOE training is more complex due to routing challenges, but major labs continue to invest in them for scalability.
Get the Snipd Podcast app to discover more snips from this episode
Get the app