Latent Space: The AI Engineer Podcast cover image

Llama 2: The New Open LLM SOTA (ft. Nathan Lambert, Matt Bornstein, Anton Troynikov, Russell Kaplan, Whole Mars Catalog et al.)

Latent Space: The AI Engineer Podcast

00:00

Trillion tokens for the 7B model: More data, less disclosure

The amount of tokens for the 7B model has increased from 1.2 trillion. The data is being kept secret due to previous issues. The context length is the same for all models. It is believed that using a smaller context length can result in better performance. The hope is that someone will discover a way to increase it now that it's open source.

Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner
Get the app