High Agency: The Podcast for AI Builders cover image

What comes after Open AI? Logan Kilpatrick on how you should prepare for the future of LLMs

High Agency: The Podcast for AI Builders

00:00

Implications of 2.5 Million Token Context Length in AI Models

Exploring factors such as latencies and costs in AI models with a large context length, highlighting the importance of context caching for efficiency and reduced expenses. Discussing strategies for shortening context to improve interactions and introducing the more cost-effective Flash model compared to other advanced models.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app