Last Week in AI cover image

#181 - Google Chatbots, Cerebras vs Nvidia, AI Doom, ElevenLabs Controversy

Last Week in AI

00:00

Efficiency through Integration

Using an 8 billion parameter model costs only 10 cents per million tokens, demonstrating the affordability of large language models (LLMs). However, achieving a high usage of tokens for inference remains challenging. The hardware's efficiency stems from its integration onto a single chip, facilitating tighter connections between logic and memory, which improves data transfer speeds during inference. This innovative approach contrasts with traditional high bandwidth memory systems, contributing to significant cost-effectiveness in processing.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app