A super specialized chip optimized for Transformers and AI workloads can process up to 1,000 trillion operations per second and handle over 500,000 tokens per second on large models. This chip outperforms GPUs significantly, with a single server equivalent to the performance of 160 Nvidia H100 GPUs. It signifies a potential shift towards specialized AI chips becoming the norm, surpassing current GPU capabilities by a significant margin.
Our 172nd episode with a summary and discussion of last week's big AI news!
With hosts Andrey Kurenkov (https://twitter.com/andrey_kurenkov) and Jeremie Harris (https://twitter.com/jeremiecharris)
Read out our text newsletter and comment on the podcast at https://lastweekin.ai/
If you would like to become a sponsor for the newsletter, podcast, or both, please fill out this form.
Email us your questions and feedback at contact@lastweekinai.com and/or hello@gladstone.ai
- (00:00:00) Intro / Banter
- Tools & Apps
- Applications & Business
- Projects & Open Source
- Research & Advancements
- Policy & Safety
- Synthetic Media & Art
- (01:49:12) Outro + AI Song