Gradient Dissent: Conversations on AI cover image

Advanced AI Accelerators and Processors with Andrew Feldman of Cerebras Systems

Gradient Dissent: Conversations on AI

CHAPTER

How to Make a Chip That Talks to Each Other

We can run sparse or dense. We harvest the sparsity by reading we get a performance boost because we're not wasting time multiplying by zero. It's published a series of blogs at NURPs showing that we could trade models that were 90% sparse to state of the accuracy, including GPT models. And they took far fewer flops to do it and could be done at much less time.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner