2min chapter

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) cover image

Vector Quantization for NN Compression with Julieta Martinez - #498

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

CHAPTER

The Permute and Quantise in Findou in a Neural Network

Nero nets have this nice property that if you have two adjacent layers, let's say a linear layer, som non linar activation, and another Linear Layer. You express the same function by having different permutations of the weights. The first thing that we do is, that's what we call the paper, permute, permute, then quantise in findou so the permute step is a nice observation. O, and then practically, how does this play out when you are using this to compress a nor on that work? Now, about efficiency and accuracy and all those kinds of things, right? Let's garinto so the question of how we do it.

00:00

Get the Snipd
podcast app

Unlock the knowledge in podcasts with the podcast player of the future.
App store bannerPlay store banner

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode

Save any
moment

Hear something you like? Tap your headphones to save it with AI-generated key takeaways

Share
& Export

Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more

AI-powered
podcast player

Listen to all your favourite podcasts with AI-powered features

Discover
highlights

Listen to the best highlights from the podcasts you love and dive into the full episode