AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
The Permute and Quantise in Findou in a Neural Network
Nero nets have this nice property that if you have two adjacent layers, let's say a linear layer, som non linar activation, and another Linear Layer. You express the same function by having different permutations of the weights. The first thing that we do is, that's what we call the paper, permute, permute, then quantise in findou so the permute step is a nice observation. O, and then practically, how does this play out when you are using this to compress a nor on that work? Now, about efficiency and accuracy and all those kinds of things, right? Let's garinto so the question of how we do it.