The Gradient: Perspectives on AI cover image

Hattie Zhou: Lottery Tickets and Algorithmic Reasoning in LLMs

The Gradient: Perspectives on AI

00:00

Is There a Way to Do Random Pruning?

One paper looked at all of these proposed methods and compared them to a version where you just shuffle the weights. It turns out that it is basically like the performance is in this Englishable from some sort of random subnet work, so that kind of suggesting that maybe none of these methods are doing really anything more than something you can do with just random pruning or like a clever version of random pruning. I haven't seen anything recently that had better performance on these things, so it seems like maybe this was just a really hard problem or maybe this is kind of not the easiest way to be doing more efficient training.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app