The Gradient: Perspectives on AI cover image

Hattie Zhou: Lottery Tickets and Algorithmic Reasoning in LLMs

The Gradient: Perspectives on AI

CHAPTER

Is There a Way to Do Random Pruning?

One paper looked at all of these proposed methods and compared them to a version where you just shuffle the weights. It turns out that it is basically like the performance is in this Englishable from some sort of random subnet work, so that kind of suggesting that maybe none of these methods are doing really anything more than something you can do with just random pruning or like a clever version of random pruning. I haven't seen anything recently that had better performance on these things, so it seems like maybe this was just a really hard problem or maybe this is kind of not the easiest way to be doing more efficient training.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner