
138 - Compositional Generalization in Neural Networks, with Najoung Kim
NLP Highlights
Structure to Sequence Models Without Structural Priors Don't Do Well on Generalization
Sequence to sequence models without structural priors don't really do well on structural generalization. Structural generalization is only 20% of the original Cogs generalization set. So this fact makes it seem like models are doing pretty well overall. But what you will notice if you look at the literature a little bit is that most of the pure sequence to sequence based approaches tend to saturate around 80 something percent. And my speculation is that it's because they're getting almost 0% on the structural cases and are getting all of the lexical cases, right? Although not often they report this like lexical structural division.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.