
138 - Compositional Generalization in Neural Networks, with Najoung Kim
NLP Highlights
Pre-Trained Models - What's the Second Setup?
The results showed that making the substitution does degrade the performance compared to what's been reported in the literature. So this is actually about like 15 to 20% point degradation across different character sampling strategiesthat we tested. And I think this does suggest that the reported results in the literature have been overestimated to some degree for not having controlled for this lexical compound, he says.
00:00
Transcript
Play full episode
Remember Everything You Learn from Podcasts
Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.