
138 - Compositional Generalization in Neural Networks, with Najoung Kim
NLP Highlights
00:00
Is There a Con Found in the Training Data?
The first one is replacing the context control lexical items like hedgehog with character sequences that have not been observed in the training data at all. The second one is actually adding novel to against to the model to the embedding layer and using those as context controlLexical items. So you're essentially making up humans and using those with the assumption that they won't occur in the pre-training data, which is probably pretty bit essential to make.
Transcript
Play full episode