
137 - Nearest Neighbor Language Modeling and Machine Translation, with Urvashi Khandelwal
NLP Highlights
00:00
The Distribution of Next Words From the Neural Language Model
I am in given that you're retrieving a finite number, probably a small number of nearest neighbors. The distribution of next words from the nearest neighbor language model is pretty sparse. Can you give us an intuition of what that distribution would look like in comparison to the distribution of the next word from the usual little language model? Yeah. I mean, it kind of depends on the, because we're saving every single token from our data set,. ItKind of depends on how sparse the signal for that token might be in our data set.
Transcript
Play full episode