137 - Nearest Neighbor Language Modeling and Machine Translation, with Urvashi Khandelwal
13/01/2023
0:00
35:56
We invited Urvashi Khandelwal, a research scientist at Google Brain to talk about nearest neighbor language and machine translation models. These models interpolate parametric (conditional) language models with non-parametric distributions over the closest values in some data stores built from relevant data. Not only are these models shown to outperform the usual parametric language models, they also have important implications on memorization and generalization in language models.
Urvashi's webpage: https://urvashik.github.io
Papers discussed:
1) Generalization through memorization: Nearest Neighbor Language Models (https://www.semanticscholar.org/paper/7be8c119dbe065c52125ee7716601751f3116844)
2)Nearest Neighbor Machine Translation (https://www.semanticscholar.org/paper/20d51f8e449b59c7e140f7a7eec9ab4d4d6f80ea)
Mais episódios de "NLP Highlights"
Não percas um episódio de “NLP Highlights” e subscrevê-lo na aplicação GetPodcast.