31 - Tying Word Vectors and Word Classifiers: A Loss Framework for Language Modeling
ICLR 2017 paper by Hakan Inan, Khashayar Khosravi…
11 Minuten
Podcast
Podcaster
Beschreibung
vor 8 Jahren
ICLR 2017 paper by Hakan Inan, Khashayar Khosravi, Richard Socher,
presented by Waleed. The paper presents some tricks for training
better language models. It introduces a modified loss function for
language modeling, where producing a word that is similar to the
target word is not penalized as much as producing a word that is
very different to the target (I've seen this in other places, e.g.,
image classification, but not in language modeling). They also give
theoretical and empirical justification for tying input and output
embeddings.
https://www.semanticscholar.org/paper/Tying-Word-Vectors-and-Word-Classifiers-A-Loss-Fra-Inan-Khosravi/424aef7340ee618132cc3314669400e23ad910ba
presented by Waleed. The paper presents some tricks for training
better language models. It introduces a modified loss function for
language modeling, where producing a word that is similar to the
target word is not penalized as much as producing a word that is
very different to the target (I've seen this in other places, e.g.,
image classification, but not in language modeling). They also give
theoretical and empirical justification for tying input and output
embeddings.
https://www.semanticscholar.org/paper/Tying-Word-Vectors-and-Word-Classifiers-A-Loss-Fra-Inan-Khosravi/424aef7340ee618132cc3314669400e23ad910ba
Weitere Episoden
30 Minuten
vor 2 Jahren
51 Minuten
vor 2 Jahren
45 Minuten
vor 2 Jahren
48 Minuten
vor 2 Jahren
36 Minuten
vor 2 Jahren
In Podcasts werben
Kommentare (0)