Authors
Sanjeev Arora, Yuanzhi Li, Yingyu Liang, Tengyu Ma, Andrej Risteski
Publication date
2016/7/1
Journal
Transactions of the Association for Computational Linguistics
Volume
4
Pages
385-399
Publisher
MIT Press
Description
Semantic word embeddings represent the meaning of a word via a vector, and are created by diverse methods. Many use nonlinear operations on co-occurrence statistics, and have hand-tuned hyperparameters and reweighting methods.
This paper proposes a new generative model, a dynamic version of the log-linear topic model of Mnih and Hinton (2007). The methodological novelty is to use the prior to compute closed form expressions for word statistics. This provides a theoretical justification for nonlinear models like PMI, word2vec, and GloVe, as well as some hyperparameter choices. It also helps explain why low-dimensional semantic embeddings contain linear algebraic structure that allows solution of word analogies, as shown by Mikolov et al. (2013a) and many subsequent papers …
Total citations
2015201620172018201920202021202220232024113036616975665410536
Scholar articles
S Arora, Y Li, Y Liang, T Ma, A Risteski - Transactions of the Association for Computational …, 2016