Authors
Peter Sutor, Yiannis Aloimonos, Cornelia Fermuller, Douglas Summers-Stay
Publication date
2019/3/28
Conference
2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR)
Pages
544-549
Publisher
IEEE
Description
Word embeddings are commonly used to measure word-level semantic similarity in text, especially in direct word-to-word comparisons. However, the relationships between words in the embedding space are often viewed as approximately linear and concepts comprised of multiple words are a sort of linear combination. In this paper, we demonstrate that this is not generally true and show how the relationships can be better captured by leveraging the topology of the embedding space. We propose a technique for directly computing new vectors representing multiple words in a way that naturally combines them into a new, more consistent space where distance better correlates to similarity. We show that this technique works well for natural language, even when it comprises multiple words, on a simple task derived from WordNet synset descriptions and examples of words. Thus, the generated vectors better represent …
Total citations
2020202120222023211
Scholar articles
P Sutor, Y Aloimonos, C Fermuller, D Summers-Stay - 2019 IEEE Conference on Multimedia Information …, 2019