Think Globally, Embed Locally - Locally Linear Meta-embedding of Words



Bollegala, Danushka, Hayashi, Kohei and Kawarabayashi, Ken-ichi
(2018) Think Globally, Embed Locally - Locally Linear Meta-embedding of Words. PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018-J. pp. 3970-3976.

[img] Text
1709.06671v1.pdf - Submitted version

Download (496kB)

Abstract

Distributed word embeddings have shown superior performances in numerous Natural Language Processing (NLP) tasks. However, their performances vary significantly across different tasks, implying that the word embeddings learnt by those methods capture complementary aspects of lexical semantics. Therefore, we believe that it is important to combine the existing word embeddings to produce more accurate and complete \emph{meta-embeddings} of words. For this purpose, we propose an unsupervised locally linear meta-embedding learning method that takes pre-trained word embeddings as the input, and produces more accurate meta embeddings. Unlike previously proposed meta-embedding learning methods that learn a global projection over all words in a vocabulary, our proposed method is sensitive to the differences in local neighbourhoods of the individual source word embeddings. Moreover, we show that vector concatenation, a previously proposed highly competitive baseline approach for integrating word embeddings, can be derived as a special case of the proposed method. Experimental results on semantic similarity, word analogy, relation classification, and short-text classification tasks show that our meta-embeddings to significantly outperform prior methods in several benchmark datasets, establishing a new state of the art for meta-embeddings.

Item Type: Article
Uncontrolled Keywords: cs.CL, cs.CL, cs.LG, cs.NE
Depositing User: Symplectic Admin
Date Deposited: 09 Jan 2018 09:19
Last Modified: 19 Jan 2023 06:46
DOI: 10.24963/ijcai.2018/552
Related URLs:
URI: https://livrepository.liverpool.ac.uk/id/eprint/3014891