Wan2Vec: Embeddings Learned on Word Association Norms

Tracking #: 2036-3249

This paper is currently under review
Helena Gomez-Adorno
Gemma Bel-Enguix
Jorge Reyes-Magaña
Gerardo Sierra

Responsible editor: 
Guest Editors Knowledge Graphs 2018

Submission type: 
Full Paper
Word embeddings are powerful for many tasks in natural language processing. In this work, we learn word embeddings using weighted graphs from word association norms (WAN) with the node2vec algorithm. Although building WAN is a difficult and time-consuming task, training the vectors from these resources is a fast and efficient process. This allows us to obtain good quality word embeddings from small corpora. We evaluate our word vectors in two ways: intrinsic and extrinsic. The intrinsic evaluation was performed with several word similarity benchmarks,WordSim-353, MC30, MTurk-287, MEN-TR-3k, SimLex-999, MTurk-771 and RG-65, and different similarity measures achieving better results than those obtained with word2vec, GloVe, and fastText, trained on a huge corpus. The extrinsic evaluation was done by measuring the quality of sentence embeddings using transfer tasks: sentiment analysis, paraphrase detection, natural language inference, and semantic textual similarity.
Full PDF Version: 
Under Review