April 30, 2019
Learning Word Vectors for 157 Languages
International Conference on Language Resources and Evaluation (LREC)
Distributed word representations, or word vectors, have recently been applied to many tasks in natural language processing, leading to state-of-the-art performance. A key ingredient to the successful application of these representations is to train them on very large corpora, and use these pre-trained models in downstream tasks. In this paper, we describe how we trained such high quality word representations for 157 languages.
By: Edouard Grave, Piotr Bojanowski, Prakhar Gupta, Armand Joulin, Tomas Mikolov
Facebook AI Research
Natural Language Processing & Speech