ELMo is a deep contextualized word representation from AllenNLP that models both characteristics of word use (e.g., syntax and semantics), and how these uses vary across linguistic contexts (i.e. to model polysemy).

All edits by  Vicki Langer 

Edits on 10 May, 2019
Vicki Langer
Vicki Langer approved a suggestion from Golden's AI on 10 May, 2019
Edits made to:
Article (+9/-9 characters)

All models except for the 5.5B model were trained on the 1 Billion Word Benchmark, approximately 800M tokens of news crawl data from WMT 2011. The ELMo 5.5B model was trained on a dataset of 5.5B tokens consisting of WikipediaWikipedia (1.9B) and all of the monolingual news crawl data from WMT 2008-2012 (3.6B). In tasks where a direct comparison was made, the 5.5B model has slightly higher performance then the original ELMo model.

Golden logo
Text is available under the Creative Commons Attribution-ShareAlike 4.0; additional terms apply. By using this site, you agree to our Terms & Conditions.