Recent results show that deep neural networks using contextual embeddings significantly outperform non-contextual embeddings on a majority of text classification task. We offer precomputed embeddings from popular contextual ELMo model for seven languages: Croatian, Estonian, Finnish, Latvian, Lithuanian, Slovenian, and Swedish. We demonstrate that the quality of embeddings strongly depends on the size of training set and show that existing publicly available ELMo embeddings for listed languages shall be improved. We train new ELMo embeddings on much larger training sets and show their advantage over baseline non-contextual FastText embeddings. In evaluation, we use two benchmarks, the analogy task and the NER task
This work highlights some important factors for consideration when developing word vector representa...
Universal embeddings, such as BERT or ELMo, are useful for a broad set of natural language processin...
Pre-trained word embeddings encode general word semantics and lexical regularities of natural langua...
The current dominance of deep neural networks in natural language processing is based on contextual ...
ELMo language model (https://github.com/allenai/bilm-tf) used to produce contextual word embeddings,...
ELMo language model (https://github.com/allenai/bilm-tf) used to produce contextual word embeddings,...
Building machine learning prediction models for a specific natural language processing (NLP) task re...
This study evaluates the robustness of two state-of-the-art deep contextual language representations...
In text processing, deep neural networks mostly use word embeddings as an input. Embeddings have to ...
International audienceWe use the multilingual OSCAR corpus, extracted from Common Crawl via language...
International audienceRecent studies in the biomedical domain suggest that learning statistical word...
Text summarization has gained a considerable amount of research interest due to deep learning based ...
The creation of word embeddings is one of the key breakthroughs in natural language processing. Word...
A family of Natural Language Processing (NLP) tasks such as part-of- speech (PoS) tagging, Named Ent...
Distributed word representations (word embeddings) have recently contributed to competitive performa...
This work highlights some important factors for consideration when developing word vector representa...
Universal embeddings, such as BERT or ELMo, are useful for a broad set of natural language processin...
Pre-trained word embeddings encode general word semantics and lexical regularities of natural langua...
The current dominance of deep neural networks in natural language processing is based on contextual ...
ELMo language model (https://github.com/allenai/bilm-tf) used to produce contextual word embeddings,...
ELMo language model (https://github.com/allenai/bilm-tf) used to produce contextual word embeddings,...
Building machine learning prediction models for a specific natural language processing (NLP) task re...
This study evaluates the robustness of two state-of-the-art deep contextual language representations...
In text processing, deep neural networks mostly use word embeddings as an input. Embeddings have to ...
International audienceWe use the multilingual OSCAR corpus, extracted from Common Crawl via language...
International audienceRecent studies in the biomedical domain suggest that learning statistical word...
Text summarization has gained a considerable amount of research interest due to deep learning based ...
The creation of word embeddings is one of the key breakthroughs in natural language processing. Word...
A family of Natural Language Processing (NLP) tasks such as part-of- speech (PoS) tagging, Named Ent...
Distributed word representations (word embeddings) have recently contributed to competitive performa...
This work highlights some important factors for consideration when developing word vector representa...
Universal embeddings, such as BERT or ELMo, are useful for a broad set of natural language processin...
Pre-trained word embeddings encode general word semantics and lexical regularities of natural langua...