To work with textual data, machine learning algorithms, in particular, neural networks, require word embeddings – vector representations of words in high-dimensional space. There are languages with a small amount of available resources. Exploiting the knowledge from the well-resourced languages for under-resourced languages is possible with cross-lingual embeddings by aligning the embeddings of one language with the vector space of another language. Existing methods for aligning embeddings are intended for context-independent embeddings, where every word has one representation. We propose a method, based on a dictionary and a parallel corpus aligns contextual embeddings, which capture more information about the context in which words appear...
After introducing the necessary background based on a review of the literature, this paper presents ...
In this work, we trained different bilingual word embeddings models without word alignments (BilBOWA...
The use of distributed vector representations of words in Natural Language Processing has become est...
To work with textual data, machine learning algorithms, in particular, neural networks, require word...
Building machine learning prediction models for a specific NLP task requires sufficient training dat...
One of the notable developments in current natural language processing is the practical efficacy of ...
The ability to accurately align concepts between languages can provide significant benefits in many ...
Cross-lingual embeddings are vector space representations where word translations tend to be co-loca...
Word embeddings - dense vector representations of a word’s distributional semantics - are an indespe...
In this paper, we present a thorough investigation on methods that align pre-trained contextualized ...
12 pagesRecent work in cross-lingual contextual word embedding learning cannot handle multi-sense wo...
Word embeddings represent words in a numeric space so that semantic relations between words are repr...
Cross-lingual embeddings are vector space representations where word translations tend to be co-loca...
Cross-lingual word embeddings aim to bridge the gap between high-resource and low-resource languages...
International audienceWe use the multilingual OSCAR corpus, extracted from Common Crawl via language...
After introducing the necessary background based on a review of the literature, this paper presents ...
In this work, we trained different bilingual word embeddings models without word alignments (BilBOWA...
The use of distributed vector representations of words in Natural Language Processing has become est...
To work with textual data, machine learning algorithms, in particular, neural networks, require word...
Building machine learning prediction models for a specific NLP task requires sufficient training dat...
One of the notable developments in current natural language processing is the practical efficacy of ...
The ability to accurately align concepts between languages can provide significant benefits in many ...
Cross-lingual embeddings are vector space representations where word translations tend to be co-loca...
Word embeddings - dense vector representations of a word’s distributional semantics - are an indespe...
In this paper, we present a thorough investigation on methods that align pre-trained contextualized ...
12 pagesRecent work in cross-lingual contextual word embedding learning cannot handle multi-sense wo...
Word embeddings represent words in a numeric space so that semantic relations between words are repr...
Cross-lingual embeddings are vector space representations where word translations tend to be co-loca...
Cross-lingual word embeddings aim to bridge the gap between high-resource and low-resource languages...
International audienceWe use the multilingual OSCAR corpus, extracted from Common Crawl via language...
After introducing the necessary background based on a review of the literature, this paper presents ...
In this work, we trained different bilingual word embeddings models without word alignments (BilBOWA...
The use of distributed vector representations of words in Natural Language Processing has become est...