We introduce Probabilistic FastText, a new model for word embeddings that can capture multiple word senses, sub-word structure, and uncertainty information. In particular, we represent each word with a Gaussian mixture density, where the mean of a mixture component is given by the sum of n-grams. This representation allows the model to share statistical strength across sub-word structures (e.g. Latin roots), producing accurate representations of rare, misspelt, or even unseen words. Moreover, each component of the mixture can capture a different word sense. Probabilistic FastText outperforms both FastText, which has no probabilistic model, and dictionary-level probabilistic embeddings, which do not incorporate subword structures, on several...
Word embedding is a feature learning technique which aims at mapping words from a vocabulary into ve...
In this paper, word sense disambiguation (WSD) ac-curacy achievable by a probabilistic classier, usi...
Rare word representation has recently enjoyed a surge of interest, owing to the crucial role that ef...
We introduce Probabilistic FastText, a new model for word embeddings that can capture multiple word ...
Distributed word representations have been widely used and proven to be useful in quite a few natura...
The digital era floods us with an excessive amount of text data. To make sense of such data automati...
There are two main types of word repre-sentations: low-dimensional embeddings and high-dimensional d...
Pretraining deep language models has led to large performance gains in NLP. Despite this success, Sc...
We propose a new word embedding model, inspired by GloVe, which is formulated as a feasible least sq...
International audienceSeveral recent studies have shown the benefits of combining language and perce...
Distributed word embeddings have yielded state-of-the-art performance in many NLP tasks, mainly due ...
Traditional natural language processing has been shown to have excessive reliance on human-annotated...
We demonstrate the benefits of probabilistic representations due to their expressiveness which allow...
The GloVe word embedding model relies on solving a global optimization problem, which can be reformu...
Pretraining deep neural network architectures with a language modeling objective has brought large i...
Word embedding is a feature learning technique which aims at mapping words from a vocabulary into ve...
In this paper, word sense disambiguation (WSD) ac-curacy achievable by a probabilistic classier, usi...
Rare word representation has recently enjoyed a surge of interest, owing to the crucial role that ef...
We introduce Probabilistic FastText, a new model for word embeddings that can capture multiple word ...
Distributed word representations have been widely used and proven to be useful in quite a few natura...
The digital era floods us with an excessive amount of text data. To make sense of such data automati...
There are two main types of word repre-sentations: low-dimensional embeddings and high-dimensional d...
Pretraining deep language models has led to large performance gains in NLP. Despite this success, Sc...
We propose a new word embedding model, inspired by GloVe, which is formulated as a feasible least sq...
International audienceSeveral recent studies have shown the benefits of combining language and perce...
Distributed word embeddings have yielded state-of-the-art performance in many NLP tasks, mainly due ...
Traditional natural language processing has been shown to have excessive reliance on human-annotated...
We demonstrate the benefits of probabilistic representations due to their expressiveness which allow...
The GloVe word embedding model relies on solving a global optimization problem, which can be reformu...
Pretraining deep neural network architectures with a language modeling objective has brought large i...
Word embedding is a feature learning technique which aims at mapping words from a vocabulary into ve...
In this paper, word sense disambiguation (WSD) ac-curacy achievable by a probabilistic classier, usi...
Rare word representation has recently enjoyed a surge of interest, owing to the crucial role that ef...