International audienceWhile contextual language models are now dominant in the field of Natural Language Processing, the representations they build at the token level are not always suitable for all uses. In this article, we propose a new method for building word or type-level embeddings from contextual models. This method combines the generalization and the aggregation of token representations. We evaluate it for a large set of English nouns in the perspective of the building of distributional thesauri for extracting semantic similarity relations. Moreover, we analyze the differences of static embeddings and type-level embeddings according to features such as the frequency of words or the type of semantic relations these embeddings account...
Most word embedding models typically represent each word using a single vector, which makes these mo...
Distributed language representation has become the most widely used technique for language represent...
Distributed language representation has become the most widely used technique for language represent...
International audienceWhile contextual language models are now dominant in the field of Natural Lang...
International audienceMany studies were recently done for investigating the properties of contextual...
A new metaphor of two-dimensional text for data-driven semantic modeling of natural language is prop...
A new metaphor of two-dimensional text for data-driven semantic modeling of natural language is prop...
This archive contains a collection of computational models called word embeddings. These are vectors...
Treball de fi de màster en Lingüística Teòrica i Aplicada. Director: Dr. Thomas BrochhagenWord embed...
Distributional semantics is built upon the assumption that the context surrounding a given word in t...
Obtaining Better Static Word Embeddings Using Contextual Embedding Models This repository contains ...
Static word embeddings that represent words by a single vector cannot capture the variability of wor...
Real-valued word embeddings have transformed natural language processing (NLP) applications, recogni...
Distributional models of semantics have proven themselves invaluable both in cog-nitive modelling of...
A fundamental principle in distributional semantic models is to use similarity in linguistic environ...
Most word embedding models typically represent each word using a single vector, which makes these mo...
Distributed language representation has become the most widely used technique for language represent...
Distributed language representation has become the most widely used technique for language represent...
International audienceWhile contextual language models are now dominant in the field of Natural Lang...
International audienceMany studies were recently done for investigating the properties of contextual...
A new metaphor of two-dimensional text for data-driven semantic modeling of natural language is prop...
A new metaphor of two-dimensional text for data-driven semantic modeling of natural language is prop...
This archive contains a collection of computational models called word embeddings. These are vectors...
Treball de fi de màster en Lingüística Teòrica i Aplicada. Director: Dr. Thomas BrochhagenWord embed...
Distributional semantics is built upon the assumption that the context surrounding a given word in t...
Obtaining Better Static Word Embeddings Using Contextual Embedding Models This repository contains ...
Static word embeddings that represent words by a single vector cannot capture the variability of wor...
Real-valued word embeddings have transformed natural language processing (NLP) applications, recogni...
Distributional models of semantics have proven themselves invaluable both in cog-nitive modelling of...
A fundamental principle in distributional semantic models is to use similarity in linguistic environ...
Most word embedding models typically represent each word using a single vector, which makes these mo...
Distributed language representation has become the most widely used technique for language represent...
Distributed language representation has become the most widely used technique for language represent...