Pre-trained language models (LMs) encode rich information about linguistic structure but their knowledge about lexical polysemy remains unclear. We propose a novel experimental setup for analysing this knowledge in LMs specifically trained for different languages (English, French, Spanish and Greek) and in multilingual BERT. We perform our analysis on datasets carefully designed to reflect different sense distributions, and control for parameters that are highly correlated with polysemy such as frequency and grammatical category. We demonstrate that BERT-derived representations reflect words' polysemy level and their partitionability into senses. Polysemy-related information is more clearly present in English BERT embeddings, but models in ...
The success of large pretrained language models (LMs) such as BERT and RoBERTa has sparked interest ...
Large pre-trained masked language models have become state-of-the-art solutions for many NLP problem...
Nowadays, contextual language models can solve a wide range of language tasks such as text classific...
Pre-trained language models (LMs) encode rich information about linguistic structure but their knowl...
The term ‘meaning’, as it is presently employed in Linguistics, is a polysemous concept, covering a ...
One of the central aspects of contextualised language models is that they should be able to distingu...
Most words in natural languages are polysemous; that is, they have related but different meanings in...
We present the MULTISEM systems submitted to SemEval 2020 Task 3: Graded Word Similarity in Context ...
Recent work has shown evidence that the knowledge acquired by multilingual BERT (mBERT) has two comp...
International audienceIn this article, we present an experiment of linguistic parameter tuning in th...
How universal is human conceptual structure? The way concepts are organized in the human brain may r...
We propose a new model for learning bilingual word representations from non-parallel document-aligne...
International audienceBERT is a recent language representation model that has surprisingly performed...
Discrimination learning (DL), a simple learning mechanism has proven to be a powerful model for desc...
International audienceMultilingual pretrained language models have demonstrated remarkable zero-shot...
The success of large pretrained language models (LMs) such as BERT and RoBERTa has sparked interest ...
Large pre-trained masked language models have become state-of-the-art solutions for many NLP problem...
Nowadays, contextual language models can solve a wide range of language tasks such as text classific...
Pre-trained language models (LMs) encode rich information about linguistic structure but their knowl...
The term ‘meaning’, as it is presently employed in Linguistics, is a polysemous concept, covering a ...
One of the central aspects of contextualised language models is that they should be able to distingu...
Most words in natural languages are polysemous; that is, they have related but different meanings in...
We present the MULTISEM systems submitted to SemEval 2020 Task 3: Graded Word Similarity in Context ...
Recent work has shown evidence that the knowledge acquired by multilingual BERT (mBERT) has two comp...
International audienceIn this article, we present an experiment of linguistic parameter tuning in th...
How universal is human conceptual structure? The way concepts are organized in the human brain may r...
We propose a new model for learning bilingual word representations from non-parallel document-aligne...
International audienceBERT is a recent language representation model that has surprisingly performed...
Discrimination learning (DL), a simple learning mechanism has proven to be a powerful model for desc...
International audienceMultilingual pretrained language models have demonstrated remarkable zero-shot...
The success of large pretrained language models (LMs) such as BERT and RoBERTa has sparked interest ...
Large pre-trained masked language models have become state-of-the-art solutions for many NLP problem...
Nowadays, contextual language models can solve a wide range of language tasks such as text classific...