Nowadays, contextual language models can solve a wide range of language tasks such as text classification, question answering and machine translation. These tasks often require the model to have knowledge about general language understanding, like how words relate to each other. This understanding is acquired through a pre-training stage where the model learn features from raw text data. However, we do not fully understand all the features the model learns through this pre-training stage. Does there exists information yet to be utilized? Can we make predictions more explainable? This thesis aims to extend the knowledge of what features a language model have acquired. We have chosen the model architecture BERT and have analyzed its word rep...
BERT model for Galician (12 layers). Marcos Garcia (2021). Exploring the Representation of Word Mea...
We present the MULTISEM systems submitted to SemEval 2020 Task 3: Graded Word Similarity in Context ...
Contextualized word embeddings, i.e. vector representations for words in context, are naturally seen...
Despite the success of contextualized language models on various NLP tasks, it is still unclear what...
The latest work on language representations carefully integrates contextualized features into langua...
When classifying texts using a linear classifier, the texts are commonly represented as feature vect...
This paper presents the first unsupervised approach to lexical semantic change that makes use of con...
International audienceDeep learning models like BERT, a stack of attention layers with an unsupervis...
Pre-trained language models such as BERT (Bidirectional Encoder Representations from Transformers) h...
Language models have become nearly ubiquitous in natural language processing applications achieving ...
International audienceBERT is a recent language representation model that has surprisingly performed...
In this paper we present a comparison between the linguistic knowledge encoded in the internal repre...
Recent work has shown evidence that the knowledge acquired by multilingual BERT (mBERT) has two comp...
Pre-trained language models (LMs) encode rich information about linguistic structure but their knowl...
Comunicació presentada al 57th Annual Meeting of the Association for Computational Linguistic (ACL 2...
BERT model for Galician (12 layers). Marcos Garcia (2021). Exploring the Representation of Word Mea...
We present the MULTISEM systems submitted to SemEval 2020 Task 3: Graded Word Similarity in Context ...
Contextualized word embeddings, i.e. vector representations for words in context, are naturally seen...
Despite the success of contextualized language models on various NLP tasks, it is still unclear what...
The latest work on language representations carefully integrates contextualized features into langua...
When classifying texts using a linear classifier, the texts are commonly represented as feature vect...
This paper presents the first unsupervised approach to lexical semantic change that makes use of con...
International audienceDeep learning models like BERT, a stack of attention layers with an unsupervis...
Pre-trained language models such as BERT (Bidirectional Encoder Representations from Transformers) h...
Language models have become nearly ubiquitous in natural language processing applications achieving ...
International audienceBERT is a recent language representation model that has surprisingly performed...
In this paper we present a comparison between the linguistic knowledge encoded in the internal repre...
Recent work has shown evidence that the knowledge acquired by multilingual BERT (mBERT) has two comp...
Pre-trained language models (LMs) encode rich information about linguistic structure but their knowl...
Comunicació presentada al 57th Annual Meeting of the Association for Computational Linguistic (ACL 2...
BERT model for Galician (12 layers). Marcos Garcia (2021). Exploring the Representation of Word Mea...
We present the MULTISEM systems submitted to SemEval 2020 Task 3: Graded Word Similarity in Context ...
Contextualized word embeddings, i.e. vector representations for words in context, are naturally seen...