International audienceThis paper presents an enhanced approach for adapting a Language Model (LM) to a specific domain, with a focus on Named Entity Recognition (NER) and Named Entity Linking (NEL) tasks. Traditional NER/NEL methods require a large amounts of labeled data, which is time and resource intensive to produce. Unsupervised and semi-supervised approaches overcome this limitation but suffer from a lower quality. Our approach, called KeyWord Masking (KWM), fine-tunes a Language Model (LM) for the Masked Language Modeling (MLM) task in a special way. Our experiments demonstrate that KWM outperforms traditional methods in restoring domain-specific entities. This work is a preliminary step towards developing a more sophisticated NER/NE...
Entity Linking (EL) systems have achieved impressive results on standard benchmarks, mainly thanks t...
Named Entity Recognition (NER) is a fundamental and important research topic for many downstream NLP...
For languages with no annotated resources, transferring knowledge from rich-resource languages is an...
We explore three different methods for improving Named Entity Recognition (NER) systems based on BER...
The health and life science domains are well known for their wealth of named entities found in large...
The health and life science domains are well known for their wealth of named entities found in large...
The health and life science domains are well known for their wealth of named entities found in large...
Data augmentation methods are often used to address data scarcity in natural language processing (NL...
In this work, we explore how to learn task-specific language models aimed towards learning rich repr...
Data augmentation methods are often used to address data scarcity in natural language processing (NL...
We introduce Named Entity (NE) Language Modelling, a stochastic finite state machine approach to ide...
The current era of natural language processing (NLP) has been defined by the prominence of pre-train...
The current era of natural language processing (NLP) has been defined by the prominence of pre-train...
In this work, we explore how to learn task-specific language models aimed towards learning rich repr...
While good results have been achieved for named entity recognition (NER) in supervised settings, it ...
Entity Linking (EL) systems have achieved impressive results on standard benchmarks, mainly thanks t...
Named Entity Recognition (NER) is a fundamental and important research topic for many downstream NLP...
For languages with no annotated resources, transferring knowledge from rich-resource languages is an...
We explore three different methods for improving Named Entity Recognition (NER) systems based on BER...
The health and life science domains are well known for their wealth of named entities found in large...
The health and life science domains are well known for their wealth of named entities found in large...
The health and life science domains are well known for their wealth of named entities found in large...
Data augmentation methods are often used to address data scarcity in natural language processing (NL...
In this work, we explore how to learn task-specific language models aimed towards learning rich repr...
Data augmentation methods are often used to address data scarcity in natural language processing (NL...
We introduce Named Entity (NE) Language Modelling, a stochastic finite state machine approach to ide...
The current era of natural language processing (NLP) has been defined by the prominence of pre-train...
The current era of natural language processing (NLP) has been defined by the prominence of pre-train...
In this work, we explore how to learn task-specific language models aimed towards learning rich repr...
While good results have been achieved for named entity recognition (NER) in supervised settings, it ...
Entity Linking (EL) systems have achieved impressive results on standard benchmarks, mainly thanks t...
Named Entity Recognition (NER) is a fundamental and important research topic for many downstream NLP...
For languages with no annotated resources, transferring knowledge from rich-resource languages is an...