peer reviewedDespite the widespread use of pre-trained models in NLP, well-performing pre-trained models for low-resource languages are scarce. To address this issue, we propose two novel BERT models for the Luxembourgish language that improve on the state of the art. We also present an empirical study on both the performance and robustness of the investigated BERT models. We compare the models on a set of downstream NLP tasks and evaluate their robustness against different types of data perturbations. Additionally, we provide novel datasets to evaluate the performance of Luxembourgish language models. Our findings reveal that pre-training a pre-loaded model has a positive effect on both the performance and robustness of fine-tuned models a...
Web site: https://camembert-model.frPretrained language models are now ubiquitous in Natural Languag...
Natural language processing of Low-Resource Languages (LRL) is often challenged by the lack of data....
International audienceBERT models used in specialized domains all seem to be the result of a simple ...
peer reviewedDespite the widespread use of pre-trained models in NLP, well-performing pre-trained mo...
peer reviewedPre-trained Language Models such as BERT have become ubiquitous in NLP where they have ...
Pre-trained language models have been dominating the field of natural language processing in recent ...
Large pretrained masked language models have become state-of-theart solutions for many NLP problems....
Large pre-trained masked language models have become state-of-the-art solutions for many NLP problem...
Multilingual language models such as mBERT have seen impressive cross-lingual transfer to a variety ...
5th Workshop on Clinical Natural Language Processing, ClinicalNLP 2023. held at ACL 2023 -- 14 July ...
Large pretrained masked language models have become state-of-the-art solutions for many NLP problems...
Currently, the most widespread neural network architecture for training language models is the so-ca...
Recently, the development of pre-trained language models has brought natural language processing (NL...
Deep neural language models such as BERT have enabled substantial recent advances in many natural la...
International audienceRecent advances in NLP have significantly improved the performance of language...
Web site: https://camembert-model.frPretrained language models are now ubiquitous in Natural Languag...
Natural language processing of Low-Resource Languages (LRL) is often challenged by the lack of data....
International audienceBERT models used in specialized domains all seem to be the result of a simple ...
peer reviewedDespite the widespread use of pre-trained models in NLP, well-performing pre-trained mo...
peer reviewedPre-trained Language Models such as BERT have become ubiquitous in NLP where they have ...
Pre-trained language models have been dominating the field of natural language processing in recent ...
Large pretrained masked language models have become state-of-theart solutions for many NLP problems....
Large pre-trained masked language models have become state-of-the-art solutions for many NLP problem...
Multilingual language models such as mBERT have seen impressive cross-lingual transfer to a variety ...
5th Workshop on Clinical Natural Language Processing, ClinicalNLP 2023. held at ACL 2023 -- 14 July ...
Large pretrained masked language models have become state-of-the-art solutions for many NLP problems...
Currently, the most widespread neural network architecture for training language models is the so-ca...
Recently, the development of pre-trained language models has brought natural language processing (NL...
Deep neural language models such as BERT have enabled substantial recent advances in many natural la...
International audienceRecent advances in NLP have significantly improved the performance of language...
Web site: https://camembert-model.frPretrained language models are now ubiquitous in Natural Languag...
Natural language processing of Low-Resource Languages (LRL) is often challenged by the lack of data....
International audienceBERT models used in specialized domains all seem to be the result of a simple ...