Pretrained multilingual contextual representations have shown great success, but due to the limits of their pretraining data, their benefits do not apply equally to all language varieties. This presents a challenge for language varieties unfamiliar to these models, whose labeled \emph{and unlabeled} data is too limited to train a monolingual model effectively. We propose the use of additional language-specific pretraining and vocabulary augmentation to adapt multilingual models to low-resource settings. Using dependency parsing of four diverse low-resource language varieties as a case study, we show that these methods significantly improve performance over baselines, especially in the lowest-resource cases, and demonstrate the importance of...
Pre-trained multilingual language models play an important role in cross-lingual natural language un...
Transfer learning has led to large gains in performance for nearly all NLP tasks while making downst...
Recently, the development of pre-trained language models has brought natural language processing (NL...
Pretrained multilingual language models have become a common tool in transferring NLP capabilities t...
Multilingual language models such as mBERT have seen impressive cross-lingual transfer to a variety ...
For many (minority) languages, the resources needed to train large models are not available. We inve...
Multilingual language models are widely used to extend NLP systems to low-resource languages. Howeve...
Large pretrained masked language models have become state-of-the-art solutions for many NLP problems...
The current dominance of deep neural networks in natural language processing is based on contextual ...
Supervised deep learning-based approaches have been applied to task-oriented dialog and have proven ...
For many (minority) languages, the resources needed to train large models are not available. We inve...
This paper investigates very low resource language model pretraining, when less than 100 thousand se...
We investigate whether off-the-shelf deep bidirectional sentence representations (Devlin et al., 201...
Thesis (Master's)--University of Washington, 2014Dependency parsing is an important natural language...
There are over 7000 languages spoken on earth, but many of these languages suffer from a dearth of n...
Pre-trained multilingual language models play an important role in cross-lingual natural language un...
Transfer learning has led to large gains in performance for nearly all NLP tasks while making downst...
Recently, the development of pre-trained language models has brought natural language processing (NL...
Pretrained multilingual language models have become a common tool in transferring NLP capabilities t...
Multilingual language models such as mBERT have seen impressive cross-lingual transfer to a variety ...
For many (minority) languages, the resources needed to train large models are not available. We inve...
Multilingual language models are widely used to extend NLP systems to low-resource languages. Howeve...
Large pretrained masked language models have become state-of-the-art solutions for many NLP problems...
The current dominance of deep neural networks in natural language processing is based on contextual ...
Supervised deep learning-based approaches have been applied to task-oriented dialog and have proven ...
For many (minority) languages, the resources needed to train large models are not available. We inve...
This paper investigates very low resource language model pretraining, when less than 100 thousand se...
We investigate whether off-the-shelf deep bidirectional sentence representations (Devlin et al., 201...
Thesis (Master's)--University of Washington, 2014Dependency parsing is an important natural language...
There are over 7000 languages spoken on earth, but many of these languages suffer from a dearth of n...
Pre-trained multilingual language models play an important role in cross-lingual natural language un...
Transfer learning has led to large gains in performance for nearly all NLP tasks while making downst...
Recently, the development of pre-trained language models has brought natural language processing (NL...