We investigate whether off-the-shelf deep bidirectional sentence representations (Devlin et al., 2018) trained on a massively multilingual corpus (multilingual BERT) enable the development of an unsupervised universal dependency parser. This approach only leverages a mix of monolingual corpora in many languages and does not require any translation data making it applicable to low-resource languages. In our experiments we outperform thebest CoNLL 2018 language-specific systems in all of the shared task’s six truly low-resource languages while using a single system. However, we also find that (i) parsing accuracy still varies dramatically when changing the training languages and (ii) in some target languages zero-shot transfer fails under all...
In this thesis, we explore the impact of M-BERT and different transfer sizes on the choice of differ...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
For many (minority) languages, the resources needed to train large models are not available. We inve...
We investigate whether off-the-shelf deep bidirectional sentence representations (Devlin et al., 201...
Natural language processing problems (such as speech recognition, text-based data mining, and text o...
Recent advances in multilingual dependency parsing have brought the idea of a truly universal parser...
Dependency parsing is an essential component of several NLP applications owing its ability to captur...
Multilingual dependency parsing encapsulates any attempt to parse multiple languages. It can involve...
International audienceThis paper studies cross-lingual transfer for dependency parsing, focusing on ...
Recent advances in multilingual language modeling have brought the idea of a truly universal parser ...
International audienceWe propose a novel approach to cross-lingual part-of-speech tagging and depend...
Pretrained multilingual contextual representations have shown great success, but due to the limits o...
As numerous modern NLP models demonstrate high-performance in various tasks when trained with resour...
International audienceWe propose UDP, the first training-free parser for Universal Dependencies (UD)...
The Conference on Computational Natural Language Learning (CoNLL) features a shared task, in which p...
In this thesis, we explore the impact of M-BERT and different transfer sizes on the choice of differ...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
For many (minority) languages, the resources needed to train large models are not available. We inve...
We investigate whether off-the-shelf deep bidirectional sentence representations (Devlin et al., 201...
Natural language processing problems (such as speech recognition, text-based data mining, and text o...
Recent advances in multilingual dependency parsing have brought the idea of a truly universal parser...
Dependency parsing is an essential component of several NLP applications owing its ability to captur...
Multilingual dependency parsing encapsulates any attempt to parse multiple languages. It can involve...
International audienceThis paper studies cross-lingual transfer for dependency parsing, focusing on ...
Recent advances in multilingual language modeling have brought the idea of a truly universal parser ...
International audienceWe propose a novel approach to cross-lingual part-of-speech tagging and depend...
Pretrained multilingual contextual representations have shown great success, but due to the limits o...
As numerous modern NLP models demonstrate high-performance in various tasks when trained with resour...
International audienceWe propose UDP, the first training-free parser for Universal Dependencies (UD)...
The Conference on Computational Natural Language Learning (CoNLL) features a shared task, in which p...
In this thesis, we explore the impact of M-BERT and different transfer sizes on the choice of differ...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
For many (minority) languages, the resources needed to train large models are not available. We inve...