Thesis (Ph.D.)--University of Washington, 2021Multilingual modeling comes up in natural language processing at any scale. High-resource language corpora train high-performing models, and can be combined with other language corpora of all sizes to make better models for low-resource languages. Projects like Universal Dependencies even make it possible to train highly multilingual models from standardized morphosyntactic labels. Multilingual (or, more generally, multi-source) training does not consistently improve modeling performance, however. With an abundance of language resources comes a difficult design choice: which corpora will train better together rather than separately? More specifically, when is it worthwhile to supplement (i....
Pretrained multilingual contextual representations have shown great success, but due to the limits o...
The Conference on Computational Natural Language Learning (CoNLL) features a shared task, in which p...
Scaling existing applications and solutions to multiple human languages has traditionally proven to ...
Thesis (Master's)--University of Washington, 2014Dependency parsing is an important natural language...
Recent research has shown promise in multilingual modeling, demonstrating how a single model is capa...
This thesis presents several studies in neural dependency parsing for typologically diverse language...
Each year the Conference on Com-putational Natural Language Learning (CoNLL)1 features a shared task...
© 2019 Association for Computational Linguistics This paper explores the task of leveraging typology...
In this thesis, we investigate the use of latent variables to model complex dependencies in natural ...
accepted to appear in the special issue on Cross-Language Algorithms and ApplicationsPeer reviewe
How cross-linguistically applicable are NLP models, specifically language models? A fair comparison ...
The Conference on Computational Natural Language Learning features a shared task, in which participa...
International audienceThis paper describes how a tokenizer can be trained from any dataset in the Un...
Multilingual dependency parsing encapsulates any attempt to parse multiple languages. It can involve...
As more and more syntactically-annotated corpora become available for a wide variety of languages, m...
Pretrained multilingual contextual representations have shown great success, but due to the limits o...
The Conference on Computational Natural Language Learning (CoNLL) features a shared task, in which p...
Scaling existing applications and solutions to multiple human languages has traditionally proven to ...
Thesis (Master's)--University of Washington, 2014Dependency parsing is an important natural language...
Recent research has shown promise in multilingual modeling, demonstrating how a single model is capa...
This thesis presents several studies in neural dependency parsing for typologically diverse language...
Each year the Conference on Com-putational Natural Language Learning (CoNLL)1 features a shared task...
© 2019 Association for Computational Linguistics This paper explores the task of leveraging typology...
In this thesis, we investigate the use of latent variables to model complex dependencies in natural ...
accepted to appear in the special issue on Cross-Language Algorithms and ApplicationsPeer reviewe
How cross-linguistically applicable are NLP models, specifically language models? A fair comparison ...
The Conference on Computational Natural Language Learning features a shared task, in which participa...
International audienceThis paper describes how a tokenizer can be trained from any dataset in the Un...
Multilingual dependency parsing encapsulates any attempt to parse multiple languages. It can involve...
As more and more syntactically-annotated corpora become available for a wide variety of languages, m...
Pretrained multilingual contextual representations have shown great success, but due to the limits o...
The Conference on Computational Natural Language Learning (CoNLL) features a shared task, in which p...
Scaling existing applications and solutions to multiple human languages has traditionally proven to ...