Traditional approaches to supervised learning require a generous amount of labeled data for good generalization. While such annotation-heavy approaches have proven useful for some Natural Language Processing (NLP) tasks in high-resource languages (like English), they are unlikely to scale to languages where collecting labeled data is di cult and time-consuming. Translating supervision available in English is also not a viable solution, because developing a good machine translation system requires expensive to annotate resources which are not available for most languages. In this thesis, I argue that cross-lingual representations are an effective means of extending NLP tools to languages beyond English without resorting to generous amounts o...
Jiawei ZhaoCurrent machine translation techniques were developed using predominantly rich resource l...
Bilingual Word Embeddings (BWEs) are one of the cornerstones of cross-lingual transfer of NLP models...
Large-scale cross-lingual language models (LM), such as mBERT, Unicoder and XLM, have achieved great...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
In the modern era of deep learning, developing natural language processing (NLP) systems require lar...
Thesis (Ph.D.)--University of Washington, 2022Modern NLP systems have been highly successful at a wi...
Abstract: The subject area of multilingual natural language processing (NLP) is concerned with the p...
NLP systems typically require support for more than one language. As different languages have differ...
State-of-the-art approaches to most Natural Language Processing (NLP) tasks have achieved near huma...
Addressing the cross-lingual variation of grammatical structures and meaning categorization is a key...
Cross-lingual Learning can help to bring state-of-the-art deep learning solutions to smaller languag...
Large-scale annotated datasets are an indispensable ingredient of modern Natural Language Processing...
This article addresses the question of how to deal with text categorization when the set of document...
In this paper, we explore a multilingual translation model with a cross-lingually shared layer that ...
Jiawei ZhaoCurrent machine translation techniques were developed using predominantly rich resource l...
Bilingual Word Embeddings (BWEs) are one of the cornerstones of cross-lingual transfer of NLP models...
Large-scale cross-lingual language models (LM), such as mBERT, Unicoder and XLM, have achieved great...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
In the modern era of deep learning, developing natural language processing (NLP) systems require lar...
Thesis (Ph.D.)--University of Washington, 2022Modern NLP systems have been highly successful at a wi...
Abstract: The subject area of multilingual natural language processing (NLP) is concerned with the p...
NLP systems typically require support for more than one language. As different languages have differ...
State-of-the-art approaches to most Natural Language Processing (NLP) tasks have achieved near huma...
Addressing the cross-lingual variation of grammatical structures and meaning categorization is a key...
Cross-lingual Learning can help to bring state-of-the-art deep learning solutions to smaller languag...
Large-scale annotated datasets are an indispensable ingredient of modern Natural Language Processing...
This article addresses the question of how to deal with text categorization when the set of document...
In this paper, we explore a multilingual translation model with a cross-lingually shared layer that ...
Jiawei ZhaoCurrent machine translation techniques were developed using predominantly rich resource l...
Bilingual Word Embeddings (BWEs) are one of the cornerstones of cross-lingual transfer of NLP models...
Large-scale cross-lingual language models (LM), such as mBERT, Unicoder and XLM, have achieved great...