Large-scale cross-lingual language models (LM), such as mBERT, Unicoder and XLM, have achieved great success in cross-lingual representation learning. However, when applied to zero-shot cross-lingual transfer tasks, most existing methods use only single-language input for LM finetuning, without leveraging the intrinsic cross-lingual alignment between different languages that proves essential for multilingual tasks. In this paper, we propose FILTER, an enhanced fusion method that takes cross-lingual data as input for XLM finetuning. Specifically, FILTER first encodes text input in the source language and its translation in the target language independently in the shallow layers, then performs cross-language fusion to extract multilingual kno...
Cross-lingual Machine Reading Comprehension (xMRC) is a challenging task due to the lack of training...
With the increasingly global nature of our everyday interactions, the need for multilingual technolo...
Accepted at EACL 2021Multilingual pretrained language models have demonstrated remarkable zero-shot ...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
The lack of publicly available evaluation data for low-resource languages limits progress in Spoken ...
Cross-lingual pre-training has achieved great successes using monolingual and bilingual plain text c...
Large pre-trained multilingual models such as mBERT and XLM-R enabled effective cross-lingual zero-s...
Language model pre-training has achieved success in many natural language processing tasks. Existing...
Prompt-based tuning has been proven effective for pretrained language models (PLMs). While most of t...
∗ Both authors contributed equally Cross-language learning allows one to use training data from one ...
International audienceSome Transformer-based models can perform crosslingual transfer learning: thos...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
In cross-lingual language understanding, machine translation is often utilized to enhance the transf...
Thesis (Master's)--University of Washington, 2020This work presents methods for learning cross-lingu...
Cross-lingual Machine Reading Comprehension (xMRC) is a challenging task due to the lack of training...
With the increasingly global nature of our everyday interactions, the need for multilingual technolo...
Accepted at EACL 2021Multilingual pretrained language models have demonstrated remarkable zero-shot ...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
The lack of publicly available evaluation data for low-resource languages limits progress in Spoken ...
Cross-lingual pre-training has achieved great successes using monolingual and bilingual plain text c...
Large pre-trained multilingual models such as mBERT and XLM-R enabled effective cross-lingual zero-s...
Language model pre-training has achieved success in many natural language processing tasks. Existing...
Prompt-based tuning has been proven effective for pretrained language models (PLMs). While most of t...
∗ Both authors contributed equally Cross-language learning allows one to use training data from one ...
International audienceSome Transformer-based models can perform crosslingual transfer learning: thos...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
In cross-lingual language understanding, machine translation is often utilized to enhance the transf...
Thesis (Master's)--University of Washington, 2020This work presents methods for learning cross-lingu...
Cross-lingual Machine Reading Comprehension (xMRC) is a challenging task due to the lack of training...
With the increasingly global nature of our everyday interactions, the need for multilingual technolo...
Accepted at EACL 2021Multilingual pretrained language models have demonstrated remarkable zero-shot ...