Cross-lingual models trained on source language tasks possess the capability to directly transfer to target languages. However, since word order variances generally exist in different languages, cross-lingual models that overfit into the word order of the source language could have sub-optimal performance in target languages. In this paper, we hypothesize that reducing the word order information fitted into the models can improve the adaptation performance in target languages. To verify this hypothesis, we introduce several methods to make models encode less word order information of the source language and test them based on cross-lingual word embeddings and the pre-trained multilingual model. Experimental results on three sequence labelin...
Languages differ from one another and must therefore be learned. Processing biases in word order can...
Word order is one of the most readily observed aspects of the syntax of human language. This thesis ...
The universal properties of human languages have been the subject of intense study across the langua...
International audienceKnowledge transfer between neural language models is a widely used technique t...
Pre-trained multilingual language models show significant performance gains for zero-shot cross-ling...
Large pre-trained multilingual models such as mBERT and XLM-R enabled effective cross-lingual zero-s...
Extending semantic parsing systems to new domains and languages is a highly expensive, time-consumin...
We compare several language models for the word-ordering task and propose a new bag- to-sequence neu...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Although multilingual pretrained models (mPLMs) enabled support of various natural language processi...
Contemporary approaches to natural language processing are predominantly based on statistical machin...
It has been established that incorporating word cluster features derived from large unlabeled corpor...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Scaling existing applications and solutions to multiple human languages has traditionally proven to ...
Languages differ from one another and must therefore be learned. Processing biases in word order can...
Word order is one of the most readily observed aspects of the syntax of human language. This thesis ...
The universal properties of human languages have been the subject of intense study across the langua...
International audienceKnowledge transfer between neural language models is a widely used technique t...
Pre-trained multilingual language models show significant performance gains for zero-shot cross-ling...
Large pre-trained multilingual models such as mBERT and XLM-R enabled effective cross-lingual zero-s...
Extending semantic parsing systems to new domains and languages is a highly expensive, time-consumin...
We compare several language models for the word-ordering task and propose a new bag- to-sequence neu...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Although multilingual pretrained models (mPLMs) enabled support of various natural language processi...
Contemporary approaches to natural language processing are predominantly based on statistical machin...
It has been established that incorporating word cluster features derived from large unlabeled corpor...
Traditional approaches to supervised learning require a generous amount of labeled data for good gen...
Scaling existing applications and solutions to multiple human languages has traditionally proven to ...
Languages differ from one another and must therefore be learned. Processing biases in word order can...
Word order is one of the most readily observed aspects of the syntax of human language. This thesis ...
The universal properties of human languages have been the subject of intense study across the langua...