It has been shown for English that discrete and soft prompting perform strongly in fewshot learning with pretrained language models (PLMs). In this paper, we show that discrete and soft prompting perform better than finetuning in multilingual cases: Crosslingual transfer and in-language training of multilingual natural language inference. For example, with 48 English training examples, finetuning obtains 33.74% accuracy in crosslingual transfer, barely surpassing the majority baseline (33.33%). In contrast, discrete and soft prompting outperform finetuning, achieving 36.43% and 38.79%. We also demonstrate good performance of prompting with training data in multiple languages other than English
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
This paper aims for a potential architectural improvement for multilingual learning and asks: Can di...
Pretrained multilingual language models have become a common tool in transferring NLP capabilities t...
It has been shown for English that discrete and soft prompting perform strongly in fewshot learning ...
Large-scale generative language models such as GPT-3 are competitive few-shot learners. While these ...
Pre-trained multilingual language models show significant performance gains for zero-shot cross-ling...
Prompt-based tuning has been proven effective for pretrained language models (PLMs). While most of t...
Supervised deep learning-based approaches have been applied to task-oriented dialog and have proven ...
Text classification is one of the most imperative tasks in natural language processing (NLP). Recent...
Recently, there has been an increasing interest in models that generate natural language explanation...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
Few-shot crosslingual transfer has been shown to outperform its zero-shot counterpart with pretraine...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Neural Machine Translation has been shown to enable in-ference and cross-lingual knowledge transfer ...
Multilingual language models are widely used to extend NLP systems to low-resource languages. Howeve...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
This paper aims for a potential architectural improvement for multilingual learning and asks: Can di...
Pretrained multilingual language models have become a common tool in transferring NLP capabilities t...
It has been shown for English that discrete and soft prompting perform strongly in fewshot learning ...
Large-scale generative language models such as GPT-3 are competitive few-shot learners. While these ...
Pre-trained multilingual language models show significant performance gains for zero-shot cross-ling...
Prompt-based tuning has been proven effective for pretrained language models (PLMs). While most of t...
Supervised deep learning-based approaches have been applied to task-oriented dialog and have proven ...
Text classification is one of the most imperative tasks in natural language processing (NLP). Recent...
Recently, there has been an increasing interest in models that generate natural language explanation...
Cross-lingual transfer learning with large multilingual pre-trained models can be an effective appro...
Few-shot crosslingual transfer has been shown to outperform its zero-shot counterpart with pretraine...
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-sh...
Neural Machine Translation has been shown to enable in-ference and cross-lingual knowledge transfer ...
Multilingual language models are widely used to extend NLP systems to low-resource languages. Howeve...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
This paper aims for a potential architectural improvement for multilingual learning and asks: Can di...
Pretrained multilingual language models have become a common tool in transferring NLP capabilities t...