The T5 model and its unified text-to-text paradigm contributed in advancing the state-of-the-art for many natural language processing tasks. While some multilingual variants of the T5 model have recently been introduced, their performances were found to provide suboptimal performances for languages other than English if compared to monolingual variants. We are motivated by these findings to introduce IT5, the first family of encoder-decoder transformer models pretrained specifically on Italian. We perform a thorough cleaning of a web-crawled Italian corpus including more than 40 billion words and use it to pretrain three IT5 models of different sizes. The performance of IT5 models and their multilingual counterparts is then evaluated on a b...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...
The T5 model and its unified text-to-text paradigm contributed in advancing the state-of-the-art for...
The T5 model and its unified text-to-text paradigm contributed in advancing the state-of-the-art for...
The T5 model and its unified text-to-text paradigm contributed in advancing the state-of-the-art for...
The use of contextualised word embeddings allowed for a relevant performance increase for almost all...
The use of contextualised word embeddings allowed for a relevant performance increase for almost all...
The emergence of attention-based architectures has led to significant improvements in the performanc...
In this paper, we present an in-depth investigation of the linguistic knowledge encoded by the trans...
In this paper, we present an in-depth investigation of the linguistic knowledge encoded by the trans...
In this paper, we present an in-depth investigation of the linguistic knowledge encoded by the trans...
In this paper, we present an in-depth investigation of the linguistic knowledge encoded by the trans...
The use of contextualised word embeddings allowed for a relevant performance increase for almost all...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...
The T5 model and its unified text-to-text paradigm contributed in advancing the state-of-the-art for...
The T5 model and its unified text-to-text paradigm contributed in advancing the state-of-the-art for...
The T5 model and its unified text-to-text paradigm contributed in advancing the state-of-the-art for...
The use of contextualised word embeddings allowed for a relevant performance increase for almost all...
The use of contextualised word embeddings allowed for a relevant performance increase for almost all...
The emergence of attention-based architectures has led to significant improvements in the performanc...
In this paper, we present an in-depth investigation of the linguistic knowledge encoded by the trans...
In this paper, we present an in-depth investigation of the linguistic knowledge encoded by the trans...
In this paper, we present an in-depth investigation of the linguistic knowledge encoded by the trans...
In this paper, we present an in-depth investigation of the linguistic knowledge encoded by the trans...
The use of contextualised word embeddings allowed for a relevant performance increase for almost all...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...
End-to-end deep learning models have pushed forward significantly many tasks of Natural Language Pro...