Traditional text classification approaches often require a good amount of labeled data, which is difficult to obtain, especially in restricted domains or less widespread languages. This lack of labeled data has led to the rise of low-resource methods, that assume low data availability in natural language processing. Among them, zero-shot learning stands out, which consists of learning a classifier without any previously labeled data. The best results reported with this approach use language models such as Transformers, but fall into two problems: high execution time and inability to handle long texts as input. This paper proposes a new model, ZeroBERTo, which leverages an unsupervised clustering step to obtain a compressed data representati...
Abstract—Object recognition systems usually require fully com-plete manually labeled training data t...
Within a situation where Semi-Supervised Learning (SSL) is available to exploit unlabeled data, this...
Humans can obtain the knowledge of novel visual concepts from language descriptions, and we thus use...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
Nowadays, owing to the superior capacity of the large pre-trained language models (PLM), the PLM-bas...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
There is a growing interest in dataset generation recently due to the superior generative capacity o...
Existing solutions to zero-shot text classification either conduct prompting with pre-trained langu...
Our research focuses on solving the zero-shot text classification problem in NLP, with a particular ...
Existing Zero-Shot Learning (ZSL) techniques for text classification typically assign a label to a p...
We propose a novel framework for zero-shot learning of topic-dependent language models, which enable...
Classifying a visual concept merely from its associated online textual source, such as a Wikipedia a...
Title from PDF of title page viewed November 5, 2020Thesis advisor: Yugyung LeeVitaIncludes bibliogr...
This paper presents the findings of the LoResMT 2020 Shared Task on zero-shot translation for low re...
Pretrained language models have shown success in various areas of natural language processing, inclu...
Abstract—Object recognition systems usually require fully com-plete manually labeled training data t...
Within a situation where Semi-Supervised Learning (SSL) is available to exploit unlabeled data, this...
Humans can obtain the knowledge of novel visual concepts from language descriptions, and we thus use...
In recent years, the community of natural language processing (NLP) has seen amazing progress in the...
Nowadays, owing to the superior capacity of the large pre-trained language models (PLM), the PLM-bas...
Pretrained language models (PLMs) have demonstrated remarkable performance in various natural langua...
There is a growing interest in dataset generation recently due to the superior generative capacity o...
Existing solutions to zero-shot text classification either conduct prompting with pre-trained langu...
Our research focuses on solving the zero-shot text classification problem in NLP, with a particular ...
Existing Zero-Shot Learning (ZSL) techniques for text classification typically assign a label to a p...
We propose a novel framework for zero-shot learning of topic-dependent language models, which enable...
Classifying a visual concept merely from its associated online textual source, such as a Wikipedia a...
Title from PDF of title page viewed November 5, 2020Thesis advisor: Yugyung LeeVitaIncludes bibliogr...
This paper presents the findings of the LoResMT 2020 Shared Task on zero-shot translation for low re...
Pretrained language models have shown success in various areas of natural language processing, inclu...
Abstract—Object recognition systems usually require fully com-plete manually labeled training data t...
Within a situation where Semi-Supervised Learning (SSL) is available to exploit unlabeled data, this...
Humans can obtain the knowledge of novel visual concepts from language descriptions, and we thus use...