Pretrained, large, generative language models (LMs) have had great success in a wide range of sequence tagging and structured prediction tasks. Casting a sequence tagging task as a Seq2Seq one requires deciding the formats of the input and output sequences. However, we lack a principled understanding of the trade-offs associated with these formats (such as the effect on model accuracy, sequence length, multilingual generalization, hallucination). In this paper, we rigorously study different formats one could use for casting input text sentences and their output labels into the input and target (i.e., output) of a Seq2Seq model. Along the way, we introduce a new format, which we show to not only be simpler but also more effective. Additional...
In sequence-to-sequence tasks, sentences with heterogeneous semantics or grammatical structures may ...
Sequence to sequence (SEQ2SEQ) models often lack diversity in their generated translations. This can...
We propose a sequence labeling framework with a secondary training objective, learning to predict su...
In Natural Language Processing (NLP), it is important to detect the relationship between two sequenc...
The task of Grammatical Error Correction (GEC) has received remarkable attention with wide applicati...
In this work, we investigated the recent sequence tagging approach for the Grammatical Error Correc...
Some machine learning tasks have a complex output, rather than a real number or a class. Those outpu...
The task of predicting dialog acts (DA) based on conversational dialog is a key component in the dev...
Cross-lingual models trained on source language tasks possess the capability to directly transfer to...
Sequence Tagging, including part of speech tagging, chunking and named entity recognition, is an imp...
With the advent of deep learning, research in many areas of machine learning is converging towards t...
Neural sequence-to-sequence models are finding increasing use in editing of documents, for example i...
Article dans revue scientifique avec comité de lecture.In natural language and especially in spontan...
In this study, we evaluate and compare state-of-the-art models on the code generation and code summa...
31st AAAI Conference on Artificial Intelligence, AAAI 2017, San Francisco, CA. USA, 4-10 February 20...
In sequence-to-sequence tasks, sentences with heterogeneous semantics or grammatical structures may ...
Sequence to sequence (SEQ2SEQ) models often lack diversity in their generated translations. This can...
We propose a sequence labeling framework with a secondary training objective, learning to predict su...
In Natural Language Processing (NLP), it is important to detect the relationship between two sequenc...
The task of Grammatical Error Correction (GEC) has received remarkable attention with wide applicati...
In this work, we investigated the recent sequence tagging approach for the Grammatical Error Correc...
Some machine learning tasks have a complex output, rather than a real number or a class. Those outpu...
The task of predicting dialog acts (DA) based on conversational dialog is a key component in the dev...
Cross-lingual models trained on source language tasks possess the capability to directly transfer to...
Sequence Tagging, including part of speech tagging, chunking and named entity recognition, is an imp...
With the advent of deep learning, research in many areas of machine learning is converging towards t...
Neural sequence-to-sequence models are finding increasing use in editing of documents, for example i...
Article dans revue scientifique avec comité de lecture.In natural language and especially in spontan...
In this study, we evaluate and compare state-of-the-art models on the code generation and code summa...
31st AAAI Conference on Artificial Intelligence, AAAI 2017, San Francisco, CA. USA, 4-10 February 20...
In sequence-to-sequence tasks, sentences with heterogeneous semantics or grammatical structures may ...
Sequence to sequence (SEQ2SEQ) models often lack diversity in their generated translations. This can...
We propose a sequence labeling framework with a secondary training objective, learning to predict su...