Structure prediction (SP) tasks are important in natural language understanding in the sense that they provide complex and structured knowledge of the text. Recently, some unified text-to-text transformer models like T5 and TANL have produced competitive results on SP tasks. These models convert SP tasks into a seq2seq problem, where a transformer is used to generate sequences with special tokens representing the extracted spans, labels, and relationships. Compared to many popular Natural Language Understanding models that are designed specifically for the task, the output of the text-to-text transformer is more flexible. With proper format, it could be trained on multiple tasks together and take advantage of the shared knowledge between t...
Unsupervised learning text representations aims at converting natural languages into vector represen...
This article describes our experiments in neural machine translation using the recent Tensor2Tensor ...
In this paper, we take the advantage of previous pre-trained models (PTMs) and propose a novel Chine...
The Transformer model is a very recent, fast and powerful discovery in neural machine translation. W...
The utility of linguistic annotation in neural machine translation seemed to had been established in...
Transformer networks have seen great success in natural language processing and machine vision, wher...
Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on...
Thesis (Master's)--University of Washington, 2021Transformer models perform well on NLP tasks, but r...
Some Transformer-based models can perform cross-lingual transfer learning: those models can be train...
We introduce a method for improving the structural understanding abilities of language models. Unlik...
Multi-Task Learning (MTL) models have shown their robustness, effectiveness, and efficiency for tran...
Neural Machine Translation (NMT) is notorious for its need for large amounts ofbilingual data. An ef...
Task-conditional architecture offers advantage in parameter efficiency but falls short in performanc...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Text classification is one of the most imperative tasks in natural language processing (NLP). Recent...
Unsupervised learning text representations aims at converting natural languages into vector represen...
This article describes our experiments in neural machine translation using the recent Tensor2Tensor ...
In this paper, we take the advantage of previous pre-trained models (PTMs) and propose a novel Chine...
The Transformer model is a very recent, fast and powerful discovery in neural machine translation. W...
The utility of linguistic annotation in neural machine translation seemed to had been established in...
Transformer networks have seen great success in natural language processing and machine vision, wher...
Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on...
Thesis (Master's)--University of Washington, 2021Transformer models perform well on NLP tasks, but r...
Some Transformer-based models can perform cross-lingual transfer learning: those models can be train...
We introduce a method for improving the structural understanding abilities of language models. Unlik...
Multi-Task Learning (MTL) models have shown their robustness, effectiveness, and efficiency for tran...
Neural Machine Translation (NMT) is notorious for its need for large amounts ofbilingual data. An ef...
Task-conditional architecture offers advantage in parameter efficiency but falls short in performanc...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
Text classification is one of the most imperative tasks in natural language processing (NLP). Recent...
Unsupervised learning text representations aims at converting natural languages into vector represen...
This article describes our experiments in neural machine translation using the recent Tensor2Tensor ...
In this paper, we take the advantage of previous pre-trained models (PTMs) and propose a novel Chine...