In sequence-to-sequence tasks, sentences with heterogeneous semantics or grammatical structures may lead to some difficulties in the model's convergence. To resolve this problem, we introduce a feature-concentrated model that focuses on each of the heterogeneous features in the input-output sequences. Building upon the encoder-decoder architecture, we design a latent-enhanced multi-filter sequence-to-sequence model (LMS2S) that analyzes the features preserved by latent space representations and constructs the outputs accordingly. We divide the latent space into subspaces using a clustering algorithm and train a set of decoders in which each decoder only concentrates on the features from its corresponding subspace. We then design a self-enha...
The field of Natural Language Processing has experienced a dramatic leap in capabilities with the re...
International audienceCurrent state-of-the-art machine translation systems are based on encoder-deco...
Most machine learning algorithms require a fixed length input to be able to perform commonly desired...
In Natural Language Processing (NLP), it is important to detect the relationship between two sequenc...
L'apprentissage profond a permis des avancées significatives dans le domaine de la traduction automa...
In encoder-decoder based sequence-to-sequence modeling, the most common practice is to stack a numbe...
Deep (recurrent) neural networks has been shown to successfully learn complex mappings between arbit...
Sibley, Kello, Plaut, and Elman (2008) proposed the sequence encoder as a model that learns fixed-wi...
We investigate input-conditioned hypernetworks for multi-tasking in NLP, generating parameter-effici...
Neural sequence models have been applied with great success to a variety of tasks in natural languag...
Recently, sequence-to-sequence models have achieved impressive performance on a number of semantic p...
State-of-the-art encoder-decoder models (e.g. for machine translation (MT) or speech recognition (AS...
The sequence memoizer is a model for sequence data with state-of-the-art per-formance on language mo...
Efficiently capturing the long-range patterns in sequential data sources salient to a given task -- ...
We present a novel semi-supervised approach for sequence transduction and apply it to semantic parsi...
The field of Natural Language Processing has experienced a dramatic leap in capabilities with the re...
International audienceCurrent state-of-the-art machine translation systems are based on encoder-deco...
Most machine learning algorithms require a fixed length input to be able to perform commonly desired...
In Natural Language Processing (NLP), it is important to detect the relationship between two sequenc...
L'apprentissage profond a permis des avancées significatives dans le domaine de la traduction automa...
In encoder-decoder based sequence-to-sequence modeling, the most common practice is to stack a numbe...
Deep (recurrent) neural networks has been shown to successfully learn complex mappings between arbit...
Sibley, Kello, Plaut, and Elman (2008) proposed the sequence encoder as a model that learns fixed-wi...
We investigate input-conditioned hypernetworks for multi-tasking in NLP, generating parameter-effici...
Neural sequence models have been applied with great success to a variety of tasks in natural languag...
Recently, sequence-to-sequence models have achieved impressive performance on a number of semantic p...
State-of-the-art encoder-decoder models (e.g. for machine translation (MT) or speech recognition (AS...
The sequence memoizer is a model for sequence data with state-of-the-art per-formance on language mo...
Efficiently capturing the long-range patterns in sequential data sources salient to a given task -- ...
We present a novel semi-supervised approach for sequence transduction and apply it to semantic parsi...
The field of Natural Language Processing has experienced a dramatic leap in capabilities with the re...
International audienceCurrent state-of-the-art machine translation systems are based on encoder-deco...
Most machine learning algorithms require a fixed length input to be able to perform commonly desired...