[EN] This paper presents a new method to reduce the computational cost when using Neural Networks as Language Models, during recognition, in some particular scenarios. It is based on a Neural Network that considers input contexts of different length in order to ease the use of a fallback mechanism together with the precomputation of softmax normalization constants for these inputs. The proposed approach is empirically validated, showing their capability to emulate lower order N-grams with a single Neural Network. A machine translation task shows that the proposed model constitutes a good solution to the normalization cost of the output softmax layer of Neural Networks, for some practical cases, without a significant impact in performance wh...
Neural machine translation (NMT) has been shown to outperform statistical machine translation. Howe...
Recent work has shown success in us-ing neural network language models (NNLMs) as features in MT sys...
This paper investigates very low resource language model pretraining, when less than 100 thousand se...
This paper presents a new method to reduce the computational cost when using Neural Networks as Lang...
[EN] Neural Network Language Models (NNLMs) are a successful approach to Natural Language Processing...
Neural network language models (NNLMs) have attracted a lot of attention recently. In this paper, we...
The quality of translations produced by statistical machine translation (SMT) systems crucially depe...
<p>For resource rich languages, recent works have shown Neural Network based Language Models (NNLMs)...
In spite of their superior performance, neural probabilistic language models (NPLMs) remain far less...
Virtually any modern speech recognition system relies on count-based language models. In this thesis...
It is today acknowledged that neural network language models outperform backoff language models in a...
In order for neural networks to learn complex languages or grammars, they must have sufficient comp...
Neural language models do not scale well when the vocabulary is large. Noise contrastive estimation ...
For resource rich languages, recent works have shown Neu-ral Network based Language Models (NNLMs) t...
Recent work has shown success in us-ing neural network language models (NNLMs) as features in MT sys...
Neural machine translation (NMT) has been shown to outperform statistical machine translation. Howe...
Recent work has shown success in us-ing neural network language models (NNLMs) as features in MT sys...
This paper investigates very low resource language model pretraining, when less than 100 thousand se...
This paper presents a new method to reduce the computational cost when using Neural Networks as Lang...
[EN] Neural Network Language Models (NNLMs) are a successful approach to Natural Language Processing...
Neural network language models (NNLMs) have attracted a lot of attention recently. In this paper, we...
The quality of translations produced by statistical machine translation (SMT) systems crucially depe...
<p>For resource rich languages, recent works have shown Neural Network based Language Models (NNLMs)...
In spite of their superior performance, neural probabilistic language models (NPLMs) remain far less...
Virtually any modern speech recognition system relies on count-based language models. In this thesis...
It is today acknowledged that neural network language models outperform backoff language models in a...
In order for neural networks to learn complex languages or grammars, they must have sufficient comp...
Neural language models do not scale well when the vocabulary is large. Noise contrastive estimation ...
For resource rich languages, recent works have shown Neu-ral Network based Language Models (NNLMs) t...
Recent work has shown success in us-ing neural network language models (NNLMs) as features in MT sys...
Neural machine translation (NMT) has been shown to outperform statistical machine translation. Howe...
Recent work has shown success in us-ing neural network language models (NNLMs) as features in MT sys...
This paper investigates very low resource language model pretraining, when less than 100 thousand se...