Neural network language models (NNLMs) have attracted a lot of attention recently. In this paper, we present a training method that can incrementally train the hierarchical softmax function for NNMLs. We split the cost function to model old and update corpora separately, and factorize the objective function for the hierarchical softmax. Then we provide a new stochastic gradient based method to update all the word vectors and parameters, by comparing the old tree generated based on the old corpus and the new tree generated based on the combined (old and update) corpus. Theoretical analysis shows that the mean square error of the parameter vectors can be bounded by a function of the number of changed words related to the parameter node. Exper...
ABSTRACT We present several modifications of the original recurrent neural network language model (R...
In recent years neural language models (LMs) have set state-of-the-art performance for several bench...
Conventional Neural Network (NN) training is done by introducing training patterns in the full input...
This paper presents a new method to reduce the computational cost when using Neural Networks as Lang...
[EN] This paper presents a new method to reduce the computational cost when using Neural Networks as...
Recent research has pointed to a limitation of word-level neural language models with softmax output...
It is today acknowledged that neural network language models outperform backoff language models in a...
International audienceDeep neural networks of sizes commonly encountered in practice are proven to c...
In spite of their superior performance, neural probabilistic language models (NPLMs) remain far less...
This work adresses the topic of neural language model acceleration. The aim of this work is to optim...
Virtually any modern speech recognition system relies on count-based language models. In this thesis...
For resource rich languages, recent works have shown Neural Network based Language Models (NNLMs) to...
In this paper, based on an asymptotic analysis of the Softmax layer, we show that when training neur...
This paper presents a framework in which hierarchical softmax is used to create a global hierarchica...
Neural network language models (NNLMs) have achieved ever-improving accuracy due to more sophisticat...
ABSTRACT We present several modifications of the original recurrent neural network language model (R...
In recent years neural language models (LMs) have set state-of-the-art performance for several bench...
Conventional Neural Network (NN) training is done by introducing training patterns in the full input...
This paper presents a new method to reduce the computational cost when using Neural Networks as Lang...
[EN] This paper presents a new method to reduce the computational cost when using Neural Networks as...
Recent research has pointed to a limitation of word-level neural language models with softmax output...
It is today acknowledged that neural network language models outperform backoff language models in a...
International audienceDeep neural networks of sizes commonly encountered in practice are proven to c...
In spite of their superior performance, neural probabilistic language models (NPLMs) remain far less...
This work adresses the topic of neural language model acceleration. The aim of this work is to optim...
Virtually any modern speech recognition system relies on count-based language models. In this thesis...
For resource rich languages, recent works have shown Neural Network based Language Models (NNLMs) to...
In this paper, based on an asymptotic analysis of the Softmax layer, we show that when training neur...
This paper presents a framework in which hierarchical softmax is used to create a global hierarchica...
Neural network language models (NNLMs) have achieved ever-improving accuracy due to more sophisticat...
ABSTRACT We present several modifications of the original recurrent neural network language model (R...
In recent years neural language models (LMs) have set state-of-the-art performance for several bench...
Conventional Neural Network (NN) training is done by introducing training patterns in the full input...