To mitigate the problem of having to traverse over the full vocabulary in the softmax normalization of a neural language model, sampling-based training criteria are proposed and investigated in the context of large vocabulary word-based neural language models. These training criteria typically enjoy the benefit of faster training and testing, at a cost of slightly degraded performance in terms of perplexity and almost no visible drop in word error rate. While noise contrastive estimation is one of the most popular choices, recently we show that other sampling-based criteria can also perform well, as long as an extra correction step is done, where the intended class posterior probability is recovered from the raw model outputs. In this work,...
Self-supervised speech recognition models require considerable labeled training data for learning hi...
In recent years recurrent neural network language models (RNNLMs) have been successfully applied to ...
Language Models (LMs) pre-trained with self-supervision on large text corpora have become the defaul...
A language model is a vital component of automatic speech recognition systems. In recent years, adva...
Neural language models do not scale well when the vocabulary is large. Noise contrastive estimation ...
This work investigates practical methods to ease training and improve performances of neural languag...
In spite of their superior performance, neural probabilistic language models (NPLMs) remain far less...
Neural language models do not scale well when the vocabulary is large. Noise-contrastive estimation ...
In recent years recurrent neural network language models (RNNLMs) have been successfully applied to ...
Long samples of text from neural language models can be of poor quality. Truncation sampling algorit...
Self-supervised representation learning (SSRL) has improved the performance on downstream phoneme re...
The development of neural vocoders (NVs) has resulted in the high-quality and fast generation of wav...
Although the word-popularity based negative sampler has shown superb performance in the skip-gram mo...
Large speech emotion recognition datasets are hard to obtain, and small datasets may contain biases....
Language model fine-tuning is essential for modern natural language processing, but is computational...
Self-supervised speech recognition models require considerable labeled training data for learning hi...
In recent years recurrent neural network language models (RNNLMs) have been successfully applied to ...
Language Models (LMs) pre-trained with self-supervision on large text corpora have become the defaul...
A language model is a vital component of automatic speech recognition systems. In recent years, adva...
Neural language models do not scale well when the vocabulary is large. Noise contrastive estimation ...
This work investigates practical methods to ease training and improve performances of neural languag...
In spite of their superior performance, neural probabilistic language models (NPLMs) remain far less...
Neural language models do not scale well when the vocabulary is large. Noise-contrastive estimation ...
In recent years recurrent neural network language models (RNNLMs) have been successfully applied to ...
Long samples of text from neural language models can be of poor quality. Truncation sampling algorit...
Self-supervised representation learning (SSRL) has improved the performance on downstream phoneme re...
The development of neural vocoders (NVs) has resulted in the high-quality and fast generation of wav...
Although the word-popularity based negative sampler has shown superb performance in the skip-gram mo...
Large speech emotion recognition datasets are hard to obtain, and small datasets may contain biases....
Language model fine-tuning is essential for modern natural language processing, but is computational...
Self-supervised speech recognition models require considerable labeled training data for learning hi...
In recent years recurrent neural network language models (RNNLMs) have been successfully applied to ...
Language Models (LMs) pre-trained with self-supervision on large text corpora have become the defaul...