Neural networks have been shown to improve performance across a range of natural-language tasks. However, design-ing and training them can be complicated. Frequently, researchers resort to repeated experimentation to pick optimal settings. In this paper, we address the issue of choosing the correct number of units in hidden layers. We introduce a method for automatically adjusting network size by pruning out hidden units through `∞,1 and `2,1 regularization. We apply this method to language modeling and demonstrate its ability to correctly choose the number of hidden units while maintaining perplexity. We also include these models in a machine translation decoder and show that these smaller neural models maintain the signif-icant improvemen...
Neural machine translation often suffers from an under-translation problem owing to its limited mode...
This paper aims to compare different regularization strategies to address a common phenomenon, sever...
[EN] Neural Network Language Models (NNLMs) are a successful approach to Natural Language Processing...
We explore the roles and interactions of the hyper-parameters governing regularization, and propose ...
Virtually any modern speech recognition system relies on count-based language models. In this thesis...
The quality of translations produced by statistical machine translation (SMT) systems crucially depe...
Currently, N-gram models are the most common and widely used models for statistical language modelin...
We explore the application of neural language models to machine translation. We develop a new model ...
Currently, N-gram models are the most common and widely used models for statistical language modelin...
We explore the application of neural language models to machine translation. We develop a new model ...
Deep neural language models like GPT-2 is undoubtedly strong at text generation, but often requires ...
Ensembling is a well-known technique in neural machine translation (NMT) to improve system performan...
Forecasting, classification, and data analysis may all gain from improved pattern recognition result...
For resource rich languages, recent works have shown Neural Network based Language Models (NNLMs) to...
This paper presents a new method to reduce the computational cost when using Neural Networks as Lang...
Neural machine translation often suffers from an under-translation problem owing to its limited mode...
This paper aims to compare different regularization strategies to address a common phenomenon, sever...
[EN] Neural Network Language Models (NNLMs) are a successful approach to Natural Language Processing...
We explore the roles and interactions of the hyper-parameters governing regularization, and propose ...
Virtually any modern speech recognition system relies on count-based language models. In this thesis...
The quality of translations produced by statistical machine translation (SMT) systems crucially depe...
Currently, N-gram models are the most common and widely used models for statistical language modelin...
We explore the application of neural language models to machine translation. We develop a new model ...
Currently, N-gram models are the most common and widely used models for statistical language modelin...
We explore the application of neural language models to machine translation. We develop a new model ...
Deep neural language models like GPT-2 is undoubtedly strong at text generation, but often requires ...
Ensembling is a well-known technique in neural machine translation (NMT) to improve system performan...
Forecasting, classification, and data analysis may all gain from improved pattern recognition result...
For resource rich languages, recent works have shown Neural Network based Language Models (NNLMs) to...
This paper presents a new method to reduce the computational cost when using Neural Networks as Lang...
Neural machine translation often suffers from an under-translation problem owing to its limited mode...
This paper aims to compare different regularization strategies to address a common phenomenon, sever...
[EN] Neural Network Language Models (NNLMs) are a successful approach to Natural Language Processing...