In this thesis, we study novel neural network structures to better model long term dependency in sequential data. We propose to use more memory units to keep track of more preceding states in recurrent neural networks (RNNs), which are all recurrently fed to the hidden layers as feedback through different weighted paths. By extending the popular recurrent structure in RNNs, we provide the models with better short-term memory mechanism to learn long term dependency in sequences. Analogous to digital filters in signal processing, we call these structures as higher order RNNs (HORNNs). Similar to RNNs, HORNNs can also be learned using the back-propagation through time method. HORNNs are generally applicable to a variety of sequence modellin...
We analyse Recurrent Neural Networks (RNNs) to understand the significance of multiple LSTM layers. ...
Recurrent Neural Networks (RNNs) continue to show outstanding performance in sequence modeling tasks...
Language modeling has been widely used in the application of natural language processing, and there...
Language acquisition is one of the core problems in artificial intelligence (AI) and it is generally...
Sequence processing involves several tasks such as clustering, classification, prediction, and trans...
Recurrent Neural Networks (RNNs) are a type of neural network that maintains a hidden state, preserv...
The vanishing gradients problem inherent in Simple Recurrent Networks (SRN) trained with back-propag...
ABSTRACT We present several modifications of the original recurrent neural network language model (R...
Recent work on language modelling has shifted focus from count-based models to neural models. In the...
Recurrent Neural Networks (RNNs) are theoretically Turing-complete and established themselves as a d...
Recurrent Neural Networks (RNNs) are powerful sequence models that were believed to be difficult to ...
Recurrent Neural Networks (RNNs) are powerful sequence models that were believed to be difficult to ...
Recurrent Neural Networks (RNNs) continue to show outstanding performance in sequence modeling tasks...
Recurrent neural networks (RNN) are efficient in modeling sequences for generation and classificatio...
The present work takes into account the compactness and efficiency of Recurrent Neural Networks (RNN...
We analyse Recurrent Neural Networks (RNNs) to understand the significance of multiple LSTM layers. ...
Recurrent Neural Networks (RNNs) continue to show outstanding performance in sequence modeling tasks...
Language modeling has been widely used in the application of natural language processing, and there...
Language acquisition is one of the core problems in artificial intelligence (AI) and it is generally...
Sequence processing involves several tasks such as clustering, classification, prediction, and trans...
Recurrent Neural Networks (RNNs) are a type of neural network that maintains a hidden state, preserv...
The vanishing gradients problem inherent in Simple Recurrent Networks (SRN) trained with back-propag...
ABSTRACT We present several modifications of the original recurrent neural network language model (R...
Recent work on language modelling has shifted focus from count-based models to neural models. In the...
Recurrent Neural Networks (RNNs) are theoretically Turing-complete and established themselves as a d...
Recurrent Neural Networks (RNNs) are powerful sequence models that were believed to be difficult to ...
Recurrent Neural Networks (RNNs) are powerful sequence models that were believed to be difficult to ...
Recurrent Neural Networks (RNNs) continue to show outstanding performance in sequence modeling tasks...
Recurrent neural networks (RNN) are efficient in modeling sequences for generation and classificatio...
The present work takes into account the compactness and efficiency of Recurrent Neural Networks (RNN...
We analyse Recurrent Neural Networks (RNNs) to understand the significance of multiple LSTM layers. ...
Recurrent Neural Networks (RNNs) continue to show outstanding performance in sequence modeling tasks...
Language modeling has been widely used in the application of natural language processing, and there...