Abstract—We address the choice of the coefficients in the cost function of a modular nested recurrent neural-network (RNN) architecture, known as the pipelined recurrent neural network (PRNN). Such a network can cope with the problem of vanishing gradient, experienced in prediction with RNN’s. Constraints on the coefficients of the cost function, in the form of a vector norm, are considered. Unlike the previous cost function for the PRNN, which included a forgetting factor motivated by the recursive least squares (RLS) strategy, the proposed forms of cost function provide “forgetting ” of the outputs of adjacent modules based upon the network architecture. Such an approach takes into ac-count the number of modules in the PRNN, through the u...
A real time recurrent learning (RTRL) algorithm with an adaptive-learning rate for nonlinear adaptiv...
We provide an analysis of nonlinear time series prediction schemes, from a common recurrent neural n...
Network pruning techniques are widely employed to reduce the memory requirements and increase the in...
We address the choice of the coefficients in the cost function of a modular nested recurrent neural-...
We present an approach for selecting optimal parameters for the pipelined recurrent neural network (...
Abstract—We present an approach for selecting optimal pa-rameters for the pipelined recurrent neural...
We provide the relationship between the learning rate and the slope of a nonlinear activation functi...
A relationship between the learning rate · in the learning algorithm, and the slope fl in the nonlin...
A relationship between the learning rate ? in the learning algorithm, and the slope ß in the nonline...
Second order properties of cost functions for recurrent networks are investigated. We analyze a laye...
Recurrent Neural Networks (RNNs) are variants of Neural Networks that are able to learn temporal rel...
An analysis of nonlinear time series prediction schemes, realised though advanced Recurrent Neural N...
A recurrent neural network (RNN) combines variable-length input data with a hidden state that depend...
In this chapter, we describe the basic concepts behind the functioning of recurrent neural networks ...
In this chapter, we present three different recurrent neural network architectures that we employ fo...
A real time recurrent learning (RTRL) algorithm with an adaptive-learning rate for nonlinear adaptiv...
We provide an analysis of nonlinear time series prediction schemes, from a common recurrent neural n...
Network pruning techniques are widely employed to reduce the memory requirements and increase the in...
We address the choice of the coefficients in the cost function of a modular nested recurrent neural-...
We present an approach for selecting optimal parameters for the pipelined recurrent neural network (...
Abstract—We present an approach for selecting optimal pa-rameters for the pipelined recurrent neural...
We provide the relationship between the learning rate and the slope of a nonlinear activation functi...
A relationship between the learning rate · in the learning algorithm, and the slope fl in the nonlin...
A relationship between the learning rate ? in the learning algorithm, and the slope ß in the nonline...
Second order properties of cost functions for recurrent networks are investigated. We analyze a laye...
Recurrent Neural Networks (RNNs) are variants of Neural Networks that are able to learn temporal rel...
An analysis of nonlinear time series prediction schemes, realised though advanced Recurrent Neural N...
A recurrent neural network (RNN) combines variable-length input data with a hidden state that depend...
In this chapter, we describe the basic concepts behind the functioning of recurrent neural networks ...
In this chapter, we present three different recurrent neural network architectures that we employ fo...
A real time recurrent learning (RTRL) algorithm with an adaptive-learning rate for nonlinear adaptiv...
We provide an analysis of nonlinear time series prediction schemes, from a common recurrent neural n...
Network pruning techniques are widely employed to reduce the memory requirements and increase the in...