Recently, a new recurrent neural network (RNN) named the Legendre Memory Unit (LMU) was proposed and shown to achieve state-of-the-art performance on several benchmark datasets. Here we leverage the linear time-invariant (LTI) memory component of the LMU to construct a simplified variant that can be parallelized during training (and yet executed as an RNN during inference), resulting in up to 200 times faster training. We note that our efficient parallelizing scheme is general and is applicable to any deep network whose recurrent components are LTI systems. We demonstrate the improved accuracy and decreased parameter count of our new architecture compared to the original LMU and a variety of published LSTM and transformer networks across s...
The random neural network (RNN) is a mathematical model for an ``integrate and fire'' spiking networ...
Recurrent Neural Networks (RNN) show a remarkable result in sequence learning, particularly in archi...
Neural networks are very powerful computational models, capable of outperforming humans on a variety...
In this paper, we investigate the memory properties of two popular gated units: long short term memo...
Deep learning has achieved great success in many sequence learning tasks such as machine translation...
In our previous work we have shown that resistive cross point devices, so called resistive processin...
Dans un problème de prédiction à multiples pas discrets, la prédiction à chaque instant peut dépendr...
In this thesis, we study novel neural network structures to better model long term dependency in seq...
The Recurrent Neural Network (RNN) is an ex-tremely powerful sequence model that is often difficult ...
Accelerating and scaling the training of deep neural networks (DNNs) is critical to keep up with gro...
International audienceIt is common to evaluate the performance of a machine learning model by measur...
Sequential models, such as Recurrent Neural Networks and Neural Ordinary Differential Equations, hav...
Recurrent neural networks (RNNs) have represented for years the state of the art in neural machine t...
Recurrent Neural Networks (RNNs) are a type of neural network that maintains a hidden state, preserv...
Recurrent neural networks (RNNs) are widely used for natural language processing, time-series predic...
The random neural network (RNN) is a mathematical model for an ``integrate and fire'' spiking networ...
Recurrent Neural Networks (RNN) show a remarkable result in sequence learning, particularly in archi...
Neural networks are very powerful computational models, capable of outperforming humans on a variety...
In this paper, we investigate the memory properties of two popular gated units: long short term memo...
Deep learning has achieved great success in many sequence learning tasks such as machine translation...
In our previous work we have shown that resistive cross point devices, so called resistive processin...
Dans un problème de prédiction à multiples pas discrets, la prédiction à chaque instant peut dépendr...
In this thesis, we study novel neural network structures to better model long term dependency in seq...
The Recurrent Neural Network (RNN) is an ex-tremely powerful sequence model that is often difficult ...
Accelerating and scaling the training of deep neural networks (DNNs) is critical to keep up with gro...
International audienceIt is common to evaluate the performance of a machine learning model by measur...
Sequential models, such as Recurrent Neural Networks and Neural Ordinary Differential Equations, hav...
Recurrent neural networks (RNNs) have represented for years the state of the art in neural machine t...
Recurrent Neural Networks (RNNs) are a type of neural network that maintains a hidden state, preserv...
Recurrent neural networks (RNNs) are widely used for natural language processing, time-series predic...
The random neural network (RNN) is a mathematical model for an ``integrate and fire'' spiking networ...
Recurrent Neural Networks (RNN) show a remarkable result in sequence learning, particularly in archi...
Neural networks are very powerful computational models, capable of outperforming humans on a variety...