Recurrent neural networks (RNNs) have achieved state-of-the-art performances on various applications. However, RNNs are prone to be memory-bandwidth limited in practical applications and need both long periods of training and inference time. The aforementioned problems are at odds with training and deploying RNNs on resource-limited devices where the memory and floating-point operations (FLOPs) budget are strictly constrained. To address this problem, conventional model compression techniques usually focus on reducing inference costs, operating on a costly pre-trained model. Recently, dynamic sparse training has been proposed to accelerate the training process by directly training sparse neural networks from scratch. However, previous spars...
Recently, sparse training methods have started to be established as a de facto approach for training...
The scalability of recurrent neural networks (RNNs) is hindered by the sequential dependence of each...
Recurrent Neural Networks (RNNs) are useful for speech recognition but their fully-connected structu...
Recurrent neural networks (RNNs) have achieved state-of-the-art performances on various applications...
Sparse neural networks have been widely applied to reduce the necessary resource requirements to tra...
Inducing sparseness while training neural networks has been shown to yield models with a lower memor...
Artificial neural networks (ANNs) have emerged as hot topics in the research community. Despite the ...
ABSTRACT We present several modifications of the original recurrent neural network language model (R...
The feed-forward neural network (FNN) has drawn great interest in many applications due to its unive...
The growing energy and performance costs of deep learning have driven the community to reduce the si...
Long Short-Term Memory (LSTM) has achieved state-of-the-art performances on a wide range of tasks. I...
Recurrent Neural Networks (RNNs) are powerful sequence models that were believed to be difficult to ...
Abstract Sparse neural networks can achieve performance comparable to fully connected networks but n...
Neural language models (LMs) based on recurrent neural networks (RNN) are some of the most successfu...
International audienceCoded recurrent neural networks with three levels of sparsity are introduced. ...
Recently, sparse training methods have started to be established as a de facto approach for training...
The scalability of recurrent neural networks (RNNs) is hindered by the sequential dependence of each...
Recurrent Neural Networks (RNNs) are useful for speech recognition but their fully-connected structu...
Recurrent neural networks (RNNs) have achieved state-of-the-art performances on various applications...
Sparse neural networks have been widely applied to reduce the necessary resource requirements to tra...
Inducing sparseness while training neural networks has been shown to yield models with a lower memor...
Artificial neural networks (ANNs) have emerged as hot topics in the research community. Despite the ...
ABSTRACT We present several modifications of the original recurrent neural network language model (R...
The feed-forward neural network (FNN) has drawn great interest in many applications due to its unive...
The growing energy and performance costs of deep learning have driven the community to reduce the si...
Long Short-Term Memory (LSTM) has achieved state-of-the-art performances on a wide range of tasks. I...
Recurrent Neural Networks (RNNs) are powerful sequence models that were believed to be difficult to ...
Abstract Sparse neural networks can achieve performance comparable to fully connected networks but n...
Neural language models (LMs) based on recurrent neural networks (RNN) are some of the most successfu...
International audienceCoded recurrent neural networks with three levels of sparsity are introduced. ...
Recently, sparse training methods have started to be established as a de facto approach for training...
The scalability of recurrent neural networks (RNNs) is hindered by the sequential dependence of each...
Recurrent Neural Networks (RNNs) are useful for speech recognition but their fully-connected structu...