Recurrent Neural Networks (RNNs) are state-of-the-art models for many machine learning tasks, such as language modeling and machine translation. Executing the inference phase of a RNN directly in edge nodes, rather than in the cloud, would provide benefits in terms of energy consumption, latency and network bandwidth, provided that models can be made efficient enough to run on energy-constrained embedded devices. To this end, we propose an algorithmic optimization for improving the energy efficiency of encoder-decoder RNNs. Our method operates on the Beam Width (BW), i.e. one of the parameters that most influences inference complexity, modulating it depending on the currently processed input based on a metric of the network's "confidence...
Increasing the capacity of recurrent neural networks (RNN) usually involves augmenting the size...
This article introduces Random Error Sampling-based Neuroevolution (RESN), a novel automatic method ...
© 2017 IEEE. This work targets the automated minimum-energy optimization of Quantized Neural Network...
Recurrent Neural Networks (RNNs) are a key technology for emerging applications such as automatic sp...
Deep learning models have reached state of the art performance in many machine learning tasks. Benef...
Deep Learning algorithms have been remarkably successful in applications such as Automatic Speech Re...
Over the past decade, Deep Learning (DL) and Deep Neural Network (DNN) have gone through a rapid dev...
The recent shift in machine learning towards the edge offers a new opportunity to realize intelligen...
Neural network language models are often trained by optimizing likelihood, but we would prefer to op...
Recurrent neural networks (RNNs) have represented for years the state of the art in neural machine t...
Lightweight neural networks that employ depthwise convolution have a significant computational advan...
Recurrent neural networks such as Long Short-Term Memories (LSTMs) learn temporal dependencies by ke...
This paper proposes a novel latency-hiding hardware architecture based on column-wise matrix-vector ...
In recent years the field of neuromorphic low-power systems gained significant momentum, spurring br...
Recurrent neural networks (RNNs) have achieved state-of-the-art performances on various applications...
Increasing the capacity of recurrent neural networks (RNN) usually involves augmenting the size...
This article introduces Random Error Sampling-based Neuroevolution (RESN), a novel automatic method ...
© 2017 IEEE. This work targets the automated minimum-energy optimization of Quantized Neural Network...
Recurrent Neural Networks (RNNs) are a key technology for emerging applications such as automatic sp...
Deep learning models have reached state of the art performance in many machine learning tasks. Benef...
Deep Learning algorithms have been remarkably successful in applications such as Automatic Speech Re...
Over the past decade, Deep Learning (DL) and Deep Neural Network (DNN) have gone through a rapid dev...
The recent shift in machine learning towards the edge offers a new opportunity to realize intelligen...
Neural network language models are often trained by optimizing likelihood, but we would prefer to op...
Recurrent neural networks (RNNs) have represented for years the state of the art in neural machine t...
Lightweight neural networks that employ depthwise convolution have a significant computational advan...
Recurrent neural networks such as Long Short-Term Memories (LSTMs) learn temporal dependencies by ke...
This paper proposes a novel latency-hiding hardware architecture based on column-wise matrix-vector ...
In recent years the field of neuromorphic low-power systems gained significant momentum, spurring br...
Recurrent neural networks (RNNs) have achieved state-of-the-art performances on various applications...
Increasing the capacity of recurrent neural networks (RNN) usually involves augmenting the size...
This article introduces Random Error Sampling-based Neuroevolution (RESN), a novel automatic method ...
© 2017 IEEE. This work targets the automated minimum-energy optimization of Quantized Neural Network...