We investigate filter level sparsity that emerges in convolutional neural networks (CNNs) which employ Batch Normalization and ReLU activation, and are trained with adaptive gradient descent techniques and L2 regularization or weight decay. We conduct an extensive experimental study casting our initial findings into hypotheses and conclusions about the mechanisms underlying the emergent filter level sparsity. This study allows new insight into the performance gap obeserved between adapative and non-adaptive gradient descent methods in practice. Further, analysis of the effect of training strategies and hyperparameters on the sparsity leads to practical suggestions in designing CNN training strategies enabling us to explore the tradeoffs bet...
The training of sparse neural networks is becoming an increasingly important tool for reducing the ...
The undeniable computational power of artificial neural networks has granted the scientific communit...
Currently, many theoretical as well as practically relevant questions towards the transferability an...
We investigate filter level sparsity that emerges in convolutional neural networks (CNNs) which empl...
The growing energy and performance costs of deep learning have driven the community to reduce the si...
Deep Convolution Neural Networks (CNNs) have been widely used in image recognition, while models of ...
Sparsity has played an important role in numerous signal processing systems. By leveraging sparse re...
The millions of filter weights in Convolutional Neural Networks (CNNs), all have a well-defined and ...
It is widely believed that the success of deep networks lies in their ability to learn a meaningful ...
Structural neural network pruning aims to remove the redundant channels in the deep convolutional ne...
Brain-inspired event-based processors have attracted considerable attention for edge deployment beca...
We analyze deep ReLU neural networks trained with mini-batch stochastic gradient decent and weight d...
Deep learning is finding its way into the embedded world with applications such as autonomous drivin...
The ever-increasing number of parameters in deep neural networks poses challenges for memory-limited...
We study the impact of different pruning techniques on the representation learned by deep neural net...
The training of sparse neural networks is becoming an increasingly important tool for reducing the ...
The undeniable computational power of artificial neural networks has granted the scientific communit...
Currently, many theoretical as well as practically relevant questions towards the transferability an...
We investigate filter level sparsity that emerges in convolutional neural networks (CNNs) which empl...
The growing energy and performance costs of deep learning have driven the community to reduce the si...
Deep Convolution Neural Networks (CNNs) have been widely used in image recognition, while models of ...
Sparsity has played an important role in numerous signal processing systems. By leveraging sparse re...
The millions of filter weights in Convolutional Neural Networks (CNNs), all have a well-defined and ...
It is widely believed that the success of deep networks lies in their ability to learn a meaningful ...
Structural neural network pruning aims to remove the redundant channels in the deep convolutional ne...
Brain-inspired event-based processors have attracted considerable attention for edge deployment beca...
We analyze deep ReLU neural networks trained with mini-batch stochastic gradient decent and weight d...
Deep learning is finding its way into the embedded world with applications such as autonomous drivin...
The ever-increasing number of parameters in deep neural networks poses challenges for memory-limited...
We study the impact of different pruning techniques on the representation learned by deep neural net...
The training of sparse neural networks is becoming an increasingly important tool for reducing the ...
The undeniable computational power of artificial neural networks has granted the scientific communit...
Currently, many theoretical as well as practically relevant questions towards the transferability an...