The strong lottery ticket hypothesis has highlighted the potential for training deep neural networks by pruning, which has inspired interesting practical and theoretical insights into how neural networks can represent functions. For networks with ReLU activation functions, it has been proven that a target network with depth L can be approximated by the subnetwork of a randomly initialized neural network that has double the target’s depth 2L and is wider by a logarithmic factor. We show that a depth L + 1 network is sufficient. This result indicates that we can expect to find lottery tickets at realistic, commonly used depths while only requiring logarithmic overparametrization. Our novel construction approach applies to a large class of ac...
We propose an optimal architecture for deep neural networks of given size. The optimal architecture ...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
In this article we present new results on neural networks with linear threshold activation functions...
The Lottery Ticket Hypothesis continues to have a profound practical impact on the quest for small s...
The lottery ticket hypothesis conjectures the existence of sparse subnetworks of large randomly init...
The lottery ticket hypothesis has sparked the rapid development of pruning algorithms that aim to re...
The strong lottery ticket hypothesis holds the promise that pruning randomly initialized deep neural...
Network pruning is an effective approach to reduce network complexity with acceptable performance co...
The strong lottery ticket hypothesis holds the promise that pruning randomly initialized deep neural...
International audienceThe lottery ticket hypothesis states that a randomly-initialized neural networ...
We solve an open question from Lu et al. (2017), by showing that any target network with inputs in $...
Large neural networks can be pruned to a small fraction of their original size, with little loss in ...
We contribute to a better understanding of the class of functions that can be represented by a neura...
Pruning is a standard technique for reducing the computational cost of deep networks. Many advances ...
Recent advances in deep learning optimization showed that just a subset of parameters are really nec...
We propose an optimal architecture for deep neural networks of given size. The optimal architecture ...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
In this article we present new results on neural networks with linear threshold activation functions...
The Lottery Ticket Hypothesis continues to have a profound practical impact on the quest for small s...
The lottery ticket hypothesis conjectures the existence of sparse subnetworks of large randomly init...
The lottery ticket hypothesis has sparked the rapid development of pruning algorithms that aim to re...
The strong lottery ticket hypothesis holds the promise that pruning randomly initialized deep neural...
Network pruning is an effective approach to reduce network complexity with acceptable performance co...
The strong lottery ticket hypothesis holds the promise that pruning randomly initialized deep neural...
International audienceThe lottery ticket hypothesis states that a randomly-initialized neural networ...
We solve an open question from Lu et al. (2017), by showing that any target network with inputs in $...
Large neural networks can be pruned to a small fraction of their original size, with little loss in ...
We contribute to a better understanding of the class of functions that can be represented by a neura...
Pruning is a standard technique for reducing the computational cost of deep networks. Many advances ...
Recent advances in deep learning optimization showed that just a subset of parameters are really nec...
We propose an optimal architecture for deep neural networks of given size. The optimal architecture ...
Recently there has been much interest in understanding why deep neural networks are preferred to sha...
In this article we present new results on neural networks with linear threshold activation functions...