The lottery ticket hypothesis conjectures the existence of sparse subnetworks of large randomly initialized deep neural networks that can be successfully trained in isolation. Recent work has experimentally observed that some of these tickets can be practically reused across a variety of tasks, hinting at some form of universality. We formalize this concept and theoretically prove that not only do such universal tickets exist but they also do not require further training. Our proofs introduce a couple of technical innovations related to pruning for strong lottery tickets, including extensions of subset sum results and a strategy to leverage higher amounts of depth. Our explicit sparse constructions of universal function families might be of...
The lottery ticket hypothesis suggests that sparse, sub-networks of a given neural network, if initi...
The lottery ticket hypothesis questions the role of overparameterization in supervised deep learning...
Recent advances in deep learning optimization showed that just a subset of parameters are really nec...
The lottery ticket hypothesis has sparked the rapid development of pruning algorithms that perform s...
The strong lottery ticket hypothesis holds the promise that pruning randomly initialized deep neural...
The Lottery Ticket Hypothesis continues to have a profound practical impact on the quest for small s...
Large neural networks can be pruned to a small fraction of their original size, with little loss in ...
International audienceThe lottery ticket hypothesis states that a randomly-initialized neural networ...
The strong lottery ticket hypothesis has highlighted the potential for training deep neural networks...
peer reviewedWe study the generalization properties of pruned models that are the winners of the lot...
The strong lottery ticket hypothesis holds the promise that pruning randomly initialized deep neural...
International audienceThe Strong Lottery Ticket Hypothesis (SLTH) states that randomly-initialised n...
Network pruning is an effective approach to reduce network complexity with acceptable performance co...
Pruning refers to the elimination of trivial weights from neural networks. The sub-networks within a...
Pruning is a standard technique for reducing the computational cost of deep networks. Many advances ...
The lottery ticket hypothesis suggests that sparse, sub-networks of a given neural network, if initi...
The lottery ticket hypothesis questions the role of overparameterization in supervised deep learning...
Recent advances in deep learning optimization showed that just a subset of parameters are really nec...
The lottery ticket hypothesis has sparked the rapid development of pruning algorithms that perform s...
The strong lottery ticket hypothesis holds the promise that pruning randomly initialized deep neural...
The Lottery Ticket Hypothesis continues to have a profound practical impact on the quest for small s...
Large neural networks can be pruned to a small fraction of their original size, with little loss in ...
International audienceThe lottery ticket hypothesis states that a randomly-initialized neural networ...
The strong lottery ticket hypothesis has highlighted the potential for training deep neural networks...
peer reviewedWe study the generalization properties of pruned models that are the winners of the lot...
The strong lottery ticket hypothesis holds the promise that pruning randomly initialized deep neural...
International audienceThe Strong Lottery Ticket Hypothesis (SLTH) states that randomly-initialised n...
Network pruning is an effective approach to reduce network complexity with acceptable performance co...
Pruning refers to the elimination of trivial weights from neural networks. The sub-networks within a...
Pruning is a standard technique for reducing the computational cost of deep networks. Many advances ...
The lottery ticket hypothesis suggests that sparse, sub-networks of a given neural network, if initi...
The lottery ticket hypothesis questions the role of overparameterization in supervised deep learning...
Recent advances in deep learning optimization showed that just a subset of parameters are really nec...