The recent lottery ticket hypothesis proposes that there is at least one sub-network that matches the accuracy of the original network when trained in isolation. Recent work shows that under SGD noise, several such tickets emerge. We build on these works and study how winning tickets derived from one fixed network differ in structural and functional terms under varying levels of stochasticity. Structurally, we show that the Hamming distance of winning tickets' shapes follow the hyper-geometric distribution. Functionally, our experiments validate that different emerging winning tickets are not disguised variants of each other, but diverge also concerning their classification outputs. Last but not least, different regimes of stochasticity aff...
Lottery tickets (LTs) is able to discover accurate and sparse subnetworks that could be trained in i...
International audienceThe theoretical analysis of deep neural networks (DNN) is arguably among the m...
Stochastic binary hidden units in a multi-layer perceptron (MLP) network give at least three potenti...
International audienceThe lottery ticket hypothesis states that a randomly-initialized neural networ...
peer reviewedWe study the generalization properties of pruned models that are the winners of the lot...
The lottery ticket hypothesis suggests that sparse, sub-networks of a given neural network, if initi...
International audienceThe Strong Lottery Ticket Hypothesis (SLTH) states that randomly-initialised n...
The lottery ticket hypothesis conjectures the existence of sparse subnetworks of large randomly init...
The Lottery Ticket Hypothesis continues to have a profound practical impact on the quest for small s...
The lottery ticket hypothesis has sparked the rapid development of pruning algorithms that aim to re...
Large neural networks can be pruned to a small fraction of their original size, with little loss in ...
Pre-training serves as a broadly adopted starting point for transfer learning on various downstream ...
Foundational work on the Lottery Ticket Hypothesis has suggested an exciting corollary: winning tick...
The strong lottery ticket hypothesis holds the promise that pruning randomly initialized deep neural...
The conventional lottery ticket hypothesis (LTH) claims that there exists a sparse subnetwork within...
Lottery tickets (LTs) is able to discover accurate and sparse subnetworks that could be trained in i...
International audienceThe theoretical analysis of deep neural networks (DNN) is arguably among the m...
Stochastic binary hidden units in a multi-layer perceptron (MLP) network give at least three potenti...
International audienceThe lottery ticket hypothesis states that a randomly-initialized neural networ...
peer reviewedWe study the generalization properties of pruned models that are the winners of the lot...
The lottery ticket hypothesis suggests that sparse, sub-networks of a given neural network, if initi...
International audienceThe Strong Lottery Ticket Hypothesis (SLTH) states that randomly-initialised n...
The lottery ticket hypothesis conjectures the existence of sparse subnetworks of large randomly init...
The Lottery Ticket Hypothesis continues to have a profound practical impact on the quest for small s...
The lottery ticket hypothesis has sparked the rapid development of pruning algorithms that aim to re...
Large neural networks can be pruned to a small fraction of their original size, with little loss in ...
Pre-training serves as a broadly adopted starting point for transfer learning on various downstream ...
Foundational work on the Lottery Ticket Hypothesis has suggested an exciting corollary: winning tick...
The strong lottery ticket hypothesis holds the promise that pruning randomly initialized deep neural...
The conventional lottery ticket hypothesis (LTH) claims that there exists a sparse subnetwork within...
Lottery tickets (LTs) is able to discover accurate and sparse subnetworks that could be trained in i...
International audienceThe theoretical analysis of deep neural networks (DNN) is arguably among the m...
Stochastic binary hidden units in a multi-layer perceptron (MLP) network give at least three potenti...