We make three related contributions motivated by the challenge of training stochastic neural networks, particularly in a PAC-Bayesian setting: (1) we show how averaging over an ensemble of stochastic neural networks enables a new class of \emph{partially-aggregated} estimators; (2) we show that these lead to provably lower-variance gradient estimates for non-differentiable signed-output networks; (3) we reformulate a PAC-Bayesian bound for these networks to derive a directly optimisable, differentiable objective and a generalisation guarantee, without using a surrogate loss or loosening the bound. This bound is twice as tight as that of Letarte et al. (2019) on a similar network type. We show empirically that these innovations make training...
International audienceWe provide a first PAC-Bayesian analysis for domain adaptation (DA) which aris...
International audienceWe provide two main contributions in PAC-Bayesian theory for domain adaptation...
We focus on a specific class of shallow neural networks with a single hidden layer, namely those wit...
Recent studies have empirically investigated different methods to train stochastic neural networks o...
PAC-Bayesian bounds are known to be tight and informative when studying the generalization ability o...
The limit of infinite width allows for substantial simplifications in the analytical study of overpa...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
This paper presents an empirical study regarding training probabilistic neural networks using traini...
International audienceWe propose a PAC-Bayesian theoretical study of the two-phase learning procedur...
We establish a disintegrated PAC-Bayesian bound, for classifiers that are trained via continuous-tim...
International audienceA learning method is self-certified if it uses all available data to simultane...
International audienceWe give a general recipe for derandomising PAC-Bayesian bounds using margins, ...
Multiclass neural networks are a common tool in modern unsupervised domain adaptation, yet an approp...
In this work, we construct generalization bounds to understand existing learning algorithms and prop...
Risk bounds, which are also called generalisation bounds in the statistical learning literature, are...
International audienceWe provide a first PAC-Bayesian analysis for domain adaptation (DA) which aris...
International audienceWe provide two main contributions in PAC-Bayesian theory for domain adaptation...
We focus on a specific class of shallow neural networks with a single hidden layer, namely those wit...
Recent studies have empirically investigated different methods to train stochastic neural networks o...
PAC-Bayesian bounds are known to be tight and informative when studying the generalization ability o...
The limit of infinite width allows for substantial simplifications in the analytical study of overpa...
International audiencePAC-Bayesian bounds are known to be tight and informative when studying the ge...
This paper presents an empirical study regarding training probabilistic neural networks using traini...
International audienceWe propose a PAC-Bayesian theoretical study of the two-phase learning procedur...
We establish a disintegrated PAC-Bayesian bound, for classifiers that are trained via continuous-tim...
International audienceA learning method is self-certified if it uses all available data to simultane...
International audienceWe give a general recipe for derandomising PAC-Bayesian bounds using margins, ...
Multiclass neural networks are a common tool in modern unsupervised domain adaptation, yet an approp...
In this work, we construct generalization bounds to understand existing learning algorithms and prop...
Risk bounds, which are also called generalisation bounds in the statistical learning literature, are...
International audienceWe provide a first PAC-Bayesian analysis for domain adaptation (DA) which aris...
International audienceWe provide two main contributions in PAC-Bayesian theory for domain adaptation...
We focus on a specific class of shallow neural networks with a single hidden layer, namely those wit...