Bayesian neural networks are theoretically well-understood only in the infinite-width limit, where Gaussian priors over network weights yield Gaussian priors over network outputs. Recent work has suggested that finite Bayesian networks may outperform their infinite counterparts, but their non-Gaussian function space priors have been characterized only though perturbative approaches. Here, we derive exact solutions for the function space priors for individual input examples of a class of finite fully-connected feedforward Bayesian neural networks. For deep linear networks, the prior has a simple expression in terms of the Meijer $G$-function. The prior of a finite ReLU network is a mixture of the priors of linear networks of smaller widths, ...
The logit outputs of a feedforward neural network at initialization are conditionally Gaussian, give...
Understanding capabilities and limitations of different network architectures is of fundamental impo...
Deep Gaussian Process (DGP) as a model prior in Bayesian learning intuitively exploits the expressiv...
Recent works have suggested that finite Bayesian neural networks may sometimes outperform their infi...
The analytic inference, e.g. predictive distribution being in closed form, may be an appealing benef...
This paper introduces a new neural network based prior for real valued functions on $\mathbb R^d$ wh...
International audienceThe connection between Bayesian neural networks and Gaussian processes gained ...
The Bayesian treatment of neural networks dictates that a prior distribution is specified over their...
Deep neural networks have bested notable benchmarks across computer vision, reinforcement learning, ...
Comparing Bayesian neural networks (BNNs) with different widths is challenging because, as the width...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
10 pages, 5 figures, ICML'19 conferenceInternational audienceWe investigate deep Bayesian neural net...
International audienceThe connection between Bayesian neural networks and Gaussian processes gained ...
The logit outputs of a feedforward neural network at initialization are conditionally Gaussian, give...
Understanding capabilities and limitations of different network architectures is of fundamental impo...
Deep Gaussian Process (DGP) as a model prior in Bayesian learning intuitively exploits the expressiv...
Recent works have suggested that finite Bayesian neural networks may sometimes outperform their infi...
The analytic inference, e.g. predictive distribution being in closed form, may be an appealing benef...
This paper introduces a new neural network based prior for real valued functions on $\mathbb R^d$ wh...
International audienceThe connection between Bayesian neural networks and Gaussian processes gained ...
The Bayesian treatment of neural networks dictates that a prior distribution is specified over their...
Deep neural networks have bested notable benchmarks across computer vision, reinforcement learning, ...
Comparing Bayesian neural networks (BNNs) with different widths is challenging because, as the width...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
10 pages, 5 figures, ICML'19 conferenceInternational audienceWe investigate deep Bayesian neural net...
International audienceThe connection between Bayesian neural networks and Gaussian processes gained ...
The logit outputs of a feedforward neural network at initialization are conditionally Gaussian, give...
Understanding capabilities and limitations of different network architectures is of fundamental impo...
Deep Gaussian Process (DGP) as a model prior in Bayesian learning intuitively exploits the expressiv...