Encoding domain knowledge into the prior over the high-dimensional weight space of a neural network is challenging but essential in applications with limited data and weak signals. Two types of domain knowledge are commonly available in scientific applications: 1. feature sparsity (fraction of features deemed relevant); 2. signal-to-noise ratio, quantified, for instance, as the proportion of variance explained. We show how to encode both types of domain knowledge into the widely used Gaussian scale mixture priors with Automatic Relevance Determination. Specifically, we propose a new joint prior over the local (i.e., feature-specific) scale parameters that encodes knowledge about feature sparsity, and a Stein gradient optimization to tune th...
We propose a novel approach for nonlinear regression using a two-layer neural network (NN) model str...
The information bottleneck framework provides a systematic approach to learning representations that...
AbstractIn this paper we show how a user can influence recovery of Bayesian networks from a database...
| openaire: EC/H2020/101016775/EU//INTERVENEEncoding domain knowledge into the prior over the high-d...
Deep neural networks have bested notable benchmarks across computer vision, reinforcement learning, ...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
In recent years, Neural Networks (NN) have become a popular data-analytic tool in Statistics, Compu...
Bayesian neural networks attempt to combine the strong predictive performance of neural networks wit...
This paper introduces a new neural network based prior for real valued functions on $\mathbb R^d$ wh...
Sparse modeling for signal processing and machine learning has been at the focus of scientific resea...
Isotropic Gaussian priors are the de facto standard for modern Bayesian neural network inference. Ho...
The Bayesian treatment of neural networks dictates that a prior distribution is specified over their...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
The need for function estimation in label-limited settings is common in the natural sciences. At the...
Deep learning is actively used in the area of sparse coding. In current deep sparse coding methods u...
We propose a novel approach for nonlinear regression using a two-layer neural network (NN) model str...
The information bottleneck framework provides a systematic approach to learning representations that...
AbstractIn this paper we show how a user can influence recovery of Bayesian networks from a database...
| openaire: EC/H2020/101016775/EU//INTERVENEEncoding domain knowledge into the prior over the high-d...
Deep neural networks have bested notable benchmarks across computer vision, reinforcement learning, ...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
In recent years, Neural Networks (NN) have become a popular data-analytic tool in Statistics, Compu...
Bayesian neural networks attempt to combine the strong predictive performance of neural networks wit...
This paper introduces a new neural network based prior for real valued functions on $\mathbb R^d$ wh...
Sparse modeling for signal processing and machine learning has been at the focus of scientific resea...
Isotropic Gaussian priors are the de facto standard for modern Bayesian neural network inference. Ho...
The Bayesian treatment of neural networks dictates that a prior distribution is specified over their...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
The need for function estimation in label-limited settings is common in the natural sciences. At the...
Deep learning is actively used in the area of sparse coding. In current deep sparse coding methods u...
We propose a novel approach for nonlinear regression using a two-layer neural network (NN) model str...
The information bottleneck framework provides a systematic approach to learning representations that...
AbstractIn this paper we show how a user can influence recovery of Bayesian networks from a database...