Choosing appropriate architectures and regular-ization strategies for deep networks is crucial to good predictive performance. To shed light on this problem, we analyze the analogous prob-lem of constructing useful priors on composi-tions of functions. Specifically, we study the deep Gaussian process, a type of infinitely-wide, deep neural network. We show that in standard architectures, the representational capacity of the network tends to capture fewer degrees of free-dom as the number of layers increases, retaining only a single degree of freedom in the limit. We propose an alternate network architecture which does not suffer from this pathology. We also examine deep covariance functions, obtained by composing infinitely many feature tra...
Khan MEE, Immer A, Abedi E, Korzepa M. Approximate Inference Turns Deep Networks into Gaussian Proce...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
© 2018 Matthew M. Dunlop, Mark A. Girolami, Andrew M. Stuart and Aretha L. Teckentrup. Recent resear...
Choosing appropriate architectures and regular-ization strategies for deep networks is crucial to go...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
Recent years have witnessed an increasing interest in the correspondence between infinitely wide net...
Many modern machine learning methods, including deep neural networks, utilize a discrete sequence of...
Understanding capabilities and limitations of different network architectures is of fundamental impo...
This article studies the infinite-width limit of deep feedforward neural networks whose weights are ...
In this paper we introduce deep Gaussian process (GP) models. Deep GPs are a deep belief net-work ba...
We propose a scalable Gaussian process model for regression by applying a deep neural network as the...
This manuscript considers the problem of learning a random Gaussian network function using a fully c...
In this paper we introduce deep Gaussian process (GP) models. Deep GPs are a deep belief network bas...
stitute two of the most important foci of modern machine learning research. In this preliminary work...
Khan MEE, Immer A, Abedi E, Korzepa M. Approximate Inference Turns Deep Networks into Gaussian Proce...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
© 2018 Matthew M. Dunlop, Mark A. Girolami, Andrew M. Stuart and Aretha L. Teckentrup. Recent resear...
Choosing appropriate architectures and regular-ization strategies for deep networks is crucial to go...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
Recent years have witnessed an increasing interest in the correspondence between infinitely wide net...
Many modern machine learning methods, including deep neural networks, utilize a discrete sequence of...
Understanding capabilities and limitations of different network architectures is of fundamental impo...
This article studies the infinite-width limit of deep feedforward neural networks whose weights are ...
In this paper we introduce deep Gaussian process (GP) models. Deep GPs are a deep belief net-work ba...
We propose a scalable Gaussian process model for regression by applying a deep neural network as the...
This manuscript considers the problem of learning a random Gaussian network function using a fully c...
In this paper we introduce deep Gaussian process (GP) models. Deep GPs are a deep belief network bas...
stitute two of the most important foci of modern machine learning research. In this preliminary work...
Khan MEE, Immer A, Abedi E, Korzepa M. Approximate Inference Turns Deep Networks into Gaussian Proce...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
© 2018 Matthew M. Dunlop, Mark A. Girolami, Andrew M. Stuart and Aretha L. Teckentrup. Recent resear...