A robust theoretical framework that can describe and predict the generalization ability of DNNs in general circumstances remains elusive. Classical attempts have produced complexity metrics that rely heavily on global measures of compactness and capacity with little investigation into the effects of sub-component collaboration. We demonstrate intriguing regularities in the activation patterns of the hidden nodes within fully-connected feedforward networks. By tracing the origin of these patterns, we show how such networks can be viewed as the combination of two information processing systems: one continuous and one discrete. We describe how these two systems arise naturally from the gradient-based optimization process, and demonstrate the c...
We attempt to demonstrate the effectiveness of multiple points of view toward neural networks. By re...
In this thesis we explore pattern mining and deep learning. Often seen as orthogonal, we show that t...
The generalization capabilities of deep neural networks are not well understood, and in particular, ...
A robust theoretical framework that can describe and predict the generalization ability of DNNs in g...
Full arxiv preprint version available here: https://arxiv.org/abs/2001.06178A robust theoretical fra...
MEng (Computer and Electronic Engineering), North-West University, Potchefstroom CampusThe ability o...
Recent theoretical results for pattern classification with thresholded real-valued functions (such a...
By making assumptions on the probability distribution of the potentials in a feed-forward neural net...
We outline a differential theory of learning for statistical pattern classification. When applied to...
Generalization is a central aspect of learning theory. Here, we propose a framework that explores an...
No framework exists that can explain and predict the generalisation ability of DNNs in general circu...
Abstract. The generalization ability of different sizes architectures with one and two hidden layers...
One of the important challenges today in deep learning is explaining the outstanding power of genera...
We present a unified framework for a number of different ways of failing to generalize properly. Dur...
There is a need to better understand how generalization works in a deep learning model. The goal of ...
We attempt to demonstrate the effectiveness of multiple points of view toward neural networks. By re...
In this thesis we explore pattern mining and deep learning. Often seen as orthogonal, we show that t...
The generalization capabilities of deep neural networks are not well understood, and in particular, ...
A robust theoretical framework that can describe and predict the generalization ability of DNNs in g...
Full arxiv preprint version available here: https://arxiv.org/abs/2001.06178A robust theoretical fra...
MEng (Computer and Electronic Engineering), North-West University, Potchefstroom CampusThe ability o...
Recent theoretical results for pattern classification with thresholded real-valued functions (such a...
By making assumptions on the probability distribution of the potentials in a feed-forward neural net...
We outline a differential theory of learning for statistical pattern classification. When applied to...
Generalization is a central aspect of learning theory. Here, we propose a framework that explores an...
No framework exists that can explain and predict the generalisation ability of DNNs in general circu...
Abstract. The generalization ability of different sizes architectures with one and two hidden layers...
One of the important challenges today in deep learning is explaining the outstanding power of genera...
We present a unified framework for a number of different ways of failing to generalize properly. Dur...
There is a need to better understand how generalization works in a deep learning model. The goal of ...
We attempt to demonstrate the effectiveness of multiple points of view toward neural networks. By re...
In this thesis we explore pattern mining and deep learning. Often seen as orthogonal, we show that t...
The generalization capabilities of deep neural networks are not well understood, and in particular, ...