Recent theoretical results for pattern classification with thresholded real-valued functions (such as support vector machines, sigmoid networks, and boosting) give bounds on misclassification probability that do not depend on the size of the classifier, and hence can be considerably smaller than the bounds that follow from the VC theory. In this paper, we show that these techniques can be more widely applied, by representing other boolean functions as two-layer neural networks (thresholded convex combinations of boolean functions). For example, we show that with high probability any decision tree of depth no more than d that is consistent with m training examples has misclassification probability no more than O i \Gamma 1 m \Gamma N e...
We consider the problem of learning in multilayer feed-forward networks of linear threshold units. W...
Full arxiv preprint version available here: https://arxiv.org/abs/2001.06178A robust theoretical fra...
For classes of concepts defined by certain classes of analytic functions depending on n parameters,...
Sample complexity results from computational learning theory, when applied to neural network learnin...
AbstractWe derive an upper bound on the generalization error of classifiers which can be represented...
This paper shows that if a large neural network is used for a pattern classification problem, and th...
To reduce the memory requirements and the computation cost, many algorithms have been developed that...
We consider the sample complexity of concept learning when we classify by using a fixed Boolean func...
We outline a differential theory of learning for statistical pattern classification. When applied to...
Abstract. The generalization ability of different sizes architectures with one and two hidden layers...
Nearest neighbor classifiers that use all the training samples for classification require large memo...
AbstractWe consider the generalization error of concept learning when using a fixed Boolean function...
This paper investigates an algorithm for the construction of decisions trees comprised of linear thr...
A robust theoretical framework that can describe and predict the generalization ability of DNNs in g...
A general relationship is developed between the VC-dimension and the statistical lower epsilon-capac...
We consider the problem of learning in multilayer feed-forward networks of linear threshold units. W...
Full arxiv preprint version available here: https://arxiv.org/abs/2001.06178A robust theoretical fra...
For classes of concepts defined by certain classes of analytic functions depending on n parameters,...
Sample complexity results from computational learning theory, when applied to neural network learnin...
AbstractWe derive an upper bound on the generalization error of classifiers which can be represented...
This paper shows that if a large neural network is used for a pattern classification problem, and th...
To reduce the memory requirements and the computation cost, many algorithms have been developed that...
We consider the sample complexity of concept learning when we classify by using a fixed Boolean func...
We outline a differential theory of learning for statistical pattern classification. When applied to...
Abstract. The generalization ability of different sizes architectures with one and two hidden layers...
Nearest neighbor classifiers that use all the training samples for classification require large memo...
AbstractWe consider the generalization error of concept learning when using a fixed Boolean function...
This paper investigates an algorithm for the construction of decisions trees comprised of linear thr...
A robust theoretical framework that can describe and predict the generalization ability of DNNs in g...
A general relationship is developed between the VC-dimension and the statistical lower epsilon-capac...
We consider the problem of learning in multilayer feed-forward networks of linear threshold units. W...
Full arxiv preprint version available here: https://arxiv.org/abs/2001.06178A robust theoretical fra...
For classes of concepts defined by certain classes of analytic functions depending on n parameters,...