We study the interaction between input distributions, learning algorithms and finite sample sizes in the case of learning classification tasks. Focusing on the case of normal input distributions, we use statistical mechanics techniques to calculate the empirical and expected (or generalization) errors for several well-known algorithms learning the weights of a singlelayer perceptron. In the case of spherically symmetric distributions within each class we find that the simple Hebb rule, corresponding to maximum-likelihood parameter estimation, outperforms the other more complex algorithms, based on error minimization. Moreover, we show that in the regime where the overlap between the classes is large, algorithms with low empirical error do w...
Abstract. We propose a novel approach for the estimation of the size of training sets that are neede...
The learning of a pattern classification rule rests on acquiring information to constitute a decisio...
We investigate to which extent one can recover class probabilities within the empirical risk minimiz...
We study the interaction between input distributions, learning algo-rithms, and finite sample sizes ...
We study model selection strategies based on penalized empirical loss minimization. We point out a...
We investigate to which extent one can recover class probabilities within the empirical risk minimiz...
Abstract The generalization ability of minimizers of the empirical risk in the context of binary cla...
Abstract. We study model selection strategies based on penalized empirical loss minimization. We poi...
Plotting a learner’s average performance against the number of training samples results in a learnin...
We compare two strategies for training connectionist (as well as non-connectionist) models for stati...
The derivation of tight generalization bounds has remained an open problem in the statistical learni...
This thesis concerns the development and mathematical analysis of statistical procedures for classi...
Abstract. The theoretical framework of Statistical Learning Theory (SLT) for pattern recognition pro...
Statistical Learning Theory studies the problem of learning an unknown relationship between observed...
In this paper we show how to extract a hypothesis with small risk from the ensemble of hypotheses ge...
Abstract. We propose a novel approach for the estimation of the size of training sets that are neede...
The learning of a pattern classification rule rests on acquiring information to constitute a decisio...
We investigate to which extent one can recover class probabilities within the empirical risk minimiz...
We study the interaction between input distributions, learning algo-rithms, and finite sample sizes ...
We study model selection strategies based on penalized empirical loss minimization. We point out a...
We investigate to which extent one can recover class probabilities within the empirical risk minimiz...
Abstract The generalization ability of minimizers of the empirical risk in the context of binary cla...
Abstract. We study model selection strategies based on penalized empirical loss minimization. We poi...
Plotting a learner’s average performance against the number of training samples results in a learnin...
We compare two strategies for training connectionist (as well as non-connectionist) models for stati...
The derivation of tight generalization bounds has remained an open problem in the statistical learni...
This thesis concerns the development and mathematical analysis of statistical procedures for classi...
Abstract. The theoretical framework of Statistical Learning Theory (SLT) for pattern recognition pro...
Statistical Learning Theory studies the problem of learning an unknown relationship between observed...
In this paper we show how to extract a hypothesis with small risk from the ensemble of hypotheses ge...
Abstract. We propose a novel approach for the estimation of the size of training sets that are neede...
The learning of a pattern classification rule rests on acquiring information to constitute a decisio...
We investigate to which extent one can recover class probabilities within the empirical risk minimiz...