We present a method for determining the globally optimal on-line learning rule for a soft committee machine under a statistical mechanics framework. This rule maximizes the total reduction in generalization error over the whole learning process. A simple example demonstrates that the locally optimal rule, which maximizes the rate of decrease in generalization error, may perform poorly in comparison
We explore the effects of over-specificity in learning algorithms by investigating the behavior of a...
On-line learning of a rule given by an N-dimensional Ising perceptron, is considered for the case wh...
We study learning from single presentation of examples (incremental or on-line learning) in single-...
We present a method for determining the globally optimal on-line learning rule for a soft committee ...
. We present a method for determining the globally optimal on-line learning rule for a soft committe...
We present a framework for calculating globally optimal parameters, within a given time frame, for o...
A method for calculating the globally optimal learning rate in on-line gradient-descent training of ...
The influence of biases on the learning dynamics of a two-layer neural network, a normalized soft-co...
In this paper we review recent theoretical approaches for analysing the dynamics of on-line learning...
We study on-line gradient-descent learning in multilayer networks analytically and numerically. The ...
We present an analytic solution to the problem of on-line gradient-descent learning for two-layer ne...
The dynamics of on-line learning is investigated for structurally unrealizable tasks in the context ...
A fast algorithm is proposed for optimal supervised learning in multiple-layer neural networks. The ...
Zero temperature Gibbs learning is considered for a connected committee machine with K hidden units....
NeurIPS 2018 - Thirty-second Conference on Neural Information Processing Systems, Montreal, Canada, ...
We explore the effects of over-specificity in learning algorithms by investigating the behavior of a...
On-line learning of a rule given by an N-dimensional Ising perceptron, is considered for the case wh...
We study learning from single presentation of examples (incremental or on-line learning) in single-...
We present a method for determining the globally optimal on-line learning rule for a soft committee ...
. We present a method for determining the globally optimal on-line learning rule for a soft committe...
We present a framework for calculating globally optimal parameters, within a given time frame, for o...
A method for calculating the globally optimal learning rate in on-line gradient-descent training of ...
The influence of biases on the learning dynamics of a two-layer neural network, a normalized soft-co...
In this paper we review recent theoretical approaches for analysing the dynamics of on-line learning...
We study on-line gradient-descent learning in multilayer networks analytically and numerically. The ...
We present an analytic solution to the problem of on-line gradient-descent learning for two-layer ne...
The dynamics of on-line learning is investigated for structurally unrealizable tasks in the context ...
A fast algorithm is proposed for optimal supervised learning in multiple-layer neural networks. The ...
Zero temperature Gibbs learning is considered for a connected committee machine with K hidden units....
NeurIPS 2018 - Thirty-second Conference on Neural Information Processing Systems, Montreal, Canada, ...
We explore the effects of over-specificity in learning algorithms by investigating the behavior of a...
On-line learning of a rule given by an N-dimensional Ising perceptron, is considered for the case wh...
We study learning from single presentation of examples (incremental or on-line learning) in single-...