We consider information-theoretic bounds on expected generalization error for statistical learning problems in a networked setting. In this setting, there are $K$ nodes, each with its own independent dataset, and the models from each node have to be aggregated into a final centralized model. We consider both simple averaging of the models as well as more complicated multi-round algorithms. We give upper bounds on the expected generalization error for a variety of problems, such as those with Bregman divergence or Lipschitz continuous losses, that demonstrate an improved dependence of $1/K$ on the number of nodes. These "per node" bounds are in terms of the mutual information between the training dataset and the trained weights at each node,...
We investigate the performance of distributed learning for large-scale linear regression where the m...
This work discusses how to derive upper bounds for the expected generalisation error of supervised l...
This work discusses how to derive upper bounds for the expected generalisation error of supervised l...
We consider information-theoretic bounds on the expected generalization error for statistical learni...
In this paper, we use tools from rate-distortion theory to establish new upper bounds on the general...
The following problem is considered: given a joint distribution P XY and an event E, bound P XY (E) ...
Bounding the generalization error of learning algorithms has a long history, which yet falls short i...
Distributed learning facilitates the scaling-up of data processing by distributing the computational...
Distributed learning facilitates the scaling-up of data processing by distributing the computational...
Distributed learning facilitates the scaling-up of data processing by distributing the computational...
In a generic distributed information processing system, a number of agents connected by communicatio...
Generalization error bounds are critical to understanding the performance of machine learning models...
We investigate the performance of distributed learning for large-scale linear regression where the m...
We investigate the performance of distributed learning for large-scale linear regression where the m...
We investigate the performance of distributed learning for large-scale linear regression where the m...
We investigate the performance of distributed learning for large-scale linear regression where the m...
This work discusses how to derive upper bounds for the expected generalisation error of supervised l...
This work discusses how to derive upper bounds for the expected generalisation error of supervised l...
We consider information-theoretic bounds on the expected generalization error for statistical learni...
In this paper, we use tools from rate-distortion theory to establish new upper bounds on the general...
The following problem is considered: given a joint distribution P XY and an event E, bound P XY (E) ...
Bounding the generalization error of learning algorithms has a long history, which yet falls short i...
Distributed learning facilitates the scaling-up of data processing by distributing the computational...
Distributed learning facilitates the scaling-up of data processing by distributing the computational...
Distributed learning facilitates the scaling-up of data processing by distributing the computational...
In a generic distributed information processing system, a number of agents connected by communicatio...
Generalization error bounds are critical to understanding the performance of machine learning models...
We investigate the performance of distributed learning for large-scale linear regression where the m...
We investigate the performance of distributed learning for large-scale linear regression where the m...
We investigate the performance of distributed learning for large-scale linear regression where the m...
We investigate the performance of distributed learning for large-scale linear regression where the m...
This work discusses how to derive upper bounds for the expected generalisation error of supervised l...
This work discusses how to derive upper bounds for the expected generalisation error of supervised l...