Bounded rational decision-makers transform sensory input into motor output under limited computational resources. Mathematically, such decision-makers can be modeled as information-theoretic channels with limited transmission rate. Here, we apply this formalism for the first time to multilayer feedforward neural networks. We derive synaptic weight update rules for two scenarios, where either each neuron is considered as a bounded rational decision-maker or the network as a whole. In the update rules, bounded rationality translates into information-theoretically motivated types of regularization in weight space. In experiments on the MNIST benchmark classification task for handwritten digits, we show that such information-theoretic regulariz...
International audienceWe investigate the consequences of maximizing information transfer in a simple...
In this paper we present a foundational study on a constrained method that defines learning problems...
How can neural networks learn to represent information optimally? We answer this question by derivin...
Rate distortion theory describes how to communicate relevant information most efficiently over a cha...
Bounded rationality investigates utility-optimizing decision-makers with limited information-process...
This paper considers the use of neural networks to model bounded rational behaviour. The underlying ...
We introduce a new, efficient, principled and backpropagation-compatible algorithm for learn-ing a p...
Generating functionals may guide the evolution of a dynamical system and constitute a possible route...
Information measures are often used to assess the efficacy of neural networks, and learning rules ca...
This article introduces the concept of optimally distributed computation in feedforward neural netwo...
Deviations from rational decision-making due to limited computational resources have been studied in...
International audienceWe prove that maximization of mutual information between the output and the in...
Inspired by the adaptation phenomenon of neuronal firing, we propose the regularity normalization (R...
An elemental computation in the brain is to identify the best in a set of options and report its val...
The ability to form abstractions and to generalize well from few samples are hallmarks of human and ...
International audienceWe investigate the consequences of maximizing information transfer in a simple...
In this paper we present a foundational study on a constrained method that defines learning problems...
How can neural networks learn to represent information optimally? We answer this question by derivin...
Rate distortion theory describes how to communicate relevant information most efficiently over a cha...
Bounded rationality investigates utility-optimizing decision-makers with limited information-process...
This paper considers the use of neural networks to model bounded rational behaviour. The underlying ...
We introduce a new, efficient, principled and backpropagation-compatible algorithm for learn-ing a p...
Generating functionals may guide the evolution of a dynamical system and constitute a possible route...
Information measures are often used to assess the efficacy of neural networks, and learning rules ca...
This article introduces the concept of optimally distributed computation in feedforward neural netwo...
Deviations from rational decision-making due to limited computational resources have been studied in...
International audienceWe prove that maximization of mutual information between the output and the in...
Inspired by the adaptation phenomenon of neuronal firing, we propose the regularity normalization (R...
An elemental computation in the brain is to identify the best in a set of options and report its val...
The ability to form abstractions and to generalize well from few samples are hallmarks of human and ...
International audienceWe investigate the consequences of maximizing information transfer in a simple...
In this paper we present a foundational study on a constrained method that defines learning problems...
How can neural networks learn to represent information optimally? We answer this question by derivin...