We investigate the behavior of two learning rules, Stochastic Best Response (SBR) and Replicator Dynamics, in a model with aggregate and time-correlated shocks to payoffs. The main difference between the two behavior of the two rules is that under SBR corners are not absorbing. We study a setting where there are two actions and many states of nature and the transition between states follows a Markov chain. We find that the SBR converges to a behavior similar to probability matching. On the other hand, the Replicator Dynamics selects the optimal action only if the average payoff of both actions is different enough. JEL Classification Number: C73
In the paper, we re-investigate the long run behavior of an adaptive learning process driven by the ...
AbstractWe address the problem of reinforcement learning in which observations may exhibit an arbitr...
In this paper, we present a brief survey of reinforcement learning, with particular emphasis on stoc...
This paper introduces a new model, i.e. state-coupled replicator dynamics, expanding the link betwee...
This paper extends the link between evolutionary game theory and multi-agent reinforcement learning ...
We investigate learning in a setting where each period a population has to choose between two action...
We investigate the stability properties of Muth's model of price movements when agents choose a prod...
We address the problem of reinforcement learning in which observations may exhibit an arbitrary form...
The paper investigates a stochastic model where two agents (persons, companies, institutions, states...
We establish how a rich collection of evolutionary games can arise as asymptotically exact descripti...
Abstract. We investigate a class of reinforcement learning dynamics in which each player plays a “re...
A population of agents plays a stochastic dynamic game wherein there is an underlying state process ...
This thesis studies a population of agents facing repeatedly the same decision problem. Each agent k...
In this handout we analyse reinforcement learning algorithms for Markov decision processes. The read...
Abstract The problem of reinforcement learning in a non-Markov environment isexplored using a dynami...
In the paper, we re-investigate the long run behavior of an adaptive learning process driven by the ...
AbstractWe address the problem of reinforcement learning in which observations may exhibit an arbitr...
In this paper, we present a brief survey of reinforcement learning, with particular emphasis on stoc...
This paper introduces a new model, i.e. state-coupled replicator dynamics, expanding the link betwee...
This paper extends the link between evolutionary game theory and multi-agent reinforcement learning ...
We investigate learning in a setting where each period a population has to choose between two action...
We investigate the stability properties of Muth's model of price movements when agents choose a prod...
We address the problem of reinforcement learning in which observations may exhibit an arbitrary form...
The paper investigates a stochastic model where two agents (persons, companies, institutions, states...
We establish how a rich collection of evolutionary games can arise as asymptotically exact descripti...
Abstract. We investigate a class of reinforcement learning dynamics in which each player plays a “re...
A population of agents plays a stochastic dynamic game wherein there is an underlying state process ...
This thesis studies a population of agents facing repeatedly the same decision problem. Each agent k...
In this handout we analyse reinforcement learning algorithms for Markov decision processes. The read...
Abstract The problem of reinforcement learning in a non-Markov environment isexplored using a dynami...
In the paper, we re-investigate the long run behavior of an adaptive learning process driven by the ...
AbstractWe address the problem of reinforcement learning in which observations may exhibit an arbitr...
In this paper, we present a brief survey of reinforcement learning, with particular emphasis on stoc...