We study the problem of finding the Nash equilibrium in a two-player zero-sum Markov game. Due to its formulation as a minimax optimization program, a natural approach to solve the problem is to perform gradient descent/ascent with respect to each player in an alternating fashion. However, due to the non-convexity/non-concavity of the underlying objective function, theoretical understandings of this method are limited. In our paper, we consider solving an entropy-regularized variant of the Markov game. The regularization introduces structure into the optimization landscape that make the solutions more identifiable and allow the problem to be solved more efficiently. Our main contribution is to show that under proper choices of the regulariz...
In this paper, we consider two-person zero-sum discounted Markov games with finite state and action ...
This paper considers the two-person zero-sum Markov game with finite state and action spaces at the ...
We study episodic two-player zero-sum Markov games (MGs) in the offline setting, where the goal is t...
We introduce a new algorithm for the numerical computation of Nash equilibria of competitive two-pla...
In this paper we present a novel method for finding the strong Nash equilibrium. The approach consis...
Algorithms designed for single-agent reinforcement learning (RL) generally fail to converge to equil...
Koller, Megiddo and von Stengel showed how to efficiently compute minimax strategies for two-player ...
This paper studies policy optimization algorithms for multi-agent reinforcement learning. We begin b...
Game theory and online optimization have a close relationship with each other. In some literature, o...
We study what dataset assumption permits solving offline two-player zero-sum Markov games. In stark ...
An ideal strategy in zero-sum games should not only grant the player an average reward no less than ...
Summarization: This paper investigates value function approximation in the context of zero-sum Marko...
International audienceThis paper reports theoretical and empirical investigations on the use of quas...
International audienceThe main contribution of this paper consists in extending several non-st...
In this paper an overview will be presented of the applicability of successive approximation methods...
In this paper, we consider two-person zero-sum discounted Markov games with finite state and action ...
This paper considers the two-person zero-sum Markov game with finite state and action spaces at the ...
We study episodic two-player zero-sum Markov games (MGs) in the offline setting, where the goal is t...
We introduce a new algorithm for the numerical computation of Nash equilibria of competitive two-pla...
In this paper we present a novel method for finding the strong Nash equilibrium. The approach consis...
Algorithms designed for single-agent reinforcement learning (RL) generally fail to converge to equil...
Koller, Megiddo and von Stengel showed how to efficiently compute minimax strategies for two-player ...
This paper studies policy optimization algorithms for multi-agent reinforcement learning. We begin b...
Game theory and online optimization have a close relationship with each other. In some literature, o...
We study what dataset assumption permits solving offline two-player zero-sum Markov games. In stark ...
An ideal strategy in zero-sum games should not only grant the player an average reward no less than ...
Summarization: This paper investigates value function approximation in the context of zero-sum Marko...
International audienceThis paper reports theoretical and empirical investigations on the use of quas...
International audienceThe main contribution of this paper consists in extending several non-st...
In this paper an overview will be presented of the applicability of successive approximation methods...
In this paper, we consider two-person zero-sum discounted Markov games with finite state and action ...
This paper considers the two-person zero-sum Markov game with finite state and action spaces at the ...
We study episodic two-player zero-sum Markov games (MGs) in the offline setting, where the goal is t...