We study the problem of convergence to a stationary point in zero-sum games. We propose competitive gradient optimization (CGO ), a gradient-based method that incorporates the interactions between the two players in zero-sum games for optimization updates. We provide continuous-time analysis of CGO and its convergence properties while showing that in the continuous limit, CGO predecessors degenerate to their gradient descent ascent (GDA) variants. We provide a rate of convergence to stationary points and further propose a generalized class of $\alpha$-coherent function for which we provide convergence analysis. We show that for strictly $\alpha$-coherent functions, our algorithm convergences to a saddle point. Moreover, we propose optimisti...
This paper considers the analysis of continuous time gradient-based optimization algorithms through ...
Cheung and Piliouras (2020) recently showed that two variants of the Multiplicative Weights Update m...
Several widely-used first-order saddle-point optimization methods yield an identical continuous-time...
We introduce a new algorithm for the numerical computation of Nash equilibria of competitive two-pla...
A core challenge in policy optimization in competitive Markov decision processes is the design of ef...
We provide several applications of Optimistic Mirror Descent, an online learning algorithm based on ...
International audienceIn this paper, we examine the equilibrium tracking and convergence properties ...
43 pages, 2 tablesInternational audienceLearning in stochastic games is a notoriously difficult prob...
In this work we look at the recent results in policy gradient learning in a general-sum game scenari...
Game theory and online optimization have a close relationship with each other. In some literature, o...
International audienceIn this paper, we examine the convergence rate of a wide range of regularized ...
© Constantinos Daskalakis and Ioannis Panageas. Motivated by applications in Game Theory, Optimizati...
This paper examines the convergence of a broad class of distributed learning dynamics for games with...
A key challenge in smooth games is that there is no general guarantee for gradient methods to conver...
We study the problem of finding the Nash equilibrium in a two-player zero-sum Markov game. Due to it...
This paper considers the analysis of continuous time gradient-based optimization algorithms through ...
Cheung and Piliouras (2020) recently showed that two variants of the Multiplicative Weights Update m...
Several widely-used first-order saddle-point optimization methods yield an identical continuous-time...
We introduce a new algorithm for the numerical computation of Nash equilibria of competitive two-pla...
A core challenge in policy optimization in competitive Markov decision processes is the design of ef...
We provide several applications of Optimistic Mirror Descent, an online learning algorithm based on ...
International audienceIn this paper, we examine the equilibrium tracking and convergence properties ...
43 pages, 2 tablesInternational audienceLearning in stochastic games is a notoriously difficult prob...
In this work we look at the recent results in policy gradient learning in a general-sum game scenari...
Game theory and online optimization have a close relationship with each other. In some literature, o...
International audienceIn this paper, we examine the convergence rate of a wide range of regularized ...
© Constantinos Daskalakis and Ioannis Panageas. Motivated by applications in Game Theory, Optimizati...
This paper examines the convergence of a broad class of distributed learning dynamics for games with...
A key challenge in smooth games is that there is no general guarantee for gradient methods to conver...
We study the problem of finding the Nash equilibrium in a two-player zero-sum Markov game. Due to it...
This paper considers the analysis of continuous time gradient-based optimization algorithms through ...
Cheung and Piliouras (2020) recently showed that two variants of the Multiplicative Weights Update m...
Several widely-used first-order saddle-point optimization methods yield an identical continuous-time...