On the O(1/T)O(1/T) Convergence of Alternating Gradient Descent–Ascent in Bilinear Games

ICLR 2026 Conference SubmissionAnonymous Authors
Two-player zero-sum gamesAlternating gradient descent-ascentPerformance estimation programming
Abstract:

We study the alternating gradient descent-ascent (AltGDA) algorithm in two-player zero-sum games. Alternating methods, where players take turns to update their strategies, have long been recognized as simple and practical approaches for learning in games, exhibiting much better numerical performance than their simultaneous counterparts. However, our theoretical understanding of alternating algorithms remains limited, and results are mostly restricted to the unconstrained setting. We show that for two-player zero-sum games that admit an interior Nash equilibrium, AltGDA converges at an O(1/T)O(1/T) ergodic convergence rate when employing a small constant stepsize. This is the first result showing that alternation improves over the simultaneous counterpart of GDA in the constrained setting. For games without an interior equilibrium, we show an O(1/T)O(1/T) local convergence rate with a constant stepsize that is independent of any game-specific constants. In a more general setting, we develop a performance estimation programming (PEP) framework to jointly optimize the AltGDA stepsize along with its worst-case convergence rate. The PEP results indicate that AltGDA may achieve an O(1/T)O(1/T) convergence rate for a finite horizon TT, whereas its simultaneous counterpart appears limited to an O(1/T)O(1/\sqrt{T}) rate.

Disclaimer
This report is AI-GENERATED using Large Language Models and WisPaper (A scholar search engine). It analyzes academic papers' tasks and contributions against retrieved prior work. While this system identifies POTENTIAL overlaps and novel directions, ITS COVERAGE IS NOT EXHAUSTIVE AND JUDGMENTS ARE APPROXIMATE. These results are intended to assist human reviewers and SHOULD NOT be relied upon as a definitive verdict on novelty.
NOTE that some papers exist in multiple, slightly different versions (e.g., with different titles or URLs). The system may retrieve several versions of the same underlying work. The current automated pipeline does not reliably align or distinguish these cases, so human reviewers will need to disambiguate them manually.
If you have any questions, please contact: mingzhang23@m.fudan.edu.cn

Overview

Taxonomy

Core-task Taxonomy Papers
40
3
Claimed Contributions
17
Contribution Candidate Papers Compared
0
Refutable Paper

Research Landscape Overview

Core task: convergence analysis of alternating gradient descent-ascent in bilinear games. The field organizes around several complementary perspectives on minimax optimization. Alternating update schemes form a central branch, examining how players take turns updating their strategies and the resulting convergence guarantees in bilinear and broader game settings. Simultaneous and optimistic gradient methods explore synchronous updates and lookahead mechanisms that can stabilize dynamics. Nonconvex-nonconcave minimax optimization tackles settings beyond the classical convex-concave regime, while Markov games and sequential decision settings extend the analysis to multi-agent reinforcement learning. Specialized algorithmic variants introduce momentum, variance reduction, and acceleration techniques, and theoretical frameworks develop tools such as performance estimation problems and co-coercivity conditions. Application-driven branches address constrained optimization and robust formulations that arise in practice. Within alternating update schemes, a particularly active line of work investigates convergence rates in bilinear settings, where the minimax objective decomposes into a simple bilinear form. Alternating GDA Bilinear[0] contributes to this cluster by analyzing how alternating updates behave in such structured games, complementing earlier studies like Local Alternating GDA[3] that examined local convergence properties and Alternating Updates Benefit[4] that highlighted advantages of the alternating schedule. These works contrast with optimistic methods such as Optimistic GDA Bilinear[7] and extragradient approaches like Unified Extragradient Analysis[5], which use lookahead or correction steps to achieve faster or more stable convergence. A recurring theme across these branches is the trade-off between algorithmic simplicity and convergence speed: alternating schemes are often straightforward to implement but may exhibit oscillatory behavior, whereas optimistic and simultaneous methods can offer improved rates at the cost of additional gradient evaluations or memory.

Claimed Contributions

O(1/T) ergodic convergence rate for AltGDA with interior Nash equilibrium

The authors prove that alternating gradient descent-ascent achieves an O(1/T) convergence rate in bilinear games with an interior Nash equilibrium using a constant stepsize. This is the first result showing that alternation improves over simultaneous GDA in the constrained setting.

2 retrieved papers
Local O(1/T) convergence rate with game-independent stepsize

The authors establish a local O(1/T) convergence rate for AltGDA in general bilinear games without requiring an interior equilibrium. The constant stepsize used does not depend on game-specific parameters.

10 retrieved papers
Performance estimation programming framework for joint stepsize and convergence rate optimization

The authors develop a PEP-based framework that simultaneously optimizes both the stepsize and worst-case convergence rate for AltGDA. This is the first instance of stepsize optimization for performance estimation problems involving primal-dual algorithms with linear operators.

5 retrieved papers

Core Task Comparisons

Comparisons with papers in the same taxonomy category

Contribution Analysis

Detailed comparisons for each claimed contribution

Contribution

O(1/T) ergodic convergence rate for AltGDA with interior Nash equilibrium

The authors prove that alternating gradient descent-ascent achieves an O(1/T) convergence rate in bilinear games with an interior Nash equilibrium using a constant stepsize. This is the first result showing that alternation improves over simultaneous GDA in the constrained setting.

Contribution

Local O(1/T) convergence rate with game-independent stepsize

The authors establish a local O(1/T) convergence rate for AltGDA in general bilinear games without requiring an interior equilibrium. The constant stepsize used does not depend on game-specific parameters.

Contribution

Performance estimation programming framework for joint stepsize and convergence rate optimization

The authors develop a PEP-based framework that simultaneously optimizes both the stepsize and worst-case convergence rate for AltGDA. This is the first instance of stepsize optimization for performance estimation problems involving primal-dual algorithms with linear operators.