JMLR

Two-Timescale Gradient Descent Ascent Algorithms for Nonconvex Minimax Optimization

Authors
Tianyi Lin Chi Jin Michael I. Jordan
Research Topics
Computational Statistics
Paper Information
  • Journal:
    Journal of Machine Learning Research
  • Added to Tracker:
    Jul 15, 2025
Abstract

We provide a unified analysis of two-timescale gradient descent ascent (TTGDA) for solving structured nonconvex minimax optimization problems in the form of $\min_x \max_{y \in Y} f(x, y)$, where the objective function $f(x, y)$ is nonconvex in $x$ and concave in $y$, and the constraint set $Y \subseteq \mathbb{R}^n$ is convex and bounded. In the convex-concave setting, the single-timescale gradient descent ascent (GDA) algorithm is widely used in applications and has been shown to have strong convergence guarantees. In more general settings, however, it can fail to converge. Our contribution is to design TTGDA algorithms that are effective beyond the convex-concave setting, efficiently finding a stationary point of the function $\Phi(\cdot) := \max_{y \in Y} f(\cdot, y)$. We also establish theoretical bounds on the complexity of solving both smooth and nonsmooth nonconvex-concave minimax optimization problems. To the best of our knowledge, this is the first systematic analysis of TTGDA for nonconvex minimax optimization, shedding light on its superior performance in training generative adversarial networks (GANs) and in other real-world application problems.

Author Details
Tianyi Lin
Author
Chi Jin
Author
Michael I. Jordan
Author
Research Topics & Keywords
Computational Statistics
Research Area
Citation Information
APA Format
Tianyi Lin , Chi Jin & Michael I. Jordan . Two-Timescale Gradient Descent Ascent Algorithms for Nonconvex Minimax Optimization. Journal of Machine Learning Research .
BibTeX Format
@article{JMLR:v26:22-0863,
  author  = {Tianyi Lin and Chi Jin and Michael I. Jordan},
  title   = {Two-Timescale Gradient Descent Ascent Algorithms for Nonconvex Minimax Optimization},
  journal = {Journal of Machine Learning Research},
  year    = {2025},
  volume  = {26},
  number  = {11},
  pages   = {1--45},
  url     = {http://jmlr.org/papers/v26/22-0863.html}
}
Related Papers