Abstract

Recently, minimax optimization has received renewed focus due to modern applications in machine learning, robust optimization, and reinforcement learning. The scale of these applications naturally leads to the use of first-order methods. However, the nonconvexities and nonconcavities present in these problems, prevents the application of typical gradient descent/ascent, which is known to diverge even in bilinear problems. Recently, it was shown that the proximal point method (PPM) converges linearly for a family of nonconvex–nonconcave problems. In this paper, we study the convergence of a damped version of the extragradient method (EGM), which avoids potentially costly proximal computations, relying only on gradient evaluation. We show that the EGM converges linearly for smooth minimax optimization problems satisfying the same nonconvex–nonconcave condition needed by the PPM. Funding: H. Lu was supported by The University of Chicago Booth School of Business Benjamin Grimmer was supported by Johns Hopkins Applied Mathematics and Statistics Department.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call