GDA-AM: On the effectiveness of solving minimax optimization via Anderson Acceleration

From MaRDI portal
Publication:6379500

arXiv2110.02457MaRDI QIDQ6379500

Author name not available (Why is that?)

Publication date: 5 October 2021

Abstract: Many modern machine learning algorithms such as generative adversarial networks (GANs) and adversarial training can be formulated as minimax optimization. Gradient descent ascent (GDA) is the most commonly used algorithm due to its simplicity. However, GDA can converge to non-optimal minimax points. We propose a new minimax optimization framework, GDA-AM, that views the GDAdynamics as a fixed-point iteration and solves it using Anderson Mixing to con-verge to the local minimax. It addresses the diverging issue of simultaneous GDAand accelerates the convergence of alternating GDA. We show theoretically that the algorithm can achieve global convergence for bilinear problems under mild conditions. We also empirically show that GDA-AMsolves a variety of minimax problems and improves GAN training on several datasets




Has companion code repository: https://github.com/hehuannb/gda-am








This page was built for publication: GDA-AM: On the effectiveness of solving minimax optimization via Anderson Acceleration

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6379500)