Coordinate Linear Variance Reduction for Generalized Linear Programming
From MaRDI portal
Publication:6382001
arXiv2111.01842MaRDI QIDQ6382001
Author name not available (Why is that?)
Publication date: 2 November 2021
Abstract: We study a class of generalized linear programs (GLP) in a large-scale setting, which includes simple, possibly nonsmooth convex regularizer and simple convex set constraints. By reformulating (GLP) as an equivalent convex-concave min-max problem, we show that the linear structure in the problem can be used to design an efficient, scalable first-order algorithm, to which we give the name emph{Coordinate Linear Variance Reduction} ( extsc{clvr}; pronounced "clever"). extsc{clvr} yields improved complexity results for (GLP) that depend on the max row norm of the linear constraint matrix in (GLP) rather than the spectral norm. When the regularization terms and constraints are separable, extsc{clvr} admits an efficient lazy update strategy that makes its complexity bounds scale with the number of nonzero elements of the linear constraint matrix in (GLP) rather than the matrix dimensions. On the other hand, for the special case of linear programs, by exploiting sharpness, we propose a restart scheme for extsc{clvr} to obtain empirical linear convergence. Then we show that Distributionally Robust Optimization (DRO) problems with ambiguity sets based on both -divergence and Wasserstein metrics can be reformulated as (GLPs) by introducing sparsely connected auxiliary variables. We complement our theoretical guarantees with numerical experiments that verify our algorithm's practical effectiveness, in terms of wall-clock time and number of data passes.
Has companion code repository: https://github.com/ericlincc/efficient-glp
This page was built for publication: Coordinate Linear Variance Reduction for Generalized Linear Programming
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6382001)