A Flexible ADMM Algorithm for Big Data Applications
We present a flexible Alternating Direction Method of Multipliers (F-ADMM) algorithm for solving optimization problems involving a strongly convex objective function that is separable into $n \geq 2$ blocks, subject to (non-separable) linear equality constraints. The F-ADMM algorithm uses a Gauss-Se...
Saved in:
Main Authors | , |
---|---|
Format | Journal Article |
Language | English |
Published |
15.02.2015
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | We present a flexible Alternating Direction Method of Multipliers (F-ADMM)
algorithm for solving optimization problems involving a strongly convex
objective function that is separable into $n \geq 2$ blocks, subject to
(non-separable) linear equality constraints. The F-ADMM algorithm uses a
Gauss-Seidel scheme to update blocks of variables, and a regularization term is
added to each of the subproblems arising within F-ADMM. We prove, under common
assumptions, that F-ADMM is globally convergent.
We also present a special case of F-ADMM that is partially parallelizable,
which makes it attractive in a big data setting. In particular, we partition
the data into groups, so that each group consists of multiple blocks of
variables. By applying F-ADMM to this partitioning of the data, and using a
specific regularization matrix, we obtain a hybrid ADMM (H-ADMM) algorithm: the
grouped data is updated in a Gauss-Seidel fashion, and the blocks within each
group are updated in a Jacobi manner. Convergence of H-ADMM follows directly
from the convergence properties of F-ADMM. Also, a special case of H-ADMM can
be applied to functions that are convex, rather than strongly convex. We
present numerical experiments to demonstrate the practical advantages of this
algorithm. |
---|---|
DOI: | 10.48550/arxiv.1502.04391 |