Linear-Time Algorithm for Learning Large-Scale Sparse Graphical Models

We consider the graphical lasso, a popular optimization problem for learning the sparse representations of high-dimensional datasets, which is well-known to be computationally expensive for large-scale problems. A recent line of results has shown-under mild assumptions-that the sparsity pattern of t...

Full description

Saved in:
Bibliographic Details
Published inIEEE access Vol. 7; pp. 12658 - 12672
Main Authors Fattahi, Salar, Zhang, Richard Y., Sojoudi, Somayeh
Format Journal Article
LanguageEnglish
Published Piscataway IEEE 2019
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:We consider the graphical lasso, a popular optimization problem for learning the sparse representations of high-dimensional datasets, which is well-known to be computationally expensive for large-scale problems. A recent line of results has shown-under mild assumptions-that the sparsity pattern of the graphical lasso estimator can be retrieved by soft-thresholding the sample covariance matrix. Based on this result, a closed-form solution has been obtained that is optimal when the thresholded sample covariance matrix has an acyclic structure. In this paper, we prove an extension of this result to generalized graphical lasso (GGL), where additional sparsity constraints are imposed based on prior knowledge. Furthermore, we describe a recursive closed-form solution for the problem when the thresholded sample covariance matrix is chordal. By building upon this result, we describe a novel Newton-Conjugate Gradient algorithm that can efficiently solve the GGL with general structures. Assuming that the thresholded sample covariance matrix is sparse with a sparse Cholesky factorization, we prove that the algorithm converges to an <inline-formula> <tex-math notation="LaTeX">\epsilon </tex-math></inline-formula>-accurate solution in <inline-formula> <tex-math notation="LaTeX">O(n\log (1/\epsilon)) </tex-math></inline-formula> time and <inline-formula> <tex-math notation="LaTeX">O(n) </tex-math></inline-formula> memory. The algorithm is highly efficient in practice: we solve instances with as many as 200000 variables to 7-9 digits of accuracy in less than an hour on a standard laptop computer running MATLAB.
ISSN:2169-3536
2169-3536
DOI:10.1109/ACCESS.2018.2890583