Literature DB >> 25419094

LSRN: A PARALLEL ITERATIVE SOLVER FOR STRONGLY OVER- OR UNDERDETERMINED SYSTEMS.

Xiangrui Meng1, Michael A Saunders2, Michael W Mahoney3.   

Abstract

We describe a parallel iterative least squares solver named LSRN that is based on random normal projection. LSRN computes the min-length solution to min x∈ℝ n ‖Ax - b‖2, where A ∈ ℝ m × n with m ≫ n or m ≪ n, and where A may be rank-deficient. Tikhonov regularization may also be included. Since A is involved only in matrix-matrix and matrix-vector multiplications, it can be a dense or sparse matrix or a linear operator, and LSRN automatically speeds up when A is sparse or a fast linear operator. The preconditioning phase consists of a random normal projection, which is embarrassingly parallel, and a singular value decomposition of size ⌈γ min(m, n)⌉ × min(m, n), where γ is moderately larger than 1, e.g., γ = 2. We prove that the preconditioned system is well-conditioned, with a strong concentration result on the extreme singular values, and hence that the number of iterations is fully predictable when we apply LSQR or the Chebyshev semi-iterative method. As we demonstrate, the Chebyshev method is particularly efficient for solving large problems on clusters with high communication cost. Numerical results show that on a shared-memory machine, LSRN is very competitive with LAPACK's DGELSD and a fast randomized least squares solver called Blendenpik on large dense problems, and it outperforms the least squares solver from SuiteSparseQR on sparse problems without sparsity patterns that can be exploited to reduce fill-in. Further experiments show that LSRN scales well on an Amazon Elastic Compute Cloud cluster.

Entities:  

Keywords:  Chebyshev semi-iterative method; LAPACK; LSQR; Tikhonov regularization; iterative method; linear least squares; minimum-length solution; over determined system; parallel computing; preconditioning; random matrix; random projection; random sampling; randomized algorithm; rank-deficient; ridge regression; sparse matrix; underdetermined system

Year:  2014        PMID: 25419094      PMCID: PMC4238893          DOI: 10.1137/120866580

Source DB:  PubMed          Journal:  SIAM J Sci Comput        ISSN: 1064-8275            Impact factor:   2.373


  2 in total

1.  A fast randomized algorithm for overdetermined linear least-squares regression.

Authors:  Vladimir Rokhlin; Mark Tygert
Journal:  Proc Natl Acad Sci U S A       Date:  2008-09-08       Impact factor: 11.205

2.  80 million tiny images: a large data set for nonparametric object and scene recognition.

Authors:  Antonio Torralba; Rob Fergus; William T Freeman
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2008-11       Impact factor: 6.226

  2 in total
  5 in total

1.  Fast Model-Fitting of Bayesian Variable Selection Regression Using the Iterative Complex Factorization Algorithm.

Authors:  Quan Zhou; Yongtao Guan
Journal:  Bayesian Anal       Date:  2018-08-29       Impact factor: 3.728

2.  Weighted SGD for ℓ p Regression with Randomized Preconditioning.

Authors:  Jiyan Yang; Yin-Lam Chow; Christopher Ré; Michael W Mahoney
Journal:  Proc Annu ACM SIAM Symp Discret Algorithms       Date:  2016-01

3.  Signal sampling for efficient sparse representation of resting state FMRI data.

Authors:  Bao Ge; Milad Makkie; Jin Wang; Shijie Zhao; Xi Jiang; Xiang Li; Jinglei Lv; Shu Zhang; Wei Zhang; Junwei Han; Lei Guo; Tianming Liu
Journal:  Brain Imaging Behav       Date:  2016-12       Impact factor: 3.978

4.  A randomized approach to speed up the analysis of large-scale read-count data in the application of CNV detection.

Authors:  WeiBo Wang; Wei Sun; Wei Wang; Jin Szatkiewicz
Journal:  BMC Bioinformatics       Date:  2018-03-01       Impact factor: 3.169

5.  A Dictionary Learning Approach for Signal Sampling in Task-Based fMRI for Reduction of Big Data.

Authors:  Bao Ge; Xiang Li; Xi Jiang; Yifei Sun; Tianming Liu
Journal:  Front Neuroinform       Date:  2018-04-12       Impact factor: 4.081

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.