A Universal Analysis of Large-Scale Regularized Least Squares Solutions
Paper in proceeding, 2017

A problem that has been of recent interest in statistical inference, machine learning and signal processing is that of understanding the asymptotic behavior of regularized least squares solutions under random measurement matrices (or dictionaries). The Least Absolute Shrinkage and Selection Operator (LASSO or least-squares with \ell_1 regularization) is perhaps one of the most interesting examples. Precise expressions for the asymptotic performance of LASSO have been obtained for a number of different cases, in particular when the elements of the dictionary matrix are sampled independently from a Gaussian distribution. It has also been empirically observed that the resulting expressions remain valid when the entries of the dictionary matrix are independently sampled from certain non-Gaussian distributions. In this paper, we confirm these observations theoretically when the distribution is sub-Gaussian. We further generalize the previous expressions for a broader family of regularization functions and under milder conditions on the underlying random, possibly non-Gaussian, dictionary matrix. In particular, we establish the universality of the asymptotic statistics (e.g., the average quadratic risk) of LASSO with non-Gaussian dictionaries.

Author

Ashkan Panahi

North Carolina State University

Babak Hassibi

California Institute of Technology (Caltech)

Advances in Neural Information Processing Systems

1049-5258 (ISSN)

3382-3391

30st Conference on Neural Information Processing Systems
Long beach, CA, USA,

Subject Categories

Information Science

Probability Theory and Statistics

Control Engineering

Signal Processing

More information

Latest update

4/14/2022