Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published 2014 | Submitted
Book Section - Chapter Open

Simple Error Bounds for Regularized Noisy Linear Inverse Problems

Abstract

Consider estimating a structured signal x_0 from linear, underdetermined and noisy measurements y = Ax_0+z, via solving a variant of the lasso algorithm: x̂ = arg min_x{∥y-Ax∥+2 + λf(x)}. Here, f is a convex function aiming to promote the structure of x_0, say ℓ_1-norm to promote sparsity or nuclear norm to promote low-rankness. We assume that the entries of A are independent and normally distributed and make no assumptions on the noise vector z, other than it being independent of A. Under this generic setup, we derive a general, non-asymptotic and rather tight upper bound on the ℓ_2-norm of the estimation error ∥x̂ - x0∥2. Our bound is geometric in nature and obeys a simple formula; the roles of λ, f and x_0 are all captured by a single summary parameter δ(λ∂f(x_0)), termed the Gaussian squared distance to the scaled subdifferential. We connect our result to the literature and verify its validity through simulations.

Additional Information

© 2014 IEEE. This work was supported in part by the National Science Foundation under grants CCF-0729203, CNS-0932428 and CIF-1018927, by the Office of Naval Research under the MURI grant N00014-08-1-0747, and by a grant from Qualcomm Inc.

Attached Files

Submitted - Simple_Error_Bounds_for_Regularized_Noisy_Linear_Inverse_Problems.pdf

Files

Simple_Error_Bounds_for_Regularized_Noisy_Linear_Inverse_Problems.pdf
Files (207.6 kB)

Additional details

Created:
August 19, 2023
Modified:
March 5, 2024