Abstract
Sparse recovery aims to reconstruct sparse signals from compressed linear measurements. In this paper, we propose a sparse recovery algorithm called multiple orthogonal least squares (MOLS), which extends the well-known orthogonal least squares (OLS) algorithm by allowing multiple $L$ indices to be selected per iteration. Owing to its ability to catch multiple support indices in each selection, MOLS often converges in fewer iterations and hence improves the computational efficiency over the conventional OLS algorithm. Theoretical analysis shows that MOLS ( $L > 1$ ) performs exact recovery of $K$ -sparse signals ( $K > 1$ ) in at most $K$ iterations, provided that the sensing matrix obeys the restricted isometry property with isometry constant $\delta _{LK} When $L = 1,$ MOLS reduces to the conventional OLS algorithm and our analysis shows that exact recovery is guaranteed under $\delta_{K +1} . This condition is nearly optimal with respect to $\delta _{K+1}$ in the sense that, even with a small relaxation (e.g., $\delta_{K + 1} = 1 / \sqrt{K}$ ), exact recovery with OLS may not be guaranteed. The recovery performance of MOLS in the noisy scenario is also studied. It is shown that stable recovery of sparse signals can be achieved with the MOLS algorithm when the signal-to-noise ratio scales linearly with the sparsity level of input signals.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.