Abstract
We establish the fundamental limits of lossless analog compression by considering the recovery of arbitrary random vectors $ {{ {\mathsf {x}}}} \in {\mathbb R} ^{m}$ from the noiseless linear measurements $ {{ {\mathsf {y}}}} = {{ {A}}} {{ {\mathsf {x}}}} $ with measurement matrix $ {{ {A}}} \in {\mathbb R} ^{n\times m}$ . Our theory is inspired by the groundbreaking work of Wu and Verdu (2010) on almost lossless analog compression but applies to the nonasymptotic, i.e., fixed- $m$ case, and considers zero error probability. Specifically, our achievability result states that, for Lebesgue-almost all $ {{ {A}}} $ , the random vector $ {{ {\mathsf {x}}}} $ can be recovered with zero error probability provided that $n>K( {{ {\mathsf {x}}}} )$ , where $K( {{ {\mathsf {x}}}} )$ is given by the infimum of the lower modified Minkowski dimension over all support sets $ {\mathcal {U}} $ of $ {{ {\mathsf {x}}}} $ (i.e., sets $ {\mathcal {U}} \subseteq {\mathbb R} ^{m}$ with $ {\mathrm {P}} [ {{ {\mathsf {x}}}} \in {\mathcal {U}} ]=1$ ). We then particularize this achievability result to the class of $s$ -rectifiable random vectors as introduced in Koliander et al. (2016); these are random vectors of absolutely continuous distribution—with respect to the $s$ -dimensional Hausdorff measure—supported on countable unions of $s$ -dimensional $C^{1}$ -submanifolds of $ {\mathbb R} ^{m}$ . Countable unions of $C^{1}$ -submanifolds include essentially all signal models used in the compressed sensing literature such as the standard union of subspaces model underlying much of compressed sensing theory and spectrum-blind sampling, smooth submanifolds, block-sparsity, and low-rank matrices as considered in the matrix completion problem. Specifically, we prove that, for Lebesgue-almost all $ {{ {A}}} $ , $s$ -rectifiable random vectors $ {{ {\mathsf {x}}}} $ can be recovered with zero error probability from $n>s$ linear measurements. This threshold is, however, found not to be tight as exemplified by the construction of an $s$ -rectifiable random vector that can be recovered with zero error probability from $n linear measurements. Motivated by this observation, we introduce the new class of $s$ -analytic random vectors, which admit a strong converse in the sense of $n\geq s$ being necessary for recovery with probability of error smaller than one. The central conceptual tools in the development of our theory are geometric measure theory and the theory of real analytic functions.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.