Skip to Main Content
We study sparse representations and sparse approximations with respect to incoherent dictionaries. We address the problem of designing and analyzing greedy methods of approximation. A key question in this regard is: How to measure efficiency of a specific algorithm? Answering this question, we prove the Lebesgue-type inequalities for algorithms under consideration. A very important new ingredient of the paper is that we perform our analysis in a Banach space instead of a Hilbert space. It is known that in many numerical problems, users are satisfied with a Hilbert space setting and do not consider a more general setting in a Banach space. There are known arguments that justify interest in Banach spaces. In this paper, we give one more argument in favor of consideration of greedy approximation in Banach spaces. We introduce a concept of M-coherent dictionary in a Banach space which is a generalization of the corresponding concept in a Hilbert space. We analyze the quasi-orthogonal greedy algorithm (QOGA), which is a generalization of the orthogonal greedy algorithm (orthogonal matching pursuit) for Banach spaces. It is known that the QOGA recovers exactly S -sparse signals after S iterations provided S <; (1+1/M)/2. This result is well known for the orthogonal greedy algorithm in Hilbert spaces. The following question is of great importance: Are there dictionaries in BBRn such that their coherence in lpn is less than their coherence in l2n for some p ∈ (1,∞)? We show that the answer to the above question is “yes.” Thus, for such dictionaries, replacing the Hilbert space l2n by a Banach space lpn , we improve an upper bound for sparsity that guarantees an exact recovery of a signal.