18.8. Restricted Isometry Property#

This section dwells deep into the implications of restricted isometry property.

Definition 18.31 (Restricted isometry property)

A matrix ΦCM×N is said to satisfy the RIP (restricted isometry property) of order K with a constant δ(0,1) if the following holds:

(18.43)#(1δ)x22Φx22(1+δ)x22

for every xΣK where

ΣK={xCN:x0K}

is the set of all K-sparse vectors in CN.

Definition 18.32 (Restricted isometry constant)

If a matrix ΦCM×N satisfies RIP of order K then the smallest value of δ (denoted as δK) for which the following holds

(18.44)#(1δ)x22Φx22(1+δ)x22xΣK

is known as the K-th restricted isometry constant for Φ. It is also written in short as K-th RIP constant. We write the bounds as in terms of δK as

(18.45)#(1δK)x22Φx22(1+δK)x22xΣK.

Some remarks are in order.

  • Φ maps a vector xΣKCN into CM as a vector Φx (usually M<N).

  • We will call ΦxCM as an embedding of xCN into CM.

  • RIP quantifies the idea as to how much the squared length of a sparse signal changes during this embedding process.

  • We can compare matrices satisfying RIP with orthonormal bases.

  • An orthonormal basis or the corresponding unitary matrix preserves the length of a vector exactly.

  • A matrix Φ satisfying RIP of order K is able to preserve the length of K sparse signals approximately (the approximation range given by δK).

  • In this sense we can say that Φ implements a restricted almost orthonormal system [19].

  • By restricted we mean that orthonormality is limited to K-sparse signals.

  • By almost we mean that the squared length is not preserved exactly. Rather it is preserved approximately.

  • An arbitrary matrix Φ need not satisfy RIP of any order at all.

  • If Φ satisfies RIP of order K then it is easy to see that Φ satisfies RIP of any order L<K (since ΣLΣK whenever L<K).

  • If Φ satisfies RIP of order K, then it may or many not satisfy RIP of order L>K.

  • Restricted isometry constant is a function of sparsity level K of the signal xCN.

Example 18.37 (Restricted isometry constant)

As a running example in this section we will use following matrix

Φ=12[11111111111111111111111111111111]R4×8.

Consider

x=(20000310)

which is a 3-sparse vector in R8. We have

y=Φx=(0133)

Now

x22=14,x2=3.7417

and

y22=19,y2=4.3589.

We note that

y22x22=1.3571.

With this much information, all we can say that δ3.3571 for this matrix Φ if Φ satisfies RIP of order 3 since we haven’t examined all possible 3-sparse vectors.

Still what is comforting to note is that for this particular example, the distance hasn’t increased by a large factor.

For a given K-sparse vector x, let J denote the support of x; i.e.,

J={1iN|xi0}.

In the running example

J={1,6,7}.

We define xJCK to be the vector formed by keeping the elements in x indexed by J and dropping of other elements (the zero elements). Note that the order of elements is preserved. In the running example,

xJ=(231).

Let ΦJ be the corresponding sub-matrix by choosing columns from Φ indexed by the set J. Note that the order of columns is preserved. In the running example

ΦJ=12[111111111111]R4×3.

It is easy to see that

y=Φx=ΦJxJ.

There are (NK) ways of choosing a K-sparse support for x. Thus we have to consider (NK) corresponding submatrices ΦJ.

For each such submatrix ΦJ, the RIP bounds can be rewritten as

(18.46)#(1δK)x22ΦJx22(1+δK)x22

for every xCK. Note that

ΦJx22=(ΦJx)H(ΦJx)=xHΦJHΦJx.

Theorem 18.44

An M×N matrix Φ cannot satisfy RIP of order K>M.

Proof. This comes from the fact that for a wide matrix rankΦM.

  1. Since every ϕjCM hence any set of M+1 columns in Φ is linearly dependent.

  2. Thus there exists a nonzero M+1 sparse signal xCN such that Φx=0 (it belongs to the null space of the chosen M+1 columns).

  3. RIP (18.43) requires that a nonzero vector be embedded as a nonzero vector.

  4. Thus Φ cannot satisfy RIP of order M+1.

  5. The argument can be easily extended for any K>M.

Theorem 18.45

If Φ satisfies RIP of order l then it satisfies RIP of order k where k<l.

Proof. Every k sparse signal is also l sparse signal. Thus if Φ satisfies RIP of order l then it automatically satisfies RIP of order k<l.

Theorem 18.46

Let Φ satisfy RIP of order k and l where k<l. Then δkδl. In other words, restricted isometry constants are non-decreasing.

Proof. Since every k sparse signal is also l sparse signal, hence for every xΣk following must be satisfied

(1δk)x22Φx22(1+δk)x22

and

(1δl)x22Φx22(1+δl)x22.

Since δk is smallest such value for which these inequalities are satisfied hence δl cannot be smaller than δk.

18.8.1. The First Restricted Isometry Constant#

We consider the simplest case where K=1. We can write Φ in terms of its column vectors

Φ=[ϕ1ϕN].
  1. Now a 1-sparse vector x consists of only one nonzero entry.

  2. Say that x is nonzero at index j.

  3. Then Φx is nothing but xjϕj.

  4. With this the restricted isometry inequality can be written as

    (1δ1)|xj|2xjϕj22(1+δ1)|xj|2.
  5. Dividing by |xj|2 we get

    (1δ1)ϕj22(1+δ1).

Let us formalize this in the following theorem.

Theorem 18.47 (Restricted isometry constants of order 1)

If a matrix Φ satisfies RIP of order K1 then the squared lengths of columns of Φ satisfy the following bounds

(18.47)#1δ1ϕj221+δ11jN.

When δ1=0 then all columns of Φ are unit norm. Now if columns of Φ span CM then Φ can also be considered as a dictionary for CM (see Definition 18.4).

Remark 18.8

A dictionary (Definition 18.4) satisfies RIP of order 1 with δ1=0.

18.8.2. Sums and Differences of Sparse Vectors#

Theorem 18.48

Let x,yCN with xΣk and yΣl; i.e., x0k and y0l. Then

(1δk+l)x±y22Φx±Φy22(1+δk+l)x±y22

as long as Φ satisfies RIP of order k+l.

Proof. We know that

x±y0x0+y0k+l.

Thus x±yΣk+l. The result follows.

18.8.3. Distance Between Sparse Vectors#

Let x,yΣK. Then clearly xyΣ2K.

The 2 distance between vectors is given by

d(x,y)=xy2=(xy)H(xy).

Now if Φ satisfies RIP of order 2K then we can see that it approximately preserves 2 distances between K-sparse vectors.

Theorem 18.49 (Approximation preservation of distances)

Let x,yΣKCN. Let Φx,ΦyCM be corresponding embeddings. If Φ satisfies RIP of order 2K, then

(1δ2K)d2(x,y)d2(Φx,Φy)(1+δ2K)d2(x,y).

Proof. Since Φ satisfies RIP of order 2K hence for every vector vΣ2K we have

(1δ2K)v22Φv22(1+δ2K)v22.

But then xyΣ2K for every x,yΣK and

d2(x,y)=xy22

and

d2(Φx,Φy)=ΦxΦy22=Φ(xy)22.

Thus we have the result.

18.8.4. RIP with Unit Length Sparse Vectors#

Sometimes it is convenient to state RIP in terms of unit length sparse vectors.

Theorem 18.50 (RIP for unit length sparse vectors)

Let x be some arbitrary unit length (i.e., x2=1) vector belonging to ΣK. A matrix Φ is said to satisfy RIP of order K if and only if the following holds

(18.48)#(1δK)Φx22(1+δK)

for every xΣK with x2=1.

Proof. If Φ satisfies RIP of order K then by putting x2=1 in (18.43) we get (18.48).

Now the converse.

  1. We assume (18.48) holds for all unit norm vectors xΣK.

  2. We need to show that (18.43) holds for all xΣK.

  3. For x=0 the bounds in (18.43) are trivially satisfied.

  4. Let xΣK be some nonzero vector.

  5. Let x^=xx2.

  6. Clearly x^ is unit length. Hence

    (1δK)Φx^22(1+δK)(1δK)Φxx222(1+δK)(1δK)x22Φx22(1+δK)x22.
  7. Thus Φ satisfies RIP of order K.

18.8.5. Singular and Eigen Values of K-Submatrices#

Consider any index set J{1,,N} with |J|=K. Let ΦJ be a sub matrix of Φ consisting of columns indexed by J. Assume KM. We define

GΦJHΦJCK×K

as the Gram matrix for columns of ΦJ (see Gram Matrices).

We consider the eigen values of G given by

Gx=λx

for some xCK and x0. We will show that eigen values of G are bounded by RIP constant.

In the running example

G=[100.5010.50.50.51].

Eigen values of G are (0.2929,1,1.7071).

Theorem 18.51

Let Φ satisfy the RIP of order K where KM. Let ΦJ be any sub matrix of Φ with K columns. Then the eigen values of G=ΦJHΦJ lie in the range [1δK,1+δK].

Proof. We note that GCK×K.

  1. Let λ be some eigen value of G.

  2. Let xCK be a corresponding (nonzero) eigenvector.

  3. Then

    Gx=λxxHGx=xHλxxHΦJHΦJx=λx22ΦJx22=λx22.
  4. From (18.46) we recall that δK RIP bounds apply for each vector in xCK for a K-column submatrix ΦJ given by

    (1δK)x22ΦJx22(1+δK)x22.
  5. Thus

    (1δK)x22λx22(1+δK)x22(1δK)λ(1+δK)

    since x0.

Corollary 18.3

Let Φ satisfy the RIP of order K where KM. Let ΦJ be any submatrix of Φ with K columns. Then the Gram matrix G=ΦJHΦJ is full rank and invertible. Moreover G is positive definite.

Proof. From Theorem 18.51, the eigen values are in range [1δK,1+δK].

  1. Since Φ satisfies RIP of order K, hence δK<1.

  2. Hence all eigenvalues of G are positive.

  3. Hence G is positive definite.

  4. Hence their product is positive.

  5. Thus det(G) is nonzero.

  6. Hence G is invertible.

Theorem 18.52

Let Φ satisfy the RIP of order K where KM. Let ΦJ be any submatrix of Φ with K columns. Then all singular values of ΦJ are nonzero and they are in the range given by

1δKσ1+δK

where σ is a singular value of ΦJ.

Proof. This is straight forward application of Lemma 4.74 and Theorem 18.51. Eigen values of ΦJHΦJ are nothing but squares of the singular values of ΦJ.

Corollary 18.4

Let Φ satisfy the RIP of order K where KM. Let ΦJ be any submatrix of Φ with k columns where kK. Then the singular values of ΦJ are nonzero and they are in the range given by

1δKσ1+δK

where σ is a singular value of ΦJ.

Proof. Let σ be a singular value of ΦJ.

  1. Since Φ satisfies RIP of order K, it also satisfies RIP of order kK.

  2. From Theorem 18.52 we have

    1δkσ1+δk.
  3. From Theorem 18.46 we have δkδK.

  4. Thus

    1δK1δk,1+δk1+δK.
  5. Thus

    1δKσ1+δK.

Theorem 18.53

Let Φ satisfy the RIP of order K where KM. Let ΦJ be any submatrix of Φ with k columns where kK. Then the eigen values of ΦJHΦJ+rI lie in the range

[1δK+r,1+δK+r].

Moreover consider Δ=ΦJHΦJI. Then

Δ2δK.

Proof. .

  1. From Theorem 18.51 eigen values of ΦJHΦJ lie in the range [1δK,1+δK].

  2. From Lemma 4.69 λ is an eigen value of ΦJHΦJ if and only if λ+r is an eigen value of ΦJHΦJ+rI.

  3. Hence the result.

Now for Δ=ΦJHΦJI

  1. The eigen values lie in the range [δK,δK].

  2. Thus for every eigen value of Δ we have |λ|δK.

  3. Since Δ is Hermitian, its spectral norm is nothing but its largest eigen value.

  4. Hence

    Δ2δK.

From previous few results we see that bound over eigen values of ΦJHΦJ given by (1δK)λ(1+δK) is a necessary condition for Φ to satisfy RIP of order K. We now show that this is also a sufficient condition.

Theorem 18.54

Let Φ be an M×N matrix with MN. Let J{1,,N} be any index set with |J|=KM. Let ΦJ be the K-column sub-matrix of Φ indexed by J. Let G=ΦJHΦJ be the Gram matrix of columns of ΦJ. Let the eigen values of G be λ. If there exists a number δ(0,1) such that

1δλ1+δ

for every eigen value of G for every K column submatrix of Φ, then Φ satisfies RIP of order K.

Alternatively, let Δ=GI. If

Δ2δ<1

for every K column submatrix of Φ then Φ satisfies RIP of order K.

Alternatively, if singular values of ΦJ satisfy

1δσ1+δ

for every ΦJ then Φ satisfies RIP of order K.

Proof. Equivalence of sufficient conditions

  1. We note that eigen values of G are related to eigen values of Δ by the relation (see Lemma 4.69)

    λG1=λΔΛG=1+λΔ.
  2. Hence

    Δ2δδλΔδ1δλG1+δ.
  3. Thus the first two sufficient conditions are equivalent.

  4. Lastly the eigen values of G are squares of singular values of ΦJ.

  5. Thus all sufficient conditions are equivalent.

Proof of sufficient condition

  1. Now let xΣK be an arbitrary vector.

  2. Let J=supp(x).

  3. Clearly |J|K. If |J|<K then augment J by adding some indices arbitrarily till we get |J|=K.

  4. Clearly xJ is an arbitrary vector in CK and Φx=ΦJxJ.

  5. Now let λ1 be the largest and λk be the smallest eigen value of G=ΦJHΦJ.

  6. G is Hermitian and all its eigen values are positive, hence it is positive definite.

  7. From Lemma 4.68 we get

    λkx22xHGxλ1x22xCK.
  8. Applying the limits on the eigen values and using xHGx=ΦJx22, we get

    (1δ)x22ΦJx22(1+δ)x22xCK.
  9. Since this holds for every index set J with |J|=K hence an equivalent statement is

    (1δ)x22Φx22(1+δ)x22xΣKCN.
  10. Thus Φ indeed satisfies RIP of order K with some δK not larger than δ.

Theorem 18.55

Let Φ satisfy the RIP of order K where KM. Let ΦJ be any submatrix of Φ with k columns where kK. Let ΦJ be its Moore-Penrose pseudo-inverse. Then the singular values of ΦJ are nonzero and they are in the range given by

11+δKσ11δK

where σ is a singular value of ΦJ.

Proof. Construction of pseudoinverse of a matrix through its singular value decomposition is discussed in Lemma 4.79.

  1. Lemma 4.80 shows that if σ is a nonzero singular value of ΦJ then 1σ is a nonzero singular value of ΦJ.

  2. From Corollary 18.4 we have that if 1σ is a singular value of ΦJ then,

    1δK1σ1+δK
  3. Inverting the terms in the inequalities we get our result.

Theorem 18.56

Eigen values of G=ΦJHΦJ provide a lower bound on δK given by

δKmax(1λmin,λmax1)

where J is some index set choosing K columns of Φ and δK is the K-th restricted isometry constant for Φ.

In other words, singular values of ΦJ provide a lower bound on δK given by

δKmax(1σmin2,σmax21)

Proof. Obvious.

In the running example, the bounds tell us that

δ30.7071.

Certainly we have to consider all possible (NK) sub-matrices ΦJ to come up with an overall lower bound on δK.

This result doesn’t provide us any upper bound on δK.

Theorem 18.57

Let Φ satisfy the RIP of order K where KM. Let ΦJ be any submatrix of Φ with k columns where kK. Then

ΦJx21+δKx2xCk.

Moreover

ΦJHy21+δKy2yCM.

Proof. We note that ΦJ is an M×k matrix.

  1. Let σ1 be the largest singular value of ΦJ.

  2. Then by Lemma 4.77 we have

    ΦJx2σ1x2xCk.

    and

    ΦJHy2σ1y2yCM.
  3. From Theorem 18.52 and Corollary 18.4 we get

    σ11+δK.
  4. This completes the proof.

First inequality is a restatement of restricted isometry property in (18.46). Second inequality is interesting. In compressive sensing terms, y is a measurement vector and we are using ΦJH to project y back into CN over a k sparse support identified by J. The inequality provides an upper bound on how much the length can increase during this operation.

Theorem 18.58

Let Φ satisfy the RIP of order K where KM. Let ΦJ be any submatrix of Φ with k columns where kK. Let ΦJ be its Moore-Penrose pseudo-inverse. Then

ΦJy211δKy2yCM.

Proof. We note that ΦJ is an k×M matrix.

  1. Let σ1 be the largest singular value of ΦJ.

  2. Then by Lemma 4.77 we have

    ΦJy2σ1y2yCM.
  3. From Theorem 18.55 we see that singular values of ΦJ satisfy the inequalities

    11+δKσ11δK.
  4. Thus

    σ111δK.
  5. Plugging it in we get

    ΦJy211δKy2yCM.

In the previous theorem we saw that back-projection using ΦJH had an upper bound on how much the length of measurement vector could increase. In this theorem we see another upper bound on how much the length of measurement vector can increase when back projected using the pseudo inverse of ΦJ.

Theorem 18.59

Let Φ satisfy the RIP of order K where KM. Let ΦJ be any submatrix of Φ with k columns where kK. Then

(1δK)x2ΦJHΦJx2(1+δK)x2xCk.

Moreover

11+δKx2(ΦJHΦJ)1x211δKx2xCk.

Proof. We note that ΦJ is a full column rank tall matrix.

  1. We recall that all singular values of ΦJ are positive and are bounded by (Corollary 18.4):

    1δKσkσ11+δK

    where σ1,,σk are the singular values of ΦJ (in descending order).

  2. We note that ΦJHΦJ is an k×k matrix which is invertible (Corollary 18.3).

  3. From Lemma 4.83 we get

    σk2x2ΦJHΦJx2σ12x2xCk.
  4. Applying the bounds on σi we get the result

    (1δK)x2ΦJHΦJx2(1+δK)x2xCk.
  5. From Lemma 4.85 we have the bounds for (ΦJHΦJ)1 given by

    1σ12x2(ΦJHΦJ)1x21σk2x2xCk.
  6. Applying the bounds on σi we get the result

    11+δKx2(ΦJHΦJ)1x211δKx2xCk.

In the sequel we will discuss that ΦHΦx can work as a very good proxy for the signal x. The results in this theorem are very comforting in this regard.

Theorem 18.60

Let Φ satisfy the RIP of order K where KM. Let ΦJ be any sub matrix of Φ with k columns where kK. Then

(ΦJHΦJI)x2δKx2xCk.

Proof. .

  1. From Theorem 18.53 we get

    ΦJHΦJI2δkδK.
  2. Thus since spectral norm is subordinate

    (ΦJHΦJI)x2ΦJHΦJI2x2δKx2xCk.

18.8.6. Approximate Orthogonality#

We are going to show that disjoint sets of columns from Φ span nearly orthogonal subspaces. This property is proved in [61].

Theorem 18.61

Let Φ satisfy the RIP of order K where KM. Let S and T denote index sets over the columns of Φ with |S|+|T|K and ST=. In other words, S and T are disjoint index sets. Let ΦS and ΦT denote corresponding sub-matrices consisting of columns indexed by S and T respectively. Then

ΦSHΦT2δK

where 2 denotes the 2-norm or spectral norm of a matrix.

Proof. Define R=ST.

  1. Consider the sub-matrix ΦR.

  2. Construct another matrix Ψ=ΦRHΦRI.

  3. The off-diagonal entries of Ψ are nothing but inner products of columns of ΦR.

  4. We note that every entry in the matrix ΦSHΦT is an entry in Ψ.

  5. Moreover, ΦSHΦT is a submatrix of Ψ.

  6. The spectral norm of a sub-matrix is never greater than the spectral norm of the matrix containing it.

  7. Thus

    ΦSHΦT2ΦRHΦRI2.
  8. From Theorem 18.53 the eigen values of ΦRHΦRI satisfy

    1δK1λ1+δK1.
  9. Thus the spectral norm of ΦRHΦRI which is its largest eigen value (see Theorem 4.145) satisfies

    ΦRHΦRI2δK.
  10. Plugging back we get

    ΦSHΦT2δK.

This result has a useful corollary. It establishes the approximate orthogonality between a set of columns in Φ and portion of a sparse vector not covered by those columns.

Corollary 18.5

Let Φ satisfy the RIP of order K where KM. Let T{1,,N} be an index set and let xCN be some vector. Let S=supp(x). Further let us assume that K|TS|. Define R=ST.

Then the following holds

ΦTHΦxR2δKxR2

where xR is obtained by keeping entries in x indexed by R while setting others to 0 (see Definition 18.11).

Proof. The set R denotes the indices at which x is nonzero but not yet covered in T. In a typical sparse recovery algorithm, one has discovered a candidate support T which may not include all of S.

  1. Since

    Φx=i=1Nϕixi

    and xR is zero on entries not indexed by R, hence

    ΦxR=ΦRxR

    where on the R.H.S. xRC|R| by dropping the 0 entries from it not indexed by R (see Definition 18.11).

  2. Thus we have

    ΦTHΦxR2=ΦTHΦRxR2.
  3. From Lemma 4.90 we know that any operator norm is subordinate.

  4. Thus

    ΦTHΦRxR2ΦTHΦR2xR2.
  5. Since K|TS| hence we have

    |R|=|ST|K.
  6. Further T and R are disjoint and |TR|K.

  7. Applying Theorem 18.61 we get

    ΦTHΦR2δK.
  8. Putting back, we get our desired result

    ΦTHΦxR2δKxR2.

18.8.7. Signal Proxy#

We can use the results so far to formalize the idea of signal proxy.

Theorem 18.62

Let x be a k-sparse signal Let Φ satisfy the RIP of order k+l or higher. Let p be defined as

p=(ΦHΦx)|l

i.e. p is obtained by keeping the l largest entries in b=ΦHΦx. Then the following holds

p2(1+δl+δk+l)x2.

Proof. Let A=supp(x) and B=supp(p).

  1. Then |A|k and |B|l.

  2. Clearly

    p=(ΦHΦx)|l=(ΦHΦx)B.
  3. From Theorem 18.20 we get

    p=ΦBHΦx.
  4. Let C=AB.

  5. Since x is supported on A only, hence we can write

    x=xB+xC.
  6. Thus from Corollary 18.1 we get (B and C are disjoint)

    Φx=ΦBxB+ΦCxC.
  7. Thus we have

    p=ΦBHΦBxB+ΦBHΦCxC.
  8. Using triangle inequality we can write

    p2ΦBHΦBxB2+ΦBHΦCxC2.
  9. Theorem 18.59 gives us

    ΦBHΦBxB2(1+δl)xB2.
  10. Since B and C are disjoint, hence Theorem 18.61 gives us

    ΦBHΦCxC2δk+lxC2.
  11. Since B and C are disjoint, hence

    xB2x2 and xC2x2.
  12. Finally

    p2(1+δl)xB2+δk+lxC2(1+δl+δk+l)x2.

18.8.8. RIP and Inner Product#

Let x and x be two different vectors in CN such that their support is disjoint. i.e. if

T=supp(x){1,,N}

and

T=supp(x){1,,N}

then TT=.

Clearly

x0=|T|

and

x0=|T|.

Since the support of x and x are disjoint hence it is straightforward that

x,x=0.

What can we say about the inner product of their corresponding embedded vectors Φx and Φx?

Following theorem provides an upper bound on the magnitude of the inner product when the signal vectors x,x belong to the Euclidean space RN. This result is adapted from [21].

Theorem 18.63

Assume that the sensing matrix ΦRM×N. For all x,xRN supported on disjoint subsets T,T{1,,N} with |T|<k and |T|<k, we have

|Φx,Φx|δk+kx2x2

where δk+k is the restricted isometry constant for the sparsity level k+k.

Proof. Let x^=xx2 and x^=xx2 be the corresponding unit norm vectors.

  1. Then

    Φx,Φx=Φx^,Φx^x2x2.
  2. Hence if we prove the bound for unit norm vectors, then it will be straightforward to prove the bound for arbitrary vectors.

  3. Let us assume without loss of generality that x,x are unit norm.

  4. We need to show that

    |Φx,Φx|δk+k.
  5. With the help of parallelogram identity (Theorem 4.93), we have

    Φx,Φx=14(Φx+Φx22ΦxΦx22).
  6. Thus

    |Φx,Φx|=14|Φx+Φx22ΦxΦx22|.
  7. Now

    x±x22=x22+x22±2x,x=x22+x22=2

    since x,x are orthogonal and unit norm.

  8. Using Theorem 18.48 we have

    (1δk+k)x±x22Φx±Φx22(1+δk+k)x±x222(1δk+k)Φx±Φx222(1+δk+k).
  9. Hence the maximum value of Φx±Φx22 can be 2(1+δk+k) while the minimum value of Φx±Φx22 can be 2(1δk+k).

  10. This gives us the upper bound

    |Φx,Φx|14(2(1+δk+k)2(1δk+k))=δk+k.
  11. Finally when x,x are not unit norm, the bound generalizes to

    |Φx,Φx|δk+kx2x2.

A variation of this result is presented below:

Theorem 18.64

Assume that the sensing matrix ΦRM×N. Let u,vRN be given and let

K=max(u+v0,uv0).

Let Φ satisfy RIP of order K with the constant δK. Then

|Φu,Φvu,v|δKu2v2.

This result is more general as it doesn’t require u,v to be supported on disjoint index sets. All it requires is them to be sufficiently sparse.

Proof. As, in the previous result, it is sufficient to prove it for the case where u2=v2=1. The simplified inequality becomes

|Φu,Φvu,v|δK.
  1. Clearly

    u±v22=u22+v22±2u,v=2±2u,v.
  2. Due to RIP, we have

    (1δK)(2±2u,v)Φ(u±v)22(1+δK)(2±2u,v).
  3. From the parallelogram identity, we have

    (18.49)#Φu,Φv=14(Φ(u+v)22Φ(uv)22).
  4. Taking the upper bound on Φ(u+v)22 and the lower bound on Φ(uv)22 in (18.49), we obtain

    Φu,Φv12((1+δK)(1+u,v)(1δK)(1u,v)).
  5. Simplifying, we get

    Φu,Φvu,v+δK.
  6. At the same time, taking the lower bound on Φ(u+v)22 and the upper bound on Φ(uv)22 in (18.49), we obtain

    Φu,Φv12((1δK)(1+u,v)(1+δK)(1u,v)).
  7. Simplifying, we get

    Φu,Φvu,vδK.
  8. Combining the two results, we obtain

    |Φu,Φvu,v|δK.

For the complex case, the result can be generalized if we choose a bilinear inner product rather than the usual sesquilinear inner product.

Theorem 18.65

Let u,vCN be given and let

K=max(u+v0,uv0).

Let the complex space CN be equipped with the bilinear inner product

u,vBRe(u,v)

i.e. the real part of the standard inner product.

Let Φ satisfy RIP of order K with the constant δK. Then

|Φu,ΦvBu,vB|δKu2v2.

Proof. Recall that the norm induced by the bilinear inner product u,vB is the usual 2 norm since

u,uB=Re(u,u)=Re(u22)=u22.
  1. Let us just work out the parallelogram identity for the complex case

    x±y22=x±y,x±yB=x,xB+y,yB±x,yB±y,xB=x,xB+y,yB±2x,yB

    due to the bilinearity of the real inner product.

  2. We can see that the rest of the proof is identical to the proof of Theorem 18.64.

18.8.9. RIP and Orthogonal Projection#

The first result in this section is presented for real matrices. The generalization for complex matrices will be done later.

  1. Let Λ{1,,N} be an index set.

  2. Let ΦRM×N satisfy RIP of order K with the restricted isometry constant δK.

  3. Assume that the columns of ΦΛ are linearly independent.

  4. We can define the pseudo inverse as

    ΦΛ=(ΦΛHΦΛ)1ΦΛH.
  5. The orthogonal projection operator to the column space for ΦΛ is given by

    PΛ=ΦΛΦΛ.
  6. The orthogonal projection operator onto the orthogonal complement of C(ΦΛ) (column space of ΦΛ) is given by

    PΛ=IPΛ.
  7. Both PΛ and PΛ satisfy the usual properties like P=PH and P2=P.

We further define

ΨΛ=PΛΦ.
  1. We are orthogonalizing the columns in Φ against C(ΦΛ).

  2. In other words, keeping the component of the column which is orthogonal to the column space of ΦΛ.

  3. Obviously the columns in ΨΛ corresponding to the index set Λ would be 0.

We now present a result which shows that the matrix ΨΛ satisfies a modified version of RIP [25].

Theorem 18.66

If Φ satisfies the RIP of order K with isometry constant δK, and Λ{1,,N} with |Λ|<K, then the matrix ΨΛ satisfies the modified version of RIP as

(18.50)#(1δK1δK)x22ΨΛx22(1+δK)x22

for all xRN such that x0K|Λ| and supp(x)Λ=.

In words, if Φ satisfies RIP of order K, then ΨΛ acts as an approximate isometry on every (K|Λ|)-sparse vector supported on Λc.

Proof. From the definition of ΨΛ, we have

ΨΛx=(IPΛ)Φx=ΦxPΛΦx.
  1. Alternatively

    Φx=ΨΛx+PΛΦx.
  2. Since PΛ is an orthogonal projection, hence the vectors PΛΦx and ΨΛx=PΛΦx are orthogonal.

  3. Thus, we can write

    (18.51)#Φx22=PΛΦx22+ΨΛx22.
  4. We need to show that Φx2ΨΛx2 or alternatively that PΛΦx2 is small under the conditions of the theorem.

  5. Since PΛΦx is orthogonal to ΨΛx, hence

    (18.52)#PΛΦx,Φx=PΛΦx,ΨΛx+PΛΦx=PΛΦx,PΛΦx+PΛΦx,ΨΛx=PΛΦx,PΛΦx=PΛΦx22.
  6. Since PΛ is a projection onto the C(ΦΛ) (column space of ΦΛ), there exists a vector zCN, such that PΛΦx=Φz and supp(z)Λ.

  7. Since supp(x)Λ=, hence x,z=0.

  8. We also note that x+z0=xz0K.

  9. Invoking Theorem 18.64, we have

    |Φz,Φx|δKz2x2.
  10. Alternatively

    |PΛΦx,Φx|δKz2x2.
  11. From RIP, we have

    1δKz2Φz2

    and

    1δKx2Φx2.
  12. Thus

    (1δK)z2x2Φz2Φx2.
  13. This gives us

    |PΛΦx,Φx|δK1δKPΛΦx2Φx2.
  14. Applying (18.52), we get

    PΛΦx22δK1δKPΛΦx2Φx2.
  15. Canceling the common term, we get

    PΛΦx2δK1δKΦx2.
  16. Trivially, we have PΛΦx20.

  17. Applying these bounds on (18.51), we obtain

    (1(δK1δK)2)Φx22ΨΛx22Φx22.
  18. Finally, using the RIP again with

    (1δK)x22Φx22(1+δK)x22

    we obtain

    (1(δK1δK)2)(1δK)x22ΨΛx22(1+δK)x22.
  19. Simplifying

    (1(δK1δK)2)(1δK)=1+δK22δKδK21δK=12δK1δK=1δK1δK.
  20. Thus, we get the intended result in (18.50).

18.8.10. RIP for Higher Orders#

If Φ satisfies RIP of order K, does it satisfy RIP of some other order K>K? There are some results available to answer this question.

Theorem 18.67

Let c and k be integers and let Φ satisfy RIP of order 2k. Φ satisfies RIP of order ck with a restricted isometry constant

δckcδ2k

if cδ2k<1.

Note that this is only a sufficient condition. Thus if cδ2k1 we are not claiming whether Φ satisfies RIP of order ck or not.

Proof. For c=1, δkδ2k. For c=2, δ2k2δ2k. These two cases are trivial. We now consider the case for c3.

  1. Let S be an arbitrary index set of size ck. Let

    Δ=ΦSHΦSI.
  2. From Theorem 18.54, a sufficient condition for Φ to satisfy RIP of order ck is that

    Δ2<1

    for all index sets S with |S|=ck.

  3. Thus if we can show that

    Δ2cδ2k

    we would have shown that Φ satisfies RIP of order ck.

  4. We note that ΦS is of size M×ck.

  5. Thus Δ is of size ck×ck.

  6. We partition Δ into a block matrix of size c×c

    Δ=[Δ11Δ12Δ1cΔ21Δ22Δ2cΔc1Δc2Δcc]

    where each entry Δij is a square matrix of size k×k.

  7. Each diagonal matrix Δii corresponds to some ΦTHΦTI where |T|=k.

  8. Thus we have (see Theorem 18.53)

    Δii2δk.
  9. The off-diagonal matrices Δij are

    Δij=ΦPHΦQ

    where P and Q are disjoint index sets with |P|=|Q|=k with |PQ|=2k.

  10. Thus from the approximate orthogonality condition (Theorem 18.61 ) we have

    Δij2δ2k.
  11. Finally we apply Gershgorin circle theorem for block matrices (Corollary 4.29).

  12. This gives us

    |Δ2Δii2|j,jiΔij for some i{1,2,,c}.
  13. Thus we have

    |Δ2δk|j,jiδ2k|Δ2δk|(c1)δ2kΔ2δk+(c1)δ2kΔ2δ2k+(c1)δ2kΔ2cδ2k.
  14. We have shown that Δ2cδ2k<1.

  15. Thus δckΔ2.

  16. Hence Φ indeed satisfies RIP of order ck.

This theorem helps us extend RIP from an order K to higher orders. If δ2k isn’t sufficiently small, the bound isn’t useful.

18.8.11. Embeddings of Arbitrary Signals#

So far we have considered only sparse signals while analyzing the embedding properties of a RIP satisfying matrix Φ. In this subsection we wish to explore bounds on the 2 norm of an arbitrary signal when embedded by Φ. This result is adapted from [61].

Theorem 18.68

Let Φ be an an M×N matrix satisfying

(18.53)#Φx21+δKx2xΣK.

Then for every signal xCN, the following holds:

Φx21+δK[x2+1Kx1].

We note that the theorem requires Φ to satisfy only the upper bound of RIP property (18.43). The proof is slightly involved.

Proof. We note that the bound is trivially true for x=0. Hence in the following we will consider only for x0.

  1. Consider an arbitrary index set Λ{1,2,,N} such that |Λ|K.

  2. Consider the unit ball in the Banach space 2(Λ) given by

    (18.54)#B2Λ={xCN|supp(x)=Λ and x21}

    i.e. the set of all signals whose support is Λ and whose 2 norm is less than or equal to 1.

  3. Now define a convex body

    (18.55)#S=conv{|Λ|KB2Λ}CN.
  4. We recall from Definition 9.9 that if x and y belong to S then their convex combination θx+(1θ)y with θ[0,1] must lie in S.

  5. Further it can be verified that S is a compact convex set with non-empty interior.

  6. Hence it is a convex body.

  7. Consider any xB2Λ1 and yB2Λ2.

  8. From (18.53) and (18.54) we have

    Φx21+δKx21+δK

    and

    Φy21+δKy21+δK.
  9. Now let

    z=θx+(1θ)y where θ[0,1].
  10. Then

    z2=θx+(1θ)y2θx2+(1θ)y2θ+(1θ)=1.
  11. Further

    Φz2=Φ(θx+(1θ)y)2Φθx2+Φ(1θ)y2=θΦx2+(1θ)Φy2θ1+δK+(1θ)1+δK1+δK.
  12. Similarly, it can be shown that for every vector xS we have x21 and Φx21+δK.

    1. Let xS.

    2. Then x=i=1rtixi such that xiB2Λi where |Λi|K, ti0 and ti=1.

    3. Hence xi21 for every i.

    4. Hence

      x2i=1rtixi2i=1rti=1.
    5. Similarly Φxi21+δK.

    6. Hence

      Φx2i=1rtiΦxi21+δK.
  13. We now define another convex body

    (18.56)#Γ={x|x2+1Kx11}CN.
  14. We quickly verify the convexity property.

    1. Let x,yΓ.

    2. Let

      z=θx+(1θ)y where θ[0,1].
    3. Then

      z+1Kz1=θx+(1θ)y2+1Kθx+(1θ)y1θx2+(1θ)y2+θKx1+(1θ)Ky1=θ[x2+1Kx1]+(1θ)[y2+1Ky1]θ+(1θ)=1.
    4. Thus zΓ.

    5. This analysis shows that all convex combinations of elements in Γ belong to Γ.

    6. Thus Γ is convex.

    7. Further it can be verified that Γ is a compact convex set with non-empty interior.

    8. Hence it is a convex body.

  15. For any xCN one can find a yΓ by simply applying an appropriate nonzero scale y=cx where the scale factor c depends on x.

  16. For a moment suppose that ΓS.

  17. Then if yΓ the following are true:

    y2+1Ky11

    and

    Φy21+δK.
  18. Now consider an arbitrary nonzero xCN.

  19. Let

    α=x2+1Kx1.
  20. Define

    y=1αx.
  21. Then

    y2+1Ky1=1α(x2+1Kx1)=1.
  22. Thus yΓ and

    Φy21+δKΦ1αx21+δKΦx21+δKαΦx21+δK(x2+1Kx1)xCN

    which is our intended result.

  23. Hence if we show that ΓS holds, we would have proven our theorem.

We will achieve this by showing that every vector xΓ can be shown to be a convex combination of vectors in S.

  1. We start with an arbitrary xΓ.

  2. Let I=supp(x).

  3. We partition I into disjoint sets of size K.

  4. Let there be J+1 such sets given by

    I=j=0JIj.
  5. Let I0 index the K largest entries in x (magnitude wise).

  6. Let I1 be next K largest entries and so on.

  7. Since |I| may not be a multiple of K, hence the last index set IJ may not have K indices.

  8. We define

    xIj(i)={x(i)if iIj;0otherwise.
  9. Thus we can write

    x=j=0JxIj.
  10. Now let

    θj=xIj2 and yj=1θjxIj.
  11. We can write

    x=j=0Jθjyj.
  12. In this construction of x we can see that 1θ0θ1θJ0.

  13. Also yjS since yj is a unit norm K sparse vector (18.55).

  14. We will show that jθj1 in a short while.

  15. This will imply that x is a convex combination of vectors from S.

  16. But since S is convex hence xS.

  17. This will imply that ΓS.

  18. The proof will be complete.

We now show that jθj1.

  1. Pick any j{1,,J}.

  2. Since xIj is K-sparse hence due to Theorem 18.17 we have

    θj=xIj2KxIj.
  3. It is easy to see that Ij1 identifies exactly K nonzero entries in x and each of nonzero entries in xIj1 is larger than the largest entry in xIj (magnitude wise).

  4. Thus we have

    xIj11=iIj1|xi|iIj1xIj=KxIj.
  5. Thus

    xIj1KxIj11.
  6. Combining the two inequalities we get

    θj1KxIj11.
  7. This lets us write

    j=1Jθjj=1J1KxIj111Kx1

    since

    x1=j=0JxIj1j=1JxIj11.
  8. Finally

    θ0=xI02x2.
  9. This gives us the inequality

    j=0Jθjx2+1Kx11

    since xΓ.

  10. Recalling our steps we can express x as

    x=θjyj

    where yjS and θj1 implies that xS since S is convex.

  11. Thus ΓS.

  12. This completes the proof.

18.8.12. A General Form of RIP#

A more general restricted isometry bound can be for an arbitrary matrix Φ can be as follows

αx22Φx22βx22

where 0<αβ<.

It is straightforward to scale Φ to match the bounds in (18.43).

Let δK=βαα+β. Then 1δK=2αα+β and 1+δK=2βα+β.

Putting in (18.43) we get

2αα+βx22Φx222βα+βx22αx22α+β2Φx22βx22.

Thus by multiplying Φ with 2/(α+β) we can transform the more general bound to the form of (18.43).

18.8.13. Finding out RIP Constants#

The optimal value of RIP constant of K-th order δK can be obtained by solving the following optimization problem.

Algorithm 18.2

minimize0<δ<1δsubject to (1δ)x22Φx22(1+δ)x22xΣK.

This problem isn’t easy to solve. In fact it has been shown in [4] that this problem is NP-hard.

18.8.14. RIP and Coherence#

Here we establish a relationship between the RIP constants and coherence of a dictionary.

Rather than a general matrix Φ, we restrict our attention to a dictionary DCN×D We assume that the dictionary is overcomplete (D>N) and full rank rank(D)=N. Dictionary is assumed to satisfy RIP of some order.

Theorem 18.69 (Coherence upper bound for RIP constant)

Let D satisfy RIP of order K. Then

δK(K1)μ(D).

Proof. We recall that δK is the smallest constant δ satisfying

(1δ)x22Dx22(1+δ)x22xΣK.
  1. Let Λ be any index set with |Λ|=K.

  2. Then

    Dx22=DΛxΛ22xCΛ.
  3. Since D satisfies RIP of order K, hence DΛ is a subdictionary (its columns are linearly independent).

  4. Recall from Theorem 18.30 that

    (1(K1)μ)v22DΛv22(1+(K1)μ)v22

    holds true for every vCK.

  5. Since δK is smallest possible constant, hence

    1+δK1+(K1)μδK(K1)μ(D).