Restricted Isometry Property
Contents
18.8. Restricted Isometry Property#
This section dwells deep into the implications of restricted isometry property.
Definition 18.31 (Restricted isometry property)
A matrix
for every
is the set of all
Definition 18.32 (Restricted isometry constant)
If a matrix
is known as the
Some remarks are in order.
maps a vector into as a vector (usually ).We will call
as an embedding of into .RIP quantifies the idea as to how much the squared length of a sparse signal changes during this embedding process.
We can compare matrices satisfying RIP with orthonormal bases.
An orthonormal basis or the corresponding unitary matrix preserves the length of a vector exactly.
A matrix
satisfying RIP of order is able to preserve the length of sparse signals approximately (the approximation range given by ).In this sense we can say that
implements a restricted almost orthonormal system [19].By restricted we mean that orthonormality is limited to
-sparse signals.By almost we mean that the squared length is not preserved exactly. Rather it is preserved approximately.
An arbitrary matrix
need not satisfy RIP of any order at all.If
satisfies RIP of order then it is easy to see that satisfies RIP of any order (since whenever ).If
satisfies RIP of order , then it may or many not satisfy RIP of order .Restricted isometry constant is a function of sparsity level
of the signal .
Example 18.37 (Restricted isometry constant)
As a running example in this section we will use following matrix
Consider
which is a
Now
and
We note that
With this much information, all we can say that
Still what is comforting to note is that for this particular example, the distance hasn’t increased by a large factor.
For a given
In the running example
We define
Let
It is easy to see that
There are
For each such submatrix
for every
Theorem 18.44
An
Proof. This comes from the fact that for a wide matrix
Since every
hence any set of columns in is linearly dependent.Thus there exists a nonzero
sparse signal such that (it belongs to the null space of the chosen columns).RIP (18.43) requires that a nonzero vector be embedded as a nonzero vector.
Thus
cannot satisfy RIP of order .The argument can be easily extended for any
.
Theorem 18.45
If
Proof. Every
Theorem 18.46
Let
Proof. Since every
and
Since
18.8.1. The First Restricted Isometry Constant#
We consider the simplest case where
Now a
-sparse vector consists of only one nonzero entry.Say that
is nonzero at index .Then
is nothing but .With this the restricted isometry inequality can be written as
Dividing by
we get
Let us formalize this in the following theorem.
Theorem 18.47 (Restricted isometry constants of order 1)
If a matrix
When
Remark 18.8
A dictionary (Definition 18.4) satisfies RIP of order 1 with
18.8.2. Sums and Differences of Sparse Vectors#
Theorem 18.48
Let
as long as
Proof. We know that
Thus
18.8.3. Distance Between Sparse Vectors#
Let
The
Now if
Theorem 18.49 (Approximation preservation of distances)
Let
Proof. Since
But then
and
Thus we have the result.
18.8.4. RIP with Unit Length Sparse Vectors#
Sometimes it is convenient to state RIP in terms of unit length sparse vectors.
Theorem 18.50 (RIP for unit length sparse vectors)
Let
for every
18.8.5. Singular and Eigen Values of -Submatrices#
Consider any index set
as the Gram matrix for columns of
We consider the eigen values of
for some
In the running example
Eigen values of G are
Theorem 18.51
Let
Proof. We note that
Let
be some eigen value of .Let
be a corresponding (nonzero) eigenvector.Then
From (18.46) we recall that
RIP bounds apply for each vector in for a -column submatrix given byThus
since
.
Corollary 18.3
Let
Proof. From Theorem 18.51,
the eigen values are in range
Since
satisfies RIP of order , hence .Hence all eigenvalues of
are positive.Hence
is positive definite.Hence their product is positive.
Thus
is nonzero.Hence
is invertible.
Theorem 18.52
Let
where
Proof. This is straight forward application of
Lemma 4.74 and Theorem 18.51.
Eigen values of
Corollary 18.4
Let
where
Proof. Let
Since
satisfies RIP of order , it also satisfies RIP of order .From Theorem 18.52 we have
From Theorem 18.46 we have
.Thus
Thus
Theorem 18.53
Let
Moreover consider
Proof. .
From Theorem 18.51 eigen values of
lie in the range .From Lemma 4.69
is an eigen value of if and only if is an eigen value of .Hence the result.
Now for
The eigen values lie in the range
.Thus for every eigen value of
we have .Since
is Hermitian, its spectral norm is nothing but its largest eigen value.Hence
From previous few results we see that bound over eigen values of
Theorem 18.54
Let
for every eigen value of
Alternatively, let
for every
Alternatively, if singular values of
for every
Proof. Equivalence of sufficient conditions
We note that eigen values of
are related to eigen values of by the relation (see Lemma 4.69)Hence
Thus the first two sufficient conditions are equivalent.
Lastly the eigen values of
are squares of singular values of .Thus all sufficient conditions are equivalent.
Proof of sufficient condition
Now let
be an arbitrary vector.Let
.Clearly
. If then augment by adding some indices arbitrarily till we get .Clearly
is an arbitrary vector in and .Now let
be the largest and be the smallest eigen value of . is Hermitian and all its eigen values are positive, hence it is positive definite.From Lemma 4.68 we get
Applying the limits on the eigen values and using
, we getSince this holds for every index set
with hence an equivalent statement isThus
indeed satisfies RIP of order with some not larger than .
Theorem 18.55
Let
where
Proof. Construction of pseudoinverse of a matrix through its singular value decomposition is discussed in Lemma 4.79.
Lemma 4.80 shows that if
is a nonzero singular value of then is a nonzero singular value of .From Corollary 18.4 we have that if
is a singular value of then,Inverting the terms in the inequalities we get our result.
Theorem 18.56
Eigen values of
where
In other words, singular values of
Proof. Obvious.
In the running example, the bounds tell us that
Certainly we have to consider
all possible
This result doesn’t provide us any upper
bound on
Theorem 18.57
Let
Moreover
Proof. We note that
Let
be the largest singular value of .Then by Lemma 4.77 we have
and
From Theorem 18.52 and Corollary 18.4 we get
This completes the proof.
First inequality is a restatement of restricted isometry property
in (18.46).
Second inequality is interesting.
In compressive sensing terms,
Theorem 18.58
Let
Proof. We note that
Let
be the largest singular value of .Then by Lemma 4.77 we have
From Theorem 18.55 we see that singular values of
satisfy the inequalitiesThus
Plugging it in we get
In the previous theorem we saw that back-projection using
Theorem 18.59
Let
Moreover
Proof. We note that
We recall that all singular values of
are positive and are bounded by (Corollary 18.4):where
are the singular values of (in descending order).We note that
is an matrix which is invertible (Corollary 18.3).From Lemma 4.83 we get
Applying the bounds on
we get the resultFrom Lemma 4.85 we have the bounds for
given byApplying the bounds on
we get the result
In the sequel we will discuss that
Theorem 18.60
Let
Proof. .
From Theorem 18.53 we get
Thus since spectral norm is subordinate
18.8.6. Approximate Orthogonality#
We are going to show that disjoint sets of columns from
Theorem 18.61
Let
where
Proof. Define
Consider the sub-matrix
.Construct another matrix
.The off-diagonal entries of
are nothing but inner products of columns of .We note that every entry in the matrix
is an entry in .Moreover,
is a submatrix of .The spectral norm of a sub-matrix is never greater than the spectral norm of the matrix containing it.
Thus
From Theorem 18.53 the eigen values of
satisfyThus the spectral norm of
which is its largest eigen value (see Theorem 4.145) satisfiesPlugging back we get
This result has a useful corollary. It establishes the approximate
orthogonality between a set of columns in
Corollary 18.5
Let
Then the following holds
where
Proof. The set
Since
and
is zero on entries not indexed by , hencewhere on the R.H.S.
by dropping the 0 entries from it not indexed by (see Definition 18.11).Thus we have
From Lemma 4.90 we know that any operator norm is subordinate.
Thus
Since
hence we haveFurther
and are disjoint and .Applying Theorem 18.61 we get
Putting back, we get our desired result
18.8.7. Signal Proxy#
We can use the results so far to formalize the idea of signal proxy.
Theorem 18.62
Let
i.e.
Proof. Let
Then
and .Clearly
From Theorem 18.20 we get
Let
.Since
is supported on only, hence we can writeThus from Corollary 18.1 we get (
and are disjoint)Thus we have
Using triangle inequality we can write
Theorem 18.59 gives us
Since
and are disjoint, hence Theorem 18.61 gives usSince
and are disjoint, henceFinally
18.8.8. RIP and Inner Product#
Let
and
then
Clearly
and
Since the support of
What can we say about the inner product of
their corresponding embedded vectors
Following theorem provides an upper bound on the magnitude
of the inner product when the signal vectors
Theorem 18.63
Assume that the sensing matrix
where
Proof. Let
Then
Hence if we prove the bound for unit norm vectors, then it will be straightforward to prove the bound for arbitrary vectors.
Let us assume without loss of generality that
are unit norm.We need to show that
With the help of parallelogram identity (Theorem 4.93), we have
Thus
Now
since
are orthogonal and unit norm.Using Theorem 18.48 we have
Hence the maximum value of
can be while the minimum value of can be .This gives us the upper bound
Finally when
are not unit norm, the bound generalizes to
A variation of this result is presented below:
Theorem 18.64
Assume that the sensing matrix
Let
This result is more general as it doesn’t require
Proof. As, in the previous result, it is sufficient to prove it
for the case where
Clearly
Due to RIP, we have
From the parallelogram identity, we have
(18.49)#Taking the upper bound on
and the lower bound on in (18.49), we obtainSimplifying, we get
At the same time, taking the lower bound on
and the upper bound on in (18.49), we obtainSimplifying, we get
Combining the two results, we obtain
For the complex case, the result can be generalized if we choose a bilinear inner product rather than the usual sesquilinear inner product.
Theorem 18.65
Let
Let the complex space
i.e. the real part of the standard inner product.
Let
Proof. Recall that the norm induced by the bilinear inner product
Let us just work out the parallelogram identity for the complex case
due to the bilinearity of the real inner product.
We can see that the rest of the proof is identical to the proof of Theorem 18.64.
18.8.9. RIP and Orthogonal Projection#
The first result in this section is presented for real matrices. The generalization for complex matrices will be done later.
Let
be an index set.Let
satisfy RIP of order with the restricted isometry constant .Assume that the columns of
are linearly independent.We can define the pseudo inverse as
The orthogonal projection operator to the column space for
is given byThe orthogonal projection operator onto the orthogonal complement of
(column space of ) is given byBoth
and satisfy the usual properties like and .
We further define
We are orthogonalizing the columns in
against .In other words, keeping the component of the column which is orthogonal to the column space of
.Obviously the columns in
corresponding to the index set would be .
We now present a result which shows that
the matrix
Theorem 18.66
If
for all
In words, if
Proof. From the definition of
Alternatively
Since
is an orthogonal projection, hence the vectors and are orthogonal.Thus, we can write
(18.51)#We need to show that
or alternatively that is small under the conditions of the theorem.Since
is orthogonal to , hence(18.52)#Since
is a projection onto the (column space of ), there exists a vector , such that and .Since
, hence .We also note that
.Invoking Theorem 18.64, we have
Alternatively
From RIP, we have
and
Thus
This gives us
Applying (18.52), we get
Canceling the common term, we get
Trivially, we have
.Applying these bounds on (18.51), we obtain
Finally, using the RIP again with
we obtain
Simplifying
Thus, we get the intended result in (18.50).
18.8.10. RIP for Higher Orders#
If
Theorem 18.67
Let
if
Note that this is only a sufficient condition.
Thus if
Proof. For
Let
be an arbitrary index set of size . LetFrom Theorem 18.54, a sufficient condition for
to satisfy RIP of order is thatfor all index sets
with .Thus if we can show that
we would have shown that
satisfies RIP of order .We note that
is of size .Thus
is of size .We partition
into a block matrix of sizewhere each entry
is a square matrix of size .Each diagonal matrix
corresponds to some where .Thus we have (see Theorem 18.53)
The off-diagonal matrices
arewhere
and are disjoint index sets with with .Thus from the approximate orthogonality condition (Theorem 18.61 ) we have
Finally we apply Gershgorin circle theorem for block matrices (Corollary 4.29).
This gives us
Thus we have
We have shown that
.Thus
.Hence
indeed satisfies RIP of order .
This theorem helps us extend RIP from an order
18.8.11. Embeddings of Arbitrary Signals#
So far we have considered only sparse signals while analyzing the embedding
properties of a RIP satisfying matrix
Theorem 18.68
Let
Then for every signal
We note that the theorem requires
Proof. We note that the bound is trivially true for
Consider an arbitrary index set
such that .Consider the unit ball in the Banach space
given by(18.54)#i.e. the set of all signals whose support is
and whose norm is less than or equal to 1.Now define a convex body
(18.55)#We recall from Definition 9.9 that if
and belong to then their convex combination with must lie in .Further it can be verified that
is a compact convex set with non-empty interior.Hence it is a convex body.
Consider any
and .From (18.53) and (18.54) we have
and
Now let
Then
Further
Similarly, it can be shown that for every vector
we have and .Let
.Then
such that where , and .Hence
for every .Hence
Similarly
.Hence
We now define another convex body
(18.56)#We quickly verify the convexity property.
Let
.Let
Then
Thus
.This analysis shows that all convex combinations of elements in
belong to .Thus
is convex.Further it can be verified that
is a compact convex set with non-empty interior.Hence it is a convex body.
For any
one can find a by simply applying an appropriate nonzero scale where the scale factor depends on .For a moment suppose that
.Then if
the following are true:and
Now consider an arbitrary nonzero
.Let
Define
Then
Thus
andwhich is our intended result.
Hence if we show that
holds, we would have proven our theorem.
We will achieve this by showing that every vector
We start with an arbitrary
.Let
.We partition
into disjoint sets of size .Let there be
such sets given byLet
index the largest entries in (magnitude wise).Let
be next largest entries and so on.Since
may not be a multiple of , hence the last index set may not have indices.We define
Thus we can write
Now let
We can write
In this construction of
we can see that .Also
since is a unit norm sparse vector (18.55).We will show that
in a short while.This will imply that
is a convex combination of vectors from .But since
is convex hence .This will imply that
.The proof will be complete.
We now show that
Pick any
.Since
is -sparse hence due to Theorem 18.17 we haveIt is easy to see that
identifies exactly nonzero entries in and each of nonzero entries in is larger than the largest entry in (magnitude wise).Thus we have
Thus
Combining the two inequalities we get
This lets us write
since
Finally
This gives us the inequality
since
.Recalling our steps we can express
aswhere
and implies that since is convex.Thus
.This completes the proof.
18.8.12. A General Form of RIP#
A more general restricted isometry bound can be for an arbitrary matrix
where
It is straightforward to scale
Let
Putting in (18.43) we get
Thus by multiplying
18.8.13. Finding out RIP Constants#
The optimal value of RIP constant of
Algorithm 18.2
This problem isn’t easy to solve. In fact it has been shown in [4] that this problem is NP-hard.
18.8.14. RIP and Coherence#
Here we establish a relationship between the RIP constants and coherence of a dictionary.
Rather than a general matrix
Theorem 18.69 (Coherence upper bound for RIP constant)
Let
Proof. We recall that
Let
be any index set with .Then
Since
satisfies RIP of order , hence is a subdictionary (its columns are linearly independent).Recall from Theorem 18.30 that
holds true for every
.Since
is smallest possible constant, hence