Singular Values
Contents
4.10. Singular Values#
In previous section we saw diagonalization of square matrices
which resulted in an eigen value
decomposition of the matrix.
This matrix factorization is very useful yet it is not applicable
in all situations.
In particular, the eigen value decomposition is useless if the square matrix
is not diagonalizable or if the matrix is not square at all.
Moreover, the decomposition
is particularly useful only for real symmetric or Hermitian matrices
where the diagonalizing
matrix is an
Fortunately there happens to be another decomposition which applies to all matrices and
it involves just
4.10.1. Singular value#
Definition 4.131 (Singular value)
A non-negative real number
and
hold.
The vectors
We first present the basic result of singular value decomposition. We will not prove this result completely although we will present proofs of some aspects.
Theorem 4.138 (Existence of singular values)
For every
and
such that
where
The non-negative real numbers
The sequence of real numbers
It is certainly possible that some of the singular values are 0 themselves.
Remark 4.22
Since
Definition 4.132 (Singular value decomposition)
The decomposition of a matrix
is known as its singular value decomposition.
Remark 4.23
When
and
Remark 4.24
There can be at most
Remark 4.25
We can also write
Remark 4.26 (SVD as a sum of rank 1 matrices)
Let us expand
Remark 4.27
Alternatively, let us expand
This gives us
Following lemma verifies that
Lemma 4.70
Let
Proof. We have
Let us expand R.H.S.
where
columns in the end appear times.Expanding the L.H.S. we get
Thus by comparing both sides we get
and
Now let us start with
Let us expand R.H.S.
where
columns appear times.Expanding the L.H.S. we get
Thus by comparing both sides we get
and
We now consider the three cases.
For
, we have . And we getThus
is a singular value of and is a left singular vector while is a right singular vector.For
, we have . We get for first vectors inFinally for remaining
vectors in , we can writeThey belong to the null space of
.For
, we have . We get for first vectors inFinally for remaining
vectors in , we can write
Lemma 4.71
where the number of
Lemma 4.72
where the number of
Lemma 4.73 (Rank from singular value decomposition)
Let
Then
In other words, the rank of
Proof. This is a straight forward application of
Lemma 4.4
and Lemma 4.5.
Further since only nonzero values in
Corollary 4.24
Let
where
Lemma 4.74 (Eigen values of the Gram matrix)
The eigen values of Hermitian matrix
Proof. Expanding the Gram matrix:
We note that
is Hermitian.Hence
is diagonalized by .The diagonalization of
is .Thus the eigen values of
are with ’s after .Clearly
Thus columns of
are the eigen vectors of .
Lemma 4.75 (Eigen values of the frame operator)
The eigen values of Hermitian matrix
Proof. Expanding the frame operator:
We note that
is Hermitian.Hence
is diagonalized by .The diagonalization of
is .Thus the eigen values of
are with ’s after .Clearly
Thus columns of
are the eigen vectors of .
Observation 4.8
The Gram matrices
4.10.2. The Largest Singular Value#
Lemma 4.76
For every
Also for every
Proof. Let us expand the term
Now since
Thus
or
The result follows.
A simpler representation of
Let
We split entries in
Then
Thus
2nd result can also be proven similarly.
Lemma 4.77 (Upper bounds on norms of matrix vector product)
Let
Moreover
Proof. We have
since
since
Similarly
since
since
There is a direct connection between the largest singular value and
Corollary 4.25
The largest singular value of
4.10.3. SVD and Pseudo Inverse#
Lemma 4.78
Let
where
Proof. Straight forward application of Lemma 4.24.
Corollary 4.26
The rank of
Proof. The number of nonzero diagonal entries in
Lemma 4.79
Let
Proof. As usual we verify the requirements for a Moore-Penrose pseudo-inverse
as per Definition 4.42. We note that
since
First requirement:
Second requirement:
We now consider
Thus
since
Thus
since
Finally we can connect the singular values of
Corollary 4.27 (Rank of pseudoinverse)
The rank of any
Proof. We have
Lemma 4.80 (Singular values of the pseudoinverse)
Let
while all other
Proof.
Since rank of
and are same, hence the number of nonzero singular values is same.Now look at
where
Clearly
.Thus expanding the R.H.S. we can get
where
and are first columns of and respectively.If we reverse the order of first
columns of and and reverse the first diagonal entries of , the R.H.S. remains the same while we are able to express in the standard singular value decomposition form.Thus
are indeed the nonzero singular values of .
4.10.4. Full Column Rank Matrices#
In this subsection we consider some specific results related to the singular value decomposition of a full column rank matrix.
We will consider
to be an matrix in with and .Let
be its singular value decomposition.From Lemma 4.73 we observe that there are
nonzero singular values of .We will call these singular values as
.We will define
Clearly
is an block matrix given bywhere the lower
is an zero matrix.From here we obtain that
is an matrix given bywhere
Lemma 4.81
Let
Proof. Since all singular values are nonzero hence
Lemma 4.82
Let
Proof. Let
We have
Now since
hence
Thus
Applying square roots, we get
We recall from Corollary 4.7
that the Gram matrix of its column vectors
Lemma 4.83 (Norm bounds for Gram matrix vector product)
Let
Proof. Expanding the Gram matrix
Let
Let
Then from previous lemma we have
Finally
Thus
Substituting we get
There are bounds for the inverse of Gram matrix also. First let us establish the inverse of Gram matrix.
Lemma 4.84 (SVD of inverse Gram matrix)
Let
where
We can now state the bounds:
Lemma 4.85 (Norm bounds for inverse Gram matrix vector product)
Let
Proof. From Lemma 4.84 we have
where
Let
Let
Then
Thus
Finally
Thus
Substituting we get the result.
4.10.5. Low Rank Approximation of a Matrix#
Definition 4.133 (Low rank matrix)
An
A matrix is low rank if the number of nonzero singular values for the matrix is much smaller than its dimensions.
Following is a simple procedure for making a low rank approximation
of a given matrix
Perform the singular value decomposition of
given by .Identify the singular values of
in .Keep the first
singular values (where is the rank of the approximation).Set all other singular values to 0 to obtain
.Compute
.