The exercise
[1, p. 34 exercise 2.4] asks to show that if
is a full rank
matrix with
>
,
is a
vector, and
is an
vector, that the effect of the linear transformation
is to project
onto the subspace spanned by the columns of
. Specifically, if
are the columns of
, the exercise
[1, p. 34 exercise 2.4] asks to show that
Furthermore it is asked why the transform
must be idempotent.
Solution: Let
be the linear transform to which the matrix
is associated to. We are asked to show that the linear transform is surjective (onto) to
.
This linear transform can be thought as the composite of the linear transforms
[2, p. 49, proposition 6.3]
,
,
to which the matrices
,
and
are associated respectively. We note that
is surjective as the rank of the matrix equals
:
. Thus the whole image
is spanned by this transform per definition. The transform
is bijective as an invertible matrix is associated with it. Therefore
is also surjective to
. From this we conclude that the standard basis
(
is the vector composed of zeros at all elements except the
that is equal to one)
is also included in the image of
.
Because any vector
can be represented as a weighted sum of the standard basis:
we obtain applying the linear transform
to this vector the following relation:
Now let
be the matrix of the transform
relative to the standard bases
with
and
,
.
If the elements of the matrix are
then the transform
can be rewritten as
[2, p. 47]:
. This is simply the vector composed of the elements of the
column of the matrix
. This vector will be denoted as
. Using matrix notation
can be written simply as
.
Thus
, which means that the image of
is composed of all linear combination of the columns of the corresponding matrix, this is per definition
and thus the composite of the transforms
is surjective (onto) to
.
Now let us proceed to show that
First we will compute the hermitian transpose of the column vector
. By using the associative property of matrix multiplication and the rule
[1, p. 23] we derive the following relations:
Now instead of multiplying
by one column
of the matrix
, as in (
3),
we will proceed compute the multiplication with the whole matrix, effectiveley using all columns
. Furthermore applying also (
4) we obtain:
A matrix
is idempotent by definition
[1, p.21] when
. Multiplying
by itself and using again the associative property of matrix multiplication results in:
which proves that the matrix is idempotent. While this finding may be used to answer the question why the matrix has to be idempotent (it has to, because it is) we will use another approach. We note by
[1, p.30,equation 2.57] and the previous analysis that
is the least squares approximation in the space spanned by the columns of
to
. Applying the matrix
to
provides thus the least square approximation to
that is in the space spanned by the columns of
. But as
resides already within the space spanned by the columns of the matrix
, the least square approximation that also resides in the same space has to be the vector
itself. But this simply means that consecutive applications of the same transform
provide the same result as the single application of the transform
. Thus
is idempotent. QED.
[1] Steven M. Kay: “Modern Spectral Estimation – Theory and Applications”, Prentice Hall, ISBN: 0-13-598582-X. [2] Lawrence J. Corwin and Robert H. Szczarba: “Calculus in Vector Spaces”, Marcel Dekker, Inc, 2nd edition, ISBN: 0824792793.
Leave a reply