The exercise
[1, p. 34 exercise 2.4] asks to show that if

is a full rank

matrix with

>

,

is a

vector, and

is an

vector, that the effect of the linear transformation
is to project

onto the subspace spanned by the columns of

. Specifically, if

are the columns of

, the exercise
[1, p. 34 exercise 2.4] asks to show that
Furthermore it is asked why the transform

must be idempotent.
Solution: Let

be the linear transform to which the matrix

is associated to. We are asked to show that the linear transform is surjective (onto) to

.
This linear transform can be thought as the composite of the linear transforms
[2, p. 49, proposition 6.3]

,

,

to which the matrices

,

and

are associated respectively. We note that

is surjective as the rank of the matrix equals

:

. Thus the whole image

is spanned by this transform per definition. The transform

is bijective as an invertible matrix is associated with it. Therefore

is also surjective to

. From this we conclude that the standard basis

(

is the vector composed of zeros at all elements except the

that is equal to one)
is also included in the image of

.
Because any vector

can be represented as a weighted sum of the standard basis:

we obtain applying the linear transform

to this vector the following relation:
Now let

be the matrix of the transform

relative to the standard bases

with

and

,

.
If the elements of the matrix are

then the transform

can be rewritten as
[2, p. 47]:

. This is simply the vector composed of the elements of the

column of the matrix

. This vector will be denoted as

. Using matrix notation

can be written simply as

.
Thus

, which means that the image of

is composed of all linear combination of the columns of the corresponding matrix, this is per definition

and thus the composite of the transforms

is surjective (onto) to

.
Now let us proceed to show that
First we will compute the hermitian transpose of the column vector

. By using the associative property of matrix multiplication and the rule
[1, p. 23] 
we derive the following relations:
Now instead of multiplying

by one column

of the matrix
![\mathbf{F}=[\mathbf{f}_1 \cdots \mathbf{f}_i \cdots \mathbf{f}_n]](https://lysario.de/wp-content/cache/tex_6d210d4ed04bcea478b5e5437ecb3cdc.png)
, as in (
3),
we will proceed compute the multiplication with the whole matrix, effectiveley using all columns

. Furthermore applying also (
4) we obtain:
A matrix

is idempotent by definition
[1, p.21] when

. Multiplying

by itself and using again the associative property of matrix multiplication results in:
which proves that the matrix is idempotent. While this finding may be used to answer the question why the matrix has to be idempotent (it has to, because it is) we will use another approach. We note by
[1, p.30,equation 2.57] and the previous analysis that

is the least squares approximation in the space spanned by the columns of

to

. Applying the matrix

to

provides thus the least square approximation to

that is in the space spanned by the columns of

. But as

resides already within the space spanned by the columns of the matrix

, the least square approximation that also resides in the same space has to be the vector

itself. But this simply means that consecutive applications of the same transform
provide the same result as the single application of the transform

. Thus

is idempotent. QED.
[1] Steven M. Kay: “Modern Spectral Estimation – Theory and Applications”, Prentice Hall, ISBN: 0-13-598582-X. [2] Lawrence J. Corwin and Robert H. Szczarba: “Calculus in Vector Spaces”, Marcel Dekker, Inc, 2nd edition, ISBN: 0824792793.
Leave a reply