Let \mathrm{{\boldsymbol X}}_1, \ldots , \mathrm{{\boldsymbol X}}_ n \in \mathbb {R}^ d denote a data set and let
\mathbb {X} = \begin{pmatrix} \longleftarrow & \mathbf{X}_1^ T & \longrightarrow \\ \longleftarrow & \mathbf{X}_2^ T & \longrightarrow \\ \vdots & \vdots & \vdots \\ \longleftarrow & \mathbf{X}_ n^ T & \longrightarrow \\ \end{pmatrix}.
Recall that the empirical covariance matrix S of this data set can be expressed as
S = \frac{1}{n} \mathbb {X}^ T H \mathbb {X}
where
H = I_ n - \frac{1}{n} \mathbf{1} \mathbf{1}^ T.
The matrix H \in \mathbb {R}^{n \times n} is an orthogonal projection .
In general, we say that a matrix M is an orthogonal projection onto a subspace S if
M is symmetric,
M^2 = M, and
S = \{ \mathrm{{\boldsymbol y}} : \, M \mathbf{x}= y \, \, \text {for some} \, \, \mathbf{x}\in \mathbb {R}^ n \}
Which of the following are true about the matrix H? (Choose all that apply.)
For any positive integer k and any vector \mathbf{x}\in \mathbb {R}^ n, we have H^ k \mathbf{x}= H \mathbf{x}.
For any positive integer k and any vector \mathbf{x}\in \mathbb {R}^ n, we have H^ k \mathbf{x}= \mathbf{x}.
The matrix H is a projection onto the subspace of vectors perpendicular to the vector \mathbf{1} \in \mathbb {R}^ n, which has all of its entries equal to 1.
The matrix H is a projections onto the subspace \{ \mathbf{x}: \frac{1}{n} \sum _{i = 1}^ n \mathbf{x}^ i = 0\} \subset \mathbb {R}^ n. (In other words, this is the set of vectors having coordinate-wise average equal to 0.)
1 answer
1. For any positive integer k and any vector 𝐱∈ℝ𝑛 , we have 𝐻^𝑘𝐱=𝐻𝐱.
This is true because 𝐻 is an orthogonal projection, so raising it to any power does not change its action on vectors.
3. The matrix 𝐻 is a projection onto the subspace of vectors perpendicular to the vector 𝟏∈ℝ𝑛 , which has all of its entries equal to 1.
This is true because 𝐻 subtracts the average of the vector from each component, effectively making the vector perpendicular to the constant vector 𝟏 .
4. The matrix 𝐻 is a projection onto the subspace {𝐱:1𝑛∑𝑖=1𝑛𝐱𝑖=0}⊂ℝ𝑛 .
(In other words, this is the set of vectors having coordinate-wise average equal to 0.)
This is true because the matrix 𝐻 subtracts the average of the vector from each component, effectively ensuring that the coordinate-wise average of the resulting vector is zero.