Involve a design matrix \(X\) with two or more columns and a least-squares
solution \(\hat{\beta}\) of \(\mathbf{y}=X \beta .\) Consider the following
numbers.
(i) \(\|X \hat{\boldsymbol{\beta}}\|^{2}-\) the sum of the squares of the
"regression term." Denote this number by \(\mathrm{SS}(\mathrm{R})\).
(ii) \(\|\mathbf{y}-X \hat{\boldsymbol{\beta}}\|^{2}-\) the sum of the squares
for error term. Denote this number by \(\mathrm{SS}(\mathrm{E})\).
(iii) \(\|\mathbf{y}\|^{2}-\) the "total" sum of the squares of the \(y\) -values.
Denote this number by \(\mathrm{SS}(\mathrm{T}) .\)
Every statistics text that discusses regression and the linear model
\(\mathbf{y}=X \boldsymbol{\beta}+\boldsymbol{\epsilon}\) introduces these
numbers, though terminology and notation vary somewhat. To simplify matters,
assume that the mean of the \(y\) -values is zero. In this case, SS(T) is
proportional to what is called the variance of the set of \(y\) -values.
Justify the equation
\(\mathrm{SS}(\mathrm{T})=\mathrm{SS}(\mathrm{R})+\mathrm{SS}(\mathrm{E})\).
[Hint: Use a theorem, and explain why the hypotheses of the theorem are
satisfied.] This equation is extremely important in statistics, both in
regression theory and in the analysis of variance.