@@ -64,177 +64,32 @@ H_1: &\; \mu \neq 0 \; .
6464$$
6565
6666
67- ** Proof:** Denote sample sizes as
67+ ** Proof:** Applying [ Cochran's theorem for two-analysis of variance ] ( /P/anova2-cochran ) , we find that the following squared sums
6868
69- $$ \label{eq:samp-size }
69+ $$ \label{eq:anova2-ss-dist }
7070\begin{split}
71- n_{ij} &- \text{number of samples in category} \; (i,j) \\
72- n_{i \bullet} &= \sum_{j=1}^{b} n_{ij} \\
73- n_{\bullet j} &= \sum_{i=1}^{a} n_{ij} \\
74- n &= \sum_{i=1}^{a} \sum_{j=1}^{b} n_{ij}
71+ \frac{\mathrm{SS}_M}{\sigma^2} &= \frac{1}{\sigma^2} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (\bar{y}_{\bullet \bullet \bullet} - \mu)^2 = \frac{1}{\sigma^2} n (\bar{y}_{\bullet \bullet \bullet} - \mu)^2 \\
72+ \frac{\mathrm{SS}_\mathrm{res}}{\sigma^2} &= \frac{1}{\sigma^2} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (y_{ijk} - \bar{y}_{i j \bullet})^2 = \frac{1}{\sigma^2} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (y_{ijk} - \bar{y}_{i j \bullet})^2
7573\end{split}
7674$$
7775
78- and denote sample means as
76+ are [ independent ] ( /D/ind ) and [ chi-squared distributed ] ( /D/chi2 ) :
7977
80- $$ \label{eq:mean-samp }
78+ $$ \label{eq:anova2-cochran-s1 }
8179\begin{split}
82- \bar{y}_{\bullet \bullet \bullet} &= \frac{1}{n} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} y_{ijk} \\
83- \bar{y}_{i \bullet \bullet} &= \frac{1}{n_{i \bullet}} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} y_{ijk} \\
84- \bar{y}_{\bullet j \bullet} &= \frac{1}{n_{\bullet j}} \sum_{i=1}^{a} \sum_{k=1}^{n_{ij}} y_{ijk} \\
85- \bar{y}_{i j \bullet} &= \frac{1}{n_{ij}} \sum_{k=1}^{n_{ij}} y_{ijk} \; .
80+ \frac{\mathrm{SS}_M}{\sigma^2} &\sim \chi^2(1) \\
81+ \frac{\mathrm{SS}_\mathrm{res}}{\sigma^2} &\sim \chi^2(n-ab) \; .
8682\end{split}
8783$$
8884
89- Assume that $\mu$ zero, according to $H_0$ given by \eqref{eq: anova2-h0 }. Under this null hypothesis, we have:
90-
91- $$ \label{eq:yijk-h0}
92- y_{ijk} \sim \mathcal{N}(\alpha_i + \beta_j + \gamma_{ij}, \sigma^2) \quad \text{for all} \quad i, j, k \; .
93- $$
94-
95- Thus, the [ random variable] ( /D/rvar ) $U_ {ijk} = (y_ {ijk} - \alpha_i - \beta_j - \gamma_ {ij})/\sigma$ [ follows a standard normal distribution] ( /P/norm-snorm )
96-
97- $$ \label{eq:Uijk-h0}
98- U_{ijk} = \frac{y_{ijk} - \alpha_i - \beta_j - \gamma_{ij}}{\sigma} \sim \mathcal{N}(0, 1) \; .
99- $$
100-
101- Now consider the following sum
102-
103- $$ \label{eq:sum-Uijk-s1}
104- \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} U_{ijk}^2 = \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} \left( \frac{y_{ijk} - \alpha_i - \beta_j - \gamma_{ij}}{\sigma} \right)^2 \\
105- $$
106-
107- which can be rewritten as follows:
108-
109- $$ \label{eq:sum-Uijk-s2}
110- \begin{split}
111- \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} U_{ijk}^2 = \frac{1}{\sigma^2} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} & \left[ (y_{ijk} - \alpha_i - \beta_j - \gamma_{ij}) - \right. \\
112- &\left. [\bar{y}_{\bullet \bullet \bullet} + (\bar{y}_{i \bullet \bullet} - \bar{y}_{\bullet \bullet \bullet}) + (\bar{y}_{\bullet j \bullet} - \bar{y}_{\bullet \bullet \bullet}) + (\bar{y}_{i j \bullet} - \bar{y}_{i \bullet \bullet} - \bar{y}_{\bullet j \bullet} + \bar{y}_{\bullet \bullet \bullet})] \right. + \\
113- &\left. [\bar{y}_{\bullet \bullet \bullet} + (\bar{y}_{i \bullet \bullet} - \bar{y}_{\bullet \bullet \bullet}) + (\bar{y}_{\bullet j \bullet} - \bar{y}_{\bullet \bullet \bullet}) + (\bar{y}_{i j \bullet} - \bar{y}_{i \bullet \bullet} - \bar{y}_{\bullet j \bullet} + \bar{y}_{\bullet \bullet \bullet})] \right]^2 \\
114- = \frac{1}{\sigma^2}\sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} & \left[ (y_{ijk} - [\bar{y}_{\bullet \bullet \bullet} + (\bar{y}_{i \bullet \bullet} - \bar{y}_{\bullet \bullet \bullet}) + (\bar{y}_{\bullet j \bullet} - \bar{y}_{\bullet \bullet \bullet}) + (\bar{y}_{i j \bullet} - \bar{y}_{i \bullet \bullet} - \bar{y}_{\bullet j \bullet} + \bar{y}_{\bullet \bullet \bullet})]) + \right. \\
115- &\left. (\bar{y}_{\bullet \bullet \bullet}) + ([\bar{y}_{i \bullet \bullet} - \bar{y}_{\bullet \bullet \bullet}] - \alpha_i) + ([\bar{y}_{\bullet j \bullet} - \bar{y}_{\bullet \bullet \bullet}] - \beta_j) \right. + \\
116- &\left. ([\bar{y}_{i j \bullet} - \bar{y}_{i \bullet \bullet} - \bar{y}_{\bullet j \bullet} + \bar{y}_{\bullet \bullet \bullet}] - \gamma_{ij}) \right]^2 \\
117- = \frac{1}{\sigma^2}\sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} & \left[ (y_{ijk} - \bar{y}_{i j \bullet}) + (\bar{y}_{\bullet \bullet \bullet}) + (\bar{y}_{i j \bullet} - \bar{y}_{\bullet \bullet \bullet} - \alpha_i - \beta_j - \gamma_{ij}) \right]^2
118- \end{split}
119- $$
120-
121- Because the following sum over $k$ is zero for all $(i,j)$
122-
123- $$ \label{eq:sum-yijk}
124- \begin{split}
125- \sum_{k=1}^{n_{ij}} (y_{ijk} - \bar{y}_{i j \bullet}) &= \sum_{k=1}^{n_{ij}} y_{ijk} - n_{ij} \bar{y}_{ij \bullet} \\
126- &= \sum_{k=1}^{n_{ij}} y_{ijk} - n_{ij} \cdot \frac{1}{n_{ij}} \sum_{k=1}^{n_{ij}} y_{ijk} \\
127- &= 0, \; (i,j) \in \left\lbrace 1, \ldots, a \right\rbrace \times \left\lbrace 1, \ldots, b \right\rbrace \; ,
128- \end{split}
129- $$
130-
131- the following sum over $(i,j,k)$ and is also zero
132-
133- $$ \label{eq:sum-yib}
134- \begin{split}
135- \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (\bar{y}_{i j \bullet} - \bar{y}_{\bullet \bullet \bullet}) &= \sum_{i=1}^{a} \sum_{j=1}^{b} n_{ij} \bar{y}_{i j \bullet} - \bar{y}_{\bullet \bullet \bullet} \sum_{i=1}^{a} \sum_{j=1}^{b} n_{ij} \\
136- &= \sum_{i=1}^{a} \sum_{j=1}^{b} n_{ij} \cdot \frac{1}{n_{ij}} \sum_{k=1}^{n_{ij}} y_{ijk} - n \cdot \frac{1}{n} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} y_{ijk} \\
137- &= 0
138- \end{split}
139- $$
140-
141- and the term $\bar{y}_ {\bullet \bullet \bullet}$ does not depend on $i$, $j$ and $k$
142-
143- $$ \label{eq:yb-const}
144- \bar{y}_{\bullet \bullet \bullet} = \text{const.} \; ,
145- $$
146-
147- non-square products in \eqref{eq: sum-Uijk-s2 } disappear and the sum reduces to
148-
149- $$ \label{eq:sum-Uijk-s3}
150- \begin{split}
151- \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} U_{ijk}^2 = \frac{1}{\sigma^2} & \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} \left[ (y_{ijk} - \bar{y}_{i j \bullet})^2 + (\bar{y}_{\bullet \bullet \bullet})^2 + (\bar{y}_{i j \bullet} - \bar{y}_{\bullet \bullet \bullet} - \alpha_i - \beta_j - \gamma_{ij})^2 + \right] \\
152- = \frac{1}{\sigma^2} & \left[ \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} \right. (y_{ijk} - \bar{y}_{i j \bullet})^2 + \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (\bar{y}_{\bullet \bullet \bullet})^2 + \\
153- & \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} \left. (\bar{y}_{i j \bullet} - \bar{y}_{\bullet \bullet \bullet} - \alpha_i - \beta_j - \gamma_{ij})^2 \right]
154- \end{split}
155- $$
156-
157- [ Cochran's theorem] ( /P/snorm-cochran ) states that, if a sum of squared [ standard normal] ( /D/snorm ) [ random variables] ( /D/rvar ) can be written as a sum of squared forms
158-
159- $$ \label{eq:cochran-p1}
160- \begin{split}
161- \sum_{i=1}^{n} U_i^2 = \sum_{j=1}^{m} Q_j \quad &\text{where} \quad Q_j = U^\mathrm{T} B^{(j)} U \\
162- &\text{with} \quad \sum_{j=1}^{m} B^{(j)} = I_n \\
163- &\text{and} \quad r_j = \mathrm{rank}(B^{(j)}) \; ,
164- \end{split}
165- $$
166-
167- then the terms $Q_j$ are [ independent] ( /D/ind ) and each term $Q_j$ follows a [ chi-squared distribution] ( /D/chi2 ) with $r_j$ degrees of freedom:
168-
169- $$ \label{eq:cochran-p2}
170- Q_j \sim \chi^2(r_j), \; j = 1, \ldots, m \; .
171- $$
172-
173- First, we define the $n \times 1$ vector $U$:
174-
175- $$ \label{eq:U}
176- U = \left[ \begin{matrix} u_{1 \bullet} \\ \vdots \\ u_{a \bullet} \end{matrix} \right] \quad \text{where} \quad u_{i \bullet} = \left[ \begin{matrix} u_{i1} \\ \vdots \\ u_{ib} \end{matrix} \right] \quad \text{where} \quad u_{ij} = \left[ \begin{matrix} (y_{i,j,1} - \alpha_i - \beta_j - \gamma_{ij})/\sigma \\ \vdots \\ (y_{i,j,n_{ij}} - \mu - \alpha_i - \beta_j)/\sigma \end{matrix} \right] \; .
177- $$
178-
179- Next, we specify the $n \times n$ matrices $B$
180-
181- $$ \label{eq:B}
182- \begin{split}
183- B^{(1)} &= I_n - \mathrm{diag}\left[ \mathrm{diag}\left( \frac{1}{n_{11}} J_{n_{11}}, \; \ldots, \; \frac{1}{n_{1b}} J_{n_{1b}} \right), \; \ldots, \; \mathrm{diag}\left( \frac{1}{n_{a1}} J_{n_{a1}}, \; \ldots, \; \frac{1}{n_{ab}} J_{n_{ab}} \right) \right] \\
184- B^{(2)} &= \frac{1}{n} J_n \\
185- B^{(3)} &= \mathrm{diag}\left[ \mathrm{diag}\left( \frac{1}{n_{11}} J_{n_{11}}, \; \ldots, \; \frac{1}{n_{1b}} J_{n_{1b}} \right), \; \ldots, \; \mathrm{diag}\left( \frac{1}{n_{a1}} J_{n_{a1}}, \; \ldots, \; \frac{1}{n_{ab}} J_{n_{ab}} \right) \right] - \frac{1}{n} J_n
186- \end{split}
187- $$
188-
189- where $J_n$ is an $n \times n$ matrix of ones, $J_ {n,m}$ is an $n \times m$ matrix of ones and $\mathrm{diag}\left( A_1, \ldots, A_n \right)$ denotes a block-diagonal matrix composed of $A_1, \ldots, A_n$. We observe that those matrices satisfy
190-
191- $$ \label{eq:U-Q-B}
192- \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} U_{ijk]^2 = Q_1 + Q_2 + Q_3 = U^\mathrm{T} B^{(1)} U + U^\mathrm{T} B^{(2)} U + U^\mathrm{T} B^{(3)} U
193- $$
194-
195- as well as
196-
197- $$ \label{eq:B-In}
198- B^{(1)} + B^{(2)} + B^{(3)} = I_n
199- $$
200-
201- and their ranks are
202-
203- $$ \label{eq:B-rk}
204- \begin{split}
205- \mathrm{rank}\left( B^{(1)} \right) &= n - a \cdot b \\
206- \mathrm{rank}\left( B^{(2)} \right) &= 1 \\
207- \mathrm{rank}\left( B^{(3)} \right) &= n - (n-ab) - 1 = a \cdot b - 1 \; .
208- \end{split}
209- $$
210-
211- Let's write down the [ explained sum of squares] ( /D/ess ) and the [ residual sum of squares] ( /D/rss ) for [ two-way analysis of variance] ( /D/anova2 ) as
212-
213- $$ \label{eq:ess-rss}
214- \begin{split}
215- \mathrm{ESS} &= \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (\bar{y}_{\bullet \bullet \bullet})^2 \\
216- \mathrm{RSS} &= \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (y_{ijk} - \bar{y}_{i j \bullet})^2 \; .
217- \end{split}
218- $$
219-
220- Then, using \eqref{eq: sum-Uijk-s3 }, \eqref{eq: cochran-p1 }, \eqref{eq: cochran-p2 }, \eqref{eq: B } and \eqref{eq: B-rk }, we find that
221-
222- $$ \label{eq:ess-rss-dist}
223- \begin{split}
224- \frac{\mathrm{ESS}}{\sigma^2} = \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} \left( \frac{\bar{y}_{\bullet \bullet \bullet}}{\sigma} \right)^2 &= Q_2 = U^\mathrm{T} B^{(2)} U \sim \chi^2(1) \\
225- \frac{\mathrm{RSS}}{\sigma^2} = \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} \left( \frac{y_{ijk} - \bar{y}_{i j \bullet}}{\sigma} \right)^2 &= Q_1 = U^\mathrm{T} B^{(1)} U \sim \chi^2(n-ab) \; .
226- \end{split}
227- $$
228-
229- Because $\mathrm{ESS}/\sigma^2$ and $\mathrm{RSS}/\sigma^2$ are also independent by \eqref{eq: cochran-p2 }, the F-statistic from \eqref{eq: anova2-fgm } is equal to the ratio of two independent [ chi-squared distributed] ( /D/chi2 ) [ random variables] ( /D/rvar ) divided by their degrees of freedom
85+ Thus, the F-statistic from \eqref{eq: anova2-fgm } is equal to the ratio of two [ independent] ( /D/ind ) [ chi-squared distributed] ( /D/chi2 ) [ random variables] ( /D/rvar ) divided by their degrees of freedom
23086
23187$$ \label{eq:anova2-fgm-ess-tss}
23288\begin{split}
233- F_M &= \frac{(\mathrm{ESS}/\sigma^2)/(1)}{(\mathrm{RSS}/\sigma^2)/(n-ab)} \\
234- &= \frac{\mathrm{ESS}/(1)}{\mathrm{RSS}/(n-ab)} \\
235- &= \frac{\sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (\bar{y}_{\bullet \bullet \bullet})^2}{\frac{1}{n-ab} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (y_{ijk} - \bar{y}_{i j \bullet})^2} \\
236- &= \frac{(\bar{y}_{\bullet \bullet \bullet})^2 \sum_{i=1}^{a} \sum_{j=1}^{b} n_{ij}}{\frac{1}{n-ab} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (y_{ijk} - \bar{y}_{i j \bullet})^2} \\
237- &= \frac{n (\bar{y}_{\bullet \bullet \bullet})^2}{\frac{1}{n-ab} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (y_{ijk} - \bar{y}_{i j \bullet})^2}
89+ F_M &= \frac{(\mathrm{SS}_M/\sigma^2)/(1)}{(\mathrm{SS}_\mathrm{res}/\sigma^2)/(n-ab)} \\
90+ &= \frac{\mathrm{SS}_M/(1)}{\mathrm{SS}_\mathrm{res}/(n-ab)} \\
91+ &\overset{\eqref{eq:anova2-ss-dist}}{=} \frac{n (\bar{y}_{\bullet \bullet \bullet} - \mu)^2}{\frac{1}{n-ab} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (y_{ijk} - \bar{y}_{i j \bullet})^2} \\
92+ &\overset{\eqref{eq:anova2-fgm}}{=} \frac{n (\bar{y}_{\bullet \bullet \bullet})^2}{\frac{1}{n-ab} \sum_{i=1}^{a} \sum_{j=1}^{b} \sum_{k=1}^{n_{ij}} (y_{ijk} - \bar{y}_{i j \bullet})^2}
23893\end{split}
23994$$
24095
0 commit comments