LINEAR MODEL GENEALOGICAL TREE
APPLICATION TO AN ODONTOLOGY EXPERIMENT
Ricardo Covas
Managing School, Polytechnic Institute of Tomar
Estrada da Serra, Quinta do Contador, 2300–313 Tomar, Portugal e-mail: ricardocovas@gmail.com
Abstract
Commutative Jordan algebras play a central part in orthogonal models. We apply the concepts of genealogical tree of an Jordan algebra associated to a linear mixed model in an experiment conducted to study optimal choosing of dentist materials. Apart from the conclu- sions of the experiment itself, we show how to proceed in order to take advantage of the great possibilities that Jordan algebras and mixed linear models give to practitioners.
Keywords: commutative Jordan algebra, binary operations, Kronecker matrix product, lattice, projectors.
2000 Mathematics Subject Classification: 17C65, 62J10.
1. Introduction
Jordan algebras were first introduced by [7] as part of a new framework
for quantum mechanics. The use of these algebras in statistical inference
started with the seminal papers of Seely, [13, 14] and [15]. This work as been
carried on by many authors, see for instance [6], [9, 10], [12] and [8]. We are
mainly interested in commutative Jordan algebras which, see for instance
[5], play a central part in the study of orthogonal models.
Consider an experiment carried out to study the differences of two different cements (C
1and C
2) which were putted in the market for tooth treatments.
These differences are measured in terms of an index S (which is the response variable) that measures the solidification of the cement. The idea is that the sooner the cement gets solidified, the better it is, so that the dentist can call in the next patient. This cement is applied conjointly with three different photopolymerizers (F
1, F
2and F
3) intended to aid the solidification of the cement. The index of solidification, S was measured at two distinct times (t
1and t
2) since depending on the treatment, some degree of solidification is enough (or, “I have only t
iminutes to spare with this patient... which cement with which polimeryzer should I use?”) in 5 disks (d
1, d
2, d
3, d
4and d
5) that gave 3 observations each (3 replicates).
The design and the analysis of the experiment is made according to the properties and the pertinent basis of the Jordan algebra that is associated to the linear mixed model used to interpret the experiment. These are well de- fined and explained in [2] and [3]. As part of these properties, we will make use of two different binary operations in Jordan algebras, the Kronecker product (⊗) and the restricted Kronecker product (?), which were first in- troduced in [5] and were developed in [2]. The most important theoretical results are resumed in the next section.
2. Theoretical results 2.1. Binary operations and the genealogical tree
We start by defining the Kronecker product between two families of matrices.
Definition 1. Given the families of matrices, M
1= {M
1i, i = 1, ..., w
1} and M
2= {M
2i, i = 1, ..., w
2}, we take
M
1⊗ M
2= {M
1i⊗ M
2j: i = 1, ..., w
1; j = 1, ..., w
2}.
Suppose that A
i= sp(M
i), i = 1, 2, are Commutative Jordan Algebras (CJA). In [13] we can see that it exists, for each CJA, a unique principal base, i.e., a base constituted by a family of mutual orthogonal orthogonal projection matrices (F M OOP M ). Let Q
1= {Q
1i, i = 1, ..., w
1} and Q
2= {Q
2i, i = 1, ..., w
2} be the principal basis of A
1and A
2. Also, with Q = Q
1⊗ Q
2we put
A
1⊗ A
2= sp(Q).
Proposition 1. A
1⊗ A
2is a CJ A and Q
1⊗ Q
2is it’s principal basis.
Let M
i∈ M = M
1⊗ M
2, then M
i= M
1i1⊗ M
2i2with M
1i1∈ M
1and M
2i2∈ M
2. Supposing
M
1i1=
w1
X
j1=1
b
1i1j1Q
1j1and M
2i2=
w2
X
j2=1
b
2i2j2Q
2j2,
we have
M
i=
w
X
1w2j=1
b
ijQ
j,
where b
ij= b
1i1j1b
2i2j2and Q
j= Q
1j1⊗ Q
2j2, with i = i
2+ (i
1− 1)w
2and j = j
2+ (j
1− 1)w
2. From here, it’s straightforward to show that the transition matrix (this matrix is the matrix such that the i-th line are the coordinates of matrix M
iwith respect to the matrices of family Q, please see [3]) between M and Q is
(1) B = B
1⊗ B
2,
where B
1is the transition matrix between M
1and Q
1and B
2is the transition matrix between M
2and Q
2.
The identity element of A
iis
(2) K
i=
wi
X
j=1
Q
ij.
Proposition 2. Given k = 1, ..., w
2− 1, the family
Q
k= {Q
1h⊗Q
2h0, h = 1, ...., w
1, h
0= 1, ..., k}∪{K
1⊗Q
2h, h = k+1, ..., w
2} is a F M OOP M .
The CJA with principal basis Q
kwill be the restricted k Kronecker product of A
1and A
2. We represent this CJA by A
1?
kA
2. When k = 1, we write A
1? A
2.
Remark that A
1?
w2A
2= A
1⊗ A
2. The operation ?
kcan, in fact, be
generalized to any two families of matrices. We are interested in the case of
when, instead of only dealing with the principal basis of CJA’s, we operate
families M
1and M
2of commuting symmetric matrices such that Q
1and Q
2are the principal basis of A
1= sp(M
1) and A
2= sp(M
2). Putting
M
1?
kM
2= {M
ih⊗ M
2h0, h = 1, ...., w
1, h
0= 1, ..., k}
∪{K
1⊗ M
2h, h = k + 1, ..., w
2}, any matrix, say M, of sp(M
1?
kM
2) will be of the form
(3) M =
w1
X
i1=1
X
k i2=1a
1i1i2M
1i1⊗ M
2i2+
w2
X
i3=k+1
a
2i3K
1⊗ M
2i3.
We now have
Proposition 3. Let M
1and M
2be two families of commuting symmetric matrices and Q
1, Q
2the principal basis of A
1= sp(M
1) and A
2= sp(M
2), assume also that A
2is segregated with separation value k, i.e.,
B
2=
"
B
110 B
21B
22# ,
where B
11is of the size k × k. Then,
A
1?
kA
2= sp(M
1?
kM
2).
Besides this proposition, it’s straightforward to see that, if A
2has segregation value k, given B
1, the transition matrix of A
1, the transition matrix of A
1?
kA
2will be
(4) B =
"
B
1⊗ B
110 1
0w1⊗ B
21B
22# .
Moreover, it’s trivial to see that, A
1?
kA
2will be segregated with separation value w
1k.
One case of singular importance, as we shall see later on, is the operation
A
1? A
2, when A
2is complete and segregated with separation value 1. In
this case, we have
B
2=
"
n
20 b B
22# ,
where b is of type (w
2−1)×1 and B
22is “almost” B
2, since it’s only missing the first line and the first column of B
2. The matrix B is then given by
(5) B =
" n
2B
10 1
0w1⊗ b B
22# .
These concepts are closely connected to linear mixed models. In [5] and [2] we may see that all crossing, nesting and replicates in a mixed linear model can be explained trough the ⊗ and ? products of CJA’s. In fact it is possible to trace back the model building until we reach singular CJA’s, drawing a genealogical tree for a model. This concept is deeply explained in [2] where a singular CJA is defined by the one of the simplest linear model, the random sample. This CJA has principal basis given by {
1nJ, ¯ J}, where J = 11
0and ¯ J = I −
n1J, and is denoted by A (n).
This procedure is useful to obtain the principal basis of CJA’s associated to models, starting from very simple input. We just write the factor by lexicographic order and, between them, we write ⊗ if the first crosses the following, or ? if the second is nested in the first. We will illustrate this procedure later on, when writing the model to interpret the experiment referred in the introduction.
2.2. Optimal estimators Let
Y ∼ N
1µ + X
mi=2
X
iβ
i,
w−1
X
j=m+1
σ
j2M
j+ σ
2I
be an orthogonal linear model. Putting M
1= 11
0, M
i= X
iX
0i, i = 2, ..., m
and M
w= I, we have the M family, {M
1, ..., M
w} and the principal basis
Q = {Q
1, ..., Q
w} of A = sp(M ) = sp(Q). In [2] we have necessary and
sufficient conditions for this last equality to hold. The transition matrix is
given by B = [b
ij] which we suppose to be segregated with separation value
m, so that
(6) B =
"
B
110 B
21B
22# , B
0=
"
B
011B
0210 B
022#
and
(7) (B
0)
−1= U =
"
U
11U
120 U
22# .
We point out that the variance covariance matrix can be rewritten as
(8) V =
X
w j=1γ
jQ
j,
where, with σ
w2= σ
2, we have γ
j= P
wi=m+1
b
ijσ
i2. The projection matrix on the range space of the mean vector is
(9) Q =
X
m i=1Q
i.
We suppose that V and Q commute and therefore, please see [16], we have the following
Theorem 4. If Cβ is estimable, d Cβ = C(X
0X)
+X
0Y is it’s BLUE.
Putting A
0= [A
01· · · A
0m], we have Q = A
0A such that we may write
(10) Xβ = A
0η,
where η = AXβ and consider these, instead of the β, as parameters of the
model. Since A and X are known, we have η b = AXb β = AXX
+QY, and,
remembering that XX
+= Q, we get
(11) η b = AY and consequently
(12) d Cη = CAY.
We can also write, for each i ∈ {1, ..., m}, η
i= AX
iβ
iand η b
i= A
iY.
Using this parameterization has some advantages, as we shall see later on.
We will now focus on equation (8). Putting σ
12= · · · = σ
m2= 0, σ
2= [σ
21· · · σ
2w]
0and γ = [γ
1· · · γ
w]
0we can write
(13) γ = B
0σ
2,
and, with
• σ
2[1]= [σ
21· · · σ
m2]
0,
• σ
2[2]= [σ
2m+1· · · σ
2w]
0,
• γ
[1]= [γ
1· · · γ
m]
0• γ
[2]= [γ
m+1· · · γ
w]
0, we have
(14) γ
[1]= B
021σ
2[2]as well
(15) σ
2[2]= U
22γ
[2].
These two last expressions are of extreme importance, since they show that once we have an unbiased estimator for γ
[2]we also have for σ
2[2]and γ
[1].
Since E[Y] = 1µ + P
mi=2
X
iβ
i, we have that E[Y] ∈ R ( L
m i=1M
i), which, due to the segregation of the transition matrix, belongs to the sub-space R ( L
mi=1
Q
i) that is orthogonal to R L
wi=m+1
Q
i. Thus
(16) E[A
iY] = 0, i = m + 1, ..., w,
where A
0iA
i= Q
i. The variance covariance matrix of A
iY, i = 1, ..., w, is
Σ(A
iY) = A
iVA
0i(17)
= A
iX
w j=1γ
jQ
jA
0i(18)
= X
w j=1γ
jA
iA
0jA
jA
0i(19)
= γ
iI
gi. (20)
Since Σ(A
iY ) = E[(A
iY − E[A
iY])(A
iY − E[A
iY])
0], for i = m + 1, ..., w, we have
(21) γ
iI
gi= E
(A
iY)(A
iY)
0.
From (16) and (21), we get
(22) E
(A
iY)
0(A
iY)
= tr(γ
iI
gi) = γ
ig
i.
Putting
(23) S
i= kA
iYk
2= (A
iY)
0(A
iY ) = Y
0Q
iY = tr(Q
iYY
0) =< Q
i, YY
0>,
we have
(24) E[S
i] = γ
ig
i,
which immediately leads us to take
(25) γ e
i= S
ig
ias an unbiased estimator of γ
i, and therefore, γ g
[2]= [ eγ
m+1· · · eγ
w]
0is an un- biased estimator of γ
[2], from which we obtain for σ
2[2]and γ
[1]the unbiased estimators
(26) σ g
2[2]= U
22g γ
[2]and
(27) γ g
[1]= B
021σ g
2[2]. Having
1. det(V) = Q
wj=1
γ
jgj2. V
−1= P
w j=1γ
j−1Q
jthe density of Y will be
(28)
n(y|µ, V) = exp −
12(y − µ)
0V
−1(y − µ) (2π)
n2Q
w j=1γ
gj 2
j
=
exp −
12P
wj=1
(y − µ)
0Q
j(y − µ)
!
(2π)
n2Q
w j=1γ
gj 2
j
.
Since Q
j= A
0jA
jand, for j > m, Q
jµ = 0, we have that
(29) (y − µ)
0Q
j(y − µ) =
kA
j(y − µ)k
2= kη
j− c η
jk
2j ≤ m kA
jYk
2= S
jj > m
,
and therefore,
(30) n(y|µ, V) = e
−12 Pm
j=1 1
γjkηj−ηcjk2+ Pw
j=m+1 Sj γj
!
(2π)
n2Q
m j=1γ
gj 2
j
.
Theorem 5. In a linear mixed normal model, the statistics η c
jand S
j, defined above, are sufficient and complete.
Given the Blackwell-Lehmann-Scheff´e theorem we then have
Corollary 6. The estimators γ g
[2], g σ
2[2], γ g
[1]and η b
j, defined above, are U M V U E.
From equations (17) to (20), we have that d Cη
j∼ N Cη
j, γ
jCC
0, j = 1, ..., m, (31)
S
j∼ γ
jχ
2(gj)
, j = m + 1, ..., w (32)
are mutually independent.
2.3. Pivot variables
According to the preceding section, we get the pivot variables
1 γ
jd Cη
j− Cη
j0
(CC
0)
+d Cη
j− Cη
j∼ χ
2(c), j = 1, ..., m, c = r(C) (33)
S
jγ
j∼ χ
2(gj)
, j = m + 1, ..., w.
(34)
Clearly, all the γ
j, j = 1, ..., m, are (would be) nuisance parameters.
From equations (14) and (15), we may write
(35) γ
[1]= B
021U
22γ
[2].
This last equation enables us to write (33) in such a way that it only depends on γ
[2]. If c
jis such that, for any given j ∈ 1, ..., m, γ
j= c
0jγ
[2], we have
(36) 1
c
0jγ
[2]d Cη
j− Cη
j0(CC
0)
+Cη d
j− Cη
j∼ χ
2(gj), j = 1, ..., m.
Writing this equation in such fashion entails an enormous advantage, since we may induce a density function for any γ
j, j = m + 1, ..., w, say f (γ
j). This is possible since
Sγjj
is an inducing pivot variable, in fact it is an invertible (with respect to γ
j) function and, moreover, given the observed value s
jof S
j, it’s invertible function is m(z) =
szj, which is measurable since it is continuous. We may read about this subject with much more detail in [1], where we have the induced density of γ
j, , j = m + 1, ..., w,
f (γ
j|s
j) = 1 Γ
g2jγ
js
j2γ
j gj2e
−sj
2γj
; γ
j> 0.
(37)
The statistics S
j, j = m + 1, ..., w, are independent, thus the joint density is
(38) f (γ
[2]|s
m+1, ..., s
w) = Y
w j=m+1f (γ
j|s
j),
with marginals
(39) f (γ
j|s
m+1, ..., s
w), j = 1, ..., m.
If ζ
j(x|s
m+1, ..., s
w, γ
j) is the density of the product of two independent random variables one with density f (γ
j|s
m+1, ..., s
w) and the other a χ
2(gj)
, since the f η
1, ..., η f
mare independent between themselves as well as from the S
m+1, ..., S
w, we may rewrite equations (33) and (34) as
(40)
g Cη
j−Cη
j0(CC
0)
+g Cη
j−Cη
j∼ ζ
j(x|s
m+1, ..., s
w, γ
j), j = 1, ..., m,
(41) γ
j∼ f (γ
j|s
j), j = m + 1, ..., w.
The density function ζ
jhas nuisance parameters, so we may apply Monte-Carlo methods.
It seems easy to obtain confidence intervals or to test hypothesis for γ
j, but for η
jit is not that evident. The work of obtaining confidence ellipsoids for η
jhas already been pursued by [4]. Taking c = r(C) we have the 1 − q level confidence ellipsoid
(g Cη
j− Cη
j)
0(CC
0)
+(g Cη
j− Cη
j) ≤ ζ
1−q,jwith ζ
1−q,jthe 1 − q quantile probability of ζ
j. By the Scheff´e Theorem,
Cη
jlies inside the previous ellipsoid if and only if
(42) \
z
|z
0η b − z
0η| ≤ q
cζ
1−q,jz
0CC
0z
,
so we obtain simultaneous confidence intervals for the z
0η
j. Whenever
|z
0η
j0− z
0η
j| >
q
cζ
1−q,jz
0CC
0z we may reject
H
0: z
0η
j= z
0η
j0with a risk less or equal than q.
3. The experiment
For better understanding of the experiment referred in the introduction, we now describe it in more detail.
The experimenter intends to evaluate the differences of two different cements (C
1and C
2) which are just now in market. These cements are intended for tooth treatments. The differences between the cements are measured in terms of an index that measures the solidification of the cement and that we take as the response variable, i.e., Y. The cements are ranked inversely to the time needed to solidification (in practice the sooner the cement is solidified, the sooner the treatment is complete and the sooner the dentist can call in the next patient, maximizing he’s profit).
The process of solidification is made under the effect of intensive light (the same for both cements), aided by the presence of a photopolymerizer.
There are a few photopolymerizers in the market, from which the three most common were taken into the experiment (F
1, F
2and F
3).
Depending on the tooth treatment made, some degree of solidification
can be enough, so the experimenter was interested in seeing if there were
differences in solidification with time. For example, if only some small grade
of solidification is needed, (meaning more time is spared), it is interesting
to ask which cement with which photopolymerizer should one use. For this
reason, the experiment was repeated at two given times (t
1and t
2).
The experiment was conducted in 5 different disks (d
1, d
2, d
3, d
4and d
5), which constitute the cells, that were big enough to give three uncorrelated observations (r
1, r
2and r
3).
The results of the experiment are resumed in Table 1, in which we present the averages of the observations in each disk.
Table 1. Averages of the disks
t
1t
2d
1d
2d
3d
4d
5d
1d
2d
3d
4d
5C
1F
126.03 28.43 27.40 26.10 26.77 29.37 30.53 30.27 29.80 29.63 F
226.10 26.47 29.90 25.60 24.17 27.13 25.97 29.20 30.77 28.60 F
39.83 10.00 10.17 10.67 11.37 9.83 10.00 10.17 10.67 11.37 C
2F
126.00 26.93 25.17 26.50 25.17 29.97 29.53 29.00 29.03 25.47 F
227.37 26.43 26.23 26.37 27.77 30.67 27.83 28.07 27.40 22.07 F
36.07 6.40 6.63 6.60 6.43 6.07 6.40 6.63 6.60 6.43
3.1. The genealogical tree and the resulting algebraic structure In this, three times replicated, experiment we have three crossed factors,
“cement” (C) which is fixed with two levels, “photopolymerizer” (F ) which is fixed with three levels and “time” (T ) that which is random with 2 levels and nests the factor “disk” (D) which is random with 5 levels.
Therefore, as referred in the second section, the genealogical tree is [C
1, C
2]
0⊗ [F
1, F
2, F
3]
0⊗ [t
1, t
2]
0? [d
1, d
2, d
3, d
4, d
5]
0? [r
1, r
2, r
3]
0and the CJA is
(A (2) ⊗ A (3) ⊗ A (2)) ? A (5) ? A (3).
This Genealogical Tree is, in fact, very practical since it allows us to get not
only the M family and the principal basis of the associated CJA, but also
the incidence matrices of the model. From the definitions of ⊗ and ?, easily
we get
1. M
1= J
2⊗ J
3⊗ J
2⊗ J
5⊗ J
32. M
2= J
2⊗ J
3⊗ I
2⊗ J
5⊗ J
33. M
3= J
2⊗ I
3⊗ J
2⊗ J
5⊗ J
34. M
4= J
2⊗ I
3⊗ I
2⊗ J
5⊗ J
35. M
5= I
2⊗ J
3⊗ J
2⊗ J
5⊗ J
36. M
6= I
2⊗ J
3⊗ I
2⊗ J
5⊗ J
37. M
7= I
2⊗ I
3⊗ J
2⊗ J
5⊗ J
38. M
8= I
2⊗ I
3⊗ I
2⊗ J
5⊗ J
39. M
9= I
12⊗ I
5⊗ J
310. M
10= I
60⊗ I
3and
1. Q
1=
12J
2⊗
13J
3⊗
12J
2⊗
15J
5⊗
13J
32. Q
2=
12J
2⊗
13J
3⊗ ¯ J
2⊗
15J
5⊗
13J
33. Q
3=
12J
2⊗ ¯ J
3⊗
12J
2⊗
15J
5⊗
13J
34. Q
4=
12J
2⊗ ¯ J
3⊗ ¯ J
2⊗
15J
5⊗
13J
35. Q
5= ¯ J
2⊗
13J
3⊗
12J
2⊗
15J
5⊗
13J
36. Q
6= ¯ J
2⊗
13J
3⊗ ¯ J
2⊗
15J
5⊗
13J
37. Q
7= ¯ J
2⊗ ¯ J
3⊗
12J
2⊗
15J
5⊗
13J
38. Q
8= ¯ J
2⊗ ¯ J
3⊗ ¯ J
2⊗
15J
5⊗
13J
39. Q
9= I
12⊗ ¯ J
5⊗
13J
310. Q
10= I
60⊗ ¯ J
3depending if we start with the basis of A (p) constituted by {J
p, I
p} or {
1pJ
p, ¯ J
p}. To get the incidence matrices of the model, it’s not difficult to see that we only have to correspond the set of the usual incidence matrices for the random sample, {1
p, I
p}, and proceed in the same way. Thus,
1. X
1= 1
2⊗ 1
3⊗ 1
2⊗ 1
5⊗ 1
32. X
2= 1
2⊗ 1
3⊗ I
2⊗ 1
5⊗ 1
33. X
3= 1
2⊗ I
3⊗ 1
2⊗ 1
5⊗ 1
34. X
4= 1
2⊗ I
3⊗ I
2⊗ 1
5⊗ 1
35. X
5= I
2⊗ 1
3⊗ 1
2⊗ 1
5⊗ 1
36. X
6= I
2⊗ 1
3⊗ I
2⊗ 1
5⊗ 1
37. X
7= I
2⊗ I
3⊗ 1
2⊗ 1
5⊗ 1
38. X
8= I
2⊗ I
3⊗ I
2⊗ 1
5⊗ 1
39. X
9= I
12⊗ I
5⊗ 1
310. X
10= I
60⊗ I
311. X = [X
1X
2X
3X
4].
The transition matrix can also be taken from the genealogical tree. For A (p) the transition matrix is given by B =
p 0 1 1
, thus from both equations
(1) and (5), we get
(43) B =
180 0 0 0 0 0 0 0 0 0
90 90 0 0 0 0 0 0 0 0
60 0 60 0 0 0 0 0 0 0
30 30 30 30 0 0 0 0 0 0
90 0 0 0 90 0 0 0 0 0
45 45 0 0 45 45 0 0 0 0
30 0 30 0 30 0 30 0 0 0
15 15 15 15 15 15 15 15 0 0
3 3 3 3 3 3 3 3 3 0
1 1 1 1 1 1 1 1 1 1
.
We have identified matrices B
11, B
21and B
22accordingly to equation (6).
Matrix U defined in equation (7) is given by (44)
U =
1 180 − 1
180 − 1 180
1
180 − 1 180
1 180
1
180 − 1
180 0 0
0 1
90 0 − 1
90 0 − 1
90 0 1
90 0 0
0 0 1
60 − 1
60 0 0 − 1
60 1
60 0 0
0 0 0 1
30 0 0 0 − 1
30 0 0
0 0 0 0 1
90 − 1 90 − 1
90 1
90 0 0
0 0 0 0 0 2
90 0 − 2
90 0 0
0 0 0 0 0 0 1
30 − 1
30 0 0
0 0 0 0 0 0 0 2
30 − 2 30 0
0 0 0 0 0 0 0 0 1
30 − 1 30
0 0 0 0 0 0 0 0 0 1
,
where we also identified U
11, U
12and U
22.
It is trivial to write a procedure to obtain each matrix M
i, Q
i, X
i, i = 1, ..., 10 and matrix B which illustrates the enormous advantage of the ge- nealogical tree.
In order to calculate one matrix P, i.e., one common diagonalizer of for all matrices M
i, i = 1, ..., w, it is easier to calculate the roots of matrices Q
i, i = 1, ..., w, which are the matrices A
i, i = 1, ..., w, referred in the previous section. In fact, it is as easy as easy as calculating a singular value decomposition of each matrix Q
i. This can be made in most matrix manipulation software packages where we can obtain matrices U
i, S
iand T
isuch that U
iS
iT
0i= Q
iwhere S
iis a diagonal matrix of the same dimension as Q
iwith nonnegative diagonal elements in decreasing order, and U and T are unitary matrices. Once the singular value decomposition is obtained, we have that A
iis constituted by the first g
ilines of the transpose of Q
iT
i, where g
iis the trace of Q
i. Observe that, in our case, these calculus are even easier to carry out since Q
iis symmetric, meaning that U
i= T
iand, since Q
ihas eigenvalues 1 or 0, S has either 0 or 1 in the diagonal.
3.2. Estimation 3.2.1. Fixed effects
The fixed effects considered in the experiment were the cement, the photopolymerizer and, therefore, the interactions between these.
Accordingly to the objectives explained before, we are interested in es- timating differences between the different levels of cement ([1 (−1)]β
2), photopolymerizer (we chose [1 (−1) 0]β
3and [0 1 (−1)]β
3) and interac- tions (we chose [1 (−1) 0 0 0 0]β
4and [0 0 0 1 (−1) 0]β
4). For this purposes, choosing
C =
1 0 0 0 0 0 0 0 0 0 0 0
0 1 −1 0 0 0 0 0 0 0 0 0
0 0 0 1 −1 0 0 0 0 0 0 0
0 0 0 0 1 −1 0 0 0 0 0 0
0 0 0 0 0 0 1 −1 0 0 0 0
0 0 0 0 0 0 0 0 0 1 −1 0
,
from Theorem 4, we get that
Cβ d =
10.5794 0.7650 3.9811 5.8700
−1.3683 1.8683
is the estimate of
µ C
1− C
2F
1− F
2F
2− F
3C
1F
1− C
1F
2C
2F
1− C
2F
2
.
To any other estimates, we just need to choose any other matrix C. We note that only contrasts are estimable.
3.2.2. Random effects
The random effects and interactions considered in the experiment are, in the design order, time (for which we want to test σ
25), the interaction time×cement (for which we want to test σ
62), the interaction time
× photopolymerizer (for which we want to test σ
27), the interaction time × cement× photopolymerizer (for which we want to test σ
82), and disk (for which we want to test σ
92). Observe that there are no interactions between nested factors and that we will also estimate σ
210= σ
2which correspond to the technical error.
Since matrices Q
i, i = 5, ..., 10, and matrices A
i, i = 5, ..., 10, are already obtained, according to equations (23) and (25), we have
g γ
[2]=
S
5g
5= tr(Q
5yy
0) tr(Q
5) S
6g
6= tr(Q
6yy
0) tr(Q
6) S
7g
7= tr(Q
7yy
0) tr(Q
7) S
8g
8= tr(Q
8yy
0) tr(Q
8) S
9g
9= tr(Q
9yy
0) tr(Q
9) S
10g
10= tr(Q
10yy
0) trQ
10
=
2.2162 × 10
31 = 2.2162 × 10
32.7534 × 10
31 = 2.7534 × 10
3898.2608
2 = 449.1304 687.3381
2 = 343.6691 173.1027
48 = 3.6063 739.3067
120 = 6.1609
which, according to expression (26) enables us to use matrix U
22to calculate
(45) U
22γ g
[2]=
−7.1408 53.5501 3.5154 22.6709
−0.8515 6.1609
which is the estimate of
σ
25σ
26σ
27σ
28σ
29σ
210
.
3.3. Testing
3.3.1. Fixed factors
The hypothesis of interest, at this point, are clear. Concerning
1. cement,
H
0C: There is no difference between C
1and C
2vs.
H
1C: There is a difference between C
1and C
2,
2. photopolymerizer,
H
0F: There are no differences between F
1, F
2and F
3vs.
H
1F: There is at least a difference between F
1, F
2or F
3,
3. interactions cement×photopolymerizer
H
0CF: There are no differences between any
interaction C
1F
1, C
1F
2, C
1F
3, C
2F
1, C
2F
2and C
3F
3vs.
H
1CF: There is at least a difference between
interactions C
1F
1, C
1F
2, C
1F
3, C
2F
1, C
2F
2or C
3F
3. Accordingly to equation (10), these hypothesis are equivalent to
1. (for cement)
H
0C: η
2= 0 vs.
H
1C: η
26= 0, 2. (for photopolymerizer)
H
0F: η
3= 0 vs.
H
1F: η
36= 0, 3. (for interactions cement×photopolymerizer)
H
0CF: η
4= 0 vs.
H
1CF: η
46= 0.
A remark is due at this point. η
2is a scalar and η
i, i = 3, 4, has two components. This is, off course, linked to the rank of the correspondent matrix A
i, i = 2, 3, 4, and is something that can be found in any introductory book of analysis of variance, see for example [11]. According to the definition of effects and interactions, their sums has to be null, i.e., C
1+ C
2= 0, F
1+ F
2+ F
3= 0, and also
F
1F
2F
3sum C
1C
1F
1C
1F
2C
1F
30 C
2C
2F
1C
2F
2C
2F
30
sum 0 0 0 0
.
This means that, for cement, there is only (2 − 1) = 1 effects “free” (or there is 1 degree of freedom), for photopolymerizer there are (3 − 1) = 2 degrees of freedom and for the interaction there are (2 − 1)(3 − 1) = 2 degrees of freedom. This is the reason why, for the cements to be equal, we only need to test if one contrast is null and for photopolymerizer and interactions we need to test if two (any two linearly independent) contrasts are simultaneous null.
The estimates of η
1(that concerns the mean value, and therefore of no interest), η
2, η
3and η
4which, geometrically, are estimates of contrasts that belong to R(Q
2), R(Q
3) and R(Q
4), can be obtained using equation (12),
η
1= −283.8763, η
2= −6.8424, η
3=
48.2270
−65.6120
and η
4=
41.7091 7.7001
.
According to equations (33) and (34) and choosing
C = 1 for cement, (46)
C = I
2for photopolymerizer and (47)
C = I
2for the interaction between them, (48)
we have that, for a significance level α,
1. under H
0C,
γ12
× 46.8180 should be smaller than the (1 − α) quantile of the chi-square distribution with 1 degree of freedom,
2. under H
0F,
γ13
×6.6308×10
3should be smaller than the (1−α) quantile of the chi-square distribution with 2 degrees of freedom,
3. under H
0CF,
γ14
× 1.7989 × 10
3should be smaller than the (1 − α) quantile of the chi-square distribution with 2 degrees of freedom.
For practical reasons, in order to apply the theory of Section 2.3, we will estimate the parameters γ
i, i = 2, 3, 4, or, in fact, use equation (35) and (39) to generate samples for each γ
i, i = 2, 3, 4. Matrix B
021U
22is
1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0 0 0 0 0 1 0 0
= [c
01c
02c
03c
04]
0,
such that γ
j= c
0jχ
2(gj)