JOINT ESTIMATION FOR NORMAL ORTHOGONAL MIXED MODELS
Vera de Jesus 1 , Sandra Saraiva Ferreira 2 and
Jo˜ ao Tiago Mexia 1
1 Department of Mathematics Faculty of Science and Technology
New University of Lisbon
Monte de Caparica 2829–516 Caparica, Portugal e-mail: veramjesus@gmail.com
e-mail: jtm@fct.unl.pt
2 Mathematics Department Universidade da Beira Interior
6201–001 Covilh˜ a, Portugal e-mail: sandra@mat.ubi.pt
Abstract
Commutative Jordan algebras are used to express the structure of mixed orthogonal models and to derive complete sufficient statistics.
From these statistics, UMVUE, (Uniformly Minimum Variance Un-
biased Estimators), are derived for the relevant parameters, first of
single models then of several such models. These models may corre-
spond to experiments designed separately so our results may be seen
as a contribution to this meta-analysis.
2000 Mathematics Subject Classification: 62K15.
1. Introduction
In this paper we consider joint estimation for normal orthogonal mixed models.
Firstly we consider single models using commutative Jordan algebras to study their structure. This study, along with the assumption of normality, leads to the obtention of complete sufficient statistics. From these statistics UMVUE for the relevant parameters are derived.
Nextly the sufficient statistics for the individual models are pooled to give sufficient statistics for the joint model. Again UMVUE are derived from these statistics.
The joined experiments may have been designed separately so our results may be seen as a contribution to their meta-analysis.
2. Single models
A linear space A constituted by symmetrical matrices will be a Jordan al- gebra if it contains the square of it’s matrices. If the matrices of an algebra commute, the Jordan algebra is said to be commutative. We will consider only commutative Jordan algebras.
Seely (1971), showed that any commutative Jordan algebra has one and only one basis constituted by pairwise orthogonal projection matrices. The basis will be the principal basis of the algebra.
We are interested in models
(1) Y n =
X w i=1
X i β c ii+ e n ,
where β c 11, ..., β c mm are fixed vectors and the β c m+1m+1, ..., β c ww and e n are ran-
dom independent vectors, with null mean vectors and covariance matrices
σ i 2 I ci, i = m + 1, ..., w, and σ 2 I n , respectively. Then Y n has mean vector
and covariance matrix
are fixed vectors and the β c m+1m+1, ..., β c ww and e n are ran-
dom independent vectors, with null mean vectors and covariance matrices
σ i 2 I ci, i = m + 1, ..., w, and σ 2 I n , respectively. Then Y n has mean vector
and covariance matrix
and e n are ran-
dom independent vectors, with null mean vectors and covariance matrices
σ i 2 I ci, i = m + 1, ..., w, and σ 2 I n , respectively. Then Y n has mean vector
and covariance matrix
(2)
µ n =
X m i=1
X i β c ii
V = X w i=m+1
σ i 2 M i + σ 2 I n ,
with M i = X i X > i , i = 1, ..., w. We will assume that X 1 = 1 n and that β 1 = µ n .
This model will be associated to a commutative Jordan algebra A if the matrices M i , i = 1, ..., w, and I n constitute a basis for A. With {Q 1 , ..., Q w+1 } the principal basis of A we have
(3)
M i =
w+1 X
j=1
b i,j Q j , i = 1, ..., w
I n =
w+1 X
j=1
Q j
.
Putting b w+1,j = 1, j = 1, ..., w + 1, the matrix B = [b i,j ] will be regular since it is a matrix for change of basis. We say that B is a transition matrix of A.
So, Y n will have normal distribution Y n ∼ N(µ n , V ), with mean vector µ n and putting σ 2 = σ 2 w+1 , we have the covariance matrix
(4) V =
X w i=m+1
σ 2 i
w+1 X
j=1
b i,j Q j
+ σ 2
w+1 X
j=1
Q j =
w+1 X
j=1
γ j Q j ,
(5) γ j =
w+1 X
i=m+1
b i,j σ 2 i = σ 2 + X w
i=1
b i,j σ i 2 , j = 1, ..., w + 1.
Moreover, if we can group the matrices from the two basis in pairs (M i , Q i ), i
= 1, ..., w + 1, so that:
(6) B =
B 1,1 0 B 2,1 B 2,2
with B 1,1 and B 2,2 square matrices, these matrices will be regular and the model, see Ferreira (2006), will be segregated with ϕ = {1, ..., m} the set of index for the fixed effects. Then the pairs of matrices (M i , Q i ), i = 1, ..., m will be associated to the fixed effects part of the model which will behave as a sub-model segregated from the global model. So, considering σ 2 (1) = 0 m , and putting
(7)
γ (1) = (γ 1 , ..., γ m ) γ (2) = (γ m+1 , ..., γ w+1 ) σ 2 (2) = (σ 2 m+1 , ..., σ w+1 2 )
we have
(8) γ (j) = B > 2,j σ 2 (2) , j = 1, 2,
so that, since B > 2,2 is regular,
(9)
σ 2 (2) =
B > 2,2 −1 γ (2)
γ (1) = B > 2,1 B > 2,2
−1 γ (2)
.
With A j the matrix whose g j row vectors are the eigenvectors of Q j corresponding to eigenvalues equal to 1, j = 1, ..., w + 1, we show, see Fonseca et al. (2006), that the statistics
(10)
S j = ke η g jjk 2 , j = 1, ..., w + 1 e
η g jj = A j Y n , j = 1, ..., m .
are sufficient, complete and independent.
With η j = A j µ n , j = 1, ..., w, and g j = rank(Q j ), j = 1, ..., m, we have
(11) η e j ∼ N(η j , γ j I gj), j = 1, ..., m,
as well as
(12) S j ∼ γ j χ 2 g
j
, j = m + 1, ..., w.
Using now the Blackwell-Lehmann-Scheff´e theorem, the estimators e η j , j = 1, ..., m, and eγ j = s j
g j , j = m+1, ..., w +1 will be UMVUE of η j , j = 1, ..., m, and of γ j , j = m + 1, ..., w + 1. So we have the UMVUE
(13) γ e (2) = (eγ m+1 , ..., γ e w+1 )
(14)
e
σ 2 (2) = (B > 2,2 ) −1 γ e (2) e
γ (1) = B > 2,1 (B > 2,2 ) −1 e γ (2) .
3. Joint estimation
Let us assume that there are L experiments involving the same factors with fixed and with random effects. The factor levels will be different from ex- periment to experiment since they were designed separately, according to orthogonal segregated mixed models.
The CJA associated to those models will have principal basis {Q 1 (l), ..., Q w (l)} and transition matrices
(15) B(l) =
B 1,1 (l) B 1,2 (l) B 2,1 (l) B 2,2 (l)
, l = 1, ..., L,
so that we will have
(16)
σ 2 (2) (l) =
B 2,2 (l) > −1
γ (2) (l), l = 1, ..., L
γ (1) (l) = B 2,1 (l) >
B 2,2 (l) > −1
γ (2) (l), l = 1, ..., L ,
as well, with Q j (l) = A j (l) > A j (l), l = 1, ..., L, j = 1, ..., w, we have
(17) η j (l) = A j (l)µ l , l = 1, ..., L, j = 1, ..., w,
where µ l , l = 1, ..., L, is the mean vector of the observations vector for the l-th experiment, l = 1, ..., L. For the different models we have the complete sufficient statistics [e η 1 (l), ..., e η m (l), S m+1 (l), ..., S w (l)] , l = 1, ..., L, which pooled together give sufficient complete statistics for the joint model.
Since the factor levels will not be the same the joint analysis will have to be based on the random effects part of the models mainly on the variance components. We are this led to take
(18) γ 2 (l) = γ 2 , l = 1, ..., L.
Actually we may validate thus assumption testing the equality of the corresponding components of the γ 2 (l), l = 1, ..., L.
Once this assumption is validated we may combine the estimators e
γ 2 (l), l = 1, ..., L. Now S m+j (l) ∼ γ m+j χ 2 g
m+j
(l) , l = 1, ..., L, j = 1, ..., w −m, so
(19) S m+j = X L l=1
S m+j (l) ∼ γ m+j χ 2 g
m+j