Probability and Statistics 33 (2013) 171–190 doi:10.7151/dmps.1153
TESTS FOR PROFILE ANALYSIS
BASED ON TWO-STEP MONOTONE MISSING DATA
Mizuki Onozawa
Department of Mathematical Information Science, Graduate School of Science Tokyo University of Science, Tokyo 162-8601, Japan
e-mail: j1412701@ed.tus.ac.jp Sho Takahashi
Chiba University Hospital Clinical Research Center Chiba 260-8677, Japan
e-mail: sho@chiba-u.jp
and Takashi Seo
Department of Mathematical Information Science, Faculty of Science Tokyo University of Science, Tokyo 162-8601, Japan
e-mail: seo@rs.kagu.tus.ac.jp
Abstract
In this paper, we consider profile analysis for the observations with two- step monotone missing data. There exist three interesting hypotheses – the parallelism hypothesis, level hypothesis, and flatness hypothesis – when comparing the profiles of some groups. The T
2-type statistics and their asymptotic null distributions for the three hypotheses are given for two- sample profile analysis. We propose the approximate upper percentiles of these test statistics. When the data do not have missing observations, the test statistics perform lower than the usual test statistics, for example, as in [8]. Further, we consider a parallel profile model for several groups when the data have two-step monotone missing observations. Under the assumption of non-missing data, the likelihood ratio test procedure is derived by [16].
We derive the test statistic based on the likelihood ratio. Finally, in order to
investigate the accuracy for the null distributions of the proposed statistics,
we perform a Monte Carlo simulation for some selected parameters values.
Keywords: Hotelling’s T
2-type statistic, likelihood ratio, profile analysis, two-step monotone missing data.
2010 Mathematics Subject Classification: 62H15.
1. Introduction
Profile analysis is a statistical method used to compare the profiles of several groups. In a normal population, the profile analysis for a two-sample problem has been discussed using Hotelling’s T
2-type statistic (see, e.g., [8]). Further, [16]
gave a profile analysis of several groups based on the likelihood ratio. For the assumption of nonnormality, [9] discussed profile analysis in elliptical populations.
Further, [7] obtained asymptotic expansions of the null distributions of some test statistics for general distributions.
At the same time, we often encounter the problem of missing data in many practical situations. For samples with observations missing at random, many statistical methods have been developed by [3, 14, 15], and [12] among others.
Moreover, when the missing observations are of the monotone-type, the test for the equality of means and simultaneous confidence intervals in repeated measures with an intraclass correlation model was discussed by [11] in a one-sample prob- lem, [5] in a two-sample problem, and [6] in a k-sample problem. For two-step monotone missing data, [2] and [10] considered tests for the mean vector in a one-sample problem. [1] obtained the maximum likelihood estimators (MLEs) of the mean vector and covariance matrix in a one-sample problem for two-step monotone missing data, and [4] discussed the distribution of these MLEs and expanded for K-step monotone missing data. In the same way as [1], the MLEs in two-sample problem have been obtained (e.g., [13]).
In this paper, we consider a profile analysis for a two-sample problem compris- ing several groups and two-step monotone missing observations. In particular, for several groups, we consider the parallelism hypothesis.
The organization of this paper is as follows. In Section 2, we consider a profile analysis for complete data. In Section 3, we derive the MLEs of µ
(i)and Σ when the missing observations are of the two-step monotone-type. In Section 4, we give the T
2-type statistics for profile analysis. In Section 5, we give the likelihood ratio test statistic for the parallelism hypothesis. In Section 6, we perform a Monte Carlo simulation to investigate the accuracy for the null distributions of these statistics. Finally, in Section 7, we conclude this study.
2. Profile analysis for complete data
In this section, we consider the test statistics when the data have non-missing
observations. Let the p-dimensional random vector x
(i)jbe independently
distributed as N
p(µ
(i), Σ) (j = 1, . . . , N
1(i), i = 1, 2), where µ
(i)= (µ
(i)1, . . . , µ
(i)p)
′. Let the i-th sample mean vector, the i-th sample covariance matrix, and the pooled sample covariance matrix be
x
(i)= 1 N
1(i)N1(i)
X
j=1
x
(i)j, S
i= 1 N
1(i)− 1
N1(i)
X
j=1
(x
(i)j− x
(i))(x
(i)j− x
(i))
′,
S = (N
1(1)− 1)S
1+ (N
1(2)− 1)S
2N
1(1)+ N
1(2)− 2 ,
respectively. When carrying out a profile analysis for two samples, we first con- sider the parallelism hypothesis that is expressed as
H
P2: Cµ
(1)= Cµ
(2)vs. A
P26= H
P2,
where C is a (p − 1) × p matrix of rank p − 1 such that C1
p= 0 and 1
pis a p-vector of ones. The test statistic for testing hypothesis H
P2can be written as
T
P c2= (x
(1)− x
(2))
′C
′( N
1(1)+ N
1(2)N
1(1)N
1(2)(CSC
′) )
−1C(x
(1)− x
(2)).
In normal populations,
T
P c2∼ (N
1(1)+ N
1(2)− 2)(p − 1) N
1(1)+ N
1(2)− p F
p−1,N1(1)+N1(2)−p
.
If the parallelism hypothesis is true, we test the level hypothesis or the flatness hypothesis. The level hypothesis is expressed as
H
L2: 1
′pµ
(1)= 1
′pµ
(2)vs. A
L26= H
L2. The test statistic for testing hypothesis H
L2can be written as
T
Lc2= (x
(1)− x
(2))
′1
p( N
1(1)+ N
1(2)N
1(1)N
1(2)(1
′pS1
p) )
−11
′p(x
(1)− x
(2)) . In normal populations,
T
Lc2∼ F
1,N1(1)+N1(2)−2
. Further, the flatness hypothesis is expressed as
H
F2: C(µ
(1)+ µ
(2)) = 0 vs. A
F26= H
F2.
The test statistic for testing hypothesis H
F2can be written as
T
F c2= x
′12C
′( 1
N
1(1)+ N
1(2)CSC
′)
−1Cx
12,
where
x
12= N
1(1)N
1(1)+ N
1(2)x
(1)+ N
1(2)N
1(1)+ N
1(2)x
(2). In normal populations,
T
F c2∼ (N
1(1)+ N
1(2)− 2)(p − 1) N
1(1)+ N
1(2)− p F
p−1,N1(1)+N1(2)−p
.
In addition, we consider a parallelism hypothesis of several groups when the data have non-missing observations. Let x
(i)1, . . . , x
(i)N1(i)
be N
1(i)independent observa- tions from N
p(µ
(i), Σ) (i = 1, . . . , k). Then we consider the primarily testing the parallelism hypothesis as follows:
H
Pk: Cµ
(1)= · · · = Cµ
(k)vs. A
Pk6= H
Pk. The MLEs of µ
(i)and Σ under A
Pkare
x
(i)= 1 N
1(i)N1(i)
X
j=1
x
(i)j, Σ b
c= 1 N
1X
k i=1N1(i)
X
j=1
(x
(i)j− x
(i))(x
(i)j− x
(i))
′,
respectively, where N
1= P
ki=1
N
1(i). In contrast, the MLEs of µ and Σ under H
Pkare
x = 1 N
1X
k i=1N
X
(i)j=1
x
(i)j, Σ e
c= 1 N
1X
k i=1N1(i)
X
j=1
(x
(i)j− x)(x
(i)j− x)
′,
respectively. For complete data, using these MLEs, we can construct the following likelihood ratio:
Λ
c= |C b Σ
cC
′|
12N1|C e Σ
cC
′|
12N1.
The likelihood ratio test statistic, −2 log Λ
c, is asymptotically distributed as a χ
2distribution with (p − 1)(k − 1) degrees of freedom as N
1(i)s tend to infinity (see
[16]). Hence, we reject H
Pkwhen −2 log Λ
c> χ
2(p−1)(k−1),α, where χ
2(p−1)(k−1),αis the upper 100α percentile of a χ
2distribution with (p − 1)(k − 1) degrees of freedom. However, convergence to the asymptotic χ
2distribution can be im- proved by considering an asymptotic expansion for the likelihood ratio statistic and deriving the modified likelihood ratio statistic as −2ρ
c1log Λ
c, where
ρ
c1= 1 − 1
2N
1(p + k + 1).
3. MLEs
We consider the case when the missing observations are of the two-step monotone- type. Observations {x
(i)ℓj} can be written in the following form:
x
(i)11· · · x
(i)1p1
x
(i)1,p1+1
· · · x
(i)1p.. . .. . .. . .. .
x
(i)N1(i)1
· · · x
(i)N1(i)p1
x
(i)N1(i),p1+1
· · · x
(i)N1(i)p
x
(i)N1(i)+1,1
· · · x
(i)N1(i)+1,p1
∗ · · · ∗
.. . .. . .. . .. .
x
(i)N(i)1
· · · x
(i)N(i)p1
∗ · · · ∗
,
where ∗ denotes missing component. Let x
(i)j≡ (x
(i)1j′, x
(i)2j′)
′(j = 1, . . . , N
1(i), i = 1, . . . , k) be a p-dimensional observation vector from the i-th group with complete data. Let x
(i)1j(j = N
1(i)+ 1, . . . , N
(i)) be p
1-dimensional vectors based on N
2(i)(=
N
(i)−N
1(i)) observations. Now, we assume the distribution of observation vectors:
x
(i)j∼ N
p(µ
(i), Σ) (j = 1, . . . , N
1(i), i = 1, . . . , k), x
(i)1j∼ N
p1(µ
(i)1, Σ
11) (j = N
1(i)+ 1, . . . , N
(i), i = 1, . . . , k), respectively, where
µ
(i)= µ
(i)1µ
(i)2! , Σ =
Σ
11Σ
12Σ
21Σ
22,
and µ
(i)and Σ are partitioned according to the blocks of the data set. Therefore, µ
(i)ℓ(ℓ = 1, 2) is a p
ℓ-dimensional vector and Σ
ℓm(ℓ, m = 1, 2) is a p
ℓ×p
mmatrix.
We give some notations for the sample mean vectors. Let x
(i)1Tbe the sample mean vector of x
(i)11, . . . , x
(i)1N(i)
. Let (x
(i)1F′, x
(i)2F′)
′be the sample mean vector of
x
(i)1, . . . , x
(i)N1(i)
, where x
(i)ℓF′: p
ℓ× 1 (ℓ = 1, 2). That is,
x
(i)1T= 1 N
(i)N
X
(i)j=1
x
(i)1j, x
(i)1F= 1 N
1(i)N1(i)
X
j=1
x
(i)1j, x
(i)2F= 1 N
1(i)N1(i)
X
j=1
x
(i)2j.
Since the MLEs based on the complete data case cannot be used, we have to estimate µ
(i)and Σ under two-step monotone missing data. Let µ b
(i)and b Σ be the MLEs of µ and Σ. These have the same patterns of partition as µ
(i)and Σ.
The likelihood function is L(µ
(i), Σ)
= Y
k i=1
N1(i)
Y
j=1
1
(2π)
p2| Σ |
12exp
− 1
2 (x
(i)j− µ
(i))
′Σ
−1(x
(i)j− µ
(i))
×
N
Y
(i)j=N1(i)+1
1
(2π)
p12| Σ
11|
12exp
− 1
2 (x
(i)1j− µ
(i)1)
′Σ
−111(x
(i)1j− µ
(i)1)
.
Let A be a p × p transformation matrix:
A =
I
p1O
−Σ
21Σ
−111I
p2. Then we have
Ax
(i)j= x
(i)1jx
(i)2j− Σ
21Σ
−111x
(i)1j!
∼ N
p(Aµ
(i), AΣA
′),
where the mean vector and the covariance matrix of transformed observation vectors are
Aµ
(i)= η
(i)= η
(i)1η
(i)2!
= µ
(i)1µ
(i)2− Σ
21Σ
−111µ
(i)1! ,
AΣA
′=
Σ
11O O Σ
22·1,
and Σ
22·1= Σ
22−Σ
21Σ
−111Σ
12. It should be noted that µ
(i)and Σ have one-to-one correspondence with η
(i)and Ψ, where
Ψ =
Ψ
11Ψ
12Ψ
21Ψ
22=
Σ
11Σ
−111Σ
12Σ
21Σ
−111Σ
22·1.
For parameters η
(1), . . . , η
(k)and Ψ, the likelihood function is L(η
(1), . . . , η
(k), Ψ)
= Const.× | Ψ
11|
−12N| Ψ
22|
−12N1× exp
− 1 2
X
k i=1N
X
(i)j=1
(x
(i)1j− η
(i)1)
′Ψ
−111(x
(i)1j− η
(i)1)
× exp
− 1 2
X
k i=1N1(i)
X
j=1
(x
(i)2j− Ψ
21x
(i)1j− η
(i)2)
′Ψ
−122(x
(i)2j− Ψ
21x
(i)1j− η
(i)2)
,
where N = P
ki=1
N
(i).
Differentiating the log likelihood function, we get that b
η
(i)1= x
(i)1T, b
η
(i)2= x
(i)2F− b Ψ
21x
(i)1F, and that
Ψ b
11= 1 N
X
k i=1N(i)
X
j=1
(x
(i)1j− x
(i)1T)(x
(i)1j− x
(i)1T)
′,
Ψ b
21=
X
k i=1N1(i)
X
j=1
z
(i)2jz
′(i)1j
X
ki=1 N1(i)
X
j=1
z
(i)1jz
′(i)1j
−1
,
Ψ b
22= 1 N
1
X
ki=1 N1(i)
X
j=1
z
(i)2jz
′(i)2j−
X
k i=1N1(i)
X
j=1
z
(i)2jz
′(i)1j
X
ki=1 N1(i)
X
j=1
z
(i)1jz
′(i)1j
−1
X
ki=1 N1(i)
X
j=1
z
(i)1jz
′(i)2j
,
z
(i)1j= x
(i)1j− x
(i)1F, z
(i)2j= x
(i)2j− x
(i)2F.
We thus obtain the MLEs of µ
(i)and Σ in general:
b
µ
(i)= µ b
(i)1b µ
(i)2!
= x
(i)1Tx
(i)2F− b Ψ
21(x
(i)1F− x
(i)1T)
! ,
Σ = b Σ b
11Σ b
12Σ b
21Σ b
22!
= Ψ b
11Ψ b
11Ψ b
12Ψ b
21Ψ b
11Ψ b
22+ b Ψ
21Ψ b
11Ψ b
12! .
4. Two-sample profile analysis with two-step monotone missing data
By using the MLEs given in Section 3, we obtain the T
2-type statistics. In this section, let k = 2. The T
2-type statistic under H
P2can be written as
T
P m2= ( µ b
(1)− µ b
(2))
′C
′{C b ΞC
′}
−1C( µ b
(1)− µ b
(2)), where b Ξ is the MLE of Ξ = {Cov[ µ b
(1)] + Cov[ µ b
(2)]},
Ξ = b
N
N
(1)N
(2)Σ b
11N
N
(1)N
(2)Σ b
12N
N
(1)N
(2)Σ b
21Cov[ d µ b
(1)2] + d Cov[ µ b
(2)2]
and
Cov[ d µ b
(1)2] + d Cov[ µ b
(2)2]
= X
2 i=1( 1 N
1(i)Σ b
22− N
2(i)N
(i)Σ b
21Σ b
−111Σ b
12!
+ N
2(i)p
1N
(i)N
1(i)(N
1(i)− p
1− 2) Σ b
22·1) .
For details of the MLEs, see [4]. T
P m2is asymptotically distributed as a χ
2distribution with p − 1 degrees of freedom when N
1(i)s are large.
The T
2-type statistic under H
L2can be written as
T
Lm2= ( µ b
(1)− µ b
(2))
′1
p{1
′pΞ1 b
p}
−11
′p( µ b
(1)− µ b
(2)).
T
Lm2is asymptotically distributed as a χ
2distribution with 1 degree of freedom when N
1(i)s are large.
When we consider the case under H
F2, we can join the two samples and regard it as a one-sample problem. The T
2-type statistic under H
F2can be written as
T
F m2= (C µ) b
′{C d Cov[ µ]C b
′}
−1(C µ), b
where b µ =
µ b
1b µ
2=
x
1Tx
2F− b Σ
21Σ b
−111(x
1F− x
1T)
,
Cov[ d µ] = b
1
N Σ b
111 N Σ b
121
N Σ b
21Cov[ d µ b
2]
,
Cov[ d µ b
2] = 1 N
1Σ b
22− N
2N Σ b
21Σ b
−111Σ b
12+ N
2p
1N N
1(N
1− p
1− 2) Σ b
22·1and
x
1T= 1 N
X
2 i=1N(i)
X
j=1
x
(i)1j, x
1F= 1 N
1X
2 i=1N1(i)
X
j=1
x
(i)1j, x
2F= 1 N
1X
2 i=1N1(i)
X
j=1
x
(i)2j,
N
2= X
ki=1
N
2(i).
These estimators are extended for the MLEs obtained by [4]. T
F m2is asymptot- ically distributed as a χ
2distribution with p − 1 degrees of freedom when N
1(i)s are large.
However, the upper percentiles of the χ
2distribution are not a good approxi- mation for the T
2-type statistic when the sample size is small, and it is difficult to obtain the exact upper percentiles of these statistics when the data have missing observations. Hence, we give the approximate upper percentiles based on the idea of [10] where it is assumed that the true upper percentiles exist between T
p−1,N2 1−p,αand T
p−1,N −p,α2. F
1,α∗can give the approximate upper percentiles of T
P mand T
F m.
F
1,α∗= T
p−1,N2 1−p,α− N p − N
2p
2N p T
p−1,N2 1−p,α− T
p−1,N −p,α2, where
T
p−1,N −p,α2= (N − 2)(p − 1)
N − p F
p−1,N −p,α, T
p−1,N2 1−p,α= (N
1− 2)(p − 1)
N
1− p F
p−1,N1−p,α,
and F
p,q,αis the upper 100α percentile of F distribution with p and q degrees of freedom. Further, F
2,α∗can give the approximate upper percentiles of T
Lm.
F
2,α∗= T
1,N2 1−2,α− N p − N
2p
2N p (T
1,N2 1−2,α− T
1,N −2,α2), where
T
1,N −2,α2= F
1,N −2,α, T
1,N2 1−2,α= F
1,N1−2,α.
5. Parallelism hypothesis for several groups with two-step monotone missing data
We have two-step monotone missing data when k ≥ 3, as in Section 3. First, we transform the observation vectors using C. Then we have
u
(i)j= Cx
(i)j∼ N
p−1(θ
(i), Γ), u
(i)1j= C
1x
(i)1j∼ N
p1−1(θ
(i)1, Γ
11),
where θ
(i)= Cµ
(i), Γ = CΣC
′, and C
1is a (p
1− 1) × p
1matrix of rank (p
1− 1) such that C
11
p1= 0 and 1
p1is a p
1-vector of ones.
θ
(i)= θ
(i)1θ
(i)2! , Γ =
Γ
11Γ
12Γ
21Γ
22.
θ
(i)and Γ are partitioned according to the blocks of the data set. It should be noted that θ
1: (p
1− 1) × 1, θ
2: p
2× 1, Γ
11: (p
1− 1) × (p
1− 1), Γ
12= Γ
′21: (p
1− 1) × p
2, and Γ
22: p
2× p
2. To construct a likelihood ratio, we obtain the MLEs of θ
(i)and Γ in general and under the hypothesis H
Pk. These can be obtained in the same way as earlier:
bθ
(i)= bθ
(i)1bθ
(i)2!
= u
(i)1Tu
(i)2F− b Φ
21(u
(i)1F− u
(i)1T)
! ,
Γ = b Γ b
11b Γ
12Γ b
21b Γ
22!
= Φ b
11Φ b
11Φ b
12Φ b
21Φ b
11Φ b
22+ b Φ
21Φ b
11Φ b
12!
,
where
u
(i)1T= 1 N
(i)N(i)
X
j=1
u
(i)1j, u
(i)1F= 1 N
1(i)N1(i)
X
j=1
u
(i)1j, u
(i)2F= 1 N
1(i)N1(i)
X
j=1
u
(i)2j,
and
Φ b
11= 1 N
X
k i=1N(i)
X
j=1
(u
(i)1j− u
(i)1T)(u
(i)1j− u
(i)1T)
′,
Φ b
21=
X
ki=1 N1(i)
X
j=1
y
(i)2jy
′(i)1j
X
ki=1 N1(i)
X
j=1
y
(i)1jy
′(i)1j
−1
,
Φ b
22= 1 N
1
X
ki=1 N1(i)
X
j=1
y
(i)2jy
′(i)2j−
X
ki=1 N1(i)
X
j=1
y
(i)2jy
′(i)1j
X
ki=1 N1(i)
X
j=1
y
(i)1jy
′(i)1j
−1
X
ki=1 N1(i)
X
j=1
y
(i)1jy
′(i)2j
,
y
(i)1j= u
(i)1j− u
(i)1F, y
(i)2j= u
(i)2j− u
(i)2F.
Similarly, the MLEs of θ and Γ under H
Pkare as follows:
eθ = eθ
1eθ
2!
=
u
1Tu
2F− e Φ
21(u
1F− u
1T)
,
Γ = e e Γ
11Γ e
12e Γ
21Γ e
22!
= Φ e
11Φ e
11Φ e
12Φ e
21Φ e
11Φ e
22+ e Φ
21Φ e
11Φ e
12! ,
where
u
1T= 1 N
X
k i=1N(i)
X
j=1
u
(i)1j, u
1F= 1 N
1X
k i=1N1(i)
X
j=1
u
(i)1j, u
2F= 1 N
1X
k i=1N1(i)
X
j=1
u
(i)2j,
and
Φ e
11= 1 N
X
k i=1N
X
(i)j=1
(u
(i)1j− u
1T)(u
(i)1j− u
1T)
′,
Φ e
21=
X
k i=1N(i)
X
j=1
w
(i)2jw
′(i)1j
X
ki=1 N(i)
X
j=1
w
(i)1jw
′(i)1j
−1
,
Φ e
22= 1 N
1X
k i=1
N
X
(i)j=1
w
(i)2jw
′(i)2j−
N(i)
X
j=1
w
(i)2jw
′(i)1j
N
X
(i)j=1
w
(i)1jw
′(i)1j
−1
N
X
(i)j=1