A. C Z A P K I E W I C Z and A. L. D A W I D O W I C Z (Krak´ ow)
THE TWO-DIMENSIONAL LINEAR RELATION IN THE ERRORS-IN-VARIABLES MODEL WITH REPLICATION OF ONE VARIABLE
Abstract. We present a two-dimensional linear regression model where both variables are subject to error. We discuss a model where one vari- able of each pair of observables is repeated. We suggest two methods to construct consistent estimators: the maximum likelihood method and the method which applies variance components theory. We study asymptotic properties of these estimators. We prove that the asymptotic variances of the estimators of regression slopes for both methods are comparable.
1. Introduction. A problem sometimes encountered in data analysis is to find a relation between two or more variables. In this paper we discuss the two-dimensional case, where both observables are not measured precisely.
Thus let us consider the model
(1) X i = s i + ε i , Y i = as i + b + δ i , i = 1, . . . , n,
where the disturbance errors ε i and δ i are independent random variables, with mean and variance equal to zero and σ ε 2 , σ δ 2 , respectively. We assume s i
to be an unknown constant. This case is known in the literature as a func- tional model (Kendall and Stuart 1979). It is well known (Reiersol 1950) that this model, with errors having normal distributions with unknown vari- ances, is nonidentifiable. To overcome this difficulty we need an additional assumption, for example, that the distribution of errors is nonnormal or that either one error variance is known or the ratio of the variances are known.
Another approach to construct consistent estimators of regression slopes in model (1) is repeating the random variables X i , Y i m i times (Cox 1976,
2000 Mathematics Subject Classification: Primary 62J05; Secondary 62F10, 62J12.
Key words and phrases: linear regression, consistent estimator.
[335]
Dolby 1976, Bunke and Bunke 1989). In this case we have
(2) X ij = s i + ε ij , Y ij = as i + b + δ ij , i = 1, . . . , n, j = 1, . . . , m i . In this paper we consider a particular case of the model with replications.
We will prove that repeating only one variable, for example Y i , enables us to construct consistent estimators of the unknown parameters of the linear relation.
We discuss the model
(3) X i = s i + ε i , Y ij = as i + b + δ ij , i = 1, . . . , n, j = 1, . . . , m.
The variables X i , Y ij are observables, the variables s i are unknown constants and ε i , δ ij are assumed to have independent normal distribution with mean zero and unknown variances σ ε 2 and σ 2 δ :
ε i ∼ N (0, σ 2 ε ), δ ij ∼ N (0, σ 2 δ ).
For constructing consistent estimators of the unknown parameters we present two methods: the maximum likelihood method and a method (Czap- kiewicz 1999) based on variance components theory. We compare these two methods by comparing the mean squared errors.
2. Maximum likelihood method
2.1. Methodology. We can express the observations X i , Y ij in (3) as z i = [X i , Y i1 , . . . , Y im ] 0 , i = 1, . . . , n.
The independent random vectors z i have means depending on i:
µ i = [s i , as i + b, . . . , as i + b] 0 and a common (m + 1) × (m + 1) covariance matrix:
Σ =
σ ε 2 0 . . . 0 0 σ 2 δ . . . 0
.. .
0 0 . . . σ 2 δ
.
The log-likelihood function has the form L(θ) = const − n ln σ ε − nm ln σ δ
− 1 2
n X
i=1
(X i − s i ) 2 σ 2 ε +
n
X
i=1 m
X
j=1
(Y ij − as i − b) 2 σ δ 2
where L(θ) = L(a, s 1 , . . . , s n , b, σ ε , σ δ ). Solving the log-likelihood equations is not easy. Cox (1976) gives the solutions for model (2) where both X i and Y i are repeated m times. When we assume that X ij = X i for each j in Cox’
model, we can use his solutions for our purposes.
To write estimators, set s yy =
n
X
i=1 m
X
j=1
(Y ij − Y i. ) 2 /(nm), b yy =
n
X
i=1
(Y i. − Y ) 2 /n,
b xx =
n
X
i=1
(X i − X) 2 /n, b xy =
n
X
i=1
(X i − X)(Y i. − Y )/n and
B(a) = b yy − 2ab xy + a 2 b xx .
Solving the likelihood equations we get estimators in terms of a:
(4)
b b = Y − aX,
σ b ε 2 = s yy + (b yy − ab xy ) 2 /B(a), σ b δ 2 = s xx + (ab xx − b xy ) 2 /B(a),
s b i = (ab xx − b xy )(Y i. − Y + aX) + (b yy − ab xy )X i /B(a).
But to get an estimator of a we must solve an equation of the fourth degree in a:
(5) −s yy (ab xx − b yx )B(a) − (b yy − ab xy )(ab xx − b xy )(b yy − a 2 b xx ) = 0.
When m > 2 we solve (5) numerically and then check whether the absolute maximum has been found.
2.2. Asymptotic behaviour of maximum likelihood estimators. In this section we look for the asymptotic properties of maximum likelihood esti- mators in the model discussed in the previous section. The random vectors z i are independent, with normal but not identical distribution. The expecta- tions of their distributions depend on i. The number of unknown parameters which we estimate increases with n.
Assume that s i , i = 1, . . . , n, belong to a bounded set as n tends to infinity and the following two limits exist:
n→∞ lim 1 n
n
X
i=1
s i and lim
n→∞
1 n
n
X
i=1
s 2 i . Then we can prove:
Lemma 1. When n → ∞ and m → ∞, the solutions of the likelihood equations give strongly consistent estimators of the unknown parameters a, b, σ δ , σ ε . For sufficiently large n and m, the variance of the vector (6) [ b a − a, b s 1 − s 1 , . . . , ( b s n − s n )(b b − b), b σ ε − σ ε , b σ δ − σ δ ] can be approximated by
(7)
− E
∂ 2
∂ξ∂φ L(θ)
−1
where ξ, φ belong to the set of unknown parameters.
This lemma may be proved by a method analogous to that described in Lehmann’s monograph (1983, p. 404, Th. 4.1). We thus have the following asymptotic variances of unknown regression slopes:
Theorem 1. When n and m are large, the asymptotic variances of b a and b b, avar( b a) and avar(b b), are
avar( b a) = ma 2 σ ε 2 + σ 2 δ m P n
i=1 (s i − s) 2 , (8)
avar(b b) = ma 2 σ ε 2 + σ 2 δ
mn ·
P n i=1 s 2 i P n
i=1 (s i − s) 2 . (9)
P r o o f. To show the formula for avar( b a), let us calculate ∂L/∂ξ∂φ where ξ, φ ∈ {a, s 1 , . . . , s n , b, σ ε , σ δ }. The matrix (7) has the form
Θ n −1 =
m
σ
δ2P s 2 i ma
σ
2δs 0 σ m2 δ
P s i 0 0
ma
σ
δ2s ma
2
σ
2ε+σ
δ2σ
ε2σ
2δI n am
σ
2δ1 n 0 0
m
σ
δ2P s i am
σ
δ21 0 n mn σ2 δ
0 0
0 . . . 0 2n σ2
ε
0
0 . . . 0 0 2nm σ2
δ