• Nie Znaleziono Wyników

TESTING HYPOTHESES IN UNIVERSAL MODELS

N/A
N/A
Protected

Academic year: 2021

Share "TESTING HYPOTHESES IN UNIVERSAL MODELS"

Copied!
23
0
0

Pełen tekst

(1)

TESTING HYPOTHESES IN UNIVERSAL MODELS

Eva Fiˇ serov´ a

Department of Mathematical Analysis and Applied Mathematics Faculty of Science, Palack´y University

Tomkova 40, 779 00 Olomouc, Czech Republic e-mail: fiserova@inf.upol.cz

Abstract

A linear regression model, when a design matrix has not full column rank and a covariance matrix is singular, is considered. The problem of testing hypotheses on mean value parameters is studied. Conditions when a hypothesis can be tested or when need not be tested are given.

Explicit forms of test statistics based on residual sums of squares are presented.

Key words: universal linear model, unbiased estimator, tests hypotheses.

2000 Mathematics Subject Classification62J05, 62F03, 62F10.

1. Introduction

Let a linear regression model be under consideration. Generally, no assump- tions on the rank of design and covariance matrices are given. When testing linear hypotheses on mean value parameters in universal (singular) models, three typical situations can occur; either a hypothesis cannot be tested, or a hypothesis need not be tested, since it is automatically true, or a hypothesis can be tested.

The aim of the paper is to investigate possible situations which can occur when testing hypothesis in universal models and to find proper test statistics based on residual sums of squares.

Supported by the Council of Czech Government MSM 6 198 959 214.

(2)

2. Notations and auxiliary statements

Let A be an m × n matrix. Let M(A) = {Au : u ∈ R

n

} ⊂ R

m

and Ker (A) = {u : u ∈ R

n

, Au = 0} ⊂ R

n

denote the column space and the null space of the matrix A, respectively. Let W be an m × m symmetric positive semidefinite matrix such that M(A) ⊂ M(W). Then P

WA

= A(A

0

WA)

A

0

W denotes a projector on M(A) in the W-seminorm.

The symbol M

WA

means I − P

WA

. If W = I (identity matrix), symbols P

A

and M

A

are used. The W-seminorm of x, x ∈ R

m

, is given by kxk

W

= √

x

0

Wx. Symbols A

and A

+

mean the g-inverse and the Moore- Penrose inverse of the matrix A, respectively.

Let N be an n × n symmetric positive semidefinite matrix. The symbol A

m(N)

denotes the minimum N-seminorm g-inverse of the matrix A, i.e., the matrix A

m(N)

satisfies equations

(1) AA

m(N)

A = A, NA

m(N)

A = A

0



A

m(N)



0

N.

One of representation of the matrix A

m(N)

is

A

m(N)

=

 

N

A

0

(AN

A

0

)

if M(A

0

) ⊂ M(N),

(N + A

0

A)

A

0

[A(N + A

0

A)

A

0

]

otherwise.

In more detail cf. [4].

Lemma 2.1. Let M(B) ⊂ M(A) and M(B

0

) ⊂ M(C). Then (2) (A − BC

B

0

)

= A

+ A

B (C − B

0

A

B)

B

0

A

.

P roof. It is an obvious consequence of Rohde theorem (cf., e.g., [1], p. 446,

Lemma 10.1.40).

(3)

3. Universal model The universal linear model is considered in the form

(3) Y ∼ N

n



Xβ, Σ 

, β ∈ R

k

,

where Y is an n-dimensional normally distributed random vector, Xβ is the mean value of Y and Σ its covariance matrix. X is a given matrix of the type n × k and Σ is a given n × n symmetric positive semidefinite matrix.

The best linear unbiased estimator (BLUE) of the function Xβ in the universal model (3) is (cf. [4], p. 148)

(4) Xβ d = X[(X

0

)

m(Σ)

]

0

Y = XD

X

0

T

Y with the covariance matrix

Var  Xβ d 

= XD

X

0

T

ΣT

XD

X

0

= X(D

− I)X

0

, where

T = Σ + XX

0

, D = X

0

T

X.

Let a null hypothesis

H

0

: h + Hβ = 0, h ∈ M(H),

where H is a given h × k matrix and h is a given h-dimensional vector, be tested in the universal model (3) against an alternative hypothesis

H

a

: h + Hβ = ξ 6= 0.

If the hypothesis is taken into account as constraints on the parameter β,the estimator of Xβ can be determined in the following way. Let β

0

be any solution of the equation h + Hβ = 0. Then the parameter β, β ∈ {u : h + Hu = 0}, can be expressed by the help of a new parameter γ

β = β

0

+ K

H

γ, γ ∈ R

k−rank(H)

,

where K

H

is a k × [k − rank(H)] matrix such that M(K

H

) = Ker(H).

(4)

The new model without constraints is in the form

Y ∼ N

n

(Xβ

0

+ XK

H

γ, Σ ) , γ ∈ R

k−rank(H)

.

Hence the BLUE of Xβ in the universal model (3) respecting the null hypothesis is

d d

Xβ = Xβ

0

+ \ XK

H

γ

= Xβ

0

+ XK

H

h

(K

0H

X

0

)

m(Σ)

i

0

(Y − Xβ

0

).

Since

M (M

H0

) = Ker(H), M (XK

H

) = M (XM

H0

) , and

P {Y − Xβ

0

∈ M(Σ)} = 1, it holds that

XK

H

h

(K

0H

X

0

)

m(Σ)

i

0

(Y − Xβ

0

) = XM

H0

h

(M

H0

X

0

)

m(Σ)

i

0

(Y − Xβ

0

) and thus

(5) Xβ d d = Xβ

0

+ XM

H0

h

(M

H0

X

0

)

m(Σ)

i

0

(Y − Xβ

0

).

The symbol b means the estimator in the universal model (3) and bb means the estimator in the universal model (3) respecting the null hypothesis.

4. Testing linear hypotheses Here approach of χ

2

-tests based on residual sums of squares

R

20

= 

Y − d Xβ 

0

h Var 

Y − d Xβ i



Y − d Xβ 

,

(5)

R

21

=



Y − Xβ d d



0

 Var



Y − Xβ d d





Y − Xβ d d



is used (cf. [3], p. 153-157, the first and the second theorems of the least squares theory).

Lemma 4.1. Let in the universal model (3) the null hypothesis be considered.

(i) Matrices Σ

, (Σ + XM

H0

X

0

)

and (Σ + XX

0

)

can be chosen as a g-inverses of the matrix Var(Y − d Xβ).

(ii) Matrices Σ

and (Σ + XM

H0

X

0

)

can be chosen as a g-inverses of the matrix Var(Y − Xβ). d d

P roof. Obviously covariance matrices of residual vectors are Var 

Y − d Xβ



= Σ − X h

(X

0

)

m(Σ)

i

0

Σ, Var



Y − Xβ d d



= Σ − XM

H0

h

(M

H0

X

0

)

m(Σ)

i

0

Σ.

Let the matrix (Σ + XM

H0

X

0

)

be chosen. Then



Σ − X h

(X

0

)

m(Σ)

i

0

Σ



(Σ + XM

H0

X

0

)

n

Σ − Σ(X

0

)

m(Σ)

X

0

o

=



I − X h

(X

0

)

m(Σ)

i

0



Σ(Σ + XM

H0

X

0

)

Σ + XM

H0

X

0

− XM

H0

X

0



× n

I − (X

0

)

m(Σ)

X

0

o

=



I − X h

(X

0

)

m(Σ)

i

0

 Σ n

I − (X

0

)

m(Σ)

X

0

o

= Var 

Y − d Xβ  .

The other statements can be proved in an analogous way.

(6)

Theorem 4.2. Let in the universal model (3) the null hypothesis be consid- ered. Let M(H

0

) ∩ M(X

0

) = {0}. Then R

21

− R

20

= 0, i.e, the hypothesis cannot be tested by the help of the statistic R

12

− R

20

.

P roof. If M(H

0

) ∩ M(X

0

) = {0}, then M(XM

H0

) = M(X), since

rank(X) + rank(H) = rank X H

!

= rank(XM

H0

) + rank(H)

(cf. [4], p. 137). Thus Y − d Xβ = n

I − X h

(X

0

)

m(Σ)

i

0

o Y = n

I − X h

(X

0

)

m(Σ)

i

0

o

(Y − Xβ

0

)

= n

I − XM

H0

h

(M

H0

X

0

)

m(Σ)

i

0

o

(Y − Xβ

0

) = Y − Xβ d d

and with respect to Lemma 4.1 we obtain R

21

− R

20

= 0.

The last theorem implies that those rows of the matrix H, which cannot be obtained from rows of the matrix X by a linear combination, cannot be used in the hypothesis. Therefore in the following text M(H

0

) ⊂ M(X

0

) is assumed. Moreover, the assumption M(H

0

) ⊂ M(X

0

) implies that the vector function Hβ is unbiasedly estimable in the universal model (3) as

Hβ d = H h

(X

0

)

m(Σ)

i

0

Y = HD

X

0

T

Y ,

Var  Hβ d 

= HD

X

0

T

ΣT

XD

H

0

= H(D

− I)H

0

.

Before a theorem on a test of a linear hypothesis some auxiliary statements must be proved.

Lemma 4.3. Let in the universal model (3) the null hypothesis be considered.

(i) As a minimum Σ-seminorm g-inverse (X

0

)

m(Σ)

of the matrix X

0

also matrices (X

0

)

m(Σ+XM

H0X0)

and (X

0

)

m(Σ+XX0)

can be chosen.

(7)

(ii) As a minimum Σ-seminorm g-inverse (M

H0

X

0

)

m(Σ)

of the matrix M

H0

X

0

also the matrix (M

H0

X

0

)

m(Σ+XM

H0X0)

can be chosen.

P roof. (i) Both matrices (X

0

)

m(Σ+XM

H0X0)

and (X

0

)

m(Σ+XX0)

are g-inverses of the matrix X

0

. Thus it suffices to prove the symmetry of matrices

Σ(X

0

)

m(Σ+XM

H0X0)

X

0

and Σ(X

0

)

m(Σ+XX0)

X

0

. The matrix

(Σ + XM

H0

X

0

)(X

0

)

m(Σ+XM

H0X0)

X

0

is symmetric with respect to definition of the matrix (X

0

)

m(Σ+XM

H0X0)

. Since

(Σ + XM

H0

X

0

)(X

0

)

m(Σ+XM

H0X0)

X

0

= Σ(X

0

)

m(Σ+XM

H0X0)

X

0

+ XM

H0

X

0

, the matrix Σ(X

0

)

m(Σ+XM

H0X0)

X

0

is symmetric too.

Analogously for the matrix Σ(X

0

)

m(Σ+XX0)

X

0

. (ii) It can be proved in the same way as (i).

Lemma 4.4. Let in the universal model (3) the null hypothesis be under consideration. Let M(X) ⊂ M(Σ + XM

H0

X

0

) and M(H

0

) ⊂ M(X

0

).

Then one choice of the g-inverse of the matrix Var( d Hβ + h) is n

H 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

o

.

P roof. The covariance matrix of the random vector d Hβ + h is Var 

Hβ d 

= H 

X

0

(Σ + XM

H0

X

0

)

X 

X

0

(Σ + XM

H0

X

0

)

Σ

× (Σ + XM

H0

X

0

)

X 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

,

(8)

since the assumption M(X) ⊂ M(Σ + XM

H0

X

0

) implies that one version of the minimum Σ-seminorm g-inverse of the matrix X

0

is

X

0



m(Σ)

= (Σ + XM

H0

X

0

)

X 

X

0

(Σ + XM

H0

X

0

)

X 

.

The last term of the expression for Var( d Hβ) is the matrix H

0

and thus it is sufficient to prove the equality

H

0

n H 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

o

Var  Hβ d 

= H

0

. It holds that

H

0

n H 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

o

Var  Hβ d 

= H

0

n H 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

o

H 

X

0

(Σ + XM

H0

X

0

)

X 

× X

0

(Σ + XM

H0

X

0

)

Σ + XM

H0

X

0

− XM

H0

X

0



× (Σ + XM

H0

X

0

)

X 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

= H

0

n H 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

o

H 

X

0

(Σ + XM

H0

X

0

)

X 

× 

X

0

(Σ + XM

H0

X

0

)

X  

X

0

(Σ + XM

H0

X

0

)

X 

H

0

= H

0

.

The following theorem deals with a test of a linear hypothesis under the special condition M(X) ⊂ M(Σ + XM

H0

X

0

).

Theorem 4.5. Let in the universal model (3) the null hypothesis be

considered. If M(H

0

) ⊂ M(X

0

) and M(X) ⊂ M(Σ + XM

H0

X

0

), the test

statistic is

(9)

R

21

− R

02

= 

Hβ d + h 

0

n H 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

o



Hβ d + h  . The statistic R

21

− R

20

has the central chi-squared distribution when the null hypothesis is true and the noncentral chi-squared distribution when the null hypothesis is not true; the parameter of noncentrality is

δ = (Hβ

+ h)

0

 H

h

X

0

Σ + XM

H0

X

0



X

i

H

0



(Hβ

+ h)

where β

is an actual value of the parameter β. Degrees of freedom are f = rank(H).

P roof. According to Lemma 4.3 the matrix (M

H0

X

0

)

m(Σ+XM

H0X0)

is one choice of the minimum Σ-seminorm g-inverse of the matrix M

H0

X

0

. Thus using relations

h

(M

H0

X

0

)

m(Σ+XM

H0X0)

i

0

= h

M

H0

X

0

Σ + XM

H0

X

0



XM

H0

i

M

H0

X

0

Σ + XM

H0

X

0



, h

M

H0

X

0

Σ + XM

H0

X

0



XM

H0

i

= h

X

0

Σ + XM

H0

X

0



X i

− h

X

0

Σ + XM

H0

X

0



X i

H

0

×

 H h

X

0

Σ + XM

H0

X

0



X i

H

0



H h

X

0

Σ + XM

H0

X

0



X i

,

(X

0

)

m(Σ)

= (X

0

)

m(Σ+XM

H0X0)

(10)

and

M(H

0

) ⊂ M(X

0

) ⇒ H h

(X

0

)

m(Σ+XM

H0X0)

i

0

0

= Hβ

0

= −h the estimator d Xβ d given by (5) can be rewritten as

d d

Xβ = d Xβ + k, where

k = −X 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

n

H 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

o

× 

Hβ d + h  . Then

R

21

− R

20

=



Y − Xβ d d



0

Σ + XM

H0

X

0





Y − d d Xβ



− 

Y − d Xβ



0

Σ + XM

H0

X

0





Y − d Xβ



= 

Y − d Xβ − k 

0

Σ + XM

H0

X

0





Y − d Xβ − k 

− 

Y − d Xβ 

0

Σ + XM

H0

X

0





Y − d Xβ 

= −2k

0

Σ + XM

H0

X

0





Y − d Xβ 

+ k

0

Σ + XM

H0

X

0



k.

It is easy to show that

k

0

(Σ + XM

H0

X

0

)



Y − d Xβ 

= 0 and

k

0

(Σ + XM

H0

X

0

)

k

= 

Hβ d + h 

0

n

H[X

0

(Σ + XM

H0

X

0

)

X]

H

0

o



Hβ d + h 

.

(11)

Further f = rank h

Var  Hβ d i

= rank n H 

X

0

(Σ + XM

H0

X

0

)

X 

X

0

(Σ + XM

H0

X

0

)

× (Σ + XM

H0

X

0

− XM

H0

X

0

)(Σ + XM

H0

X

0

)

× X 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

o

= rank n H 

X

0

(Σ + XM

H0

X

0

)

X 

H

0

− HM

H0

H

0

o

= rank(H).

The rest of the proof is obvious.

The condition M(X) ⊂ M(Σ + XM

H0

X

0

), which enable us to utter the statement on testing linear hypotheses in the classical form, is an obstacle in a general solution of the problem. Therefore we shall investigate a situation in which we assume M(H

0

) ⊂ M(X

0

), however M(X) ⊂ M(Σ+XM

H0

X

0

), is not assumed.

Theorem 4.6. Let in the universal model (3) the null hypothesis be consid- ered. Let M(H

0

) ⊂ M(X

0

). Then the BLUE of the vector 

X

H

 β is

\ \ X H

!

β = Xβ d Hβ d

!

− X

H

!

(D

− I)H

0



H(D

− I)H

0





Hβ d + h  .

The expression  \ \

X

H



β is invariant with respect to the choice of the g-inverse.

P roof. The universal model (3) with the null hypothesis can be written in the form

Y

−h

!

∼ N

n+q

"

X H

!

β, Σ, 0 0, 0

!#

.

(12)

Then the sought estimator is

\ \ X H

!

β = X H

! h

(X

0

, H

0

)

m

Σ, 0 0, 0



i

0

Y

−h

! .

Further h

(X

0

, H

0

)

m

Σ, 0 0, 0



i

0

=

"

(X

0

, H

0

) T, XH

0

HX

0

, HH

0

!

X H

!#

(X

0

, H

0

) T, XH

0

HX

0

, HH

0

!

=

"

(X

0

, H

0

) Q, R S, U

! X H

!#

(X

0

, H

0

) Q, R S, U

! ,

where (cf. Rohde theorem, e.g., in [1], p. 446, Lemma 10.1.40) Q = T

+ T

XH

0

[H(I − D)H

0

]

HX

0

T

, R = −T

XH

0

[H(I − D)H

0

]

,

S = −[H(I − D)H

0

]

HX

0

T

, U = [H(I − D)H

0

]

.

Then the expression

(X

0

, H

0

) Q, R S, U

!

can be rewritten as

X

0

T

− (I − D)H

0

[H(I − D)H

0

]

HX

0

T

; (I − D)H

0

[H(I − D)H

0

]



(13)

and

"

(X

0

, H

0

) Q, R S, U

0

! X H

!#

= n

D + (I − D)H

0

[H(I − D)H

0

]

H (I − D) o

. Further, with respect to formula (2) we have

n

D + (I − D)H

0

[H(I − D)H

0

]

H (I − D) o

= D

−D

(I−D)H

0

h

H (I−D)H

0

+H(I−D)D

(I−D)H

0

i

H (I−D)D

and using DD

H

0

= H

0

we obtain

H (I − D)H

0

+ H(I − D)D

(I − D)H

0

= HH

0

+ HD

H

0

− 2HDD

H

0

= H(D

− I)H

0

. Now the expression for the BLUE of the vector 

X

H



β is obvious.

The statement on an arbitrary choice of a g-inverse is a consequence of the relationship

P

( Y

−b

!

∈ M

"

Σ, 0 0, 0

!

+ X

H

!

(X

0

, H

0

)

#)

= 1.

Theorem 4.7. Let in the universal model (3) the null hypothesis be consid- ered. If M(H

0

) ⊂ M(X

0

), then the test statistic is

R

12

− R

20

= 

Hβ d + h 

0



H(D

− I)H

0





Hβ d + h  ,

where H(D

− I)H

0

= Var(d Hβ + h).

(14)

The statistic R

21

− R

20

has the central chi-squared distribution when the null hypothesis is true and the noncentral chi-squared distribution when the null hypothesis is not true; the parameter of noncentrality is

δ = (Hβ

+ h)

0



H(D

− I)H

0



(Hβ

+ h) ,

where β

is an actual value of the parameter β. The degrees of freedom are rank 

H(D

− I)H

0

 .

P roof. With respect to Lemma 4.1 as a g-inverse of both matrices Var



Y − Xβ d d



and Var 

Y − d Xβ 

the matrix W

= (Σ + XM

H0

X

0

)

can be chosen and at the same time the quadratic forms



Y − Xβ d d



0

 Var



Y − Xβ d d





Y − Xβ d d



and 

Y − d Xβ 

0

h Var 

Y − d Xβ i



Y − d Xβ  are invariant with respect to the choice of the g-inverse. Let

d d

Xβ = d Xβ − Xa, where

a = (D

− I)H

0



H(D

− I)H

0





Hβ d + h  . Then



Y − Xβ d d



0

W



Y − Xβ d d



= 

Y − d Xβ



0

W



Y − d Xβ



− 2 

Y − d Xβ



0

W

Xa + a

0

X

0

W

Xa.

(15)

In the following consideration the matrix W

+

will be used instead of W

. Thus we can proceed in simpler way.

At first the validity of the equality (Y − d Xβ)

0

W

+

Xa = 0 will be proved.

Using the relationship (2), the identity

(Σ + XM

H0

X

0

)

+

= (T − XP

H0

X

0

)

+

can be rewritten as

T − XP

H0

X

0



+

= T

+

+ T

+

XP

H0

I − P

H0

X

0

T

+

XP

H0



+

P

H0

X

0

T

+

. Then we obtain



Y − d Xβ



0

W

+

Xa = Y

0

I − T

+

XD

X

0



× h

T

+

+ T

+

XP

H0

I − P

H0

X

0

T

+

XP

H0



+

P

H0

X

0

T

+

i

Xa = 0, since

I − T

+

XD

X

0



T

+

X = 0.

Further a

0

X

0

W

+

Xa

= 

Hβ d + h 

0



H(D

− I)H

0



H(D

− I)X

0

Σ + XM

H0

X

0



+

× X(D

− I)H

0



H(D

− I)H

0





Hβ d + h 

= 

Hβ d + h 

0



H(D

+

− P

H0

)H

0



+

H(D

+

− P

H0

)X

0

Σ + XM

H0

X

0



+

× X(D

+

− P

H0

)H

0



H(D

+

− P

H0

)H

0



+



Hβ d + h 

= 

Hβ d + h 

0



H(D

+

− I)H

0



+



Hβ d + h 

= 

Hβ d + h 

0



H(D

− I)H

0





Hβ d + h 

,

(16)

since

X

0

Σ + XM

H0

X

0



+

X

= X

0



T

+

+ T

+

XP

H0

(I − P

H0

DP

H0

)

+

P

H0

X

0

T

+

 X

= D + DP

H0

(I − P

H0

DP

H0

)

+

P

H0

D = (D

+

− P

H0

)

+

. Finally

Var 

Hβ d + h 

= HD

X

0

T

ΣT

XD

H

0

= H(D

− I)H

0

. The other statements are obvious.

Theorem 4.8. Let in the universal model (3) the null hypothesis be consid- ered. If M(H

0

) ⊂ M(X

0

M

Σ

), then the hypothesis need not be tested since in this case P{d Hβ + h = 0} = 1.

P roof. This statement is implied by the fact that Var( d Hβ) = 0. It is a consequence of the relationship

M(H

0

) ⊂ M(X

0

M

Σ

) ⇒ ∃E : H

0

= X

0

M

Σ

E, which implies

Var  Hβ d 

= Var



E

0

M

Σ

X h

(X

0

)

m(Σ)

i

0

Y



= E

0

M

Σ

X h

(X

0

)

m(Σ)

i

0

Σ(X

0

)

m(Σ)

X

0

M

Σ

E

= E

0

M

Σ

X h

(X

0

)

m(Σ)

i

0

ΣM

Σ

E = 0.

If the null hypothesis is not true, the test statistic R

21

−R

20

has the noncentral chi-squared distribution with f degrees of freedom and the parameter of noncentrality is

δ = (Hβ

+ h)

0

h Var 

Hβ d + h i

(Hβ

+ h) ,

(17)

where β

is a true value of the parameter β. The power of the test at the point ξ is

p (ξ) = P 

R

21

− R

20

≥ χ

2f

(0, 1 − α)| H

a

: Hβ + h = ξ 6= 0 .

Here χ

2f

(0, 1 − α) is (1 − α)-quantile of the central chi-squared distribution with f degrees of freedom.

The random variable R

21

− R

20

∼ χ

2f

(δ) can be approximated by (cf. [2]) χ

2f

(δ) ≈ c

2

χ

2g

(0),

where

c

2

= f + 2δ

f + δ , g = (f + δ)

2

f + 2δ .

Remark 4.9. It is to be pointed out that in practical computing it is nec- essary to be very careful since in some situations some derived formulae can be numerically unstable. For example, small numbers on the main diagonal of the matrix Σ can caused the covariance matrix H(D

− I)H

0

numerically unstable. In practice it is useful to compute with both expressions

Var 

Hβ d + h 

= H(D

− I)H

0

,

Var 

Hβ d + h 

= HD

X

0

T

ΣT

XD

H

0

and to compare obtained results. If they are different, one can use for example the substitution Σ → kΣ, X → √

kX, where k > 0 is a sufficiently large number, and to compute them once more. This substitution does not influence the result of the original covariance matrix.

Another problem can occur when degrees of freedom are computed.

Here, e.g., expressions rank[(Σ + XX

0

)

X] or rank[(Σ + XM

H0

X

0

)

X] can be numerically unstable.

5. Example

Example 5.1. Let a linear part of high-speed lane be under consideration.

One of the safety conditions is that rails are in the line. For the sake of

simplicity let the problem be studied in plane only. An experiment for

the verification that rails are in the straight line can be done, e.g.,

(18)

in the following way. Firstly, points X

i

, i = 1, . . . , 4, are chosen elsewhere on rails. Then other points Z

1

, Z

2

, Z

3

are chosen around rails such that all distances Z

i

X

j

and Z

i

Z

k

, i, k = 1, 2, 3, i 6= k, j = 1, . . . , 4, can be observed. Finally points X

i

, Z

j

are put into proper coordinates system (the map), see Figure 1. Let each distance be measured just once. Let distances Z

i

X

j

, i, j = 1, 2, and Z

1

Z

2

have been measured in previous experiment by V¨ais¨al¨a interferometer, i.e., the accuracy of measurement is practically σ

1

= 0. (cf. [5], p. 50). Let other distances be measured by optical range- finder with the accuracy σ = 0.01 m. The problem is to test a hypothesis that all points X

i

, i = 1, . . . , 4, are located on a straight line.

0 200 400 600 800 1000

300 400 500 600 700 800 900 1000 1100 1200

x (m)

y (m)

Z1

Z2 Z3

X1

X2

X3

X4

y1 y2 y3 y4

y5 y6 y7

y8 y9

y10 y11 y12

y13 y14

y15

Figure 1. The design of the experiment

(19)

Let the notation

• y = (y

1

, . . . , y

15

)

0

. . . a vector of observed distances Z

i

X

j

, Z

i

Z

k

, i, k = 1, 2, 3, i 6= k, j = 1, . . . , 4,

• β = (β

1

, . . . , β

14

)

0

. . . a vector of unknown coordinates of points X

i

= [β

2i−1

, β

2i

]

0

, i = 1, . . . , 4, and Z

i

= [β

2i−1

, β

2i

]

0

, i = 5, 6, 7, be used. The mentioned process of measurement can be modelled by

Y ∼ N

15

(f (β), Σ) , where, e.g.,

f

1

(β) = p

9

− β

1

)

2

+ (β

10

− β

2

)

2

and Σ is a diagonal matrix given by

Σ = Diag

 

 0, 0, σ

2

, σ

2

, 0, 0, σ

2

, . . . , σ

2

| {z }

, 0, σ

2

, σ

2

 

 . The linear version of the model can be written in the form

Y − f  β

(0)



∼ N

15

(X∆β, Σ) , ∆β = β − β

(0)

, where β

(0)

are approximate values of the vector β and

X = ∂f(u)

∂u

0

u=β(0)

.

Let straight line p

i

intersects points X

i

and X

i+1

, i = 1, 2, 3. Straight lines p

i

can be expressed as

p

i

: y = a

i

+ b

i

x, i = 1, 2, 3, where

a

i

= β

2i

− β

2i−1

β

2i

− β

2i+2

β

2i−1

− β

2i+1

, b

i

= β

2i

− β

2i+2

β

2i−1

− β

2i+1

.

(20)

The problem is to test the null hypothesis H

0

: p

1

= p

2

= p

3

against the alternative hypothesis

H

a

: ∃i 6= j : p

i

6= p

j

, i, j ∈ {1, 2, 3}.

Straight lines p

1

and p

2

are identical if and only if a

1

= a

2

and b

1

= b

2

, i.e., b

1

= b

2

= b ⇔ β

2

− β

4

β

1

− β

3

= β

4

− β

6

β

3

− β

5

= b and

a

1

= a

2

⇔ β

2

− β

1

b = β

4

− β

3

b ⇔ β

2

− β

4

β

1

− β

3

= b.

Analogously for p

2

, p

3

and p

1

, p

3

. Thus

p

1

= p

2

= p

3

⇔ g(β) = 0, where

g

i

(β) = (β

2i

− β

2i+2

)(β

2i+1

− β

2i+3

) − (β

2i−1

− β

2i+1

)(β

2i+2

− β

2i+4

), i = 1, 2, and

g

3

(β) = (β

2

− β

4

)(β

5

− β

7

) − (β

1

− β

3

)(β

6

− β

8

).

Linear version of the null hypothesis can be written as H

0

: H∆β = 0,

where

H = ∂g(u)

∂u

0

u=β(0)

and the alternative hypothesis as

H

a

: H∆β = ξ 6= 0.

(21)

Let approximate values β

(0)

have been chosen as (in meters):

Z

1(0)

= 100 400

!

, Z

2(0)

= 800 650

!

, Z

3(0)

= 250 1100

! ,

X

1(0)

= 200 660

!

, X

2(0)

= 320 696

!

, X

3(0)

= 400 720

!

, X

4(0)

= 510 753

! .

In this case, terms in linearized model are the following one:

f β(0)

= [ 278.56777, 368.80347, 438.63424, 541.02588, 600.08333, 482.19913, 406.07881, 307.74827, 442.83180, 410.01951, 408.53396, 433.60005, 743.30344, 715.89105, 710.63352 ]0,

the design matrix X = (X

1

, X

2

), where

X

1

=

 

 

 

 

 

 

 

 

 

 

 

 

 

 

5.99149, 15.57786, 0, 0, 0, 0, 0

0, 0, 11.45579, 15.41325, 0, 0, 0

0, 0, 0, 0, 14.32419, 15.27913, 0

0, 0, 0, 0, 0, 0, 17.62686

−24.49320, 0.40822, 0, 0, 0, 0, 0

0, 0,−21.85889, 2.09481, 0, 0, 0

0, 0, 0, 0,−19.84974, 3.47370, 0

0, 0, 0, 0, 0, 0,−16.53104

−2.37602,−20.90900, 0, 0, 0, 0, 0

0, 0, 3.45697,−19.95166, 0, 0, 0

0, 0, 0, 0, 7.42125,−18.80050, 0

0, 0, 0, 0, 0, 0, 12.48615

0, 0, 0, 0, 0, 0, 0

0, 0, 0, 0, 0, 0, 0

0, 0, 0, 0, 0, 0, 0

 

 

 

 

 

 

 

 

 

 

 

 

 

 

,

(22)

X

2

=

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

0, −5.99149, −15.57786, 0, 0, 0, 0

0, −11.45579, −15.41325, 0, 0, 0, 0

0, −14.32419, −15.27913, 0, 0, 0, 0

15.17629, −17.62686, −15.17629, 0, 0, 0, 0

0, 0, 0, 24.49320, −0.40822, 0, 0

0, 0, 0, 21.85889, −2.09481, 0, 0

0, 0, 0, 19.84974, −3.47370, 0, 0

5.87137, 0, 0, 16.53104, −5.87137, 0, 0

0, 0, 0, 0, 0, 2.37602, 20.90900

0, 0, 0, 0, 0, −3.45697, 19.95166

0, 0, 0, 0, 0, −7.42125, 18.80050

−16.66421, 0, 0, 0, 0, −12.48615, 16.66421

0, −25.67527, −9.16974, 25.67527, 9.16974, 0, 0 0, −5.60619, −26.16222, 0, 0, 5.60619, 26.16222 0, 0, 0, 20.63193, −16.88067, −20.63193, 16.88067

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

and the matrix H = (H

1

, 0

3×6

), where

H

1

=

 

24, −80, −60, 200, 36, −120, 0, 0

0, 0, 33, −110, −57, 190, 24, −80 33, −110, −33, 110, −36, 120, 36, −120

  .

Let the simulated data of observed distances be (in meters):

y= [ 278.56778, 368.80346, 438.63423, 541.02588, 600.07120, 482.18593, 406.08812, 307.74839, 442.82535, 410.02757, 408.53628, 433.59015, 743.31683, 715.89395, 710.64831 ]0.

Cytaty

Powiązane dokumenty

(b) What is the size of a test performed by a different researcher, who tests the same hypotheses but rejecting the null when all seedlings sprout?. (c) A third researcher tests

Verify if this sample allows to assume that the mean price for this type of vehicle is equal to 3 (thousand USD), or rather more, for a 0.05 significance level.. A producer of

In a group of 180 random workers of the first factory, the mean absence in work during the analyzed year was equal to 16 days, with a standard deviation of 62. In a group of 200

by taking for them the likelihood which would have produced as posterior, given a flat prior, the PDF one believes the nuisance is distributed from. This differs from the Tevatron

Apostolskiej Mości, Franci- szkowi Józefowi Pierwszemu, z Bożej Łaski Cesarzowi Austrii itd., i po- tomkom z Jego krwi i rodu, że niezwłocznie dochowasz zasadniczych ustaw

This study simulated the aforementioned LWI development case to evaluate biases regarding the regression coefficients that each of the 6 missing-data handling methods led to,

If we consider the zero order image in an off- plane mounting (object out of the XZ plane) with a spherical grating the choice of the pupil coordinates normal and parallel to