• Nie Znaleziono Wyników

NOTE ON UNBIASED ESTIMABILITY OF THE LARGER OF TWO MEAN VALUES

N/A
N/A
Protected

Academic year: 2021

Share "NOTE ON UNBIASED ESTIMABILITY OF THE LARGER OF TWO MEAN VALUES"

Copied!
7
0
0

Pełen tekst

(1)

L. G A J E K (Warszawa and L´ od´ z)

NOTE ON UNBIASED ESTIMABILITY OF THE LARGER OF TWO MEAN VALUES

Abstract. An unbiased estimator of the larger of two mean values is constructed provided that the number of observations is random.

1. Introduction. This paper is motivated by the result of Kumar and Sharma (1993). They gave general conditions under which there does not exist an unbiased estimator of the larger of the two components of the expectation of a two-dimensional random vector. They applied this result to show unbiased inestimability of the larger of the two location parameters in the case of uniform as well as double exponential densities.

In Section 2 we give a general scheme for constructing unbiased estima- tors with a random sample size which can be applied to several situations where there does not exist an unbiased estimator of a given parameter. For example, a somewhat similar idea was investigated earlier by Rychlik (1995) [see also Rychlik (1990)] in order to construct an unbiased estimator of the unknown density of a given random variable.

In Section 3 we consider the problem of unbiased estimability of the larger of two mean values.

2. Unbiased estimation from the sample of a random size. As- sume that T 1 , T 2 , . . . is a sequence of k-dimensional random vectors with distributions depending on some parameter θ ∈ Θ, where Θ is a subset of R k . The question is how to construct an unbiased estimator of the param- eter θ when it is known that no T i is unbiased. The following proposition characterizes the situation when randomizing the sample size provides an unbiased estimator of θ.

1991 Mathematics Subject Classification: Primary 62F10.

Key words and phrases: unbiased estimation, ordered parameters.

Research partially supported by the NRC–NAS Grant.

[239]

(2)

Proposition 2.1. Let N be a random variable with the distribution in- dependent of θ and such that P (N = n) = p n , p n > 0 for n = 1, 2, . . . and P ∞

n=1 p n = 1. Then there exists a sequence of statistics {θ n , n = 1, 2, . . .}

independent of N and such that

N = θ, for every θ ∈ Θ,

if and only if there exists an asymptotically unbiased estimator T n , n = 1, 2, . . . , of the parameters θ ∈ Θ which is independent of N and such that

X

n=1

E θ |T n − T n−1 | < ∞.

P r o o f. We first prove the “if” part. Assume that T n is an asymptoti- cally unbiased estimator of θ ∈ Θ, i.e.

n→∞ lim E θ T n = θ.

Assume that N is a random variable independent of {T 1 , T 2 , . . .}. For every n = 1, 2, . . . , define

θ n = T n − T n−1 p n

, where T 0 ≡ 0. Then

E θ |θ N | = E θ (E θ (|θ N | | N )) =

X

n=1

E θ



T N − T N −1 p N

N = n

 p n

=

X

n=1

E θ |T n − T n−1 | < ∞.

Moreover,

E θ θ N = E θ (E θ (θ N | N )) =

X

n=1

E θ

 T N − T N −1 p N

N = n

 p n

= lim

n→∞ E θ T n = θ,

which proves the “if” part. Now assume that there is a sequence of statistics {θ n , n = 1, 2, . . .} independent of N and such that

E θ θ N = θ, for every θ ∈ Θ.

For every n = 1, 2, . . . , define

T n = θ 1 p 1 + . . . + θ n p n . Since

E θ |θ N | =

X

n=1

E θ |T n − T n−1 |,

(3)

where T 0 = 0, the series on the right side converges. Moreover, {T n , n = 1, 2, . . .} is asymptotically unbiased. Indeed,

n→∞ lim E θ T n = lim

n→∞

n

X

i=1

p i E θ θ i = E θ (E θ (θ N | N )) = θ.

So in order to construct an unbiased estimator of some parameter it is enough to find an asymptotically unbiased sequence of estimators (T n , n = 1, 2, . . .) for which P ∞

n=1 E θ |T n −T n−1 | < ∞. The next section concerns finding an unbiased estimator of the greater of two expectations.

3. Unbiased estimability of the larger of two expectations. We start with two simple but useful lemmas.

Lemma 3.1. Let Z 1 , . . . , Z n be i.i.d. random variables with EZ 1 = m and σ 2 = Var Z 1 < ∞. Let ξ n = n −1 P n

i=1 Z i , ξ n + = max(ξ n , 0) and ξ n = max(−ξ n , 0). Then

Var |ξ n | = σ 2 n − 4 R

ξ n (ω) dP (ω) h R

ξ n (ω) dP (ω) + m i (1)

= σ 2 n − 4 R

ξ n + (ω) dP (ω) h R

ξ n + (ω) dP (ω) − m i . (2)

P r o o f. Observe that Eξ n 2 = σ 2 /n + m 2 and Var |ξ n | = Eξ n 2 − (E|ξ n |) 2 . Since E|ξ n | = Eξ n + 2 R ξ n (ω) dP (ω), therefore

Var |ξ n | = σ 2 /n + m 2 − h

m + 2 R

ξ n (ω) dP (ω) i 2

= σ 2 /n − 4 R

ξ n (ω) dP (ω) h

m + R

ξ n (ω) dP (ω) i

. The second equality can be proven in an analogous way.

R e m a r k 3.2. In order to evaluate R ξ n dP when m > 0 (or R ξ n + dP when m < 0), one could apply first the Cauchy inequality and next the Chebyshev inequality. Indeed, from the Cauchy inequality we get

h R

ξ n (ω) dP (ω) i 2

≤ R

n (ω)] 2 dP (ω) P (ξ n < 0).

Now, observe that P (ξ n < 0) = P  1

n

n

X

i=1

Z i − m < −m



≤ P

 1 n

n

X

i=1

Z i − m

> m

 . Hence, by the Chebyshev inequality, we obtain

P (ξ n < 0) ≤ σ 2

nm 2 .

(4)

And finally,

R ξ n (ω) dP (ω) ≤  σ 2 n + m 2

 1/2

σ m √

n .

So when m > 0, R ξ n dP decreases at least as fast as n −1/2 . The lemma below shows that the rate of decrease is in fact n −1 .

Lemma 3.3. Under the assumptions of Lemma 3.1,

R ξ n (ω) dP (ω) ≤ 1 2

p m 2 + σ 2 /n − 1

2 m if m > 0;

R ξ n + (ω) dP (ω) ≤ 1 2

p m 2 + σ 2 /n + 1

2 m if m < 0.

P r o o f. Suppose that m > 0. From the equality (1) of Lemma 3.1 it follows that x ≡ R ξ n dP satisfies the inequality

(3) σ 2

n − 4x(m + x) ≥ 0.

From (3) we get the bound

R ξ n (ω) dP (ω) ≤ 1 2

p m 2 + σ 2 /n − 1 2 m.

The case m < 0 can be treated analogously by applying (2) instead of (1).

Let X 1 , X 2 , . . . be i.i.d. random variables with a common expectation θ 1

and variance σ 2 1 . Let Y 1 , Y 2 , . . . be i.i.d. random variables with a common expectation θ 2 and variance σ 2 2 . Assume that {X i , i = 1, 2, . . .} and {Y i , i = 1, 2, . . .} are mutually independent. The problem of our interest is to estimate ϑ = max(θ 1 , θ 2 ). Let us define the following estimator of ϑ:

(4) ϑ b n = max  1

n

n

X

i=1

X i , 1 n

n

X

i=1

Y i

 .

Let ν denote the vector of parameters (θ 1 , θ 2 ). The lemma below provides evaluations of the bias and of the variance of b ϑ n .

Lemma 3.4. Under the above assumptions the following inequalities hold : (i) 0 ≤ E ν ϑ b n − ϑ ≤ 1 2 {[(σ 1 2 + σ 2 2 )/n + (θ 1 − θ 2 ) 2 ] 1/2 − |θ 1 − θ 2 |};

(ii) Var ν ϑ b n ≤ (σ 2 1 + σ 2 2 )/n.

P r o o f. (i) Observe that max(a 1 , a 2 ) = 1 2 (a 1 + a 2 ) + 1 2 |a 1 − a 2 | for any a 1 , a 2 ∈ R. Hence

(5) E ν ϑ b n − ϑ = 1 2

 E ν

1 n

n

X

i=1

(X i − Y i )

− |θ 1 − θ 2 |



.

Clearly, (5) implies the lower bound in (i).

(5)

Now we will prove the upper bound. Let ξ n = n −1 P n

i=1 (X i − Y i ).

Assume for the time being that m ≡ θ 1 − θ 2 > 0 and observe that E ν |ξ n | = R

ξ n + (ω) dP (ω) + R

ξ n (ω) dP (ω) (6)

= θ 1 − θ 2 + 2 R

ξ n (ω) dP (ω).

From (5), (6) and Lemma 3.3 the assertion follows. The case θ 1 < θ 2 can be treated analogously.

(ii) Observe that Var ν ϑ b n = 1

4 E ν

 1 n

n

X

i=1

(X i + Y i ) + 1 n

n

X

i=1

(X i − Y i ) (7)

− θ 1 − θ 2 − E ν 1 n

n

X

i=1

(X i − Y i )

 2

= 1 4

 E ν

 1 n

n

X

i=1

(X i + Y i ) − θ 1 − θ 2

 2

+ Var ν |ξ n |

+ 2E ν

 1 n

n

X

i=1

(X i + Y i ) − θ 1 − θ 2



[|ξ n | − E νn |]



≤ 1 4

 r

σ 1 2 + σ 2 2

n + p

Var ν |ξ n |

 2

,

where the last inequality follows from the Cauchy inequality. From Lem- ma 3.1 it follows that Var ν |ξ n | ≤ (σ 2 1 + σ 2 2 )/n. Applying this to (7), we get the assertion.

From Proposition 2.1 and Lemma 3.4 we get the following result.

Proposition 3.5. Let N be a random variable independent of the se- quences (X 1 , X 2 , . . .) and (Y 1 , Y 2 , . . .) and such that P (N = n) = p n , p n > 0 for all n = 1, 2, . . . and P ∞

n=1 p n = 1. Let k n be a sequence of integers such that k n % ∞ as n → ∞. Define a sequence of statistics {T 0 , T 1 , . . .} by

T 0 = 0 and T n = b ϑ k

n

for n = 1, 2, . . . , where the b ϑ k

n

are defined by (4).

(i) If P ∞ n=1 1/ √

k n < ∞, then ϑ N = (T N − T N −1 )/p N is an unbiased estimator of ϑ = max(θ 1 , θ 2 ).

(ii) If , additionally, P ∞

n=1 (k n p n ) −1 < ∞, then ϑ N has a finite variance.

P r o o f. (i) From Lemma 3.4(i), it follows that {T n , n = 1, 2, . . .} is asymptotically unbiased. Since for n > 1,

E ν |T n − T n−1 | ≤ p

Var ν T n + p

Var ν T n−1 + E ν T n − ϑ + E ν T n−1 − ϑ,

(6)

using Lemma 3.4(i), (ii), we get E ν |T n − T n−1 | ≤ 2

s

σ 2 1 + σ 2 2 k n−1

+  σ 1 2 + σ 2 2 k n−1

+ (θ 1 − θ 2 ) 2

 1/2

− |θ 1 − θ 2 |

≤ 2 s

σ 2 1 + σ 2 2 k n−1

+ σ 2 1 + σ 2 2 2|θ 1 − θ 2 |k n−1 .

From the above inequality and the assumption of (i), it follows that

X

n=1

E ν |T n − T n−1 | < ∞ and, by Proposition 2.1, ϑ N is unbiased.

(ii) Notice that (8) E ν (ϑ N ) 2 = E ν

 E ν

 T N − T N −1

p N

 2

N



=

X

n=1

E ν (T n − T n−1 ) 2 p n

. From Lemma 3.4, we get

(9) E ν (T n − T n−1 ) 2

= E ν [T n − E ν T n − (T n−1 − E ν T n−1 ) + E ν T n − E ν T n−1 ] 2

≤ Var ν T n + Var ν T n−1 + (E ν T n − E ν T n−1 ) 2 − 2 Cov ν (T n , T n−1 )

≤ ( p

Var ν T n + p

Var ν T n−1 ) 2 + [E ν (T n − T n−1 )] 2

≤ 4  σ 1 2 + σ 2 2 k n−1

+

 σ 1 2 + σ 2 2 2k n−1 |θ 1 − θ 2 |

 2  .

From (8), (9) and the assumption of (ii), it follows that E ν (ϑ N ) 2 < ∞.

R e m a r k 3.6. One can easily give examples of sequences of integers k n

satisfying the assumption of (ii) of Proposition 3.5 and such that P ∞ n=1 k n p n

< ∞, which means that the average number of observations needed to con- struct an unbiased estimator of ϑ is finite. In view of Proposition 3.5(ii) in order to have a finite variance, ϑ N needs a random number of observations which does not have a finite expectation. Therefore the following problem arises: does there exist a sequential estimator b ϑ M of ϑ = max(θ 1 , θ 2 ) based on a random number M of observations for which both E ν M and Var ν ϑ b M

are finite?

Acknowledgements. I wish to thank Tomek Rychlik very much for

providing me with two of his papers and for a helpful discussion.

(7)

References

S. K u m a r and D. S h a r m a (1993), Unbiased inestimability of the larger of two parame- ters, Statistics 24, 137–142.

T. R y c h l i k (1995), A class of unbiased kernel estimators of a probability density function, Appl. Math. (Warsaw) 22, 485–497.

T. R y c h l i k (1990), Unbiased nonparametric estimation of the derivative of the mean, Statist. Probab. Lett. 10, 329–333.

LES lAW GAJEK

INSTITUTE OF MATHEMATICS INSTITUTE OF MATHEMATICS

POLISH ACADEMY OF SCIENCES TECHNICAL UNIVERSITY OF L ´ OD´ Z

P.O. BOX 137 AL. POLITECHNIKI 11

00-950 WARSZAWA, POLAND 90-924 L ´ OD´ Z, POLAND

Received on 15.12.1994

Cytaty

Powiązane dokumenty

Mutation rates had a more substantial effect on MRSE (line colors in Fig. 5) than the maximum number of allelic states at a locus (line types in Fig. 5; Table S1), except for H MLE

Properties of order statistics (o.s.) for fixed sample size n were widely investigated, while a literature on this subject in the case when n is a value of random variable V is not

We have introduced a new parameter which is useful, together with moments vector λ M solution of (2.1) and µ CW , or µ CCW , as in (1.14), to characterize attractivity for

Trajectoi-y and kinematic parameters: Hough and Ordway model without side force and with different side force models... Trajectory and kinematic parameters: comparison among

We propose a class of unbiased and strongly consistent non- parametric kernel estimates of a probability density function, based on a random choice of the sample size and the

Random matrices, Circular Unitary Ensemble, Tensor product, Sine point process, Poisson point process.... We refine this result and investigate what happens when n becomes large with

W języku słoweńskim wskazanie na właściwy desygnat wymaga dookreślenia: „[…] spoj- rzałem na zegarek” (s. Dwie różne postawy wpisane są w oba języki. W języku

These conditions are described in terms of data of the problem and allow one to identify those problems for which linear state feedback stochastic robust control can be constructed..