Optimum Statistical Test Procedure

Published on December 2016 | Categories: Documents | Downloads: 57 | Comments: 0 | Views: 321
of 16
Download PDF   Embed   Report

Optimum Statistical Test Procedure

Comments

Content


Optimum Statistical Test Procedure
Rajesh Singh
Department of Statistics, BHU, Varanasi (U.P.), India
[email protected]
Jayant Singh
Department of Statistics
Rajasthan University, Jaipur, India
[email protected]
Florentin Smarandache
Chair of Department of Mathematics, University of New Mexico, Gallup, USA
[email protected]


Introduction
Let X be a random variable having probability distribution ), / X ( P θ Θ ∈ θ . It is
desired to test
0 0
: H Θ ∈ θ against
0 1 1
: H Θ − Θ = Θ ∈ θ . Let S denote the sample space of
outcomes of an experiment and ) x , , x , x ( x
n 2 1
− − − = denote an arbitrary element of S. A test
procedure consists in diving the sample space into two regions W and S – W and deciding to
reject H
0
if the observed x W ∈ . The region W is called the critical region. The function
θ
= θ γ P ) ( ( x W ∈ ) = ) W ( P
θ
, say, is called the power function of the test.
We consider first the case where
0
Θ consists of a single element,
0
θ

and its
complement
1
Θ

also has a single element
1
θ . We want to test the simple hypothesis
H
0
:
0
θ = θ against the simple alternative hypothesis H
1
: .
1
θ = θ
Let L
0
= L(X/H
0
) and L
1
= L(X/H
1
) be the likelihood functions under H
0
and H
1

respectively. In the Neyman – Pearson set up the problem is to determine a critical region W
such that
, dx L ) W ( P ) (
w
0 0
0
α = = = θ γ

θ an assigned value (1)
and
dx L ) W ( P ) (
w
1 1
1 ∫
= = θ γ
θ is maximum (2)

compared to all other critical regions satisfying (1).
If such a critical region exists it is called the most powerful critical region of level α.
By the Neyman-Pearson lemma the most powerful critical region W
0
for testing E
0
: 0 = 0
0

against E
1
: 0 = 0
1
is given by
} kL L : x { W
0 1 0
≥ =
provided there exists a k such that (1) is satisfied.
For this test α = θ γ ) (
0
and 1 ) (
1
→ θ γ as n α → .
But for any good test we must have 0 ) (
0
→ θ γ and 1 ) (
1
→ θ γ as n ∞ → because complete
discrimination between the hypotheses H
0
and H
1
should be possible as the sample size
becomes indefinitely large.
Thus for a good test it is required that the two error probabilities α and β should
depend on the sample size n and both should tend to zero as n ∞ → .
We describe below test procedures which are optimum in the sense that they
minimize the sum of the two error probabilities as compared to any other test. Note that
minimizing ( β + α ) is equivalent to maximising
1 - ( β + α ) = α − β − ) 1 ( = Power – Size.
Thus an optimum test maximises the difference of power and size as compared to any other
test.
Definition 1 : A critical region W
0
will be called optimum if

∫ ∫ ∫ ∫
− ≥ −
0 0
w w
0
w
1
w
0 1
dx L dx L dx L dx L (3)
for every other critical region W.
Lemma 1: For testing H
0
:
0
θ = θ against H
1
:
1
θ = θ

the region
} L L : x { W
0 1 0
≥ = is optimum.
Proof : W
0
is such that inside W
0
,
0 1
L L ≥ and outside W
0
,
0 1
L L < . Let W be any other
critical region.
∫ ∫
− − −
0
W W
0 1 0 1
dx ) L L ( dx ) L L (
=
∫ ∫
− − −
c
0
c
0
W W W W
0 1 0 1
dx ) L L ( dx ) L L (
I I
,
by subtracting the integrals over the common region W
0
I W.
≥ 0
since the integrand of first integral is positive and the integrand under second integral is
negative.
Hence (3) is satisfied and W
0
is an optimum critical region.
Example 1 :
Consider a normal population ) , ( N
2
σ θ where
2
σ is known.
It is desired to test H
0
:
0
θ = θ against H
1
:
1
θ = θ

,
0 1
θ > θ .
L( θ x )=









π σ
= σ
θ −

n
1 i
2
2
i
2
) x (
n
e
2
1

2
2
0 i
2
2
1 i
2
) x (
2
) x (
0
1
e
e
L
L
σ
θ −

σ
θ −



=
The optimum test rejects H
o

if 1
L
L
0
1

i.e. if 0
L
L
log
0
1


i.e. if 0
2
) x (
2
) x (
2
2
0 i
2
2
1 i

σ
θ −
+
σ
θ −

∑ ∑

i.e. if
∑ ∑
≥ θ + μ − θ − θ 0 n x 2 n x 2
2
0 i 0
2
1 i 1

i.e. if ( ) ( )

θ − θ ≥ θ − θ
2
0
2
1 i 0 1
n x 2 2
i.e. if
2 n
x
0 1
i
θ + θ



i.e. if
2
x
0 1
θ + θ

Thus the optimum test rejects H
0
if
2
x
0 1
θ + θ

We have






⎡ θ + θ
≥ = α
2
x P
0 1
H
0

= ( )








σ
θ − θ

σ
θ −
2
n
n /
x
H
0 1 0
0
P
Under H
o
,






σ
θ −
n
x
0
follows N(0,1) distribution.
·
( )








σ
θ − θ
Φ − = α
2
n
1
0 1

where Φ is the c.d.f. of a N(0,1) distribution.






⎡ θ + θ
< = β
2
x P
0 1
H
1
=
1
H
P
( )










σ
θ − θ −
<
σ
θ −
2
n
n
x
0 1 1

Under H
1
,






σ
θ −
n
x
1
follows N(0,1) distribution.

( )








σ
θ − θ
Φ − = β
2
n
1
0 1

Here β = α .
It can be seen that 0 → β = α as ∞ → n .
Example 2 : For testing H
0
:
0
θ = θ against H
1
:
1
θ = θ

,
0 1
θ < θ , the optimum test rejects
H
0
when
2
x
0 1
θ + θ
≤ .
Example 3 : Consider a normal distribution ( ), , N
2
σ θ θ known.
It is desired to test H
0
:
2
0
2
σ = σ against H1 :
2
0
2
1
2
1
2
, σ > σ σ = σ .
We have
( )
( )









πσ
= σ
σ
θ −

2
2
i
2
x
2
n
2
2
e
2
1
x L
( )
( )
2
0
2
i
2
1
2
i
2
x
2
x
2
n
2
1
2
0
0
1
e
e
L
L
σ
θ −

σ
θ −











σ
σ
=

( )
( ) ( )
2
0
2
i
2
1
2
i 2
0
2
1
0
1
2
x
2
x
log log
2
n
L
L
log
σ
θ −
+
σ
θ −
− σ − σ − =
∑ ∑

= ( )
( )
2
0
2
1
2
0
2
1
2
i 2
0
2
1
2
x
log log
2
n
σ σ
σ − σ
θ −
+ σ − σ −


The optimum test rejects H
0


I
1
I
0
¸ 1

i.e. if ( ) ( )
2
0
2
1
2
i
2
0
2
1
2
0
2
1
log log
2
n
x
2
σ − σ ≥ θ −
σ σ
σ − σ


i.e. if
( )
( )
2
0
2
1
2
0
2
1
2
1
2
0
2
i
log log
n
x
σ − σ
σ − σ
σ

σ
θ −


i.e. if nc
x
2
0
i









σ
θ −



where ( )
2
0
2
1
2
0
2
1
2
1
log log c σ − σ
σ − σ
σ
=
Thus the optimum test rejects H
0
if ∑[
x
i
-0
c
0
¸
2
¸ nc.
Note that under
0
i
0
x
: H
σ
θ −

follows N(0,1) distribution. Hence









σ
θ −
2
0
i
x
follows, under
H
0
, a chi-square distribution with n degrees of freedom (d.f.).
Here [ ] nc P nc
x
P
2
) n (
2
0
0 i
H
0
≥ χ =

















σ
θ −
= α


and 1 – β =

















σ
θ −

nc
x
P
2
0
i
H
1

=








σ
σ









σ
θ −

2
1
2
0
2
1
i
H
nc x
P
1

=








σ
σ
≥ χ
2
1
2
0 2
) n (
H
nc
P
1

Note that unuei B1, ∑[
x
i
-0
c
1
¸
2
follows a chi-square distribution with n d.f.
It can be seen that 0 → α and 0 → β as . n ∞ →

Example 4 : Let X follow the exponential family distributions
( ) ( )
( )
) x ( h e c x f
) x ( T Q θ
θ = θ
It is desired to test H
0
:
0
θ = θ against H
1
:
1
θ = θ


[ ]


θ = θ
θ
i
i
) x ( T ) ( Q n
) x ( h e ) ( c ) x ( L
i

The optimum test rejects H
o
when
log
I
1
I
0
¸ u
i.e. when [ ]
) ( c
) ( c
log n ) x ( T ) ( Q ) ( Q
1
0
i 0 1
θ
θ
≥ θ − θ


i.e. when
[ ] ) ( Q ) ( Q
) ( c
) ( c
log n
) x ( T
0 1
1
0
i
θ − θ
θ
θ


if 0 ) ( Q ) ( Q
0 1
> θ − θ
and rejects H
o
,
when
[ ] ) ( Q ) ( Q
) ( c
) ( c
log n
) x ( T
0 1
1
0
i
θ − θ
θ
θ


if 0 ) ( Q ) ( Q
0 1
< θ − θ
Locally Optimum Tests:
Let the random variable X have probability distribution P(x 0 ⁄ ). We are interested in
testing H
0
:
0
θ = θ against H
1
:
0
θ > θ . If W is any critical region then the power of the test
as a function of 0 is
y(0) = P
0
(w) = ] I(X 0 ⁄ )Jx
w

We want to determine a region W for which
y(0) -y(0
0
) = ] I(x 0 ⁄ )Jx -] I(x 0 ⁄ )
w w

is a maximum.
When a uniformly optimum region does not exist, there is not a single region which is
best for all alternatives. We may, however, find regions which are best for alternatives close
to the null hypothesis and hope that such regions will also do well for distant alternatives. We
shall call such regions locally optimum regions.
Let y(0) admit Taylor expansion about the point 0 = 0
0
. Then
y(0) = y(0
0
) +(0 -0
0
)ý (0
0
) + o where o -u as 0 -0
0

· y(0) -y(0
0
) = (0 - 0
0
)ý (0
0
) +o
If |0 -0
0
| is small y(0) -y(0
0
) is maximised when ý (0
0
)is maximised.
Definition2 : A region W
0
will be called a locally optimum critical region if
] I
´
(x 0
0
⁄ )Jx ¸
w
0
] I
´
(x 0
0
⁄ )Jx
w
(4)
For every other critical region W.
Lemma 2 : Let W
0
be the region |x: I
´
(x 0
0
⁄ ) ¸ I(x 0
0
⁄ )|. Then W
0
is locally optimum.
Proof: Let W
0
be the region such that inside it I
´
(x 0
0
⁄ ) ¸ I(x 0
0
⁄ ) and outside it
I
´
(x 0
0
⁄ ) < I(x 0
0
⁄ ). Let W be any other region.
] I
´
(x 0
0
⁄ )Jx -
w
0
] I
´
(x 0
0
⁄ )Jx
w

= ] I
´
(x 0
0
⁄ )Jx -
w
0rW
c
] I
´
(x 0
0
⁄ )Jx
w
0
c
rw

= ] I
´
(x 0
0
⁄ )Jx +
w
0rW
c
] I
´
(x 0
0
⁄ )Jx
w
0
c
Uw
(*)
¸ ] I(x 0
0
⁄ )Jx +
w
0rW
c
] I(x 0
0
⁄ )Jx
w
0
c
Uw

since I
´
(x 0
0
⁄ ) ¸ I(x 0
0
⁄ ) inside both the regions of the integrals.
¸ 0, since I(x 0
0
⁄ ) ¸ u in all the regions.
Hence ] I
´
(x 0
0
⁄ )Jx ¸
w
0
] I
´
(x 0
0
⁄ )Jx
w
for every other region W.
To prove (*):
We have ] I(x 0
0
⁄ )Jx +
R
] I(x 0
0
⁄ )Jx = 1
R
c
for every region R.
Differentiating we have
_ I
´
(x 0
0
⁄ )Jx
R
+ _ I
´
(x 0
0
⁄ )Jx = u
R
c

_ I
´
(x 0
0
⁄ )Jx
R
= -_ I
´
(x 0
0
⁄ )Jx = u
R
c

In (*), take R
C
=w
0
c
r w and the relation is proved.
Similarly if the alternatives are H
1
:
0
θ < θ

, the locally optimum critical region is
|x: I
´
(x 0
0
⁄ ) ¸ I(x 0
0
⁄ )|.
Example 5: Consider N(0, o
2
) distribution, o
2
known.
It is desired to test H
0
:
0
θ = θ against H
1
:
0
θ > θ

I(x 0 ⁄ ) = _
1
o√2n
]
n
c
-∑(x
i
-0)
2
2c
2

logI(x 0 ⁄ ) = nlog _
1
o√2n
] -
∑(x
ì
-0)
2
2o
2

I
´
(x 0 ⁄ )
I(x 0 ⁄ )
=
ologI(x 0 ⁄ )
o0
=
∑(x
ì
-0)
o
2
=
n(x -0)
o
2

·
I
´
(x 0
0
⁄ )
I(x 0
0
⁄ )
=
n(x - 0
0
)
o
2

The locally optimum test rejects H
0
,if
n(x -0
0
)
o
2
¸ 1
i.e. x ¸ 0
0
+
c
2
n

Now,
o = P
H
0
_x ¸ 0
0
+
o
2
n
_
= P
H
0
_
x -0
0
o
√n
,
¸
o
√n
,
_
= 1 - Φ_
o
√n
,
], since under H
0
,
x -0
0
c
√n
,
follows N(0,1) distribution.
1 -[ = P
H
1
_x ¸ 0
0
+
o
2
n
,
_
= P
H
1
_
x -0
1
o
√n
,
¸ -
0
1
- 0
0
o
√n
,
+
o
√n
_
= 1 -Φ_
-(0
1
- 0
0
√n)
o
+
o
√n
_
since under H
1
,
x -0
1
c
√n
,
follows N(0,1) distribution.
Exercise : If 0
0
= 1u, 0
1
= 11, o = 2, n = 16, then o = 0.3085, 1-[ = 0.9337
Power - Size = 0.6252
If we reject H
0
when
x -0
0
c
√n
,
> 1.64, then o = 0.05, 1-[ = 0.6406
Power – Size = 0.5906
Hence Power – Size of locally optimum test is greater than Power – size of the usual test.
Locally Optimum Unbiased Test:
Let the random variable X follows the probability distribution P(x 0 ⁄ ). Suppose it is
desired to test H
0
:
0
θ = θ against H
1
:
0
θ ≠ θ . We impose the unbiasedness restriction
y(0) ¸ y(0
0
), 0 = 0
0

and y(0) - y(0
0
) is a maximum as compared to all other regions. If such a region does not
exist we impose the unbiasedness restriction ý (0
0
) = u.
Let y(0) admit Taylor expansion about the point 0 = 0
0
.
Then y(0) = y(0
0
) + (0 -0
0
)ý (0
0
) +
(0-0
0
)
2
2
( )
0
θ γ ′ ′ +p
where p - u as 0 - u.
· y(0) -y(0
0
) = (0 -0
0
)ý (0
0
) +
(0-0
0
)
2
2
( )
0
θ γ ′ ′ +p
Under the unbiasedness restriction ý (0
0
) = 0, if |0 -0
0
| is small y(0) -y(0
0
) is
maximised when ( )
0
θ γ ′ ′ is maximised.
Definition 3 : A region W
0
will be called a locally optimum unbiased region if
ý (0
0
) = ] I
´
(x 0
0
⁄ )Jx = u
w
0
(5)
and ( )
0
θ γ ′ ′ = ( ) ( )dx X L dx X L
W
0
W
0
0
∫ ∫
θ ′ ′ ≥ θ ′ ′ (6)
for all other regions W satisfying (5).
Lemma 3 : Let W
0
be the region ( ) ( ) { }
0 0
x L x L : x θ ≥ θ ′ ′
Then W
0
is locally optimum unbiased.

Proof : Let W be any other region
] I
ii
(x 0
0
⁄ )Jx - ] I
ii
(x 0
0
⁄ )Jx
w w
0

= _ I
ii
(x 0
0
⁄ )Jx + _ I
ii
(x 0
0
⁄ )Jx
w
0
c
rw w
0
rw
c

by subtracting the common area of W and W
0
.
= _ I
ii
(x 0
0
⁄ )Jx + _ I
ii
(x 0
0
⁄ )Jx,
w
0
rw
c w
0
rw
c

since I
ii
(x 0
0
⁄ ) ¸ I(x 0 ⁄ ) inside W
0
and outside W.
¸ u since I(x 0 ⁄ ) ¸ u.
Example 6:
Consider N(0, o
2
) distribution, o
2
known .
E
0
: 0 = 0
0
, E
1
: 0 = 0
0

I = _
1
o√2n
]
n
c
-
∑(x
i
-0)
2
2c
2

I"(x 0 ⁄ )
I(x 0 ⁄ )
=
n
2
(x - 0)
2
o
4
-
n
o
2

Locally optimum unbiased test rejects H
0


n
2
(x - 0
0
)
2
o
4
-
n
o
2
¸ 1
i. c.
n(x - 0
0
)
2
o
2
¸ 1 +
o
2
n

unJcr E
0,

n(x -0
0
)
2
c
2
follows _
(1)
2
distribution.
Testing Mean of a normal population when variance is unknown.
Consider N(0, o
2
) distribution, o
2
known.
For testing H
0
:
0
θ = θ against H
1
:
1
θ = θ
,
the critical function of the optimum test is
given by
¢
m
(x) = ]
1 i¡ I(x 0
1
⁄ ) ¸ I(x 0
0
⁄ )
u otbcrwisc

On simplification we get
¢
m
(x) = _
1 i¡ x ¸
0
0
+ 0
1
2
i¡ p
1
> p
0
u otbcrwisc

¢
m
(x) = _
1 if x <
θ
0

1
2
if µ
1
< µ
0
u otheiwise

Consider the case when o
2
is unknown.
For this case we propose a test which rejects H
0
when
I
`
(x 0
1
⁄ )
I
`
(x 0
0
⁄ )
¸ 1
where I
`
(x 0
ì
⁄ ), (i=0,1) is the maximum of the likelihood under H
i
obtained from I(x 0
ì
⁄ )
by replacing o
2
by its maximum likelihood estimate
o
i
2 ¯
=
1
n
∑ (x
]
-0
ì
)
2
n
]=1
; i=0,1.
Let ¢
p
(x) denote the critical function of the proposed test, then
¢
p
(x) = _
1 i¡ I
`
(x 0
1
⁄ ) ¸ I
`
(x 0
0
⁄ )
u otbcrwisc

On simplification we get
¢
p
(x) = _
1 i¡ x ¸
0
0
+0
1
2
i¡ 0
1
> 0
0
o otbcrwsic

and ¢
p
(x) = _
1 i¡ x <
0
0
+0
1
2
i¡ 0
1
< 0
0
o otbcrwsic

Thus the proposed test ¢
p
(x) is equivalent to ¢
m
(x) which is the optimum test that
minimizes the sum of two error probabilities (o + [). Thus we see that one gets the same test
which minimises the sum of the two error probabilities irrespective of whether o
2
is known
or unknown.
Acknowledgement : Authors are thankful to Prof. Jokhan Singh for his guidance in writing
this article.

References
Pandey, M. and Singh,J. (1978) : A note on testing mean of a normal population when
variance is unknown. J.I.S.A. 16, pp. 145-146.
Pradhan,M. (1968) : On testing a simple hypothesis against a simple alternative making the
sum of the probabilities of two errors minimum. J.I.S.A. 6, pp. 149-159.
Rao, C.R.(1973) : Linear Statistical Inference and its Applications. John Wiley and Sons
Singh,J. (1985): Optimum Statistical Test Procedures. Vikram Mathematical Journal Vol V,
pp. 53-56.

Sponsor Documents

Or use your account on DocShare.tips

Hide

Forgot your password?

Or register your new account on DocShare.tips

Hide

Lost your password? Please enter your email address. You will receive a link to create a new password.

Back to log-in

Close