text stringlengths 270 6.81k |
|---|
8 < q 0 then the likelihood function of ✓ is : 1 2 (x ✓)2 for x ✓ otherwise, L(✓) = n i=1 r Y 2 ⇡ e 1 2 (xi ✓)2 , where x1 ✓, x2 ✓,..., xn ✓. This likelihood function simplifies to n L(✓) = n 2 1 2 e i=1 X (xi ✓)2, 2 ⇡ ✓. Taking the natural logarithm of L(✓) and where min{x1, x2,..., xn} maximizing, we obtain the maximum likelihood estimator of ✓ as the first order statistic of the sample X1, X2,..., Xn, that is ✓ = X(1), b Techniques for finding Interval Estimators of Parameters 498 where X(1) = min{X1, X2,..., Xn}. Suppose the true value of ✓ = 1. Using the maximum likelihood estimator of ✓, we are trying to guess this value of ✓ based on a random sample. Suppose X1 = 1.5, X2 = 1.1, X3 = 1.7, X4 = 2.1, X5 = 3.1 is a set of sample data from the above population. Then based on this random sample, we will get ✓ML = X(1) = min{1.5, 1.1, 1.7, 2.1, 3.1} = 1.1. b If we take another random sample, say X1 = 1.8, X2 = 2.1, X3 = 2.5, X4 = ✓ = 1.8 3.1, X5 = 2.6 then the maximum likelihood estimator of this ✓ will be based on this sample. The graph of the density function f (x; ✓) for ✓ = 1 is shown below. b From the graph, it is clear that a number close to 1 has higher chance of getting randomly picked by the sampling process, then the numbers that are substantially bigger than 1. Hence, it makes sense that ✓ should be estimated by the smallest sample value. However, from this example we see that the point estimate of ✓ is not equal to the true value of |
✓. Even if we take many random samples, yet the estimate of ✓ will rarely equal the actual value of the parameter. Hence, instead of finding a single value for ✓, we should report a range of probable values for the parameter ✓ with certain degree of confidence. This brings us to the notion of confidence interval of a parameter. 17.1. Interval Estimators and Confidence Intervals for Parameters The interval estimation problem can be stated as follow: Given a random ↵, find a pair of statistics U such that the sample X1, X2,..., Xn and a probability value 1 L = L(X1, X2,..., Xn) and U = U (X1, X2,..., Xn) with L Probability and Mathematical Statistics 499 probability of ✓ being on the random interval [L, U ] is 1 P (L ✓ U ) = 1 ↵. ↵. That is Recall that a sample is a portion of the population usually chosen by method of random sampling and as such it is a set of random variables X1, X2,..., Xn with the same probability density function f (x; ✓) as the population. Once the sampling is done, we get X1 = x1, X2 = x2, · · ·, Xn = xn where x1, x2,..., xn are the sample data. Definition 17.1. Let X1, X2,..., Xn be a random sample of size n from a population X with density f (x; ✓), where ✓ is an unknown parameter. The interval estimator of ✓ is a pair of statistics L = L(X1, X2,..., Xn) and U such that if x1, x2,..., xn is a set of sample U = U (X1, X2,..., Xn) with L data, then ✓ belongs to the interval [L(x1, x2,...xn), U (x1, x2,...xn)]. The interval [l, u] will be denoted as an interval estimate of ✓ whereas the random interval [L, U ] will denote the interval estimator of ✓. Notice |
that the interval estimator of ✓ is the random interval [L, U ]. Next, we define the 100(1 ↵)% confidence interval for the unknown parameter ✓. Definition 17.2. Let X1, X2,..., Xn be a random sample of size n from a population X with density f (x; ✓), where ✓ is an unknown parameter. The interval estimator of ✓ is called a 100(1 ↵)% confidence interval for ✓ if P (L ✓ U ) = 1 ↵. The random variable L is called the lower confidence limit and U is called the ↵) is called the confidence coefficient upper confidence limit. The number (1 or degree of confidence. There are several methods for constructing confidence intervals for an unknown parameter ✓. Some well known methods are: (1) Pivotal Quantity Method, (2) Maximum Likelihood Estimator (MLE) Method, (3) Bayesian Method, (4) Invariant Methods, (5) Inversion of Test Statistic Method, and (6) The Statistical or General Method. In this chapter, we only focus on the pivotal quantity method and the MLE method. We also briefly examine the the statistical or general method. The pivotal quantity method is mainly due to George Bernard and David Fraser of the University of Waterloo, and this method is perhaps one of the most elegant methods of constructing confidence intervals for unknown parameters. Techniques for finding Interval Estimators of Parameters 500 17.2. Pivotal Quantity Method In this section, we explain how the notion of pivotal quantity can be used to construct confidence interval for a unknown parameter. We will also examine how to find pivotal quantities for parameters associated with certain probability density functions. We begin with the formal definition of the pivotal quantity. Definition 17.3. Let X1, X2,..., Xn be a random sample of size n from a population X with probability density function f (x; ✓), where ✓ is an unknown parameter. A pivotal quantity Q is a |
function of X1, X2,..., Xn and ✓ whose probability distribution is independent of the parameter ✓. Notice that the pivotal quantity Q(X1, X2,..., Xn, ✓) will usually contain both the parameter ✓ and an estimator (that is, a statistic) of ✓. Now we give an example of a pivotal quantity. Example 17.1. Let X1, X2,..., Xn be a random sample from a normal population X with mean µ and a known variance 2. Find a pivotal quantity for the unknown parameter µ. Answer: Since each Xi ⇠ N (µ, 2), X ⇠ N µ, ✓ 2 n. ◆ Standardizing X, we see that X µ pn ⇠ N (0, 1). The statistics Q given by Q(X1, X2,..., Xn, µ) = µ X pn is a pivotal quantity since it is a function of X1, X2,..., Xn and µ and its probability density function is free of the parameter µ. There is no general rule for finding a pivotal quantity (or pivot) for a parameter ✓ of an arbitrarily given density function f (x; ✓). Hence to some extents, finding pivots relies on guesswork. However, if the probability density function f (x; ✓) belongs to the location-scale family, then there is a systematic way to find pivots. Probability and Mathematical Statistics 501 Definition 17.4. Let g : IR any µ and any > 0, the family of functions! IR be a probability density function. Then for F = f (x; µ, ) = 1 2 1 ), (0, ) 1 2 ✓ ⇢ ◆ is called the location-scale family with standard probability density f (x; ✓). The parameter µ is called the location parameter and the parameter is called the scale parameter. If = 1, then F is called the location family. If µ = 0, then F is called the scale family It should be noted that each member f (x; µ, ) of the location-scale, then family is a probability density function. If we take g(x) = 1 p2⇡ the normal density function 2 x |
2 e 1 f (x; µ, ) = p2⇡ 2 e 1 2 ( x µ )2 , < x < 1 1 belongs to the location-scale family. The density function f (x; ✓) = 1 ✓ e x ✓ 8 < 0 if 0 < x < 1 otherwise, belongs to the scale family. However, the density function : ✓ x✓ 1 if 0 < x < 1 f (x; ✓) = 8 < 0 otherwise, does not belong to the location-scale family. : µ µ, where It is relatively easy to find pivotal quantities for location or scale parameter when the density function of the population belongs to the location-scale family F. When the density function belongs to location family, the pivot for the location parameter µ is µ is the maximum likelihood is the maximum likelihood estimator of , then the pivot estimator of µ. If b for the scale parameter is when the density function belongs to the scale µ and the pivot for the scale parameter is when the density function belongs to location-scale famb ily. Sometime it is appropriate to make a minor modification to the pivot b obtained in this way, such as multiplying by a constant, so that the modified pivot will have a known distribution. family. The pivot for location parameter µ is b b b b µ Techniques for finding Interval Estimators of Parameters 502 Remark 17.1. Pivotal quantity can also be constructed using a sufficient statistic for the parameter. Suppose T = T (X1, X2,..., Xn) is a sufficient statistic based on a random sample X1, X2,..., Xn from a population X with probability density function f (x; ✓). Let the probability density function of T be g(t; ✓). If g(t; ✓) belongs to the location family, then an appropriate a(✓) is a pivotal quantity for the location parameter constant multiple of T ✓ for some suitable expression a(✓). If g(t; ✓) belongs to the scale family, then an appropriate constant multiple of T b(✓) is a pivotal quantity for the scale parameter ✓ for some suitable expression b( |
ile) of a standard P (Z z↵) = 1 ↵, where ↵ 0.5 (see figure below). Note that ↵ = P (Z z↵) if ↵ 0.5. 1- α α Zα A 100(1 ↵)% confidence interval for a parameter ✓ has the following interpretation. If X1 = x1, X2 = x2,..., Xn = xn is a sample of size n, then ↵)% confidence interval [l, u] based on this sample we construct a 100(1 which is a subinterval of the real line IR. Suppose we take large number of samples from the underlying population and construct all the corresponding ↵)% of these 100(1 intervals would include the unknown value of the parameter ✓. ↵)% confidence intervals, then approximately 100(1 In the next several sections, we illustrate how pivotal quantity method can be used to determine confidence intervals for various parameters. 17.3. Confidence Interval for Population Mean At the outset, we use the pivotal quantity method to construct a confidence interval for the mean of a normal population. Here we assume first the population variance is known and then variance is unknown. Next, we construct the confidence interval for the mean of a population with continuous, symmetric and unimodal probability distribution by applying the central limit theorem. Let X1, X2,..., Xn be a random sample from a population X N (µ, 2), where µ is an unknown parameter and 2 is a known parameter. First of all, we need a pivotal quantity Q(X1, X2,..., Xn, µ). To construct this pivotal ⇠ Techniques for finding Interval Estimators of Parameters 504 quantity, we find the likelihood estimator of the parameter µ. We know that N (µ, 2), the distribution of the sample mean is µ = X. Since, each Xi ⇠ given by b X N µ,. ⇠ 2 n ✓ It is easy to see that the distribution of the estim |
� U ). One can find infinitely many pairs L, U such that ↵ = P (L 1 ✓ U ). Hence, there are infinitely many confidence intervals for a given parameter. However, we only consider the confidence interval of shortest length. If a confidence interval is constructed by omitting equal tail areas then we obtain what is known as the central confidence interval. In a symmetric distribution, it can be shown that the central confidence interval is of the shortest length. Example 17.2. Let X1, X2,..., X11 be a random sample of size 11 from a normal distribution with unknown mean µ and variance 2 = 9.9. If 11 i=1 xi = 132, then what is the 95% confidence interval for µ? Answer: Since each Xi ⇠ P by N (µ, 9.9), the confidence interval for µ is given X pn z ↵ 2, X + z ↵ 2. pn Since ✓ ✓ 11 i=1 xi = 132, the sample mean x = 132 11 = 12. Also, we see that ◆ ◆ P 2 n = 9.9 11 r r = p0.9. Further, since 1 ↵ = 0.95, ↵ = 0.05. Thus z ↵ 2 = z0.025 = 1.96 (from normal table). Using these information in the expression of the confidence interval for µ, we get that is 12 h 1.96 p0.9, 12 + 1.96 p0.9 i [10.141, 13.859]. Techniques for finding Interval Estimators of Parameters 506 Example 17.3. Let X1, X2,..., X11 be a random sample of size 11 from a normal distribution with unknown mean µ and variance 2 = 9.9. If 11 i=1 xi = 132, then for what value of the constant k is P 12 h k p0.9, 12 + k p0.9 i a 90% con� |
�dence interval for µ? Answer: The 90% confidence interval for µ when the variance is given is x ✓ pn ◆ z ↵ 2, x + pn ✓ z ↵ 2. ◆ Thus we need to find x, 2 n and z ↵ 2 corresponding to 1 q x = 11 i=1 xi 11 = P 132 11 = 12. ↵ = 0.9. Hence 2 n r = 9.9 11 r = p0.9. z0.05 = 1.64 (from normal table). Hence, the confidence interval for µ at 90% confidence level is 12 (1.64) p0.9, 12 + (1.64) p0.9. i Comparing this interval with the given interval, we get h k = 1.64. and the corresponding 90% confidence interval is [10.444, 13.556]. Remark 17.3. Notice that the length of the 90% confidence interval for µ is 3.112. However, the length of the 95% confidence interval is 3.718. Thus higher the confidence level bigger is the length of the confidence interval. Hence, the confidence level is directly proportional to the length of the confidence interval. In view of this fact, we see that if the confidence level is zero, Probability and Mathematical Statistics 507 then the length is also zero. That is when the confidence level is zero, the confidence interval of µ degenerates into a point X. Until now we have considered the case when the population is normal with unknown mean µ and known variance 2. Now we consider the case when the population is non-normal but its probability density function is continuous, symmetric and unimodal. If the sample size is large, then by the central limit theorem X µ pn ⇠ N (0, 1) as n.! 1 Thus, in this case we can take the pivotal quantity to be Q(X1, X2,..., Xn, µ) = µ, X |