@daixuan1996
2015-01-07T23:45:16.000000Z
字数 6726
阅读 1257
概率统计
Given a parameter of interest, such as a population mean μ or population proportion p, the objective of point estimation is to use a sample to compute a number that represents in some sense a good guess for the true value of the parameter. The resulting number is called a point estimate.
Some General Concepts of Point Estimation
Statistical inference is almost always directed toward drawing some type of conclusion about one or more parameters (population characteristics). To do so requires that an investigator obtain sample data from each of the populations under study. Conclusions can then be based on the computed values of various sample quantities.
When discussing general concepts and methods of inference, it is convenient to have a generic symbol for the parameter of interest. We will use the Greek letter θ for this purpose. The objective of point estimation is to select a single number, based on sample data, that represents a sensible value for θ.
- A point estimate(点估计值) of a parameter
θ is a single number that can be regarded as a sensible value forθ .- A point estimate is obtained by selecting a suitable statistic and computing its value from the given sample data. The selected statistic is called the point estimator(点估计量) of
θ .- In the best of all possible worlds, we could find an estimator
θ^ for whichθ^=θ always. However,θ^ is a function of the sampleX′is , so it is a random variable. For some samples,θ^ will yield a value larger thanθ , whereas for other samplesθ^ will underestimateθ .- If we write
θ^=θ+ error of estimation
Then an accurate estimator would be one resulting in small estimation errors, so that estimated values will be near the true value.- An estimator that has the properties of unbiasedness(无偏性) and minimum variance will often be accurate in this sense.
Unbiased Estimators(无偏估计量)
- A point estimator
θ^ is said to be an unbiased estimator ofθ ifE(θ^)=θ for every possible value ofθ .
Ifθ^ is not unbiased, the differenceE(θ^)−θ is called the bias(偏差) ofθ^ .- Thus,
θ^ is unbiased if its probability distribution is always "centered" at the true value of the parameter. Note that "centered" here means that the expected value, not the median, of the distribution ofθ^ is equal toθ .
- When X is a binomial rv with parameters n and p, the sample proportion
p^=X/n is an unbiased estimator of p.- When choosing among several different estimators of
θ , select one that is unbiased.- Let
X1,X2,…,Xn be a random sample from a distribution with meanμ and varianceσ2 . Then the estimator
σ^2=S2=∑(Xi−X¯¯¯)2n−1
is an unbiased estimator ofσ2 .
Estimators with Minimum Variance
- Among all estimators of θ that are unbiased, choose the one that has minimum variance. The resulting
θ^1 is called the minimum variance unbiased estimator - MVUE ofθ .
- Let
X1,X2,…,Xn be a random sample from a normal distribution with parametersμ andσ . Then the estimatorμ^=X¯¯¯ is the MVUE ofμ .
Some Complications
- For distributions other than normal, it is possible that the estimator
X¯¯¯ is not the best choice in estimating a population meanμ .- The very important moral here is that the best estimator for μ depends crucially on which distribution is being sampled. In particular.
- If the random sample comes from a normal distribution, then
X¯¯¯ is the best of the four estimators, since it is the MVUE.- If the random sample comes from a Cauchy distribution(柯西分布), then
X¯¯¯ andXe¯¯¯¯ (the average of the two extreme observations) are terrible estimators forμ , whereasX^ is quite good;X¯¯¯ is bad because it is very sensitive to outlying observations, and the heavy tails of the Cauchy distribution make a few such observation likely to appear in any sample.- If the underlying distribution is uniform, the best estimator is
X¯¯¯ ; this estimator is greatly influenced by outlying observations, but the lack of tails makes such observations impossible- The trimmed mean is best in none of these three situations, but works reasonably well in all three. That is,
X¯¯¯tr(10) does not suffer too much in any of the three situations.
Reporting a Point Estimate: The Standard Error(标准误差)
- Besides reporting the value of a point estimate, some indication of its precision should be given. The usual measure of precision is the standard error of the estimator used.
- The standard error of an estimator is its standard deviation
σθ^=V(θ^)−−−−√ .
If the standard error itself involves unknown parameters whose values can be estimated, substitution of these estimates intoσθ^ yields the estimated standard error (estimated standard deviation) of the estimator. The estimated standard error can be denoted either byσθ^ or bysθ^ .
Methods of Point Estimation
The Method of Moments(矩估计)
- Let
X1,X2,…,Xn be a random sample from a pmf or pdf f(x).(独立同分布) For k = 1, 2, 3, …, the kth population moment(k阶总体矩), or kth moment of the distribution f(x), isE(Xk) . The kth sample moment(k阶样本矩) is(1/n)∑ni=1Xki .
(用样本矩代替总体矩,有多少参数就用多少阶)- Let
X1,X2,…,Xn be a random sample from a distribution with pmf or pdff(x;θ1,…,θm) , whereθ1,…,θm are parameters whose values are unknown. Then the moment estimators(矩估计量)θ^1,…,θ^m are obtained by equating the first m sample moments (expressions of some statistics) to the corresponding first m population moments (expressions of the unknown parameters) and solving forθ1,…,θm (the solutions are the moment estimators).
Maximum Likelihood Estimation(最大似然估计)
- Let
X1,X2,…,Xn have joint pmf or pdff(x1,x2,...,xn;θ1,...,θm) where the parametersθ1,...,θm have unknown values. Whenx1,…,xn are the observed sample values and f is regarded as a function ofθ1,...,θm , it is called the likelihood function(似然函数).
The maximum likelihood estimates - mle’s(最大似然估计值)θ^1,...,θ^m are those values of theθ′is that maximize the likelihood function, so that
f(x1,x2,...,xn;θ^1,...,θ^m)≥f(x1,x2,...,xn;θ1,...,θm)
When theX′is are substituted in place of thex′is , the maximum likelihood estimators(最大似然估计量) result.
Estimating Functions of Parameters
- Let
θ^1,...,θ^m be the mle’s of the parametersθ1,...,θm . Then the mle of any functionh(θ1,...,θm) of these parameters is the functionh(θ^1,...,θ^m) of the mle’s.
Large Sample Behavior of the MLE
- Under very general conditions on the joint distribution of the sample, when the sample size n is large, the maximum likelihood estimator of any parameter θ is approximately unbiased
[E(θ^)≈θ] and has variance that is nearly as small as can be achieved by any estimator. Stated another way, the mleθ^ is approximately the MVUE ofθ .
Some Complications
Copyright © 2015 by Xuan Dai. All rights reserved.