[关闭]
@daixuan1996 2015-01-07T23:45:16.000000Z 字数 6726 阅读 1257

概率统计笔记 CH6

Point Estimation

点估计

概率统计

Given a parameter of interest, such as a population mean μ or population proportion p, the objective of point estimation is to use a sample to compute a number that represents in some sense a good guess for the true value of the parameter. The resulting number is called a point estimate.


  1. Some General Concepts of Point Estimation

    • Statistical inference is almost always directed toward drawing some type of conclusion about one or more parameters (population characteristics). To do so requires that an investigator obtain sample data from each of the populations under study. Conclusions can then be based on the computed values of various sample quantities.
      When discussing general concepts and methods of inference, it is convenient to have a generic symbol for the parameter of interest. We will use the Greek letter θ for this purpose. The objective of point estimation is to select a single number, based on sample data, that represents a sensible value for θ.

      • A point estimate(点估计值) of a parameter θ is a single number that can be regarded as a sensible value for θ.
      • A point estimate is obtained by selecting a suitable statistic and computing its value from the given sample data. The selected statistic is called the point estimator(点估计量) of θ.
      • In the best of all possible worlds, we could find an estimator θ^ for which θ^=θ always. However, θ^ is a function of the sample Xis, so it is a random variable. For some samples, θ^ will yield a value larger than θ , whereas for other samples θ^ will underestimate θ.
      • If we write θ^=θ+ error of estimation
        Then an accurate estimator would be one resulting in small estimation errors, so that estimated values will be near the true value.
      • An estimator that has the properties of unbiasedness(无偏性) and minimum variance will often be accurate in this sense.
    • Unbiased Estimators(无偏估计量)

      • A point estimator θ^ is said to be an unbiased estimator of θ if E(θ^)=θ for every possible value of θ.
        If θ^ is not unbiased, the difference E(θ^)θ is called the bias(偏差) of θ^.
      • Thus, θ^ is unbiased if its probability distribution is always "centered" at the true value of the parameter. Note that "centered" here means that the expected value, not the median, of the distribution of θ^ is equal to θ.
        biase
      • When X is a binomial rv with parameters n and p, the sample proportion p^=X/n is an unbiased estimator of p.
      • When choosing among several different estimators of θ, select one that is unbiased.
      • Let X1,X2,,Xn be a random sample from a distribution with mean μ and variance σ2. Then the estimator
        σ^2=S2=(XiX¯¯¯)2n1

        is an unbiased estimator of σ2.
    • Estimators with Minimum Variance

      • Among all estimators of θ that are unbiased, choose the one that has minimum variance. The resulting θ^1 is called the minimum variance unbiased estimator - MVUE of θ.
        mvue
      • Let X1,X2,,Xn be a random sample from a normal distribution with parameters μ and σ. Then the estimator μ^=X¯¯¯ is the MVUE of μ.
    • Some Complications

      • For distributions other than normal, it is possible that the estimator X¯¯¯ is not the best choice in estimating a population mean μ.
      • The very important moral here is that the best estimator for μ depends crucially on which distribution is being sampled. In particular.
        1. If the random sample comes from a normal distribution, then X¯¯¯ is the best of the four estimators, since it is the MVUE.
        2. If the random sample comes from a Cauchy distribution(柯西分布), then X¯¯¯ and Xe¯¯¯¯(the average of the two extreme observations) are terrible estimators for μ, whereas X^ is quite good; X¯¯¯ is bad because it is very sensitive to outlying observations, and the heavy tails of the Cauchy distribution make a few such observation likely to appear in any sample.
        3. If the underlying distribution is uniform, the best estimator is X¯¯¯; this estimator is greatly influenced by outlying observations, but the lack of tails makes such observations impossible
        4. The trimmed mean is best in none of these three situations, but works reasonably well in all three. That is, X¯¯¯tr(10) does not suffer too much in any of the three situations.
    • Reporting a Point Estimate: The Standard Error(标准误差)

      • Besides reporting the value of a point estimate, some indication of its precision should be given. The usual measure of precision is the standard error of the estimator used.
      • The standard error of an estimator is its standard deviation σθ^=V(θ^).
        If the standard error itself involves unknown parameters whose values can be estimated, substitution of these estimates into σθ^ yields the estimated standard error (estimated standard deviation) of the estimator. The estimated standard error can be denoted either by σθ^ or by sθ^.
  2. Methods of Point Estimation

    • The Method of Moments(矩估计)

      • Let X1,X2,,Xn be a random sample from a pmf or pdf f(x).(独立同分布) For k = 1, 2, 3, …, the kth population moment(k阶总体矩), or kth moment of the distribution f(x), is E(Xk). The kth sample moment(k阶样本矩) is (1/n)ni=1Xki.
        (用样本矩代替总体矩,有多少参数就用多少阶)
      • Let X1,X2,,Xn be a random sample from a distribution with pmf or pdf f(x;θ1,,θm), where θ1,,θm are parameters whose values are unknown. Then the moment estimators(矩估计量) θ^1,,θ^m are obtained by equating the first m sample moments (expressions of some statistics) to the corresponding first m population moments (expressions of the unknown parameters) and solving for θ1,,θm (the solutions are the moment estimators).
        moment
    • Maximum Likelihood Estimation(最大似然估计)

      • Let X1,X2,,Xn have joint pmf or pdf f(x1,x2,...,xn;θ1,...,θm) where the parameters θ1,...,θm have unknown values. When x1,,xn are the observed sample values and f is regarded as a function of θ1,...,θm, it is called the likelihood function(似然函数).
        The maximum likelihood estimates - mle’s(最大似然估计值) θ^1,...,θ^m are those values of the θis that maximize the likelihood function, so that
        f(x1,x2,...,xn;θ^1,...,θ^m)f(x1,x2,...,xn;θ1,...,θm)

        When the Xis are substituted in place of the xis, the maximum likelihood estimators(最大似然估计量) result.
        likelihood1
        likelihood2
    • Estimating Functions of Parameters

      • Let θ^1,...,θ^m be the mle’s of the parameters θ1,...,θm. Then the mle of any function h(θ1,...,θm) of these parameters is the function h(θ^1,...,θ^m) of the mle’s.
    • Large Sample Behavior of the MLE

      • Under very general conditions on the joint distribution of the sample, when the sample size n is large, the maximum likelihood estimator of any parameter θ is approximately unbiased [E(θ^)θ] and has variance that is nearly as small as can be achieved by any estimator. Stated another way, the mle θ^ is approximately the MVUE of θ.
    • Some Complications
      likelihood3
      likelihood4


Copyright © 2015 by Xuan Dai. All rights reserved.

添加新批注
在作者公开此批注前,只有你和作者可见。
回复批注