Comments about Efron and Gong's 1983 article in The American Statistician


I'm not providing a summary of the entire article here, or even the sections that I address. Rather, I'm giving some comments about certain parts of the article, where I think some comments may be helpful to you.

Note: There is a minor error in the paper --- 2 lines below expression (3) on the 1st page, the bar is ommited from the x on the right side of the equation.

Section 1

The 3rd paragraph contains the sentence "A good parametric analysis, when appropriate, can be more efficient than its nonparametric counterpart." This means that when the assumptions of the parametric analysis are completely met, that for a given sample size, a parametric estimator can outperform a nonparametric estimation method like bootstrapping or jackknifing. (Bootstrapping and jackknifing are nonparametric because they are used the same way to obtain estimates of bias and standard error no matter what the underlying distribution of the observations is --- the estimation procedure and the general formula used to obtain the estimate doesn't depend on what the underlying distribution is.) In some cases, for example, estimating the standard error of the sample mean, bootstrapping and jackknifing do just about as good as a parametric procedure --- the jackknife estimate of the standard error of the sample mean is exactly the same as the usual parametric estimate, and the bootstrap estimate differs only slightly. But in other cases, a good parametric procedure can greatly outperform bootstrapping and jackknifing. For an example see the right portion of Table 2 in the paper --- the parametric procedure has a much smaller MSE than the bootstrapping and jackknifing procedures.

The last sentence of the 3rd paragraph suggests that bootstrapping and jackknifing have some strengths too! They can yield trustworthy estimates when one cannot confidently select a good parametric model, whereas parametric procedures can do very badly if their parametric assumptions are not met.

The middle portion of the section contains material similar to what I covered in class.

The last portion of the section pertains to cross validation. Don't bother too much this this now --- I'll discuss cross validation later this summer.

Section 2

The first example dealt with in the section is a bit different than most of the examples I discussed in class, since the sample is of bivariate data. But one can note that the bootstrapping method is carried out the same way as usual --- the only difference is that resampling is done from the empirical distribution estimate of a bivariate distribution.

Table 1 isn't explained as well as it should be. Here's what they did to get the values for the left side of the table. They generated some number, M, of samples of size 15 from the standard normal distribution. (For Table 2 they indicate that M = 200, but I don't see where they give the value of M for Table 1.) From each of the M samples, the bootstrap and jackknife estimates of the standard error of the 25% trimmed mean were computed. The sample mean of the M standard error estimates is given for each estimation procedure in the Ave column, and the sample standard deviation of the M standard error estimates is given for each estimation procedure in the Sd column. (These things tend to be a bit confusing at times. The value of .084 in the Sd column is the estimated standard deviation of the jackknife estimator of the standard error of a 25% trimmed mean. That is, .084 is the estimated standard error of an estimator of standard error.) At the bottom of the Ave column is the true standard error of the 25% trimmed mean --- the estimand for the bootstrap and jackknife estimation procedures. The value of .287 is an estimate of the expected value of the bootstrap estimator. (We can refer to the bootstrap procedure as an estimator even though we can't write the estimator as a closed form expression involving the observations.) Since the estimand is .286, it appears that the bootstrap estimator is unbiased or nearly unbiased (and I'll guess it's not exactly unbiased). Table 1 suggests that the bootstrap estimator of standard error is superior to the jackknife estimator in this setting, since it appears to be less biased and also have a smaller standard error. (The larger the value of M is, the more seriously we can take the Monte Carlo results.) The better performance of the bootstrap estimator isn't surprising in this setting since the 25% trimmed mean is not a smooth estimator. It would have been better to have also had a column giving the estimated MSE (or RMSE (the square root of the MSE)) for the two estimators, as was done in Table 2. (To estimate the MSE of the bootstrap estimator using the M Monte Carlo trials, one would just use the average of the M squared deviations of the M estimates from the known estimand.)

It can be seen from Table 2 that once again the bootstrap outperformed the jackknife when estimating the standard error. (The comparison of performances is made easier in Table 2 since estimates of the RMSE are given and so we don't have to compute the bias using the Ave value and the true value of the estimand, and then correctly combine the bias with the standard error estimate in the Sd column to obtain an estimate of the MSE.) Again, this isn't suprising for the settings considered since the standard error estimates are for estimators which aren't linear estimators.

Some of the various bootstrap estimators considered in Table 2 are smoothed bootstrap estimators. I didn't have time to discuss the smoothed bootstrap in class, so it's nice that the article provides you with some information.

Section 3

The 1st paragraph of the section states that the jackknife "performs less well than the bootstrap in Tables 1 and 2, and in most cases investigated by the author," but to be fair to the jackknife it should be pointed out that there are some circumstances in which it can perform just about as well (or even slightly better) than the bootstrap. Indeed, the result of the Theorem given in the section implies that if an estimator is a linear estimator, so that a linear approximation of the estimator would in fact be the estimator, then there can be very little difference between the jackknife and bootstrap estimates of standard error. (My web page describes the situations in which jackknife and bootstrap estimates of bias and standard error differ little.)