5.4 Bootstrap

1.1 Nonparametric Estimation

Setting: X1,,Xni.i.dP, P is any distribution. Want to do inference on some "parameter" (functional) θ(P), e.g.

Recall the empirical distribution of X1,,Xn is P^n=1ni=1nδXi (P^n(A)=#{i:XiA}n). The plug-in estimator of θ(P) is θ^=θ(P^n):

Does plug-in estimator work? Depends.
Does P^npP? Depends on what sense of convergence.
Does P^n(A)pP(A) for all A? Yes.
By Glivenko-Cantelli, supx|P^n((,x])P((,x])|p0 for XR. But

P^n is non-parametric estimator for P, e.g. empirical distribution.
Plug-in estimator θ^=θ(P^n) is called bootstrap estimator.

θ(P)=medianP[1minij|XiXj|], or θ(P)=E[max1inXi].
Say determine the dam's height that has a 99% chance higher than the storm that year.

1.2 Bootstrap CI

Rn(X,P)=θ^n(X)θ(P), Gn,P(r)=PP(θ^(X)θ(P)r). r1,r2 are α/2,1α/2 quantiles of Gn,p. Then C(X)=[θ^(X)r^2,θ^(X)r^1].
We can take r^1=Gn,P^n1(α2).

2 Double Bootstrap

Might have theory telling us, e.g., supa<b|Gn,P^n([a,b])Gn,P([a,b])|p0.
Let γn,P(α)=PP(Cn,α(X)θ(P))1α. But in finite samples might have γn,P(α)<1α (or more).
If γn,P(0,1)=0.87<0.9, double bootstrap:

  1. Estimate γn,P() with plug-in γn,P^n().
  2. Use Cn,α^(X) where γ^(α^)=1α.
    α^=γ^1(1α), γ^n,P^n()=γ^().
Bootstrap CI Algorithm (MC version)

Want to estimate r1,P=Gn,P1(α2) which is the value of r that PP(θ^(X)θ(P)r)=α2.
So calculate plug-in estimator r^1=r^1,P^n=Gn,P^n1(α2)=PP^n(θ^(X)θ(P^n)r)=α2. Gn,P^n is the distribution of θ^n(X)θ(P^n).

For b=1,,B,

  • X1b,,Xnbi.i.dP^n.
  • Rnb=θ^(Xb)θ(P^n).

Gn,P^n(r)=1Bb=1B1{Rnbr}.
r^1=Gn,P^n1(α2).
C(X)=[θ^(X)r^2,θ^(X)r^1].

2.1 Double Bootstrap CI

Double Bootstrap CI

Want to estimate γn,P(α)=P(C(X)θ(P)). So calculate γ^n,P^n(α)=PP^n(C(Xθ(Pn^))).

For α=1,,B,

  • X1b,,Xnbi.i.dP^n
  • Cn,αb=Cn,α(Xb)

γ^(α)=1Bb=1B1{Cn,αbθ(P^n)}.
α^=γ^1(1α).
C(X)=Cn,α^(X).