Consider the "Bootstrap Method" (https://en.wikipedia.org/wiki/Bootstrapping_(statistics)) in Probability and Statistics.
As I understand, the Bootstrap Method is a useful procedure that can be used to estimate the "empirical distribution" of some "statistic" (e.g. mean) for some observed data. In the Bootstrap Method:
- First, we take a random sample (e.g. 70%) of the collected data and calculate the "statistic" from this random sample.
- Next, repeat this above step many times. Each time, you will have a "version" of this "statistic" corresponding to each random resample.
- Finally, rank all these "versions" from smallest to largest - by taking the "version" corresponding to the 5th percentile and the 95th percentile from this ranked list, you can effectively place "Confidence Intervals" on this "statistic".
The Bootstrap Method is said to be particularly advantageous as it allegedly "works" in many otherwise difficult circumstances where a closed-form distribution for the "statistic" of interest might not be readily known or available. Our professor demonstrated that the Bootstrap Method does in fact work, and showed us some examples with randomly simulated data where the closed-form distributions for the "statistic" of interest is known - and it is easy to compare the solutions generated from the Bootstrap Simulations and the analytical answer. But in the back of my mind, I always play "Devils Advocate" and wonder - how do I know that the Bootstrap "just happens" to work in this example, and perhaps in the next example, we might not be as lucky.
I tried asking one of my professors as to why exactly the Bootstrap Method works - but the professor replied that its because of the Law of Large Numbers (https://en.wikipedia.org/wiki/Law_of_large_numbers). While this is probably true, I was hoping to find a more "detailed reason" as to why the Bootstrap Method works. As an example, (my understanding of) the Law of Large Numbers applies in situations where you have access to the entire population and can you can resample this population an infinite number of times - whereas in situations where the Bootstrap is used, you have a (possibly imperfect) sample from the original population, and can only resample this sample of the population. This makes me a bit unsure if extending the use of the Law of Large Numbers to justify the correctness of the Bootstrap is legitimate.
I found what seems to be a very informative University Lecture on this subject (https://www.stat.cmu.edu/~larry/=sml/Boot.pdf) in which proofs are even provided - but I don't think my knowledge of mathematics is currently adequate enough to understand this proof by myself.
I was hoping that perhaps someone here might be able to walk me through a simplified version of this proof - or perhaps provide another simplified version of a similar proof which demonstrates why the Bootstrap Method "works".
Thanks!