Introduction
Researchers typically encounter many situations in which parametric statistical techniques are less than ideal. The t-statistic, for example, assumes that the data were sampled from a normal distribution. Of course, much real-world data follow distributions that are far from normal, and may in fact be quite skewed. Suppose a researcher is investigating data that is known to follow an exponential distribution. Clearly, it would take an extremely large sample and a great deal of manipulation (e.g., averages of averages), for the central limit theorem to apply. In many cases, there is no parametric test for the measurement of interest because the sampling distribution of that measurement may be unknown and thus there would be no tractable analytic formulas for estimating such measures, for example, the difference between two medians (Mooney and Duval 1993, p. 8).
There are a number of nonparametric statistical techniques that do not rely on distributional assumptions and...
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Cho, K. (1997). Performance assessment through bootstrap. IEEE Transactions on Pattern Analysis and Machine Intelligence, 19, 1185–1198.
Efron, B. (1981). Nonparametric estimates of standard error: The jackknife, the bootstrap and other methods. Biometrika, 68, 589–599.
Efron, B. (1982). The jackknife, the bootstrap and other resampling plans. Philadelphia: SIAM.
Efron, B., & Tibshirani, R. (1991). Statistical data analysis in the computer age. Science, 253, 390–395.
Fan, X., & Jacoby, W. G. (1995). Bootsreg: An SAS matrix language program for bootstrapping linear regression models. Educational and Psychological Measurement, 55, 764–768.
Friedman, L. W., & Friedman, H. H. (1995). Analyzing simulation output using the bootstrap method. Simulation, 64, 95–100.
Jeske, D. R. (1997). Alternative prediction intervals for pareto proportions. Journal of Quality Technology, 29, 317–326.
Jochen, V. A. (1997). Using the bootstrap method to obtain probabilistic reserves estimates from production data. Petroleum Engineer International, 70, 55 +.
Kim, Y. B., Willemain, T. R., Haddock, J., & Runger, G. C. (1993). The threshold bootstrap: A new approach to simulation output analysis. Proceedings of the 1993 Winter Simulation Conference, 498–502.
LeBaron, B. (1998). A bootstrap evaluation of the effect of data splitting on financial time series. IEEE Transactions on Neural Networks, 9, 213–220.
Mooney, C. Z., & Duval, R. D. (1993). Bootstrapping: A nonparametric approach to statistical inference. Newbury Park, CA: Sage Publications.
Seppala, T. (1995). Statistical process control via the subgroup Bootstrap. Journal of Quality Technology, 27, 139–153.
Shimshoni, Y. (1998). Classification of seismic signals by integrating ensembles of neural networks. IEEE Transactions on Signal Processing, 46, 1194–1120.
Simon, J. L. (1995). Resampling stats user’s guide. Arlington, VA: Resampling Stats.
Willemain, T. R. (1994). Bootstrap on a shoestring: Resampling using spreadsheets. The American Statistician, 48, 40–42.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer Science+Business Media New York
About this entry
Cite this entry
Friedman, L.W., Friedman, H.H. (2013). Bootstrapping: Resampling Methodology. In: Gass, S.I., Fu, M.C. (eds) Encyclopedia of Operations Research and Management Science. Springer, Boston, MA. https://doi.org/10.1007/978-1-4419-1153-7_84
Download citation
DOI: https://doi.org/10.1007/978-1-4419-1153-7_84
Published:
Publisher Name: Springer, Boston, MA
Print ISBN: 978-1-4419-1137-7
Online ISBN: 978-1-4419-1153-7
eBook Packages: Business and Economics