Abstract
In this study, we adapt sufficient bootstrap into the jackknife-after-bootstrap (JaB) algorithm. The performances of the sufficient and conventional JaB methods have been compared for detecting influential observations in linear regression. Comparison is based on two real-world examples and an extensive designed simulation study. Design includes different sample sizes and various modeling scenarios. The results reveal that proposed method is a good competitor for conventional JaB method with less standard error and amount of computation.
Similar content being viewed by others
References
Andrews DF, Pregibon D (1978) Finding the outliers that matter. J R Stat Soc B 40:85–93
Bae W, Hwang S, Kim C (2008) Influence diagnostics in the varying coefficient model with longitudinal data. Comput Stat 23:185–196
Belsley DA, Kuh E, Welsch RE (1980) Regression diagnostics. Wiley, New York
Beyaztas U, Alin A (2013) Jackknife-after-bootstrap method for detection of influential observations in linear regression models. Commun Stat Simul 42:1256–1267
Chatterjee S, Hadi AS (1986) Influential observations, high leverage points, and outliers in linear regression. Stat Sci 1:379–416
Cook RD (1977) Detection of influential observations in linear regression. Technometrics 19:15–18
Cook RD (1979) Influential observations in linear regression. J Am Stat Assoc 74:169–174
Cook RD, Weisberg S (1980) Characterization of an empirical influence function for detecting influential cases in regression. Technometrics 22:495–508
Cook RD, Weisberg S (1982) Residuals and influence in regression. Chapman & Hall, New York
Dale JR (1986) Asymptotic normality goodness of fit statistics for sparse product multinomials. J R Stat Soc B 41:48–59
Efron B (1992) Jackknife-after-bootstrap standard errors and influence functions. J R Stat Soc 54:83–127
Hossain A, Naik DN (1991) A comparative study on detection of influential observations in linear regression. Stat Pap 32:55–69
Lachos WH, Vilca F, Gale M (2007) Influence diagnostics for Grubbs’s model. Stat Pap 48:419–436
Martin MA, Roberts S (2010) Jackknife-after-bootstrap regression influence diagnostics. J Nonparametr Stat 22:257–269
Rousseeuw PJ, Leroy AM (1987) Robust regression and outlier detection. Wiley, New York
Singh S, Sedory SA (2011) Sufficient bootstrapping. Comput Stat Data Anal 55:1629–1637
Acknowledgments
The authors are grateful to Michael Martin and an anonymous referee for their comments which improved the original version of the paper significantly. The authors also thank to “R Development Core Team (2009). R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. ISBN 3-900051-07-0, URL: http://www.R-project.org”, for using R coding in the simulation study.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Beyaztas, U., Alin, A. Sufficient jackknife-after-bootstrap method for detection of influential observations in linear regression models. Stat Papers 55, 1001–1018 (2014). https://doi.org/10.1007/s00362-013-0548-4
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00362-013-0548-4