Abstract
Conducting an EEG-based neuroergonomics experiment can be a daunting task for novice researchers. This chapter provides an overview of three aspects of EEG research which we hope will help novice researchers efficiently produce meaningful and replicable results: power analysis, data preprocessing, and reporting. We explain why power analysis and sample size estimation are critical yet often overlooked aspects of experimental research and describe the most common measures of effect size likely to be encountered, Cohen’s d and eta-squared. We also provide a list of powerful (and free) power analysis tools to facilitate the actual calculations. We also provide step-by-step instructions for data preprocessing with EEGLAB which can be used in preparation for subsequent ERP or connectivity analyses. This includes filtering, artifact removal and correction, independent component analysis, and source localization. Finally, we condense EEG reporting guidelines into a checklist which can be used to ensure that your manuscript draft follows best practices.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Acunzo, D. J., MacKenzie, G., & van Rossum, M. C. W. (2012). Systematic biases in early ERP and ERF components as a result of high-pass filtering. Journal of Neuroscience Methods, 209(1), 212–218. https://doi.org/10.1016/j.jneumeth.2012.06.011.
Bezeau, S., & Graves, R. (2003). Statistical power and effect sizes of clinical neuropsychology research. Journal of Clinical and Experimental Neuropsychology, 23(3), 399–406. https://doi.org/10.1076/jcen.23.3.399.1181.
Bigdely-Shamlo, N., Mullen, T., Kothe, C., Su, K.-M., & Robbins, K. A. (2015). The PREP pipeline: Standardized preprocessing for large-scale EEG analysis. Frontiers in Neuroinformatics, 9, 1–20. https://doi.org/10.3389/fninf.2015.00016.
Cohen, B. H., & Lea, R. B. (2004). Essentials of statistics for the social and behavioral sciences. Hoboken: Wiley. ISBN 0-471-22031-0.
Cohen, J. (1988). Statistical power analysis for the behavioral sciences. New York: Routledge. ISBN 978-1-134-74270-7.
Cohen, J. (1992). A power primer. Psychological Bulletin, 112(1), 155–159. https://doi.org/10.1038/141613a0.
Delorme, A., Mullen, T., Kothe, C., Akalin Acar, Z., Bigdely-Shamlo, N., Vankov, A., & Makeig, S. (2011). EEGLAB, SIFT, NFT, BCILAB, and ERICA: New tools for advanced EEG processing. Computational Intelligence and Neuroscience, 2011, 1–12. https://doi.org/10.1155/2011/130714.
Dien, J. (1998). Issues in the application of the average reference: Review, critiques, and recommendations. Behavior Research Methods, Instruments, and Computers, 30(1), 34–43. https://doi.org/10.3758/BF03209414.
Faul, F., Erdfelder, E., Lang, A., & Buchner, A. (2007). G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. IEEE International Symposium on Information Theory Proceedings, 39(2), 175–191. https://doi.org/10.1109/ISIT.2013.6620417.
Guo, Y., Logan, H. L., Glueck, D. H., & Muller, K. E. (2013). Selecting a sample size for studies with repeated measures. BMC Medical Research Methodology, 13(1). https://doi.org/10.1186/1471-2288-13-100.
Ioannidis, J. (2008). Why most discovered true associations are inflated. Epidemiology, 19(6), 640–648. https://doi.org/10.1097/EDE.ObO.
Ioannidis, J. P. A. (2005). Why most published research findings are false. PLoS Medicine, 2(8), 0696–0701. https://doi.org/10.1371/journal.pmed.0020124.
Johnson, V. E., Payne, R. D., Wang, T., Asher, A., & Mandal, S. (2017). On the reproducibility of psychological science. Journal of the American Statistical Association, 112(517), 1–10. https://doi.org/10.1080/01621459.2016.1240079.
Keil, A., Debener, S., Gratton, G., Junghöfer, M., Kappenman, E. S., Luck, S. J., … Yee, C. M. (2014). Committee report: Publication guidelines and recommendations for studies using electroencephalography and magnetoencephalography. Psychophysiology, 51(1), 1–21. https://doi.org/10.1111/psyp.12147.
Kim, J., & Seo, B. S. (2013). How to calculate sample size and why. Clinics in Orthopedic Surgery, 5(3), 235–242. https://doi.org/10.4055/cios.2013.5.3.235.
Lakens, D. (2013). Calculating and reporting effect sizes to facilitate cumulative science: A practical primer for t-tests and ANOVAs. Frontiers in Psychology, 4(November), 1–12. https://doi.org/10.3389/fpsyg.2013.00863.
Larson, M. J., & Carbine, K. A. (2017). Sample size calculations in human electrophysiology (EEG and ERP) studies: A systematic review and recommendations for increased rigor. International Journal of Psychophysiology, 111, 33–41. https://doi.org/10.1016/j.ijpsycho.2016.06.015.
Lenth, R. V. (2001). Some practical guidelines for effective sample size determination. The American Statistician, 55(3), 187–193.
Luck, S. J. (2004). Ten simple rules for designing and interpreting ERP experiments. Components, 249, 17–32. Retrieved from http://www.ncbi.nlm.nih.gov/pubmed/7783201.
Luck, S. J. (2005). An introduction to the event-related potential technique. Cambridge, MA: MIT Press.
Mullen, T. R. (2014). The dynamic brain: Modeling neural dynamics and interactions from human electrophysiological recordings (Order No. 3639187). Available from Dissertations & Theses @ University of California; ProQuest Dissertations & Theses A&I. (1619637939).
Mullen, T. R., Kothe, C. A. E., Chi, Y. M., Ojeda, A., Kerth, T., Makeig, S., … Cauwenberghs, G. (2015). Real-time neuroimaging and cognitive monitoring using wearable dry EEG. IEEE Transactions on Biomedical Engineering, 62(11), 2553–2567. https://doi.org/10.1109/TBME.2015.2481482.
Olejnik, S., & Algina, J. (2003). Generalized eta and omega squared statistics: Measures of effect size for some common research designs. Psychological Methods, 8(4), 434–447. https://doi.org/10.1037/1082-989X.8.4.434.
Picton, T., Ritter, W., Ruchkin, D., Rugg, M., Taylor, M., Bentin, S., … Miller, G. (2000). Guidelines for using human event-related potentials to study cognition: Recording standards and publication criteria. Psychophysiology, 37(2), 127–152. https://doi.org/10.2113/gselements.13.1.29.
Richardson, J. T. E. (2011). Eta squared and partial eta squared as measures of effect size in educational research. Educational Research Review, 6(2), 135–147. https://doi.org/10.1016/j.edurev.2010.12.001.
Sawilowsky, S. S. (2017). New effect size rules of thumb. Journal of Modern Applied Statistical Methods, 8(2), 597–599. https://doi.org/10.22237/jmasm/1257035100.
Seth, A. K., Barrett, A. B., & Barnett, L. (2015). Granger causality analysis in neuroscience and neuroimaging. Journal of Neuroscience, 35(8), 3293–3297. https://doi.org/10.1523/jneurosci.4399-14.2015.
Widmann, A., Schröger, E., & Maess, B. (2015). Digital filter design for electrophysiological data–a practical approach. Journal of neuroscience methods, 250, 34–46.
Winkler, I., Debener, S., Muller, K. R., & Tangermann, M. (2015). On the influence of high-pass filtering on ICA-based artifact reduction in EEG-ERP. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS (pp. 4101–4105) 2015–November. https://doi.org/10.1109/EMBC.2015.7319296.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Sanders, N., Choo, S., Nam, C.S. (2020). The EEG Cookbook: A Practical Guide to Neuroergonomics Research. In: Nam, C. (eds) Neuroergonomics. Cognitive Science and Technology. Springer, Cham. https://doi.org/10.1007/978-3-030-34784-0_3
Download citation
DOI: https://doi.org/10.1007/978-3-030-34784-0_3
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-34783-3
Online ISBN: 978-3-030-34784-0
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)