Abstract
This chapter continues the discussion of errors that can be avoided without requiring an extraordinary amount of time or resources but that would, when not avoided, seriously weaken the study design. The errors discussed in this chapter are typically made when researchers want to speed up the evaluation process by spending less time on applied tasks such as collecting test data or installing comparison systems. These errors are made when researchers reuse test data during the evaluation or when they use a demo or default version of comparison systems. Both cases result in an unfair comparison of the new system with another: the first because the system has been tuned to the test data and will perform unusually well, and the second because the comparison system is used in a substandard version and will perform unusually poorly. Then, two other types of errors are discussed, one when no time is spent verifying that the randomization of subjects was successful, and lastly when carelessly constructed Likert scales are used.
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Friedman CP, Wyatt JC (2000) Evaluation methods in medical informatics. Springer-Verlag, New York
Brender J (2006) Handbook of evaluation methods for health informatics (trans: Carlander L). Elsevier Inc, San Diego
Vaughan L (2001) Statistical methods for the information professional: a practical, painless approach to understanding, using, and interpreting statistics. Commercial statistics. Information Today, Inc, New Jersey
Rosenthal R, Rosnow RL (1991) Essentials of behavioral research: methods and data analysis. McGraw-Hill, Boston
Kirk RE (1995) Experimental design: procedures for the behavioral sciences, 3rd edn. Brooks/Cole Publishing Company, Pacific Grove
Kurtz NR (1999) Statistical analysis for the social sciences. Social sciences - statistical methods. Allyn & Bacon, Needham Heights
van Doormaal JE, Mol PGM, Zaal RJ, van den Bemt PMLA, Kosterink JGW, Vermeulen KM, Haaijer-Ruskamp FM (2010) Computerized physician order entry (CPOE) system: expectations and experiences of users. J Eval Clin Pract (Online). doi:10.1111/j.1365-2753.2009.01187.x
Garamendi E, Pesudovs K, Stevens MJ, Elliott DB (2006) The refractive status and vision profile: evaluation of psychometric properties and comparison of Rasch and summated Likert-scaling. Vis Res 46:1375–1383
Dunn-Rankin P, Knezek GA (2004) Scaling methods, 2nd edn. Lawrence Erlbaum Associates, Inc., Mahwah
Carr SC, Bishop DMG (1995) Attitude assessment in non-Western countries: critical modifications to Likert-scaling. Psychologia 39:55–59
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2011 Springer-Verlag London Limited
About this chapter
Cite this chapter
Leroy, G. (2011). Other Errors to Avoid. In: Designing User Studies in Informatics. Health Informatics. Springer, London. https://doi.org/10.1007/978-0-85729-622-1_16
Download citation
DOI: https://doi.org/10.1007/978-0-85729-622-1_16
Published:
Publisher Name: Springer, London
Print ISBN: 978-0-85729-621-4
Online ISBN: 978-0-85729-622-1
eBook Packages: MedicineMedicine (R0)