The internal consistency and precedence of key process areas in the capability maturity model for software
- 200 Downloads
- 3 Citations
Abstract
Evaluating the reliability of maturity level (ML) ratings is crucial for providing confidence in the results of software process assessments. This study investigates the dimensions underlying the maturity construct in the Capability Maturity Model (CMM) for Software (SW-CMM) and estimates the internal consistency of each dimension. The results suggest that SW-CMM maturity is a three-dimensional construct, with “Project Implementation” representing the ML 2 key process areas (KPAs), “Organization Implementation” representing the ML 3 KPAs, and “Quantitative Process Implementation” representing the KPAs at MLs 4 and 5. The internal consistency for each of the three dimensions as estimated by Cronbach’s alpha exceeds the recommended value of 0.9. Based on those results, this study builds and tests a theoretical model which posits that the achievement of lower ML KPAs sustains the implementation of higher ML KPAs. Results of path analysis using partial least squares (PLS) support the theoretical model and provide detailed understanding of the process improvement path. The analysis is based on 676 CMM-Based Appraisal for Internal Process Improvement (CBA IPI) assessments.
Keywords
Cronbach’s alpha Convergent and discriminant validities Dimensionality Factor analysis Internal consistency Partial least squares SW-CMMNotes
Acknowledgements
The authors wish to acknowledge the assessors, sponsors, and others who participated in the assessments of the SW-CMM. This work would not be possible without the information that they regularly provide to the SEI. Thanks to Mike Zuccher, Kenny Smith, and Xiaobo Zhou for their support in extracting the data on which the study is based. The authors would also like to thank Sheila Rosenthal for her expert support with our bibliography, and Lauren Heinz for helping improve the readability of the document. The authors express their thanks to our SEI colleagues, Will Hayes, Mike Konrad, Keith Kost, Steve Masters, Jim McCurley, Mark Paulk, Mike Phillips, and Dave Zubrow. Thanks also to Khaled El-Emam, Robin Hunter, and Hyung-Min Park for their valuable comments on earlier drafts. Many thanks to the anonymous referees for the valuable comments and suggestions to improve the presentation of the paper. This study was supported by a Korea University Grant (2006). This support is gratefully acknowledged.
References
- Anderson JC, Gerbing DW (1991) Predicting the performance of measures in a confirmatory factor analysis with a pretest assessment of their substantive validities. J Appl Psychol 76(5):732–740CrossRefGoogle Scholar
- Bollinger T, McGowan C (1991) A critical look at software capability evaluation. IEEE Softw 8(4):25–41CrossRefGoogle Scholar
- Brodman JG, Johnson DI (1996) Return on investment from software process improvement as measured by U.S. industry. Crosstalk 9(4): 23–29. http://www.stsc.hill.af.mil/crosstalk/1996/04/index.html
- Campbell DT, Fiske DW. (1959) Convergent and discriminant validation by the multitrait–multimethod matrix. Psychol Bull 56(1):81–105CrossRefGoogle Scholar
- Carmines EG, Zeller RA (1979) Reliability and validity assessment. Sage University paper series on quantitative applications in social sciences. Sage, Newbury Park, CAGoogle Scholar
- Chin WW (1998) Issues and opinion on structural equation modeling. MIS Quarterly 22(1):vii–xviMathSciNetGoogle Scholar
- Chin WW, Newsted PR (1999) Structural equation modeling analysis with small samples using partial least squares. In: Hoyle R (ed) Statistical strategies for small sample research. Sage, Thousand Oaks, CA, pp 307–341Google Scholar
- Chin WW, Marcolin BL, Newsted P (2003) A partial least squares latent variable modeling approach for measuring interaction effects: results from a Monte Carlo simulation study and an electronic-mail emotion/adoption study. Inf Syst Res 14(2):189–217CrossRefGoogle Scholar
- Clark B (1997) The effect of software process maturity on software development effort. Ph.D. Thesis. University of Southern California, Los Angeles, CAGoogle Scholar
- Coffman A, Thompson K (1997) Air force software process improvement report. Crosstalk 10(1):25–27. http://www.stsc.hill.af.mil/crosstalk/1997/01/index.html Google Scholar
- Cohen J (1988) Statistical power analysis for the behavior sciences, 2nd edn. Erlbaum, Hillsdale, NJGoogle Scholar
- Comrey A (1973) A first course on factor analysis. Academic, LondonGoogle Scholar
- Cronbach L (1951) Coefficient alpha and the internal structure of tests. Psychometrika 16(3):297–334CrossRefGoogle Scholar
- Curtis B (1996) Factor structure of the CMM and other latent issues. Proceedings of the 1996 SEPG Conference, Atlantic City, NJ, USAGoogle Scholar
- Dunaway D, Baker M (2001) Analysis of CMM-based appraisal for internal process improvement (CBA IPI) assessment feedback. Technical report CMU/SEI-2001-TR-021. Software Engineering Institute, Carnegie Mellon University, Pittsburgh, PA. http://www.sei.cmu.edu/publications/documents/01.reports/01tr021.html
- El-Emam K (1998) The internal consistency of the ISO/IEC 15504 software process capability scale. Proceedings of the 5th International Symposium on Software Metrics, Los Alamitos, CA, USA, pp 72–81Google Scholar
- El-Emam K, Goldenson D (1995) SPICE: an empiricist’s perspective. Proceedings of the Second IEEE International Software Engineering Standards Symposium, Los Alamitos, CA, USA, pp 84–97Google Scholar
- El-Emam K, Goldenson D (2000) An empirical review of software process assessments. Adv Comput 5(3):319–423Google Scholar
- El-Emam K, Madhavji N (1995) The reliability of measuring organizational maturity. Softw Process Improv Pract 1(1):3–25Google Scholar
- El-Emam K, Simon J-M, Rousseau S, Jacquet E (1998) Cost implications of interrater agreement for software process assessment. Proceedings of the 5th International Symposium on Software Metrics, Los Alamitos, CA, USA, pp 38–51Google Scholar
- Fayad M, Laitinen M (1997) Process assessment considered wasteful. Commun ACM 40(11):125–128CrossRefGoogle Scholar
- Fornell C, Larcker DF (1981) Evaluating structural equation models with unobservable variables and measurement errors. J Mark Res 18(1):39–50CrossRefGoogle Scholar
- Fusaro P, El-Emam K, Smith B (1998) The internal consistencies of the 1987 SEI maturity questionnaire and the SPICE capability dimension. Empirical Software Engineering 3(2):179–210CrossRefGoogle Scholar
- Gefen D, Straub D (2005) A practical guide to factorial validity using pls-graph: tutorial and annotated example. Communications of the Association for Information Systems 16:91–109Google Scholar
- Goldenson D, El-Emam K (2000) What should you measure first? Lessons learned from the software CMM. Software Engineering Symposium, September 2000Google Scholar
- Gray E, Smith W (1998) On the limitations of software process assessment and the recognition of a required re-orientation for global process improvement. Softw Qual J 7(1):21–34Google Scholar
- Hattie J (1985) Methodology review: assessing unidimensionality of tests and items. Appl Psychol Meas 9:139–164CrossRefGoogle Scholar
- Herbsleb J, Zubrow D, Goldenson D, Hayes W, Paulk M (1997) Software quality and the capability maturity model. Commun ACM 40(6):30–40CrossRefGoogle Scholar
- Humphrey W, Curtis B (1991) Comments on ‘A Critical Look’. IEEE Softw 8(4):42–46CrossRefGoogle Scholar
- Igbaria M, Zinatelli N, Cragg P, Cavaye A (1997) Personal computing acceptance factors in small firms: a structural equation model. MIS Quarterly 21(3):279–302CrossRefGoogle Scholar
- ISO, ISO/IEC PDTR 15504. 1996. Information technology—software process assessment: part 1–part 9. ISO/IEC JTC1/SC7/WG10Google Scholar
- Jöreskog KG, Sörbom D (1993) LISREL 8: structural equation modeling with the SIMPLIS command language. SSI Scientific Software International, ChicagoGoogle Scholar
- Jung H-W, Hunter R (2003) Evaluating the SPICE rating scale with regard to the internal consistency of capability measures. Softw Process Improv Pract 8(3):169–178CrossRefGoogle Scholar
- Jung H-W, Hunter R, Goldenson D, El-Emam K (2001) Findings from phase 2 of the SPICE trials. Softw Process Improv Pract 6(2):205–242CrossRefGoogle Scholar
- Kotrlik J, Williams H (2003) The incorporation of effect size in information technology, learning, and performance research. Inf Technol Learn Perform J 21(1):1–7Google Scholar
- Krishnan MS, Kellner MI (1999) Measuring process consistency: implications for reducing software defects. IEEE Trans Softw Eng 25(6):800–815CrossRefGoogle Scholar
- Kuder GF, Richardson MW (1937) The theory of the estimation of test reliability. Psychometrika 2(3):151–160CrossRefGoogle Scholar
- Kwok WC, Sharp DJ (1998) A review of construct measurement issues in behavior accounting research. J Account Lit 17:137–174Google Scholar
- Likert R, Roslow S (1934) The effects upon the reliability of attitude scales of using three, five, seven alternatives. Working paper, New York University, New YorkGoogle Scholar
- Lissitz RW, Green SB (1975) Effects of the number of scale points on reliability: a Monte Carlo approach. J Appl Psychol 60(1):10–13CrossRefGoogle Scholar
- Marcoulides GA, Saunders C (2006) PLS: a silver bullet? MIS Quarterly 30(2):iii–xGoogle Scholar
- McIver JP, Carmines GE (1981) Unidimensional scaling. Sage University paper series on quantitative applications in social sciences. Sage, Newbury Park, CAGoogle Scholar
- Nunnally JC, Bernstein IH (1994) Psychometric theory, 3rd edn. McGraw-Hill, New YorkGoogle Scholar
- Paulk M, Webber C, Curtis B, Chrissis MB (1994) The capability maturity model: guidelines for improving the software process. Addison-Wesley, New YorkGoogle Scholar
- Saiedian H, Kuzara R (1995) SEI capability maturity models’ impact on contractors. Computer 28(1):16–26CrossRefGoogle Scholar
- SEI (2006) CMMI® for Development (CMMI-DEV), Version 1.2, Technical Report, CMU/SEI-2006-TR-008, Software Engineering Institute, Carnegie Mellon University, Pittsburgh, PAGoogle Scholar
- Sharma S (1996) Applied multivariate techniques. Wiley, New YorkGoogle Scholar
- Spector PF (1992) Summated rating scale construction: an introduction. Sage University paper series on quantitative applications in social sciences. Sage, Newbury Park, CAGoogle Scholar
- Trochim WM (2001) The research methods knowledge base, 2nd edn. http://automicdogpublishinh.com
- Van DeVen AH, Ferry DL (1980) Measuring and assessing organizations. Wiley, New YorkGoogle Scholar
- Werts CE, Linn RL, Jöreskog KG (1974) Intraclass reliability estimates: testing structural assumptions. Educ Psychol Meas 34:25–33CrossRefGoogle Scholar
- Wold H (1982) Soft modeling: intermediate between traditional model building and data analysis. Mathematical Statistics 6:333–346MathSciNetGoogle Scholar
- Zeller RA, Carmines EG (1980) Measurement in the social sciences: the link between theory and data. Cambridge University Press, CambridgeGoogle Scholar