Skip to main content
Log in

Abstract

In this paper I provide a preliminary sketch of the types of logics of evaluation in the third sector. I begin by tracing the ideals that are evident in three well-articulated yet quite different third sector evaluation practices: the logical framework, most significant change stories, and social return on investment. Drawing on this analysis, I then tentatively outline three logics of evaluation: a scientific evaluation logic (systematic observation, observable and measurable evidence, objective and robust experimental procedures), a bureaucratic evaluation logic (complex, step-by-step procedures, analysis of intended objectives), and a learning evaluation logic (openness to change, wide range of perspectives, lay rather than professional expertise). These logics draw attention to differing conceptions of knowledge and expertise and their resource implications, and have important consequences for the professional status of the practitioners, consultants, and policy makers that contribute to and/or are involved in evaluations in third sector organizations.

Résumé

Dans cet article j’offre une esquisse préliminaire des divers types de logique d’évaluation du tiers-secteur. Je commence par décrire les idéaux que trois pratiques d’évaluation du tiers-secteur bien définies et distinctes font clairement apparaître. A partir de cette analyse, je trace ensuite les contours de trois logiques d’évaluation : une logique d’évaluation scientifique (observation scientifique, preuves observables et mesurables, procédures d’expérimentation objectives et solides), une logique d’évaluation bureaucratique (procédures pas-à-pas complexes, analyse des objectifs choisis) et une logique d’évaluation didactique (ouverture au changement, grande variété de perspectives, préférence pour l’expertise non-professionnelle). Ces logiques attirent notre attention sur des conceptions divergentes du savoir et de l’expertise et leurs implications en termes de ressources. Elles ont aussi des conséquences importantes pour le statut professionnel des praticiens, des consultants et des décideurs, qui contribuent à et/ou sont partie intégrante des évaluations des organisations du tiers-secteur.

Zusammenfassung

In diesem Beitrag präsentiere ich einen vorläufigen Entwurf über Formen der Bewertungslogiken im Dritten Sektor. Zunächst verfolge ich die Leitbilder, die in drei gut verständlichen, jedoch sehr unterschiedlichen Bewertungspraktiken des Dritten Sektors ersichtlich sind: das logische Rahmenwerk, die wichtigsten Hintergründe für Veränderungen und die Sozialrendite. Unter Bezugnahme auf diese Analyse präsentiere ich sodann eine vorläufige Darstellung dreier Bewertungslogiken: eine wissenschaftliche Bewertungslogik (systematische Beobachtung, beobachtbare und messbare Nachweise, objektive und robuste experimentelle Verfahren), eine bürokratische Bewertungslogik (umfangreiche, schrittweise Verfahren, Analyse beabsichtigter Ziele) und eine lernende Bewertungslogik (Offenheit für Veränderungen, viele unterschiedliche Perspektiven, Laienexpertise statt professioneller Expertise). Diese Logiken lenken die Aufmerksamkeit auf unterschiedliche Auffassungen von Wissen und Expertise und deren Auswirkungen auf die Ressourcen. Sie haben weitreichende Konsequenzen für den professionellen Status der Praktizierenden, Berater und Entscheidungsträger, die in Organisationen des Dritten Sektors zu Bewertungen beitragen oder sich mit diesen befassen.

Resumen

En el presente documento, proporciono un esbozo preliminar de los tipos de lógica de evaluación en el sector terciario. Comienzo por seguir el rastro a los ideales que son evidentes en las tres prácticas, bien articuladas pero muy diferentes, de evaluación del sector terciario: el marco lógico, las historias de cambio más significativas y la rentabilidad social de la inversión. Echando mano de este análisis, esbozo después provisionalmente tres lógicas de evaluación: una lógica de evaluación científica (observación sistemática, evidencia observable y mensurable, procedimientos experimentales objetivos y robustos), una lógica de evaluación burocrática (procedimientos complejos, paso a paso, análisis de los objetivos perseguidos) y una lógica de evaluación de aprendizaje (apertura al cambio, amplia gama de perspectivas, experiencia técnica secular en vez de profesional). Estas lógicas llaman la atención sobre diferentes concepciones del conocimiento y de la experiencia técnica y sobre las implicaciones de sus recursos, y tienen consecuencias importantes para el estatus profesional de los profesionales, asesores y aquellos que toman las decisiones que contribuyen a, y/o están implicados en, evaluaciones en organizaciones del sector terciario.

摘要

本文对评估第三产业的逻辑类型进行了初步探讨。首先,我针对三种论述缜密却又彼此迥异的第三产业评估方法——逻辑框架、最显著的改变案例和社会投资回报,分别梳理了每种方法的思路。在此分析的基础上,我试着描摹出三种评估逻辑的大致梗概:一种是科学型的评估逻辑(系统观察、可观察和度量的证据、客观严密的实验步骤),一种是官僚型的评估逻辑(复杂、按部就班、分析预定目标),还有就是学习型的评估逻辑(接受改变,广阔的视角,重视外行而非专业人士的意见)。在如何看待知识和专业技能及资源的意义这个问题上,奉行不同逻辑的人会有不同的理解。同时,对那些负责和/或参与第三产业评估的从业者、顾问和政策制定者来说,上述逻辑对其专业身份亦有重要影响。.

ملخص

أقدم في هذا البحث وصف تمهيدي لأنواع تقييم المنطق في القطاع الثالث. أبدأ من خلال تتبع المثل العليا التي هي واضحة في ثلاث معالم واضحة مع ذلك مختلفة تماما˝ ممارسات تقييم القطاع الثالث: الإطار المنطقي، قصص أهم تغيير، والعائد الاجتماعي على الإستثمار. بالإعتماد على هذا التحليل، أنا بعد ذلك حددت مبدئيا˝ ثلاثة منطق للتقييم:منطق التقييم العلمي،(أدلة ملحوظة وقابلة للقياس ، إجراءات موضوعية وتجريبية قوية)، منطق التقييم البيروقراطي (إجراءات معقدة، خطوة بخطوة ، تحليل أهداف مقصود) وتعلم منطق التقييم (الإنفتاح للتغيير، مجموعة واسعة من وجهات النظر، وضع بدلا من الخبرة المهنية). هذا المنطق لفت الإنتباه إلى إختلاف مفاهيم المعرفة والخبرة و آثار الموارد الخاصة بها ، ولها آثار هامة على الوضع المهني للممارسين، والإستشاريين، وصانعي السياسات التي تسهم في و/أو تشارك في عمليات التقييم في منظمات القطاع الثالث.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

Notes

  1. Whilst the analysis is focused upon these texts, I also examined a variety of other guide books and texts for each of the techniques. For the LFA: DFID (2009) Guidance on using the revised logical framework; SIDA (2004) The Logical framework approach; BOND (2003) Logical framework analysis; W.K. Kellogg Foundation (2004) Logic model development guide; World Bank (undated) The Logframe handbook. For the MSC: Dart and Davies (2003b) MSC Quick Start Guide; Clear Horizons (2009) Quick start guide MSC design; Davies (1998) An evolutionary approach to organizational learning. For the SROI: Olsen and Nicholls (2005) A framework for approaches to SROI analysis; NEF (2008) Measuring value: a guide to social return on investment (SROI) 2nd edition; NPC (2010) Social return on investment position paper; Office of the Third Sector UK (2009) A guide to social return on investment.

  2. For further information about the MSC, see http://mande.co.uk/special-issues/most-significant-change-msc/, and www.clearhorizon.com.au/flagship-techniques/most-significant-change/. More information, including example reports and applications, can be obtained by joining the MSC Yahoo group: http://groups.yahoo.com/group/MostSignificantChanges/.

  3. The academic journals examined included Nonprofit and Voluntary Sector Quarterly, Voluntas, Nonprofit Management and Leadership, American Journal of Evaluation, Evaluation, Public Administration and Development, and the Journal of International Development.

  4. Here the term ‘scientific’ functions as a descriptive label to characterze a particular approach to evaluation. The use of this term makes no claims about the scientific merits or value of this approach relative to other evaluation approaches.

  5. I thank an anonymous reviewer for suggesting this line of reasoning.

References

  • Abma, T. A. (1997). Playing with/in plurality: revitalizing realities and relationships in Rotterdam. Evaluation, 3, 25–48.

    Article  Google Scholar 

  • Acumen Fund (2007). The Best Available Charitable Option. Retrieved September 19, 2011 from http://www.acumenfund.org/uploads/assets/documents/BACO%20Concept%20Paper%20final_B1cNOVEM.pdf.

  • Bagnoli, L., & Megali, C. (2011). Measuring performance in social enterprises. Nonprofit and Voluntary Sector Quarterly, 40, 149–165.

    Article  Google Scholar 

  • Barman, E. (2007). What is the bottom line for third sector organizations? A history of measurement in the British voluntary sector. VOLUNTAS International Journal of Voluntary and Nonprofit Organizations, 18, 101–115.

    Article  Google Scholar 

  • Benjamin, L. M. (2008). Account space: How accountability requirements shape nonprofit practice. Nonprofit and Voluntary Sector Quarterly, 37, 201–223.

    Article  Google Scholar 

  • Blalock, A. N. (1999). Evaluation research and the performance management movement: from estrangement to useful integration? Evaluation, 5, 117–149.

    Article  Google Scholar 

  • BOND. (2003). Logical framework analysis. Guidance notes No.4.

  • Bouchard, J. M. (2009a). The worth of the social economy. In J. M. Bouchard (Ed.), The worth of the social economy: An international perspective (pp. 11–18). Brussels: P.I.E. Peter Lang.

  • Bouchard, J. M. (2009b). The evaluation of the social economy in Quebec, with regards to stakeholders, mission and organizational identity. In J. M. Bouchard (Ed.), The worth of the social economy: An international perspective (111–132). Brussels: P.I.E. Peter Lang.

  • Campos, J. G., Andion, C., Serva, M., Rossetto, A., & Assumpe, J. (2011). Performance evaluation in non-governmental organizations (NGOs): An analysis of evaluation models and their applications in Brazil. VOLUNTAS International Journal of Voluntary and Nonprofit Organizations, 22, 238–258.

    Article  Google Scholar 

  • Carman, J. G. (2007). Evaluation practice among community-based organizations: research into the reality. American Journal of Evaluation, 28, 60–75.

    Article  Google Scholar 

  • Charities Evaluation Service. (2008). Accountability and learning: Developing monitoring and evaluation in the third sector. London: Charities Evaluation Service.

  • Clear Horizons. (2009). Quick start guide: MSC design. Retrieved November 1, 2012 from http://www.clearhorizon.com.au/wp-content/uploads/2009/02/quick-start_feb09.pdf.

  • Conlin, S., & Stirrat, R. L. (2008). Current challenges in development evaluation. Evaluation, 14, 193–208.

    Article  Google Scholar 

  • Crewe, E., & Harrison, E. (1998). Whose development? An ethnography of aid. London: Zed Books.

    Google Scholar 

  • Dart, J., & Davies, R. (2003a). A dialogical, story-based evaluation tool: The most significant change technique. American Journal of Evaluation, 24, 137–155.

    Article  Google Scholar 

  • Dart, J., & Davies, R. (2003b). Quick-start guide: A self-help guide for implementing the most significant change technique (MSC). Retrieved November 1, 2012 from http://www.clearhorizon.com.au/wp-content/uploads/2008/12/dd-2003-msc_quickstart.pdf.

  • Davies, R. (1998). An evolutionary approach to organisational learning: an experiment by an NGO in Bangladesh. In D. Mosse, J. Farrington & A. Few (Eds.), Development as process: Concepts and methods for working with complexity (pp. 68–83). London: Routledge.

  • Davies, R., & Dart, J. (2005). The ‘most significant change’ (MSC) technique: A guide to its use. Retrieved November 1, 2012 from http://www.mande.co.uk/docs/MSCGuide.pdf.

  • Department for International Development (DFID). (2009). Guidance on using the revised logical framework. DFID Practice Paper.

  • Durie, S., Hutton, E., & Robbie, K. (2007). Investing in Impact: Developing Social Return on Investment. Retrieved September 14, 2012 from http://www.socialimpactscotland.org.uk/media/1200/SROI-%20investing%20in%20impact.pdf.

  • Easterling, D. (2000). Using outcome evaluation to guide grantmaking: Theory, reality, and possibilities. Nonprofit and Voluntary Sector Quarterly, 29, 482–486.

    Article  Google Scholar 

  • Ebrahim, A. (2002). Information struggles: The role of information in the reproduction of NGO-funder relationships. Nonprofit and Voluntary Sector Quarterly, 31, 84–114.

    Article  Google Scholar 

  • Ebrahim, A. (2005). Accountability myopia: Losing sight of organizational learning. Nonprofit and Voluntary Sector Quarterly, 34, 56–87.

    Article  Google Scholar 

  • Ebrahim, A., & Rangan, V. K. (2011). Performance measurement in the social sector: a contingency framework. Social Enterprise Initiative, Harvard Business School, working paper.

  • Eckerd, A., & Moulton, S. (2011). Heterogeneous roles and heterogeneous practices: Understanding the adoption and uses of nonprofit performance evaluations. American Journal of Evaluation, 32, 98–117.

    Article  Google Scholar 

  • Eme, B. (2009). Miseries and worth of the evaluation of the social and solidarity-based economy: For a paradigm of communicational evaluation. In J. M. Bouchard (Ed.), The worth of the social economy: An international perspective (pp. 63–86). Brussels: P.I.E. Peter Lang.

  • Enjolras, B. (2009). The public policy paradox. Normative foundations of social economy and public policies: Which consequences for evaluation strategies? In J. M. Bouchard (Ed.), The worth of the social economy: An international perspective (pp. 43–62) Brussels: P.I.E. Peter Lang.

  • Fine, A. H., Thayer, C. E., & Coghlan, A. T. (2000). Program evaluation practice in the nonprofit sector. Nonprofit Management and Leadership, 10(3), 331–339.

    Article  Google Scholar 

  • Fowler, A. (2002). Assessing NGO performance: Difficulties, dilemmas and a way ahead. In M. Edwards & A. Fowler (Eds.), The earthscan reader on NGO management (pp. 293–307). London: Earthscan.

  • Friedland, R., and Alford, R. R. (1991). Bringing society back in: Symbols, practices, and institutional contradictions. In W. W. Powell & P. J. DiMaggio (Eds.), The new institutionalism in organizational analysis (pp. 232–266). Chicago: University of Chicago Press.

  • Gasper, D. (2000). Evaluating the ‘logical framework approach’: Towards learning-oriented development evaluation. Public Administration and Development, 20, 17–28.

    Article  Google Scholar 

  • Greene, J. C. (1999). The inequality of performance measurements. Evaluation, 5, 160–172.

    Article  Google Scholar 

  • Guba, E. G., & Lincoln, Y. S. (1989). Fourth generation evaluation. Newbury Park: Sage Publications.

    Google Scholar 

  • Hoefer, R. (2000). Accountability in action? Program evaluation in nonprofit human service agencies. Nonprofit Management and Leadership, 11(2), 167–177.

    Article  Google Scholar 

  • Howes, M. (1992). Linking paradigms and practice: Key issues in the appraisal, monitoring and evaluation of British NGO projects. Journal of International Development, 4, 375–396.

    Article  Google Scholar 

  • Keystone (undated). Impact Planning, Assessment and Learning. Retrieved September 19, 2011 from http://www.keystoneaccountability.org/sites/default/files/1%20IPAL%20overview%20and%20service%20offering_0.pdf.

  • Jacobs, A., Barnett, C., & Ponsford, R. (2010). Three approaches to monitoring: Feedback systems, participatory monitoring and evaluation and logical frameworks. IDS Bulletin, 41, 36–44.

    Article  Google Scholar 

  • Kaplan, R. S. (2001). Strategic performance measurement and management in third sector organizations. Nonprofit Management and Leadership, 11(3), 353–371.

    Article  Google Scholar 

  • LeRoux, K., & Wright, N. S. (2011). Does performance measurement improve strategic decision making? Findings from a national survey of nonprofit social service agencies. Nonprofit and Voluntary Sector Quarterly (in press).

  • Lingane, A., & Olsen, S. (2004). Guidelines for social return on investment. California Management Review, 46, 116–135.

    Article  Google Scholar 

  • Lounsbury, M. (2008). Institutional rationality and practice variation: new directions in the institutional analysis of practice. Accounting, Organizations and Society, 33, 349–361.

    Article  Google Scholar 

  • Marquis, C., & Lounsbury, M. (2007). Vive la resistance: Competing logics and the consolidation of U.S. community banking. Academy of Management Journal, 50, 799–820.

    Article  Google Scholar 

  • Marsden, D., & Oakley, P. (1991). Future issues and perspectives in the evaluation of social development. Community Development Journal, 26, 315–328.

    Article  Google Scholar 

  • McCarthy, J. (2007). The ingredients of financial transparency. Nonprofit and Voluntary Sector Quarterly, 36, 156–164.

    Article  Google Scholar 

  • New Economics Foundation (NEF) (2007). Measuring real value: A DIY guide to social return on investment. London: New Economics Foundation.

  • New Economics Foundation (NEF) (2008). Measuring value: A guide to social return on investment (SROI) (2nd ed.). London: New Economics Foundation.

  • New Philanthropy Capital (NPC). (2010). Social return on investment. Position paper. Retrieved November 1, 2012 from http://www.thinknpc.org/publications/social-return-on-investment-position-paper/.

  • Nicholls, A. (2009). We do good things, don’t we? ‘Blend value accounting’ in social entrepreneurship. Accounting, Organizations and Society, 34, 755–769.

    Article  Google Scholar 

  • Office of the Third Sector (UK Cabinet Office). (2009). A guide to social return on investment. London: Office of the Third Sector.

  • Olsen, S., & Nicholls, J. (2005). A framework for approaches to SROI analysis. Retrieved November 1, 2012 from http://ccednet-rcdec.ca/sites/ccednet-rcdec.ca/files/ccednet/pdfs/2005-050624_SROI_Framework.pdf.

  • Porter, M. E., & Kramer, M. R. (1999). Philanthropy’s new agenda: creating value. Harvard Business Review, November/December, 121–130.

  • Reed, E., & Morariu, J. (2010). State of evaluation 2010: Evaluation practice and capacity in the nonprofit sector. Innovation Network. Retrieved November 1, 2012 from http://www.innonet.org/client_docs/innonet-state-of-evaluation-2010.pdf.

  • Roberts Enterprise Development Fund (REDF) (2001). Social return on investment methodology: Analyzing the value of social purpose enterprise within a social return on investment framework. Retrieved November 1, 2012 from http://www.redf.org/learn-from-redf/publications/119.

  • Roche, C. (1999). Impact assessment for development agencies: learning to value change. Oxford: Oxfam GB.

    Book  Google Scholar 

  • Rosenberg, L. J., & Posner, L. D. (1979). The logical framework: a manager’s guide to a scientific approach to design and evaluation. Washington DC: Practical Concepts.

  • Scott, J. C. (1998). Seeing like a state: How certain schemes to improve the human condition have failed. New Haven: Yale University Press.

    Google Scholar 

  • SIDA. (2004). The logical framework approach: A summary of the theory behind the LFA method. Stockholm: SIDA.

  • SIDA. (2006). Logical framework approach—with an appreciative approach. Stockholm: SIDA.

  • Stark, D. (2009). The sense of dissonance: Accounts of worth in economic life. Princeton: Princeton University Press.

    Google Scholar 

  • Urban Institute. (2006). Building a Common Outcome Framework to Measure Nonprofit Performance. Retrieved September 19, 2011 from http://www.urban.org/UploadedPDF/411404_Nonprofit_Performance.pdf.

  • W. K. Kellogg Foundation. (2004). Logic model development guide: Using logic models to bring together planning, evaluation and action. Battle Creek: W. K. Kellogg Foundation.

    Google Scholar 

  • Wallace, T., Bornstein, L., & Chapman, J. (2007). The aid chain: Coercion and commitment in development NGOs. Rugby, UK: Practical Action Publishing.

    Google Scholar 

  • Waysman, M., & Savaya, R. (1997). Mixed method evaluation: A case study. American Journal of Evaluation, 18, 227–237.

    Article  Google Scholar 

  • Weber, M. (1904/1949). The methodology of the social sciences (E. A. Shils & H. A. Finch (Eds. and Trans.). Free Press: New York.

  • World Bank. (2006). Conducting quality impact evaluations under budget, time and data constraints. Washington, DC: Independent Evaluation Group.

    Google Scholar 

  • World Bank. (undated). The LogFrame handbook: A logical framework approach to project cycle management. Retrieved November 1, 2012 from http://www.wau.boku.ac.at/fileadmin/_/H81/H811/Skripten/811332/811332_G3_log-framehandbook.pdf.

Download references

Acknowledgments

The author thanks the anonymous reviewers, Emily Barman, Alnoor Ebrahim, David Lewis, and participants at the 2011 ‘Impact of evaluations in the third sector: values, structures and relations’ conference and the ‘Rationalization and professionalization of the nonprofit sector’ track at EGOS 2012.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Matthew Hall.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Hall, M. Evaluation Logics in the Third Sector. Voluntas 25, 307–336 (2014). https://doi.org/10.1007/s11266-012-9339-0

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11266-012-9339-0

Keywords

Navigation