Abstract
Several different EKE protocols are reviewed in this volume, each with their pros and cons, but any is only as good as the quality of the experts and their judgments. In this chapter a structured approach to the selection of experts for EKE is presented that is grounded in psychological research.
In Part I various definitions of expertise are considered, and indicators and measures that can be used for the selection of experts are identified. Next, some ways of making judgements of uncertain quantities are discussed, as are factors influencing judgment quality.
In Part II expert selection is considered within an overall policy-making process. Following the analysis of Part I, two new instruments are presented that can help guide the selection process: expert profiles provide structure to the initial search, while a questionnaire permits matching of experts to the profiles, and assessment of training needs. Issues of expert retention and documentation are also discussed.
It is concluded that although the analysis offered in this chapter constitutes a starting point there are many questions still to be answered to maximize EKE’s contribution. A promising direction is research that focusses on the interaction between experts and the tasks they perform.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
This is a manifestation of the ‘paradox of expertise’ (e.g. Dror 2011), which is that experts become worse in some respects as they become more expert e.g. less flexible, creative and responsive; more biased etc. This is because knowledge and reasoning become ‘fossilized’: less amenable to inspection, change and communication.
- 2.
In the Delphi procedure, ‘groups’ of experts—who never meet or interact directly, and are anonymous to each other (all to reduce sources of social bias)—are polled for their opinions. These opinions are usually point estimates or forecasts of event occurrence (see e.g. Chap. 5 for a discussion of the elicitation and evaluation of such judgments) but can also be judgments of uncertain quantities expressed as probability intervals or distributions (see Bolger et al. 2014): reasons for judgments are also often elicited. Once experts have individually expressed their opinions they are collated by a facilitator and fed back to the expert panel (most normally quantitative estimates are averaged in some manner, and qualitative responses summarized, although individual responses may also be fed back if the group is not too large). The experts are then invited to revise their opinions in the light of the feedback and resubmit them to the facilitator. The process continues through a number of iterations usually until opinion change ceases. Normally the aggregated judgements from the final round are the output although partially aggregated or disaggregated judgements can be submitted if the process fails to lead to consensus.
- 3.
- 4.
For instance, stock price movements have been characterized as random (e.g. Fama 1965). Although more recent research suggests that stock markets are, in fact, predictable in the long term (e.g. De Bondt and Thaler 1989) it is still agreed that it is not in the short-term, contrary to the beliefs of ‘day-traders’. It may often be the case that ‘experts’ believe there to be predictability where there is not, or it is rather low. In such situations, there can, of course, be little or no expertise (see e.g. forecasting of GDP growth, Budescu and Chen 2015) nor variation in performance. Further, perceived ability where there is none is another name for ‘overconfidence’ (more generally, insensitivity to task difficulty will lead to miscalibration).
- 5.
Bolger and Rowe (2015a) identify a number of problems with this approach, including finding a sufficient number of suitable seeds—ones that draw on the same expert knowledge as the target. They also comment that this ‘Classical Method’? is atheoretical, in that it is not founded on any particular conceptualization of expert knowledge, and propose a cue-based approach that would provide a reasoned basis for the selection of similar seeds (i.e. those that are related by the cues used to judge them). This cue-based approach is outlined in Sections “Cue-Based Judgements” and 16.2.4.2 below.
- 6.
However it must be stressed that this is just one study (which fails to report all the potentially relevant correlations). Further, we do not know the extent to which the tests of substantive expertise are good measures of actual expert performance on, for example, a real-world risk-assessment or forecasting task.
- 7.
The other two traits being extroversion and neuroticism.
- 8.
It is possible that the advantages of some personality traits are protocol-dependent. For example, conscientiousness might be good for a remotely administered elicitation such as is often the case in Delphi, while agreeableness might be particularly helpful in protocols that require face-to-face interaction. Openness to experience is probably a useful characteristic in both protocols as it should assist opinion change towards the true value.
- 9.
Indeed, the ‘Theory of Errors’ (Dalkey 1975; Parenté and Anderson-Parenté 1987), which is the leading account for why the Delphi technique works, assumes those who stick are on average closer to the truth than those who shift—Bolger and Wright (2011) propose that in order to achieve ‘virtuous opinion change’—i.e. opinion change towards the truth—rationales for opinions should be fed back between Delphi rounds rather than confidence as the former will be better indicators that the expert is knowledgeable about the topic than the latter.
- 10.
There is some empirical support for the suggestion that relative frequency is a more natural way of representing uncertainty than probability (Gigerenzer and Hoffrage 1995) thus posing questions as relative frequencies rather than probabilities might be an alternative to training.
- 11.
In principle, many of these things could be done remotely but I am not aware of any existing protocols or software to support this.
- 12.
- 13.
Support Theory also is well-developed with regard to uncertainty assessment but its primary focus is the assessment of the likelihood of categorical assessments (e.g. of event occurrences, correctness, or truth) rather than values of continuous variables.
- 14.
Although as already indicated, there is still an ongoing debate as to the source and degree (and even existence) of overconfidence in such tasks—see Olsson 2014, for a recent review.
- 15.
Bazerman and Moore (2008) suggest that experts also need to have coherent models (mental or formal) in order to make good quality (probability) judgements. So models may be added to learnability and ecological validity of tasks as a criterion for well-calibrated experts.
- 16.
For example, ‘bootstrap’ linear models—models derived from regressing expert judgments onto the cues that they are presumed to use—make better predictions of a criterion than do the original unaided experts (e.g. Goldberg 1970) because they apply the judgment model more consistently.
- 17.
In contrast to the ‘crowdsourcing’ approach mentioned above whereby a large number of people with little or no domain expertise are polled (see, e.g. Budescu and Chen 2015).
- 18.
These experts will be referred to later as ‘super-experts’ since they have an overview of the problem as a whole and are responsible for recruitment, selection and management of any other experts used in the process.
- 19.
- 20.
However, in some other important respects, expert selection is not like job selection i.e. you want to find people with the right skills rather than reject people with the wrong skills… so it is more akin to head hunting.
- 21.
Potentially, scores for weighting could be derived from this questionnaire but this has not been attempted yet, let alone any validation of weights thus derived, therefore this is something for future research.
References
Alvarado-Valencia J, Barrero LH, Onkal D, Dennerlein JT (2017) Expertise, credibility of systems forecasts and integration of methods in judgmental demand forecasting. Int J Forecast 33(1):298–313
Anderson JR (1982) Acquisition of cognitive skill. Psychol Rev 89:369–406
Aspinall W (2010) A route to more tractable expert advice. Nature 463:294–295
Bargh JA (1994) The four horsemen of automaticity: intention, awareness, efficiency, and control as separate issues. In: Wyer R, Srull T (eds) Handbook of social cognition. Lawrence Erlbaum, Hillsdale, NJ, pp 1–40
Barron G, Erev I (2003) Small feedback-based decisions and their limited correspondence to description-based decisions. J Behav Decis Mak 16(3):215–233
Bayindir M, Bolger F, Say B (2017) An investigation of the role of some person and situation variables in multiple cue probability learning. Q J Exp Psychol 70(1):36–52
Bazerman MH, Moore DA (2008) Judgment in managerial decision making, 7th edn. Wiley, New York
Bolger F, Rowe G (2015a) The aggregation of expert judgment: do good things come to those who weight? Risk Anal 35(1):5–11
Bolger F, Rowe G (2015b) There is data, and then there is data: only experimental evidence will determine the utility of differential weighting of expert judgment. Risk Anal 35(1):21–26
Bolger F, Wright G (1993) Coherence and calibration in expert probability judgement. OMEGA Int J Manag Sci 21:629–644
Bolger F, Wright G (1994) Assessing the quality of expert judgment: issues and analysis. Decis Support Syst 11:1–24
Bolger F, Wright G (2011) Improving the Delphi process: lessons from social psychological research. Technol Forecast Soc Chang 78:1500–1513
Bolger F, Wright G (2017) Use of expert knowledge to anticipate the future: issues, analysis and directions. Int J Forecast 33(1):230–243
Bolger F, Wright G, Rowe G, Gammack J, Wood R (1989) LUST for life: developing expert systems for life assurance underwriting. In: Shadbolt N (ed) Research and development in expert systems VI. Cambridge University Press, Cambridge, pp 128–139
Bolger F, Hanea A, Mosbach-Schulz O, Oakley J, O’Hagan A, Rowe G, Wenholt M (2014) Guidance on expert knowledge elicitation in food and feed safety risk assessment. European Food Safety Authority (EFSA), Parma
Brenner LA (2003) A random support model of the calibration of subjective probabilities. Organ Behav Hum Decis Process 90(1):87–110
Brunswik E (1955) Representative design and probabilistic theory in a functional psychology. Psychol Rev 62(3):193
Budescu DV, Chen E (2015) Identifying expertise to extract the wisdom of crowds. Manag Sci 61(2):267–280
Burgman MA, McBride M, Ashton R, Speirs-Bridge A, Flander L et al (2011) Expert status and performance. PLoS One 6(7):E22998
Chase WG, Simon HA (1973) Perception in chess. Cogn Psychol 4(1):55–81
Clemen RT (2008) Comment on Cooke’s classical method. Reliab Eng Syst Saf 93:760–765
Cooke RM (2014) Validating expert judgments with the classical model. In: Martini C, Boumans M (eds) Experts and consensus in social science—critical perspectives from economics, sociology, politics, and philosophy. Ethical economy—studies in economic ethics and philosophy. Springer, Heidelberg, pp 191–121
Cooke RM, Goossens LLHJ (2008) TU Delft expert judgment database. Reliab Eng Syst Saf 93:657–674
Cooke RM, ElSaadany S, Huang X (2008) On the performance of social network and likelihood-based expert weighting schemes. Reliab Eng Syst Saf 93(5):745–756
Dalgleish LI (1988) Decision making in child abuse cases: Applications of social judgment theory and signal detection theory. In: Brehmer B, Joyce CB (eds) Human judgment: the SJT view. Elsevier, Amsterdam, pp 317–360
Dalkey NC (1975) Toward a theory of group estimation. In: Linstone HA, Turoff M (eds) The Delphi method: techniques and applications. Addison-Wesley, Reading, MA, pp 236–261
De Bondt WF, Thaler RH (1989) Anomalies: a mean-reverting walk down Wall Street. J Econ Perspect 3(1):189–202
De Groot AD (1965) Thought and choice in chess. The Hague, Mouton
Dhami MK, Wallsten TS (2005) Interpersonal comparison of subjective probabilities: toward translating linguistic probabilities. Mem Cogn 33(6):1057–1068
Dror IE (2011) The paradox of human expertise: why experts get it wrong. In: Kapur N (ed) The paradoxical brain. Cambridge University Press, Cambridge, pp 177–188
Eden C (1988) Cognitive mapping. Eur J Oper Res 36(1):1–13
Edwards W (1968) Conservatism in human information processing. In: Kleinmuntz B (ed) Formal representation of human judgment, CMU cognition series, vol 3. Wiley, New York, pp 17–51
Eggstaff JW, Mazzuchi TA, Sarkani S (2014) The effect of the number of seed variables on the performance of Cooke’s classical model. Reliab Eng Syst Saf 121:72–82
Ellis HC, Ashbrook PW (1989) The state of mood and memory research: a selective review. J Soc Behav Pers 4(2):1–21
European Food Safety Authority (2013) Technical meeting of the EFSA Scientific Network on EFSA Scientific. Network for risk assessment in Animal Health and Welfare – Risk of introduction of Rift Valley fever into the Southern Mediterranean area through undocumented movement of infected animals. EFSA Supporting Documents, 10 (4)
Fama EF (1965) The behaviour of stock market prices. J Bus 38:34–105
Ferrell WR, McGoey PJ (1980) A model of calibration for subjective probabilities. Organ Behav Hum Perform 26(1):32–53
Frederick S (2005) Cognitive reflection and decision making. J Econ Perspect 19(4):25–42
Genre V, Kenny G, Meyler A, Timmermann A (2013) Combining expert forecasts: can anything beat the simple average? Int J Forecast 29:108–121
Germain ML (2006) Development and preliminary validation of a psychometric measure of expertise: the generalized expertise measure (GEM). Unpublished doctoral dissertation. Barry University, Florida, USA
Germain M-L, Tejeda MJ (2012) A preliminary exploration on the measurement of expertise: an initial development of a psychometric scale. Hum Resour Dev Q 23(2):203–232
Gibbons AM, Sniezek JA, Dalal RS (2003) Antecedents and consequences of unsolicited versus explicitly solicited advice. In: Budescu D (Chair), Symposium in Honor of Janet Sniezek. Symposium presented at the annual meeting of the Society for Judgment and Decision Making, Vancouver, BC (2003 November)
Gigerenzer G, Hoffrage U (1995) How to improve Bayesian reasoning without instruction: frequency formats. Psychol Rev 102(4):684
Gigerenzer G, Hoffrage U, Kleinbölting H (1991) Probabilistic mental models: a Brunswikian theory of confidence. Psychol Rev 98(4):506–528
Goldberg LR (1970) Man versus model of man: a rationale, plus some evidence, for a method of improving on clinical inferences. Psychol Bull 73(6):422
Griffin D, Brenner L (2004) Perspectives on probability judgment calibration. In: Koehler DJ, Harvey N (eds) Blackwell handbook of judgment and decision making. Blackwell, Malden (MA), pp 177–199
Griffin D, Tversky A (1992) The weighing of evidence and the determinants of confidence. Cogn Psychol 24(3):411–435
Hardman DK (2009) Judgment and decision making: psychological perspectives. Wiley, Chichester
Hayes-Roth F, Waterman D, Lenat D (1983) Building expert systems. Addison-Wesley, Boston
Hertwig R, Barron G, Weber EU, Erev I (2004) Decisions from experience and the effect of rare events in risky choice. Psychol Sci 15(8):534–539
Hodge R, Evans M, Marshall J, Quigley J, Walls L (2001) Eliciting engineering knowledge about reliability during design-lessons learnt from implementation. Qual Reliab Eng Int 17(3):169–179
Hoefling A, Strack F (2010) Hunger induced changes in food choice. When beggars cannot be choosers even if they are allowed to choose. Appetite 54(3):603–606
Howard RA, Matheson JE (2005) Influence diagrams. Decis Anal 3:127–143
Isen AM, Erez A (2002) The influence of positive affect on the components of expectancy motivation. J Appl Psychol 87(6):1055–1067
Ivlev I, Kneppo P, Bartak M (2015) Method for selecting expert groups and determining the importance of experts’ judgments for the purpose of managerial decision-making tasks in health systems. Ekon Manag 2(18):57–72
Janis IL (1982) Groupthink: psychological studies of policy decisions and fiascoes. Houghton Mifflin, Boston
Juslin P (1993) An explanation of the hard-easy effect in studies of realism of confidence in one’s general knowledge. Eur J Cogn Psychol 5(1):55–71
Juslin P (1994) The overconfidence phenomenon as a consequence of informal experimenter-guided selection of almanac items. Organ Behav Hum Decis Process 57(2):226–246
Keren G (1991) Calibration and probability judgements: conceptual and methodological issues. Acta Psychol 77(3):217–273
Klahr D, Langley P, Neches R (1987) Production system models of learning and development. MIT Press, Cambridge, MA
Klein G (1998) Sources of power: how people make decisions. MIT Press, Cambridge, MA
Kleindorfer PR, Kunreuther H, Schoemaker PJ (1993) Decision sciences: an integrative perspective. Cambridge University Press, Cambridge, p 484
Kliger D, Levy O (2003) Mood and judgment of subjective probabilities: evidence from the US index option market. Eur Finan Rev 7(2):235–248
Koehler D, Brenner L, Griffin D (2002) The calibration of expert judgment: heuristics and biases beyond the laboratory. In: Gilovich T, Griffin D, Kahneman D (eds) Heuristics and biases: the psychology of intuitive judgment. Cambridge University Press, Cambridge, pp 686–715
Koriat A, Lichtenstein S, Fischhoff B (1980) Reasons for confidence. J Exp Psychol Hum Learn Mem 6(2):107–118
Lave J, Wenger E (1991) Situated learning: legitimate peripheral participation. Cambridge University Press, Cambridge
Lichtenstein S, Fischhoff B, Phillips L (1982) Calibration of probabilities: the state of the art to 1980. In: Kahneman D, Slovic P, Tversky A (eds) Judgment under uncertainty: heuristics and biases. Cambridge University Press, Cambridge, pp 306–334
Lin SW, Bier VM (2008) A study of expert overconfidence. Reliab Eng Syst Saf 93:711–721
Lin SW, Cheng CH (2009) The reliability of aggregated probability judgments obtained through Cooke’s classical method. J Model Manag 4:149–161
Lohr S (2009) Sampling: design and analysis. Nelson Education, Toronto
MacGregor DG, Lichtenstein S (1991) Problem structuring aids for quantitative estimation. J Behav Decis Mak 4:101–116
MacGregor DG, Lichtenstein S, Slovic P (1988) Structuring knowledge retrieval: an analysis of decomposed quantitative judgments. Organ Behav Hum Decis Process 42:303–323
MacIntyre A (2007) After virtue: a study in moral theory, 3rd edn. Duckworth, London
Malkiel BG (2011, October) The efficient-market hypothesis and the financial crisis. In: Rethinking finance: perspectives on the crisis (Proceedings of a conference). Russel Sage Foundation
Mannes AE, Soll JB, Larrick RP (2014) The wisdom of select crowds. J Pers Soc Psychol 107(2):76–299
Meyer MA, Booker JM (1991) Eliciting and analyzing expert judgment: a practical guide. Academic, London
Moore G, Beadle R (2006) In search of organizational virtue in business: agents, goods, practices, institutions and environments. Organ Stud 27:369–389
Moreira C, Wichert A (2013) Finding academic experts on a multisensor approach using Shannon’s entropy. Expert Syst Appl 40:5740–5754
Murphy AH, Brown BG (1985) A comparative evaluation of objective and subjective weather forecasts in the United States. In: Wright G (ed) Behavioral decision making. Plenum, New York, pp 329–359
Newell A, Simon HA (1972) Human problem solving. Prentice-Hall, Englewood Cliffs, NJ
Newell A, Shaw JC, Simon HA (1959) The processes of creative thinking. Rand Corporation, Santa Monica, CA
Nisbett RE, Wilson TD (1977) Telling more than we can know: verbal reports on mental processes. Psychol Rev 84(3):231–259
Oliver RM, Smith JQ (1990) Influence diagrams, belief nets and decision analysis. Wiley, New York
Olsson H (2014) Measuring overconfidence: methodological problems and statistical artifacts. J Bus Res 67:1766–1770
Oskamp S (1965) Overconfidence in case study judgments. J Consult Psychol 29:261–265
Parenté FJ, Anderson-Parenté JK (1987) Delphi inquiry systems. In: Wright G, Ayton P (eds) Judgmental forecasting. Wiley, Chichester, pp 129–156
Phillips LD (1987) On the adequacy of judgmental forecasts. In: Wright G, Ayton P (eds) Judgmental forecasting. Wiley, Oxford, pp 11–30
Phillips JM (1999) Antecedents of leader utilization of staff input in decision-making teams. Organ Behav Hum Decis Process 77:215–242
Price PC, Stone ER (2004) Intuitive evaluation of likelihood judgment producers: evidence for a confidence heuristic. J Behav Decis Mak 17(1):39–57
Rakow T, Newell BR (2010) Degrees of uncertainty: an overview and framework for future research on experience-based choice. J Behav Decis Mak 23(1):1–14
Rowe G, Bolger F (2016) The identification of food safety priorities using the Delphi technique. EFSA Supporting Publications, 13 (3). EN-1007, 141 pp
Rowe G, Wright G (1996) The impact of task characteristics on the performance of structured group forecasting techniques. Int J Forecast 12:73–90
Rowe G, Wright G (2001) Differences in expert and lay judgments of risk: myth or reality? Risk Anal 21(2):341–356
Rowe G, Wright G, McColl A (2005) Judgment changes during Delphi-like procedures: the role of majority influence, expertise, and confidence. Technol Forecast Soc Chang 72:217–238
Saunders EMJ (1993) Stock prices and Wall Street weather. Am Econ Rev 83:1337–1345
Solomon I, Ariyo A, Tomasini LA (1985) Contextual effects on the calibration of probabilistic judgments. J Appl Psychol 70:528–532
Tetlock P, Gardner D (2016) Superforecasting: the art and science of prediction. Random House, New York
Toulmin S (1958) The uses of argument. Cambridge University Press, Cambridge, p 272
Tversky A, Koehler DJ (1994) Support theory: a nonextensional representation of subjective probability. Psychol Rev 101:547–567
Walls L, Quigley J, Marshall J (2006) Modeling to support reliability enhancement during product development with applications in the UK aerospace industry. IEEE Trans Eng Manag 53(2):263–274
Wallsten TS, Budescu DV (1995) A review of human linguistic probability processing: general principles and empirical evidence. Knowl Eng Rev 10(1):43–62
Waterman DA, Hayes-Roth F (1978) Pattern-directed inference systems. Academic, New York
Wenger E (1998) Communities of practice: learning, meaning and identity. Cambridge University Press, Cambridge
Wright G, Rowe G, Bolger F, Gammack J (1994) Coherence, calibration and expertise in judgmental probability forecasting. Organ Behav Hum Decis Process 57:1–25
Wright G, Bolger F, Rowe G (2002) An empirical test of the relative validity of expert and lay judgments of risk. Risk Anal 22:1107–1122
Wright G, Rowe G, McColl A (2004) A framework for future study of expert and lay differences in the judgment of risk. Risk Decis Policy 9(2):91–106
Acknowledgements
I wish to acknowledge the support of the European Food Safety Authority (EFSA) who convened the “Working Group on Guidance on Expert Knowledge Elicitation in Food and Feed Safety Risk Assessment”: the research conducted for which laid the foundations for this chapter. I also wish to thank Working Group members, Anca Hanea, Anthony O‘Hagan, Jeremy Oakley, Gene Rowe and Meike Wentholt; and EFSA staff, Elisa Aiassa, Fulvio Barizzone, Eugen Christoph, Andrea Gervelmeyer, Olaf Mosbach-Schulz, and Sara Tramontini for their contributions to the original research. Finally, I am grateful to David Budescu, John Quigley, and Gene Rowe for their comments on earlier drafts of this manuscript.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG
About this chapter
Cite this chapter
Bolger, F. (2018). The Selection of Experts for (Probabilistic) Expert Knowledge Elicitation. In: Dias, L., Morton, A., Quigley, J. (eds) Elicitation. International Series in Operations Research & Management Science, vol 261. Springer, Cham. https://doi.org/10.1007/978-3-319-65052-4_16
Download citation
DOI: https://doi.org/10.1007/978-3-319-65052-4_16
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-65051-7
Online ISBN: 978-3-319-65052-4
eBook Packages: Business and ManagementBusiness and Management (R0)