Abstract
A longstanding finding in the judgment and decision-making literature is that human decision performance can be improved with the help of a mechanical aid. Despite this observation and celebrated advances in computing technologies, recently presented evidence of algorithm aversion raises concerns about whether the potential of human-machine decision-making is undermined by a human tendency to discount algorithmic outputs. In this chapter, we examine the algorithm aversion phenomenon and what it means for judgment in predictive analytics. We contextualize algorithm aversion in the broader human vs. machine debate and the augmented decision-making literature before defining algorithm aversion, its implications, and its antecedents. Finally, we conclude with proposals to improve methods and metrics to help guide the development of human-machine decision-making.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
- 2.
The other primary opposition came decades later in Gary Klein’s (1993, 1997, 2008) studies on naturalistic decision-making. These studies promote reliance on expert intuition by focusing on real world contexts marked by time pressure and high-stake consequences, rather than artificial experiments. While Kahneman acted as his contemporary adversary, they reconciled their positions by agreeing that the comparative performance of human versus machine judgment depends on the environment in which it takes place (Kahneman & Klein, 2009).
- 3.
Bayes’ theorem is a mathematical formula for calculating a conditional probability that a hypothesis is true given some evidence (for an in depth review see Joyce, 2003).
- 4.
- 5.
These five factors fall under the fives themes presented in Burton et al. (2020): expectations and expertise, decision autonomy, incentivization, cognitive compatibility, and divergent rationalities.
- 6.
We note that in experimental set ups using variants of the WOA measure (e.g., Logg et al., 2019; Önkal et al., 2009), decision accuracy is often artificially accounted for because the machine aid’s judgment (or advice) is guaranteed to be highly accurate by the experimenter. However, this is not necessarily guaranteed in the real world, where achieving a high WOA may be undesirable in some circumstances (i.e, complacency bias or automation bias, Baudel et al., 2021; Parasuraman & Manzey, 2010; Zerilli et al., 2019).
References
Ægisdóttir, S., White, M. J., Spengler, P. M., Maugherman, A. S., Anderson, L. A., Cook, R. S., Nichols, C. N., Lampropoulos, G. K., Walker, B. S., Cohen, G., & Rush, J. D. (2006). The meta-analysis of clinical judgment project: Fifty-six years of accumulated research on clinical versus statistical prediction. The Counseling Psychologist, 34(3), 341–382. https://doi.org/10.1177/0011000005285875
Alavi, M., & Henderson, J. C. (1981). An evolutionary strategy for implementing a decision support system. Management Science, 27(11), 1309–1323.
Alexander, V., Blinder, C., & Zak, P. J. (2018). Why trust an algorithm? Performance, cognition, and neurophysiology. Computers in Human Behavior. https://doi.org/10.1016/j.chb.2018.07.026
Almaatouq, A., Griffiths, T. L., Suchow, J. W., Whiting, M. E., Evans, J., & Watts, D. J. (2022). Beyond playing 20 questions with nature: Integrative experiment design in the social and behavioral sciences. Behavioral and Brain Sciences, 2022, 1–55. https://doi.org/10.1017/s0140525x22002874
Arkes, H. R., Dawes, R. M., & Christensen, C. (1986). Factors influencing the use of a decision rule in a probabilistic task. Organizational Behavior and Human Decision Processes, 37, 93–110.
Arkes, H. R., Shaffer, V. A., & Medow, M. A. (2007). Patients derogate physicians who use a computer-assisted diagnostic aid. Med Decis Making, 27(2), 189–202. https://doi.org/10.1177/0272989X06297391
Ashton, A. H., Ashton, R. H., & Davis, M. N. (1994). White-collar robotics: Levering managerial decision making. California Management Review, 37, 83–109.
Baudel, T., Verbockhaven, M., Cousergue, V., Roy, G., & Laarach, R. (2021). ObjectivAIze: Measuring performance and biases in augmented business decision systems. In C. Ardito, R. Lanzilotti, A. Malizia, H. Petrie, A. Piccinno, G. Desolda, & K. Inkpen (Eds.), Human-computer interaction – INTERACT 2021 (Vol. 12934, pp. 300–320). Springer. https://doi.org/10.1007/978-3-030-85613-7_22
Benbasat, I., & Taylor, R. N. (1978). The impact of cognitive styles on information system design. MIS Quarterly, 2(2), 43–54.
Berger, B., Adam, M., Rühr, A., & Benlian, A. (2021). Watch me improve—Algorithm aversion and demonstrating the ability to learn. Business & Information Systems Engineering, 63(1), 55–68. https://doi.org/10.1007/s12599-020-00678-5
Bigman, Y. E., & Gray, K. (2018). People are averse to machines making moral decisions. Cognition, 181, 21–34. https://doi.org/10.1016/j.cognition.2018.08.003
Blattberg, R. C., & Hoch, S. J. (1990). Database models and managerial intuition: 50% model + 50% Manager. Management Science, 36(8), 887–899. https://doi.org/10.1287/mnsc.36.8.887
Bonaccio, S., & Dalal, R. S. (2006). Advice taking and decision-making: An integrative literature review, and implications for the organizational sciences. Organizational Behavior and Human Decision Processes, 101(2), 127–151. https://doi.org/10.1016/j.obhdp.2006.07.001
Brown, R. V. (2015). Decision science as a by-product of decision-aiding: A practitioner’s perspective. Journal of Applied Research in Memory and Cognition, 4, 212–220. https://doi.org/10.1016/j.jarmac.2015.07.005
Brown, D. L., & Jones, D. R. (1998). Factors that influence reliance on decision aids: A model and an experiment. Journal of Information Systems, 12(2), 75–94.
Brunswik, E. (1955). Representative design and probabilistic theory in a functional psychology. Psychological Review, 62(3), 193–217. https://doi.org/10.1037/h0047470
Burton, J. W., Stein, M., & Jensen, T. B. (2020). A systematic review of algorithm aversion in augmented decision making. Journal of Behavioral Decision Making, 33(2), 220–239. https://doi.org/10.1002/bdm.2155
Burton, J. W., Almaatouq, A., Rahimian, M. A., & Hahn, U. (2021a). Rewiring the wisdom of the crowd. Proceedings of the Annual Meeting of the Cognitive Science Society, 43, 1802–1808. Retrieved from https://escholarship.org/uc/item/7tj34969
Burton, J. W., Hahn, U., Almaatouq, A., & Rahimian, M. A. (2021b). Algorithmically mediating communication to enhance collective decision-making in online social networks. ACM Collective Intelligence Conference, 2021(9), 1–3. Retrieved from https://www.acm-ci2021.com/program
Cadario, R., Longoni, C., & Morewedge, C. K. (2021). Understanding, explaining, and utilizing medical artificial intelligence. Nature Human Behaviour. https://doi.org/10.1038/s41562-021-01146-0
Camerer, C. (1981). General conditions for the success of bootstrapping models. Organizational Behavior and Human Performance, 27(3), 411–422. https://doi.org/10.1016/0030-5073(81)90031-3
Carey, J. M., & Kacmar, C. J. (2003). Toward a general theoretical model of Computerbased factors that affect managerial decision making. Journal of Managerial Issues, 15(4), 430–449.
Castelo, N., Bos, M. W., & Lehmann, D. R. (2019). Task-dependent algorithm aversion. Journal of Marketing Research, 56(5), 809–825. https://doi.org/10.1177/0022243719851788
Christin, A. (2017). Algorithms in practice: Comparing web journalism and criminal justice. Big Data & Society, 1–14. https://doi.org/10.1177/2053951717718855
Dana, J., & Thomas, R. (2006). In defense of clinical judgment … and mechanical prediction. Journal of Behavioral Decision Making, 19(5), 413–428. https://doi.org/10.1002/bdm.537
Dawes, R. M. (1971). A case study of graduate admissions: Application of three principles of human decision making. American Psychologist, 26(2), 180–188. https://doi.org/10.1037/h0030868
Dawes, R. M. (1979). The robust beauty of improper linear models in decision making. American Psychologist, 34(7), 571–582.
Dawes, R. M., Faust, D., & Meehl, P. E. (1989). Clinical versus actuarial judgment. Science, 243(4899), 1668–1674. https://doi.org/10.1126/science.2648573
Dietvorst, B. J., & Bharti, S. (2020). People reject algorithms in uncertain decision domains because they have diminishing sensitivity to forecasting error. Psychological Science, 31(10), 1302–1314. https://doi.org/10.1177/0956797620948841
Dietvorst, B. J., Simmons, J. P., & Massey, C. (2015). Algorithm aversion: People erroneously avoid algorithms after seeing them err. Journal of Experimental Psychology: General, 144(1), 114–126. https://doi.org/10.1037/xge0000033
Dietvorst, B. J., Simmons, J. P., & Massey, C. (2016). Overcoming algorithm aversion: People will use imperfect algorithms if they can (even slightly) modify them. Management Science, 64(3), 1155–1170. https://doi.org/10.1287/mnsc.2016.2643
Dijkstra, J. J. (1999). User agreement with incorrect expert system advice. Behaviour & Information Technology, 18(6), 399–411. https://doi.org/10.1080/014492999118832
Dijkstra, J. J., Liebrand, W. B. G., & Timminga, E. (1998). Persuasiveness of expert systems. Behaviour & Information Technology, 17(3), 155–163. https://doi.org/10.1080/014492998119526
Eastwood, J., Snook, B., & Luther, K. (2012). What people want from their professionals: Attitudes toward decision-making strategies. Journal of Behavioral Decision Making, 25, 458–468. https://doi.org/10.1002/bdm.741
Edwards, W. (1954). The theory of decision making. Psychological Bulletin, 51(4), 380–417.
Edwards, W. (1962). Dynamic decision theory and probabilistic information processings. Human Factors, 4(2), 59–74. https://doi.org/10.1177/001872086200400201
Efendić, E., Van de Calseyde, P. P. F. M., & Evans, A. M. (2020). Slow response times undermine trust in algorithmic (but not human) predictions. Organizational Behavior and Human Decision Processes, 157, 103–114. https://doi.org/10.1016/j.obhdp.2020.01.008
Einhorn, H. J. (1972). Expert measurement and mechanical combination. Organizational Behavior and Human Performance, 7(1), 86–106. https://doi.org/10.1016/0030-5073(72)90009-8
Einhorn, H. J. (1986). Accepting error to make less error. Journal of Personality Assessment, 50(3), 387–395. https://doi.org/10.1207/s15327752jpa5003_8
Eining, M. M., Jones, D. R., & Loebbecke, J. K. (1997). Reliance on decision aids: An examination of auditors’ assessment of management fraud. Auditing: A Journal of Practice & Theory, 16(2), 1–19.
Er, M. C. (1988). Decision support systems: A summary, problems, and future trends. Decision Support Systems, 4, 355–363.
Fildes, R., & Petropoulos, F. (2015). Improving forecast quality in practice. Foresight: The International Journal of Applied Forecasting, 36, 5–12.
Gigerenzer, G., & Gaissmaier, W. (2015). Decision making: Nonrational theories. In International Encyclopedia of the Social & Behavioral Sciences (pp. 911–916). Elsevier. https://doi.org/10.1016/B978-0-08-097086-8.26017-0
Ginsberg, J., Mohebbi, M. H., Patel, R. S., Brammer, L., Smolinski, M. S., & Brilliant, L. (2009). Detecting influenza epidemics using search engine query data. Nature, 457(7232), 1012–1014. https://doi.org/10.1038/nature07634
Goldberg, L. R. (1965). Diagnosticians vs. diagnostic signs: The diagnosis of psychosis vs. neurosis from the MMPI. Psychological Monographs: General and Applied, 79(9), 1–28. https://doi.org/10.1037/h0093885
Goldberg, L. R. (1970). Man versus model of man: A rationale, plus some evidence, for a method of improving on clinical inferences. Psychological Bulletin, 73(6), 422–432. https://doi.org/10.1037/h0029230
Green, G. I., & Hughes, C. T. (1986). Effects of decision support systems training and cognitive style on decision process attributes. Journal of Management Information Systems, 3(2), 83–93. https://doi.org/10.1080/07421222.1986.11517764
Grønsund, T., & Aanestad, M. (2020). Augmenting the algorithm: Emerging human-in-the-loop work configurations. The Journal of Strategic Information Systems, 29(2), 101614. https://doi.org/10.1016/j.jsis.2020.101614
Grove, W. M., & Meehl, P. E. (1996). Comparative efficiency of informal (subjective, impressionistic) and formal (mechanical, algorithmic) prediction procedures: The clinical–statistical controversy. Psychology, Public Policy, and Law, 2(2), 293–323. https://doi.org/10.1037/1076-8971.2.2.293
Grove, W. M., Zald, D. H., Lebow, B. S., Snitz, B. E., & Nelson, C. (2000). Clinical versus mechanical prediction: A meta-analysis. Psychological Assessment, 12(1), 19–30. https://doi.org/10.1037/1040-3590.12.1.19
Hafenbrädl, S., Waeger, D., Marewski, J. N., & Gigerenzer, G. (2016). Applied decision making with fast-and-frugal heuristics. Journal of Applied Research in Memory and Cognition, 5, 215–231. https://doi.org/10.1016/j.jarmac.2016.04.011
Hahn, U., & Harris, A. J. L. (2014). What does it mean to be biased: Motivated reasoning and rationality. In Psychology of learning and motivation (Vol. 61, pp. 41–102). Elsevier. https://doi.org/10.1016/B978-0-12-800283-4.00002-2
Hammond, K. R. (1955). Probabilistic functioning and the clinical method. Psychological Review, 62(4), 255–262.
Highhouse, S. (2008). Stubborn reliance on intuition and subjectivity in employee selection. Industrial and Organizational Psychology, 1(3), 333–342. https://doi.org/10.1111/j.1754-9434.2008.00058.x
Hilbert, M., & Lopez, P. (2011). The World’s technological capacity to store, communicate, and compute information. Science, 332(6025), 60–65. https://doi.org/10.1126/science.1200970
Hindman, M. (2015). Building better models: Prediction, replication, and machine learning in the social sciences. The Annals of the American Academy of Political and Social Science, 659(1), 48–62. https://doi.org/10.1177/0002716215570279
Holst, A. (2021). Volume of data/information created, captured, copied, and consumed worldwide from 2010 to 2025. Statista. Retrieved from https://www.statista.com/statistics/871513/worldwide-data-created/
Holt, R. R. (1958). Clinical and statistical prediction: A reformulation and some new data. The Journal of Abnormal and Social Psychology, 56(1), 1–12. https://doi.org/10.1037/h0041045
Holt, R. R. (1970). Yet another look at clinical and statistical prediction: Or, is clinical psychology worthwhile? American Psychologist, 25(4), 337–349. https://doi.org/10.1037/h0029481
Holt, R. R. (1986). Clinical and statistical prediction: A retrospective and would-be integrative perspective. Journal of Personality Assessment, 50(3), 376–386. https://doi.org/10.1207/s15327752jpa5003_7
Hovey, H. B., & Stauffacher, J. C. (1953). Intuitive versus objective prediction from a test. Journal of Clinical Psychology, 9(4), 349–351.
Jarrahi, M. H. (2018). Artificial intelligence and the future of work: Human-AI symbiosis in organizational decision making. Business Horizons, 61(4), 577–586. https://doi.org/10.1016/j.bushor.2018.03.007
Joyce, J. (2003). Bayes’ Theorem. In The Stanford Encyclopedia of philosophy (Fall 2021). Retrieved from https://plato.stanford.edu/archives/fall2021/entries/bayes-theorem/.
Jung, M., & Seiter, M. (2021). Towards a better understanding on mitigating algorithm aversion in forecasting: An experimental study. Journal of Management Control. https://doi.org/10.1007/s00187-021-00326-3
Jussupow, E., Benbasat, I., & Heinzl, A. (2020). Why are we averse towards algorithms? A comprehensive literature review on algorithm aversion. ECIS 2020 Proceedings, 2020, 1–18.
Kahn, B. E., & Baron, J. (1995). An exploratory study of choice rules favored for high-stakes decisions. Journal of Consumer Psychology, 4(4), 305–328.
Kahneman, D. (2013). Thinking, fast and slow (1st ed.). Farrar, Straus & Giroux.
Kahneman, D., & Klein, G. (2009). Conditions for intuitive expertise: A failure to disagree. American Psychologist, 64(6), 515–526. https://doi.org/10.1037/a0016755
Klein, G. (1993). A recognition-primed decision (RPD) model of rapid decision making. In Decision making in action: Models and methods (pp. 138–147). Ablex Publishing.
Klein, G. (1997). Developing expertise in decision making. Thinking & Reasoning, 3(4), 337–352. https://doi.org/10.1080/135467897394329
Klein, G. (2008). Naturalistic decision making. Human Factors: The Journal of the Human Factors and Ergonomics Society, 50(3), 456–460. https://doi.org/10.1518/001872008X288385
Kleinmuntz, B. (1990). Why we still use our heads instead of formulas: Toward an integrative approach. Psychological Bulletin, 107(3), 296.
Klimoski, R., & Jones, R. G. (2008). Intuiting the selection context. Industrial and Organizational Psychology, 1(3), 352–354. https://doi.org/10.1111/j.1754-9434.2008.00061.x
Knight, F. H. (1921). Risk, uncertainty, and profit. Houghton Mifflin.
Kuncel, N. R. (2008). Some new (and old) suggestions for improving personnel selection. Industrial and Organizational Psychology, 1(3), 343–346. https://doi.org/10.1111/j.1754-9434.2008.00059.x
Kuncel, N. R., Klieger, D. M., Connelly, B. S., & Ones, D. S. (2013). Mechanical versus clinical data combination in selection and admissions decisions: A meta-analysis. Journal of Applied Psychology, 98(6), 1060–1072. https://doi.org/10.1037/a0034156
Lawrence, M., Edmundson, R. H., & O’Connor, M. J. (1986). The accuracy of combining judgemental and statistical forecasts. Management Science, 32(12), 1521–1532. https://doi.org/10.1287/mnsc.32.12.1521
Lawrence, M., Goodwin, P., O’Connor, M., & Önkal, D. (2006). Judgmental forecasting: A review of progress over the last 25 years. International Journal of Forecasting, 22(3), 493–518. https://doi.org/10.1016/j.ijforecast.2006.03.007
Lazer, D., & Kennedy, R. (2015). What we can learn from the epic failure of Google flu trends. Wired. Retrieved from https://www.wired.com/2015/10/can-learn-epic-failure-google-flu-trends/
Lim, J. S., & Connor, M. O. (1996). Judgmental forecasting with interactive forecasting support systems. Decision Support Systems, 16, 339–357.
Lin, W., Kim, S. H., & Tong, J. (2022). What drives algorithm use? An empirical analysis of algorithm use in type 1 diabetes self-management. https://doi.org/10.2139/ssrn.3891832
Logg, J. M., Minson, J. A., & Moore, D. A. (2019). Algorithm appreciation: People prefer algorithmic to human judgment. Organizational Behavior and Human Decision Processes, 151, 90–103. https://doi.org/10.1016/j.obhdp.2018.12.005
Mackay, J. M., & Elam, J. J. (1992). A comparative study of how experts and novices use a decision aid to solve problems in complex knowledge domains. Information Systems Research, 3(2), 150–172. https://doi.org/10.1287/isre.3.2.150
McNemar, Q. (1955). Review of the book clinical versus actuarial prediction. American Journal of Psychology, 68, 510.
Meehl, P. E. (1954). Clinical versus statistical prediction: A theoretical analysis and a review of the evidence. University of Minnesota Press.
Meehl, P. E. (1957). When shall we use our heads instead of the formula? Journal of Counseling Psychology, 4(4), 268–273. https://doi.org/10.1037/h0047554
Meehl, P. E. (1986). Causes and effects of my disturbing little book. Journal of Personality Assessment, 50(3), 370–375. https://doi.org/10.1207/s15327752jpa5003_6
Miller, G. A. (1956). The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psychological Review, 101(2), 343–352. https://doi.org/10.1037/0033-295X.101.2.343
Mittelstadt, B. D., Allo, P., Taddeo, M., Wachter, S., & Floridi, L. (2016). The ethics of algorithms: Mapping the debate. Big Data & Society, 3(2), 205395171667967. https://doi.org/10.1177/2053951716679679
Montazemi, A. L. I. R. (1991). The impact of experience on the design of user interface. International Journal of Man-Machine Studies, 34(5), 731–749.
Muir, B. M. (1987). Trust between humans and machines, and the design of decision aids. International Journal of Man-Machine Studies, 27(5–6), 527–539. https://doi.org/10.1016/S0020-7373(87)80013-5
Newell, S., & Marabelli, M. (2015). Strategic opportunities (and challenges) of algorithmic decision-making: A call for action on the long-term societal effects of ‘datification’. The Journal of Strategic Information Systems, 24(1), 3–14. https://doi.org/10.1016/j.jsis.2015.02.001
Önkal, D., Goodwin, P., Thomson, M., Gonul, S., & Pollock, A. (2009). The relative influence of advice from human experts and statistical methods on forecast adjustments. Journal of Behavioral Decision Making, 22, 390–409. https://doi.org/10.1002/bdm.637
Pankoff, L. D., & Roberts, H. V. (1968). Bayesian synthesis of clinical and statistical prediction. Psychological Bulletin, 70(6), 762–773. https://doi.org/10.1037/h0026831
Parasuraman, R., & Manzey, D. H. (2010). Complacency and bias in human use of automation: An attentional integration. Human Factors: The Journal of the Human Factors and Ergonomics Society, 52(3), 381–410. https://doi.org/10.1177/0018720810376055
Parasuraman, R., Sheridan, T. B., & Wickens, C. D. (2000). A model for types and levels of human interaction with automation. IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans, 30(3), 286–297. https://doi.org/10.1109/3468.844354
Partnership on AI. (2019). Report on algorithmic risk assessment tools in the U.S. Criminal Justice System. Partnership on AI. Retrieved from https://partnershiponai.org/paper/report-on-machine-learning-in-risk-assessment-tools-in-the-u-s-criminal-justice-system/
Pescetelli, N., Rutherford, A., & Rahwan, I. (2021). Modularity and composite diversity affect the collective gathering of information online. Nature Communications, 12(1), 3195. https://doi.org/10.1038/s41467-021-23424-1
Prahl, A., & Van Swol, L. (2017). Understanding algorithm aversion: When is advice from automation discounted? Journal of Forecasting, 36, 691–702. https://doi.org/10.1002/for.2464
Premack, D., & Woodruff, G. (1978). Does the chimpanzee have a theory of mind? Behavioral and Brain Sciences, 1(4), 515–526. https://doi.org/10.1017/S0140525X00076512
Promberger, M., & Baron, J. (2006). Do patients trust computers? Journal of Behavioral Decision Making, 19(5), 455–468. https://doi.org/10.1002/bdm.542
Rebitschek, F. G., Gigerenzer, G., & Wagner, G. G. (2021). People underestimate the errors made by algorithms for credit scoring and recidivism prediction but accept even fewer errors. Scientific Reports, 11(1), 20171. https://doi.org/10.1038/s41598-021-99802-y
Renier, L. A., Schmid Mast, M., & Bekbergenova, A. (2021). To err is human, not algorithmic – Robust reactions to erring algorithms. Computers in Human Behavior, 124, 106879. https://doi.org/10.1016/j.chb.2021.106879
Rosenberg, L., Pescetelli, N., & Willcox, G. (2017). Artificial Swarm Intelligence amplifies accuracy when predicting financial markets. In 2017 IEEE 8th Annual Ubiquitous Computing, Electronics and Mobile Communication Conference (UEMCON), pp. 58–62. https://doi.org/10.1109/UEMCON.2017.8248984.
Sage, A. P. (1981). Behavioral and organizational considerations in the design of information systems and processes for planning and decision support. IEEE Transactions on Systems, Man, and Cybernetics, 11(6), 640–678.
Sanders, G. L., & Courtney, J. F. (1985). A field study of organizational factors influencing DSS success. MIS Quarterly, 9(1), 77–93.
Sanders, N. R., & Manrodt, K. B. (2003). The efficacy of using judgmental versus quantitative forecasting methods in practice. Omega, 31(6), 511–522. https://doi.org/10.1016/j.omega.2003.08.007
Sawyer, J. (1966). Measurement and prediction, clinical and statistical. Psychological Bulletin, 66(3), 178–200. https://doi.org/10.1037/h0023624
Scherer, L. D., de Vries, M., Zikmund-Fisher, B. J., Witteman, H. O., & Fagerlin, A. (2015). Trust in deliberation: The consequences of deliberative decision strategies for medical decisions. Health Psychology, 34(11), 1090–1099. https://doi.org/10.1037/hea0000203
Sieck, W. R., & Arkes, H. A. L. R. (2005). The recalcitrance of overconfidence and its contribution to decision aid neglect. Journal of Behavioral Decision Making, 53, 29–53.
Sills, D., Driedger, N., Greaves, B., Hung, E., & Paterson, R. (2009). ICAST: A prototype thunderstorm nowcasting system focused on optimization of the human-machine mix. In Proceedings of the World Weather Research Programme Symposium on Nowcasting and Very Short Range Forecasting, pp. 2, 16.
Simon, H. A. (1955). A behavioral model of rational choice. The Quarterly Journal of Economics, 69(1), 99. https://doi.org/10.2307/1884852
Simon, H. A. (1956). Rational choice and the structure of the environment. Psychological Review, 63(2), 129–138. https://doi.org/10.1037/h0042769
The Parliamentary Office of Science and Technology. (2020). Interpretable machine learning. UK Parliament POST. Retrieved from https://researchbriefings.files.parliament.uk/documents/POST-PN-0633/POST-PN-0633.pdf.
Thurman, N., Moeller, J., Helberger, N., & Trilling, D. (2019). My friends, editors, algorithms, and I: Examining audience attitudes to news selection. Digital Journalism, 7(4), 447–469. https://doi.org/10.1080/21670811.2018.1493936
Tutt, A. (2017). An FDA for algorithms. Administrative Law Review, 69(1), 83–123. http://www.jstor.org/stable/44648608
Watts, D. J. (2017). Should social science be more solution-oriented? Nature Human Behaviour, 1(1), 0015. https://doi.org/10.1038/s41562-016-0015
Whitecotton, S. M. (1996). The effects of experience and confidence on decision aid reliance: A causal model. Behavioral Research in Accounting, 8, 194–216.
Wolfe, C., & Flores, B. (1990). Judgmental adjustment of earnings forecasts. Journal of Forecasting, 9(4), 389–405. https://doi.org/10.1002/for.3980090407
Worthen, B. (2003). Future results not guaranteed; contrary to what vendors tell you, computer systems alone are incapable of producing accurate forecasts. Retrieved from http://www2.cio.com.au/article/168757/future_results_guaranteed/.
Yaniv, I., & Hogarth, R. M. (1993). Judgmental versus statistical prediction: Information asymmetry and combination rules. Psychological Science, 4(1), 58–62. https://doi.org/10.1111/j.1467-9280.1993.tb00558.x
Zellner, M., Abbas, A. E., Budescu, D. V., & Galstyan, A. (2021). A survey of human judgement and quantitative forecasting methods. Royal Society Open Science, 8(2), 201187.
Zerilli, J., Knott, A., Maclaurin, J., & Gavaghan, C. (2019). Algorithmic decision-making and the control problem. Minds and Machines, 29(4), 555–578. https://doi.org/10.1007/s11023-019-09513-7
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Burton, J.W., Stein, MK., Jensen, T.B. (2023). Beyond Algorithm Aversion in Human-Machine Decision-Making. In: Seifert, M. (eds) Judgment in Predictive Analytics. International Series in Operations Research & Management Science, vol 343. Springer, Cham. https://doi.org/10.1007/978-3-031-30085-1_1
Download citation
DOI: https://doi.org/10.1007/978-3-031-30085-1_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-30084-4
Online ISBN: 978-3-031-30085-1
eBook Packages: Business and ManagementBusiness and Management (R0)