Notes
Greene et al. (2009). The study offers a solution to the descriptive part of the trolley problem (which principles govern our responses to trolley scenarios?), not to its moral part (which responses are correct?).
Greene et al. (2009, p. 365).
Musen and Greene, MS; see Singer (1972).
Uhlmann et al. (2009).
For another study of this sort, see Gino et al. (2010).
The moral (ir)relevance of intention, the second of the two factors identified by Greene and colleagues, is less clear (see Greene 2014, pp. 720–721). Note also that Greene and Musen’s study was challenged by Nagel and Waldmann (2013). They found that the apparent distance effect is really due to various confounding variables, such as informational directness, group membership and the relative efficaciousness of one’s helping efforts. But as Greene rightly points out, at least some of these factors are morally irrelevant, too (Greene 2013, p. 378).
Greene (2016a, p. 176).
Note that Uhlmann and colleagues confirmed in a pre-test that ethnicity and nationality are typically not considered morally relevant.
As previously observed by Kumar and Campbell (2012, pp. 317–318).
At one point, Greene seems reluctant to explicitly endorse this assumption (Greene 2014, p. 713), but he has to if the argument is to be a precise argument from irrelevance against deontology. Due to the mentioned symmetry, he cannot attack the deontological intuition directly by claiming that it, but not the consequentialist response, is triggered by the irrelevant factor. For whenever our case-specific responses vary in response to an irrelevant factor, both responses are sensitive to this irrelevant factor. The deontological intuition is triggered by the irrelevant fact that the victim is pushed rather than killed by hitting a switch (or black rather than white). But the consequentialist intuition is likewise triggered by the irrelevant fact that the victim is killed by hitting a switch rather than pushed (or white rather than black).
Greene (2010, p. 21), my emphasis.
Greene (2008, p. 39). Elsewhere, he defines deontological judgments “as ones that are naturally justified in deontological terms (in terms of rights, duties, etc.) and that are more difficult to justify in consequentialist terms”, and consequentialist judgments “as ones that are naturally justified in consequentialist terms (by impartial cost–benefit reasoning) and that are more difficult to justify in deontological terms” (Greene 2014, p. 699). In this paper, I go along with Greene’s definition.
Kumar and Campbell (2012, p. 318).
Kumar and Campbell (2012, p. 322).
Singer (1972) and Unger (1997). Singer’s argument may be an example of a successful precise argument from moral irrelevance. Surely, no one would want to suggest that we should resolve the conflict by concluding that we must not save the drowning child. See Campbell and Kumar (2012) for a related discussion.
Unger has, however, “[i]nformally and intermittently […] asked many students, colleagues and friends” for their intuitions (Unger 1997, p. 31).
Rini (2013, p. 267). Notice, though, that the morally irrelevant factors that Rini discusses also include factors that are external to the moral scenarios themselves, such as framing effects and psychological manipulation. I return to this point below.
Unger (1997, pp. 11–12).
There is the third possibility of claiming that case-specific intuitions have priority over general ones. Preservationists like Frances Kamm are associated with this view. Ethical particularists, who do not believe in moral principles in the first place, are naturally inclined towards this view, too. While in principle conceivable, this would entail that arguments from moral irrelevance do not work at all, as they require that an intuition at a higher level of generality can override the case-specific intuitions. The third possibility is therefore not an option for proponents of arguments from moral irrelevance. These three ways of understanding the relation between general and case-specific intuitions are also distinguished by Kagan (1998, pp. 13–14) and Kamm (1993, pp. 5–7).
Nye (2015).
Singer (1972, p. 236). The principle Singer refers to is: “[I]f it is in our power to prevent something very bad from happening, without thereby sacrificing anything of comparable moral importance, we ought, morally, to do it.” (Singer 1972, p. 231). This raises an interpretative question: Is the duty to give to charity entailed by our obligation to save the drowning child and the moral irrelevance of spatial distance? Or is it directly entailed by the above principle (which would render the other argument obsolete)? I won’t address this interpretative question here (refer e.g. to Nye 2015, p. 630). Singer is inspired by Henry Sidgwick (1981), another utilitarian who favors intuitions at a high level of generality.
This is not to deny that there are many uncontroversial cases, in which Greene’s deontological/consequentialist distinction does not apply. He explains: “When it comes to uncontroversial moral questions, these terms have little meaning. Disapproving of child abuse is both ‘characteristically consequentialist’ and ‘characteristically deontological’, or neither—take your pick. […] These two ‘characteristically’ labels are not very meaningful outside the context of moral dilemmas in which considerations about rights and duties, at least superficially, appear to conflict with an impartial cost–benefit analysis.” (Greene 2016b, p. 179) But ethical inquiry is all about truly controversial questions that involve precisely such a conflict, and there are therefore plenty genuine ‘characteristically deontological’ intuitions that challenge consequentialism.
Kumar and Campbell (2012, p. 313).
Liao et al. (2012), Petrinovich and O'Neill (1996), Schnall et al. (2008), Schwitzgebel and Cushman (2012, 2015), Tobia et al. (2013a, b), Wheatley and Haidt (2005) and Wiegmann et al. (2012). In light of the replication crisis, findings like these should be taken with a grain of salt (see e.g. Landy and Goodwin 2015).
In fact, even Kamm, the leader of the preservationist camp, acknowledges, if only by lip-service, that a moral principle derived from case-specific intuitions may stand in need of further validation: We must “consider the principle on its own, to see if it expresses some plausible value or conception of the person or relations between persons. This is necessary to justify it as a correct principle, one that has normative weight, not merely one that makes all of the case judgments cohere.” (Kamm 2007, p. 5, see also pp. 346, 379). In practice, Kamm shows relatively little interest in whether a principle considered on its own is plausible (Nye 2015, p. 627).
I am here bracketing the problem that intention does seem to be a morally relevant factor.
See again Campbell and Kumar (2012) for a related discussion.
Greene (2014).
Greene (2008, p. 68).
On this see Königs (2018a, p. 197).
References
Berker, S. (2009). The normative insignificance of neuroscience. Philosophy & Public Affairs,37(4), 293–329.
Bruni, T., Mameli, M., & Rini, R. A. (2014). The science of morality and its normative implications. Neuroethics,7(2), 159–172.
Campbell, R., & Kumar, V. (2012). Moral reasoning on the ground. Ethics,122(2), 273–312.
DePaul, M. (2006). Intuitions in moral inquiry. In D. Copp (Ed.), The Oxford Handbook of Ethical Theory (pp. 595–623). New York: Oxford University Press.
Ernst, Z. (2007). The liberationists’ attack on moral intuitions. American Philosophical Quarterly,44(2), 129–142.
Gino, F., Shu, L. L., & Bazerman, M. H. (2010). Nameless + harmless = blameless: When seemingly irrelevant factors influence judgment of (un)ethical behavior. Organizational Behavior and Human Decision Processes,111(2), 93–101.
Greene, J. (2008). The secret joke of Kant’s soul. In W. Sinnott-Armstrong (Ed.), Moral psychology (Vol. 3): The neuroscience of morality: Emotion, brain disorders, and development (pp. 35–80). Cambridge, MA: MIT Press.
Greene, J. (2010). Notes on ‘The normative insignificance of neuroscience’ by Selim Berker. (unpublished manuscript)
Greene, J. (2013). Moral tribes: Emotion, reason, and the gap between us and them. New York: Penguin Press.
Greene, J. (2014). Beyond point-and-shoot morality: Why cognitive (Neuro)science matters for ethics. Ethics,124(4), 695–726.
Greene, J. (2016a). Solving the Trolley problem. In J. Sytsma & W. Buckwalter (Eds.), A Companion to Experimental Philosophy (pp. 175–189). Malden, MA: Blackwell.
Greene, J. (2016b). Reply to Driver and Darwall. In S. M. Liao (Ed.), Moral brains: The neuroscience of morality (pp. 170–181). New York: Oxford University Press.
Greene, J., Cushman, F. A., Stewart, L. E., Lowenberg, K., Nystrom, L. E., & Cohen, J. D. (2009). Pushing moral buttons: The interaction between personal force and intention in moral judgment. Cognition,111(3), 364–371.
Huemer, M. (2005). Ethical Intuitionism. Basingstoke: Palgrave Macmillan.
Huemer, M. (2008). Revisionary intuitionism. Social Philosophy and Policy,25(1), 368–392.
Kagan, S. (1989). The limits of morality. Oxford: Oxford University Press.
Kagan, S. (1998). Normative ethics. Boulder, CO: Westview.
Kagan, S. (2016). Solving the trolley problem. In E. Rakowski (Ed.), The Trolley problem mysteries (pp. 151–168). New York: Oxford University Press.
Kahane, G. (2013). The armchair and the trolley: An argument for experimental ethics. Philosophical Studies,162(2), 421–445.
Kamm, F. (1993). Morality, mortality: Volume 1: Death and Whom to Save from It. New York: Oxford University Press.
Kamm, F. (2007). Intricate ethics: Rights, responsibilities, and permissible harm. Oxford: Oxford University Press.
Kamm, F. (2009). Neuroscience and moral reasoning: A note on recent research. Philosophy & Public Affairs,37(4), 330–345.
Kauppinen, A. (2014). Ethics and empirical psychology—Critical remarks one empirical informed ethics. In M. Christen, C. Van Schaik, J. Fischer, M. Huppenbauer, & C. Tanner (Eds.), Empirically informed ethics: Morality between facts and norms (pp. 279–305). Cham: Springer.
Königs, P. (2018a). On the normative insignificance of neuroscience and dual-process theory. Neuroethics,11(2), 195–209.
Königs, P. (2018b). Two types of debunking arguments. Philosophical Psychology,31(3), 383–402.
Kumar, V., & Campbell, R. (2012). On the normative significance of experimental moral psychology. Philosophical Psychology,25(3), 311–330.
Landy, J. F., & Goodwin, G. P. (2015). Does incidental disgust amplify moral judgment? A meta-analytic review of experimental evidence. Perspectives on Psychological Science,10(4), 518–536.
Liao, M. S., Wiegmann, A., Alexander, J., & Vong, G. (2012). Putting the trolley in order: Experimental philosophy and the loop case. Philosophical Psychology,25(5), 661–671.
McMahan, J. (2013). Moral intuition. In H. LaFollette & I. Persson (Eds.), The Blackwell Guide to Ethical Theory (2nd ed., pp. 103–120). Malden, MA: Blackwell.
Mihailov, E. (2016). Is deontology a moral confabulation? Neuroethics,9(1), 1–13.
Musen, J. D., & Greene, J. (MS). Mere spatial distance weakens perceived moral obligation to help those in desperate need. (unpublished manuscript)
Nagel, J., & Waldmann, M. R. (2013). Deconfounding distance effects in judgments of moral obligation. Journal of Experimental Psychology,39(1), 237–252.
Nye, H. (2015). Directly plausible principles. In C. Daly (Ed.), The Palgrave Handbook of Philosphical Methods. New York: Palgrave Macmillan.
Paulo, N. (2019). In search of Greene’s argument. Utilitas,31(1), 38–58.
Petrinovich, L., & O’Neill, P. (1996). Influence of wording and framing effects on moral intuitions. Ethology and Sociobiology,17(3), 145–171.
Rini, R. A. (2013). Making psychology normatively significant. The Journal of Ethics,17(3), 257–274.
Sandberg, J., & Juth, N. (2011). Ethics and intuitions: A reply to Singer. The Journal of Ethics,15(3), 209–226.
Sauer, H. (2012). Morally irrelevant factors: What’s left of the dual process-model of moral cognition? Philosophical Psychology,25(6), 783–811.
Sauer, H. (2018). Debunking Arguments in Ethics. Cambridge: Cambridge University Press.
Schnall, S., Haidt, J., Clore, G. L., & Jordan, A. H. (2008). Disgust as embodied moral judgment. Personality and Social Psychology Bulletin,34(8), 1096–1109.
Schwitzgebel, E., & Cushman, F. (2012). Expertise in moral reasoning? Order effects on moral judgment in professional philosophers and non-philosophers. Mind and Language,27(2), 135–153.
Schwitzgebel, E., & Cushman, F. (2015). Philosophers’ biased judgments persist despite training, expertise and reflection. Cognition,141, 127–137.
Sidgwick, H. (1981). The methods of ethics. Indianapolis: Hackett Publishing.
Singer, P. (1972). Famine, affluence, and morality. Philosophy & Public Affairs,1(3), 229–243.
Sinnott-Armstrong, W. (2008). Framing moral intuitions. In W. Sinnott-Armstrong (Ed.), Moral psychology: The cognitive science of morality: Intuition and diversity (Vol. 2, pp. 47–76). Cambridge, MA: MIT Press.
Tobia, K., Buckwalter, W., & Stich, S. (2013a). Moral intuitions: Are philosophers experts? Philosophical Psychology,26(5), 629–638.
Tobia, K., Chapman, G. B., & Stich, S. (2013b). Cleanliness is next to morality, even for philosophers. Journal of Consciousness Studies,20(11–12), 195–205.
Uhlmann, E. L., Pizarro, D. A., Tannenbaum, D., & Ditto, P. H. (2009). The motivated use of moral principles. Judgment and Decision Making,4(6), 476–491.
Unger, P. (1997). Living high and letting die: Our illusion of innocence. New York: Oxford University Press.
Weinberg, J. M., Gonnerman, C., Buckner, C., & Alexander, J. (2010). Are philosophers expert intuiters? Philosophical Psychology,23(3), 331–355.
Wheatley, T., & Haidt, J. (2005). Hypnotic disgust makes moral judgments more severe. Psychological Science,16(10), 780–784.
Wiegmann, A., Horvath, J., & Meyer, K. (forthcoming). Intuitive expertise and irrelevant options. In T. Lombrozo, J. Knobe, & S. Nichols (Eds.), Oxford Studies in Experimental Philosophy (Vol. 3). Oxford: Oxford University Press
Wiegmann, A., Okan, Y., & Nagel, J. (2012). Order effects in moral judgment. Philosophical Psychology,25(6), 813–836.
Wilson, T. D. (2002). Strangers to ourselves: Discovering the adaptive unconscious. Cambridge, MA: Harvard University Press.
Acknowledgements
I wish to thank the anonymous referee, Chiara Brozzo, Christian Seidel, Emilian Mihailov, Hanno Sauer, Irina Schumski, Katharina Brecht, Leo Menges, Michael W. Schmidt, Nora Heinzelmann, Norbert Paulo and audiences in Karlsruhe, Porto, Saarbrücken and Tübingen for their valuable comments.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Königs, P. Experimental ethics, intuitions, and morally irrelevant factors. Philos Stud 177, 2605–2623 (2020). https://doi.org/10.1007/s11098-019-01330-z
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11098-019-01330-z