Notes
This is not to say that political philosophers could not help with this question. There are lots of questions that should have as their research centre some other department, but to which philosophers can usefully help. Indeed, the examples from economic methodology and evolutionary explanation I just mentioned are two more such questions.
Philosophers sometimes understate the importance of independent checks. We can know a scale is working, but if we want to check its reliability we don’t use it, we use something else. I suspect that a certain amount of theory-independence is part of the explanation of the value of intuitions.
Relatedly, I have not seen Liverpool get awarded an undeserved free kick for about that long.
There is an ambiguity in Cappelen’s text that I am not sure I am interpreting the right way. Let’s assume that someone intuits that in a particular case, \( c \) doesn’t cause \( e \). Call the content of that intuition, i.e., what is intuited, \( p_{d} \). And call the proposition that the person has this intuition, i.e., the event of the intuiting, \( p_{g} \). Plausibly both \( p_{d} \) and \( p_{g} \) could be evidence in the right cases, though most of the time the salient evidence will be \( p_{d} \). I think \( p_{d} \) can be an unjustified justifier in the sense that other beliefs, e.g., that a particular theory of causation is false, can be justified on the basis of \( p_{d} \), but no other beliefs the agent has justify \( p_{d} \). But you might want a stronger sense of ‘unjustified’, where it means not just not justified by anything else, but not justified at all. I think in these kinds of cases, \( p_{d} \) is justified, just not justified by anything else. And the justification is, as I’ll get to below, strong but fragile. If when Cappelen says that intuitions, according to Centrality, are unjustified justifiers he means that the belief that \( p_{d} \) is unjustified, then I am not defending Centrality. I just mean that the agent need not have any other mental states which justify the belief \( p_{d} \), or indeed any access to anything that justifies \( p_{d} \). But for all that it might be that the belief that \( p_{d} \) is justified, and the grounds for the justification include what the agent learned about causation as a child, plus perhaps her competence in distinguishing causes from non-causes.
I am simplifying a little here. My preferred position is that intuiteds provide strong but fragile evidence, while intuitings provide weak but resilient evidence. The reason this is relevant is related to footnote 7.
At one point in Ben Levinstein’s doctoral dissertation (Levinstein 2013), he considers whether there’s a general rule for deciding which of two conflicting sources we can trust. There turns out to be very little in general one can say. In particular, trust the more reliable source turns out not in general to be good advice. If sources have characteristic errors, it might be that given what the two sources have said, it is better on this occasion to trust the less reliable source, because the verdicts the sources deliver provide evidence that we are seeing one of the characteristic errors of the more reliable source. It takes more space than I have here to fill in the details of this argument, and most of the details I’d include would be Levinstein’s not mine. But here’s the big conclusion. Assume that intuitions are often wrong, but rarely dramatically wrong. The reason for that is that heuristics are bad at getting things exactly right, and good at getting in the ballpark. And that careful reasoning is often right, but sometimes dramatically wrong. This is trickier to motivate, but I think true. Then when intuition dramatically diverges from theory, and we don’t have independent reason to think that intuition is mistaken about the kind of case that’s in question, we should trust the intuition more than the theory.
There is interesting work to be done on the relative role of intuitions and arguments about principles, but I’m going to leave that for another day, and focus here on cases. The principles/cases distinction can be a bit slippery, but paradigm cases are easy to identify, and we’ll be working with fairly paradigmatic cases here.
References
Carruthers, P. (1990). The metaphysics of the Tractatus. Cambridge: Cambridge University Press.
Carruthers, P. (2011). The opacity of mind: An integrative theory of self-knowledge. Oxford: Oxford University Press.
Chudnoff, E. (2011). What should a theory of knowledge do? Dialectica, 65(4), 561–579. doi:10.1111/j.1746-8361.2011.01285.x.
Devitt, M. (2010). Experimental semantics. Philosophy and Phenomenological Research, 82(2), 418–435.
Devitt, M., & Sterelny, K. (1987). Language and reality: An introduction to the philosophy of language. Cambridge, MA: MIT Press.
Fodor, J. A. (1983). The modularity of mind. Cambridge, MA: MIT Press.
Fodor, J. (2000). The mind doesn’t work that way. Cambridge, MA: MIT Press.
Friedman, M. (1953). The methodology of positive economics. In Essays in positive economics (pp. 3–43). Chicago: University of Chicago Press.
Gladwell, M. (2005). Blink: The power of thinking without thinking. New York: Little Brown.
Goldman, A. I. (1976). Discrimination and perceptual knowledge. The Journal of Philosophy, 73(20), 771–791.
Gopnik, A. (2009). The philosophical baby: What children’s minds tell us about truth, love, and the meaning of life. New York: Farrar, Straus and Giroux.
Harman, G. (1973). Thought. Princeton: Princeton University Press.
Harman, G. (1986). Change in view. Cambridge, MA: Bradford.
Hetherington, S. (2001). Good knowledge, bad knowledge: On two dogmas of epistemology. Oxford: Oxford University Press.
Ichikawa, J., Ishani, M., & Brian, W. (2012). In defence of a Kripkean dogma. Philosophy and Phenomenological Research, 85(1), 56–68. doi:10.1111/j.1933-1592.2010.00478.x.
Kahneman, D. (2011). Thinking fast and slow. New York: Farrar, Straus and Giroux.
Kilkarni, S., & Harman, G. (2011). An elementary introduction to statistical learning theory. Hoboken, NJ: Wiley.
Klein, G. A. (1999). Sources of power. Cambridge, MA.: MIT Press.
Levinstein, B. (2013). Accuracy as epistemic utility. PhD Dissertation, Rutgers University.
Lewis, D. (1981). What puzzling Pierre does not believe. Australasian Journal of Philosophy, 59(3), 283–289. doi:10.1080/00048408112340241.
Lewis, D. (1988). The trap’s dilemma. Australasian Journal of Philosophy, 66(2), 220–223.
Lewis, D. (1996). Elusive knowledge. Australasian Journal of Philosophy, 74(4), 549–567. doi:10.1080/00048409612347521.
Lewis, D. (1997a). Finkish dispositions. Philosophical Quarterly, 47(187), 143–158.
Lewis, D. (1997b). Naming the colours. Australasian Journal of Philosophy, 75(3), 325–342. doi:10.1080/00048409712347931.
Lewis, D. (2004a). Causation as influence. In J. Collins, N. Hall, & L. A. Paul (Eds.), Causation and counterfactuals (pp. 75–106). Cambridge: MIT Press.
Lewis, D. (2004b). Void and object. In J. Collins, N. Hall, & L. A. Paul (Eds.), Causation and counterfactuals (pp. 277–290). Cambridge: MIT Press.
Machery, E., Mallon, R., Nichols, S., & Stich, S. (2012). If folk intuitions vary, then what? Philosophy and Phenomenological Research. doi:10.1111/j.1933-1592.2011.00555.x.
Nagel, J. (2007). Epistemic intuitions. Philosophy Compass, 2(6), 792–819. doi:10.1111/j.1747-9991.2007.00104.x.
Nagel, J. (2013). Intuitions and experiments: A defense of the case method in epistemology. Philosophy and Phenomenological Research, 85(3), 495–527.
Nesbø, J. (2009). The Redeemer. (trans: Bartlett, D.). London: Vintage.
Nolan, D. (2005). David lewis. Chesham: Acumen Publishing.
Pryor, J. (2000). The sceptic and the dogmatist. Noûs, 34(4), 517–549. doi:10.1111/0029-4624.00277.
Ruetsche, L. (2011). Interpreting quantum theories. Oxford: Oxford University Press.
Sainsbury, M. (1995). Vagueness, ignorance and margin for error. British Journal for the Philosophy of Science, 46, 589–601.
Sartwell, C. (1992). Why knowledge is merely true belief. Journal of Philosophy, 89(4), 167–180.
Schwarz, W. (2009). David Lewis: Metaphysik und analyse. Paderborn: Mentis-Verlag.
Sterelny, K. (2012). The evolved apprentice: How evolution made humans unique. Cambridge, MA.: Bradford.
Sugden, R. (2000). Credible worlds: The status of theoretical models in economics. Journal of Economic Methodology, 7(1), 1–31. doi:10.1080/135017800362220.
Sugden, R. (2009). Credible worlds, capacities and mechanisms. Erkenntnis, 70(1), 3–27. doi:10.1007/s10670-008-9134-x.
Tribe, K. (2002). The Cambridge economics tripos 1903–1955 and the training of economists. The Manchester School, 68(2), 222–248. doi:10.1111/1467-9957.00191.
Warfield, T. A. (2005). Knowledge from falsehood. Philosophical Perspectives, 19, 405–416. doi:10.1111/j.1520-8583.2005.00067.x.
Weatherson, B. (2003a). Many many problems. Philosophical Quarterly, 53(213), 481–501.
Weatherson, B. (2003b). What good are counterexamples? Philosophical Studies, 115(1), 1–31. doi:10.1023/A:1024961917413.
Weatherson, B. (2004). Luminous margins. Australasian Journal of Philosophy, 82(3), 373–383.
Whewell, W. (1840). The philosophy of the inductive sciences, founded upon their history. London: John W. Parker.
Williamson, T. (1994). Vagueness. London: Routledge.
Williamson, T. (2000). Knowledge and its limits. Oxford: Oxford University Press.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Weatherson, B. Centrality and marginalisation. Philos Stud 171, 517–533 (2014). https://doi.org/10.1007/s11098-014-0289-9
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11098-014-0289-9