1 Section 1

Models of the attitudes span a range from the not-so-structured — merely sets of hypotheses compatible with what the agent Vs — to commitment to particular mental symbols, and access to as much structure as they may have.Footnote 1 This paper identifies interesting intermediate models, that posit some structure but are noncommittal about how it’s concretely realized in our cognitive architecture. For some explanatory purposes this seems a preferable level of abstraction. States as described by these models can then be shared between creatures who realize that structure differently.

The kind of structure I target is not Boolean structure, such as a difference between believing P and believing not-PP. Rather I aim to model the property attitudes have when they purport to be about a common object (or common sequence of objects), as when Hob believes that someone i blighted Bob’s mare, and wonders whether she i killed Cob’s sow.Footnote 2

The framework I’ll present has some resemblance to what Recanati and others have formulated in terms of “mental files,” the topic of this volume. What follows could be understood as one way to (abstractly) develop some of their framework. I’ll address some possible differences in Section 8. My strategy even more closely resembles Ninan’s “multi-centered” view of mental contentFootnote 3; and as I’ll explain, there is an underlying unity with Heck’s 2012.

A deeply-entrenched idea in theorizing about mental states is that they can be individuated by the pair of their content plus the attitude we hold towards it. What exactly that idea requires is unclear, because (at least) what counts as content is up for negotiation. But despite some pockets of readiness when externalism is discussed to allow that identity of content may be non-transparent, I think most philosophers are reluctant to allow thoughts with the same content, or the contents themselves, to differ merely numerically, as Max Black’s spheres are alleged to differ. A subsidiary aim of this paper is to join Heck 2012 in encouraging more sympathy for that possibility, one that my framework is well-suited to model.Footnote 4 On some natural ways to talk of “content,” we could call this the possibility of believing the content that P (simultaneously) twice, rather than once. I’ll present cases I find to speak most compellingly for this in Section 8. The earlier parts of this paper are devoted to developing the models that can even represent the idea.Footnote 5

2 Section 2

Alice believes her local baseball team has several Bobs on it. In fact, they only have one Bob. Maybe this happened because Alice met that same player repeatedly, and didn’t realize she was getting reacquainted with someone already familiar. Or maybe she only met him once, but through some cognitive glitch she mistakenly opened several “mental files” for that single player. Either way, let’s suppose she has now forgotten the specifics of how she met “each” Bob; and let’s suppose that the memories, information, and attitudes she associates with “each” of them is the same — or as close to this as possible. (We will discuss how close it is possible to get later.)

Presently, Alice is thinking how the voting for her team’s Most Valuable Player will go. There’s a thought she can entertain that we might articulate, on her behalf, like this:

  1. (1)

    Bob i will vote for Bob j .

where ij.Footnote 6 The indices are meant only to be suggestive; we’ll have to work out what it amounts to for Alice to have such a thought.

Even with only this suggestive start, already we can formulate an interesting question, whether Alice would also be in a position to entertain a distinct thought, articulable like this:

  1. (2)

    Bob j will vote for Bob i .

The story I’ve told certainly doesn’t put Alice in a position to have more justification for accepting one of these thoughts over the other; but that doesn’t settle whether there are two distinct thoughts available here for her to entertain. (If we think there aren’t, perhaps the way she’d have to express her single thought is One of the Bobs will vote for the other.)

Whatever our views on that question, though, most will want to distinguish the thought we’ve articulated with (1) from a thought articulable like this:

  1. (3)

    Bob i will vote for Bob i .

It’s only this last thought whose acceptance we’d expect to rationally commit Alice to the thought that anyone self-voted. And in my story, Alice doesn’t seem to be so committed. Yet she was only ever thinking of a single player throughout. It’s just that when thinking (1) — but not when thinking (3) — she’d unknowingly be thinking of this single player twice, or via two different presentations.

This is a familiar contrast; but there is some awkwardness finding a vocabulary perfectly suited to characterize it.

One phrase I just employed was “unknowingly thinking of a single player twice.” But how should we describe the contrasting case (3)? Shall we say that there Alice knowingly thinks of a single player twice? That over-intellectualizes what’s going on in case (3). Nothing I said required Alice to have any opinion about her own thoughts; and absent some controversial assumptions about the transparency or luminosity of introspection, it’s not obvious that if she did have opinions about her own thoughts, they’d have to be correct.

Another phrase I employed was “two different presentations.” That might be okay if we understand it to mean two numerically different presentations; but it’s too easily heard as implying that the presentations must differ in some qualitative, that is, not-merely numerical, way. And that further step should be a substantive one. As I mentioned above, I will be arguing that it should sometimes be resisted: Alice’s two presentations in case (1) needn’t differ except numerically. But even if I’m wrong, the fact that this can be intelligibly disputed speaks for favoring a vocabulary that doesn’t prejudice the question.

One recently popular way to characterize the difference between cases (1) and (3) is to say that in the latter but not the former, Alice thinks about Bob in a way that coordinates the two argument roles her thought ascribes to him.Footnote 7 This can also be misconstrued, but it’s the vocabulary I will adopt.

Fans of structure are willing to count thoughts as distinct even when their contents are “logically equivalent.” One form this can take is to say, here are different contents for thoughts to have:Footnote 8

(1′):

(λ x.x will vote for Bob) Bob

(3′):

(λ x.x will vote for x) Bob

If one’s sympathetic to that, then one might describe our case (1) as one where even if Alice thinks contents of form (1′), and others like it, she doesn’t think contents of form (3′). In case (3), she does think contents of form (3′).Footnote 9 That characterization doesn’t seem to commit you either way on the question whether Alice’s two thought-tokenings of Bob in (1) themselves have different contents. And so it might seem an attractively minimal way to capture the difference illustrated by our cases (1) and (3). (If you wanted to go on to distinguish case (1) from case (2), then you’d have to say things less minimal.)

I’ll defer for another discussion a detailed consideration of how much explanatory coverage the idea just sketched can achieve. I’ll just announce my conviction that the phenomena we’ve been intuitively characterizing extend beyond cases where they correlate with variable binding structure. To take just one example:

(4a):

Carol will vote for Bob, but Bob won’t.

is best understood as having the binding structure of:

(4b):

(λ x.x will vote for Bob) Carol, but ¬(\(\underline {\qquad \,\,}\) Bob)

where the \(\underline {\qquad \,\,}\) is occupied by the same predicate expressed by the lambda term on the left.Footnote 10 Rather than:

(4c):

(λ x.x will vote for Bob) Carol, but ¬ ((λ x x will vote for x) Bob)

or:

(4d):

(λ y (λ x.x will vote for y) Carol, but ¬(\(\underline {\qquad \,\,}\,y\))) Bob Footnote 11

Yet, the thought’s having the structure of (4b) should intuitively still be compatible with its being a “coordinated” way of thinking about Bob — one where the thinker uses only a single mental presentation for him.

So in this discussion, I will assume that the difference between coordinated and uncoordinatedFootnote 12 thinking can’t (always) be attributed to the binding structure of the predicates one’s thoughts ascribe. Instead, I will sketch a different way to represent this difference as grounded in the structure of one’s thinking.

3 Section 3

I’ve already declared that the models I’ll be developing are abstract, that is, less committal about their cognitive implementation than some alternatives. What’s also true is that I’ll only be presenting a general framework for modeling a subject’s thoughts. As we’ll see, the framework can be specifically deployed in different ways. Some of these ways are more finely structured (more “linguisticy”); others less finely structured (more “logicy”). Which of those variants is more appropriate will depend on the explanatory work the model is supposed to perform.

To get us started, though, it may help to sketch a fine-grained cousin of my view. This begins with the idea that thought contents have predicative and logical structure, as championed by King and others.Footnote 13 On their views, the content of Alice’s thought in case (1) would be something like the syntactic structure of the sentence we use to articulate it, except with the leaves labeled by objects and relations, rather than by expression-types:

figure a

To that base, we add the further refinement of “wires” connecting the two occurrences of Bob when Alice is thinking a coordinated thought:

figure b

and the absence of such “wires” (or “unwires”) when Alice is thinking an uncoordinated thought:

figure c

(Alternatively, one could tag all the leaves with indices, understanding leaves to be wired when they have the same object and index, and unwired when they have the same object but different indices.)

A natural question that arises is what is the relation between the content in the wired tree and the content:

figure d

Are they the same, at least in this simple case? Does either entail the other? Those are good questions; but we will not try to answer them now.

If a view of this sort allowed “wires” and “unwires” among all of the occurrences of Bob within and across all of Alice’s beliefs (and also her other attitudes), then it would turn out to coincide with the finest-grained versions of the general framework I’ll be proposing.Footnote 14

However, for many explanatory purposes we’ll want a notion of “content” that abstracts from some syntactic detail. For example, we may want to count the thoughts articulated by:

(5a):

Bob will vote for Carol; and Carol will vote for Bob.

(5b):

Carol will vote for Bob; and Bob will vote for Carol.

(5c):

Bob and Carol will vote for each other.

as having the same (coordinated) contents. We may want to count thoughts like these:

(6a):

Bob was afraid to vote.

(6b):

Bob feared voting.

(6c):

Voting frightened Bob.

as having the same contents, too, despite their varying syntax. The general framework I’ll propose can also be deployed in such coarser-grained ways, and indeed those are the variants I’m more sympathetic to.

Rather than syntactic tree-structures, we’ll instead be working with (a generalized form of) directed graphs. And rather than representing coordinated thoughts with “wires” between multiple occurrences of Bob, we’ll instead have Bob occurring just once in Alice’s thought-graph (for each presentation Alice has of him). The directed edges will have to “come back” to that single occurrence each time they represent a coordinated thinking about him.

Thus, when Alice thinks the coordinated thought she does in case (3), we’ll understand her to have thoughts with the following structure:

  1. (7)
    figure e

and when she thinks the uncoordinated thought she does in case (1), we’ll instead have something like this:

  1. (8)
    figure f

To develop this properly, though, we should first review some elementary graph theory.

4 Section 4

A graph is made up of one or more vertices joined by zero or more edges. In some applications, the edges are assumed to be undirected; in others, not. Graphs where all the edges are directed are often called digraphs. Graphs may or may not be connected; here is an unconnected digraph, with three components:

figure g

and some vertices may be isolated, as the leftmost vertex in the above diagram. But usually edges are not allowed to be disconnected in an analogous way; they are usually assumed to always begin and end at specific vertices. In some generalizations, though, that we will consider later, this requirement on edges can be relaxed.The edges in a graph may or may not exhibit cycles, as this digraph does:

figure h

But usually self-loops starting and ending at the same vertex are prohibited:

figure i

That’s not because there’s anything incoherent about self-loops; it’s just that for many customary applications, it’s easiest to assume these are excluded. Similarly, for many applications, it’s easiest to assume there is at most one edge (in a given direction) joining a given pair of vertices. In the following diagram:

figure j

the edges on the right would violate this requirement. Again, this is not because there’s anything incoherent about multiple edges joining a given pair of vertices: indeed, the famous Bridges of Königsberg problem, which launched the discipline of graph theory, makes essential use of such. It’s just that for many customary applications, it’s easiest to assume these are excluded. For our purposes, though, we will want to allow self-loops and multiple edges.Footnote 15

Another requirement we’ll want to relax is that edges must always join exactly two vertices. For some applications, one instead wants to work with structures like these:

figure k

Notice there is no vertex in the middle of the diagram. This graphlike structure has five vertices and a single edge that joins all of them. Graphlike structures of this sort are called hypergraphs. For these structures, we assume that each edge joins one or more vertices, rather than always exactly two.

The hypergraph in the preceding diagram had an undirected edge; we might instead designate some of the vertices it joins as initial and others as terminal, thus getting a directed hypergraph:

figure l

I want to refine these structures in a slightly different way. Instead of sorting an edge’s vertices merely as initial or terminal, let’s assign them a linear order. Thus we will instead have something like this hypergraph, with an ordered length 4 edge:

figure m

That’s to be distinguished from this simple digraph with a directed path of 4 edges of length 1:

figure n

Think of the latter structure as involving four different flight bookings; but the former as involving a single booking with three intermediate layovers. For our purposes, we will need the more general hypergraph structure that allows for both of these possibilities.

For some applications, we want to associate further information with a graph’s vertices or edges. For example, in this (undirected) graph the labels on the edges may represent the cost of traveling between the cities that label the vertices:

figure o

Notice we don’t hesitate to assign one and the same label to distinct edges. What’s trickier to understand is that we needn’t hesitate to assign one and the same label to distinct vertices, either. This is because the information labeling the vertex shouldn’t be thought of as constituting the vertex. Rather, we should just think of the vertex as constituted by its position in the graph structure.Footnote 16

For example, consider the phrase-structure tree of a sentence where the same word appears multiple times:

  1. (9)

    If dating Carol was stupid , living with her will be more stupid.

A phrase-structure tree can be thought of as a special kind of vertex-labelled digraph, with an additional left-to-right structure imposed on its leaves:

figure p

Notice that the single word-type “stupid” labels distinct vertices in this graph. (So too do the categories S, NP, and VP.)

Now, you might consider saying instead that it’s the distinct tokens of “stupid” in our inscription of sentence (9) that should label the vertices of its phrase-structure tree. I don’t know what account you’d then give of the syntax of (9)’s sentence-type. But this idea falters even as an account of the syntax of tokens. Consider the following inscription:

(9′):
figure q

This contains only a single concrete instance of the word “stupid,” yet it too should have the same syntactic type as displayed above. But here we only have a single token of “stupid” to go around. So whether it be type or token, inevitably we’ll have to admit that one and the same object sometimes labels distinct positions in a phrase-structure tree.

Some philosophers distinguish what I’ve called a token from what we call an occurrence of the word-type “stupid.” Even if that word has only a single concrete tokening in (9′), they’d say it has two occurrences, one for each of the syntactic positions where the word appears. I have no complaint with that. But don’t think that such occurrences are given to us as distinct objects in advance, out of which we could build the vertices of our phrase-structure tree. Rather, the distinctness of the occurrences instead already assumes the distinctness of those syntactic positions. And having been given the distinctness of the latter, there is no harm in letting it be the single word-type “stupid” that labels them both.Footnote 17

Hence, labeled digraphs like this:

  1. (10)
    figure r

should be clearly distinguished from labeled digraphs like this:

(11a):
figure s

or this:

(11b):
figure t

even when it’s one and the same object, Bob, which labels both of the vertices on the right-hand side of diagram (10). The graphs are different because Bob appears twice (that is, he labels two vertices) in (10) where he only appears once in (11a) and in (11b).Footnote 18

5 Section 5

So how can we use these tools to model Alice’s thoughts?

Let’s suppose she turns her thoughts from baseball to the denizens of Marvel Comics. She doesn’t realize that the individual named “Peter Parker” and the individual named “Spider-Man” are one and the same. Her thoughts may have a structure like this:

  1. (12)
    figure u

This represents Alice as thinking twice about a single individual (here designated as Peter), who she thinks of in qualitatively different ways. As presented one way, the individual seems to her to be hated by Otto, and (only) to bear the name “Spider-Man”; as presented another way, he seems to her (only) to bear the name “Parker.”

So far, we’re only modeling thoughts whose content involves the atomic predication of binary relations. Since Alice thinks about Otto standing in multiple relations to Spider-Man, we need to allow her mental graph to join the upper-left vertices with multiple labeled edges. We should also be prepared for her thoughts to join some vertices to themselves: for example, she may believe that Otto praises himself:

  1. (13)
    figure v

Here then would be Alice thinking in a coordinated way about Peter, that both Otto and he will vote him Most Inadequate Hero:

  1. (14)
    figure w

By contrast, here would be Alice thinking about Peter in an uncoordinated way:

  1. (15)
    figure x

This picture needs to be extended in various ways. But first let’s talk about what these graphs are meant to capture. When representing Alice’s thoughts, we might aim to model just her fully-committed beliefs, or perhaps those and her desires and intentions, or perhaps a wider range of propositional attitudes. This is what I do aim to do; we will discuss shortly how to bring more attitudes than belief into the framework. Additionally, one might aim to capture states that aren’t propositional but are equally manifest in Alice’s mental life, as some would construe states like seeking, imagining, and admiring, and others would construe phenomenal consciousness. Let’s suppress all such states for the purposes of this discussion. Finally, one might also aim to capture lower-level facts about Alice’s mental life. Perhaps one of her presentations of Peter comes quicker to mind when she considers who Otto interacts with. I am not here aiming to capture any facts like these, either. I’m assuming that for some explanatory purposes we want to model facts about Alice’s thinking that she shares with other thinkers, where those subcontentful details may be different. That’s the level of abstraction that these mental graphs are meant to work at.

Okay, so let’s consider how to refine the graph framework towards that goal. One question is how to represent the difference between Alice’s believing and her intending that Otto will defeat Spider-Man. The graph:

  1. (16)
    figure y

doesn’t seem to distinguish these. Neither does it distinguish her beliefs from her deliberate withholdings of belief (which is not the same as a mere lack of belief and disbelief). A quick fix would be to label the edges not merely with the relation Alice’s thought ascribes, but also with the attitude she ascribes it with; but we will suppress this issue for now and give a more considered response later.

A second question is how to represent non-atomic thought contents. Here too, we will give a more considered response a bit later. At present, let’s just allow the possibility of Alice’s thoughts ascribing (singly) negated relations:

  1. (17)
    figure z

A third question is how to represent thought contents that ascribe unary properties, or that ascribe relations with arity greater than two. To handle the latter, we can help ourselves to the ordered hyperedges we saw in Section 4. For example, here is Alice thinking that Otto will denounce Spider-Man (but not Peter Parker) to May Parker:

  1. (18)
    figure aa

Notice that it’s a single (hyper)edge that leads from Otto, makes an intermediate layover at the upper vertex labeled Peter, then proceeds to May; and likewise a single edge that leads from Otto to the lower Peter to May. These two edges are labeled differently (only one is negated), because Alice thinks of Otto, Peter, and May as differently related when she thinks of Peter under each presentation.

Similarly, unary predicates can be represented by labeled (hyper)edges that “join” only a single vertex: that is, which start there and point off to nowhere:

(19a):
figure ab

Or perhaps we should draw it as:

(19b):
figure ac

A fourth question is how to represent Frege-type cases that concern the predicative parts of Alice’s thinking, rather than the objects she’s thinking of. One natural move here will shift the ascribed relation from its position labeling an edge into a new, distinguished kind of vertex. That is, suppose that the relations of denouncing and of excoriating are one and the same, but that Alice hasn’t realized this. Then she might think that Otto excoriated Hammerhead while denying that he denounced him:

  1. (20)
    figure ad

Notice that the (identical) relations of excoriating and denouncing here occupy specially marked vertices. We are not conflating Alice’s thoughts with the second-order thoughts that Otto and Hammerhead did and didn’t instantiate the relations of excoriating and denouncing. Those would be represented differently:

  1. (21)
    figure ae

Here it’s the higher-order relation of instantiating that occupies a special vertex, whereas Alice thinks of the relation of excoriating as just another object. This quick sketch is all I’ll give to indicate how the framework might be extended to model Frege-type cases for the predicative parts of thought; for the remainder, I’ll suppress such issues.

A fifth question is how to represent Alice’s thoughts when they’re partly “empty,” that is, some of their expected objects are missing. A quick fix would be to represent her thoughts in those cases as joining vertices that lack any label. However, we will spend a little time on this and instead give a more systematic proposal, that can be repurposed when we address the task of modeling Alice’s quantificational thoughts.

6 Section 6

The tool we need for this is the notion of a disjoint sum or union. This is a counterpart to the more familiar notion of a Cartesian product. The product of two sets Γ and Δ is a collection of objects each of which designates both an element from Γand an element from Δ. The sum of Γ and Δ, on the other hand, will be a collection of objects each of which designates either an element from Γor an element from Δ. Initially, you might think this role could be played just by the ordinary set-theoretic union of Γ and Δ. But recall that when Γ and Δ overlap, we say that (u,v) and (v,u) may be different members of their product. In somewhat an analogous way, it will be useful for the counterpart notion of a sum to distinguish between a member that designates u qua element of Γ and a member that designates u qua element of Δ. So we define a disjoint sum not as a simple union, but instead as a collection of pairs, whose first element is a tag or index indicating which of the operand sets (Γ or Δ) the second element is taken from:

$${\Gamma} \uplus {\Delta} = (\{\mathsf{the\,\, tag}\,\,\text{\textsc{left}}\} \times {\Gamma}) \cup (\{\mathsf{the\,\, tag}\,\,\text{\textsc{right}}\} \times {\Delta}) $$

It does not matter what objects we use as our tags left and right, so long as they are distinct. These objects may even belong to the sets Γ or Δ; no confusion need result.Footnote 19

The notion so defined is a basic building block in the type theories of functional programming languages like Haskell and OCaml. There, instead of:

(22a):

(left, u), or

(22b):

(right, u),

one would instead write:

(23a):

Left u, or

(23b):

Right u.

In the special case where the lefthand operand is a singleton set containing a special object supposed to represent a failure, Haskell writes:

(24a):

Nothing, or

(24b):

Just u

instead of:

(25a):

Left failure, or

(25b):

Right u.

Similarly, OCaml writes None or Some u. I will follow the Haskell convention.

How can we apply this to the case of Alice’s empty thoughts? Let’s suppose that, though Spider-Man and his kin do exist, non-Marvel-approved substances like kryptonite and phlogiston do not. But Alice doesn’t know this. She thinks, confusedly, that Otto hoards kryptonite. She does not think that he hoards phlogiston. At least, that’s how we’d like to put it. Can we represent this difference in Alice’s thoughts, without any objects such as kryptonite or phlogiston to label her mental vertices?

What we will do is label Alice’s vertices using not the bare objects of her thought Otto, Peter, and so on — elements of the domain Marvel — but instead using members of the disjoint sum {failure} ⊎Marvel:Footnote 20

  1. (26)
    figure af

Notice that the labels of these vertices are no longer the objects of Alice’s thought, even when there is such an object. Alice is not thinking about the abstract construction Just Otto. She’s thinking about Otto. Neither is she thinking he hoards the abstract construction Nothing. There isn’t any object she’s genuinely thinking is hoarded, when she has the thought she’d articulate as:

  1. (27)

    Otto hoards kryptonite.

Neither is there any object for her to genuinely think about and call “phlogiston.” But we haven’t here represented her as also thinking:

  1. (28)

    Otto hoards phlogiston.

(Neither does (26) represent her as thinking Otto doesn’t hoard phlogiston, but a different mental graph could do so.)

If we want to represent “empty” thoughts in this way, then we’ll need to go back and relabel all our graphs (12)–(21) using labels like the abstract construction Just Otto instead of simply Otto. Let’s take this as having been done.

Another refinement will be to use as labels not members of the disjoint sum {failure} ⊎Marvel, but instead members of the slightly different sum {failure, variable} ⊎Marvel. The role of this special object variable will be to represent not thoughts that lack objects but bound argument positions in quantified thoughts, which we will discuss below. Before we get there, let me help myself to an extension of the convention by which we express (25ab) as (24ab). Instead of:

(25a):

Left failure, or

(25b):

Right u, or

(25c):

Left variable,

I will now write:

(24a):

Nothing, or

(24b):

Just u, or

(24c):

Var.

That concludes the basic pieces of the framework I want to sketch. What remains is to discuss:

  • different ways the framework can be extended to handle non-atomic thoughts (including quantified thoughts)

  • the differences between this framework and neighboring views, such as neo-Fregean or mental file models of thought

  • how this framework can be put to good use, for example in giving a semantics of folk attitude ascriptions

Sections 79 are devoted to the first two topics. I leave the third, ambitious, topic for later work. Section 10 summarizes what I hope to have achieved in this discussion.

7 Section 7

Let’s start now to expand these models to logically complex thoughts. We can’t just model disjunctive thoughts by sets of graphs, because that would introduce indeterminacy about which vertices in one graph purport to be about the same objects as vertices in another. Our account needs to be more integrated than that.Footnote 21

As I said at the start of Section 3, what I’m offering is deliberately incomplete. How we should continue here depends on our explanatory needs. In some settings we may want to model Alice’s thinking in a fine-grained way. Perhaps not every syntactic detail from how she articulates her thoughts should be mirrored in the thoughts themselves; but we may want a model that tends in that direction.

If that’s what we want, then we can extend the framework sketched so far as follows. First, connectives like o r and a n d can be modeled by introducing labeled edges that join not vertices but other edges.Footnote 22 Here might be Alice’s thought that one of Otto or Spider-Man will eventually defeat the other:

  1. (29)
    figure ag

Using the tools introduced in Section 6, we could model the quantificational thought that most villains will either defeat or be defeated by Spider-Man like this:

  1. (30)
    figure ah

This quick sketch should give a feel for the idea.Footnote 23 Rather than developing it further, let’s instead think about other ways to extend the framework, that are more coarse-grained and encode less language-like structure into Alice’s thinking — though they do still preserve the difference between coordinated and uncoordinated thinking. In some settings, this may suit our explanatory needs better.

Let’s suppose the coarseness of grain we’re aiming for is to identify all of Alice’s thoughts (under a given attitude) having the same prenex/disjunctive normal form. Thus, the thought she’d articulate as:

  1. (31)

    If there is a villain who likes pumpkins, many children will be delighted but Spider-Man will not.

might be equivalent to:

  1. (32)

    x. Many c. (¬Villainous x) ∨ (¬Likes pumpkins x) ∨ (Child cDelighted c ∧¬Delighted Spider-Man)

For this variation, rewind to graphs as described in Sections 5 and 6. So the edges only represent Alice’s atomic thought contents, except we’re also allowing them to represent (single) negations of such. Each (hyper)edge joining n vertices has a direction or order for those vertices, as well as what we might call a “polarity,” indicating whether the n-ary relation it’s labeled with is thought by Alice to hold or to not hold. Let’s extend this by also supposing each edge to have one or more “colors.” My idea here is to impose a cover on all the edges in Alice’s mental graph. A cover of E is a collection of non-empty subsets of E whose union exhausts E; that is, it is like a partition except that cells in the cover are permitted to overlap. The metaphor of “coloring” the edges is just a vivid way to think of this. Each “color” had by an edge represents an additional cover-cell to which the edge belongs. (We don’t rely on any other facts about the “colors” besides their identity and distinctness.) The point of these cover-cells will emerge in a moment.

If you look at the form of Alice’s thought (32), you’ll see that it involves a sequence of quantifiers, around a disjunction each of whose disjuncts is a conjunction (of one or more terms). Each conjunction can be identified with a set of atomic (or singly negated) predications. We will associate each such conjunction with a new color. (If a given conjunction recurs in other thoughts, the same color should be re-used.) Then the whole disjunction can be identified with the set of those colors.Footnote 24

Hence, the quantifiers aside, Alice’s thought (32) might be modeled like this:

  1. (33)
    figure ai

One of the V a r vertices corresponds to the bound variable x in (32), and the other to the bound variable c. The complex disjunction which is inside the quantifiers in (32) can be represented by the graph displayed in (33), together with the set of colors {red, green, blue}. To get Alice’s whole thought, we have to also specify the surrounding quantifiers. This can be represented by an ordered sequence of zero or more pairs, each pair consisting of a quantifier and some Var-labeled vertex in the graph. And we should also specify the attitude under which Alice thinks the thought in question: does she believe it? intend for it to be true? remember it? or what?

Altogether, then, we can define an attitude triplet to be a triple of an attitude, a sequence of zero or more (quantifier, Var-labeled vertex) pairs, and a set of colors.Footnote 25 Given a graph as in (33), each such triplet will represent one of Alice’s thoughts; and a set of them can represent all her thoughts. Coordination relations can be identified between her thoughts, as well as within them, because those thoughts will involve edges joining a common vertex in the underlying graph. Again, this is just a quick sketch, but it should give enough of a feel for the idea.

I’m deliberately not advocating one of these variations over the other, because I think Alice’s thinking will plausibly have each of the structures described: both the fine-grained one and the coarser-grained one — and presumably others too. Different theories may want to make use of different structural properties of her thought. What unites the different models I’ve sketched here is their shared strategy of representing the difference between coordinated and uncoordinated thinking in terms of how many mental vertices are present in the underlying graph, as opposed to the “wires” and “unwires” from the start of Section 3.Footnote 26

8 Section 8

The framework I’ve presented looks a bit different from the “mental file” and neo-Fregean theories we’re more familiar with.Footnote 27 But perhaps these differences are merely superficial?

As I mentioned in Section 1, the machinery developed here needn’t be understood as competing with a “mental files” theory. The most prominent difference is that in my framework, Alice’s memories, information, and attitudes are distributed across thought structures, that may involve multiple “files” or vertices, rather than being contained inside each file. This gives a more satisfying representation of thoughts that predicate (n > 1)-ary relations.Footnote 28 If we had to bundle Alice’s relational thoughts about Otto into a single file, how should that bundled information latch onto her presentation of Peter as Spider-Man rather than his presentation as Peter Parker? Minimally, we’d have to take care to distinguish her thoughts about Otto’s relation to the person Peter Parker, as presented by some other file, from her thoughts about that mental file or presentation.Footnote 29 More awkward would be questions why Alice’s Otto file is entitled to populate itself with information that makes use of other mental files in a way that’s itself immune to reference-failure and Frege mistakes.Footnote 30 (If mental files have access to such a representational system, why can’t we?) These issues are not insuperable for a mental file theorist, but I expect that satisfactory answers to them are just going to reproduce the structure of the framework I’ve presented. They may be willing to take this on as a development of their initial account. If so, I have no complaint.

If my framework were construed as a variant of a mental files account, it would be one that allows files with no referent, files that are not even believed to have a referent (like the V a r-labeled files in (30) and (33)), and the possibility of distinct but indiscernible files. Recanati 2012 explicitly allows the first two of these. He doesn’t address the third, and some of his arguments trend against it, but on the whole it seems to me compatible with his fundamental picture.Footnote 31

In Section 1, I also cited Ninan’s “multi-centered” view of mental content. The graph models sketched above use different theoretical apparatus than Ninan, but I think our proposals are fundamentally similar. Ninan needs an arbitrary ordering of the objects of an agent’s thought, which we might avoid (though the most common mathematical constructions of graphs don’t bother to be very hygienic about this, see note 16, above). Also, Ninan’s proposals about mental content are intertwined with his proposals about the semantics of attitude reports; and here I want to diverge from him more substantially. But I cannot adequately address the semantics of reports in this paper.Footnote 32

Neo-Fregeans can also agree with some parts of my framework. Or at least, my framework seems capable of representing things they want to say. Information that Alice associates with Peter, when presented in a given way, is represented in her graph by the edges incident with that Peter-labeled vertex. For example, the presentation of Peter deployed in her thought (31)/(32) is connected with the information that he is named “Spider-Man” and that he slings webs — even though this additional information isn’t itself part of that thought. (That’s why those edges in graph (33) weren’t colored red, green, or blue.) Nor are Alice’s associations limited to beliefs: the edge that associates Peter with web-slinging could be part of a suspicion, or an intention, or some other attitude. So this framework seems hospitable to the neo-Fregean’s ambition to capture the informational and attitudinal differences between Alice’s different presentations of Peter.

Still, I count it as an advantage of my framework that is does not force us to say that Alice associates rich bodies of information with every mental vertex. Her presentations can manage to be multiple even when the information she associates with them is comparatively thin, and in particular falls short of being uniquely identifying.

Indeed, her vertices or presentations can be multiple even when they’re associated with all the same information and attitudes. As I suggested at the outset, Alice’s two presentations of Bob in case (1) might differ merely numerically, as do the vertices on the right-hand sides of diagrams (10′) and (34):

(10′):
figure aj
  1. (34)
    figure ak

So, though my framework can accommodate all the associations and fineness of grain a neo-Fregean wants, adopting it may undermine, or at least relieve the motivational pressure for, paradigmatic neo-Fregean claims about how much associated information must be available.Footnote 33

Should we want our models of Alice’s thinking to have the capacity to represent merely numerical differences in her presentation of a single object? Is it really intelligible for her to have multiple presentations of Bob that are “qualitatively indiscernible” — that is, unaccompanied by any qualitative difference in her associated memories, information, or attitudes? (As in (34), we allow there to be networks of further numerical differences in what other, also qualitatively indiscernible, vertices they are adjacent to. The upper Bob Jr vertex is joined only to the upper, and not also to the lower, Bob Sr vertex.)

Here are several cases to motivate the intelligibility of this.

Suppose Alice has a vivid dream of a distinctive-looking woman, and remembers the experience. The dream did not present itself to her as being of anyone she was already familiar with. We can suppose also that it did not present itself to her as being dreamt, or at any rate that over time she forgets that this experience came from a dream. Later in life she has a perceptual experience that exactly qualitatively matches the dream experience. Most will agree that it’s possible for the dream not to have been a prescient presentation of the same person she later perceives.Footnote 34 Intuitively, neither do these two experiences need to be coordinated. It can be an open epistemic question for Alice whether these two presentations were of a single individual. Twisting the case, we can further suppose that the later experience was in fact itself a hallucination. That doesn’t seem to render the two presentations any less distinct. We can suppose further that as time passes, Alice forgets which experience came first, whether one was dreamt and the other apparently perceptual, or what. They might nonetheless remain distinct presentations to her. Yet no merely qualitative information contained in the presentations may differentiate them. For a final twist, we can suppose that Ellen is a mental duplicate of Alice as she is now, but in Ellen’s case, the two memories of a distinctive-looking woman never were generated from distinct experiences. She only ever had a single experience, but through some signaling glitch in her brain, now has cognitively distinct memory presentations, just like Alice. If Alice or Ellen were to revise one of their memories, it would change autonomously. The other presentation would remain as it is.

Suppose Flugh is an alien whose many eyes are on long stalks, which can wriggle through the maze of twisty little passages where he lives. The experiences generated by his eyes are co-conscious, but not always spatially integrated. In particular, on one day he has two qualitatively matching experiences of a homogeneous sphere, without any presentation of how the spheres are spatially related to each other. Neither experience even seems to be above, or to the right, of the other. As it turns out, Flugh is seeing only a single sphere.Footnote 35 In fact, it may be that he’s only seeing a single sphere with a single eye, but through some signaling glitch in his brain, he now has cognitively distinct mental presentations. That is, Flugh is in a position to wonder:

  1. (35)

    Is this sphere the same as that one?

Perhaps Flugh would be in a position to introduce some difference in how he thinks of the two spheres. Perhaps he could start thinking of one of them as the sphere I’m attending to in memory right now. That may make his thought problematically private and ephemeral, but regardless, I don’t see why Flugh must be thinking of the spheres that way. Couldn’t his two sphere presentations in fact be qualitatively exactly alike, but nonetheless be two presentations?Footnote 36

Also relevant is some interesting empirical work by Anthony Marcel. Following some informal reports in Schilder 1935, Marcel established that retinal disparity and splitting attention between vision and touch could generate an illusory tactile experience as of being touched twice on one’s fingertip, when in fact there was only a single touch. Interestingly, subjects were confused about the felt location, or even the spatial relation, of these two (apparent) touches. Marcel describes the effect as “perception of tactile two-ness without determinate perception of each of the two touches” (p.c.). The phenomenon needs to be investigated further, but one intelligible possibility is that the touches and their felt locations were presented as two, without there being any difference in the qualities they were presented as having. Neither touch felt to be to the right of the other; subjects were not even sure on what axis the touches were felt to be aligned. That’s why identifying and comparing the felt locations was so difficult. Marcel does say that when he underwent the experiment himself, he “came to be able to perceive one touch as lighter than the other but was unsure which was which.” But that leaves open the possibility that before he achieved this, or in the experience of other subjects, there was no qualitative difference between the two touch experiences.

In response to such cases, some of my interlocutors protest that the subjects would only be thinking plurally about a pair of (putative) women, or a pair of spheres or touches, and wouldn’t be in a position anymore to think about each element of the pair separately.Footnote 37 I acknowledge there can be cases of that sort. But I don’t see why our cases must take that form. This diagnosis sounds especially forced for the dreamt woman case, where the two presentations never need to have occupied Alice’s attention simultaneously.

Others may insist that in my examples there must be some underlying difference in how Alice’s or Flugh’s twin vertices are implemented. That may be; but at the level of abstraction we’re modeling their mental reality, which can be shared between them and other physiologically different subjects, I don’t see why any such implementation differences need to be manifest.

Consider also this. Flugh may know that spheres have invisible properties, let’s call them even-centeredness and odd-centeredness, which are roughly equally distributed. If he’s confident that the spheres he seemed to see are distinct, then it seems that two lines of hypothetical reasoning should be open to him. One would go: Suppose at least one of the spheres is even-centered, then the probability that they’re both even is \(\frac {1}{3}\), because in \(\frac {1}{3}\) of the cases where at least one of two independent spheres is even-centered, they both are. But there’s also another line of reasoning available to him. It would go: Call one of the spheres X. (Perhaps here he calls to mind and attends to one of his sphere images; or perhaps he doesn’t.) Suppose it is even-centered. Then the probability that they’re both even is \(\frac {1}{2}\) — again because the centeredness of the other sphere is an independent matter. Both of these seem to be legitimate courses of reasoning. They are marred only by Flugh’s ignorance of the fact (and giving no credence to the possibility that) the two spheres are one and the same. It may be a subtle matter for Flugh to sort out when he should deploy the one line of reasoning and when the other. But even making sense of the intelligibility of the second line of reasoning suggests that Flugh will have, at least temporarily and suppositionally, the ability to think about his “two” spheres independently.Footnote 38

9 Section 9

Some questions and replies.

First, suppose neo-Fregeans are prepared to accept the possibility of merely numerical differences in Alice’s presentations of Bob. How much of a departure need this be from more familiar versions of their view?

Some neo-Fregeans will already insist on the possibility that modes of presentation be “non-descriptive.” Often what this seems to mean is only that the subject may lack concepts or the language to articulate information that her presentations can nonetheless ascribe to an object. I haven’t made any effort to exclude or include this in my framework. If we did want to specially keep track of such “non-descriptive” modes of presentation, though, they’d presumably have to do with what labels the edges of Alice’s graph, not with the number of indiscernible vertices she can have. As I’d put it, this is only an issue about which qualities Alice’s presentations can attribute, or perhaps the formats in which she grasps those qualities. What I’m after is more radical: the possibility of presentational differences that exceed any differences in attributed qualities.

Some neo-Fregeans like Evans will also already insist on the possibility that presentations may be qualitatively indiscernible but distinct, because they are presentations of different objects — or at least, presentations grounded in “different ways of gaining information” from an object. This is a step in the direction I’m after. But I want to go further in that direction than I expect Evans would. In some variants of the examples in the previous section, Ellen and Flugh only gained information in a single “way,” but due to signaling glitches that information was cognitively “forked,” and they then had multiple presentations downstream.

Let’s suppose the neo-Fregeans are willing to accommodate all of this. All my examples motivate, they may say, is the possibility that senses may differ solo numero. This may be a relatively novel variety of neo-Fregeanism, but why shouldn’t it belong to their family of views for all that?

Indeed, that would grant most of what I’m after. If you think of yourself as a neo-Fregean, and you can understand the proposals I’m making as refinements of your own view, I see no compelling need to resist that. The only reservation I have about this way of framing my proposals is this. If Alice really does have two numerically distinct but qualitatively indiscernible presentations of Bob, and these presentations are reasonably thought of as two particular senses, then mightn’t it make sense to talk about two (or more) possible situations, in which these particular senses are permuted. I don’t think that ought to make sense. There doesn’t have to be anything in the mental reality we’re modeling — at the level of abstraction we’re modeling it, which may be shared between Alice and others who realize the same mental structure in concretely different ways — that captures such differences. At this level of abstraction, the permutations aren’t meaningful.Footnote 39 Properly understood, that’s the result a graph framework for talking about Alice’s mind should deliver. There is no graph-theoretical difference between our graph (10′) and a “different” graph in which the two right-hand vertices have been permuted. If we talk about particular senses, though, I worry that these permutational differences might be genuine possibilities. At least, I think we’d have to do work to earn the right to say that permuting the senses makes no difference. If we can achieve an understanding of indiscernible but numerically different senses where such permutations aren’t genuinely different possibilities, then okay, we’re just doing graph theory in other vocabulary.

Second, in presenting this framework and sketching different ways it may be fleshed out, I’ve only been specifying what it is for an agent to have a given set of attitudes at a single time. I haven’t said anything about what it would be for the agent to retain some of those attitudes while revising others. Nor have I said what it would be for the agent to share some of those attitudes, but not others, with other agents. I point this out to counter the impression that the framework developed here must give holistic answers to those questions: that the only kind of mental overlap it would allow would require sharing entire mental graphs. Not so. This framework can be developed that way; but it is also compatible with various stories about when agents share more local, specific mental structures.

One especially hard version of the preceding questions concerns the mental relations between Alice, who has two presentations of Bob (and Peter and other objects), and another subject Daniel who has only a single presentation. When Daniel thinks about Alice’s thoughts, should he associate her Bob i presentation with his own unitary presentation of Bob? What then should he associate her other, Bob j presentation with? These are complex issuesFootnote 40; and I mention them here only to disavow any ambition for the present framework on its own to be adequate for answering them. An answer will have to come from a fuller story about the semantics of our talk and thought about others’ minds. I do believe the present framework is a useful component to be used in such a fuller story, but that will have to be shown elsewhere.

For the time being, then, I’m only offering a model of a single agent’s mind, at a single time, concerning non-semantic and non-intentional matters.

Third, I said earlier that coordinated thinking should be represented with a common vertex, like this:

  1. (7)
    figure al

and uncoordinated thinking with multiple vertices, like this:

  1. (8)
    figure am

But what if the agent has multiple vertices but believes the objects so presented to be identical? That is, what is the difference between (7) and:

  1. (8′)
    figure an

(Here we meet again some “good questions” we raised in Section 3, but never answered.)

I expect that subjects can have attitudes like that, attributing identity to objects thought of in uncoordinated ways.Footnote 41 So I’ll agree that (8′) represents a possible mental state, distinct from (7). Depending on what attitude it represents the subject as having, it may even be a rational state — for example, if the agent isn’t fully, indefeasibly certain that the identity holds. If she were that fully, indefeasibly certain that the identity held, but persisted in thinking of the objects with two vertices rather than one, then perhaps she’d thereby be manifesting some irrationality.Footnote 42 So: (8′) might represent even a possible state of maximally-committed belief, distinct from (7), though arguably not an “ideally rational” one.Footnote 43

There can also occur cases where the agent does aim to think of an object coordinately, but has unknowingly latched onto multiple referential candidates. In such cases, I’d posit a single vertex, labeled in some special way to indicate that the referent is indeterminate between the several candidates. (Here we can again use the machinery introduced in Section 6.) That would need to be clearly separated from vertices with plural referents, which we will also sometimes want. The framework sketched here naturally lends itself to such amplifications. The difficulties hereabouts concern what theoretical constraints and predictions we want to make with these various notions, not how to adjust the formal model to express them.

10 Section 10

What I’ve presented here is only a programmatic framework; and as I’ve said, it may be developed further in different ways.

I observed some commonality between my framework and views advanced by Heck, Ninan, and Recanati, or at least with some core parts of their proposals. The divergences between us are interesting but are compatible with a shared fundamental spirit.

I’ve also remarked on how opposed a neo-Fregean need or needn’t be to my framework.

Richard 1990, Fine 2007, and Pinillos 2011 have come up a few times in our discussion, and to those who know those works (and Richard 1983, 1987, and Taschek 1995, 1998), a broad consonance with my framework will also be evident. I will discuss those proposals in detail elsewhere; I haven’t addressed them more specifically here because the connections and differences aren’t straightforward. I am sympathetic to the semantic ideas they share, but unlike these authors (and Ninan too), I think it’s more productive — at least, at this stage of inquiry — to deal with the mental and the linguistic separately. This paper aims only to give a framework for modeling and theorizing about our mental states. What linguistic mechanisms subjects may have to express, or talk about, coordinated thinking, I’ve deferred for later discussion.

I want to dwell a moment longer on the differences between my framework and a range of nearby positions, that I’ve called “symbolist” and cited examples of in note 1, above. (See also the symbolist interpretations of mental files and discourse referents in note 31.) These views will invoke particular mental symbols, or words in a language of thought, or private representational vehicles, where I have only posited graph-theoretic structure. The symbolists needn’t resist much (if anything) of what I’ve said, since the machinery they invoke does also have graph-theoretic structure. They could just be regarded as taking on stronger commitments. To compensate, they can claim greater familiarity, and perhaps more explanatory power.

But are these views really explanatorily better?

I’m not sure exactly what cognitive architecture we commit to when we talk of “mental symbols,” but symbolists generally allow that it’s possible for subjects to share propositional attitudes while having different, or no, mental symbol systems. And as I said a few times earlier, if we want explanations that are driven by what we have in common with other such possible thinkers, we should have models that abstract away from the details of how each subject happens to implement the structures they share. The details of how things are represented in one’s own cognitive architecture — not just the neural details, but whether the implementation has any syntax-like structure at all — is arguably only important for other, more parochial, explanatory aims.

Consider the following analogy. Suppose I appeal to the algebraic notion of a ring for some explanatory purposes, with which you’re initially unfamiliar. After learning about rings, you say, “Oh I see, like the integers. Why don’t you you just use integers in your model?” I respond: Integers are a more specific kind of structure than a ring. They may be more familiar, and more easily axiomatized (if you’re not concerned to emphasize how they are specialized rings and so on). But they are heavier-duty, less minimal structures than rings because of their more specific commitments. If we make explanatory use of integers, we should keep track of when those more specific commitments are appropriate or justified. We have to make sure we don’t illegitimately help ourselves to stuff in our model whose presence hasn’t yet been fairly motivated.

How does that bear on our present discussion? The symbolists bring more to the explanatory table than just the graph structure of their symbols. But I don’t think we’re in a position to affirm that these extras are really anything that our explanatory models of the mind, with the aim to capture what we share with other possible thinkers, are entitled to. The “permutation worry” I voiced in Section 9, above, emerges from such doubts. (See also the artifacts cited in note 39.) I expect that making a model of the mind abstract enough to avoid these worries will inevitably have to bring in the kind of graph-theoretical descriptions I’ve employed, or something that is not substantially any different.

To the extent that the framework presented here really does differ substantially from similar-seeming views, and isn’t inter-translatable with them, the take-home story is that my framework aims to be more minimal and abstract, capturing interesting structure the other views posit, while avoiding their commitments about how that structure is implemented. What my framework includes is already enough on its own to model the paradigm Frege phenomena in thought that we’ve accumulated over the past century and a quarter.