Long-time reader writes,
I was going through some of your posts from earlier this month (Belief, Designation, and Substitution, January 10, 2017) and was interested in seeing your comment that "[l]inguistic reference is built upon, and nothing without, thinking reference, or intentionality.". . . I have to say that your above sentence was the first time I've heard anyone articulate what you have articulated in such a direct manner. It's something that certainly makes the most sense to me in terms of thinking about some of the broad discussion points in the field, but I'm surprised, actually, that no one I've come across has articulated this, and I'm curious whether that lacuna has to do with the analytic tradition's anti-metaphysical tendencies (of a more robust type of metaphysics, in any event): if one moves the object of analysis from questions about how language refers to how the mind refers, perhaps it gets one into hoary metaphysical waters that people back in the day would rather have left alone. Is this actually the case or am I missing something or is the whole thing simply too obvious for most people to bother mentioning?
The Primacy of the Intentional Over the Linguistic
Following Chisholm, et al. and as against Sellars, et al. I subscribe to the broadly logical primacy of the intentional over the linguistic.
But before we can discuss the primacy of the intentional, we must have some idea of (i) what intentionality is and (ii) what the problem of intentionality is. Very simply, (mental) intentionality is object-directedness, a feature of some (if not all) of our mental states. (The qualifier 'mental' leaves open the epistemic possibility of what George Molnar calls physical intentionality which transpires, if it does transpire, below the level of mind. I take no position on it at the moment. Dispositionality would count as physical intentionality.)
Suppose a neighbor asks me about Max Black, a stray cat of our mutual acquaintance, who we haven't seen in a few weeks. The asking occasions in me a thought of Max, with or without accompanying imagery. The problem of intentionality is to provide an adequate account of what it is for my thought of Max to be a thought of Max, and of nothing else. Simply put, what makes my thought of Max a thought of Max? How is object-directedness (intentionality, the objective reference of episodes of thinking) possible? How does it work? How does the mental act of thinking 'grab onto' a thing whose existence does not depend on my or anyone's thinking?
Why should there be a problem about this? Well, an episode of thinking is a datable event in my mental life. But a cat is not. First of all, no cat is an event. Second, no cat is a content of consciousness. It's an object of consciousness but not a content of consciousness. Cats ain't in the head or in the mind. Obviously, no cat is spatially inside my skull, or spatially inside my nonspatial mind, and it is only a little less obvious that no cat depends for its existence on my mind: it's nothing to Max, ontologically speaking, if me and my mind cease to exist. He needs for his existence my thinking of him as little as my thinking needs to be about him. We are external to each other. Cats are physical things out in the physical world. And yet my thinking of Max 'reaches' beyond my mind and targets -- not some cat or other, but a particular cat. How is this possible? What must our ontology include for it to be possible?
To get the full flavor of the problem, please observe that my thinking of Max would be unaffected if Max were, unbeknownst to me, to pass out of existence while I was thinking of him. (He's out on the prowl and a hungry coyote kills him while I am thinking of him.) It would be the very same thought with the very same content and the very same directedness. But if Max were to cease to exist while a flea was biting him, then the relation of biting would cease to obtain. So if the obtaining of a relation requires the existence of all its relata, it follows that intentionality is not a relation between a thinker (or his thought) and an external object. But if intentionality is not a relation, then how are we to account for the fact that intentional states refer beyond themselves to objects that are (typically) transcendent of the mind?
How is it that the act of thinking and its content 'in the mind' hooks onto the thing 'in the world' and in such a way that true judgments can be made about the thing, judgments that articulate the nature and existence of thing as it is in itself apart from any (finite) thinking directed upon it?
Now it seems to me that any viable solution must respect the primacy of the intentional over the linguistic. This thesis consists of the following subtheses:
A. Words, phrases, clauses, sentences, paragraphs and the like, considered in their physical being as marks on paper or sounds in the air or carvings in stone (etc.) are entirely lacking in any intrinsic referential, representative, semantic, or intentional character. There is nothing in the nature of the mark 'red' that makes it mean red. After all, it doesn't mean red to a speaker of German. It doesn't mean anything to a speaker of German qua speaker of German. In German 'rot' means red while in English the same sign is in use but has a different meaning. Clearly, then, marks on paper, pixels on screen, etc. have no intrinsic sense or reference grounded in their very nature. It is a matter of conventional that they mean what they mean. And that brings minds into the picture.
B. So any sense or reference linguistic signs have must be derivative and relational as opposed to intrinsic: whatever intentionality they have they get from minds that are intrinsically intentional. Mind is the source of all intelligibility. Linguistic signs in and of themselves as mere marks and sounds (etc.) are unintelligible.
C. There can be mind without language, but no language without mind. Laird Addis puts it like this:
Conscious states can and do occur in beings with no language, and in us with no apparent connection to the fact that we are beings with language. Thus we may say that "mind explains language" in a logical or philosophical sense: that while it is perfectly intelligible to suppose the existence of beings who have no language but have much the same kinds of conscious states that we have, including introspections of other conscious states, it is unintelligible to suppose the existence of beings who are using language in all of its representative functions and who are also lacking in conscious states. The very notion of language as a representational system presupposes the notion of mind, but not vice versa. (Natural Signs: A Theory of Intentionality, Temple University Press, 1989, pp. 64-65)
These considerations strike me as decisive. Or are there counter-considerations that 'cancel them out'?
>>Note to the Astute Opponent:
Sorry only just spotted this. I have been down with the winter virus (not pleasant).
>>Simply put, what makes my thought of Max a thought of Max?
1. Start by assuming that you express your thought, saying ‘Max is black’.
2. If this really does express your thought, then it must be a thought about whatever ‘Max’ refers to.
3. So we can dispense with the intentionality of thoughts, and instead ask about the intentionality of reference.
4. So the question is, what makes ‘Max is black’ a statement about Max?
I.e. if ‘Max is black’ truly expresses a thought of yours, then the object you are thinking about must be the referent of ‘Max’
How does that sound so far? More to come.
Posted by: Astute opponent | Thursday, January 26, 2017 at 04:35 AM
Sorry about the virus. Did you get a flu shot in the fall? I have found that that is a very good idea especially as we age.
I am thinking about the precis of your book, and I will try to respond before long.
What you say above sounds self-contradictory. You assume that there is a thought which then is expressed. But this is precisely not to "dispense with the intentionality of thoughts."
Besides, I needn't express my thought in language.
By the way if you scroll down, there is a short entry on mortality that I'd like your comment on.
Posted by: BV | Thursday, January 26, 2017 at 04:58 AM
>>But this is precisely not to "dispense with the intentionality of thoughts."
I mean, precisely, that we don't have to worry about what thought is, hire a clairvoyant etc. On the assumption that the object of the thought = the referent of the singular subject term, we can now ask the same question about the referent. This in my view makes the problem tractable.
This does not rule out the possibility of reference being intentional, and I am sure you would claim it is.
Posted by: Astute opponent | Thursday, January 26, 2017 at 05:13 AM
Tying this to the precis thing:
The two fundamental assumptions of the book are (i) that the proper name ‘Max’ in (2) refers back to the first sentence that introduces the name. It could equally be replaced by ‘he’. And (ii) that the same phenomenon is in play in (3). I.e. (3) is true so long as the ‘Max’ also back refers to sentence (2). The thesis is that reference statements which appear to assert a relation between language and reality (e.g. sentence (2) and Max himself) are grammatically misleading. What makes them true is an intralinguistic relation.Your view so far has been to scoff and assert that ‘Max’ in fact refers to Max. I reply: that is certainly true, and ‘Max’ really really does refer to Max. So we are not disagreeing there. So what are we disagreeing about?
Your other argument is that only pronouns back-refer. OK, but you need to show why sentence (2) does not involve back-reference.
Posted by: Astute opponent | Thursday, January 26, 2017 at 05:46 AM
Bill,
What does he mean by all of its representative functions? Would he allow a being using some of the representative functions of language that was lacking in conscious states?Addis says,
Posted by: David Brightly | Friday, January 27, 2017 at 04:45 AM
David,
No. It would have been better had he written 'any of its representative functions.'
Suppose you have a robot with a voice synthesizer. The robot is performing tasks in a hot environment. Suddenly the environment gets hotter, the robot senses this, and produces the sounds 'If it gets hotter, my circuitry will melt down.'
The robot lacks conscious states. Addis would say that the robot's sounds do not represent anything to the robot, although we who are conscious will take the sounds as words representing something. The robot is not using language in any representative way.
To me this is obvious. I fear you will disagree completely.
Posted by: BV | Friday, January 27, 2017 at 05:40 AM
That your robot lacks conscious states is obvious to me too, Bill. But you are making the stronger statement that any use of language to represent must involve consciousness, on pain of unintelligibility. I'm not convinced of this necessary connection. Obviously linguistic representation and consciousness tend to go together in us. But we don't find the puzzles of consciousness in linguistic representation. If the latter is a way of linearising and manipulating a model which guides behaviour then it need not imply our kind of consciousness at all.
Posted by: David Brightly | Saturday, January 28, 2017 at 09:13 AM