Keyboard Shortcuts?

×
  • Next step
  • Previous step
  • Skip this slide
  • Previous slide
  • mShow slide thumbnails
  • nShow notes
  • hShow handout latex source
  • NShow talk notes latex source

Click here and press the right key for the next slide (or swipe left)

also ...

Press the left key to go backwards (or swipe right)

Press n to toggle whether notes are shown (or add '?notes' to the url before the #)

Press m or double tap to slide thumbnails (menu)

Press ? at any time to show the keyboard shortcuts

\title {Joint Action \\ Lecture 11}
 
\maketitle
 

Lecture 11

Joint Action

\def \ititle {Lecture 11}
\def \isubtitle {Joint Action}
\begin{center}
{\Large
\textbf{\ititle}: \isubtitle
}
 
\iemail %
\end{center}
 
\section{Aggregate Subjects: Recap}
 
\section{Aggregate Subjects: Recap}
In what ways might the notion of an aggregate subject help us to understand which forms of shared agency underpin our social nature? To answer this question, we first need to consider a more basic question: How can aggregate subjects exist? In a previous lecture we considered the suggestion (developed by Pettit and List among others) that aggregate subjects arise as a conseuqence of individuals representing them. Why do should we consider other suggestions?

Which forms of shared agency underpin our social nature?

What distinguishes joint action from parallel but merely individual action?

In the first part of this course we saw that there are reasons to reject both the Simple View and Bratman’s view (a counterexample, no less), and significant challenges to extracting an answer to these questions from Gilbert’s work.
When we look at the leading accounts, it’s not much of an exaggeration to talk about a train wreck.
Maybe we can get further by adopting a more radical approach.

How?

aggregate subject

The more radical approach we are in the middle of exploring hinges on aggregate subjects. An aggregate subject is a subject with proper parts who are themselves subjects.
I compared aggregate subjects to an aggregate animal, the Portuguese man o' war (Physalia physalis), which is composed of polyps.
Here you can say that ‘the group [of polyps] itself’ is engaged in action which is not just a matter of the polyps all acting.
But how can such a thing exist? Humans do not mechanically attach themselves in the way that the polyps making up that jellyfish do.
So (as we saw in the last lecture) we have to ask, How are aggregate agents possible?
It was striking that many people took the view that such things just couldn’t exist; just as Searle did. One thing I aim to do in this lecture is to convince you that their existence is not quite as strange as you might think.
But first I want to remind you of a couple of themes from the last lecture.
 

Team Reasoning

 
\section{Team Reasoning}
 
\section{Team Reasoning}
‘You and another person have to choose whether to click on A or B. If you both click on A you will both receive £100, if you both click on B you will both receive £1, and if you click on different letters you will receive nothing. What should you do?’ (Bacharach 2006, p. 35) Team reasoning is a game-theoretic attempt to explain what makes your both choosing A rational. But what is team reasoning? And how is team reasoning relevant to questions about joint action? This unit provides the barest outline. The aim is not to explain, but to make you aware that there’s a body of research in this area.

How to create an aggregate subject?

1. self-representation (just done)

2. team reasoning?

There is further, alternative motivation for considering team reasoning. Some have claimed that it will provide us with a ground-level account of shared intention, and thereby of shared agency ...

‘The key difference between the two kinds of intention is not a property of the intentions themselves, but of the modes of reasoning by which they are formed. Thus, an analysis which starts with the intention has already missed what is distinctively collective about it’

\citep{Gold:2007zd}
So these researchers, unlike Pettit and List are aiming to capture a basic form of shared agency rather than to build on a prior account of shared intention. But there’s more ...

‘collective intentions are the product of a distinctive mode of practical reasoning, team reasoning, in which agency is attributed to groups.’

\citep{Gold:2007zd}
So these researchers are aiming to build a kind of aggregate subject.
They think, in a nutshell, that aggregate subjects are not only a consequence of self-reflection, but can also arise through (a special mode of) reasoning about what to do.

Gold and Sugden (2006)

But what is team reasoning? I’m so glad you asked ...

‘somebody team reasons if she works out the best possible feasible combination of actions for all the members of her team, then does her part in it.’

\citep[p.~121]{Bacharach:2006fk}

Bacharach (2006, p. 121)

These are the questions you would want to answer if you were going to pursue team reasoning. In this lecture series I decided there isn’t time to do that this year.

1. What is team reasoning?

2. In what sense does team reasoning give rise to aggregate agents?

3. How might team reasoning be used in constructing a theory of shared agency?

 

Schmid’s ‘Charlie Brown Phenomenon’

 
\section{Schmid’s ‘Charlie Brown Phenomenon’}
 
\section{Schmid’s ‘Charlie Brown Phenomenon’}
The the Charlie Brown phenomenon is ‘the possibility of being intentionally engaged in a joint action in a situation in which overwhelming evidence to assume that the others will not do their part is recognized to exist’ (Schmid 2013, p. 54). What does the existence of this phenomenon suggest about shared agency?

‘participants in joint action are usually focused on whatever it is they are jointly doing rather than on each other. Where joint action goes smoothly, the participants are not thinking about the others anymore than they are thinking about themselves’

\citep[p.~37]{Schmid:2013}

Schmid (2013, p. 37)

I don’t think this observation is an argument or an objection, but it is suggestive.
In Pacherie’s account, we need beliefs about the others and their team reasoning.
In Bratman’s account we need intentions about others’ intentions.

‘cooperators normatively expect their partners to cooperate; they do not predict their cooperation’

Dominant View: ‘the representation of the participation of the others has a mind-to-world direction of fit.’

Alternative View: ‘the representation of the participation of the others has a world-to-mind direction of fit.’

\citep[p.~38]{Schmid:2013}

Schmid (2013, p. 38)

explain
Ok, this is just an assertion. What’s the argument for it ...
‘As his intention was to hit the ball rather than just to try to hit the ball, Charlie’s intention either represents Lucy (cognitively) as doing her part (holding the ball steady) or at least is incompatible with the belief that Lucy will not hold the ball. Already in the 1950s, it becomes increasingly clear that Lucy will pull the ball away. This evidence is further corroborated over the following decades. Therefore, Charlie should not be optimistic. By continuing to intend to kick the ball, it seems that Charlie violates the sufficient reason condition. As he has reason to believe that Lucy will pull away the ball, there is insufficient reason for optimism that he will be able to kick it.
‘Thus, in the view developed so far, and endorsed by such authors as John Searle (2010), Raimo Tuomela, and Michael Bratman, there must be something structurally wrong with Charlie’s intentionality; in his right mind, he cannot intend to kick the ball. According to this line of analysis of joint action, Charlie is simply unreasonable.’
BUT: ‘People think that Lucy rather than Charlie is at fault’ (Schmid 2013, p. 46)

Team reasoning

Schmid

Participants reason about what team-directed preferences require (so do not distinguish themselves from others).

‘participants in a joint action represent their partners as doing their parts in the same way as individual intentions implicitly represent the agent as continuing to be willing and able to perform the action until the intention’s conditions of satisfaction are reached’

‘individual agents of temporally extended actions “represent” their own future intentions and actions in the same way in which cooperators represent their partners’ intentions and actions.’

\citep[p.~49]{Schmid:2013}

Schmid (2013, p. 49)

How do cooperators represent their partners’ actions?

‘this representation is neither (purely) cognitive nor (purely) normative, but rather a very peculiar combination of the two. ’

\citep[p.~50]{Schmid:2013}

Compare representing your own actions:

‘An individual with a purely cognitive stance toward his own future self’s behavior and no normative expectation is a predictor of his behavior rather than an intender of his future action; similarly, an individual with a purely normative stance toward his own future behavior is a judge over [...] his future behavior rather than an agent.’

\citep[p.~50]{Schmid:2013}

Schmid (2013, p. 50)

I don’t want to go all the way with Schmid. In particular, I don’t quite accept his suggestion that ‘A cooperator’s basic attitude toward his partner is such that he (implicitly) assumes that by representing the other as doing his part he makes it more likely that the other will in fact do his part because it provides the other with a motivating and a normative reason to do so.’ (p. 50)

How can we turn these observations into a theory?

Recall that we want a theory in order to be able to distinguish genuine joint action from parallel but merely individual action.

PS: Are we still talking about aggregate agents?

Yes: from the point of view of the agents. (If Schmid is right, the basic attitude I have towards your actions does not distinguish your actions from mine.)
 

Interim Summary

 
\section{Interim Summary}
 
\section{Interim Summary}
Our question is, What distinguishes genuine joint actions from parallel but merely individual actions? So far we have considered objections to some of the leading attempts to answer this question. What has gone wrong?
Summary so far

What distinguishes genuine joint actions from parallel but merely individual actions?

The Simple View (but counterexample)

The shared intention strategy

--- Bratman (but counterexample)

--- Gilbert (but are there joint commitments?)

--- Ludwig, Pacherie, Searle, Tollefsen, ... (not covered)

The ‘aggregate subject’ strategy [nb plural vs aggregate]

--- reflectively constituted (Pettit and List)

--- emotionally constituted (Helm; not covered)

Btw, it seems to me that there’s an ethical issue. If there are aggregate subjects, are we under obligations not to end their lives? Or is the life of an aggregate subject somehow not as valuable as the life of a non-aggregate one?

Team reasoning (not covered)

Diagnosis

Too reflective!

Grounded merely on intuitive contrasts!

I think our problems have at least two sources.