Thinking in patterns:

advertisement
Thinking in patterns:
representations in the neural basis of theory of mind
by
Jorie Koster-Hale
B.A. Linguistics and Cognitive Science
Pomona College, 2009
Submitted to the Department of Brain and Cognitive Sciences
in partial fulfillment of the requirements for the degree of
Doctor of Philosophy in Neuroscience
at the
MASSACHUSETTS INSTITUTE OF TECHNOLOGY
September 2014
© 2014 Jorie Koster-Hale. All rights reserved.
The author hereby grants MIT permission to reproduce and to distribute publicly paper
and electronic copies of this thesis document in whole or in part in any medium now known or hereafter
created.
Signature of Author:______________________________________________________
Jorie Koster-Hale
Department of Brain and Cognitive Sciences
August 16, 2014
Certified by: __________________________________________________________
Rebecca R. Saxe
Professor of Brain and Cognitive Sciences
Thesis Supervisor
Accepted by:__________________________________________________________
Matthew A. Wilson
Sherman Fairchild Professor of Neuroscience
Director of Graduate Education for Brain and Cognitive Science
1 of 179
This page intentionally left blank
2 of 179
Thinking in patterns:
representations in the neural basis of theory of mind
by
Jorie Koster-Hale
Submitted to the Department of Brain and Cognitive Sciences
on August 6, 2014, in partial fulfillment of the requirements for the degree of
Doctor of Philosophy in Neuroscience
Abstract
Social life depends on understanding other people’s behavior: why they do the things they
do, and what they are likely to do next. These actions are just observable consequences of
an unobservable, internal causal structure: the person’s intentions, beliefs, and goals. A
cornerstone of the human capacity for social cognition is the ability to reason about these
invisible causes; having a “theory of mind”. A remarkable body of evidence has
demonstrated that social cognition reliably and selectively recruits a specific group of brain
regions. Yet, we have little insight into how these neural substrates function at a
computational level. This thesis lays the groundwork to address that question, both
empirically and theoretically, first by demonstrating that functional neuroimaging can find
behaviorally relevant features of mental state representation within the cortical regions that
support social cognition, and second by proposing a theoretical framework to interpret
activity in these brain regions. In Chapter 1, I review the literature of the last 15 years, and
argue that a key next step in understanding the neural basis of social cognition is
characterizing the neural representations and computations supported by “social” brain
regions. In Chapter 2, I demonstrate in four experiments that functional neuroimaging can
be used to find neural representations of distinct features of mental states. Specifically, I
show that multivoxel pattern analysis (MVPA) can detect features of mental state
representations (e.g., intent), and that these neural patterns are behaviorally relevant,
including in autism spectrum disorders. In Chapter 3, I demonstrate that these brain regions
contain explicit, abstract representations of another feature of others’ mental states:
perceptual source. I find that these representations persist in the face of drastic changes in
developmental history (congenital blindness), providing evidence that these representations
emerge even in the absence of relevant first-person experience. In Chapter 4, I demonstrate
that these cortical regions contain representations of epistemic and emotional features of
others’ beliefs, and that these features are represented along continuous, abstract
dimensions. Finally, in Chapter 5, I extend a model from vision and neuroeconomics —
predictive coding — and explore its application to the neural basis of social cognition.
Together, this work provides a key next step to understanding the neural basis of theory of
mind, by demonstrating that it is possible to find abstract, behaviorally relevant features of
mental state inferences inside cortical regions that support social cognition, and taking a first
step in characterizing their content and format.
Thesis Supervisor: " Rebecca R. Saxe
"
"
"
Associate Professor of Brain and Cognitive Sciences
3 of 179
This page intentionally left blank
4 of 179
Contents
ACKNOWLEDGEMENTS!
9
CHAPTER 1. FUNCTIONAL NEUROIMAGING OF THEORY OF MIND!
11
Introduction!
11
Theory of Mind and the Brain!
12
Theory of mind brain regions respond to mental state content"
Specificity"
Links to behavior"
Summary"
A Strong Hypothesis!
13
22
24
26
27
Objections from theoretical considerations"
Empirical objections"
Summary"
What Now?!
27
33
35
36
Differences between theory of mind regions"
Information within theory of mind regions"
Patterns within theory of mind regions
Magnitude: “more” theory of mind
A First Step!
36
38
41
CHAPTER 2. DECODING MORAL JUDGMENTS FROM NEURAL REPRESENTATIONS OF
INTENTIONS !
45
Introduction!
46
Methods!
48
Participants"
fMRI Protocol and Task"
Experiments 1 and 4
Experiment 2
Experiment 3
Theory of Mind Localizer task
Acquisition and Preprocessing"
Motion and Artifact Analysis"
fMRI Analysis"
Functional Localizer: Individual subject ROIs
Within-ROI Magnitude Analysis
Within-ROI Pattern Correlation Analysis
Whole Brain Pattern Correlation Analysis (Searchlight)
Results!
48
48
52
52
52
55
fMRI task: Behavioral Results"
Motion and Artifact Analysis Results"
fMRI Results: Functional Localizer"
55
55
55
5 of 179
fMRI Results: Response Magnitude"
fMRI Results: Pattern Analysis for Neurotypical"
Within ROI Pattern Analysis for Experiment 1
Within ROI Pattern Analysis for Experiment 2 and 3
Behavioral and Neural Correlation
Whole Brain Searchlight Analysis for Experiment 1,2,3 (NT)
fMRI Results: Pattern Analysis for ASD"
Within ROI Pattern Analysis for Experiment 4 (ASD)
Behavioral and Neural Correlation (ASD)
Whole Brain Searchlight for Experiment 4 (ASD)
ASD symptom severity in Experiment 4 (ASD)
Group Comparison: NT vs ASD"
Within ROI Pattern Analysis
Behavioral and Neural Correlation
Decoding true vs. false beliefs?"
Discussion!
56
58
61
61
62
63
MVPA discriminates accidental and intentional harms in RTPJ of neurotypical adults"
Pattern discrimination of intentional versus accidental harm is correlated with moral
judgment"
High functioning adults with ASD show atypical neural activity in RTPJ"
Conclusion"
63
64
65
66
Acknowledgments!
67
Appendix!
67
SI Behavioral Results: Results from Raw Behavioral Responses"
Supplementary Results: Pairwise matched subsets of ASD and NT"
Independent effects of magnitude and pattern"
A note on separating the phases of Experiment 1"
67
68
69
70
CHAPTER 3. THINKING ABOUT SEEING: PERCEPTUAL SOURCES OF KNOWLEDGE ARE
ENCODED IN THE THEORY OF MIND BRAIN REGIONS OF SIGHTED AND BLIND ADULTS ! 71
Introduction!
72
Methods!
76
Participants"
Comparison to Bedny et al. (2009)"
fMRI Protocol and Task"
Acquisition and Preprocessing"
Motion and Artifact Analysis"
fMRI Analysis"
Functional Localizer: Individual subject ROIs
Within-ROI Pattern Correlation Analysis
Whole Brain Pattern Correlation Analysis (Searchlight)
Results!
76
76
77
78
79
79
82
fMRI task: Behavioral Results"
Sighted participants
82
6 of 179
Blind participants
Group Comparison
Motion and Artifact Analysis Results"
fMRI Results: Functional Localizer"
fMRI Results: Within ROI Pattern Analysis"
Sighted participants
Blind participants
Across Groups
Whole Brain Searchlight Analysis"
Source (seeing vs. hearing)
Valence (good vs. bad):
Discussion!
83
83
83
86
88
MVPA reveals features of mental state representations"
Knowledge source and theory of mind"
The valence of others’ mental states"
Blind adults represent knowledge source"
Learning about sight"
Conclusion"
Acknowledgments!
88
89
90
91
92
93
94
CHAPTER 4. MENTALIZING REGIONS REPRESENT CONTINUOUS, ABSTRACT DIMENSIONS
OF OTHERS’ BELIEFS !
95
Introduction!
96
Methods!
98
Participants"
Stimuli"
fMRI Task"
Theory of Mind Localizer"
Acquisition and Preprocessing"
Motion and Artifact Analysis"
fMRI Analysis"
Modeling
Whole-brain random effects (univariate)
Defining Individual subject ROIs
Within-ROI Multivariate Analysis (MVPA)
Item-Wise classification: Testing for continuous representations
Results!
98
98
102
102
102
103
103
108
fMRI task: Behavioral Results"
fMRI task: Classification Results"
Quality
Modality
Quality vs Modality
Directness
Valence
Generalization to other manipulations of evidence quality"
7 of 179
108
108
111
Stimuli
Methods and Analyses
Results
Discussion!
114
Evidence Quality is explicitly represented in RTPJ and RMSTS"
The modality of others’ evidence is explicitly represented in right and left TPJ"
No region could decode directness: first-person experience versus hearsay"
The valence of others’ emotions is represented in VMPFC "
Neural representations of abstract features"
Why track epistemic features of beliefs?"
Summary and Conclusion"
Acknowledgments!
114
115
115
116
116
118
119
119
CHAPTER 5. THEORY OF MIND: A NEURAL PREDICTION PROBLEM !
121
Introduction!
122
A predictive coding framework!
123
The sources of predictions!
126
Predicting goal directed action"
Predicting beliefs and desires"
Predicting preferences and personalities"
Beyond error!
128
130
132
133
Distinguishing prediction from attention"
Finding the prediction signal"
Using predictive coding to test the neural computations underlying theory of mind"
Acknowledgements!
133
136
137
138
CHAPTER 6. GENERAL DISCUSSION!
Summary!
139
139
Where We Are!
140
Content in theory of mind brain regions"
A feature-based approach to mental state representations"
Beyond Features"
Continuous, abstract representations
Inputs to representations
Replication and generalization
What next?!
140
141
143
146
Information between regions, and between networks"
Development"
Links to behavior: Content and structure of mental state inference
Beyond neuroimaging"
REFERENCES !
146
147
148
151
8 of 179
Acknowledgements
This thesis came into existence with the help and support of many people.
The first thanks goes to my advisor, Rebecca Saxe. Without her thoughtfulness,
intelligence, generosity, support, and excitement, my work would not be what it is, and I
would be not the scientist that I am.
Many thanks to my committee, Nancy Kanwisher, for her clarity in thinking, impeccable
standards, and enthusiasm; Josh Tenenbaum, for always seeing and pushing for the
bigger picture (and making “prior” part of my non-scientific vocabulary), and Fiery
Cushman, for his sharp insights, excitement for science, and clear sense of how things
fit together. Thanks to Amy Skerry, Stefano Anzellotti, Hilary Richardson, and Martin
Hackl who read through, edited, and thought about many portions of this dissertation,
and to Todd Thompson, who spent many hours writing with me in cafes around
Cambridge.
Many thanks to all the people who made MIT the best place possible to get a PhD.
Thanks to people in Saxelab: Marina, Liane, Emile, Mina, Lindsey, Dorit, Stefano, Hyo,
Amy, Ben, Alex, Hilary, David, Alek, Nick, Julianne, Nir, Adele, Laura, Swetha, Natalia
and Mika. Thanks to BCS folk who made our department fun and interesting, especially
folks in Nancy’s lab, Laura’s lab, and Josh’s lab, and to the imaging team downstairs:
Steve, Sheeba, and Atsushi. Thanks to my non-MIT collaborators, Naomi, Rachel M.,
Rachel B., Jenny, Jennie, and Amy, and thanks to the people who nudged me in the
right direction when I arrived at MIT, especially Irene Heim, Adam Albright, and Jim
DiCarlo.
Thanks in particular to Liane and Marina, who are not only great co-authors, but
amazing role models; to Nir and Nick for answering random Matlab questions at 2am; to
Hilary for being a fabulous officemate, coauthor, and coffee partner; to Amy for always
being there to talk through ideas (scientific and otherwise); to Todd for writing dates,
walks, talks, and The Epic Computer Rescue of 2014, to Aeriel for 27 years of cookie
and coffee date, to Sarah for surprisingly effective mindmelding across hundreds of
miles, and to Natasha for reminding me that the best part of getting a PhD is getting to
wear wizard robes for a day.
Thanks to my family, who not only put up with me writing this dissertation in the middle
of a family reunion, but supported me with a continual stream of baked goods. (On the
note of sustenance, thanks also to Area Four for putting good caffeine within walking
distance, and to Tatte for butter cookies.) Thanks to my dad, William Earl, for always
thinking my work was so cool. Thanks to my moms, Pamela and Mindy, for being the
best parents ever, and thanks to Martin for being Most Excellent.
9 of 179
This page intentionally left blank
10 of 179
Chapter 1.
Functional neuroimaging of theory of mind1
Introduction
Social life depends on developing an understanding of other people’s behavior: why
they do the things they do, and what they are likely to do next. Critically, though, actions
are just observable consequences of an unobservable, internal causal structure: the
person’s goals and intentions, beliefs and desires, preferences and personality traits. A
cornerstone of the human capacity for social cognition is the ability to reason about
these invisible causes: if a person checks her watch, is she uncertain about the time or
bored with the conversation? And is she chronically rude or just unusually frazzled? The
ability to reason about these questions is sometimes called having a “theory of mind
(ToM)”.
In the last decade, the number of papers that use human neuroimaging tools to
investigate the neural basis of theory of mind has exploded from four to, as of 2014, well
over 500. Studying ToM with neuroimaging works. Unlike many aspects of higher-level
cognition, which tend to produce small and highly variable patterns of responses across
individuals and tasks, ToM tasks generally elicit activity in an astonishingly robust and
reliable group of brain regions. In fact, convergence came almost immediately. By 2000,
Frith and Frith (2000) concluded that “studies in which volunteers have to make
inferences about the mental states of others activate a number of brain areas, most
notable the medial [pre]frontal cortex [(mPFC)] and temporo-parietal junction [(TPJ)].”
These regions remain the focus of most neuroimaging studies of ToM and social
cognition, more than a decade later (Adolphs, 2009; 2010; Carrington, 2009; Van
Overwalle & Van Overwalle, 2009). This consensus is one of the most remarkable
scientific contributions of human neuroimaging, and the one least foreshadowed by a
century of animal neuroscience.
Nevertheless, most of the fundamental questions about how our brains allow us to
understand other minds remain unanswered; we have mainly discovered where to look.
This dissertation offers the first step in closing that gap: demonstrating that not only do
the parts of cortex that support social cognition contain abstract, behaviorally relevant
features of mental state inference, but that we can find them using fMRI.
In this chapter, I offer a perspective on the contribution that neuroimaging has made to
the science of ToM in the last decade, some thoughts on the contribution that it could
make in the next one, and lay out the specific contributions of this thesis. The chapter
has four sections. Theory of Mind and the Brain reviews the existing evidence for a
basic association between thinking about people’s thoughts and feelings and activity in
1A
version of this chapter appeared as Koster-Hale, J., & Saxe, R. R. (2013). Functional neuroimaging of
theory of mind. In S. Baron-Cohen, H. Tager-Flusberg, & M. V. Lombardo, Understanding Other Minds
Perspectives from developmental social neuroscience (pp. 132–163). Oxford University Press.
11 of 179
this group of brain regions. A Strong Hypothesis discusses some objections, both
theoretical and empirical, to a strong interpretation of this association. What Next?
highlights newer approaches to functional imaging data, how they can contribute to the
future neuroscience of ToM, and their strengths and limitations. A First Step lays out
the goal of the thesis and provides a roadmap for the subsequent chapters.
Theory of Mind and the Brain
Over the course of development, human children make a remarkable discovery: other
people have minds, both similar to and distinct from their own. Other people see the
world from a different angle, have different desires and preferences, and acquire
different knowledge and beliefs. Children learn that other people’s minds contain
representations of the world which are often true and reasonable, but which may be
strange, incomplete, or even entirely false. These discoveries (i.e. “building a Theory of
Mind”) help children to make sense of some otherwise mystifying behaviors: why mom
would eat broccoli, even though there is chocolate cake available, (e.g. Repacholi &
Gopnik, 1997), or why she is looking for the milk in the fridge, even though dad just put
it on the table (e.g. Wimmer & Perner, 1983).
Developmental psychologists historically focused on one key transition in this
developmental process—when and how children come to understand false beliefs.
Assessing understanding of false beliefs has been taken to be a good measure of ToM
capacity because it requires a child to understand both that someone can maintain a
representation of the world, and that this representation may not match the true state of
reality or the child’s own beliefs. In a standard version of the false belief task, children
might see that while their mother thinks the milk is in the fridge (having put it there 5
minutes ago), it is now actually on the table. The children are asked: “Where will she
look for the milk?” or “Why is she looking in the fridge?”. Five-year-old children, like
adults, usually predict that she will look in the fridge, because that is where she thinks
the milk is (Wellman, Cross, & Watson, 2001). Three-year-olds, however, predict that
she will look on the table, explaining that she wants the milk and the milk is on the table
(Heyes, 2014 for further discussion of ToM behavior in pre-verbal children; at least when
asked explicitly; see e.g. Onishi, Baillargeon, & Baillargeon, 2005; Saxe, 2013;
Southgate, Senju, & Csibra, 2007). In fact, when three-year-olds see her look in the
fridge instead, some will go so far as to fabricate belief-independent explanations,
stating that she no longer wants the milk, and must be looking for something else
(Wellman et al., 2001; Wimmer & Perner, 1983)
Building off decades of experience in developmental psychology, the first neuroimaging
studies of ToM also used versions of false belief tasks. Adults, lying in positron emission
tomography (PET) or magnetic resonance imaging (MRI) scanners, read short stories
describing a person’s action (see Figure 1.1), and were asked to explain that action
(usually silently to themselves, to avoid motion artifacts). These early studies revealed
increased levels of blood oxygen and glucose uptake (indirect measures of metabolic
activity, henceforth called “activity”), in a small, but consistent group of brain regions—
left and right TPJ, anterior regions of the superior temporal sulcus (STS), down to the
temporal poles, mPFC, and medial parietal cortex (precuneus, PC; Figure 1.1).
12 of 179
Activity during the false belief task is, of course, far from sufficient evidence that these
brain regions have any role in understanding other minds. This proposition becomes
more compelling after reviewing the range of different experimental tasks and
paradigms that have been used successfully over the years, across laboratories and
countries, each aiming to elicit some aspect of “understanding other minds.” While
some studies used complex verbal narratives, other researchers have used simple
sentences or non-verbal cartoons; some studies explicitly instruct participants to think
about a person’s thoughts, and others have elicited ToM spontaneously. The
heterogeneity of methods, materials, and participant demographics makes it especially
impressive that these studies have converged on the same regions of the brain. In the
next section, I review a sample of these different procedures.
Theory of mind brain regions respond to mental state content
In the original theory of mind neuroimaging experiments, both the content of the
materials and the explicit instructions focused participants’ attention on (or away from)
thinking about someone’s mind. For example, in an early PET study, Fletcher Happe,
Frith, Baker, Dolan, Frackowiak, et al. (1995) told participants that they would be
reading different kinds of verbal passages. Just before each item, the participant was
told what kind of story was coming next. If it was a “mental” story, participants were
instructed that it was “vital to consider the thoughts and feelings of the characters,” and
then shown a story revolving around someone's mental states (see example in Table
1.1). If the story was a “physical” story, participants were first instructed that thinking
about thoughts and feelings was irrelevant and undesirable, then shown a control story
(see example in Table 1.1). After each story, participants were instructed to silently
answer an action-explanation question, such as “Why did the prisoner say that?”.
Glucose consumption increased in the theory of mind brain regions while people read
the mental stories, relative to the control stories.
The same design has been used with non-verbal stimuli. In an early fMRI experiment
(Gallagher et al., 2000), participants both read the stories used by Fletch et al. (1995)
and were shown cartoons depicting visual jokes that either relied on ToM (in which
understanding the joke depended on attribution of either a false belief or ignorance), or
other types of humor (such as puns, idioms, and physical humor). Again, participants
were cued in advance about whether to expect a mental or control cartoon (for
examples, see Figure 1.2). For both types of cartoons, they were asked to silently
contemplate the meaning; for mental cartoons, they were also explicitly instructed to
consider the thoughts and feelings of the characters. With less than 20 minutes of
scanning for each task, Gallagher and colleagues found that the same group of brain
regions showed increased activity for both verbal and non-verbal ToM stimuli; these
regions include the bilateral temporal-parietal junction and the middle prefrontal cortex.
13 of 179
Figure 1.2: Examples from Gallagher et al. (2000). ToM cartoon, non-ToM cartoon, jumbled picture
This activation profile does not depend on explicit instruction to attend to mental state
information. Some experiments elicit thinking about thoughts simply by describing those
thoughts in words. For example, participants can answer questions about people’s
mental characteristics, such as “How likely is Queen Elizabeth to think that keeping a
diary is important?” vs. their physical traits, such as “How likely is Queen Elizabeth to
sneeze when a cat is nearby?” (Lombardo et al., 2010 Table 1.1); or they can read
single sentences describing thoughts (“He thinks that the nuts are rancid”) or facts (“It is
likely that the nuts are rancid;” Zaitchik et al., 2010). In both cases, the items related to
mental states elicited more activity in ToM regions than the control conditions.
Other experiments elicit thinking about thoughts indirectly. Saxe & Kanwisher (2003)
used verbal stories based on either inferences about false beliefs or about physical
events, similar to Fletcher et al. (1995). Participants were not given any explicit
instructions about the different kinds of stories, with a common task across conditions.
In Experiment 1, participants did not give any response, while in Experiment 2, they
responded to fill-in-the-blank questions about details in the stories. Similarly, Mason &
Just (2011) had participants read short stories about actions, and then answer simple
comprehension questions. Critically, the stories elicited spontaneous inferences about
unstated, but implied, events; some of these inferences were about a character’s
thoughts (mental), and some about purely physical events (control). Compared with the
original Fletcher et al. (1995) stories, the stories in these experiments (see examples in
Table 1.1) were shorter, and included less (or no) explicit description of thoughts and
feelings (see also Bruneau & Saxe, 2012). Instead, the thoughts and feelings of the
characters had to be inferred. Listening to the mental stories elicited strong activation in
ToM regions relative to control stories, suggesting consideration of the character’s
thoughts despite the absence of explicit instruction.
14 of 179
Mental
Control
During'the'war,'the'Red'army'capture'a'member'of'the'
Blue'army.'They'want'him'to'tell'them'where'his'armies''
tanks'are.'They'know'that'they'are'either'by'the'sea'or'in'
the'mountains.'They'know'that'the'prisoner'will'not'
want'to'tell'them,'he'will'want'to'save'his'army,'and'so'
he'will'certainly'lie'to'them.'The'prisoner'is'very'brave'
and'very'clever,'he'will'not'let'them'find'his'tanks.'The'
tanks'are'really'in'the'mountains.'Now'when'the'other'
side'ask'him'where'his'tanks'are'he'says,'"They'are'in'the'
mountains."
Two'enemy'powers'have'been'at'war'for'a'very'long'
time.'Each'army'has'won'several'battles,'but'no'the'
outcome'could'go'either'way.'The'forces'are'equally'
matched.'However,'the'Blue'army'is'stronger'than'the'
Yellow'army'in'air'foot'soldiers'and'artillery.'But'the'
yellow'army'is'stronger'than'the'Blue'army'in'air'power.'
On'the'day'of'the'final'battle,'which'will'decide'the'
outcome'of'the'war,'there'is'heavy'fog'over'the'
mountains'where'the'fighting'is'about'to'occur.'LowG
lying'clouds'hang'over'the'soldiers.'By'the'end'of'the'
day,'the'Blue'army'has'won.
Brad'had'no'money,'but'just'had'to'have'the'beautiful'
ruby'ring'for'his'wife.'Seeing'no'salespeople'around,'he'
quietly'made'his'way'closer'to'the'counter.'He'was'seen'
running'out'the'door.
While'playing'in'the'waves,'Sarah’s'Frisbee'went'flying'
toward'the'rocks'in'the'shallow'water.'While'searching'
for'it,'she'stepped'on'a'piece'of'glass.'Sarah'had'to'wear'
a'bandage'on'her'foot'for'a'week.
The'path'to'the'castle'leads'via'the'lake.'But'children'tell'
the'tourists:'‘‘The'way'to'the'castle'goes'through'the'
woods.’’'The'tourists'now'think'that'the'castle'is'via'the.'
woods/lake?
The'sign'to'the'monastery'points'to'the'path'through'the'
woods.'While'playing'the'children'make'the'sign'point'to'
the'golf'course.According'to'the'sign'the'monastery'is'
now'in'the'direction'of'the.'golf'course/woods?
How'likely'is'Queen'Elizabeth'to'think'that'keeping'a'
diary'is'important?
How'likely'is'Queen'Elizabeth'to'sneeze'when'a'cat'is'
nearby?
John'was'on'a'hike'with'his'girlfriend.'He'had'an'
engagement'ring'in'his'pocket'and'at'a'beautiful'
overlook'he'proposed'marriage.'His'girlfriend'said'that'
she'could'not'marry'him'and'began'crying.'John'sat'on'a'
rock'and'looked'at'the'ring.
Joe'was'playing'soccer'with'his'friends.'He'slid'in'to'steal'
the'ball'away,'but'his'cleat'stuck'in'the'grass'and'he'
rolled'over'his'ankle,'breaking'his'ankle'and'tearing'the'
ligaments.'His'face'was'flushed'as'he'rolled'over.
That'morning,'people'sat'around'looking'at'each'other,'
wondering'if'they'were'dreaming,'because'everything'
looked'purple.'Some'people'were'shocked.'Some'people'
thought'that'it'was'funny'to'see'everybody'all'purple.'
But'even'the'smartest'scientists'didn’t'know'what'had'
happened.'
The'whole'world'had'turned'purple'overnight.'Just'about'
everything'was'purple,'included'the'sky'and'the'ocean'
and'the'mountains'and'the'trees.'The'tallest'skyscrapers'
and'the'tiniest'ants'were'all'purple.'The'bicycles'and'
furniture'and'food'were'purple.'Even'the'candy'was'
purple.'
In'spite'of'her'neighborhood,'Erica'has'a'strong'dislike'of' Erica'lives'in'Los'Angeles.'One'night'recently'she'was'in'a'
violence,'and'believes'that'conflicts'can'usually'be'
bar'where'a'fight'broke'out'between'two'drunk'men'and'
resolved'without'fists.'
she'was'caught'in'between.
Sam'thinks'he'can'grow'trees'with'fruit'that'taste'like'
pizza.'How'likely'is'it'that'Sam'wants'these'trees'for'a'
treehouse'too?
In'the'backyard'are'trees'with'fruit'that'taste'like'pizza'
when'ripe.'How'likely'is'it'that'these'trees'can'be'used'
for'building'a'treehouse?
Table 1.1. Samples of experiments that elicit thinking about thoughts and feelings by
manipulating the content of the stimuli. Sample stimuli from Fletcher et al. 1995, Mason & Just 2010,
Perner et al. 2006, Lombardo et al. 2010, Bruneau et al. in 2013, Saxe et al. 2009, Saxe & Wexler 2005,
Young et al. 2010.
15 of 179
In fact, when participants read a story, they appear to automatically represent the
thoughts and feelings of the characters in order to make sense of the plot, even if
instructed to perform an orthogonal task. For example, Koster-Hale & Saxe (2011) had
participants read short verbal stories, and then make a delayed-match-to-sample
judgment, indicating whether a single probe word occurred in the story (match) or not
(non-match); half of the stories described a false belief and half described physical
representations. Despite the word-level task (and no mention of ToM in the explicit
instructions), the contrast revealed activation across the ToM network. Similarly, in two
fMRI experiments with children aged 5–12 years, children heard child-friendly verbal
stories, describing characters’ thoughts and feelings vs. physical events; children
answered orthogonal (delayed-match-to-sample) questions about each story. As with
adults, there was increased activation in the ToM network when children were listening
to stories involving thoughts and feelings (Gweon, Dodell-Feder, Bedny, & Saxe, 2012;
Saxe, Whitfield-Gabrieli, Scholz, & Pelphrey, 2009).
Sommer, Döhnel, Sodian, Meinhardt, Thoermer, & Hajak (2007) also use non-verbal
stimuli to focus participants’ attention on the thoughts of a character, but without
explicitly cuing the condition. They showed participants a series of cartoon images
depicting a story—Betty hides her ball, Nick moves it, and then Betty comes back to
look for her ball. In half of the trials, Betty looks into the box where she thinks the ball is
(the expected condition); in the other half, she looks into the other box (the unexpected
condition). Participants judged whether, based on the character’s beliefs, the
character’s action was expected or unexpected. Rather than explicitly labeling the
mental and control conditions, a key contrast was between the beginning of the trial,
before mental state inferences were possible, and the end of the trial, when participants
had presumably made belief inferences to complete the task. The second key contrast
was between trials in which Betty had a false belief (so that predicting her action
required considering her thoughts) and trials in which Betty knew where the ball was all
along (so her action could be predicted based on the actual location of the ball).2 Both
contrasts revealed activity in ToM brain regions (Figure 1.3).
As well as depicting complex stories lines, non-verbal stimuli can be endowed with
mentalistic content by altering the movements of simple geometric shapes (Heider &
Simmel, 1944). For example, in a PET study, Castelli, Happé, Frith, & Frith (2000)
showed participants animations of two triangles moving around. Participants were
instructed that while some of the triangles “just move about with random movement [...]
disconnected from each other” (the control condition), other animations would show
“two triangles doing something more complex together, as if they are taking into account
their reciprocal feelings and thoughts [...] for example, courting each other,” (the mental
2
In the original study, this contrast could have been due to a difference between false vs. true beliefs, or
between representing a belief (required for the false trials) and making a prediction based solely the
actual location of the ball (possible for the true beliefs). Subsequent work has shown that activation
observed by Sommer et al. 2007 is due to the latter: individuals often simply reduce the information they
need to process by choosing not to represent true beliefs as a mental state (Apperly et al., 2007), and
when this is controlled for, neuroimaging has revealed indistinguishably high activation for true and false
beliefs (Döhnel, Schuwerk, Meinhardt, Sodian, Hajak, & Sommer, 2012; Jenkins & Mitchell, 2010; Young
et al., 2010b).
16 of 179
condition). Participants watched the animations, and then described what the triangles
were doing. ToM animations elicited more activity than the random animations in the
TPJ, the nearby superior temporal sulcus (STS), and the mPFC.
Figure 1.3: A False Belief trial from Sommer et al. (2007)
Another procedure for eliciting spontaneous ToM in the scanner was developed by
Spiers & Maguire (2007). Participants engaged in naturalistic actions (e.g. driving a
taxicab through bustling London streets) in a rich virtual reality environment (Figure
1.4). After the scan and without prior warning, participants reviewed their performance,
and were asked to recall their spontaneous thoughts during each of the events. These
recollections were coded for content concerning the thoughts and intentions of the
taxicab customers and the other drivers and pedestrians on the road (e.g. “I reckon that
she’s going to change her mind”) and used these coded events to predict neural
responses. They found that when participants were thinking about someone else’s
intentions, but not during other events, regions in the ToM network showed increased
activity.
Other studies have targeted spontaneous consideration of others’ intentions by asking
participants to make moral judgments (see Young & Waytz, 2013 for a review). Morally
relevant facts appear to rely in part on consideration of intentions (Cushman, 2008), and
evoke increased activity in the ToM network, relative to other facts in a story (Young &
Saxe, 2009). The same brain regions are recruited when participants are forced to
17 of 179
choose between acting on a personal desire vs. a conflicting moral principle, compared
to deciding between two conflicting personal desires (Sommer et al., 2010). These
regions are also recruited in children watching an animation of one person intentionally
harming another, compared to animations of other painful and non-painful situations
(Decety, Michalska, & Akitsuki, 2008).
Figure 1.4: Example stimuli from Spiers and Maguire (2006)
18 of 179
Figure 1.5: Examples of cartoons used in Schnell et al. (2010) and Walter et al. (2010). For each
frame, participants were asked to decide whether (a) the mood of the protagonist got better or worse, or
(b) the number of living things they could see increased or decreased.
Thinking about thoughts and feelings can be also manipulated by changing the task
while holding the stimuli constant. In an early PET study, Goel, Grafman, Sadato, &
Hallett (1995) showed participants sets of 75 photographs of objects, some modern and
familiar, and some from pre-fifteenth century North American aboriginal culture.
Participants either judged whether the object was elongated along the principle axis (the
control task) or whether “someone with the background knowledge of Christopher
Columbus could infer the [object’s] function” (the mental task). They found increased
ToM activation when participants were considering Christopher Columbus, but not when
making the physical judgments. Similarly, Walter and colleagues showed participants
sequences of three cartoon images (Schnell, Bluschke, Konradt, & Walter, 2010; Walter
et al., 2010). Participants either judged, on each picture, whether “the protagonist feels
worse/equal/better, compared to the previous picture” (the mental task) or whether “the
number of living beings [in the image is] smaller/equal/greater, compared to the
previous picture” (the control task, Figure 1.5).
female
male
Figure 1.6: Reading the Mind in the Eyes (Baron-Cohen, 1991; Adams et al. 2009)
In another set of studies, Baron-Cohen and others (Adams et al., 2010; BaronCohenMO'Riordan, 1999; Platek, 2004) showed participants pictures of a person’s
eyes. Participants pressed a button to indicate either the mental/emotional state of the
19 of 179
person in the picture (e.g. embarrassed, flirtatious, worried; the mental task) or their
gender (the control task, Figure 1.6). Mitchell, Banaji, & MacRae (2005) asked
participants to either judge either how happy a person was to be photographed (mental)
or how symmetric their face was (control). In all of these cases, the mental tasks
activated the ToM regions more than the control tasks.
Spunt and colleagues have recently developed a clever paradigm for eliciting ToM using
a simple task manipulation. In their first experiment, Spunt, Satpute, & Lieberman
(2010) showed participants pictures of simple human actions, (e.g. a person riding a
bike), and instructed them to silently answer one of three questions: why the person is
doing the action (e.g. to get exercise), what the person is doing (e.g. riding a bike), or
how the person is doing it (e.g. holding handlebars, Figure 1.7). These questions
require successively less consideration of the mind of the person and, correspondingly,
showed successively less ToM region activity. Spunt & Lieberman (2012) replicated the
result using a similar paradigm with brief naturalistic film clips of facial expressions of
emotions. Participants judged either how the person is expressing their emotion (e.g.
“looking down and away,” the control task) or why she is feeling that emotion (e.g. “she
is confused because a friend let her down,” the mental task). Again, ToM regions were
recruited more when thinking about why than how.
Figure 1.7 Spunt et al. (2010): parametric design of why, what and how?
Using these types of parametric designs and analyzing the continuous magnitude of
response in ToM regions may provide a powerful tool for studying the neural basis of
ToM, especially in combination with computational models of ToM, which offer
quantitative predictions of both when and how much (or how likely) people are thinking
about others’ thoughts. For example, Bhatt, Lohrenz, Camerer, & Montague (2010)
created a competitive buying and selling game in which participants could try to bluff
about the value of an object. The authors predicted that reliance on ToM would increase
in proportion to the riskiness of the bluff, i.e. the discrepancy between the object’s true
value and the proposed price. Consistent with this idea, a region near the right TPJ
showed activity correlated with bluff riskiness across trials. They suggested that
participants may be more likely to engage in ToM, engage in more ToM, or engage in
ToM for longer, when they are making a riskier bluff relative to a less risky bluff, and this
20 of 179
relationship is continuous over a large range of possible risks. Similarly, Behrens et al.
(2008) manipulated the trustworthiness of a confederate over time: the other player
provided “advice” to the participant; this advice shifted over the experiments, so that it
was reliable in some phases, and unreliable in others. The response in ToM regions
tracks with trial-by-trial error in expectations about the informant’s reliability.
Finally, along with manipulating thinking about thoughts across stimuli and tasks, it is
also possible to look at when participants are thinking about thoughts within a single
ongoing stimulus and task. Stories about human actions and beliefs can be broken
down into sections, separating the description of the background and set-up from the
specific sentence that describes or suggests a character’s mental states. Thinking about
other minds can thus be pinned to a specific segment within an ongoing story. The right
temporo-parietal junction, in particular, shows activity at the point within a single story
when a character’s thoughts are mentioned (Mason & Just, 2011; Saxe & Wexler, 2005;
Young & Saxe, 2008). A similar manipulation, dividing a 60-second story into 20-second
segments, only one of which has mental information, has been used in children (Saxe et
al., 2009).
In sum, neuroimaging experiments on understanding other minds produce similar
results, across a wide range of participants, methods, and materials. Similar
experiments have been conducted in Britain, the USA, Japan, Germany, China, the
Netherlands, and Italy (e.g. Leshinskaya & Caramazza, 2014; Moriguchi et al., 2006
Markus van Ackeren, personal communication; Perner, Aichhorn, Kronbichler, Staffen, &
Ladurner, 2006; Schnell et al., 2010). The same brain regions are found in participants
ranging from 5 years old (e.g. Decety et al., 2008; Gweon et al., 2012; Saxe et al.,
2009) to at least 65 years old (e.g. Bedny, Pascual-Leone, & Saxe, 2009; Fletcher et al.,
1995). These regions recruited in adults with high functioning autism (Dufour et al.,
2013), adults who have been completely blind since birth (Bedny et al., 2009); ongoing
work by my colleagues and I suggests they are found in congenitally deaf adults as well.
As described above, the same set of regions responds whether the stimuli is presented
in text or with pictures, visually or aurally. Participants can be thinking about the
thoughts and feelings of a real person or a fictional character.3 The stimuli can include
complex narratives, or just a single thought; participants can be instructed to consider
others’ thoughts and feelings, or be led to do so spontaneously.
The range of tasks, stimuli, and populations make it all the more striking that these
experiments converge on the same conclusions. A consistent group of brain regions
shows increased metabolic activity across all of these experiments in the “mental” or
“theory of mind” condition: regions in bilateral temporo-parietal junction, medial parietal/
precuneus, medial prefrontal cortex, and anterior superior temporal sulcus. Though this
generalization is striking on its own, a key question is: why? Which cognitive process,
invoked by all of these diverse tasks, is specifically necessary and sufficient to elicit
activity in these brain regions?
3
Interestingly, participants can also be assigning hypothetical thoughts and preferences to themselves
(e.g. Lombardo et al., 2010; Vogeley et al., 2001). Note, however, that not all metacognition elicits ToM
activity. The link between attributing hypothetical thoughts to the self, vs. other kinds of metacognition, is
not completely clear (see Saxe & Offen, 2009).
21 of 179
Specificity
During the initial discovery of the ToM brain regions, the first experiments (e.g. Fletcher
et al., 1995; Gallagher et al., 2000) compared two conditions that differed on multiple
dimensions. Compared with the control stories,
the stories about false beliefs also included
more individual characters, more specific
human actions, more implied human emotions,
more invisible causal mechanisms, more social
roles, more unexpected events, more demand
to consider false representations of the world,
different syntax, and so on. In fact, an inherent
risk of such complex stimuli is that there may
be hidden dimensions along which the stimuli
grouped into separate conditions differ and that
it is these differences, rather than the intended
manipulation, that lead to differential brain
activity.
Given all these dimensions, how can we infer
which are the necessary and sufficient features
that led to activity in each region during a given
task? One approach is to try to design an
experiment that contrasts minimal pairs: stimuli
and tasks that differ only in one key dimension,
but are exactly identical on all other
dimensions. Taking this approach, Saxe,
Schulz, and Jiang (2006) were able to match
Figure 1.8: Example stimuli from Saxe et
both the stimulus and the participant’s
al. (2006)
response, using task instructions to change just
how the participants construed the stimulus.
The stimulus was a stick-figure animation of a girl. Modeled after a false-belief transfer
(change of location) task, a bar of chocolate moved from one box to another, while the
girl either faced toward the transfer or away. In the first half of the experiment, rather
than introducing the task as a false-belief task, participants were trained to treat the
stick-figure as a physical cue to the final location of the chocolate bar using three rules,
including the critical Rule 1: ‘Facing = last; Away = first. If the girl is facing the boxes at
the end of the trial, press the button for the last box. If the girl is looking away from the
boxes, press the button for the first box.” (Figure 1.8) Participants were accurate in the
task, but found it difficult and unnatural. In the second half of the experiment,
participants were then told that for Rule 1, another strategy was possible: namely to
view the stick-figure as a person and to consider that person’s thoughts. Rule 1 was
equivalent to judging, based on what the character had seen, where she thought the
chocolate was. Both ways of solving “Rule 1” generate the same behavioral responses,
but only in the second half of the experiment were participants construing Rule 1 as
referring to a character’s thoughts. Saxe et al. (2006) found that, though the stimuli and
the responses were identical across tasks, right TPJ activity was significantly higher in
22 of 179
the second half, when participants were using ToM to solve the puzzle, rather than the
simple association rule.
Another approach to dealing with the many dimensions of ToM stimuli is to
systematically vary or match each of these dimensions in a long series of experiments.
Although each experiment has many differences between the mental and control
conditions, across the whole set of experiments, most other kinds of differences are
eliminated, leaving only one systematic factor: thinking about thoughts.
For example, both Gallagher et al. (2000) and Kobayashi, Glover & Temple (2007)
showed that none of the low-level features of the original verbal stimuli (e.g. number of
nouns, number of straight edges, retinal position) are necessary to elicit activity in these
brain regions, because they found the same patterns of activity in response to verbal
false belief stories and non-verbal false belief cartoons. Within verbal stories, it is not
necessary to explicitly state a character’s thoughts or beliefs: there is activity in these
regions both when people read about a character’s thoughts and when they infer those
thoughts from the character’s actions (Mason & Just, 2011; Young & Saxe, 2009). Nor is
it necessary that the beliefs in question be false: true beliefs, false beliefs, and beliefs
whose veracity is unknown are all sufficient to elicit robust activity in these brain regions
(Döhnel et al., 2012; Jenkins & Mitchell, 2010; Young, Nichols, & Saxe, 2010b).
Other experiments showed that the presence of a human character in the stimuli is not
sufficient: stories that describe a character’s physical appearance, or even their internal
(but not mental) experiences, like hunger or queasiness or physical pain, elicit much
less response than stories about the character’s beliefs, desires, and emotions (Bedny
et al., 2009; Bruneau & Saxe, 2012; Lombardo, Chakrabarti, Bullmore, Baron-Cohen, &
Consortium, 2011; Saxe & Powell, 2006). It is also not sufficient for a story to describe
invisible causal mechanisms (like melting and rusting, Saxe & Kanwisher, 2003),
unexpected events (like a ball of dough that rises to be as big as a house, Young,
Dodell-Feder, et al., 2010b; Gweon et al., 2012), or people’s stable social roles
(including kinship and professional relationships, Saxe & Wexler, 2005; Gweon et al.,
2012), if the story does not also invoke thinking about a person’s thoughts.
One particularly important dimension to test was whether considering any
representation of the world, mental or otherwise, would be sufficient to elicit activity in
these brain regions. Understanding other minds often requires the ability to suspend
one’s own beliefs and knowledge, and consider the world as it would seem from another
perspective. These cognitive processes have been called meta-representation (the
ability to conceive of distinct representations of the world; Aichhorn, Perner, Weiss,
Kronbichler, Staffen, & Ladurner, 2009; Perner, 1991; Perner et al., 2006), and
decoupling (the ability to suspend one’s own knowledge in order to respond from a
different perspective, Leslie & Frith, 1990; Liu, Sabbagh, & Gehring, 2004). Since metarepresentation and decoupling are such essential ingredients of understanding other
minds, and especially understanding false beliefs, many scientists initially hypothesized
that brain regions recruited by false belief tasks most likely performed one of these two
functions. To test this hypothesis, we need stimuli or tasks that require metarepresentation and decoupling, but are not about understanding other minds. Currently,
the best such example are stories about “false signs” and “false maps” (Zaitchik, 1990).
23 of 179
Like beliefs, signs and maps represent (and sometimes misrepresent) reality. Thinking
about the world as depicted in a map requires the capacity for meta-representation;
when the map is wrong, reasoning about the world as it seems in the map requires
decoupling from one’s own knowledge of reality. Nevertheless, stories about false signs
and maps elicit much less activity in these brain regions (especially right TPJ) than
stories about beliefs (Aichhorn et al., 2009; Perner et al., 2006; Saxe & Kanwisher
2003).
In sum, tasks and stimuli that require, or robustly suggest, thinking about thoughts lead
to activity in these regions. Thinking about thoughts can be manipulated by changing
participants’ instructions for the same stimuli, or by changing the stimuli with the same
instructions. Very similar stimuli and tasks, however, which focus on physical objects,
physical representations, or externally observable properties, do not lead to activity in
these regions.
Links to behavior
The review in the previous section shows that tasks and stimuli that evoke thinking
about thoughts also elicit metabolic activity in the ToM brain regions. However, there is
even stronger evidence for a link between activity in these regions and understanding
other minds: across trials, across individuals, and across development, performance on
behavioral tests of ToM is related to brain activity in these same regions.
Most adults pass standard laboratory ToM tasks 100% of the time, leaving little room for
inter-individual variability in accuracy. However, by using tasks with no simple right
answer, it is possible to reveal individual differences in mental state attribution. Imagine,
for example, learning about a girl Grace, who was on a tour of a chemical plant. While
making coffee, Grace found a jar of white powder, labeled “sugar,” next to the coffee
machine. She put the white powder, which was actually a dangerous toxic poison, in
someone else’s coffee, who drank the poison and got sick. Is Grace morally
blameworthy? These scenarios require weighing what Grace intended (her mental
state) against what she did (the outcome), and participants disagree in their judgments;
some people think she is completely innocent (because she believed the powder was
sugar), whereas others assign some moral blame (because she hurt someone). This
difference is correlated with neural activity during moral judgments across individuals;
the more activity there was in a participant’s right TPJ, in particular, the more the
participant forgave Grace for her accidental harms (Young & Saxe 2009b).
An alternative strategy is to measure the quantity and quality of people’s spontaneous
mentalistic attributions to ambiguous stimuli. For example, when viewing the simple
animations of a small and large moving triangle (Castelli et al., 2000), people generate
very rich mentalistic interpretations from the simple movements depicted in these stimuli
(e.g. “the child is pretending to do nothing, to fool the parent”). People differ in the
amount, and appropriateness, of the thoughts and feelings that they infer from the
animations. People who have more activity in ToM brain regions, while watching the
animations give more appropriate descriptions of the triangles’ thoughts and feelings
after the scan (Moriguchi et al., 2006). Relatedly, Wagner, Kelley, & Heatherton (2011)
showed participants still photographs of natural scenes, approximately a quarter of
24 of 179
which contained multiple people in a social interaction. Participants performed an
orthogonal categorization task (“animal, vegetable, mineral?”). Individuals who scored
high on a separate questionnaire, measuring tendencies to think about others’ thoughts
and feelings (the “empathizing quotient”; Baron-Cohen & Wheelwright 2004; Lawrence,
Shaw, Baker, Baron-Cohen, & David, 2004) also showed higher activity in mPFC in
response to photographs of interacting people.
Differences in ToM are easier to find in young children, who are still learning how to
understand other minds. Interestingly, two recent studies from our laboratories suggest
that getting older, and getting better at understanding other minds, is associated,
overall, not with more activity in “ToM brain regions,” but with more selective activity.
Children from 5 to 12 years old all have adult-like neural activity when listening to
stories about characters’ thoughts and feelings. What is different is that ToM regions in
younger children show similarly high activity when listening to any information about
characters in the story, including the characters’ physical appearance or social
relationships (Saxe et al., 2009; Gweon et al., 2012), whereas in older children and
adults, the ToM brain regions are recruited only when listening to information about
thoughts and feelings (Saxe & Powell, 2006; Saxe et al., 2009). This developmental
difference in the selectivity of the ToM brain regions is correlated with age, but also with
performance outside the scanner on difficult ToM tasks (Gweon et al., 2012). Moreover,
the correlation between neural “selectivity” and behavioral task performance remains
significant in the right TPJ, even after accounting for age.
One limitation of all the foregoing studies is that they are necessarily correlational. The
strongest evidence that some brain regions are involved in a cognitive task is to show
that disrupting those regions leads to biases or disruption in task performance.
Transcranial magnetic stimulation (TMS) offers a tool for temporarily disrupting a
targeted brain region. Young, Camprodon, Hauser, Pascual-Leone, & Saxe (2010a)
compared people’s moral judgments following TMS to either right TPJ or a control brain
region 5 cm away. TMS to right TPJ, but not the control region, produced moral
judgments temporarily biased away from considerations of mental state information.
Innocent accidents appeared more blameworthy, while failed attempts appeared less
blameworthy, as though it mattered less what the agent believed she was doing, and
more what she actually did. People didn’t lose the ability to make moral judgments
altogether; they still judged that it is completely morally wrong to intentionally kill, and
not wrong at all to simply serve someone soup. Disrupting the right TPJ thus appears to
leave moral judgment overall intact, but impairs people’s ability to integrate
considerations of the character’s thoughts into their moral judgments. Converging
evidence comes from another TMS study: TMS to right TPJ made adults slower to
recognize a false belief, in a simple (non-moral) false belief task (Costa, Torriero, &
Oliveri, 2008).
Another way to study the necessary contributions of a brain region is to work with
people who have suffered permanent focal (i.e. local) damage to that region, typically
due to a stroke. Samson, Apperly, and colleagues (Apperly & Butterfill 2009; Apperly,
Samson, & Humphreys, 2005; Samson, Apperly, & Humphreys, 2007) have conducted
a series of elegant studies using this approach. Initially, these authors tested a large
group of people, with damage to many different brain regions, on a set of carefully
25 of 179
controlled tasks. They then identified individuals with a specific pattern of performance:
individuals who passed all the control tasks (e.g. measuring memory, cognitive control,
etc.), but still failed to predict a character’s actions based on their false beliefs. Next, the
scientists used a lesion-overlap analysis to ask which brain region was damaged in all,
and only, the patients with this diagnostic profile of performance. The answer was the
left TPJ, one of the same brain regions identified by fMRI4.
Summary
The literature from the last 15 years thus suggests a generalization—there are cortical
regions in the human brain where activity is associated with understanding other minds
in three ways:
1.
Metabolic measures of activity reliably increase when the participant is thinking
about thoughts, across a wide range of stimuli and tasks, but not in response to a
variety of similar control tasks and stimuli.
2.
Activity is correlated with behavioral measures of thinking about thoughts.
3.
Disrupting activity leads to deficits in thinking about thoughts.
So far, these claims are relatively uncontroversial. As noted above, there is a broad
consensus in social cognitive neuroscience. However, much controversy remains about
the proper interpretation of these data.
Exactly what is the nature of these regions, their functions, and their contribution to
thinking about thoughts? Here’s a strong hypothesis: one or more of these regions has
the specific cognitive function of representing people’s mental states and experiences—
that is, of thinking about thoughts. Whenever we are thinking about thoughts, there are
neurons in these regions firing. These neurons are gathered in spatial proximity (i.e. into
a “region”) because they have related computational properties, that are distinct from
the computation properties of neurons in the surrounding cortex.5 The pattern of firing,
in space and time, of these neurons encodes aspects of someone’s thoughts. As an
analogy, consider the way MT neurons encode speed and direction of motion, and face
area neurons encode aspects of facial features that are relevant to face identity
(Freiwald, Tsao, & Livingstone, 2009; Georgopoulos, Schwartz, & Kettner, 1986). In the
proposed hypothesis, the neurons in the ToM brain regions encode aspects and
dimensions of inferred thoughts. Scrambling the pattern of activity in these neurons
would therefore lead to an inability to discriminate one inferred mental state from
another, for example, making all minds appear homogenous: people might all seem to
have the same desires and preferences, and the same knowledge and beliefs. More
4
It is worth noting that the effects of lesions to the right TPJ, one of the regions argued to be most
selective for ToM, haven’t yet been effectively tested. The candidate participants all had extensive and
diffuse damage to the right hemisphere, and failed the control tasks, making it impossible to test ToM
deficits specifically.
5
These distinct properties may derive entirely from patterns of connectivity, not from the structure of the
neurons themselves.
26 of 179
serious damage to these regions might make it impossible to think about other minds at
all, without similarly impairing the rest of cognition.
There certainly is not enough evidence to prove that this strong hypothesis is right; a
more immediate question is whether it is obviously wrong. There are at least two
classes of potential objections: theoretical arguments, based on general principles of
how the brain works, and empirical arguments, based on the results of other
experiments in cognitive neuroscience. In the next section, I describe some of these
objections, and some of our responses to them.
A Strong Hypothesis
Objections from theoretical considerations
Many authors have expressed discomfort with the project of trying to link specific
cognitive functions with delineated brain regions. For example, a decade after their
original chapter, Chris and Uta Frith wrote: “We passionately believe that social
cognitive neuroscience needs to break away from a restrictive phrenology that links
circumscribed brain regions to underspecified social processes” (Frith & Frith, 2012).
Others have echoed this accusation of phrenology; for example, Bob Knight criticizes
the “phrenological notion that a given innate mental faculty is based solely in just one
part of the brain” (Knight, 2007), and William Uttal recently argued that “any studies
using brain images that report single areas of activation exclusively associated with any
particular cognitive process should a priori be considered to be artifacts of the arbitrary
thresholds set by investigators and seriously questioned” (Uttal, 2011). Most such
theoretical objections include variations on three themes: social cognitive
neuroscientists are accused of (incorrectly) viewing regions as (1) functioning in
isolation, (2) internally functionally homogenous, and (3) spatially bounded and distinct.
Here, I address each of these concerns in turn.
First, does claiming that a region has a specific function (e.g. in thinking about thoughts)
entail suggesting that this region functions in isolation? To put it more extremely, does
the hypothesis claim that, for example, the right temporo-parietal junction (RTPJ) could
pass a false belief task on its own? Obviously not. Performing any cognitive task
necessarily depends on many different cognitive and computational processes, and
therefore brain regions. No interesting behavioral task can be accomplished by a single
region. The tasks of the mind and brain—recognizing a friend, understanding a
sentence, deciding what to eat for dinner—must all be accomplished by a long
sequence of processing steps, passing information between many different regions or
computations, from sensory processing all the way to motor action. The function of a
neuron or a brain region should never be identified with completing a cognitive task.
Thus, for example, “passing a false belief task” is not even a candidate function of a
brain region. Any time an individual passes a false belief task, many brain regions—
involved in perceiving the stimuli, manipulating ideas in working memory, making a
decision, and producing a response—will all be required (Bloom & German, 2000).
More generally, functions of regions (or neural population, regardless of spatial
organization) are likely to receive a class of inputs, and transform them into output,
27 of 179
which make different information relatively explicit. Therefore, the specific questions
about any neural population should include: what input does it receive, what output
does it produce, and what information is made explicit in that transformation? Of course,
the answers to these questions will require us to understand the position of this neural
population within a larger network, especially when characterizing a region’s input and
output. In the case of the ToM regions, a related question concerns the relationships
between the different regions within the network. At least five cortical regions are
commonly recruited during many different social cognitive tasks: how is information
passed between, and transformed by, each of these spatially distinct regions?
Thus, studying the function of a brain region means studying in isolation one component
of a system that could never function in isolation. This description may sound ominous,
but scientific progress frequently requires us to break complex systems into component
parts. While the pieces could not function in isolation, understanding their isolated
contributions is necessary to understanding the function of the integrated system. For
any given neural population, it is reasonably to ask: what classes of stimuli and tasks
predictably and systematically elicit increased activity in the population as a whole?
Which dimensions of stimuli lead to activity in distinct subpopulations of neurons? Both
traditional and new fMRI methods help answer these questions, albeit somewhat
indirectly (more on this, in What Now?).
The second objection is that studying brain regions leads to the false assumption that
groups of spatially adjacent neurons are functionally homogenous. The regions we
study in fMRI are orders of magnitude larger than the true computational units of brain
processing, the neurons. Changes in blood oxygenation measured by fMRI inevitably
reflect averages over the activity of many thousands of individual neurons. Why is it
useful to study oxygen flow to chunks of cortex approximately 1–5 cm2 in size, which
are so much bigger than neurons, but so much smaller than the networks required to
complete a task?
Our suggestion is that there is no reason, a priori. It just happens, as a matter of
empirical fact, that many interesting computational properties of the brain can be
detected by studying the organization of neural responses on this scale. Aggregating
the responses of neighboring neurons often produces informative population averages.
Results obtained via fMRI reflect the same distinctions found in directly observable
population codes (e.g. Kamitani & Tong, 2005; Kriegeskorte, Bandettini, & Bandettini,
2007). This empirical fact may have a theoretical explanation. Neurons with similar or
related functions may be spatially clustered to increase the computational efficiency of
frequent comparisons (e.g. lateral inhibition). Blood-oxygen delivery to the cortex may
follow the contours of neural computations, to increase the hemodynamic efficiency of
simultaneously getting oxygen to all of the neurons that need it (Kanwisher, 2010).
However, these arguments are not necessary premises; for fMRI to be useful, we only
need the empirically observable fact that useful and reliable generalizations can be
made for hemodynamic activity in patches of cortex at the spatial scale of millimeters.
Of course, though, neurons within a region or an fMRI voxel are never functionally
homogenous. Consider the analogy of primary visual cortex—neurons in V1 have visual
receptive fields, meaning that activity can be induced by a pattern of bars of light falling
28 of 179
on a specific region of the retina. However, neurons in V1 differ from one another in
where on the retina one should place the light (retinotopy), how large the pattern should
be (size and spatial frequency preferences), and the orientation to which the bars
should be rotated (orientation selectivity), to elicit a maximal response. There is also a
separate (but systematically interleaved) population of neurons for which the response
depends on color, but not orientation or size. Furthermore, some neurons primarily send
information to subsequent regions of visual cortex (e.g. excitatory neurons) whereas
other neurons primarily modulate the response of neighboring neurons in V1 (e.g.
inhibitory interneurons). As far as we know, the metabolic activity measured by fMRI
reflects a combination of activity in all of these different populations. Consequently, we
should never assume that the amount of “activity” we measure in a region with fMRI
represents (or would correlate very well with) the rate of firing of any individual neuron
inside that region. Similarly, we cannot assume that if two different stimuli or tasks elicit
similar magnitudes of activity in a region, then they are eliciting responses in the same,
or even shared, neural populations. Completely non-overlapping subpopulations of
neurons could produce the same magnitude of fMRI activity within a region. Any
interpretation of fMRI data must be sensitive to this possibility. In fact, studying the
organization of functional subpopulations within a region (e.g. which dimensions of
stimuli are represented by distinct subpopulations within a region) may be one of the
most powerful ways that fMRI will contribute to the neuroscience of ToM. I describe
these methods in greater detail in What Now?.
The third potential objection is that studying regions with fMRI leads researchers to
imagine boundaries between discrete regions, when the truth is a continuous
distribution of neural responses over the cortical sheet. The data we described in the
first section shows that cortex is not functionally homogenous with regard to theory of
mind, and regional distinctions are not all “artifacts of arbitrary thresholds.” Still, there is
a legitimate reason why cognitive neuroscientists may be reluctant to call any reliable
functional regularity discovered by fMRI a “region.” These “regions” may turn out to be
just one piece of a larger continuous functional map over cortex, not computationally
distinct areas of their own (Kriegeskorte, Goebel, & Bandettini, 2006).
Cortical responses at scales measurable by fMRI are organized along multiple
orthogonal spatial principles. One is the division of cortex into cytoarchitectonic “areas,”
like primary visual cortex (V1) and primary auditory cortex (A1). Orthogonal to the
division of cortex into areas are topographic principles. Most visual regions, for
example, are organized by retinotopy: moving across the cortical sheet, the region of
the visual field eliciting a maximal response varies smoothly, covering the whole visual
field from fovea to periphery, top to bottom, and left to right. Likewise, multiple distinct
motor areas are organized by somatotopy, and auditory areas by tonotopy.
These orthogonal principles of cortical organization create a challenge for cognitive
neuroscientists, because in charting new territory, away from well-understood sensory
and motor systems, we may claim to discover new functional regions associated with
higher-order cognitive processes, which are really just one end of a larger map (cf.
Konkle & Oliva, 2012). To make the puzzle concrete, imagine looking at functional
responses to visual stimuli across occipital cortex for the first time, without the benefit of
the history of visual neuroscience. One tempting way to divide the occipital cortex into
29 of 179
functional “regions” might be by retinotopy—one group of patches that responds to
foveal stimuli, and a different group of patches that responds to peripheral stimuli. This
foveal vs. peripheral difference is highly robust, replicable within and across subjects,
within and across tasks, and correlated with behavior (i.e. visual performance in
corresponding regions of the visual field). Nevertheless, other considerations, such as
cytoarchitecture, connectivity, and processing time, suggest that this is the wrong
division for capturing functional and computational regularities. Identifying a robust
functional regularity that divides one patch of cortex from another is not the same thing
as identifying a true cortical area—a region that is computationally distinct from its
neighbors, with distinct cytoarchitecture, connectivity, and topography (Friston, Holmes,
Worsley, Poline, Frith, & Frackowiak, 1995; Kanwisher, 2010; Kriegeskorte et al., 2006;
Worsley, Evans, Marrett, & Neelin, 1992). Instead of studying the “peripheral patches,”
neuroscientists divide the cortex into V1, V2, V3, MT, etc., each of which is then
internally organized (in part) by retinotopy.
Imagine you are a social cognitive neuroscientist, looking at a new bit of cortex, and you
see a new functional regularity—a patch of cortex that shows a high response when the
individual is thinking about stories, cartoons, or movies depicting the contents of another
mind. Which kind of functional regularity is this? On the one hand, these data could
signal the discovery of a true computational area, like V1. On the other hand, the
observed functional regularity might be more like “peripheral patches,” one part of a
stimulus space or dimension that is mapped across cortex, but crosscuts multiple
computational areas.
We believe that current fMRI data cannot resolve this puzzle directly. One approach
may therefore be to suspend judgment until other sources of evidence are available. If
patches of cortex involved in understanding other minds are true cortical areas, it will be
possible to distinguish them from their anatomical neighbors by cytoarchitecture,
connectivity, and/or topography. Non-invasive functional imaging tools do not yet have
high enough resolution to reveal cytoarchitecture in vivo. To study the links between
function and cytoarchitecture with current technology, we would need to collect
functional data to identify the regions in vivo, and then analyze the neuroanatomy of the
same individual post mortem.
A weaker, but more accessible, source of evidence is patterns of connectivity. In some
cases, cortical areas can be differentiated by their profiles of connectivity. It has become
increasingly possible to measure the pattern of connections between regions using
neuroimaging. Diffusion imaging (DTI), which looks at the predominant direction of
water diffusion, allows us to visualize the dominant pathways of axons connecting brain
regions. Using diffusion, we can ask whether a patch of cortex involved in
understanding other minds shows a different pattern of connectivity than its neighbors to
the rest of the brain. Initial evidence suggests it does, at least in the case of the right
TPJ. Mars, Sallet, Schüffelgen, Jbabdi, Toni, & Rushworth (2012) found that the broad
area of right temporo-parietal cortex (BA 39/40) can be sub-divided into three clusters,
based on DTI connectivity alone, and one of these clusters (the posterior one) is
functionally correlated (during a resting baseline) with the other ToM regions, including
medial prefrontal and medial parietal cortex. Bzdok et al. (2013) corroborated and
extended this finding, using resting state connectivity and meta-analyses. They found
30 of 179
similar anatomical parcellation, and also tested functional profile: the region of TPJ that
correlated with other ToM regions was also the region that responded most to social
cognition tasks. Similarly, in a recent meta-analysis, Schurz et al. (2014) found that the
same posterior region of TPJ defined by tractography responded more to false belief
tasks that any other sort of social cognition task (such as trait inference, violations of
rational action, or strategic games); while the region anterior to it (posterior STS)
responded most to violations of rational actions. In contrast, regions MPFC respond
strongly to trait inference tasks.
Currently, however, neither cytoarchitecture nor connectivity analyses give a definitive
evidence that patches of cortex recruited by mental state reasoning tasks are true
cortical areas. An alternative approach might therefore be to consider the alternative
hypothesis directly—these regions are one part of a larger, continuous map. If we
compare the cortical patches we are studying to their anatomical neighbors, is there a
plausible higher-level “stimulus space,” which could unite these responses into one
map?
Again, the right TPJ is an interesting example. The right TPJ region that is activated by
ToM tasks (as described in “Theory of mind and the brain”) has two very close
anatomical neighbors. One neighbor (up toward parietal cortex in the right inferior
intraparietal sulcus (IPS), but confusingly sometimes also called RTPJ) is recruited by
unexpected events that demand attention. These events may be unexpected because
they are rare, or because a generally reliable cue was misleading on this occasion.
Redirecting attention toward an unexpected event leads to metabolic activity in this
region (Corbetta & Shulman, 2002; Mitchell, 2008; Serences, Shomstein, Leber, Golay,
Egeth, & Yantis, 2005). Damage to this region makes it hard for objects and events in
the contralateral visual field to attract attention, producing left hemifield neglect
(Corbetta, Patel, & Shulman, 2008). Some authors have noted that false belief tasks
typically involves an unexpected event (Corbetta et al., 2008; Decety & Lamm, 2007;
Mitchell, 2008): for example, false belief tasks often hinge on an object unexpectedly
changing location while the protagonist is out of room, and require the participant to shift
attention between both locations. In fact, the region that is recruited during exogenous
attention tasks is so close to the region recruited during ToM tasks that some concluded
that these are actually just two different ways to identify the same region (Corbetta,
Kincade, Ollinger, McAvoy & Shulman, 2000; Mitchell, 2008). More recently, however,
connectivity analyses (Bzdok et al., 2012; Krall et al., 2014; Mars et al., 2012), metaanalyses (Schurz et al, 2014; Decety & Lamm, 2007; Kubit & Jack, 2013), highresolution scanning within individual subjects (Scholz, Triantafyllou, Whitfield-Gabrieli,
Brown, & Saxe, 2009) suggest that there are actually two distinct cortical regions (or
“patches”), and the region recruited by attention tasks is approximately 10 mm superior
to the region recruited by ToM tasks (Krall et al., 2014; Scholz et al., 2009). Also, the
region that is recruited during false belief tasks is not recruited by unexpected transfers
of location in stories about false maps and physical representations, as described above
young (Young, Dodell-Feder, & Saxe, 2010a). Nevertheless, it remains an interesting
possibility that the exogenous attention response and the ToM response are part of a
larger continuous map across cortex, a topography of different kinds of unexpected
attentional shifts. The more superior end of this map could direct attention toward
31 of 179
unexpected positions in space and time, while the more inferior end of the map could
direct attention toward unexpected people, actions, or inferred thoughts.
A second topographical “stimulus space” of which the RTPJ could be a part runs
anteriorly, through the right superior temporal sulcus (STS), and down toward the
temporal pole. As with the RTPJ and the right IPS, the RTPJ and the posterior STS
were initially conflated, but have subsequently been shown to be functionally distinct
(Perner & Leekam, 2008; Perner & Roessler, 2012; Schurz, Radua, Aichhorn, Richlan,
& Perner, 2014; Gobbini, Koralek, Bryan, KJ, & Haxby, 2007). Multiple parts of the STS
are recruited when participants observe other people’s actions in photographs, film
clips, point light walkers or animations (Pelphry, Mitchell, McKeown, Goldstein, Allison,
& McCarthy, 2003; Pelphrey, Morris, Michelich, Allison, & McCarthy, 2005; Hein &
Knight, 2008). The STS is large, and Kevin Pelphrey and colleagues (Pelphrey et al.,
2005; Pelphrey & Morris 2006) propose that it contains a pseudo-somatotopic map of
observed actions: others’ mouth motions represented most anteriorly, followed by hand
and body movements, followed by head and eye movements represented most
posteriorly. An intriguing possibility is that the temporo-parietal junction, which is at the
most posterior end of the superior temporal sulcus, is part of this same map. Whereas
hand and body movements convey information about what a person is doing and
intending, head and eye movements convey information about what a person is looking
at and seeing. Thus, the STS may contain a map of others actions that move from
externally observable body movements (anterior end) toward invisible mental states
(posterior end), culminating in the RTPJ, which responds to thinking about what a
person is thinking.
In principle, either of these mapping hypotheses, or both, could be true. Evidence that a
ToM region, e.g. the RTPJ, is part of a larger cortical map might come in the form of a
response that moves continuously across contiguous patches of cortex, modulated by
continuous changes in a stimulus dimension (Konkle & Oliva, 2012). For example, if the
RTPJ is one end of an attention map, we might expect to see that surprising facts about
physical entities elicit activity relatively far from the TPJ, but that as the unexpected
stimulus becomes more social and interpersonal, activity moves continuously"
across the map, ending at the RTPJ. Similarly, if the RTPJ is the “abstract” or “head”
end of a map of observable social actions, we’d expect to see continuous changes in
the location of activity, as depictions of human actions become either more abstract or
physically higher on the body. Finally, both could be true. The RTPJ could exist at its
precise location because that is where a region that deals with unexpected information
converges with a region that deals with human actions. We would be enthusiastic to see
someone test these hypotheses further.
For now, however, we suggest that it does not matter for any of our empirical purposes
whether the RTPJ (or any other patch of cortex recruited in ToM tasks) is a true cortical
area, or whether it is one end of a larger map, as defined by cytoarchitecture,
connectivity, and topography. In either case, there is a robust functional regularity
replicable within and across subjects, within and across tasks, and correlated with
behavior: a bounded, adjacent patch of cortex where activity is high during a range of
ToM tasks. Without committing to whether they are true computational “areas,” or just
the equivalent of “Peripheral Patches,” we believe that it is fruitful to continue to study
32 of 179
these coherent patches as “regions,” in order to discover the computations and
representations that underlie thinking about other minds.
Given everything else we know about the brain, it is not surprising that systematic
response profiles are linked to regions of cortex much bigger than a neuron and much
smaller than a network. The challenge now is to integrate the huge, and growing, list of
empirical discoveries and to construct hypotheses about the computations and
representations in the neural populations we are studying. These empirical results form
the basis of a different set of potential objections to the hypothesis that there is a strong
specific link between cortical regions, like the TPJ, mPFC, and PC, and thinking about
thoughts.
Empirical objections
An empirical objection to our argument in “Theory of mind and the brain” might begin by
pointing out that our review of the literature was selective. In addition to the dozens of
articles we cited, there are dozens of others that claim so-called “ToM regions” are
active in tasks that do not involve thinking about thoughts, or are not active in tasks that
do involve thinking about thoughts. How can we integrate these other data into a
coherent hypothesis?
First, what about claims that “ToM regions” are active in tasks that do not involve
thinking about thoughts? The RTPJ is again a useful example. As we mentioned above,
some authors initially believed that the same region of RTPJ involved in false belief
tasks was also recruited during any exogenous shift of attention and/or biological motion
perception. Other literature suggests that the RTPJ is involved in maintaining a
representation of one’s own body, by integrating multi-sensory information and locating
the body in space (Blanke et al., 2005; Blanke & Arzy, 2005; Tsakiris et al., 2008).
Experiments that ask people to mentally rotate their own body or imagine their body in
different parts of space, as well as those that induce changes in bodily self-perception
(with e.g. rubber hands) find activation in this region (Arzy et al., 2006; Blanke & Arzy,
2005). TMS and intracranial stimulation to this region has been shown to lead to out-ofbody experiences, confusion of the body vs. the environment, and illusory changes in
the orientation of body parts (Blanke et al., 2002, 2005; Tsakiris et al., 2008).
How do we interpret these results, which seem to contradict our theory? In general, we
could consider five possibilities. The first is that one group of scientists has made an
error, leaving an unnoticed confound in their experimental paradigm. Could the body
representation tasks also involve thinking about thoughts? Could the false belief task
accidentally induce updating maintaining a representation of one’s own body? These
are empirical questions, but we consider both possibilities highly unlikely. The second
option is that there is some deep common computation, served by the same neural
population that is required by these different classes of tasks. One option here would be
“decoupling” (Gallagher & Frith, 2003; Leslie & Frith 1990; Liu, Sabbagh, & & Gehring,
2004), maintaining distinct representational reference frames, e.g. for one’s own and
other minds, for current vs. imaginary body positions, and so on. The third option is that
the same neurons can assume distinct and even unrelated functional roles, depending
on the context and the pattern of activity in other neural populations (e.g. Miller &
33 of 179
Cohen, 2001). On this view, thinking about thoughts and maintaining a body
representation are cognitively unrelated, in spite of being implemented by the same
neurons. The fourth option is that distinct neuronal populations are involved in thinking
about thoughts and maintaining a body representation, but these neurons are
interleaved within the RTPJ, just as color- and orientation-sensitive neural populations
are interleaved in V1.
Finally, the fifth option is that distinct neuronal populations are involved in thinking about
thoughts vs. maintaining one’s body representation, exogenous attention shifts, and
biological motion perception. These neural populations are not interleaved: they are
contained in distinct regions that are merely nearby on the cortical sheet. We find this
last option most plausible. Standard fMRI methods, which involve extensive spatial
blurring at three stages (acquisition, preprocessing, and group averaging; Fedorenko &
Kanwisher, 2009; Logothetis, 2008), are strongly biased to conflate neighboring regions
that are truly distinct. Even so, the existing evidence suggests that the region involved in
body representations is lateral (MNI x-coordinates typically around 64mm) to the region
involved in thinking about thoughts (x-coordinates around 52 mm). As described above,
this was also true of the regions involved in biological motion perception (Gobbini et al.,
2007) and exogenous attention (Decety & Lamm, 2007; Scholz et al., 2009); these
almost completely non-overlapping in individual subjects.
A different kind of challenge arises from examples of tasks that apparently do involve
thinking about thoughts, but do not elicit activity in TPJ, mPFC, or PC. Two interesting
examples are visual perspective-taking tasks, and recognition of facial expressions of
emotions from photographs. In visual perspective taking tasks, the participant sees an
image of a character in a 3D space, and is asked to imagine the view of the room from
the viewpoint of the character. For example, participants may be asked to report how
many dots the character can see (the third person perspective), versus how many dots
the participants themselves can see (including those that are out of the character’s view,
the first person perspective). This perspective-taking task clearly involves thinking about
the character’s visual access, which could be construed as a mental state.
Nevertheless, this task typically does not elicit activity in the same regions as thinking
about thoughts (Aichhorn et al., 2006; Vogeley May, Ritzl, Falkai, Zilles, & Fink, 2004).
Similarly, participants in hundreds of fMRI experiments have viewed photographs of
human faces expressing various basic emotional expressions (e.g. sad, afraid, angry,
surprised, happy, neutral). Although these images do depict evidence of another
person’s emotional experience, they also typically do not elicit activity in the same
regions as thinking about thoughts (Costafreda et al., 2008; Lamm, Batson, & Decety,
2007; Vuilleumier et al., 2001).
What should we conclude from these examples? One option is to make a forward
inference. Using the evidence from these tasks to change our hypothesis about the
brain regions’ functions. Here, the forward inference might be that the ToM brain regions
are responsible for only a subset of mental state processing. We could conclude that
different brain regions are involved in thinking about different classes of internal
experiences: bodily states, emotional states, perceptual states, or epistemic states (e.g.
thinking, knowing, doubting, etc.). The so-called ToM brain regions might be specifically
34 of 179
involved in representing epistemic states, while regions of insula represent others’
emotions and regions of parietal cortex represent others’ perceptual states.
Another option is to make a reverse inference: using the pattern of neural activity to
change our analysis of the cognitive processes required by the task. Reverse inferences
are risky because they require a lot of confidence in the functional specificity of the brain
region(s) involved (Poldrack, 2006a). However, we think ToM brain regions are good
candidates to support reverse inference, given the converging evidence across the
many experiments described above. In this case, a reverse inference might be that
these visual perspective taking and emotional face tasks do not actually elicit thinking
about thoughts, and instead are solved by alternative computational strategies. For
example, rather than truly considering someone’s perceptual experiences, line-of-sight
tasks may be solved using mental rotation and geometric calculation. Emotional facial
expressions may be recognized (akin to object recognition) without always requiring a
representation of the person’s internal state.
In these particular cases, we are open to either the forward or reverse inference; only
further experimentation will tell which is the better generalization. In general, we believe
that in this domain, inferences can be made in both directions, forward and reverse. The
absence of activation in perspective taking and emotion recognition tasks provides an
important constraint on the possible functions of ToM brain regions (the forward
inference). At the same time, the fact that visual perspective-taking and emotion
recognition rely on different brain regions from reading stories about thoughts provides
evidence that these tasks depend on different cognitive processes (the reverse
inference). It’s the give and take of these two kinds of inferences, as evidence
accumulates, that allows us to build a coherent understanding of both cognitive and
neural function.
Summary
Taken together, these first two sections illustrate the main contributions of the first
decade of neuroimaging the understanding other minds. We have discovered a robust,
replicable functional regularity in the human brain: regions that have increased activity
when participants think about thoughts. These regions may be true cortical areas or
parts of larger topographical maps, but in either case, understanding other minds is a
major organizing principle of responses over cortex. The function of these regions is not
to complete a task, but to transform some class of input into some output; and the class
of input has something to do with thinking about minds, and not bodies or abstract
representations. Of course, this description remains unsatisfying and largely
underspecified. Nevertheless, we are optimistic that the second decade of this research
program will continue to improve our specifications. In particular, we are excited about
newly emerging methods for fMRI data analysis that focus on the second aspect of a
region’s function: the features within its preferred stimulus class that organize differential
responses within each of the ToM regions.
35 of 179
What Now?
Differences between theory of mind regions
One step in specifying the computations performed by ToM regions will be
understanding the division of labor and information transfer between the different
regions. Overall, ToM regions show similar profiles to most of the contrasts we
described. However, research in the last 5 years has begun to tease apart the functional
profiles of these regions, and the differences are intriguing, though much work still
remains to be done to form a coherent view of how they all fit together.
The most striking contrast comes from a task that elicits very robust activity in the
medial ToM regions (mPFC and precuneus), but not the lateral ToM regions (TPJ and
anterior STS): thinking about personality traits, especially of the self and close others
(Whitfield-Gabrieli et al., 2011; Moran et al., 2011a; Saxe, Moran, Scholz, & Gabrieli,
2006a; Krienen, Tu, & Buckner, 2010). In a typical version of the experiment,
participants in the scanner see single words describing personality traits (e.g. “lazy,”
“talkative”, “ambitious”) and judge either whether each one is desirable or undesirable
(the semantic control condition), is true of a famous person (the other control condition),
or is true of themselves (the self condition). MPFC and precuneus regions show much
higher activity during the self condition; moreover, within the self condition, activity in
mPFC is higher for the words that participants say are true of themselves (Moran et al.,
2011a), and the amount of activity in mPFC for each item presented in the self task (but
not in the semantic or other tasks) predicts participants’ subsequent memory for those
items on a surprise memory test (Mitchell, Macrae, & Banaji, 2006; Jenkins & Mitchell,
2009). These data are compelling to us: the mPFC, but not the TPJ, is involved in
reflection about one’s own stable traits and attributes (Lombardo et al., 2010). Similarly,
elaborating one’s own autobiographical memories leads to activity in medial ToM
regions, whereas imagining someone else’s experiences on similar occasions elicits
activity in bilateral TPJ (Rabin, Gilboa, Stuss, Mar, & Rosenbaum, 2010).
In fact, coding information in terms of similarity to the self may be a key computation of
mPFC. In one series of studies (Tamir & Mitchell, 2010), participants judged the likely
preferences of strangers (e.g. is this person likely to “fear speaking in public” or “enjoy
winter sports”) about whom they had almost no background information. Under those
circumstances, the response of the mPFC (but not TPJ) was predicted by the
discrepancy between the attributions made to the target and the participant’s own
preference for the same items: the more another person was perceived as different from
the self, for a specific item, the larger the response in mPFC.
Another distinction, supported by multiple studies, suggests that sub-regions of mPFC
are most recruited when thinking about someone’s negative emotions or bad intentions,
whereas the TPJ makes no distinction based on valence. For example, Bruneau, Pluta,
& Saxe (2011) found that only the mPFC showed a higher response to stories about
very sad events (e.g. a person proposes marriage and is rejected) compared to neutral
or positive events (e.g. the marriage proposal is accepted). In a PET study, Hayashi,
Abe, Ueno, Shigemune, Mori, Tashiro, et al. (2010) found that a region in mPFC was
recruited when people were considering an actor’s dishonesty as a factor in moral
36 of 179
judgments; and in an fMRI study, Young & Saxe (2009a) found that a region in ventral
mPFC was correlated with moral judgments of attempted harms, which are morally
wrong only because of the actor’s negative intentions. Converging with these
neuroimaging studies, lesion studies suggest that focal damage to ventral mPFC
creates disproportionate difficulty in understanding bad intentions, and in integrating
those intentions into moral judgments (Koenigs et al., 2007).
In this vein, further work has been done to examine the response profile of mPFC. The
“regions” implicated in ToM are very large, especially in the mPFC, and there may be
multiple sub-divisions, each with different response profiles. Proposed distinctions along
the ventral-dorsal axis of mPFC include: similarity to self (such that self-relevant
processes elicit responses more ventrally (e.g. Mitchell et al., 2006; Jacques. Conway,
Lowder, & Cabeza, 2011), interpersonal closeness (people who are closer, or more
important to the self elicit responses more ventrally, Krienen et al., 2010), or affective
content (“hot” affective states elicit responses more ventrally (Ames, Jenkins, Banaji, &
Mitchell, 2008; D’Argembeau et al., 2007; Mitchell & Banaji, 2005), while “cool” cognitive
states elicit responses more dorsally (Kalbe, Schlegel, Sack, Nowak, Dafotakis,
Bangard, et al., 2010; Shamay-Tsoory & Aharon-Peretz, 2007; Shamay-Tsoory, Tomer,
Berger, Goldsher, & Aharon-Peretz, 2005). Thus, while there may be a convergent
theoretical account of the mPFC, and its responses to other people’s negative intentions
and emotions, one’s own personality traits, and ambiguous inferences about
preferences, another possibility is that these response profiles reflect distinct subregions within mPFC, each contributing a distinct computation to understanding other
minds.
Intriguingly, none of these distinctions have shown to affect the lateral ToM regions. In
contrast, one dimension that seems to influence the magnitude of response in TPJ more
than mPFC is whether someone’s thought or feeling is unexpected, given the other
information you have about that person. Saxe & Wexler (2005) introduced characters
whose social background was mundane (e.g. New Jersey) or unusual (e.g. a
polyamorous cult). Participants then read about that character’s thoughts and feelings
(e.g. a husband who believed it would be either fun or awful if his wife had an affair). On
their own, neither the background nor the content of the belief affected the magnitude of
response in the TPJ, but there was a significant interaction: whichever thought was
unlikely, given the character’s social background, elicited a larger response in the right
TPJ. Recently, Cloutier, Gabrieli, O'Young, & Ambady (2011) provided a conceptual
replication of this result: participants saw photographs of people labeled as Democratic
or Republican, paired with opinions that were either typical of their political affiliation or
typical of the opposite affiliation. Opinions that were unexpected given the protagonist’s
political background (e.g. a Republican wanting liberal Supreme Court judges) elicited a
higher response in most of the ToM regions, including bilateral TPJ and mPFC. Finally,
a third study suggests that the conflict between background and belief is necessary for
increased activation, not just sufficient. In the absence of specific background
information about the believer, there is no difference in the response of any ToM region
to absurd vs. commonsense beliefs (e.g. “John believes that swimming in the pool is a
good way to grow fins/cool off,” Young, et al., 2010b).
37 of 179
Although they are preliminary, we find these results exciting because they are
consistent with the idea that activity in TPJ reflects a process of forming a coherent
model of another’s mind. We expect other people to be coherent, unified entities, and
strive to resolve inconsistencies with that expectation (see Hamilton & Sherman, 1996).
These results suggest that while TPJ activity may be related to the discrepancy
between a thought or feeling and other information about the protagonist, mPFC and
PC activity may be related to discrepancies between the protagonist’s thoughts or
feelings and the participant’s own thoughts or feelings on the same topic (see Chapter 5
for more discussion of this idea). Thus, whereas TPJ may be involved in integrating a
belief or preference into a coherent model of another’s mind (e.g. Young et al., 2010c),
mPFC and precuneus activity may reflect a different “anchor-and-adjust” strategy, that
helps identify other people’s thoughts and preferences by starting with one’s own
preferences, and then adjusting them as necessary (Tamir & Mitchell, 2010). Taken
together, these results suggest that medial and lateral ToM regions support distinct
computations within ToM. These distinctions help us separate ToM into its real (i.e.
neurally-realized) component parts, and formulate hypotheses about each of the more
specific functions of individual regions within the group.
Information within theory of mind regions
An equally important step in specifying the computations performed by ToM regions will
be understanding what is happening inside each region. Until now, we asked simply
whether ToM brain regions do or do not show activity in response to a task or stimulus.
However, a single “region” millions of neurons: activation within a region is clearly
neither binary nor uniform.
Patterns within theory of mind regions
One strategy is to look for divisions of functional responses across the neural
populations within each region. Two relatively novel methods for analyzing fMRI data
may allow neuroscientists to look inside ToM regions. Distinctions between neural
subpopulations within regions may provide clues to how these regions function.
The first method is repetition-suppression, also called functional adaptation. Repetitionsuppression analyses take advantage of the observation that after processing a
stimulus or task once, activity in a neuron or brain region in response to an identical
stimulus or task is suppressed, or adapted (Grill-Spector, Henson, & Martin, 2006). By
manipulating the features of the repeated stimulus, so that some are identical and some
are different from the original stimulus, it is possible to ask what counts as the same
stimulus for a particular brain region (Kourtzi & Kanwisher, 2001). If the repeated
stimulus is effectively the same with regard to the features represented by the brain
region, then the region’s response will be suppressed or adapted. On the other hand, if
a feature that is represented in the brain region has been sufficiently modified, the brain
region’s response will “recover” from adaptation.
This method has been used frequently and effectively to study visual representations of
objects and places (Poldrack, 2006b). One a few studieshave taken advantage of this
approach to test hypotheses about ToM. Jenkins, Macrae, & Mitchell (2008) asked
38 of 179
whether attributing a preference (e.g. “enjoys winter sports”) to oneself, a similar
stranger, or a dissimilar stranger, depend on the same neural subpopulations within
mPFC. They found that thinking about a similar other person after thinking about the
self led to repetition suppression, while thinking about the self and then a dissimilar
other led to recovery from adaption. These results support the hypothesis that the
mPFC represents other minds in terms of their similarity to the self. Similarly, Ma and
colleagues (Ma et al., 2013a; Ma, Baetens, Vandekerckhove, Van der Cruyssen, & Van
Overwalle, 2013b) find evidence of representation suppression in MPFC to stimuli that
suggest similarly personality traits, but not repeated object traits.
Currently, we are using a similar strategy to investigate the components of mental state
attributions. People read short stories in which a key mental state was repeated twice,
with some elements changed. After the first mental state sentence (e.g. “Megan thinks
that Julie is being too flirty”), the repetition either changed the agent (e.g. "Gina thinks”),
the attitude verb (e.g. "Megan worries that"), the content (e.g. “Megan thinks that Julie
should be more flirty"), all three, or none of these elements. In preliminary data, we find
that ToM brain regions recover from adaptation for each kind of change on its own
(compared to no change), suggesting that these regions encode all three elements of a
mental state attribution. Interestingly, while the mPFC shows the most recovery when
the content of the mental state changes, the left temporoparietal junction (LTPJ) shows
the most recovery when the attitude verb changes, and the RTPJ shows equal recovery
for any kind of change. If these results hold up to further analyses, they may provide
clues about the contributions of each ToM brain region to thinking about the minds of
others.
The second method, multi-voxel pattern analysis (MVPA), looks for subtle, but reliable
spatial patterns within a single region (or local neighborhood) of cortex. By looking at
spatial separability, MVPA provides a more direct measure of the existence of
functionally separable sub-populations of neurons than repetition suppression. Each
“voxel” (the fMRI equivalent of a pixel) may have some, possibly very small, preference
for one kind of stimulus over another due to biases in the neuronal populations toward
one type of information-processing vs. another; pattern analyses measure the
similarities and differences between these patterns across space (Haxby, 2001). A few
studies have begun to use pattern analyses to study ToM (e.g. Gilbert et al., 2008). In
one promising example, Peelen, Wiggett, & Downing (2006) found that the pattern of
activation in the mPFC reliably reflected the content of another person’s emotion (e.g.
sad vs. angry), independent of the stimulus modality (e.g. vocal expressions vs. body
posture). In Chapters 2, 3, and 4, I show that this technique can used to find
distinctions which represented within a region, and point toward the underlying
distinctions and computations in ToM.
Magnitude: “more” theory of mind
As well as asking how information is distributed across a region, is it equally important
to understand what drives activity, in a region, in a voxel, and in a neuron. The
magnitude of activity, over tasks or stimuli, could reveal not only what class of stimuli is
processed in a region, but also which dimensions of those stimuli and tasks elicit more
or less processing.
39 of 179
Interestingly, initial attempts to modulate activation in the ToM network mostly
discovered features that do not elicit differential magnitudes of response. For example,
most of the ToM regions show an equally high response to explicit descriptions of
beliefs that are true or false (Young & Saxe, 2008), justified or unjustified (Young,
Nichols, & Saxe, 2010c), and well-intentioned or bad (i.e. a girl who believes she is
putting poison in her friends coffee, vs. believes that she is putting sugar in the coffee,
Young, Cushman, Hauser, & Saxe, 2007). In the TPJ, at least, it also does not matter to
whom the thought or feeling is attributed: there is an equally strong response to beliefs
attributed to similar or dissimilar others (Saxe & Wexler, 2005), or to members of one’s
own group vs. an enemy group (Bruneau & Saxe, 2010; Bruneau, Dufour & Saxe,
2012).
Part of the reason for this lack of success may be that we do not yet have satisfactory
cognitive or computational theories of ToM that allow us to predict when “more” ToM
processing will be required, or even exactly what “more” means. Some intuitive
possibilities have already proven empirically false. For example, making it harder to
infer what a character believes, by making the available evidence more ambiguous,
does not lead to more activity in TPJ regions (Jenkins & Mitchell 2009). In fact, we
(Dodell-Feder et al., 2011) found that, while some stories about thoughts systematically
elicit more activity than others, in each ToM region, we could not find any feature (e.g.
vividness, length, syntactic complexity) that predicted these differences in activity, with
the partial exception of the precuneus, which showed greater activity to stories that
involved more people.
Researchers with a background in computer science or game theory often suggest one
particular dimension for “more” ToM processing—the depth of embedding of one mental
state within another. Thus, many people intuit, reasoning about an embedded belief
(e.g. “Carla believes that Ben thinks that she eats too much junk food”) should require
more ToM processing that reasoning about a simple belief (e.g. “Carla believes that she
eats too much junk food.”) When we tested this hypothesis directly using verbal stories,
we found that no ToM regions showed greater activity for the more embedded beliefs
(Koster-Hale & Saxe, 2011). Other brain regions did show more activity—regions
involved in language processing and regions involved difficult memory and cognitive
control tasks, like dorsolateral prefrontal cortex (DLPFC). Our interpretation of these
results is that embedding beliefs inside other beliefs makes the reasoning problem
harder, but does not lead to greater ToM processing, per se. This finding converges with
results from patient populations and aging adults showing that failure to pass secondorder false belief tasks may, in fact, be due to domain-general impairment, rather than
diminished ToM processing (Slessor, Phillips, & Bull, 2007; Zaitchik, et al., 2006).
An alternative approach draws from a model put forward in other domains, such as
vision and decision making. Predictive coding posits that neural systems make forwardlooking predictions about incoming information. In this model, neural signals contain
information not about the currently perceived stimulus, but about the difference between
the observed and the predicted stimulus. There has recently been increasing interest in
the idea of predictive coding as a unifying framework for understanding neural
computations across many domains (e.g., Clark 2013). In Chapter 5, I consider this
theoretical framework in more detail, and explore its application to theory of mind.
40 of 179
A First Step
Human social life depends on understanding other people’s behavior: why they do what
they do, and what they are likely to do next. However, these actions are merely the
output of an unobservable, internal causal structure: the person’s goals and intentions,
beliefs and desires, preferences and personality traits. A cornerstone of the human
capacity for social cognition is the ability to reason about these invisible causes: having
a “theory of mind”. The literature reviewed in this chapter presents converging evidence
that reasoning about the minds of other people reliably and selectively recruits specific
cortical regions. Yet, while we have an understanding of which parts of the brain support
social inference, we have remarkably little insight into how these neural substrates
function at a computational level. A key next question in social cognitive neuroscience is
to investigate how neural populations encode these concepts.
To move past where in the brain, to how, a first step is asking whether we have the tools
and theoretical understanding to successfully examine brain regions that support social
cognition in more detail. The goal of this thesis is to demonstrate that it is possible to
use functional neuroimaging to ask questions about neural representations in social
cognition. I present evidence that neuroimaging can find abstract, behaviorally relevant
features of mental state inferences inside cortical regions that support social cognition,
by demonstrating:
(i)
that the neural code in brain regions that support social cognition contains
information which distinguishes between mental state representations (rather than
mental state representations from other types of representation), and that this
code is detectable using fMRI (Chapter 2),
(ii)
that these neural distinctions are behaviorally relevant: they reflect distinctions
that we know matter, based on cognitive and developmental work on theory of
mind, and are sensitive to individual differences in social cognition (Chapter 2),
(iii)
that these neural distinctions are robust in the face of profound differences in
developmental experience (Chapter 3),
(iv)
that these neural distinctions are part of a more complex feature space of mental
state representation, and are coded along independent, continuous dimensions
(Chapter 4),
(v)
that an existing framework of neural coding, predictive coding, can provide a
unifying framework for interpreting — and making new predictions about — the
kinds of representations we find in these brain regions.
In Chapter 2, I demonstrate that functional neuroimaging of the human brain can be
used to find neural representations of distinct features of mental states. As a first case
study, I consider intent — whether someone’s goal is aligned with the results of their
actions — when causing harm. Intent is a clear test case to look for neural
representations of mental state features. First, reasoning about intent has previously
been shown to rely on ToM brain regions: distinguishing between intentional harms and
accidents depends on the capacity for mental state reasoning and disrupting activity in
RTPJ can impair mental state reasoning for moral judgment (Young et al, 2010).
41 of 179
Second, intent has clear behavioral relevance: intentional harms are judged to be
morally worse and more punishable than accidental harms (Cushman, 2008; Baird &
Astington, 2004). And third, there are reliable individual and group differences in
reasoning about intent: individuals differ in the extent to which they weigh intent versus
outcome (Young & Saxe, 2009b; Moran et al., 2011); moreover, high-functioning
individuals with Autism Spectrum Disorders (ASD) make moral judgments based less on
intent information, compared to neurotypical (NT) participants (Grant, 2005; Moran et
al., 2011; Yang, Baillargeon, & Baillargeon, 2013). In four experiments, using multi-voxel
pattern analysis (MVPA), I find that (i) in NT adults the RTPJ shows reliable and distinct
spatial patterns of responses across voxels for intentional versus accidental harms, and
(ii) individual differences in this neural pattern predict differences in participants’ moral
judgments. These effects are specific to RTPJ. By contrast, (iii) this distinction is absent
in adults with ASD. Together, these data provide a proof of concept: fMRI can be used
to find neural representations inside ToM brain regions that reliably track differences in
behavior, both at the individual level, and the clinical level.
Having found that information about different types of mental states can be found in ToM
brain regions, we need to ask, can we find other features of mental state inference, and
can we learn anything about their format and developmental origin? In Chapter 3, I
show that ToM brain regions contain explicit, abstract representations of another feature
of others’ mental states: perceptual source. I find that in the RTPJ of sighted adults, the
pattern of neural response distinguished the source of the mental state (did the
protagonist see or hear something?). Moreover, I find that these representations persist
in the face of profound differences in developmental history (congenital blindness):
neural representations of perceptual source were preserved in congenitally blind adults,
who have no first-person experience with sight. The fact that blind people’s inferences
about how other people see rely on a neural code that parallels that of sighted people
provides a window into fundamental questions about the human capacity to think about
one another’s thoughts: we can ask both what kinds of representations people form
about others’ minds, and how much these representations depend on the observer
having had similar mental states themselves.
In Chapter 4, I take a first step towards looking for a feature space of mental state
inference, by simultaneously testing four unique features of beliefs. I find that that social
brain regions contain representations of epistemic features (why does someone believe
what they believe?) and emotional features (how do they feel based on what they
believe?) of others’ beliefs. Using MVPA, I find that the patterns of activity in RTPJ and
right STS contain complementary neural codes, which reliably distinguish stimuli in
which a person’s belief is based on reliable and unambiguous evidence, from stimuli in
which their belief based on unreliable and ambiguous evidence. Moreover, I show that
the neural code tracks item-wise differences along these abstract dimensions:
classification accuracy for each stimulus was predicted by independent ratings of the
strength of the evidence in that stimulus. I also find evidence for a representation of
evidence modality (visual vs auditory) in bilateral TPJ, and of emotional valence in
ventral MPFC. Together, these data suggest that the cortical regions implicated in
mental state inference contain neural representations of epistemic and emotional
42 of 179
features of others’ beliefs, and that these features are represented along continuous,
abstract dimensions. Finally, in Chapter 5, I take a more global view, and consider how we might understand
the activity we observe in these regions, and how we might go about generating new,
testable hypotheses. I extend a model from vision and neuroeconomics — predictive
coding — and explore its application to the neural basis of social cognition. Predictive
coding posits that neural systems make forward-looking predictions about incoming
information. Neural signals contain information not about the currently perceived
stimulus, but about the difference between the observed and the predicted stimulus. I
review evidence that, across ToM brain regions, neural responses to depictions of
human behavior, from biological motion to trait descriptions, exhibit a key signature of
predictive coding: reduced activity to predictable stimuli. I propose a series of
experiments to distinguish predictive coding from alternative explanations of this
response profile, and argue that this approach could both further our understanding of
the representations inside ToM brain regions, and our understanding of the distribution
of labor across them.
Taken together, this work provides a key next step to understanding the neural basis of
theory of mind: demonstrating that it is possible to find abstract, behaviorally relevant
features of mental state inferences inside cortical regions that support social cognition.
In Chapter 6, I discuss how these results lay the groundwork that will allow us to pursue
the research lines suggested above, and lay out some of the next steps in this research
program.
43 of 179
This page intentionally left blank
44 of 179
Chapter 2.
Decoding moral judgments from neural
representations of intentions 6
Intentional harms are typically judged to be morally worse than accidental harms.
Distinguishing between intentional harms and accidents depends on the capacity for
mental state reasoning (i.e., reasoning about beliefs and intentions), which is
supported by a group of brain regions including the right temporo-parietal junction
(RTPJ). Prior research has found that interfering with activity in RTPJ can impair
mental state reasoning for moral judgment, and that high-functioning individuals with
Autism Spectrum Disorders (ASD) make moral judgments based less on intent
information, compared to neurotypical (NT) participants. Three experiments, using
multi-voxel pattern analysis (MVPA) find that (i) in NT adults the RTPJ shows reliable
and distinct spatial patterns of responses across voxels for intentional versus
accidental harms, and (ii) individual differences in this neural pattern predict
differences in participants’ moral judgments. These effects are specific to RTPJ. By
contrast, (iii) this distinction was absent in adults with ASD. We conclude that MVPA
can detect features of mental state representations (e.g., intent), and that the
corresponding neural patterns are behaviorally and clinically relevant.
6
A version of this chapter was published as Koster-Hale, J., Saxe, R. R., Dungan, J., & Young, L. L.
(2013). Decoding moral judgments from neural representations of intentions. PNAS, 110(14), 5648–5653.
doi:10.1073/pnas.1207992110
45 of 179
Introduction
Thinking about another’s thoughts increases metabolic activity in a specific group of
brain regions. These regions, which comprise the ‘theory of mind network’, include the
medial prefrontal cortex (MPFC), precuneus (PC), right superior temporal sulcus
(RSTS), and bilateral temporal-parietal junction (TPJ). While many studies have
investigated the selectivity and domain specificity of these brain regions for theory of
mind (e.g., Aichhorn et al., 2009; Saxe & Kanwisher, 2003), a distinct but fundamental
question concerns the computational roles of these regions: which features of people’s
beliefs and intentions are represented, or made explicit, in these brain regions? Prior
work has focused on where in the brain mental state reasoning occurs, whereas the
present research builds on this work to investigate how neural populations encode
these concepts.
A powerful approach for understanding neural representation in other domains has been
to ask which features of a stimulus can be linearly decoded from a population of
neurons. For example, in the ventral visual stream (involved in object recognition), lowlevel stimulus properties like line-orientation and shading are linearly decodable from
small populations of neurons in early visual areas (e.g., V1), whereas in higher-level
regions the identity of an object becomes linearly decodable, and invariant across
viewing conditions (DiCarlo, Zoccolan, & Rust, 2012; Kamitani & Tong, 2005). These
results suggest that as information propagates through the ventral pathway, the neural
response is reformatted to make features that are relevant to object identity more
explicit to the next layer of neurons (DiCarlo et al., 2012).
A decoding approach can be similarly applied to functional magnetic resonance image
(fMRI) data, using multi-voxel pattern analysis (MVPA) to examine the spatial pattern of
neural response within a brain region. If a distinction between cognitive tasks, stimulus
categories, or stimulus features is coded in the population of neurons within a brain
region, and if the subpopulations within the region are (at least partially) organized into
spatial clusters or maps over cortex (Dehaene & Cohen, 2007; Formisano et al., 2003),
then the target distinction may be detectable in reliable spatial patterns of activity
measurable with fMRI (Haynes & Rees, 2006; Kriegeskorte, Bandettini, & Bandettini,
2007; Norman, Polyn, Detre, & Haxby, 2006). MVPA has therefore been used to identify
categories and features that are represented within a single region, (e.g., Kamitani &
Tong, 2006; Mahon & Caramazza, 2010; Peelen, Wiggett, & Downing, 2006), and to
relate these representations to behavioral performance (Haynes & Rees, 2006; Norman
et al., 2006; Raizada et al., 2010).
Compared to object recognition, much less is known about the cognitive and neural
mechanisms that support theory of mind. However, linear separability of the neural
response could serve as a diagnostic measure of the core features and local
computations even in this abstract domain. We therefore asked whether the spatial
pattern of response in theory of mind brain regions could be used to decode a feature
that has previously been shown to be critical for theory of mind: whether an action was
performed intentionally or accidentally.
46 of 179
The distinction between intentional and accidental acts is particularly salient in the case
of moral cognition. Adults typically judge the same harmful act (e.g., putting poison in a
drink, failing to help someone who is hurt, making an insensitive remark) to be more
morally wrong and more deserving of punishment when committed intentionally versus
accidentally (e.g., Cushman, 2008). These moral judgments depend on individuals’
ability to consider another person’s beliefs, intentions, and knowledge, and emerge
relatively late in childhood, around age 6-7 years (e.g., Baird & Astington, 2004).
Individuals with Autism Spectrum Disorders (ASD), who are disproportionately impaired
on tasks that require them to consider people’s beliefs and intentions (Baron-Cohen,
1995; Peterson, Wellman, & Liu, 2005), are also impaired in using information about an
innocent intention to forgive someone for accidentally causing harm (Grant, 2005;
Moran et al., 2011; Yang, Baillargeon, & Baillargeon, 2013) but see (Baez et al., 2012).
The RTPJ is particularly implicated in these moral judgments. In prior research, (i)
increased TPJ activation is related to greater consideration of mitigating intentions and
more lenient punishment (Buckholtz et al., 2008; Yamada et al., 2012), (ii) individual
differences in the forgiveness of accidental harms are correlated with the magnitude of
activity in the RTPJ at the time of the judgment (Young & Saxe, 2009b), and (iii)
interfering with activity in the RTPJ shifts moral judgments away from reliance on mental
states (Young, Camprodon, Hauser, Pascual-Leone, & Saxe, 2010a).
Given the importance of intent for moral judgments of harms, we predicted that one or
more of the brain regions in the theory of mind network would explicitly encode this
feature of others’ mental states in neurotypical (NT) adults. That is, we predicted that (i)
while participants read about a range of harmful acts, we would be able to decode
whether the harm was intentional or accidental based on the spatial pattern of activity
within theory of mind brain regions. We tested this prediction in three experiments with
NT adults. We also investigated (ii) whether the robustness of the spatial pattern within
individuals would predict those individuals’ moral judgments, and (iii) whether, in a fourth
experiment, high-functioning adults with ASD, who make atypical moral judgments of
accidental harms, would show atypical patterns of neural activity in pattern or
magnitude.
In all four experiments, participants in the scanner read short narratives in which
someone caused harm to another individual, intentionally or accidentally, as well as
narratives involving no harm. Participants in Experiments 1 and 2 made a moral
judgment about the action. Participants in Experiment 3 made true/false judgments
about facts from the narratives. In Experiment 4, high-functioning adults with ASD read
and made moral judgments about the same narratives as in Experiment 1.
47 of 179
Methods
Participants
Experiment 1: Twenty-three right-handed members of the local community (aged 18-50
years, mean=27, 7 women). Experiment 2: Sixteen right-handed college
undergraduate students (aged 18-25 years, 8 women). Experiment 3: Fourteen righthanded college undergraduate students (aged 18-25 years, 8 women). Experiment 4:
Sixteen individuals diagnosed with Autism Spectrum Disorders (aged 20-46 years,
mean=31, 2 women). Participants in Exp.4 were recruited via advertisements placed
with the Asperger’s Association of New England. All participants were prescreened
using the Autism Quotient questionnaire (AQ; Baron-Cohen, 2001). ASD participants
(mean=32.6) scored significantly higher on the AQ than the NT participants from Exp.1
(mean=17.3; t(25.5)=6.4, p<0.0001).
ASD participants underwent both the Autism Diagnostic Observation Schedule (Joseph,
Tager-Flusberg, & Lord, 2002; ADOS; Lord et al., 2000) and impression by a clinician
trained in both ADOS administration and diagnosis of ASD. All ASD participants
received a diagnosis of an Autism Spectrum Disorder based on their social ADOS score
(criterion of ≥ 4; mean=6.4), communication ADOS score (criterion of ≥ 2; mean=3.1),
and total ADOS score (criterion ≥ 7; mean=9.5), and on clinical impression based upon
the diagnostic criteria of the DSM-IV (American Psychiatric Association & American
Psychiatric Association Task Force on DSM-IV, 2000). The NT (Exp.1) and ASD (Exp.4)
groups did not differ in age (NT mean=27, ASD mean=31, t(35.26)=1.32, p=0.2) or IQ
(NT mean=121; ASD=120; t(28.3)=0.26, p=0.8). Additionally, all analyses were run with
one-to-one matched subsets of participants (both n=15, see Appendix).
All participants were native English speakers, had normal or corrected-to-normal vision,
gave written informed consent in accordance with the requirements of Institutional
Review Board at MIT, and received payment. Data from Exp.2 and 3 have previously
been published in papers analyzing the magnitude but not the pattern of response in
each region (Young & Saxe, 2008; 2009a).
fMRI Protocol and Task
Experiments 1 and 4
Participants were scanned while reading 60 stories (Figures 2.1): 12 described harm
caused intentionally (e.g. knowingly kicking someone in the face), 12 described harm
caused accidentally (e.g. kicking without seeing the person), 12 described neutral
actions (e.g. eating lunch), and 24 stories describing disgusting but not harmful actions
(e.g. smearing feces on one’s own face, not analyzed here). Stories were presented in 4
cumulative segments, describing the background (6s), action (+4s), outcome (+4s) and
intention (+4s). After each story, participants made a moral judgment of the action (“How
much blame should you get?”) from “none at all” (1) to “very much” (4), using a button
press. Behavioral data from 5 ASD and 3 NT participants were lost due to error (n=4) or
to theft of experimental equipment (n=4). Ten stories were presented in each 5.5 min
run; the total experiment, six runs, lasted 33.2 min.
48 of 179
Experiment*1+4
Accidental
Intentional
Experiment*2
Accidental*Harm
Intentional*Harm
Background
Action
Outcome
Intent
You%had%absolutely*no*idea*
Your%family%is%over% You%grind%up% Your%cousin,%
about%your%cousin's%allergy%
for%dinner.%You%wish%
some%
one%of%your%
when%you%added%the%peanuts.
to%show%off%your%
peanuts,%add%
dinner%
culinary%skills.%For% them%to%that% guests,%is%
one%of%the%dishes,%
dish,%and%
severely%
You%knew%about%your%cousin's%
adding%peanuts%will%
serve%
allergic%to%
peanut%allergy%when%you%added%
really%bring%out%the%
everyone.
peanuts.
the%peanuts%to%the%dish.
flavor.
Background
Foreshadow
Belief
The%container%is%
labeled%“sugar”,%
Grace%and%her%friend%
so%Grace%believes%
are%taking%a%tour%of%a%
The%white%
that%the%white%
chemical%plant.%When%
powder%is%a%very% powder%is%regular*
Grace%goes%over%to%the%
toxic%substance%
sugar.
coffee%machine%to%pour%
left%behind%by%a%
some%coffee,%Grace’s%
The%container%is%
scientist,%and%
friend%asks%for%some%
labeled%“toxic”,%
deadly%when%
sugar%in%hers.%There%is%
so%Grace*believes%
ingested.
white%powder%in%a%
that%the%white%
container%by%the%coffee.
powder%is*a*toxic*
Outcome
Grace%puts%the%
substance%in%her%
friend’s%coffee.%
Her%friend%drinks%
the%coffee%and%
dies.%
substance.*
Experiment*3
Accidental*Harm
Intentional*Harm
Background
Belief
Outcome
Peter%believes%that%it%is*safe%to%
wade%in%the%pond%because*other*
Peter%is%traveling%in%
Africa%with%a%friend.%His% tourists*around*them*are*doing* Malarial%mosquitoes%
it*too*and*seem*to*be*enjoying*
friend%sees%a%pond%and%
actually%live%in%the%
themselves.*
wants%to%go%wading%in%it%
pond.%A%single%bite%is%
because%it%is%very%hot.%
Peter%believes%that%it%is*unsafe*to% enough%to%create%an%
Peter%encourages%his%
wade%in%the%pond*because%Africa* infection,%so%the%pond%
friend%to%wade%in%the%
is%unsafe%to%wade%in.
is*known*for*malarial*
pond.%
mosquitoes,*and*mosquitoes*
congregate*around*water.*
Figure 2.1. Sample stimuli from Exp. 1 and 4 (top), Exp. 2 (middle), and Exp. 3 (bottom)
Participants were scanned while reading 60 stories (Figure 2.1). Stories were
presented in the second person, using present tense. Each participant read 12 stories
describing someone harming another individual caused accidentally, 12 describing
someone harming another individual intentionally, 12 neutral actions (described below),
and 24 stories describing disgusting but not harmful actions (e.g., consensual incest,
drinking blood, smearing feces on one’s own face, not analyzed here).
Each story was displayed in four cumulative segments (i.e. earlier segments remained
on the screen as later segments were added). Cumulative segments partially reduce the
49 of 179
variability in reading time across participants and items, by slowing down faster
participants and shorter items, without cutting off reading early too often in longer
segments or for slower readers.
For accidental and intentional harms, the initial three segments of the stories were
identical: information about the Background (6s), Action (+4s), and harmful Outcome
(+4s). Harmful outcomes included both physical harms (kicking someone in the face,
feeding them an allergen) and psychological harms (insulting or humiliating someone).
The only information that distinguished between these two conditions was the final
sentence, describing the Intent (+4s), which was innocent in the accidental harm
condition (e.g. “you did not see”, “you did not know” that the act would cause harm) and
negative in the intentional harm condition (e.g. “you saw”, “you realized” that the act
would cause harm). The accidental and intentional sentences were matched pairwise by
changing the fewest possible words in the Intent segment, thus preserving length (see
Figure 2.1 for examples; mean=14 words, standard deviation=1.9, range=9-18). Across
subjects, each scenario (i.e. Background+Action+Outcome) was equally likely to occur
in the Intentional or Accidental condition. Each participants saw either the intentional or
the accidental version of each scenario, but not both.
For the neutral actions, a separate set of scenarios described a neutral Background and
Action (e.g. giving a class presentation, eating lunch). The Outcome was mildly positive
(the teacher liked the graph) or negative (a little spilled ketchup); and the Intent
described either knowledge or no knowledge of the outcome (e.g. “you did/did not
realize” that you spilled ketchup). Because there were only 12 neutral stories in total, we
used the averaged the response for all neutral stories in the analyses.
In the scanner, after each story, participants made a moral judgment of the action (“How
much blame should you get?”) from “none at all” (1) to “very much” (4), using a button
press. Participants were given 4 seconds to respond before the screen was blanked.
Thus, the whole trial lasted 22 seconds. A 10 s rest block (blank screen) occurred after
each trial.
Stories were presented in a pseudorandom order, such that no condition was
immediately repeated, each condition was equally likely to occur in each position in the
run, and each condition was equally likely to follow each other condition. Ten stories
were presented in each 5.5 min run; the total experiment, six runs, lasted 33.2 min.
Stories were projected onto a screen via Matlab 5.0 running on an Apple MacBook Pro
in 40-point white font.
Experiment 2
Young et al.(Young & Saxe, 2008) report results from 17 participants; data from one
participant have been lost, leaving 16 participants.
Participants were scanned while reading 48 stories. Stories were presented in the third
person, using past tense. Each participant read stories describing 12 intentional harms,
12 accidental harms, 12 failed attempts to harm, and 12 neutral actions (Figure 2.1; for
the complete list see Young & Saxe, 2008). The focus of the current manuscript is the
accidental and intentional harm stories. All harms were physical harms, resulting in
someone’s death or serious injury.
50 of 179
Stories were presented in four cumulative segments: Background (6s); Foreshadow
(+6s); Belief (+6s); and Outcome (+6s). Half of the stories in each run were presented
with Foreshadow before Intent; the other half were presented with Intent before
Foreshadow. The Background, Foreshadow, and Outcome segments of intentional and
accidental harms were identical. The Belief segments described the content of the
character’s belief (e.g. “Leah believes that the child is about to dive into shallow water
and break his neck”) and the reason for that belief (e.g. “Because of a warning sign at
the side of the pool”). The difference between conditions was created by manipulating
the content of the belief. In the context of the foreshadow (“The child is about to dive
into the shallow end and smack his head very hard”) the action (“Leah walks by, without
saying anything to the child”), and the negative outcome, these beliefs determined
whether the harm (“The child dives in and breaks his neck”) was caused intentionally or
accidentally.
After each story, participants made moral judgments (“How morally permissible was X’s
action?”) of the action on a 3-point scale, from “forbidden (1) to “permissible” (3), using
a button press. Participants were given 4 seconds to respond before the screen was
blanked. All trials were followed by a 14s rest block. Stories were projected onto a
screen via Matlab 5.0 running on an Apple G4 laptop in 24-point white font. Behavioral
data were available for 10 participants. The methods and materials from Experiment 2
have been published in (Young & Saxe, 2008).
Experiment 3
Participants were scanned while reading 48 stories. Stories were presented in the third
person, using present tense. Each participant read stories describing 8 intentional
harms, 8 accidental harms, 8 failed attempts to harm, 8 neutral actions, and 16 actions
for which the outcome was unknown (a morally irrelevant fact was presented in place of
the outcome; Figure 2.1 and Young & Saxe, 2009a Experiment 2). Stories were
presented in cumulative segments: Background (6s), Belief (+6s), and Fact (+6s). The
focus of the current manuscript is the accidental and intentional harm stories; the
Background and Fact segments were identical for these conditions. Belief segments
described the content of the character’s belief (e.g. “Steve believes the ground beef is
safe to eat”) and the reason for that belief (e.g. “because of the expiration date.”). The
difference between conditions was created by manipulating the content of the belief. In
the context of the fact (e.g. “The meat has some invisible but deadly bacteria”) and the
action (Steve “makes a large meatloaf for all the children”), these beliefs led to either
accidental or intentional harm, although the harmful outcome itself (i.e. the children
becoming ill) was only implied, not explicit stated. After each story, participants
answered a true/false question about the Fact segment (e.g. “The meat is unsafe for
consumption: True / False.”). Participants were given 4 seconds to respond before the
screen was blanked. All stories were followed by a 14s rest block. Stories were
projected onto a screen via Matlab 5.0 running on an Apple G4 laptop in 24-point white
font. The methods and materials from Experiment 3 have been published in (Young &
Saxe, 2009a).
51 of 179
Theory of Mind Localizer task
All subjects participated in a theory of mind localizer task, contrasting stories requiring
inferences about mental state representations (e.g., thoughts, beliefs) versus physical
representations (e.g., maps, signs, photographs). Mental state stimuli versus physical
representation stimuli were similar in their meta-representational and logical complexity;
the key difference was that for the mental state stories the reader had to build a
representation of someone else’s mental state. See (Dodell-Feder, Koster-Hale, Bedny,
& Saxe, 2011; Saxe & Kanwisher, 2003) for further discussion. Stimuli and experimental
design are available at http://saxelab.mit.edu/superloc.php. This experiment was
modeled treating each trial as a block, with a boxcar lasting 14 seconds, the whole
period from the initial presentation of the story to the end of the question presentation.
Beta values were estimated, in each voxel, for stories describing mental states (“Belief”)
or physical representations (“Photo”). Then a simple contrast map was produced in
each subject, identifying voxels responding more to Belief than Photo stories, at
p<0.001 uncorrected for multiple comparisons.
Acquisition and Preprocessing
In all four experiments, fMRI data were collected in a 3T Siemens scanner at the
Athinoula A. Martinos Imaging Center at the McGovern Institute for Brain Research at
MIT, using a 12-channel head coil. Using standard echoplanar imaging procedures,
blood oxygen level dependent (BOLD) signal was acquired in 26 near axial slices using
3x3x4 mm voxels (TR=2 s, TE=40 ms, flip angle=90°). To allow for steady state
magnetization, the first 4 seconds of each run were excluded. Data processing and
analysis were performed using SPM8 (Exp.1 and 4; (http://www.fil.ion.ucl.ac.uk/spm),
SPM2 (Exp.2 and 3) and custom software. The data were motion corrected, realigned,
normalized onto a common brain space (Montreal Neurological Institute (MNI)
template), spatially smoothed using a Gaussian filter (full-width half-maximum 5 mm
kernel) and high-pass filtered (128 Hz).
Motion and Artifact Analysis
For the NT (Exp.1) and ASD (Exp.4) participants, we calculated the total motion per run
(sum of translation in three dimensions), and the number of timepoints that either (a)
deviated from the global mean signal by more than 3SD, or (b) included TR-to-TR
motion of more than 2mm.
fMRI Analysis
All fMRI data were modeled using a boxcar regressor, convolved with a standard
hemodynamic response function (HRF). The general linear model was used to analyze
the BOLD data from each subject as a function of condition. The model included
nuisance covariates for run effects, global mean signal, and an intercept term. A slow
event-related design was used. An event was defined as a single story, beginning with
the onset of text on screen and ending after the response prompt was removed.
52 of 179
Functional Localizer: Individual subject ROIs
Functional regions of interest (ROIs) were defined in right and left temporo-parietal
junction (RTPJ, LTPJ), medial precuneus (PC), and dorsal medial prefrontal cortex
(DMPFC). For each participant, we found the peak voxel in the contrast image for each
region. ROIs were then defined as all voxels within a 9mm radius of the peak voxel that
passed threshold in the contrast image (Belief > Photo, p<0.001 uncorrected, k>10);
Within-ROI Magnitude Analysis
We measured the response to each condition in each ROI. Baseline response in each
region was the average BOLD response, in that region, at at all timepoints when there
was no stimulus on the screen, excluding the first 4 seconds after the off-set of each
stimulus (to allow the hemodynamic response to decay). The percent signal change
(PSC) relative to baseline was calculated for each time point in each condition,
averaging across all voxels in the ROI and across all blocks in the condition, where
PSC(at time t)=100 × ((average BOLD magnitude for condition (at time t) – average
BOLD magnitude for fixation) / average BOLD magnitude for fixation). We averaged the
PSC across the entire presentation (offset 6s from presentation time to account for
hemodynamic lag) to estimate a single PSC for each condition, in each ROI, in each
participant (Poldrack, 2006). Using custom software to visualize the percent signal
change in a region, we extract the full timecourse in the region (i.e. set of voxels), and
then create an event-related average, by aligning and averaging the raw BOLD
response at each time point after the trial onset, per condition (rather than extracting the
beta value for the whole block). Other software packages (e.g. SPM) calculate the
baseline response as the mean over the run, which is undesirable because it will overestimate the response at baseline in a region with a strong positive response to most
blocks, and will under-estimate the response at baseline in a region that deactivates
during most blocks.
In addition, because the stimuli in the current experiment were presented cumulatively,
for Experiments 1 & 4 we separately analyzed the PSC for two phases of the trial: (1)
the initial phase (Background + Action + Outcome, first 14 s), in which intentional and
accidental harms could not be distinguished, and (2) the final phase (Intent + Decision,
final 8 s), after intentional and accidental harms were distinguished.
Within-ROI Pattern Correlation Analysis
In all four experiments, we conducted within-ROI pattern analyses. Following Haxby et
al. (2001), each participant's data were divided into even and odd runs (‘partitions’) and
the mean response (beta value) of every voxel in the ROI was calculated for each
condition. The pattern of activity was defined as the vector of beta values across voxels
within the ROI. To calculate the within-condition correlation, the pattern in one (e.g.,
even) partition was compared to the pattern for the same condition in the opposite (e.g.,
odd) partition; to calculate the across-condition correlation, the pattern was compared to
the opposite condition, across partitions. For each individual, an index of classification
was calculated as the z-scored within-condition correlation minus the z-scored acrosscondition correlation. A region successfully classified a difference in conditions if, across
individuals, the within-condition correlation was higher than the across-condition
53 of 179
correlation, using a Student's T complementary cumulative distribution function. Note
that in this procedure (unlike other machine learning approaches, e.g., support vector
machines) differences in the spatial patterns across conditions are independent of
differences in the average magnitude of response. This procedure implements a simple
linear decoder, which is, while in principle less flexible and less powerful than non-linear
decoding, preferable both theoretically and empirically (e.g., DiCarlo et al., 2012;
Kamitani & Tong, 2005; Norman et al., 2006).
Whole Brain Pattern Correlation Analysis (Searchlight)
Searchlight. In this analysis, rather than using a predefined ROI, a Gaussian kernel
(14mm FWHM, corresponding approximately to the size of the functional ROIs) is
moved iteratively across the brain. Using the same logic as the ROI-based MVPA, we
compute the spatial correlation, in each kernel, of the neural response (i.e. betas) within
conditions and across conditions; we transform the correlations using Fisher’s Z, and
subtract the across-condition from the within-condition correlation to create an index of
classification. Thus, for each voxel, we obtain an index of how well the spatial pattern of
response in the local region (i.e. the area centered on that voxel) can distinguish
between the two conditions of interest. The use of a Gaussian kernel smoothly deemphasizes the influence of voxels at increasing distances from the reference voxel
(Fedorenko, Nieto-Castañón, & Kanwisher, 2012). We created whole brain maps of the
index of classification for each subject. These individual correlation maps were
subjected to a second-level analysis using a one-sample t-test (thresholded at p<0.001,
voxelwise, uncorrected). Note that because each voxel’s discrimination index reflects
the spatial pattern in the surrounding region, the map of spatial discrimination is highly
smooth, reducing the effective number of comparisons. "
This classification procedure implements a simple linear decoder. Linear decoding,
while in principle less flexible and less powerful than non-linear decoding, is preferable
both theoretically and empirically. A non-linear classifier can decode nearly any arbitrary
feature contained implicitly within an ROI, reflecting properties of the pattern analysis
algorithm rather than the brain, which makes successful classification largely
uninformative (Cox & Savoy, 2003; DiCarlo & Cox, 2007; Goris & Op de Beeck, 2009;
Kamitani & Tong, 2005; Norman et al., 2006). Moreover, linear codes have been argued
to be more neurally plausible, reflecting the information made available to the next layer
of neurons (Bialek, Rieke, Van Steveninck, & Warland, 1991; Butts et al., 2007; DiCarlo
& Cox, 2007; Naselaris, Kay, Nishimoto, & Gallant, 2011; Rolls & Treves, 2011).
54 of 179
Results
fMRI task: Behavioral Results
Because participants used the scales in different ways (e.g., some using largely 2-4,
others using largely 1-3), we z-scored the behavioral data (see Appendix for
untransformed data).
Experiment 1: Intentional harms (1.2±0.03) were rated as more blameworthy than
accidental harms (-0.38±0.04; t(19)=-25, p<0.001), and both were rated more
blameworthy than neutral acts (-0.8±0.03; t(19)=50, p<0.001; t(19)=6.9, p<0.001).
Experiment 2: Replicating the results in Exp.1, participants judged intentional harms
(1.0±.05) to be worse than accidental harms (-0.53±.11; t(9)=16.0, p<0.001).
Experiment 3: Participants in Exp.3 did not make moral judgments of the scenarios, but
instead answered true/false questions about the content of the fact. Reaction time (RT)
was analyzed using a one way, repeated measures ANOVA, revealing a main effect of
condition, (F(1,13)=4.56, p<0.02). Post hoc t-tests revealed that RTs for answering
factual questions following non-harm stories (mean RT=2.4 sec) were not different from
RTs following accidental harms (mean 2.3 sec; t(13)=1.65 p=0.12), or intentional harms
(mean 2.5 s, t(13)=1.49 p=0.16), but that responses following intentional harms were
slower than responses following accidental harms (t(13)=2.87, p< 0.01).
Experiment 4: When making moral judgments, ASD participants, like NT participants
from Exp.1, judged intentional harms (1.0±0.09) more blameworthy than accidental
harms (-0.23±0.08, t(10)=8.0 p<0.0001), and both intentional and accidental harms
were rated worse than neutral acts (-0.77±0.08, t(10)=11.5, p<0.0001; t(10)=4.3,
p=0.0015).
Group Comparison: A mixed effects ANOVA crossing Group (NT in Exp.1, ASD in Exp.
4) by Condition (accidental, intentional, z-scored ratings) yielded a main effect of
Condition (F(1,29)=446.9, p<0.0001) and a Group by Condition interaction (F(1,29)=4.7,
p=0.03). A planned comparison t-test (Moran et al., 2011) revealed that ASD adults
assigned more blame for accidental harms than NT adults (t(29)=1.9, p=0.03, onetailed). An additional post-hoc t-test revealed that ASD adults also assigned less blame
to intentional actions (t(29)=2.1, p=0.04).
Motion and Artifact Analysis Results
There was no difference in total motion between NT (Exp.1, mean=0.24mm/run) and
ASD participants (mean=0.23mm/run, t(37)=0.24, p=0.81), or in the number of outliers
per run (NT: 3.7±.86; ASD: 3.4±1.2, t(37)=0.2, p=0.8).
fMRI Results: Functional Localizer
Replicating studies using a similar functional localizer task (e.g., Saxe & Kanwisher,
2003), we localized four theory of mind brain regions showing greater activation for
mental state stories (e.g., describing false beliefs) compared to physical state stories
(e.g., describing “false” or outdated physical representations, such as photographs;
p<0.001, k>10) in the majority of individual participants. Four ROIs were identified for
55 of 179
each subject, based on the contrast “belief > photo” thresholded at P < 0.001. For
experiments 1–3 (NT), these were as follows: RTPJ (52/53), LTPJ (51/53), pre- cuneus
(PC; 51/53), and dorsal medial prefrontal cortex (DMPFC; 43/53). For experiment 4
(ASD), these were as follows: RTPJ (16/16), LTPJ (15/16), PC (16/16), and DMPFC
(7/16) Table 2.1 reports the average position of the peak and the average and SD of the
ROI size for each ROI for each experiment. All subsequent analyses are conducted
using individually-defined regions of interest (ROIs).
Experiment*1
Region
size
Experiment*2
Average-PeakVoxel
x
y
z
size
Experiment*3
Average-PeakVoxel
x
y
z
size
Experiment*4*(ASD)
Average-PeakVoxel
x
y
z
size
Average-PeakVoxel
x
y
z
RTPJ
260'
(17)
54
G54
22
151'
(18)
56
G56
22
106'
(22)
53
G55
21
194'
(24)
55
G55
24
LTPJ
144'
(26)
G51
G56
22
121'
(21)
G50
G63
26
178'
(23)
G54
G58
19
156'
(35)
G51
G60
26
PC
253'
(25)
0
G58
35
183'
(25)
G1
G58
39
192'
(20)
1
G58
37
260'
(23)
1
G54
33
DMPFC
142'
(23)
3
44
24 87'(17)
G2
58
29
130'
(19)
2
54
36
175'
(55)
6
54
31
Table 2.1: Individual subject ROIs. Individual subject ROIs are defined as all voxels within a 9mm sphere
around the peak voxel (T of mental-physical from the localizer) identified in each region. Size: average
number ± standard deviation of voxels included in individual ROIs. Average Peak Voxel: average
coordinates of the peak voxels, across participants, reported in Montreal Neurological Institute (MNI)
space.
fMRI Results: Response Magnitude
Experiment 1 (NT): Averaged over the whole trial, harmful actions elicited a higher
response than neutral acts in all four ROIs (RTPJ, LTPJ, PC, DMPFC; all t>4.6,
p<0.0003).
In the initial phase only (Background, Action, Outcome), we compared the response to
harms versus neutral acts. In all four ROIs, harmful actions elicited a higher response
than neutral acts (all t>4, p<0.001).
In the final 8 s of the trial, after the intention had been revealed, the response in the
RTPJ of NT adults was higher for accidental than intentional harms (mean percent
signal change from rest, accidental: 0.1±0.04, intentional: 0.01±0.04, t(21)=3.59
p=0.002). LTPJ showed a similar trend (t(21)=1.82 p=0.08); there was no difference in
the other two regions. There was no difference in PC or DMPFC (both t<0.2, p>0.3).
Experiment 4 (ASD): Averaged over the whole trial, harmful actions elicited a higher
response than neutral acts in all four ROIs (RTPJ, LTPJ, PC, DMPFC; all t>3, p<0.02).
In the initial phase only (Background, Action, Outcome), we compared the response to
harms versus neutral acts. In all four ROIs, harmful actions elicited a higher response
than neutral acts (all t>3.5, p<0.0012). In the final phase only, after the intention had
56 of 179
been revealed, in ASD adults, no region showed a significant difference between
accidental and intentional harms (all t<1, p>0.3).
Group Comparison: We directly compared the magnitude of responses across groups
(ASD, NT) in three ways. First we compared the response to all harms vs all neutral
acts, averaged across the whole trial. No region showed a main effect of Group or a
Group by Condition interaction (all F<2.6, all p>0.12).
RTPJ
0.6
Harm
Accidental
Percent Signal Change
Intentional
0.4
NT
ASD
0.2
0.0
intent information
−0.2
0
5
10
Seconds
15
20
25
Figure 2.2: Percent Signal Change in the RTPJ of ASD and NT adults, for accidental harms,
intentional harms, and neutral stories. NT: n=22, ASD: n=16. Error bars indicate within-participant
SEM. In the initial phase (Background, Action, and Outcome, seconds 0-14, plus 4 s for hemodynamic
lag), both NT and ASD show a significant increase in mean activity in RTPJ for harms stories, relative to
rest. After information about intent was revealed (second 14 plus 4 s for hemodynamic lag), NT showed
increased activity for accidental harms, relative to intentional harms, while ASD showed no difference.
Second, we compared the response to all harms vs all neutral acts during just the initial
phase (Background, Action, and Outcome) of each trial. In PC, the response trended
towards being overall higher in the ASD group; an ANOVA crossing Group (ASD, NT) by
Condition (harms, neutral) revealed a marginal main effect of group (F(1,35)=3.37,
p=0.08). There were no other main effects of Group or any Group by Condition
interactions (all F<1.7, p>0.2)."
Finally, we compared the response to accidental vs intentional harms, specifically during
the final phase of the trial. A Group (NT, ASD) by Condition (accidental, intentional)
ANOVA, revealed a Group by Condition interaction in RTPJ (F(1,36)=5.37, p=0.03),
such that NT showed a difference between accidental and intentional harms, while ASD
did not). No other effects of Group, or Group by Condition interactions, were significant
(all F<1.3, p>0.25). See Figure 2.2 for the Percent Signal Change for both groups.
57 of 179
Experiment 2 & 3: See (Young & Saxe, 2008; 2009a) for analysis of the response
magnitude in Experiments 2 and 3.
fMRI Results: Pattern Analysis for Neurotypical
Within ROI Pattern Analysis for Experiment 1
Harm vs. Neutral: Multi-voxel pattern analyses revealed reliably distinct patterns of
neural activity for harmful (intentional, accidental) versus neutral acts — the pattern
generated by stories in one category (i.e., harmful or neutral) was more correlated with
the pattern from other stories in the same category than in the opposite category — in
three of four ROIs: RTPJ, LTPJ, and PC (RTPJ: within=1.3±0.10, across=1.1±0.10,
t(21)=3.2, p=0.002; LTPJ: within=1.6±0.05, across=1.4±0.06, t(21)=3.3, p=0.002; PC:
within=1.1±0.09, across=0.86±0.1, t(21)=3.4, p=0.001). DMPFC showed the same trend
(DMPFC: within=1.2±0.09, across=1±0.09, t(17)=1.7, p=0.056; Figure 2.3 A). All
correlations are Fisher Z transformed to allow statistical comparisons with parametric
tests; see Appendix for an analysis of untransformed data.
C
NT
Within
Across
ASD
RTPJ
Within
Across
2
.5 1 1.5
0
2
Exp. 4 ASD
.5 1 1.5
Exp.1 NT
∗
Exp. 1 NT
Exp. 4 ASD
∗
∗
Exp. 2 NT
Exp. 3 NT
0
∗
correlation (Z)
∗
correlation (Z)
2
.5 1 1.5
B Accidental vs Intentional
0
correlation (Z)
A Moral vs Neutral
Figure 2.3: MVPA results from Experiment 1-4. (A) Neurotypical (n=23) and ASD adults (n=16) show
pattern discrimination in the RTPJ for moral vs. neutral actions, with higher within-condition correlations
than across- condition correlations. (B) NT adults show pattern discrimination for accidental vs. intentional
harms in RTPJ, a finding replicated across Experiments 2 and 3 (n=16,14), but adults with ASD do not
show discrimination of intent. High overall correlation in ASD, combined with matched motion parameters
for both groups, suggest that this is not due to noise, but rather a stereotyped response across both
accidental and intentional harms in adults with ASD. (C) The RTPJ in a single participant. Error bars
indicated SEM.
58 of 179
NT
ASD
W
Accidental vs. Intentional: Only in RTPJ did the pattern of activity distinguish between
accidental and intentional harms (within=1.2±0.12, across=1.1±0.12, t(21)=2.2, p=0.02).
No other regions showed sensitivity to intent (all correlation differences<0.02, all p>0.3,
Figure 2.3 B).
Within ROI Pattern Analysis for Experiment 2 and 3
Exp.2 and 3 replicate Exp.1. In only the RTPJ did MVPA reveal reliably distinct neural
patterns for intentional and accidental harms (Exp.2 RTPJ: within=1.1±0.13,
across=0.91±0.10, t(15)=2.6, p=0.01; Exp.3 RTPJ: within=0.42±0.10, across =
0.24±0.10, t(13)=2, p=0.034; all other regions: correlation differences<0.1, p>0.1,
Figure 2.3 B).
RTPJ
LTPJ
PC
MPFC
C
Moral Judgment
1 1.5 2
.5
B
Pattern Correlation
0 .5 1 1.5
A
∗∗
Within
Across
∗∗
r2=.28
Within
Across
∗
n = 29
r2=.14
Within
Across
ns
n = 29
r2=.04
Within
Across
ns
n = 29
-0.7 -0.3 1 0.5 0.9
-0.6 -0.3 0 0.3 0.6
-0.8 -0.4 0 0.4 0.8
Pattern Discrimination
Pattern Discrimination
Pattern Discrimination
r2=.02
-0.7 -0.4
n = 26
0
0.4 0.7
Pattern Discrimination
Figure 2.4: MVPA results from Experiments 1, 2 & 3. (A) theory of mind regions from a single
participant. (B) Across three experiments, neurotypical adults (n=53) show pattern discrimination for
accidental vs. intentional harms in the right TPJ, but not in any other theory of mind region. (C) Moreover,
individual differences in pattern discrimination (within condition correlation minus across condition
correlation) predict differences in behavioral moral judgment (ratings of intentional minus accidental) in
NT adults in RTPJ and LTPJ (n=29, Exp. 1&2), such that adults with more discriminable patterns are
more forgiving of accidental harms relative to intentional harms. Error bars indicated SEM.
Pooling the data across all three experiments increased our power to detect results in
neural regions beyond RTPJ. Again, MVPA revealed distinct patterns for accidental and
intentional harms in RTPJ (within=0.96±0.08, across=0.81±0.08, t(51)=3.9, p=0.0002)
and no other region (all differences<0.1, p>0.1, Figure 2.4 B).
59 of 179
Behavioral and Neural Correlation
In Exp.1 and 2, NT participants provided moral judgments of each scenario in the
scanner, allowing us to determine whether behavioral responses were related to neural
pattern. In both experiments, we found that only in the RTPJ, the difference between
intentional and accidental harms in individuals’ moral judgments was correlated with the
neural classification index (within-across condition correlation; Exp.1: r(17)=0.6;
p=0.007; Exp.2: r(8)=0.65; p=0.04). The correlation between neural pattern and
behavior was also significant after combining the data from both experiments (r(27)=0.6,
p=0.0005; Figure 2.4 C). Though neither Exp.1 nor Exp.2 showed a significant
correlation in LTPJ alone, combining data from both revealed a significant correlation in
LTPJ as well (r(27)=0.38, p=0.04). The other regions showed no significant correlation
with behavioral judgments (all r<0.4, p>0.1).
Whole Brain Searchlight Analysis for Experiment 1,2,3 (NT)
Combining across all NT participants (n=53, Exp 1, 2, 3) we found only one small region
that discriminated (p<0.001, voxel-wise, uncorrected) between accidental and
intentional harms in the fusiform gyrus (peak voxel MNI coordinates: [30,-54,-14],
Figure 2.5 (top): Whole Brain Searchlight Analysis in Neurotypical Adults for Accidental vs.
Intentional Harms. (n=53, p<0.001, uncorrected) Sagittal slice, showing region in left fusiform gyrus,
and, and axial slices 40- 110.
Figure 2.5 (bottom): Whole Brain Searchlight Analysis in Behaviorally Sensitive Neurotypical
Adults for Accidental vs. Intentional Harms. (n=15, p<0.001, uncorrected) Surface, showing region in
anterior RTPJ, and slices 40-110. Participants were median-split based on the difference in their moral
judgments between intentional and accidental harms (Exp.1&2). RTPJ is the only region that
discriminated intentional and accidental harms in the group with the largest difference in their ratings of
accidental and intentional harms.
60 of 179
Figure 2.5). We then median-split participants based on the difference in their moral
judgments between intentional and accidental harms (Exp.1&2). In participants showing
a larger behavioral effect (n=15) the only region that discriminated between intentional
and accidental harms was in RTPJ (peak voxel MNI coordinates: [52, -60, 28], Figure
2.5). In the remaining participants (n=15), no significant voxels were found.
fMRI Results: Pattern Analysis for ASD
Within ROI Pattern Analysis for Experiment 4 (ASD)
Harm vs. Neutral: As in NT controls, pattern analyses revealed a separation in the
pattern of response for harmful versus neutral acts in ASD. Using individual ROIs,
significant discrimination was found in RTPJ, LTPJ, and PC (RTPJ: within=1.4±0.13,
across=1.2±0.16, t(15)=4.1, p=0.0005; LTPJ: within=1.5±0.10, across=1.3±0.16,
t(14)=2.1, p=0.027; PC: within=1.2±0.14, across=1.1±0.14, t(14)=1.9, p=0.04, Figure
2.3). The effect was non-significant in DMPFC (within=1.2±0.11, across=1.2±0.12,
t(6)=0.21, p=0.42), possibly because an ROI for DMPFC was defined in only 7/16
participants.
Accidental vs. Intentional: Accidental vs. Intentional: Accidental and intentional harms
did not elicit distinct patterns in any theory of mind ROI in ASD: RTPJ: within=1.3±0.11,
across=1.3±0.12, t(15)=1.1, p=0.86; LTPJ: within=1.4±0.12, across=1.4±0.14, t(14)=1.1,
p=0.86; PC: within=1.2±0.14, across=1.2±0.15, t(14)=1, p=0.16; DMPFC:
within=1±0.15, across=1.1±0.08±, t(6)=0.59, p=0.71, Figure 2.3).
Behavioral and Neural Correlation (ASD)
There was a marginal negative correlation between participants’ moral judgments of
accidental vs intentional harms and pattern discrimination in RTPJ (r(9)=-.54, p=.09).
Note that this effect is in the reverse direction of the correlation observed in NT adults.
There was no correlation with behavior in any other region (all r<0.2, p>0.5). There was
also no correlation with symptom severity (ADOS score) in any region.
Whole Brain Searchlight for Experiment 4 (ASD)
No regions discriminated between accidental and intentional harms (p<0.001, voxelwise, uncorrected).
ASD symptom severity in Experiment 4 (ASD)
Some prior work has found correlations between symptom severity and neural
information, either in mean signal (e.g., Wang, 2006) or in pattern discriminability (e.g.,
Coutanche, Thompson-Schill, & Schultz, 2011). However, in this data set, we found no
significant correlation between neural pattern and ADOS symptom severity scores(Lord
et al., 2000)in any region (RTPJ: r(16)=0.33, p=0.22; LTPJ: r(15)=0.05, p=0.85; PC:
r(15)=.14, p=.61; DMPFC: r(7)=0.04, p=0.93).
Group Comparison: NT vs ASD
Within ROI Pattern Analysis
61 of 179
Moral vs. Neutral: A Group (ASD, NT) x Pattern (within, across) ANOVA revealed that
NT and ASD participants show strong, and equally robust, neural discrimination in
response to moral violations versus neutral actions in their RTPJ, LTPJ, and PC, with a
main effect of Pattern (within > across), no effect of Group, and no interaction in all
three ROIs (RTPJ: Pattern (F(1,36)=25.9, p<0.0001), Group (F(1,36)=0.5, p=0.5),
interaction (F(1,36)=0.9, p=0.3); LTPJ: Pattern (F(1,35)=13.3, p = 0.0008), Group
(F(1,35) = 1.2, p = 0.3), interaction F(1,35)=0.3, p=0.6; PC: Pattern, (F(1,35)=15.1,
p=0.0004), Group (F(1,35)=1.7, p=0.2), interaction (F(1, 35) =1.18, p=0.3)). There were
no significant effects in DMPFC (Pattern (F(1,23)=2.7, p=0.1), Group, (F(1,23)=0.14,
p=0.7), interaction (F(1,23) = 0.7, p =0.4)).
Accidental vs. Intentional: In the RTPJ, accidental and intentional harms were more
discriminable in NT than ASD participants, reflected in a significant Group x Pattern
interaction (F(2,36)=4.9, p=.03), with no main effect of Pattern (F(1,36)=1.7, p=0.2) or
Group (F(1,36)=.71, p=0.4). The same interaction was observed in one-to-one matched
subsets of participants (F(1,28)=5.9, p=.02, see Appendix). There was no Group x
Pattern interaction in any other region (all F<0.9, all p>0.3).
Behavioral and Neural Correlation
The correlation of pattern discrimination in RTPJ with behavioral responses was
significantly larger in NT than in ASD participants (full sample, difference of correlations,
z=3.0, p=0.003; matched subsets z=2.6, p=0.009).
Decoding true vs. false beliefs?
One open question about the current results is whether the stimulus feature that we
successfully decoded is actually accidental vs. intentional harm, or another feature of
the beliefs in the scenarios which is confounded with this distinction. Two features are
confounded with accidental versus intentional: (i) false beliefs / ignorance versus true
beliefs / knowledge, and (ii) good / neutral versus bad intentions. In all of our stimuli,
accidental harms were actions based on a false belief or ignorance about the outcome
of the act and a neutral or good intention, while intentional harms were actions based on
a true belief or foreknowledge about the outcome of the act and a negative intention.
To test which of these features is represented by the patterns of neural response, we
analyzed additional data from Experiments 2 and 3. In these experiments, in addition to
accidental and intentional harms, participants read about failed attempts to harm (i.e.
neutral outcome, false belief, negative intention) and neutral actions (i.e. neutral
outcome, true belief, neutral intention). Like the comparison between accidental and
intentional harms, this contrast thus holds the outcome constant, and manipulates
whether the belief is true or false and whether the intention is good/neutral or bad.
We found that even when combining data from Experiments 2 and 3 (for maximum
power, n=30), the patterns in response to failed attempts and neutral actions were not
different in RTPJ (within=0.74(0.10), across=0.78(0.08), t(29)=-0.8, p=0.8). These
results suggest that the pattern distinction observed in RTPJ for accidental vs.
intentional harms, in the same participants, is not just due to the difference between true
and false beliefs, or whether the intention is good/neutral or bad.
62 of 179
Discussion
A central aim of this study was to ask whether the difference between accidental and
intentional harms could be decoded from the pattern of neural response within theory of
mind brain regions. Across three experiments, using different stimuli, paradigms, and
participants, we found converging results: in neurotypical (NT) adults, stories about
intentional versus accidental harms elicited spatially distinct patterns of response within
the right temporo-parietal junction (RTPJ). Moreover, this neural response mirrored
behavioral judgments: individuals who showed more distinct patterns in RTPJ also
made a larger distinction between intentional and accidental harms in their moral
judgments. Notably, in the fourth experiment, this pattern was absent in adults with
ASD: we found no neural difference between accidental and intentional harms in pattern
or mean signal, and behavioral judgments showed a marginal negative correlation with
neural pattern.
MVPA discriminates accidental and intentional harms in RTPJ of
neurotypical adults
The current results suggest that the RTPJ contains an explicit representation that
distinguishes intentional from accidental harms. This representation was apparent in
reliable but distinct spatial patterns of activity. The current results are among the first to
extend this method to high-level cognition and abstract stimulus features (see also
Fedorenko et al., 2012; Hampton & O'Doherty, 2007; Peelen et al., 2006; Tusche, Bode,
& Haynes, 2010). In particular, this is the first evidence of a feature of mental state
reasoning explicitly represented in a ToM brain region.
The convergence across experiments provides strong evidence that intentional and
accidental harms can be discriminated, using MVPA, in RTPJ. Designed to test a series
of separate questions, the three experiments differed in story content, voice of the
narrative (2nd or 3rd person), tense (past or present), the severity of harm caused (mild
in Exp.1, extreme in Exp.2, unspecified in Exp.3), the order and timing of the story
segments, the number of stories per condition, and the participants’ explicit task.
Perhaps most importantly, the cues to intent were different across experiments. In
Experiment 1, the same mental state content (e.g., your cousin’s allergy to peanuts)
was described as known or unknown (e.g., “you had no idea” vs. “you definitely knew”).
By contrast, in Exp.2 and 3, sentences with the same syntax and mental state verbs
were used to describe beliefs with different content (e.g. “Steve believes the ground
beef is safe / rotten”). Nevertheless, the spatial pattern of response was reliable and
distinct for intentional versus accidental harms specifically in the RTPJ, across all three
experiments. The converging results across experiments suggests that, rather than
being driven by superficial stimulus features or task demands, the distinct neural
patterns reflect an underlying distinction in the representation of accidental and
intentional harms.
The pattern difference found in the current work suggests that the distinction between
intentional and accidental harm is encoded in the neural representation in RTPJ.
Interestingly, the pattern in RTPJ did not distinguish between true and false beliefs, or
63 of 179
between negative and neutral intentions, in the context of non-harmful acts (i.e. the
difference between neutral acts and failed attempts to harm), suggesting that the
representations underlying the difference between accidental and intentional harm are
relatively specific.
This evidence that one feature of mental states is explicitly represented in the neural
pattern opens the door to many future studies. Many other features may also be
decodable. For example, patterns of response in theory of mind brain regions may
discriminate between attributing beliefs that are justified or unjustified (Young, Nichols, &
Saxe, 2010c), plausible or crazy (Young, Dodell-Feder, & Saxe, 2010b), attributed to
friends or enemies (Bruneau & Saxe, 2012), “constrained” or “open-ended” (Jenkins &
Mitchell, 2010), or first-order or higher-order (Koster-Hale & Saxe, 2011). Important
challenges for future research will include (i) determining the full set of mental state
features that can be decoded from the RTPJ response using MVPA, and (ii) identifying
the features of social stimuli that can be decoded from other regions within and beyond
the theory of mind network.
Pattern discrimination of intentional versus accidental harm is
correlated with moral judgment
The distinctness of the spatial patterns in the RTPJ was correlated with individuals’
moral judgments. Neurotypical adults differed in the amount of blame they assign to
accidental harms: some weighed intent more strongly (i.e., forgive more based on
innocent intentions), while others weighed outcome more strongly (i.e., blamed more
based on bad outcomes, Nichols & Ulatowski, 2007; Young & Saxe, 2009a). These
differences in moral judgment were predicted by individual differences in neural pattern
discriminability in the RTPJ, and more weakly in the LTPJ. Though Experiment 1 used a
blameworthiness scale (“How much blame should you get?”), and Experiment 2 used a
permissibility scale (“How permissible was Steve’s action?”), the same result emerged
in both studies: the individuals who encoded the difference between accidental and
intentional harms most strongly in their RTPJ also showed the greatest difference in
their moral judgments of these acts. These findings were corroborated by the wholebrain searchlight results: in the participants whose moral judgments were most sensitive
to the difference between intentional and accidental harm, only a region within RTPJ
discriminated between intentional and accidental harms.
Note that mental states (e.g., beliefs, intentions) represent just one of many inputs to
moral judgment. Decisions about moral blame and permissibility depend on many
features of the event, including the agent’s beliefs and desires (Cushman, 2008), the
severity of the harm (Greene, Sommerville, Nystrom, Darley, & Cohen, 2001), the
agent’s prior record (e.g., Woolfolk, Doris, & Darley, 2006), the means of the harm
(Cushman, Young, & Hauser, 2006; Greene, Nystrom, Engell, Darley, & Cohen, 2004),
the external constraints on the agent (e.g., coercion, self-defense, Woolfolk et al., 2006;
Young & Phillips, 2011), and more (e.g., Young et al., 2010c). Thus, activity in the RTPJ
reflects the representation of one input to moral judgment, rather than the judgment
itself.
64 of 179
In addition to finding distinct neural patterns in RTPJ, we also found a difference in the
magnitude of response when participants were reading about intentions and making
moral judgments: participants showed a higher level of RTPJ activity for accidental
harms relative to intentional harms. The current results converge with, and extend, prior
reports that RTPJ is recruited in the face of mitigating mental state information such as
innocent intent (Buckholtz et al., 2008; Yamada et al., 2012; e.g., Young & Saxe,
2009b). We find the effect in response magnitude is independent of the difference in
neural pattern, suggesting that these effects are sensitive to different aspects of RTPJ
function. A higher magnitude of response to accidental harms suggests that NT adults
use mental state reasoning in their moral judgments more when faced with relevant
information about the mind of the perpetrator. The stable difference in neural pattern
suggests that additionally the distinction between intentional and accidental harm is
encoded in the neural representation in RTPJ.
High functioning adults with ASD show atypical neural activity in
RTPJ
In ASD adults, we found distinct patterns for harmful and neutral acts in all theory of
mind regions, but no distinction between intentional and accidental harms in the RTPJ
or any other region. Similarly, mean signal did not differentiate between accidental and
intentional harms in any region. Finally, unlike in NT adults, behavioral responses
showed a marginal negative correlation with neural discrimination in the RTPJ,
The results of the current study match the behavioral profile of high-functioning
individuals with ASD. Individuals with ASD do not have impairments in moral judgment
as a whole: children with ASD make typical distinctions between moral and conventional
transgressions (Blair, 1996), and between good and bad actions (Leslie, Mallon, &
DiCorcia, 2006). However, they are delayed in using information about innocent
intentions to forgive accidents (Grant, 2005). Furthermore, even very high-functioning
adults with ASD, who pass traditional tests of understanding (false) beliefs, neglect
beliefs and intentions in their moral judgments compared to NT adults (Moran et al.,
2011; Yang et al., 2013); although in the current sample, this effect was observed more
strongly in z-scored behavioral data (Appendix and Baez et al., 2012).
Two prior papers have used MVPA to study neural mechanisms of social cognition
(though not specifically theory of mind) in ASD. Gilbert et al. (2008) measured the
magnitude and pattern of activity in MPFC while participants performed a task that did
or did not elicit thinking about another person. The magnitude of response during the
two tasks was equivalent in participants with ASD and controls, but participants with
ASD showed a less reliable and distinct pattern of activation in the MPFC during the
‘person’ task. Coutanche and colleagues (2011) measured the pattern and magnitude of
response in ventral visual areas while viewing faces versus houses. Again, the
magnitude of response in these regions was not different in typical adults and those with
ASD, but individuals with ASD showed less discriminable patterns in response to faces,
a social category, compared to houses. Disorganization of the pattern of activity for
faces versus houses was correlated with ASD symptom severity.
65 of 179
Broadly consistent with this prior work, the current results suggest that ASD affects the
organization (i.e., pattern) of information in theory of mind brain regions. One concern
might be that the reduced pattern information is merely a result of more noisy or
heterogeneous neural responses in ASD (e.g., Dinstein et al., 2010). Our results do not
favor this interpretation: rather than lower within-condition spatial correlations (i.e.,
noisier or more idiosyncratic responses), we found numerically higher within-condition
correlations in participants with ASD. That is, participants with ASD seemed to show a
reliable pattern of response in the RTPJ in response to all harmful acts, regardless of
whether the act was intentional or accidental.
We also found a difference between groups in the magnitude of response. NT adults
showed a higher response to accidental than intentional harms in the RTPJ, suggesting
increased activity the face of mitigating mental state information. In contrast, ASD adults
showed equal activation to both types of stories, suggesting that accidental harms did
not elicit more consideration of mental states than intentional harms. Given the
independence of the observed differences in magnitude and pattern, our results provide
converging evidence from two different types of analyses of atypical representations of
intentional versus accidental harms in the RTPJ of adults with ASD.
Prior studies investigating the magnitude of response in theory of mind regions of
individuals with ASD have found inconsistent results. Some studies suggest that theory
of mind regions are hypo-active (i.e., produce a smaller or less selective response,
(Kennedy & Courchesne, 2008; Lombardo et al., 2011), while other studies find no
difference between ASD and NT individuals (Gilbert et al., 2008; Nieminen-von Wendt et
al., 2003), and still others find the opposite pattern, hyper-activation, in ASD (Mason,
Williams, Kana, Minshew, & Just, 2008; Tesink et al., 2009; Wang, 2006). Studies using
tasks that elicit spontaneous or implicit social processing may be more likely to find
hypo-activation (Adolphs, 2001; Groen et al., 2010; Wang, 2006). Because increases in
magnitude may reflect either successful representation of a stimulus, or effortful but
ineffectual processing, differences in the magnitude of response between groups can be
difficult to interpret.
MVPA may therefore offer a sensitive tool for measuring neural differences in ASD that
are related to social impairments. Note, though, that due to the demands of the task and
scanning environment, the present ASD participants (as in previous task-oriented
neuroimaging studies e.g., Brieber et al., 2010; Koshino et al., 2007), and (see Philip et
al., 2010 for a review) are very high-functioning, which may limit the generalizability of
the results to lower-functioning individuals. Nevertheless, the individuals in the current
study do experience disproportionate difficulties with social interaction and
communication; therefore, the current results provide a window into the neural
mechanism underlying these difficulties.
Conclusion
In summary, using multivoxel pattern analyses across four experiments, we found that
(i) the difference between accidental and intentional harms is linearly decodable from
stable and distinct spatial patterns of neural activity in RTPJ; (ii) individual differences in
neural discrimination in RTPJ predict individual differences in moral judgment; and (iii)
66 of 179
these neural patterns are not detectable in high functioning adults with ASD.
Considerable neuroimaging work on theory of mind suggests that the RTPJ plays some
role in thinking about others’ thoughts; the current evidence suggests that one aspect of
this role is to make explicit, in the population response of its neurons, features of beliefs
that are most relevant for inference and decision-making — for example, encoding the
intent behind a harmful act.
Acknowledgments
This material is based upon work supported by the National Institute of Health under
grant 1R01MH096914-01A1, the Simons Foundation, the National Science Foundation
under grant 095518, a John Merck Scholars Grant, the Dana Foundation, and a
National Science Foundation Graduate Research Fellowship, grant 0645960. We thank
our participants, Lee Marvos and Caitlin Malloy for recruiting help, Nancy Kanwisher, Ev
Fedorenko, Hilary Richardson, Nick Dufour, members of the MIT Saxelab, the BC
Morality Lab, and the Brown Moral Psychology Research lab for useful comments and
discussion. Data were collected at the Athinoula A. Martinos Imaging Center at the
McGovern Institute for Brain Research, MIT.
Appendix
SI Behavioral Results: Results from Raw Behavioral Responses
In the main text, we report behavioral data and statistics that have been z-scored to
account for variability in the participants’ use of the rating scales. Here we provide the
same analyses, using the untransformed button press responses. We find largely
converging results using untransformed behavioral data.
Experiment 1: From no blame at all (1) to very much blame (4), participants judged
intentional harms (3.4 ± 0.09) to be more blameworthy than accidental harms (1.5 ±
0.08; t(19)=19, p<0.0001), both of which were judged to be worse than neutral stories
(1.1 ± .03, t(19)=30, p<0.0001; t(19)=6.9, p<0.0001).
Experiment 2: Replicating the results in Exp.1, on a scale of Forbidden (1) to
Permissible (3), participants judged intentional harms (1.1 ± .04) as less permissible
than accidental harms (2.2 ± .11; t(9)=11, p<0.001).
Experiment 3: Participants in Exp.3 did not make moral judgments of the scenarios, but
instead answered true/false questions about the content of the fact.
Experiment 4: When making moral judgments, ASD participants, like NT participants
from Exp.1, judged intentional harms (3.4 ± 0.18) to be worse than accidental harms
(1.9±0.17; t(10)=7.5; p<0.0001), both of which were judged to be worse than neutral
stories (1.2±.06; t(10)=10.7, p=8.3e-07; t(10)=4.3, p=0.0016).
Group comparison: A mixed effects ANOVA crossing Group (NT in Exp. 1, ASD in Exp.
4) by Condition (accidental, intentional raw behavioral ratings) yielded a main effect of
67 of 179
Condition (F(1,29)=381.1, p<0.0001), with no effect of Group F(1,29)=2.7, p=0.1, and
no interaction (F(1,29)=1.1, p=0.3). The planned comparison t-test revealed that ASD
adults assigned more blame for accidental harms than NT adults (t(14.2)=1.7, p=0.05,
one-tailed; see Moran et al., 2011).
Behavioral and Neural Correlation: In Exp.1 and 2, NT participants provided moral
judgments of each scenario in the scanner, allowing us to determine whether behavioral
responses were related to the spatial pattern of the neural response in RTPJ or any
other region. For each participant, we calculated the difference in moral judgments for
intentional versus accidental harms. We tested whether this difference score was
correlated, across participants, with the index of classification in each region (intentional
vs. accidental, within-condition correlation minus across-condition correlation). In both
experiments, we found that only in the RTPJ was the difference between intentional and
accidental harms in individuals’ moral judgments correlated with the neural classification
index (Exp. 1: r(17)=0.55; p=0.016; Exp. 2: r(8)=0.71; p=0.02). The correlation between
neural pattern and behavior remained significant after combining the data from both
experiments, using the moral judgments converted to the same scale 7 (r(27)=0.56,
p=0.002). No other regions showed a significant correlation with behavioral judgments,
through there was a strong trend in LTPJ (r(27)=0.35, p=0.06).
Supplementary Results: Pairwise matched subsets of ASD and NT
In the main manuscript, we mostly focus on the comparison of the full sample of NT
(n=23) and ASD (n=16) participants. Here we report the results of the key analyses
when conducted in a subset of each group that were matched in pairs. These analyses
find the same key pattern of results; importantly, the group by condition interaction in the
neutral pattern of RTPJ remains significant.
Participants: Of the 16 ASD participants, we were able to create one-to-one matches
on gender, age, and IQ for 15 participants, resulting in two groups (NT and ASD), both
n=15, 13 male and 2 female. These pairs are matched in age (NT (mean ± SD)=30
years ±10; ASD=30 years ± 8; maximum difference between pairs 7 years, average
absolute difference 3.4 years, t(26.5)=0.26, p=0.8) and IQ (NT mean=121 ± 11.8;
ASD=120 ± 12.7, maximum difference 12 points, average absolute difference 4.2
points, t(27.1)=0.32, p=0.7). As in the full sample, ASD participants scored significantly
higher than NT participants on the AQ (NT (mean ± SD)=18.5 years ± 6.6; ASD=32.5 ±
7.2; t(20.2)=4.8, p < 0.0001).
Behavioral Results: Behavioral data were available for 11 ASD and 13 NT participants
from the matched subsets (NT: accidental -0.39 ± 0.05, intentional 1.17 ± 0.03; ASD:
accidental -0.23 ± 0.07, intentional 1.01 ± 0.09). In these subjects, a mixed effects
ANOVA crossing Group (NT, ASD) by Condition (accidental, intentional z-scored ratings)
yielded a main effect of Condition (F(1,22)=304, p<0.0001), and a marginal Group by
7
Raw Data: Behavioral data (key presses) from Experiments 1 and 2 were translated to the same scale
for comparison. Specifically, responses from Experiment 2 were inverted to make “forbidden / very much
blame” the top end of the scale for all experiments by subtracting each response from 4 (e.g., a rating of 1
becomes a rating of 3), and then scaling linearly from a 3-point scale to a 4-point scale.
68 of 179
Condition interaction (F(1,22)=3.93, p=0.059). A planned comparison (see Moran et al.,
2011) t-test revealed that, in the matched subsets, ASD adults showed the same trend
of assigning more blame for accidental harms than NT adults (t(22)=1.6, p=0.06, onetailed).
Motion and Artifact: The subsets of NT and ASD participants did not differ in the
number of motion artifacts per run (NT: 4.3 ± 1.2; ASD: 3.6 ± 1.3, t(27.9)=0.35, p=0.7),
in the number of global signal outliers per run (NT: 2.5 ± 0.4; ASD: 2.7 ± 0.3,
t(26.1)=0.50, p=0.6), or in the total vector translation (NT: 0.25 ± 0.02), ASD: 0.23 ±
0.02, t(28)=0.73, p=0.47.
Voxel-Wise Pattern Results: Harm vs. Neutral: As in the full sample, the neural pattern
in the RTPJ of both NT and ASD participants discriminated harms from neutral actions
(NT: within=1.3±0.11, across=1.2±0.13, t(14)=2.0, p=0.03; ASD: within=1.5±0.14,
across=1.2±0.16, t(14)=3.7, p=0.001). A Group (ASD, NT) x Pattern (within, across)
ANOVA revealed that matched NT and ASD participants show equally robust neural
discrimination in response to harmful versus neutral actions in their RTPJ, with a main
effect of Pattern (F(1,28)=16.7, p=0.0003), no effect of Group (F(1,28)=0.3, p=0.5), and
no interaction (F(1,28)=1.7, p=0.2).
Accidental vs. Intentional: As in the full sample, the neural pattern in the RTPJ of NT
participants discriminated accidental from intentional harms (within=1.1±0.14,
across=0.99±0.15, t(14)=2.2, p=0.02), while matched ASD participants did not
(within=1.3±0.11, across=1.3±0.13, t(14)=1.1, p=0.8). Matched NT participants
discriminated between accidental and intentional harms to a greater extent than ASD
participants, reflected in a significant Group x Pattern interaction (F(1,28)=5.9, p=.02),
with no main effect of Group (F(1,28)=1.93, p=0.18) or Pattern (F(1,28)=1.53, p=0.23).
Behavioral and Neural Correlation: The correlation of pattern discrimination in RTPJ
with behavioral responses was significantly larger in NT (r(11)=0.56; p=0.047) than in
ASD participants (r(9)=0.54, p=0.08; z=2.6, p=0.009). Note that the marginal effect in
ASD is in the reverse direction: stronger neural discrimination in individuals with more
similar behavioral moral judgments.
Independent effects of magnitude and pattern
In neurotypical adults, we observed differences between accidental and intentional
harms in the pattern of response over the whole trial and in the magnitude response in
the final phase. Specifically in the RTPJ, both of these effects contrasted with the
pattern observed in ASD adults, with significant Condition by Group interactions. When
interpreting these results it is important to know whether these two different ways of
measuring the same effect or converging evidence of two different aspects of atypical
RTPJ function in ASD. To address this question, we asked whether these two effects
were correlated, across individuals. We found these two effects were not correlated
across either NT (r(20)=0.11, p=0.62) or ASD (r(14)=-0.16, p=0.55, suggesting that the
effects are sensitive to different aspects of RTPJ function. Moreover, our analysis
techniques are sensitive to different features of the data: because correlations are not
sensitive to overall magnitude, the pattern analyses used here are not driven simply by
differences in the magnitude (see discussion in the next section).
69 of 179
A note on separating the phases of Experiment 1
Over the full duration of the trial, in the RTPJ we found no difference in the magnitude of
response to intentional versus accidental harms (accidental: 0.26±0.05; intentional:
0.22±0.05; accidental vs. intentional: t(21)=1.4, p=0.18). By contrast, we could reliably
decode the difference between these two conditions from the spatial patterns of the
betas (within=1.2±0.12, across=1.1±0.12, t(21)=2.2, p=0.02). When we tried to
separately estimate the response to only the second phase of the trial (Intent, Decision),
we found a higher magnitude of response in RTPJ for accidental than intentional harms
(accidental: 0.27±0.06; intentional: 0.22±0.07; accidental vs. intentional: t(21)=2.95,
p=0.008). However, we could not reliably the decode the difference between these two
conditions from the spatial patterns of the betas (within=0.98±0.12, across=0.92±0.11,
t(21)=1.3, p=0.11).
These results illustrate, first, that the current techniques for estimating the magnitude
and pattern of neural responses are independent. Because correlations depend only on
the spatial pattern of the response (i.e. which voxels show relatively higher betas, and
which show relatively lower betas), it is possible to find a pattern difference in the
absence of a magnitude difference, and it is also possible to find a magnitude difference
in the absence of a pattern difference. This is a difference between the current MVPA
technique (focusing exclusively on spatial patterns based on Haxby, 2001) and other
machine learning algorithms used commonly in the literature (e.g support vector
machines).
Second, these results illustrate the different sensitivity of analyses based on average
percent signal change, versus estimated betas of modeled hemodynamic response
functions. Because there was no jitter in the timing (i.e. the onset of the Intent segment
was always perfectly predictable from the onset of the story) separate regressors for the
first and second phases of each trial are partially co-linear. As a result, the withincondition correlations of betas (a measure of our ability to estimate reliable betas) were
notably lower for the same subjects in the last-segment-only (e.g. RTPJ within=0.98
±0.12), compared to the full trial (e.g. RTPJ within=1.2±0.12). Note that these
correlations reflect the reliability of the spatial pattern across trials, not the magnitude of
the betas. Because the discriminating information all occurs in the second phase of the
trial, successful decoding from the whole trial but not the second phase alone must
reflect the unreliable estimates of second-phase betas.
70 of 179
Chapter 3.
Thinking about Seeing: perceptual sources of
knowledge are encoded in the theory of mind brain
regions of sighted and blind adults 8!
Blind people’s inferences about how other people see provide a window into
fundamental questions about the human capacity to think about one another’s
thoughts. By working with blind individuals, we can ask both what kinds of
representations people form about others’ minds, and how much these
representations depend on the observer having had similar mental states
themselves. Thinking about others’ mental states depends on a specific group of
brain regions, including the right temporo-parietal junction (RTPJ). We investigated
the representations of others’ mental states in these brain regions, using multivoxel
pattern analyses (MVPA). We found that, first, in the RTPJ of sighted adults, the
pattern of neural response distinguished the source of the mental state (did the
protagonist see or hear something?) but not the valence (did the protagonist feel
good or bad?). Second, these neural representations were preserved in congenitally
blind adults. These results suggest that the temporo-parietal junction contains
explicit, abstract representations of features of others’ mental states, including the
perceptual source. The persistence of these representations in congenitally blind
adults, who have no first-person experience with sight, provides evidence that these
representations emerge even in the absence of relevant first-person perceptual
experiences.
8
A version of this chapter was published as Koster-Hale, J., Bedny, M., & Saxe, R. R. (2014). Thinking
about seeing: Perceptual sources of knowledge are encoded in the theory of mind brain regions of
sighted and blind adults, Cognition 133(1), 65–78.
71 of 179
Introduction
Imagine a friend tells you that last night, looking out the window onto a dark, rainy
street, she saw her boyfriend get into a car with a strange woman, and drive away. Your
reaction will depend on many inferences about her thoughts and feelings. You will
recognize that she believes her boyfriend is being unfaithful, and feels betrayed. You
might also note the source of her belief, and question how clearly she could see at a
distance and in the dark. Perhaps she was mistaken about the identity of the man
getting into the car, or about the driver; maybe the driver was actually her boyfriend’s
sister. Knowing how she got her information might strongly affect how you reason about
her beliefs and experiences — do you yourself believe that her boyfriend is being
unfaithful? How strongly do you think she believes it? What is she likely to do next?
Now imagine that you are congenitally blind. Would your inferences be any different?
Clearly, a blind adult would understand the emotional toll of discovering a lover’s
possible betrayal, but could a blind person make the same inferences about the visual
source of the discovery? How much would a blind person understand about the
experience of seeing a familiar person and a strange woman, from afar, in the dark?
Blind people’s inferences about how other people see provide a window into a
fundamental question about the human capacity to think about one another’s thoughts:
what are the mechanisms used to think about someone else’s mind? One possibility is
that we think about someone’s experience by invoking our own relevant experiences
and sensations. In this view, thinking about someone else’s experience of seeing
requires (among other things) a first-person experience of sight. In contrast, if people
use an intuitive “theory” of mind, composed of relationships among abstract mental
state concepts, to reason about others’ experiences, then experience of sight is not
always necessary for reasoning about seeing. In many cases, these views are hard to
disentangle; however, they predict different outcomes in blindness. If first-person
experience is necessary to understand others’ experiences, blind people should have
only a fragmentary, limited, or metaphorical understanding of seeing. By asking how
blind individuals represent other’s experiences of sight, we can place important limits on
our theories of mental state inference: to what extent does theory of mind depend on
the observer having had similar sensations, experiences, beliefs, and feelings as the
target?
The first possibility is that people understand another’s mind by trying to replicate it, by
imagining themselves in a similar situation, or by re-experiencing a similar past event of
their own lives. You would understand your friend’s feelings of sadness by imagining
your own feelings in response to a lover’s betrayal, recreating in your emotional system
a version of your friend’s experience. Similarly, you would understand your friend’s
experience of seeing her boyfriend get into the car by recreating the visual scene in
your own mind’s eye (Gordon, 1989; Stich & Nichols, 1992). Understanding others’
minds thus depends on the observer having experienced a relevantly similar mental
state to the target (Gallese & Goldman, 1998; Goldman, 1989; 2006; Gordon, 1989;
Harris, 1992; Nichols, Stich, Leslie, & Klein, 1996; Stich & Nichols, 1992). Simulationbased accounts do not necessarily posit that people can only think about exactly those
72 of 179
experiences that they themselves have had; rather, mental state representation could
be a composition of one’s existing relevantly similar first-person experiences, composed
flexibly to simulate a novel experience. Still, because simulation depends on similar
experiences, the extent to which we can simulate the minds of others depends on “the
interpersonal sharing of the same kind of neural and cognitive resources. When this
sharing is limited (or even missing), people are not fully able (or are not able at all) to
map the mental states or processes of others because they do not have suitable mental
states or processes to reuse” (Gallese & Sinigaglia, 2011). Because congenitally blind
people lack the mental states and processes involved in seeing, their representations of
sight are predicted to be limited or unreliable.
Neuroimaging experiments provide evidence that first-person sensorimotor
representations are “reused” during observation of others’ actions and sensations.
Similar brain regions are recruited when experiencing physical pain compared to
observing another person experience similar pain (e.g., Botvinick et al., 2005;
Immordino-Yang, McColl, Damasio, & Damasio, 2009; Singer et al., 2004); and when
experiencing a tactile sensation compared to observing another person being touched
in the same way (Blakemore, 2005; Keysers:2004dj Gazzola & Keysers, 2009). More
importantly, neural activity during some observation tasks depends on the observer’s
own specific first-person experiences. For example, motor activation in dancers during
observation of dance moves is enhanced for specific movements that the observers
themselves have frequently executed (Calvo-Merino, Glaser, Grèzes, Passingham, &
Haggard, 2005; Calvo-Merino, Grèzes, Glaser, Passingham, & Haggard, 2006). If this
type of reuse extends to mental state representation, typical representations of other’s
experiences of seeing should depend on, or be profoundly affected by, having first seen
yourself.
However, many authors (Gopnik & Meltzoff, 1997; Gopnik & Wellman, 1992; Perner,
1993; Saxe, 2005) have suggested an alternative mechanism for understanding other
minds: namely, that people have an intuitive theory of other minds. An intuitive theory
includes causal relations among abstract concepts (like beliefs and desires), and can be
learned from many sources of evidence, not limited to first-person experiences. One
source of evidence children could use to build a theory of mind is the testimony of
others: verbal labels and descriptions of mental states and experiences, often including
mental state verbs like think and see (Harris, 2002b; 1992). Thus a congenitally blind
child, growing up in a world full of sighted people, might develop an intuitive theory that
includes concepts of vision, to explain everyone else’s behavior (e.g. reacting to objects
at a distance) and testimony (e.g. saying “I see your toy on the top shelf!”). This intuitive
theory would then allow a blind child to predict how a sighted person would act in a
given environment, and what that person would be likely to infer based on what she
could see.
To test these theories, we investigated how blind people think about sight. Observation
and behavioral studies suggest that even young blind children know that other people
can see with their eyes, and can understand basic principles of vision: e.g. that objects
can be seen from a distance and are invisible in the dark (Bigelow, 1992; Landau &
Gleitman, 1985; Peterson, & Webb, 2000). By adulthood, congenitally blind people
know the meanings of verbs of sight, including fine-grained distinctions, such as the
73 of 179
difference between verbs like peer, gaze, and gawk (Landau & Gleitman, 1985; Lenci,
Baroni, Cazzolli, & Marotta, 2013, Koster-Hale et al., in prep). Blind adults are thus
sensitive to subtle distinctions in how sighted people gather information visually.
This behavioral evidence alone, however, cannot answer the question of whether a
blind person uses the same cognitive mechanisms as a sighted person to understand
sight. Any surface similarity between a blind and sighted person’s verbal descriptions of
sight could be the product of compensatory mechanisms in the blind. For example,
some authors have suggested that blind people mimic the words used by sighted
people, without being able to fully access their meaning (so-called “verbalisms,” Rosel,
Caballer, Jara, & Oliver, 2005) or integrate them into their conceptual understanding.
Thus, a blind person who hears the sentence “I saw my boyfriend getting into the car
from the window,” may have only a limited or metaphorical understanding of the
experience it describes.
This methodological challenge affords an opportunity for cognitive neuroscience:
functional neuroimaging can provide an online, unobtrusive measure of ongoing
psychological processes and thus offers an alternative strategy to ask if two groups of
people are performing a cognitive task using similar or different mechanisms. Here we
use neuroimaging to ask whether blind and sighted people rely on similar cognitive
mechanisms when they reason about seeing.
Previous studies have shown that thinking about someone else's thoughts (including
those based on visual experiences, Bedny et al., 2009) increases metabolic activity in a
specific group of brain regions often called the 'mentalizing' or theory of mind network.
These regions include the medial prefrontal cortex (MPFC), the precuneus (PC), and
the bilateral temporal-parietal junction (TPJ), (e.g., Aichhorn et al., 2009; Bedny,
Pascual-Leone, & Saxe, 2009; Lombardo, Chakrabarti, Bullmore, Baron-Cohen, &
Consortium, 2011; Mason & Just, 2011; Rabin, Gilboa, Stuss, Mar, & Rosenbaum, 2010;
Saxe & Kanwisher, 2003; Saxe & Powell, 2006; Spunt, Satpute, & Lieberman, 2010).
However, very little is known about which aspects of mental states are represented in
these brain regions. Mental experiences have many features, including the content
(“boyfriend in stranger’s car”), the valence (very sad), and the modality or source of the
experience (seen from afar). Neurons, or groups of neurons, within theory of mind brain
regions may represent any or all of these features. Thus, two important questions
remain open. First, in sighted people, do neurons in any theory of mind region
specifically represent the perceptual source of another person’s belief? Second, if so,
are similar representations present in blind people?
A powerful approach for understanding neural representation is to ask which features of
a stimulus are represented by distinct subpopulations of neurons within a region. For
example, within middle temporal visual, subpopulations of neurons that respond to
visual stimuli moving orientations are spatially organized at a fine spatial scale. Although
no individual fMRI voxel (which includes hundred of thousands of neurons) shows an
orientation selective response, and therefore overall early visual cortex shows equal
average magnitude of response to all orientations, it is possible to detect reliably distinct
spatial patterns of response across cortex that do distinguish between orientations
(Kamitani & Tong, 2006). This technique of looking for reliable spatial patterns of fMRI
74 of 179
response within a brain region is called multivoxel pattern analysis (MVPA; (Haynes &
Rees, 2006; Kriegeskorte & Bandettini, 2007; Norman, Polyn, Detre, & Haxby, 2006).
MVPA can reveal how stimulus categories are processed within a functional region
(Peelen et al, 2006; Haynes & Rees, 2006). MVPA has been successfully used to probe
the neural basis of many different types of representation, including subjectively
perceived directions of motion in ambiguous stimuli, semantic category, emotional
affect, and intent when causing harm (Chapter 2, Koster-Hale, Saxe, Dungan, & Young,
2013; Mahon & Caramazza, 2010; e.g., Norman et al., 2006; Peelen, Wiggett, &
Downing, 2006; Serences & Boynton, 2007).
Here, we ask if source modality (seeing vs hearing) is a relevant feature of the neural
representations of mental states. If one set of neurons responds more to stories about
seeing than hearing, while another (partially distinct) set responds more to stories about
hearing than seeing, we have evidence that seeing and hearing are being represented
in different ways in that brain region. Measuring the average magnitude of activity in
theory of mind brain regions cannot be used to address this question, because the
average magnitude may obscure distinct subpopulations of neurons within the regions.
We therefore used multivoxel pattern analysis (MVPA) to look for differences in the
neural response to stories about hearing and seeing.
In this study, we first asked whether stories about someone else’s hearing and seeing
experiences evoke different spatial patterns of response within theory of mind brain
regions in sighted individuals. Because very little is known about how theory of mind is
represented, cognitively or neurally, this is itself a fundamental question about mental
state representation. We then asked whether the same patterns are observed in
congenitally blind people. Finding these patterns would provide support for the notion
that blind individuals represent mental experiences of seeing in a qualitatively similar
manner to sighted individuals. For comparison, we also tested whether theory of mind
regions encode a feature of mental states that should not differ between sighted and
blind people: the valence (feeling good versus bad).
75 of 179
Methods
Participants
Thirteen sighted members of the larger MIT community participated (8 women; mean
age ± SD, 52 years ± 16), all with normal or corrected-to-normal vision. Ten blind
individuals participated (5 women; mean age ± SD, 50 years ± 7). Nine blind
participants were born blind and one lost sight between the ages of 2 and 3 years. Due
to possible effects of early vision, this person was excluded from all analyses. All blind
participants reported having at most faint light perception and no pattern discrimination.
None were able to perceive shapes, colors, or motion.
One blind participant was ambidextrous and one was left-handed; one sighted
participant was left-handed. All were native English speakers and gave written informed
consent in accordance with the requirements of the Institutional Review Board at MIT.
Participants were compensated $30/hour for their time.
Comparison to Bedny et al. (2009)
These data have previously been published, analyzing the magnitude but not the
pattern of response in each region, in Bedny et al. (2009). Bedny et al. (2009) found that
theory of mind regions showed equally high responses to stories about hearing and
seeing, in both sighted and blind participants. However, measuring the average
magnitude across the entire region may obscure distinct subpopulations of neurons
within a region. Specifically, the equally high magnitude of response to stories about
seeing and hearing observed by Bedny et al. (2009) is consistent with three
possibilities: (i) neurons in theory of mind brain regions do not distinguish between
hearing and seeing in blind or sighted people, (ii) distinct subpopulations of neurons
within theory of mind brain regions respond to stories about seeing versus hearing in
both sighted and blind participants, reflecting a common representation of the
perceptual source of other’s mental states, or (iii) distinct subpopulations of neurons
respond to seeing versus hearing in sighted but not blind individuals, reflecting different
representations in the two groups, depending on their first person experiences. The
current analyses allowed us to distinguish between these possibilities.
Note that the current results exclude one blind participant who was included in the
previous paper: this participant was born with cataracts and had some light and shape
perception during the first ten years of his life. To be conservative, we exclude his data
from the current analyses.
Finally, the previous paper included an analysis of reaction time to the behavioral
portion of the task (judging how good or bad the protagonist felt at the end of the story),
looking at the seeing events, hearing events, and additional control events, all collapsed
across valence. In this paper, we are also treating valence as a dimension of interest in
the neural data, and so break down the behavioral data by both modality and valence to
report reaction time and rating data.
76 of 179
fMRI Protocol and Task
Source task: Participants heard 32 stories (each 13 sec long): four stories in each of
eight conditions. Stories in the four conditions of interest described a protagonist’s
mental experiences, characterized by both a modality-specific source (something seen
vs heard) and a modality-independent valence (whether the protagonist felt good or
bad). The “seeing” stories described the protagonist coming to believe something as a
result of a visual experience, such as seeing a friend’s worried face or recognizing
someone’s handwriting. The “hearing” stories described the protagonist coming to
believe something as a result of an auditory experience, such as hearing a friend’s
worried voice or recognizing someone’s footsteps. The stories with negative valence
described an event that would make the protagonist feel bad, such as receiving criticism
or losing a game; the stories with positive valence described a good event, such as
receiving praise or winning a game (Figure 3.1). The remaining four control conditions,
which did not describe mental experiences, are not analyzed here (see Bedny et al.,
2009 for details).
Figure 3.1: Sample stimuli.
77 of 179
For each narrative context (e.g. a job interview, dinner with parents-in-law, cleaning a
dorm room), we constructed four endings, one in each condition. Thus the stories
formed a matched and counterbalanced 2x2 (seeing vs hearing, positive vs negative)
design. Individual participants saw each context only once; every context occurred in all
conditions, across participants. Word count was matched across conditions (mean
length ± SD, 32 words ± 4). Stories were presented in a pseudorandom order, condition
order was counterbalanced across runs and subjects, and no condition was immediately
repeated. Rest blocks of 11 sec were presented after each story. Eight stories were
presented in each 4 min 34 sec run. The total experiment, four runs, lasted 18 min 18
sec.
After each story, participants indicated whether the main character in the story felt very
bad, a little bad, a little good, or very good, using a button press (1-4). Reaction time
was measured from the onset of each question.
Theory of Mind Localizer task: Participants also completed a theory of mind and
language localizer task. Participants listened to 48 short verbal stories from two
conditions: 24 stories requiring inferences about mental state representations (e.g.,
thoughts, beliefs) and 24 stories requiring inferences about physical representations
(e.g., maps, signs, photographs). These conditions were similar in their metarepresentational and logical complexity but differ in whether the reader is building a
representation of someone else’s mental state (See Saxe & Kanwisher, 2003, DodellFeder et al. 2011, and Bedny et al., 2009 for further discussion). After each story,
participants answered a true/false question about the story. As a control condition,
participants listed to 24 blocks of “noise,” unintelligible backwards speech created by
playing the stories backwards. The task was performed in 6 runs with 12 items per run
(4 belief, 4 physical, and 4 backward-speech). Each run was 6 min and 12 sec long.
The stimuli for the localizer and both experiments were digitally recorded by a female
speaker at a sampling rate of 44,100 to produce 32-bit digital sound files.
Acquisition and Preprocessing
fMRI data were collected in a 3T Siemens scanner at the Athinoula A. Martinos Imaging
Center at the McGovern Institute for Brain Research at MIT, using a 12-channel head
coil. Using standard echoplanar imaging procedures, we acquired T1-weighted
structural images in 128 axial slices with 1.33 mm isotropic voxels (TR = 2 ms, TE =
3.39 ms), and functional blood oxygen level dependent (BOLD) data in 30 near-axial
slices using 3x3x4 mm voxels (TR=2 s, TE=40 ms, flip angle=90°). To allow for steady
state magnetization, the first 4 seconds of each run were excluded.
Data processing and analysis were performed using SPM2 (http://www.fil.ion.ucl.ac.uk/
spm) and custom software. The data were motion corrected, realigned, normalized onto
a common brain space (Montreal Neurological Institute (MNI) template), spatially
smoothed using a Gaussian filter (full-width half-maximum 5 mm kernel) and subjected
to a high-pass filter (128 Hz).
78 of 179
Motion and Artifact Analysis
To estimate motion and data quality, we used three measures for each participant:
mean translation was defined as the average absolute translation for each TR across
the x, y, and z plane; mean rotation was defined as the average absolute rotation per
TR across yaw, pitch, and roll; and number of outliers per run was defined as the
average number of time points per run in which (a) TR-to-TR head movement exceeded
2mm of translation, or (b) global mean signal deviated by more than three standard
deviations of the mean. One blind participant moved excessively during the scan (mean
translation of 1.2 mm, mean rotation of 1.4°, and mean outliers per run = 20.7,
compared to other blind participants’ means of 0.33 mm of rotation, 0.35° of rotation,
and 2.5 outliers), so his results were dropped from the analyses.
fMRI Analysis
All fMRI data were modeled using a boxcar regressor, convolved with a standard
hemodynamic response function (HRF). The general linear model was used to analyze
the BOLD data from each subject, as a function of condition. The model included
nuisance covariates for run effects, global mean signal, and an intercept term. A slow
event-related design was used. An event was defined as a single story, the event onset
was defined by the onset of the story sound file, and offset as the end of the story.
Functional Localizer: Individual subject ROIs
In each participant, functional regions of interest (ROIs) were defined in right and left
temporo-parietal junction (RTPJ, LTPJ), medial precuneus (PC), dorsal medial
prefrontal cortex (DMPFC 9), and ventral medial prefrontal cortex (VMPFC3). Each
subject’s contrast image (Belief > Photo) was masked with each of the regions’ likely
locations, using probabilistic maps created from a separate dataset.The peak voxel that
occurred in a cluster of 10 or more voxels significant at p<0.001 was selected. All voxels
within a 9mm radius of the peak voxel, individually significant at p<0.001, were defined
as the ROI.
Within-ROI Pattern Correlation Analysis
Split-half correlation-based MVPA asks whether the neural pattern in a region is
sensitive to a category-level distinction. Specifically, we ask whether the neural patterns
generated by items within a condition are more similar to each other (“within-condition
correlation”) than to the neural patterns generated by items in the other conditions
(“across-condition correlation”). If we find that within-condition correlations are reliably
higher than across-condition correlations, we can conclude that there are reliable neural
pattern generated by different items within a condition, and that these neural patterns
are distinct from one condition to another. Together, this suggests that the region is
sensitive to the category distinction — items within a category are coded in a similar
way, with distinguishable codes for different categories.
9
Note that in Bedny et al., 2009, DMPFC was called SOFC, and VMPFC was called OFC.
79 of 179
Here, we conducted within-ROI pattern analyses, independently testing for information
about belief source and valence in the regions identified in the independent functional
localizer. To compare seeing and hearing beliefs, we collapsed across good and bad
valence; to compare good
and bad valence, we
collapsed across seeing
and hearing.
Following Haxby et. al.
(2001), each participant's
data were divided into
even and odd runs
(‘partitions’) and then the
mean response (beta
value) of every voxel in
the ROI was calculated
for each condition.
Because each participant
read 8 stories about
hearing, seeing, feeling
good, and feeling bad,
each partition contained
the average response to 4
individual stories. The
"pattern" of response was
Figure 3.2: Split half MVPA analysis. The data from each condition the vector of beta values
(e.g. hearing and seeing) were divided by run, and for each across voxels within the
participant, we asked whether the within condition correlations (blue participants individual
arrows) were higher than the across conditions correlations (grey ROI. To determine the
arrows). Data from one blind participant: within-condition correlation w i t h i n - c o n d i t i o n
= 1.3; across-condition correlation = 0.8.
correlation, the pattern in
one (e.g. even) partition
was correlated with the pattern for the same condition in the opposite (e.g. odd)
partition; to determine the across-condition correlations the pattern was compared to the
opposite condition, across partitions (Figure 3.2).
For each condition pair (e.g. seeing vs hearing) in each individual, an index of
classification was calculated as the within-condition correlation (e.g.the correlation of
one half of the seeing stories to the other half of seeing stories, averaged with the
correlation of one half of the hearing to the other half of hearing stories) minus the
across-condition correlation (e.g. the correlation of seeing stories compared to hearing
stories). To allow for direct comparison of correlation coefficients, we transformed all r
values using Fisher’s Z transform. A region successfully classified a category of stimuli
if, across individuals, the within-condition correlation was higher than the acrosscondition correlation, using a Student's T complementary cumulative distribution
function.
80 of 179
This procedure implements a simple linear decoder. Linear decoding, while in principle
less flexible and less powerful than non-linear decoding, is preferable both theoretically
and empirically. A non-linear classifier can decode nearly any arbitrary feature contained
implicitly within an ROI, reflecting properties of the pattern analysis algorithm rather than
the brain, which makes successful classification largely uninformative (Cox & Savoy,
2003; DiCarlo & Cox, 2007; Goris & Op de Beeck, 2009; Kamitani & Tong, 2005;
Norman et al., 2006). Moreover, linear codes have been argued to be a more neurally
plausible way of making information available to the next layer of neurons (Bialek,
Rieke, Van Steveninck, & Warland, 1991; Butts et al., 2007; DiCarlo & Cox, 2007;
Naselaris, Kay, Nishimoto, & Gallant, 2011; Rolls & Treves, 2011).
Whole Brain Pattern Correlation Analysis (Searchlight)
To ask whether any other part of the brain encoded the difference between other
people's experiences of seeing and hearing, or between good and bad events, we used
a searchlight analysis to look across the brain. In the searchlight analysis, rather than
using a predefined ROI, a Gaussian kernel (14mm FWHM, corresponding
approximately to the observed size of the functional ROIs) was moved iteratively across
the brain. Using the same logic as the ROI-based MVPA, we computed the spatial
correlation, in each kernel, of the neural response (i.e. betas) within conditions and
across conditions. We then transformed the correlations using Fisher’s Z, and
subtracted the across-condition correlation from the within-condition correlation to
create an index of classification. Thus, for each voxel, we obtained an index of how well
the spatial pattern of response in the local region (i.e. the area centered on that voxel)
can distinguish between the two conditions. The use of a Gaussian kernel smoothly deemphasizes the influence of voxels at increasing distances from the reference voxel
(Fedorenko, Nieto-Castañón, & Kanwisher, 2012). We created whole brain maps of the
index of classification for each subject. These individual-subject correlation maps were
then subjected to a second-level analysis using a one-sample t-test (thresholded at
p<0.001, voxelwise, uncorrected).
81 of 179
Results
fMRI task: Behavioral Results
Sighted participants
We performed a 2x2 ANOVA, crossing valence (good/bad) and modality (seeing/
hearing), on both goodness ratings and reaction time data. Using a 1 (very bad) to 4
(very good) scale, sighted participants rated protagonists who experienced positive
events as feeling better than the protagonists experiencing negative events, with no
effect of modality and no interaction (hearing-bad: 1.84±0.14, hearing-good: 3.43±0.14,
seeing-bad: 1.88±0.12, seeing-good: 3.65±0.11; main effect of valence: F(1,12)=223,
p<0.001, partial η2=0.79; modality: F(1,12)=1.2 p=0.3, partial η2=0.02; modality by
valence interaction: F(1,12)=0.4, p=0.5, partial η2=0.01).
Sighted participants showed small, but significant effects of condition on reaction time
(measured from onset of the question), with a marginal main effect of modality and a
small but significant interaction (hearing-bad: 5.63±0.23, hearing-good: 5.2±0.21,
seeing-bad: 5.03±0.3, seeing-good: 5.28±0.23; main effect of modality: F(1,12)=4.4,
p=0.06, partial η2=0.02; valence: F(1,12)=0.3, p=0.5, partial η2<0.01; modality by
valence interaction: F(1,12)=6.8, p=0.002, partial η2=0.04). Post-hoc t-tests reveal that
sighted participants responded more slowly to stories about negative experiences
based on hearing than based on seeing (t(12)=3.3, p=0.007); there was no effect of
modality on responses to positive events (t(12)=0.47, p=0.65).
Blind participants
Congenitally blind participants also rated protagonists in positive stories as feeling
significantly better than those in negative stories, with no effect of modality and no
interaction (hearing-bad: 1.81±0.15, hearing-good: 3.51±0.13, seeing-bad: 1.95±0.25,
seeing-good: 3.43±0.11; main effect of valence: F(1,8)=72, p<0.001, partial η2=0.72;
modality: F(1,8)=0.02, p=0.9, partial η2 <0.01; modality by valence interaction:
F(1,8)=0.5, p=0.5, partial η2<0.02).
In reaction time of the blind adults, there were no significant effects of modality or
valence, and no interaction (hearing-bad: 5.51±0.35, hearing-good: 5.02±0.41, seeingbad: 5.47±0.33, seeing-good: 5.64±0.3; all F<3.1, all p>0.1).
Group Comparison
We found no differences in the ratings across groups. We performed a 2x2x2 repeated
measures ANOVA crossing valence and modality as within-subjects factors with group
(blind/sighted) as a between-subjects factor. All participants rated the protagonist as
feeling worse in the negative valence stories (F(1,20)=264, p<0.001, partial η2=0.76),
with no main effect of modality or group, and no interactions (all F<1, p>0.3).
In the reaction time data, there were no main effects of modality, valence, or group (all
F<1, p>0.4). We found a small but significant modality by valence interaction
(F(1,20)=7.0, p=0.02, partial η2=0.03): participants responded more slowly to stories
82 of 179
about negative experiences based on hearing than seeing; and more slowly to stories
about positive experiences based on seeing than hearing.
There was also a small group by modality interaction (F(1,20)=7.4, p=0.01, partial
η2=0.02): Sighted adults respond faster than blind adults to stories about seeing; blind
adults respond faster than sighted adults to stories about hearing. Post-hoc t-tests
comparing groups within modality (e.g. RTs of blind vs sighted for seeing stories)
revealed no differences in reaction time between groups in either modality (all t<1.1, all
p>0.2).
Motion and Artifact Analysis Results
Sighted participants and blind participants showed no difference in the mean translation
per run (sighted mean + sd = 0.22mm ± 0.03, blind = 0.23mm ± 0.04), t(20)=0.13,
p=0.9), mean rotation per run (sighted = 0.26 degrees ± 0.04, blind = 0.24 ± 0.04,
t(20)=0.37, p=0.71, or the mean number of outliers per run (sighted = 0.98±0.46, blind =
0.53±0.35, t(20)=0.72, p=0.48) Together, these data suggest the groups were well
matched in motion and scanner noise.
fMRI Results: Functional Localizer
Replicating many studies using a similar functional localizer task (e.g. Saxe &
Kanwisher, 2003), we localized five theory of mind brain regions showing greater
activation for false belief stories compared to false photograph stories in the majority of
participants (uncorrected, p<0.001, k>10): Sighted participants: RTPJ 13/13
participants, LTPJ, 12/13, PC 13/13, DMPFC 11/13, VMPFC 11/13; Blind participants:
RTPJ 8/9 participants, LTPJ 9/9, PC 9/9 and DMPFC 9/9, VMPFC 9/9 (Figure 3.3 A).
As reported in (Bedny et al., 2009), sighted and blind participants did not differ in the
activation or anatomical loci of any active regions (all p>0.1), nor did blind participants
show more variability in spatial location or size of ROIs.
fMRI Results: Within ROI Pattern Analysis
Sighted participants
Source (seeing vs. hearing): Multi-voxel pattern analyses revealed reliably distinct
patterns of neural activity for stories about seeing versus hearing in the RTPJ and LTPJ,
but not PC, DMPFC, or VMPFC, of sighted adults. Note that correlations are Fisher Z
transformed to allow statistical comparisons with parametric tests. Across partitions of
the data, the pattern generated by stories in one category (seeing or hearing) was more
correlated with the pattern for the same category than with the pattern for the opposite
category, in the RTPJ (within condition correlation ± standard error, z=1.1±0.2, across
condition correlation, z=0.95±0.2, t(12)=2.0, p=0.03) and LTPJ (within = 0.82±0.2,
across = 0.64±0.2, t(11)=2.0, p=0.04), but not in the PC (within = 0.86±0.2, across =
0.78±0.2, t(12)=0.93, p=0.19), DMPFC (within=0.7±0.1, across=0.5±0.2, t(10)=1.2,
p=0.12) or VMPFC (within = 0.58±0.2, across = 0.47±0.2, t(10)=0.75, p=0.23, Figure
3.3 B, Table 3.1).
83 of 179
Figure 3.3: MVPA results. (A) Canonical theory of mind brain regions; for all analyses, theory of mind
brain regions were individual defined in each participant. (B) Sighted adults (n=13) show pattern
discrimination for beliefs based on seeing vs. hearing in the right and left TPJ, but not in any other
theory of mind region. (C) These representations of seeing and hearing persist in the RTPJs of blind
adults (n=9).
To test whether the difference between ROIs itself was significant, we conducted a 1x5
repeated measures ANOVA. Because baseline differences in correlations across ROIs
are hard to interpret (higher overall correlation in one region compared to another could
be due to many factors, including the distance of a region from the coils, amount of
vascularization, or region size; see Smith, Kosillo, & Williams, 2011), we used difference
84 of 179
scores (within condition correlation - across condition correlation) as the dependent
variable. We found a main effect of ROI (F(4,28)=3.03, p=0.03, partial η2=0.27),
suggesting that the regions contain varying amounts of information about source
modality in their neural pattern.
Valence (good vs. bad): In no region did the pattern of response distinguish between
good and bad valence (all correlation differences t<0.2, all p>0.1, Figure 3.3 B).
Blind participants
Source (seeing vs. hearing): Like in seeing adults, the pattern generated by stories
within a condition were more correlated with other stories in the same condition
compared to stories in the opposite condition in the RTPJ (within condition correlation ±
standard error, z=1.1± 0.2, across condition correlation z=0.93±0.3, t(7)=2.0, p=0.04),
but not in the LTPJ (within=1.3±0.3, across=1.3±.03, t(8)=0.2 p=0.4), PC
(within=0.71±0.2, across=0.59±0.2, t(8)=1.3, p=0.12), DMPFC (within=0.83±0.2,
across=0.73±0.1, t(8)=1.4, p=0.11) or VMPFC (within=0.56±0.2, across=0.34±0.2,
t(8)=1.4, p=0.11, Figure 3.3 C, Table 3.1). This difference between ROIs in
discrimination was significant (F(4,28) = 4.0, p=0.01, partial η2 =0.36).
Valence (good vs. bad): As in sighted adults, the pattern of response in congenially
blind adults did not distinguish between good and bad valence in any theory of mind
region (all correlation differences t<0.2, all p>0.1, Figure 3.3 C).
Sighted
Blind
Region
Within
Across
Significance
RTPJ
1.1±0.2
0.95±0.2
t(12)=2.0, p=0.03*
LTPJ
0.82±0.2
0.64±0.2
t(11)=2.0, p=0.04*
PC
0.86±0.2
0.78±0.2
t(12)=0.93, p=0.19
DMPFC
0.7±0.1
0.5±0.2
t(10)=1.2, p=0.12
VMPFC
0.58±0.2
0.47±0.2
t(10)=0.75, p=0.23
RTPJ
1.1± 0.2
0.93±0.3,
t(7)=2.0, p=0.04*
LTPJ
1.3±0.3
1.3±.03
t(8)=0.2 p=0.4
PC
0.71±0.2
0.59±0.2
t(8)=1.3, p=0.12
DMPFC
0.83±0.2
0.73±0.1
t(8)=1.4, p=0.11
VMPFC
0.56±0.2
0.34±0.2
t(8)=1.4, p=0.11
Table 3.1: Z scores for within versus across condition comparisons of seeing and hearing. * p < .05
85 of 179
Across Groups
Source (seeing vs. hearing): Comparing across groups, we looked for three things: (a)
a main effect of discrimination, driven by differences between the within-condition and
across-condition correlations in that region: evidence of reliable discrimination between
conditions; (b) a main effect of group, which indicates that one group has overall higher
correlations, due to higher overall inter-trial correlations independent of condition (and
thus not suggestive of interpretable group differences), and (c) an interaction between
discrimination and group, such that one group shows a larger difference between the
within- and across-condition correlations: evidence that one group shows more
sensitivity to condition differences than the other. Finding an interaction would be the
key piece of evidence to show that blind and sighted people have difference sensitivity
to the condition manipulation.
Overall, we found that blind and sighted adults showed very similar neural patterns, with
no evidence that either group was more sensitive to the distinction of seeing versus
hearing. We found evidence of distinct neural patterns for seeing and hearing in both
blind and sighted adults in the TPJ and no other regions. Specifically, blind and sighted
participants show equally robust neural discrimination of seeing versus hearing in the
RTPJ, with a main effect of discrimination (F(1,19) = 7.8, p=0.01, partial η2 = 0.3), no
effect of group (F(1,19)=0.001, p=0.9), and, critically, no interaction (F(1,19)=0.02,
p=0.89). There were no significant effects of group, discrimination, or their interaction in
any other ROI. LTPJ showed a trend towards both a main effect of group (overall higher
inter-trial correlations in blind participants, F(1,19)=3.2, p=0.09) and a main effect of
discrimination (F(1,19)=3.6, p=0.07), with no interaction. Precuneus, DMPFC, and
VMPFC showed no effects (PC: group: F(1,20)=0.5, p=0.5, discrimination: F(1,20)=2.3,
p=0.14, interaction: F(1,20)=0.1, p=0.7; DMPFC: group: F(1,18)=0.9, p=0.4,
discrimination: F(1,18)=2.7, p=0.12, interaction: F(1,18)=0.3, p=0.6; VMPFC: group:
F(1,18)=0.1, p=0.8, discrimination: F(1,18)=2.1, p=0.16, interaction: F(1,18)=0.3, p=0.6).
In summary, this suggests that both blind and sighted adults code the difference
between other people's experiences of seeing and hearing in the RTPJ.
Valence (good vs. bad): We found no evidence of neural code distinguishing good
valence from bad in any brain region of either group. Comparing patterns for valence,
we found a main effect of group in the LTPJ (F(1,18)=5.1, p=0.04, partial η2=0.22), due
to higher overall inter-trial correlations in the blind, independent of condition (and thus
not suggestive of sensitivity to differences in valence). There were no significant effects
of group, discrimination, or their interaction in any other ROI. "
Whole Brain Searchlight Analysis
Source (seeing vs. hearing)
The results of the searchlight converge with the ROI analyses, suggesting a
representation of seeing and hearing in the RTPJ of both groups. Because of the
similarity across groups in the ROI analyses, we combined the data from blind and
sighted participants for increased power. Converging with the results of the ROI
analyses, the whole brain analysis revealed that only the RTPJ distinguished between
86 of 179
seeing versus hearing beliefs (n = 22, peak voxel at [60, -48, 12], p<0.001, uncorrected
peak T = 3.6).
Second, a two-sample T-test across groups revealed a significant group difference in
the left dorsolateral PFC (BA45/46, peak voxel at [-58, 30, 8], p<0.001 uncorrected,
peak T = 3.7). In this DLPFC region, sighted participants showed a greater difference
between within and across condition correlations than blind participants. No regions
showed stronger decoding in blind participants relative to sighted ones.
Valence (good vs. bad):
Combining across groups for power, we find that only the anterior cerebellum
distinguishes between good and bad emotional valence (n = 22, peak voxel at [12, -42,
-32], p<0.001 uncorrected, peak T = 4.2).
87 of 179
Discussion
MVPA reveals features of mental state representations
For neuroimaging research to have cognitive implications, a key challenge is to go
beyond where a cognitive function occurs in the brain, to provide a window into neural
representations and computations. Dozens of neuroimaging studies suggest a
hypothesis for where in the brain key aspects of theory of mind are processed (e.g.,
Aichhorn et al., 2009; Bedny, Pascual-Leone, & Saxe, 2009; Lombardo, Chakrabarti,
Bullmore, Baron-Cohen, & Consortium, 2011; Mason & Just, 2011; Rabin, Gilboa,
Stuss, Mar, & Rosenbaum, 2010; Saxe & Kanwisher, 2003; Saxe & Powell, 2006;
Spunt, Satpute, & Lieberman, 2010). The right and left TPJ, the PC, and MPFC show
increased hemodynamic responses to stimuli that require participants to think about
other people’s mental states, and many studies have investigated the selectivity and
domain specificity of these brain regions for theory of mind. These brain regions
therefore provide a prime opportunity to probe the neural computations of mental state
representation: What features of people’s beliefs and intentions are represented, or
made explicit, in these brain regions?
A powerful, if simplifying, assumption about neural representation is that a feature is
explicitly represented by a population of neurons if that feature can be decoded linearly
from the population’s response. Different subpopulations of neurons within a region may
contribute to a common task by representing different features or aspects of the
stimulus or task. A well-studied example is object recognition in the ventral pathway of
macaques. Low-level features of the stimulus, like retinotopic position and motion
energy, can be linearly decoded from the population response in early visual cortex,
whereas higher-level properties, like object identity (invariant to size and position), can
be linearly decoded from the population response in IT, a region further along the
processing stream (DiCarlo et al. 2012; Kamitani & Tong 2005).
We can take advantage of this property of neural populations to identify features of
human neural representations using fMRI. When the subpopulations of neurons that
respond to different stimulus features are at least partially organized into spatial clusters
or maps over cortex (Dehaene & Cohen, 2007; Formisano et al., 2003), those features
may be detectable in reliable spatial patterns of activity measurable with fMRI (Haynes
& Rees, 2006; Kriegeskorte & Bandettini, 2007; Norman et al., 2006). Multi-voxel
pattern analyses therefore offer an exciting opportunity for studying the representations
underlying human theory of mind. In spite of many studies on the neural basis of theory
of mind, very little is known about which features or dimensions of mental states are
represented by the neural populations in the candidate brain regions. Here we identified
two dimensions of others’ mental states that are common and salient: the source
modality of the other person’s experience, and the emotional valence. To ask whether
either of these dimensions are explicitly represented in these brain regions, we then
tested whether either of these dimensions could be linearly decoded from the spatial
pattern of the response in any theory of mind brain region.
We found that both the right and left TPJ of sighted people showed distinct spatial
patterns of responses to stories that described seeing versus hearing. Two independent
88 of 179
groups of stories, similar only in the source modality of the character’s beliefs, elicited
reliably similar spatial patterns of response in the TPJ, in spite of the wide heterogeneity
of the stories in many other respects (e.g. the physical environment, the type of event,
the character’s name, age, gender, social status, etc). Furthermore, we replicated these
results in the right TPJ in an independent group of congenitally blind adults. We
therefore suggest that spatially distinguishable neural populations in the TPJ explicitly
represent the perceptual source of another person’s mental states. Together with prior
evidence, our findings suggest that these representations develop in the absence of first
person experience.
Knowledge source and theory of mind
Why might the TPJ code the distinction between seeing and hearing mental states
described in stories? One possibility is that the patterns of activation we observed were
driven simply by a response to the mental state attitude verb used in the story (i.e.
“sees” vs “hears”). However, we consider this interpretation unlikely, for three reasons.
First, mental state verbs on their own elicit little to no response in the TPJ, both relative
to rest and relative to non-mental verbs (e.g. “to think” vs “to run” or “to rust,” Bedny,
McGill, & Thompson-Schill, 2008). Robust responses are best elicited in TPJ by full
propositional attitudes (e.g., a person’s mental attitude towards some content), ideally
embedded in an ongoing narrative. Second, in previous research we directly
manipulated the mental state verb in a sentence and did not find distinct patterns of
response in TPJ (e.g. “believe” vs “hope”, Paunov and Saxe, personal communication).
Third, in an other experiment, we found distinct patterns of response in TPJ for distinct
mental states that were described using the same verb (i.e. intentional vs accidental
harm, both described by “believed”, Chapter 2, Koster-Hale et al., 2013). In all, we
suggest that the patterns of activity observed here are most likely responses to features
of the character’s mental state, rather than to specific verbs in the stories. Still, future
research should test this possibility by independently manipulating the source of the
mental state and the verb used to describe it, as well as by looking at the pattern
evoked by these verbs outside of a mental state context.
Based on the current findings, we hypothesize that the TPJ contains an explicit
representation of the perceptual source of mental states. Why might theory of mind
brain regions code this information? Considering the source of others’ knowledge is
central to inferences about other people’s minds. Knowing how another person got their
information can influence the inferences we make about what they know, how well they
know it, and whether we should believe it ourselves. For example, we are more likely to
trust another person’s testimony if they acquired their knowledge by direct visual
observation (an “eye-witness”) than through gossip or hearsay (Bull Kovera, Park, &
Penrod, 1992; Peter Miene, Bordiga, & Park, 1993; Peter Miene, Park, & Bordiga,
1992); eyewitness but not hearsay testimony is admitted as evidence in courts of law
(Park, 1987). Similarly, recalling how one acquired one’s own information plays an
important role in evaluating and justifying a belief, and deciding how readily it should be
discarded (O'neill & Gopnik, 1991).
89 of 179
Recognizing sources of knowledge is so cognitively salient that some languages
explicitly mark source syntactically. In languages with evidential marking, such as
Turkish, Bulgarian, Tibetan and Quechua, utterances are syntactically marked to
indicate how the information was acquired (Aikhenvald, 2004; J. G. de Villiers, Garfield,
Gernet-Girard, Roeper, & Speas, 2009; Faller, 2002; Fitneva, 2001; Johanson & Utas,
2000; Smirnova, 2013). Some languages mark direct versus indirect experience, or
hearsay versus everything else; others mark whether the information was gained from
inference, hearsay, or sensory experience. Languages such as Tucano and Fasu
encode aspects of sensory modality, including distinctions between visual versus nonvisual sensory information; one language, Kashaya, has a separate marker for auditory
evidence (Aikhenvald, 2004). Acquisition of linguistic evidential marking by children
appears to depend on the development of theory of mind: during acquisition of Turkish,
for example, children’s earlier performance on explicit theory of mind tasks concerning
sources of knowledge predict their later comprehension of linguistic evidentials, but not
vice versa (Papafragou & Li, 2001).
In sum, the source of others’ knowledge is both behaviorally relevant and cognitively
salient; the current results suggest that it is also explicitly encoded by distinct neural
populations within the TPJ. An open question for future work is whether patterns in the
TPJ can distinguish only the modality or also other aspects of another person’s source
of knowledge. For example, sources of knowledge can be direct (perceptual
observation), inferential (by induction from clues or patterns), or from hearsay (from
other people’s report). Direct perceptual sources may be more or less reliable,
depending on the situation (e.g. observed from nearby or at a distance, based on a
quick glance or a long stare) and the observer (e.g. an expert, an amateur, or a child).
Future research should investigate the neural representation of these other features of
knowledge source, and the relationship between these features and perceptual source.
The valence of others’ mental states
In contrast to perceptual source, we failed to decode the valence of another person’s
mental experience from the pattern of activity in any theory of mind region. This is
especially noteworthy, considering that the behavioral task — judging how good or bad
someone felt at the end of the story — specifically drew attention towards valence, while
ignoring source information. However, a prior study successfully decoded positive
versus negative valence of others’ emotions, expressed in emotional body movements,
facial expressions, and tones of voice, from patterns of activity in MPFC (Peelen,
Atkinson, & Vuilleumier, 2010; Skerry & Saxe, in prep). One possibility is that emotional
experience is more effectively conveyed in non-verbal stimuli. However, null results from
MVPA should be interpreted with caution. The neural sub-populations that respond to
different types of valence could be intermingled such that we would be unable to detect
the distinct neural populations at the spatial scale of fMRI voxels; the distinction would
only be revealed by techniques with higher spatial resolution.
90 of 179
Blind adults represent knowledge source
We found similar patterns of neural activity in blind and sighted adults. Most notably, we
found that other people’s visual versus auditory experiences are encoded distinctly by
both sighted and blind adults. Blind individuals specifically lack first person experience
of visual aspects of mental states, and thus allow us to test the role of first person,
sensory experience in the development of theory of mind representation: what does it
take to be able to think about someone else’s experience?
One set of theories of how we understand others’ mental experiences posits that we
vividly simulate, or imagine, having the same experience ourselves. For example,
Gallese & Sinigaglia (2011) describe the process of simulation as: “People reus[ing]
their own mental states or processes in functionally attributing them to others, where the
extent and reliability of such reuse and functional attribution depend on the simulator’s
bodily resources and their being shared with the target’s bodily resources.” (p. 518).
That is, we understand other people’s mental states using a template of our own firstperson experiences. This view is difficult to reconcile with evidence that blind people
use similar neural mechanisms to reason about seeing as do sighted people. In the
initial analyses of these data, Bedny et al. (2009) showed that blind adults can represent
mental states acquired by vision, without additional costs or compensatory
mechanisms. The present analyses further show that in these same blind and sighted
adults, not only are the same brain regions recruited to think about others’ mental states
based on seeing and hearing, but these regions represent the difference between visual
and auditory sources of belief. These results suggest the representation of perceptual
source in the TPJ is not dependent on first-person experience.
Importantly, however, our data do not show that people never use first-person
experience to make inferences about other people’s minds and goals. Both in the motor
and sensory domains, people do seem to use sensory-motor information to understand
other’s actions and experiences. For example, interference between other people’s
sensory-motor experience and our own suggests that we use partially shared
representations for action observation and action execution. Observing someone else’s
actions can interfere with executing actions yourself (Kilner, Paulignan, & Blakemore,
2003), and observing what another person can see interferes with the ability to report on
what you yourself are seeing (Samson, Apperly, Braithwaite, Andrews, & Bodley Scott,
2010). Similarly, personal experience specifically helps children learn about others’
visual experiences: infants who experienced a transparent blindfold follow the gaze of a
blindfolded adult, but infants who experienced an opaque blindfold do not (Meltzoff,
2004; 2007; Meltzoff & Brooks, 2007).
In all, understanding of other minds is likely guided by both first-person experience,
which provides rich and detailed input about the character of specific experiences, and
an intuitive theory of mind, which allows individuals to form representations of
experiences they have not had, and indeed could never have, such as blind adults’
representations of experiences of sight. In some cases, such as action prediction, first
person sensory machinery may play a direct role in representing information about other
people; in other cases, first person experience provides a rich source of data about how
people feel and act, playing an important role in building causal theories of other’s
91 of 179
minds. The best “theory of mind” will both be able to deal with abstract generalizations
and make use of data from first-person experience (Apperly, 2008; Keysers & Gazzola,
2007; Kilner & Frith, 2007).
Learning about sight
The present findings, along with prior evidence, suggest that blind people can have
impressively rich knowledge about sight (Landau & Gleitman, 1985; Marmor, n.d.;
Shepard & Cooper, 1992). How do blind individuals acquire this extensive knowledge
about seeing? We propose that one source of information is others’ testimony. Blind
children live immersed in a world of sighted people, who talk constantly about
experiences of seeing. In general, children acquire much of their knowledge of the
invisible causal structure of the world through testimony (Harris, 2002a; 2002b; Harris &
Koenig, 2006; Koenig, Clément, & Harris, 2004). Language and testimony are a
particularly powerful tool for learning about the invisible contents of other minds (Bedny
& Saxe, 2012; Robinson & Whitcombe, 2003; Robinson, Haigh, & Nurmsoo, 2008;
Urwin, 1983); the absence of linguistic access to other minds can significantly delay
theory of mind development, for example in deaf children born to non-signing parents
(Figueras-Costa & Harris, 2001; Meristo et al., 2007; de Villiers, 2005; Moeller & Schick,
2006; Peterson & Wellman, 2010; Peterson & Siegal, 1999; Pyers & Senghas, 2009;
Schick, de Villiers, de Villiers, & Hoffmeister, 2007; Woolfe & Want, 2002). In contrast,
conversational access seems to give blind children a detailed representation of visual
mental states, and the inferences they afford.
The importance of testimony to the understanding of other minds is highlighted by the
contrast between how much blind people know about sight and how little sighted people
know about blindness. Blind individuals can never directly experience vision, but they
are constantly exposed to testimony about seeing. By contrast, sighted people do
experience the perceptual aspects of blindness (when in the dark or with their eyes
closed). Nevertheless, the average sighted person has little understanding of blindness.
Most sighted people do not know how blind individuals navigate their environment or
perform daily tasks such as dressing and eating. This lack of knowledge can lead to
incorrect and sometimes problematic inferences about blind people’s incapability to
work, learn, and parent. Given how much blind people know about seeing, such
misunderstandings are unlikely to result from an in principle incapacity to understand
experiences different from our own. While blind people are surrounded by sighted
people, most sighted individuals have never met anyone who is blind. The ignorance of
sighted people is thus another dramatic example of the importance of testimony for
learning about other minds.
However, these experiments have just begun to probe sight understanding in blind
individuals. In the current stimuli (and unlike our initial example of the woman at the
window) neither the content nor the reliability of the mental states depended on the
perceptual modality. The belief content (e.g. “his team lost”) was matched across
perceptual sources, and always reliably inferred from the perceptual evidence. Thus,
successfully encoding the story did not depend on specific knowledge about sight, e.g.,
how difficult it is to recognize a face in the dark. It is possible that blind people would
92 of 179
show different behavior and neural processes for such inferences. Additionally,
participants’ task (judging how good or bad the protagonist would feel) depended on
encoding the content but not the source of beliefs. Despite this, participants
spontaneously encoded the perceptual source of the belief described in the story.
However, as a consequence, the current results cannot determine whether blind and
sighted people would show similar neural patterns (or behavioral performance) if
specific details about the perceptual source were more relevant in the task. Future
research should investigate whether blind and sighted adults have similar knowledge
and representations of the details of sight, such as how knowledge based on vision will
vary with distance, occlusion, and darkness, and whether the neural patterns observed
here encode similar information about seeing and hearing. Such specific details may
show greater influences of first person experiences.
Finally, while the data here suggest a common endpoint in the development of an adult
theory of mind, the processes by which blind and sighted children acquire this theory
may be different. Although blind adults appear to have typical theory of mind, blind
children are delayed on a variety of theory of mind tasks, including both tasks that rely
directly on inferences about vision, such as perspective taking tasks (Bigelow, 1991;
1992; Millar, 1976), and also tasks that do not depend directly on visual perspective,
including auditory versions of the false belief task (McAlpine & Moore, 1995; Minter, &
Hobson, 1998; Peterson et al., 2000). Using a battery of theory of mind tasks designed
specifically for blind children, Brambring & Asbrock (2010) concluded that blind children
are delayed an average of 1-2 years, relative to blindfolded sighted control children,
though these children catch up in adolescence (see Peterson & Wellman, 2010 for a
similar argument about deaf children). Thus an open question is whether the neural
mechanisms for theory of mind, which are similar in blind and sighted adults, develop
differently in blind and sighted children.
Conclusion
In summary, we find that (i) theory of mind brain regions (specifically, the TPJ) encode
perceptual source and sensory modality of others’ mental states, and (ii) these
representations are preserved in the RTPJ of congenitally blind adults. Considerable
neuroimaging work on theory of mind suggests that the RTPJ plays a role in thinking
about others’ thoughts; we find that one aspect of this role is to make explicit, in the
population response of its neurons, features of beliefs — in this case, the perceptual
source of the belief. The persistence of these representations in congenitally blind
adults, provides evidence that theory of mind brain regions come to encode these
aspects of mental states even in the absence of first-person experience.
93 of 179
Acknowledgments
We thank the Athinoula A. Martinos Imaging Center at the McGovern Institute for Brain
Research at MIT, and the Saxelab, especially Amy Skerry and Hilary Richardson for
helpful discussion. We also gratefully acknowledge support of this project by an NSF
Graduate Research Fellowship (#0645960 to JKH) and an NSF CAREER award
(#095518), NIH (1R01 MH096914-01A1), and the Packard Foundation (to RS).
94 of 179
Chapter 4.
Mentalizing regions represent continuous, abstract
dimensions of others’ beliefs 10
The human capacity to think about others’ thoughts includes not only the capacity to
infer what another person is thinking, but critically also the ability to evaluate why the
other person believes what they do. In two experiments, we investigate how the
sources of others’ beliefs are coded in brain regions involved in mental state
inference, including bilateral temporo-parietal junction (TPJ), superior temporal
sulcus (STS) and medial prefrontal cortex (MPFC). We probed three epistemic
features of others’ beliefs (i.e. distinctions relevant to the source of another person’s
belief), and one salient but non-epistemic distinction (the valence of the person's
resulting emotion). Using multivoxel pattern analyses (MVPA), we find that the
patterns of activity in right TPJ and right STS contain complementary neural codes,
which reliably distinguish stories in which a person comes to a conclusion based on
reliable and unambiguous evidence, from stories in which they come to their
conclusion based on unreliable and ambiguous evidence. The neural classification
accuracy for each story was predicted by independent ratings of the strength of the
evidence in that story. We also find evidence for a representation of
evidence modality (visual vs auditory) in bilateral TPJ, and of emotional valence in
ventral MPFC. Together, these data suggest that the cortical regions implicated in
mental state inference contain neural representations of epistemic and emotional
features of others’s beliefs, and that these features are represented along
continuous, abstract dimensions. 10
Koster-Hale, J., Richardson, H., Velez-Alicia, N, Asaba, M, Young, L., & Saxe, R. R. (in prep). Thinking
about seeing: Mentalizing regions represent continuous, abstract dimensions of others’ beliefs
95 of 179
Introduction
Othello is a tragedy not only because Othello is wrong about Desdemona — she is not
unfaithful to him — but because he is wrong for the wrong reasons. Seeing a lost
handkerchief, hearing insinuations from a jealous and vengeful subordinate: Othello has
only poor, indirect, and unreliable evidence against his wife. And yet, as Iago tells the
audience, “trifles light as air/ are to the jealous confirmations strong/ as proofs of holy
writ.” (Othello, 3.3.322-324). For the audience, it is almost unbearable to watch Othello
allowing himself to believe, and to act on, such trifles.
Understanding why someone believes what they do — their evidence and reasons for
believing — is a key component of human social cognition. Observers readily make
inferences about the source of other people’s beliefs (Does the evidence really justify
their conclusion? Did they learn the information themselves, or did someone tell them?).
We can use information about about the source of others’ beliefs to facilitate learning
and cooperation (Robinson & Whitcombe, 2003; Smith & Ellsworth, 1987; Toglia, Ross,
Ceci, & Hembrooke, 1992; Whitcombe & Robinson, 2000); to judge and to punish
(Young, Nichols, & Saxe, 2010), and to decide who to trust in the future (Jaswal &
Neely, 2006; Koenig & Harris, 2005a; Scofield & Behrend, 2008). We more readily
believe people with knowledge from direct visual observation (an “eye-witness”) than
from inference or hearsay (i.e. other people’s report; Bull Kovera, Park, & Penrod, 1992;
Peter Miene, Bordiga, & Park, 1993; Peter Miene, Park, & Bordiga, 1992; Robinson,
Champion, & Mitchell, 1999) and we forgive those who made justifiable mistakes, based
on the evidence available to them, even when they cause harm (Kornblith, 1983; Young
et al., 2010). Yet, despite its prominent role in human social cognition, little is known
about how the brain tracks and represents the relationship that other people have to the
evidence available in the world.
Previous work suggests a set of cortical regions that may represent abstract features of
others’ minds. Regions implicated in mental state inference, including bilateral temporal
parietal junction (right and left TPJ), right superior temporal sulcus (STS), precuneus
(PC), and medial prefrontal cortex (MPFC), collectively called the “theory of mind (ToM)
network,” show selective activity while participants are thinking about other people and
especially their minds, and these effects are robust across stimuli, tasks, and
populations (for a review see Chapter 1, and Koster-Hale & Saxe, 2013). There is thus
extensive evidence concerning when, how much, and how selectively these regions are
recruited when thinking about other minds. Yet, relatively little is know about the
representations and computations supported by the populations of neurons in these
regions.
A powerful approach for understanding abstract neural representation is to ask which
features of a stimulus can be decoded from a population of neurons. Many high-level,
abstract features are coded in the pattern of response in a population of neurons (e.g.
Averbeck, Latham, & Pouget, 2006; DiCarlo, Zoccolan, & Rust, 2012; Quian Quiroga &
Panzeri, 2009; Tudusciuc & Nieder, 2007). Though fMRI data pools over hundreds of
thousands of neurons, if these subpopulations of neurons are spatially organized into
clusters or maps over cortex (Dehaene & Cohen, 2007; Formisano et al., 2003; Konkle
96 of 179
& Caramazza, 2013), then these representations may be detectable with fMRI, using a
class of analysis techniques called multi-voxel pattern analyses (MVPA; Haynes &
Rees, 2006; Kriegeskorte, Bandettini, & Bandettini, 2007; Norman, Polyn, Detre, &
Haxby, 2006). MVPA has been successfully used to probe the neural basis of many
different types of representation, including emotional affect and person identity
(Anzellotti, Fairhall, & Caramazza, 2013; Chikazoe, Lee, Kriegeskorte, & Anderson,
2014). Support vector machines (SVMs) are a particularly powerful version of linear
classifiers, because their relatively few free parameters and robustness against
overfitting (Hearst, Dumais, Osman, Platt, & Scholkopf, 1998; Steinwart & Christmann,
2008; V. N. Vapnik & Vapnik, 1998), and can yield trial-by-trial information, allowing us
to relate features of individual stimuli to the neural response they generate. Here, we
use these methods to probe the neural representations of others’ beliefs.
Specifically, we ask whether regions implicated in social cognition contain neural
populations that track features of others’ beliefs, including three epistemic distinctions
(i.e. those that are relevant to the source of others’ beliefs), and contrast this with one
salient but non-epistemic distinction: the valence of the person’s resulting emotion. We
independently manipulated three features of other people’s evidence: First, and mostly
critically, the quality of other people’s evidence: how good or bad was the evidence, for
the conclusion? Second, the perceptual modality of the evidence: was the other
person’s evidence seen or heard? Third, the directness of the evidence: was the other
person's evidence experienced first-hand or learned indirectly from another person’s
report? Finally, for comparison, we included a non-epistemic feature of other people’s
mental states: whether the resulting emotion is positive or negative (emotional valence).
To answer the question of how these features are represented, we first examined
whether population responses within theory of mind regions could reliably discriminate
between categories: good versus bad evidence; visual versus auditory evidence; firstperson experience versus hearsay, and positive versus negative emotion. We then
examined whether these representations were sensitive to continuous measures of
these dimensions that vary across individual stimuli. Lastly, we examined whether these
representations replicated and extended in a second experiment, manipulating the
quality of others’ evidence in the context of moral judgments.
97 of 179
Methods
Participants
Twenty right-handed members of the larger MIT community participated (11 women;
mean age ± SD, 28.5 years ± 6.8). All participants were native English speakers, had
normal or corrected-to-normal hearing and vision, gave written informed consent in
accordance with the requirements of Institutional Review Board at MIT, and received
payment. Two participants were dropped for excessive head motion, and one failed to
complete the experiment, leaving 17 in all analyses.
Stimuli
Experiment 1 featured 50 stories in 5 conditions. 10 stories were control stories,
describing physical events and scenes (such as gardens full of colorful flowers). In the
remaining 40 target stories, the protagonist came to hold a belief about the world, based
on one of four different types of evidence (i.e. four epistemic conditions). In Seeing
Good Evidence (SG) stories, the protagonist makes an inference based on clear visual
access to reliable and unambiguous evidence (e.g., a woman infers that a wolf is
outside, based on seeing large, fresh paw-prints, in good light). In Seeing Bad
Evidence (SB) stories, the protagonist makes the same inference based on indistinct
and unreliable visual evidence (e.g., the woman sees some old markings in the dirt in
dim light). In Hearing First-person Evidence (HF) stories, the protagonist draws her
conclusion based on good aural evidence (e.g., the woman hears a distinctive growl on
a quiet afternoon). Finally, in Hearsay Evidence (HS) stories, the protagonist comes to
her conclusion based on someone else’s report of evidence (e.g., the woman is told by
her friend that there are large, fresh paw prints in the dirt). In half of the stories in each
condition, the protagonist felt a negative emotion based on her inference (e.g. that
there is wolf outside), whereas in the other half of the stories, the protagonist would feel
a positive emotion based on her inference (e.g. that she is getting a puppy for her
birthday, Figure 4.1).
Evidence quality was manipulated in two ways. First, we changed the characters’
perceptual access to the evidence. In Seeing Good-Evidence stories, protagonists saw
evidence from close up, in bright lighting, and in still, visually clean environments, while
in Seeing Bad-Evidence stories, protagonists saw evidence through visual barriers
(such as dirty windows, or foggy air), in poor lighting, from far away, and in cluttered
environments. Second, we changed the extent to which the evidence supported the
conclusion. SG stories contained evidence that relatively directly and unambiguously
supported the protagonist’s conclusion (such as a pirate inferring his treasure has been
stolen after seeing a large hole and shovels where he buried his treasure), while in SB
stories the evidence was more ambiguous (the pirate reaching the same conclusion
after seeing a ship sailing off on the horizon line before he gets to his treasure site). The
bad evidence could both be visually ambiguous (concluding that a good friend has
arrived, from seeing only the outline of a person, compared to directly seeing his face)
and conceptually ambiguous (concluding that your favorite pen was stolen after seeing
two bullies give each other high-fives, compared to seeing them pocket something silver
98 of 179
and pen-sized). Directness was always the difference between the protagonist hearing
direct evidence with their own ears, versus being told about the evidence. The reported
evidence was always the same evidence presented in the SG or HF condition. Thus,
even in the HS stories, the protagonist is drawing their own conclusion based on the
evidence they were told about. Finally, modality was manipulated by changing whether
the protagonist received the information visually or aurally.
Individual participants saw each target story in only one of the four epistemic conditions,
forming a matched and counterbalanced design; every target story occurred in all four
conditions across participants. The valence of the protagonist’s emotion was
manipulated across stories; half of the stories in each condition led to a positive
emotion, and half led to a negative emotion.
Seeing
Good*Evidence*(SG)
Background*
(16s)
Evidence
(4?16s)
Seeing
Bad*Evidence*(SB)
Hearing*
First?person*(HF)
Hearing*
Hearsay*(HS)
Julie'had'wanted'a'puppy'of'her'very'own'for'years.'Every'holiday,'she'asked'her'parents'to'give'
her'a'puppy,'but'they'had'always'said'that'they'couldn’t'afford'it.'On'Julie’s'ninth'birthday,'she'
woke'up'and'ran'downstairs'to'her'parents.
In'the'cluBered'In'the'bright-kitchen,'
kitchen,'she'saw'a'
she'saw'a'newlyfew-small-toys-on-theinstalled-doggy-door.'
floor.
In'the'quiet'kitchen,'
she'heard'a'quiet,excited-yipping.'
Conclusion
(2?4s)
Julie'asked'her'mom,'"Did'we'get'a'new'puppy?!"
Task
Happy'or'Sad?
In'the'kitchen,'hermom-told-her'that'
there'was'a'newlyinstalled-doggy-door.'
Figure 4.1: Example stimuli from Experiment 1. In the fMRI task, participants listened to audio
recordings of 40 target stories and 10 physical control stories. Stories lasted from 22 to 36 seconds.
After each story, participants heard the question “Happy or sad?”, and indicated whether the main
character in the story felt happy or sad, using a button press (left/right).
Online Behavioral and Language Norming
To ensure that the condition manipulations were effective, we piloted 50 possible target
stories, each appearing in all four epistemic conditions (SG, SB, HF, HS; 200 items
total). Stories were rated by 521 adults in an online behavioral study using Amazon’s
Mechanical Turk (www.mturk.com), an online crowd-sourcing website that has been
found to yield valid and reliable data (Buhrmester, Kwang, & Gosling, 2011). All adults
were from the United States, and gave informed consent in accordance with the
requirements of the Institutional Review Board at MIT.
At the start of the norming study, participants were shown two example stories and
answers to ensure comprehension of the task. Participants then read the target story.
After reading each story, participants answered six questions. First, we verified that
participants could infer the modality of the evidence, by asking “At the end of the story,
[character] concludes that [conclusion]. Why did they think that?” Participants had a
99 of 179
forced choice response of “Something they heard” and “Something they saw”. Second,
we asked about the quality of the evidence, by asking “At the end of the story,
[character] concludes that [conclusion]. How good was his/her evidence?” (Scale 1 (not
good at all) to 7 (very good)). Third, we asked “How engaging was the story?” (1 (not at
all) to 7 (very engaging)). Fourth, we asked “How easy was it to read the story?” (1 (not
at all) to 7 (very easy)). Fifth, because the explicit task for fMRI participants was to rate
the character’s emotions, we asked “How happy does character feel at the end of the
story?” 1 (not at all happy) to 7 (very happy). Finally, we included a catch question,
asking “What is the name of the main character in the story?”
The study was designed so that participants never rated the same story in multiple
conditions. Four participants were excluded for (a) failing to follow basic instructions, (b)
leaving more than 50 percent of questions blank, or (c) failing 50 percent of the catch
questions, across items. Additionally, 182 ratings on single items were excluded
because (a) more than 50% of questions were left blank or (b) the catch question was
not answered correctly, leaving 521 participants who generated 3938 ratings.
Participants rated an average of 7.6 items, split evenly across conditions (SG: 2.36±1.9;
SB: 2.44±1.9; HF: 2.44±1.9; HS: 2.43±1.9). Each item was rated by an average of 19.7
people (16-21 people), split evenly across conditions (SG: 19.66±1.08; SB: 19.6±1.05;
HF: 19.8±0.88; HS: 19.7±1.02). Item-wise means were calculated as the average of all participant ratings for each item,
in each condition. Based on the behavioral ratings, 40 of the 50 target stories were
included in the fMRI experiment. These stories were rated as easy and engaging to
read, and were selected to maximize the difference in rated evidence quality between
SG and SB, and preserve the match in evidence quality between seeing and hearing,
and HF and HS, while keeping valence ratings, word count, reading ease, and
engagingness matched between conditions (Table 4.1).
Additionally, to match items in their reading ease and lexical frequency, we ran all
stories through the Coh-Metrix, an online tool that calculates a number of metrics
designed to measure computational cohesion and text difficulty (http://cohmetrix.com/’;
Graesser & McNamara, 2011; McNamara, Graesser, McCarthy, & Cai, 2014). For each
stimulus, we analyzed the Flesch reading ease (Flesch, 1948), Flesch-Kincaid Grade
Level (Kincaid, Fishburne, Rogers, & Chissom, 1975), and total word count and log
lexical frequency, both of which affect comprehension and reading time (Rayner & Duffy,
1986).
The behavioral ratings verified that our manipulation of evidence quality and valence
was effective: Seeing stories with Good evidence were rated as having better quality
evidence than Seeing stories with Bad evidence (t(77.2)=7.15, p<0.001, d=1.6 (large)).
In contrast, there was no difference in the quality of evidence between hearing firstperson and hearsay (t(77.2)=0.94, p=0.35), nor between seeing and hearing
100 of 179
(t(149.4)=-1.05, p=0.29)11 . The four epistemic conditions were matched on emotional
valence (F(3,156)=0.03, p=0.99). The protagonists of positive stories were rated as
significantly happier than the protagonists of negative stories (Positive: 6.13±0.06;
Negative: 1.69±0.03; t(110.5)=63.35, p<0.001, d=10.02 (large)).
Seeing Good
Evidence
Seeing Bad
Evidence
Hearing First
Person
Hearing
Hearsay
Evidence Quality
5.21±0.13
3.98±0.12
4.82±0.13
4.66±0.12
Valence
3.97±0.36
3.83±0.35
3.91±0.37
3.95±0.36
Engagingness
4.96±0.06
4.78±0.08
4.96±0.07
4.78±0.06
Reading Ease
6.31±0.03
6.23±0.04
6.28±0.04
6.16±0.04
Flesch Reading Ease
83.0 ± 6.6
82.5 ± 6.1
81.4 ± 6.8
82.5 ± 6.8
Flesch-Kincaid Grade Level
4.3 ± 1.2
4.4 ± 1.1
4.5 ± 1.1
4.5 ± 1.2
Word Count
68.2 ± 4.6
68.5 ± 4.4
67.5 ± 4.4
68.7 ± 4
Log Lexical Frequency
2.93 ± 0.13
2.95 ± 0.12
2.98 ± 0.13
3.01 ± 0.14
Table 4.1: descriptive statistics of fMRI stimuli used in Experiment 1. 521 participants on Amazon’s
Mechanical turk rated evidence quality, emotional valence, engagingness, and reading ease
on a scale from 1-7. Flesch Reading ease (range of 0 to 120), Flesch-Kincaid Grade Level (range of 1
to 12), word count, and log lexical frequency were calculated using Coh-Metrix, an online tool for
analyzing text.
Key Comparisons
This design allows us to test four key comparisons: First, by comparing Seeing GoodEvidence stories and Seeing Bad-Evidence stories, we can look for sensitivity to the
quality of others’ evidence (good versus bad), holding constant the modality (seeing)
and the directness (first-person). Second, by comparing Hearing First-Person stories to
Hearsay stories, we can look for sensitivity to the directness of others’ evidence (firstperson versus third-person), while holding constant the quality of the evidence and the
modality. Third, by comparing the two seeing stories (SG and SB) to the two hearing
stories (HF and HS), we can look for sensitivity to the modality of others’ evidence
(hearing vs. seeing) while holding constant the quality of the evidence (see evidence
from behavioral norming). Finally, by comparing positive to negative stories, we can look
for sensitivity to the valence of others’ emotions.
11
Note that there was a reliable difference in the quality of evidence between the Seeing Good-Evidence
stories and the First-person Hearing stories (SG: 5.21±0.13; HF: 4.82±0.13; t(78)=2.14, p=0.04, d=0.48
(small)). As a result, this specific comparison is confounded between modality and quality. All analyses of
the effect of modality use the full set of seeing and hearing stories, which are matched on evidence
quality.
101 of 179
fMRI Task
Participants were scanned while listening to audio recordings of 40 target stories and 10
physical control stories. All target stories had 16 seconds of background information
(identical across condition, 32 words ± 2.5), 6-14 seconds of evidence-gathering (in four
conditions; 25 words ± 2.3; matched across conditions; SG: 25.2±2.5, HF: 24.5±2.2, SB:
25.4±2.4, HS: 25.6±2.3), 2-4 seconds of conclusion (identical across condition, 10
words ± 2.4), and 6 seconds of response time (Figure 4.1). Stories were presented in a
pseudorandom order, condition order was counterbalanced across runs and subjects,
and no condition was immediately repeated. Rest blocks of 12 seconds occurred four
times in each run. The total experiment of five runs (6.2 minutes each) lasted 31
minutes. The stimuli were digitally recorded by 11 female speakers at a sampling rate of
44,100 to produce 32-bit digital sound files. During the stories, one of eight colorful
abstract images was displayed.
After each story, participants heard a recording of a male speaker ask “Happy or sad?”
Participants indicated whether the main character in the story felt happy or sad, using a
button press (left/right). During the response period, a screen with a happy (left) and
sad (right) face appeared, to remind participants which button to press. Reaction time
was measured from the onset of the question “Happy or sad?”. In each condition, 5
stories were happy (the protagonist felt good at the end of the story) and 5 stories were
sad.
Theory of Mind Localizer
Participants in both experiments were also scanned on a localizer task designed for
identifying theory of mind brain regions in individual participants (Dodell-Feder, KosterHale, Bedny, & Saxe, 2011; Saxe & Kanwisher, 2003); stimuli are available at http://
saxelab.mit.edu/superloc.php. Participants read verbal narratives in two conditions:
stories that described a situation in which someone held a false belief about the world,
such as mistaking white paint for cream or not knowing that her car had been moved
(belief condition); and stories describing a situation in which there was a false physical
representation of the world, such as an out-of-date photograph or advertisement
(physical condition). Both kinds of stories require the reader to consider incorrect or
outdated representations of the world, and so are similar in their meta-representational
and logical complexity; however, they differ in whether the reader is considering
someone else’s mental state, and thus comparing them serves to localize regions
recruited during mental state inference. Stories were presented in white, 24-point font
on a black background and were presented for 10s, followed by a true or false question
(4s) about the story. Each run (4.53min) consisted of 10 trials interleaved with 12s rest
periods. Participants saw two runs of the localizer (10 trials of each condition); The
orders of stimulus type (Belief or Photo) and correct answer (True or False) were
counterbalanced within and across runs.
Acquisition and Preprocessing
fMRI data were collected in a 3T Siemens scanner at the Athinoula A. Martinos Imaging
Center at the McGovern Institute for Brain Research at MIT, using a 32-channel phased
102 of 179
array head coil. We collected a high-resolution (1mm isotropic) T-1 weighted MPRAGE
anatomical scan, followed by functional images acquired with a gradient-echo EPI
sequence sensitive to blood-oxygen-level-dependent (BOLD) contrast (repetition time
(TR)=2s, echo time (TE)=40ms, flip angle=90°, voxel size 3x3x3mm, matrix 64x64, 26
near-axial slices). Slices were aligned with the anterior/posterior commissure and
provided whole-brain coverage (excluding the cerebellum). For steady state
magnetization, the first 4 seconds of each run were excluded.
Data processing and analysis were performed using SPM8, freesurfer, and custom
software. Data were motion corrected, realigned, normalized onto a common brain
space (rigid rotation and translation, MNI template), spatially smoothed using a
Gaussian filter (FWHM 5mm) and subjected to a high-pass filter (128 Hz).
Motion and Artifact Analysis
To estimate motion and data quality, we used three measures for each participant:
mean translation was defined as the average absolute translation for each TR across
the x, y, and z plane; mean rotation was defined as the average absolute rotation per
TR across yaw, pitch, and roll; and number of outliers per run was defined as the
average number of time points per run in which (a) TR-to-TR head movement exceeded
2mm of translation, or (b) global mean signal deviated by more than three standard
deviations of the mean. All outlier time points were excluded from both magnitude and
pattern analyses (using an impulse regressor of one TR). Runs missing more than 1/3
of the total time points due to artifact or motion were dropped, and participants with
fewer than 4 (of 5) remaining runs were dropped. Using these criteria, two participants
from Experiment 1 were excluded from all analyses.
fMRI Analysis
Modeling
All fMRI data were modeled using a boxcar regressor, convolved with a standard
hemodynamic response function (HRF). A general linear model was used to analyze the
BOLD data from each subject, as a function of condition, using a slow event related
design. The model included nuisance covariates for run effects, global mean signal, and
an intercept term, and impulse regressors to remove motion-artifact outliers. The
localizer was modeled as a 14s boxcar (the full length of the story and question period).
For the main experiment, we modeled the key sections of the stories separately: the
task was modeled using a single regressor for the first 16s of background (one
regressor for all 5 conditions), a single regressor for the conclusion and response time
(one regressor across all 5 conditions), and 5 condition regressors, modeling the 6-14
seconds of manipulated story (the “evidence” portion of the epistemic conditions).
Whole-brain random effects (univariate)
A second-level random effects analysis was performed on the contrast images
generated for each individual in the localizer task, by performing Monte Carlo
permutation tests using SnPM5 (Hayasaka & Nichols, 2004; http://www.sph.umich.edu/
103 of 179
ni-stat/SnPM/; T. Nichols, 2002), with a false positive rate controlled at p < 0.05, with
5000 permutations and a variance smoothing of 3mm.
Defining Individual subject ROIs
The whole-brain random effects analysis revealed eight main regions that showed
greater activation for false belief stories compared to false photograph stories (corrected
p < 0.05, pseudo-T=5.44): right and left temporo-parietal junction (RTPJ, LTPJ), right
middle superior temporal sulcus (RMSTS), right and left anterior superior temporal
sulcus (RASTS, LASTS), medial precuneus (PC), dorsal medial prefrontal cortex
(DMPFC), middle medial prefrontal cortex (MMPFC), ventral medial prefrontal cortex
(VMPFC).
Region
MNI
Size (Voxels)
Peak T
# Found
RTPJ
[54,-54,19]
277±67
11.01±2.44
17/17
LTPJ
[-47,-61,23]
251±81
9.69±2.05
17/17
RMSTS
[58,-25,-8]
158±76
8.19±1.71
17/17
RASTS
[56,-2,-18]
129±52
7.92±1.69
17/17
LASTS
[-54,1,-22]
123±88
6.56±1.74
17/17
PC
[1,-56,36]
300±63
9.60±2.00
17/17
DMPFC
[1,55,26]
145±102
7.57±2.29
15/17
MMPFC
[5,58,14]
157±108
7.87±2.33
16/17
VMPFC
[0,54,-14]
125±89
6.71±2.27
14/17
Table 4.2: Individual subject ROIs. Individual subject ROIs are defined as all voxels within a 9mm
sphere around the peak voxel (T of mental-physical from the localizer) identified in each region. MNI:
average coordinates of the peak voxels, across participants, reported in Montreal Neurological Institute
(MNI) space. Size: average number ± standard deviation of voxels included in individual ROIs. Peak:
average T value ± standard deviation of peak voxel for mental>physical. # Found: number of participants
in which an ROI was identified.
Based on the results of the whole-brain analysis, these functional regions of interest
(ROIs) were defined for each individual. For each participant, we masked the individual
T-map (one sample T-test of mental>physical) with hypothesis spaces defined from
random effects analysis on independent subjects (Dufour et al., 2013). In each
hypothesis space, we found the peak voxel in the contrast image for each region. ROIs
were then defined as all voxels within a 9mm radius of the peak voxel that passed
threshold in the contrast image (Belief > Photo, p<0.001 uncorrected, k>10 contiguous
voxels). Identifying the set of brain regions that are considered a core part of the theory
of mind network, these results replicate a number of studies using a similar functional
localizer task (Fletcher, Happé, Frith, & Baker, 1995; Gallagher et al., 2000; Gobbini,
Koralek, Bryan, KJ, & Haxby, 2007; Goel, Grafman, Sadato, & Hallett, 1995; Perner,
104 of 179
Aichhorn, Kronbichler, Staffen, & Ladurner, 2006; Saxe & Kanwisher, 2003; Van
Overwalle & Van Overwalle, 2009). These nine ROIs were identified in most individual
subjects (Table 4.2).
Within-ROI Multivariate Analysis (MVPA)
To ask whether any theory of mind ROIs contained information about our condition
manipulations, we conducted within-ROI multivoxel pattern analysis (MVPA) in
individual subject ROIs. The data from each individual ROI were classified using a
cross-validated linear support vector machine (SVM), which plausibly models the
readout of neural populations (Bialek, Rieke, Van Steveninck, & Warland, 1991; Butts et
al., 2007; DiCarlo & Cox, 2007; Hung, Kreiman, Poggio, & DiCarlo, 2005; Naselaris,
Kay, Nishimoto, & Gallant, 2011; Rolls & Treves, 2011; Seung & Sompolinsky, 1993;
Shamir & Sompolinsky, 2006), and performs as well as most non-linear classifiers (Bray,
Chang, & Hoeft, 2009; Cox & Savoy, 2003).
An individual ROI (Figure 4.2 A) had to have at least 20 voxels to be included in the
analysis. In each voxel, the data were high-pass filtered (within run; 128 Hz) and linearly
detrended (across runs). For each trial, we calculated the average BOLD signal in each
voxel in the ROI, measured during the “evidence” portion of the story (starting at 17 s
into the story, offset by 4s to account for hemodynamic lag, ranging from 4-16 s in
length). To normalize voxel response, BOLD responses in each voxel were z-scored
relative to the mean signal in that voxel across trials. This procedure resulted in a neural
“pattern” for each trial: a vector of the z-scored BOLD responses in each voxel inside
the ROI (Figure 4.2 B).
To reduce the dimensionality of the number of features (voxels) relative to the number of
items, feature selection was used to find voxels most likely to contain task-related signal
(De Martino et al., 2008; T. M. Mitchell et al., 2004; Pereira & Mitchell, 2009). Using data
from all runs, a task vs rest ANOVA identified the 100 voxels in each ROI in which the
response to all conditions differed most reliably from rest (ranked by F-statistic; T. M.
Mitchell et al., 2004). Because all conditions are modeled together, this selection
criterion did not bias the outcome of classification on condition pairs (“peeking”) and
allowed us to use the same voxels across cross-validation folds.
In this analysis, each item is treated as a single point in high-dimensional voxel space
(n=100); linear SVMs find a hyperplane through this space which maximizes the
distance between labeled classes (the margin), while minimizing the number of mislabeled data points. We conducted all analyses with a “soft” fixed regularization
parameter (C=1), allowing for a wider margin, at the cost of more mislabeled data in the
training set. After the model is built, new data can be classified based on the neural
pattern they elicit. Thus, successful classification requires that items generate neural
patterns that are reliably similar within a condition, and relatively different between
conditions.
The SVM algorithm was implemented with libSVM (http://www.csie.ntu.edu.tw/~cjlin/
libsvm/; C.-C. Chang & Lin, 2011). The data were partitioned into cross-validation folds
by run. A classification model was built iteratively on all runs but one (using labeled
data), and then used to predict the condition labels of items in the remaining run.
105 of 179
(Figure 4.2 C). Accurate predictions matched the actual stimulus label, and inaccurate
ones did not.
To determine subject-wise
c l a s s i fi c a t i o n a c c u r a c y,
accuracy was averaged across
items within a run, and then
averaged across runs, to yield
a single classification accuracy
for each participant. A region
could reliably distinguish
between two conditions if
subject-wise classification
accuracies were significantly
above chance (.50; onesample, one-tailed t-test12).
Figure 4.2 (A) RTPJ (B) Training and testing data. Matrix
shows mean activation in each voxel (rows) to each item
(columns). A classification model is built from the labeled
training data. The model finds a linear hyperplane (solid line)
through voxel space that best separates the two conditions
(light and dark dots) based on their neural activation (dashed
lines). The model is then applied to the left out testing data
(circled dots). Test items are labeled based on which side of
the hyperplane they fall; accuracy is computed as the
percentage of correctly labeled test items. (C) A two-voxel
model with 100% classification accuracy in the testing data; (D)
A two-voxel model with 75% classification accuracy.
12
Item-Wise classification:
Te s t i n g f o r c o n t i n u o u s
representations
In addition to looking at
subject-wise classification
accuracies, we also calculated
item-wise classification for
each condition. For each item,
classification of quality is
defined as the proportion of
times an item was classified as
“good evidence”, across
participants; modality
classification is defined as the
proportion of time an item was
classified as visual evidence.
Based on the subject-wise
results, we asked whether
ROIs that show a binary
distinction (e.g., successful
classification of good versus
bad evidence quality across
subjects) also represent graded
Parametric tests are not always appropriate for testing classification accuracy (Stelzer, Chen, & Turner,
2013). Here however, all classification data were independent (from separate subjects, rather than from
folds on overlapping data), and classification accuracies were normally distributed around the mean for
most ROIs. Where there was significant non-normality (using Lilliefor’s test for normality) in an ROI, we
report non-parametric statistics.
106 of 179
information about the distinction (e.g., very good evidence versus somewhat ambiguous
evidence across items).
We asked if item-wise classification scores (e.g., proportion of times an item was
classified as “good”) were predicted by item-wise behavioral ratings (1-7 scale of “how
good is the evidence”?). Because of the high collinearity between behavioral ratings and
condition labels, we used regularized linear regression (Hastie, Tibshirani, & Friedman,
2011; Hoerl & Kennard, 1970; ridge regularization; Krämer, Schäfer, & Boulesteix,
2009). Like multiple linear regression, ridge regression learns a beta weight for each
regressor, and uses the weighted sum of regressors to predict an outcome. However,
ridge regression simultaneously minimizes the sum of the squared prediction error and
the sum of the squared beta — effectively limiting the size of the coefficients, which are
often inflated by collinear regressors. In each model, the regularization parameter λ
determines the bias towards solutions with smaller beta values. As λ goes to 0, the
algorithm becomes identical to multiple linear regression (OLS). To avoid over-fitting λ,
for each model, we used cross-validation, iterating across 8 folds. Items were evenly
and randomly distributed into folds. In each iteration, items from one fold were set aside
for testing. In the remaining items, leave-one-out cross validation was used to build a
series of regularized linear models. Ridge regression was performed on the left-in
training items, using 1000 logarithmically spaced values of λ, ranging from from 1*10-8
to 30, implemented in R 3.1.1 (http://www.R-project.org R Core Team, 2013) using
packages parcor (Krämer et al., 2009) and ridge (Cule, 2012; Cule & De Iorio, 2013).
For each of these models, the predicted value for the withheld training item was
calculated. For each value of λ, the total residual was calculated as the sum of the
square of the predicted value minus the actual value for the withheld training items,
averaged across folds. The value of λ that minimized the overall error in the withintraining cross-validation was used to model the testing data. To determine how well the
model fit overall, betas for each regressor (condition labels and behavior) were
averaged across folds.
"
107 of 179
Results
fMRI task: Behavioral Results
After each story, participants answered whether the main character in the story felt
happy or sad. There were no differences in accuracy or response time between Seeing
Good-Evidence stories and Seeing Bad-Evidence stories (Accuracy (mean proportion
correct ± standard error): SG: 0.94±0.02; SB: 0.92±0.02; t(34.6)=0.55, p=0.59; RT
(mean time in seconds ± standard error): SG: 1.38±0.09; SB: 1.52±0.11; t(36.9)=-1,
p=0.32), between hearing first-person and hearsay (Accuracy: HF: 0.92±0.02; HS:
0.96±0.02; t(35.2)=-1.46, p=0.15; RT: HF: 1.39±0.08; HS: 1.37±0.09; t(37.6)=0.22,
p=0.83), or between seeing and hearing (Accuracy: seeing: 0.93±0.02; hearing:
0.94±0.02; t(38)=-0.45, p=0.65, d=-0.14; RT: seeing: 1.45±0.09; hearing: 1.38±0.09;
t(37.6)=0.54, p=0.59).
fMRI task: Classification Results
The key question we asked in Experiment 1 was whether any of the classic “ToM brain
regions” contains distinct spatial patterns of response to different features of others’
beliefs. Confirming that these regions play a role in ToM, 7/9 regions of interest showed
reliably distinct patterns of neural responses to stories describing the protagonist’s belief
versus stories describing physical events: RTPJ (accuracy=0.67(0.03), t(16)=5.7,
p<0.001), LTPJ (accuracy=0.61(0.04), t(16)=2.7, p=0.0074), RMSTS (accuracy =
0.62(0.03), t(16)=3.6, p=0.001), RASTS (accuracy = 0.57(0.03), t(16)=2.4, p=0.015), PC
(accuracy=0.62(0.03), t(16)=4.1, p<0.001), DMPFC (accuracy=0.59(0.03), t(14)=3.3,
p=0.0029), and MMPFC (accuracy=0.57(0.04), t(15)=1.8, p=0.049), with a marginal
effect in LASTS (accuracy=0.55(0.03), t(14)=1.7,p=0.06; Figure 4.3 A).
We next tested whether these regions distinguish between different features of beliefs,
including three epistemic distinctions (i.e. those that are relevant to the source of
others’ knowledge), and one salient but non-epistemic distinction (the valence of the
protagonist’s resulting emotion).
Quality
First, we tested whether any region represents others’ beliefs in terms of the quality of
evidence. Overall, the pattern of neural responses could successfully classify stories
describing good versus bad visual evidence in two regions: RTPJ (accuracy=0.56(0.03),
t(16)=2.4, p<0.014) and RMSTS (accuracy=0.58(0.04), t(16)=2.2, p=0.023, Figure 4.3
B), but in no other region (all p>0.1).
Independent behavioral ratings of the quality of the evidence in each story predicted
how likely that story was to be classified as depicting good evidence in both RTPJ
(r(78)=.29, p=0.009, Figure 4.3 C), and RMSTS (r(78)=.28, p=0.01, Figure 4.3 C). In
both regions, the behavioral ratings contributed significant information to neural
classification accuracies even after accounting for the binary condition labels (RTPJ:
Condition: beta=0.16±.1, t=1.7, p=0.11; Behavioral rating: beta=0.22±.1, t=2.3, p=0.038;
RMSTS: Condition: beta=0.2±.08, t=2.2, p=0.035; Behavioral rating: beta=0.17±.08,
t=2.1, p=0.048).
108 of 179
These results suggest that RTPJ and RMSTS not only encode the distinction between
good and bad evidence overall, but also encode continuous differences in evidence
quality. Interestingly, RMSTS and RTPJ seem to encode different aspects of the quality
of evidence depicted in the stories: there was no correlation between the classification
accuracies in RTPJ and RMSTS item-wise (r(78)=.18, p=0.12) or subject-wise
(r(15)=-0.19, p=0.48), and the classification accuracy of RTPJ and RMSTS made
independent contributions to the behavioral ratings (RTPJ: beta=1.0±.4, t=2.3, p=.02;
RMSTS: beta=1.1±.5, t=2.2, p=.03).
Modality
Second, we tested whether any region represents others’ beliefs in terms of the
modality of evidence. Replicating a prior study (Chapter 3, Koster-Hale, Bedny, & Saxe,
2014), we found distinct patterns of response to beliefs based on visual versus aural
evidence in RTPJ (accuracy=0.55(0.03), t(16)=2.1, p=0.027) and LTPJ
(accuracy=0.59(0.03), t(16)=2.8, p=0.007, Figure 4.3 B). The RTPJ and LTPJ appear to
contain very similar representations of the modality of evidence: classification scores for
modality in RTPJ and LTPJ were correlated across items (r(157)=0.21, p=0.009). Also,
across participants, individual differences in classification accuracies for modality were
correlated between RTPJ and LTPJ (r(15)=0.51, p=0.04). There were marginal trends
towards successful classification by modality in PC (accuracy=0.54(0.03), t(16)=1.4,
p=0.093) and in MMPFC (accuracy=0.53(0.02), t(15)=1.5, p=0.072).
Quality vs Modality
The RTPJ thus showed significant classification accuracy for both quality and modality
of evidence. These two dimensions are often correlated in real-world situations (e.g. the
expression “seeing is believing”), although they were deliberately made orthogonal in
our stimuli, so one possibility is that the RTPJ actually encodes only one of these two
dimensions, and the other dimension is distinguished by proxy. We tested this
hypothesis in two ways. First, we asked whether an item that was rated as depicting
better evidence was more likely to be classified as “visual” based on the pattern of
response in the RTPJ. We found no such correlation, either in all 160 items, or within
the 80 visual items alone (all items: r(158)=0.04, p=0.6; visual items: r(78)=-0.01,
p=0.9). Behavioral ratings of evidence quality were better predictors of neural
classification by evidence quality than by evidence modality (difference of correlations,
z(78)=1.93, p=0.05). Second, we asked whether there were shared individual
differences, across participants, in classification accuracies for evidence quality and
evidence modality in the RTPJ. We found no relationship between the classification
accuracy in a participant’s RTPJ for modality versus quality (r(15)=-.22, p=0.4).
Together, these results suggest that RTPJ contains independently significant and
orthogonal representations of the quality and modality of the source of others’ beliefs.
Directness
Third, we tested whether any region distinguishes between others’ beliefs based on
first-person (auditory) evidence versus hearsay (another person’s report), independent
109 of 179
Figure 4.3 (A) Individual ROIs (B) Classification accuracy in Experiment 1 for Mental: mental vs.
physical stories, Quality: good vs. bad quality evidence (SG, SB), Modality: visual vs auditory
modality (S,H), Directness: first person vs hearsay (HF, HS), and valence: positive vs negative
emotion, (C) Classification accuracy in the replication study for Quality: good vs. bad quality
evidence. (D). RTPJ, RSTS, VMPFC: Correlation between item-wise classification scores and
behavioral ratings. LTPJ: density plot of classification accuracies by condition. In RTPJ and
RMSTS there were significant correlations between quality classification scores (how many times
an item was scored as “good”, across participants) and behavioral judgments (how strong is the
character’s evidence?). In VMPFC, there was a significant correlation between valence
classification score (how many times an item was scores as “positive”) and behavioral judgements
(how happy does the character feel?). Modality is not a continuous feature and thus not compared
to behavior; rather we observe striking similarity in the classification accuracy in LTPJ between
seeing conditions (red: SG and SB) and hearing conditions (purple: HF HS). * p < 0.05
110 of 179
of quality and modality. However, we could not classify stories into these categories in
any region (all p>0.14).
Valence
Finally, we tested whether any ToM region distinguishes stories in which the protagonist
feels happy versus sad as a result of their conclusion. The pattern of neural response in
one region, the VMPFC, could reliably classify stories resulting in positive versus
negative emotions (accuracy=0.57(0.04), t(12)=1.9, p=0.038, Figure 4.3 B).
Independent behavioral ratings of the valence of the protagonist’s emotion in each story
predicted how likely that story was be to be classified as positive valence by the VMPFC
(r(158)=0.21, p=0.009, Figure 4.3 C). The behavioral ratings contributed significant
information to neural classification accuracies even after accounting for the binary
condition labels (Condition: beta=0.13±.07, t=2.2, p=0.047; Behavioral rating:
beta=0.15±.07, t=2.2, p=0.02).
Generalization to other manipulations of evidence quality
A key novel result of Experiment 1 is the evidence of a representation of the quality of
others’ evidence in RTPJ and RMSTS. To test the replicability of this result, and its
robustness to variation in the stimulus and task, we reanalyzed published data that
contained a similar manipulation.
Young et al. (2010) asked participants to make moral judgments of actions (e.g. “Grace
puts the powder in her friend’s coffee”) based on justified or unjustified beliefs. Belief
justification was manipulated by describing the belief as based on good evidence (e.g.
“Grace thinks the white powder is sugar, because the container is labeled ‘sugar’”) or
based on insufficient evidence (e.g. Grace thinks the white powder is sugar, though
there's no label on the container.”). To look for neural representations of belief
justification, Young et al. (2010) analyzed the average magnitude of response in ToM
brain regions, and found no difference in the magnitude of response to justified versus
unjustified beliefs in any region. However, they did not conduct any pattern analyses. In
the current re-analysis, we tested whether the pattern of response in any ToM region
(particularly the RTPJ and RMSTS) distinguished between beliefs based on good
versus insufficient evidence.
Stimuli
In this study, the “bad evidence” consisted not of obscured and ambiguous evidence,
but of missing or misleading evidence (Figure 4.4). In some stories, the protagonist
came to a conclusion without any evidence at all (believing a tree house is safe, despite
having never been inside, or believing that a watch is waterproof without seeing any
markings on it). In some stories, the protagonist is inattentive to key details (failing to
notice that someone next to them is choking); in others, the protagonist seems to
actively ignore good evidence (failing to read a note from a pet-owner specifying what
their kittens can eat, or failing to slow down to check whether a runner stopped on the
side of the road needs help). Finally, in some stories, the protagonist draws a
111 of 179
conclusion that seems to go directly against the available evidence (believing that a
sleeping roommate has heard a fire alarm even though he didn’t move, or believing
lunch meat is safe to eat despite its bad smell.) The manipulations of evidence quality in
this experiment are very different than Experiment 1, which manipulated the
protagonist’s perceptual access and the extent to which the evidence unambiguously
supported the conclusion. Thus, we can use these data to ask not only whether the
neural representations are found in a new set of participants, with new stimuli and a
new task, but also whether they extend to conceptually different manipulations of
evidence quality.
Good*Reason
Bad*Reason
Background
(6s)
Ray'and'his'girlfriend'are'on'a'weekend'hike.'They'come'across'a'narrow'bridge'that'spans'a'
rocky'creek.'
Evidence
(4s)
Ray'thinks'that'the'bridge'is'safe'to'walk'across Ray'thinks'that'the'bridge'is'safe'to'walk'across
because-it-isthough-it’s-not
marked'with'the'same'nadonal'park'symbol'as' marked'with'the'same'nadonal'park'symbol'as'
all'the'other'footbridges.
all'the'other'footbridges
AcHon*
and*
Outcome
(6s)
Ray'encourages'his'girlfriend'to'walk'across'the'bridge.
The'bridge'is'actually'very'unstable.'It'is'a'makeshie'bridge'put'together'by'other'hikers'without'
much'care.'Unfortunately,'it'collapses,'and'Ray's'girlfriend'falls'and'shagers'her'ankle.'
Task*(4s)
How*morally*blameworthy*is*Ray*for*encouraging*
his*girlfriend*to*walk*across*the*bridge?
Figure 4.4: Example stimuli from replication.
Methods and Analyses
In this study, 20 participants (12 women; mean age ± SD, 21.25 years ± 2.6) read 54
stories in the scanner. Each story was broken into three sections: (1) the background,
which set the stage (identical across conditions; 6 seconds), (2) the agent’s belief
(identical across conditions) and the justification for the belief (bad, good or
unspecified; 4 s), and (3) the conclusion, which detailed the agent’s action and the final
outcome (bad or neutral; 6 s; Figure 4.4). Only stories describing good or bad evidence
are analyzed here. Also, we collapse across outcome: 1/3 of the stories ended with a
bad outcome (the protagonist’s action led to physical injury or death of another
character), and 2/3 of the stories ended with a neutral outcome (no harm to anyone).
Each story appeared once in each condition, forming a matched and counterbalanced
design. Individual participants saw each story in only one of the 9 conditions; every
story occurred in all conditions, across participants. Stories were presented in a
pseudorandom order; conditions were counterbalanced across runs and participants,
and 14 s rest blocks were interleaved between stories. After each story, participants
responded to the question “How morally blameworthy is [the agent] for [performing the
action]?” on a 4-point scale (1-not at all, 4-very much), using a button press. Stories
were presented in white, 24-point font on a black background. Nine stories were
112 of 179
presented in each 5.1 minute run, with six runs total (30.6 minutes). Each participant
also completed the theory of mind localizer (four runs, 20 stimuli per condition).
Data were collected and analyzed using the same parameters as Experiment 1, though
with a 12 channel head coil. No participants were excluded due to motion. Individual
ROIs were picked using the data generated by the ToM localizer (Table 4.3). SVM
classification was performed in individual ROIs, on the data generated during the 4
second evidence section of each story (offset 4 s for hemodynamic lag).
Results
As in Experiment 1, the neural
pattern of response in the RTPJ
could successfully classify beliefs
based on good versus insufficient
evidence (accuracy=0.54(0.02),
t(17)=1.8, p<0.05, Figure 4.3 B:
quality 2). No other ROI showed this
d i s t i n c t i o n ( a l l c l a s s i fi c a t i o n
accuracies < 51%, p>0.3), including
the RMSTS (accuracy = 0.47(0.03),
t(15)=-0.99, p=0.83; Wilcoxon
signed-rank W = 6, p=0.70 13).
RMSTS carried reliably less
information about evidence quality
than RTPJ (Wilcoxon signed-rank W
= 86.5, p=0.04). Moreover, RMSTS
carried significantly less information
than it did in Experiment 1 (Wilcoxon
signed-rank W = 75.5, p-value =
0.03), suggesting that RMSTS
carries reliably less measurable
information in the replication than in
Experiment 1.
Exp 2
MNI
Size
Coordinate
(Voxels)
s
Peak T
# Found
RTPJ
[54,-51,19]
231±97
7.22±1.65
18/20
LTPJ
[-54,-58,25] 173±111
6.05±1.42
19/20
RMSTS
[57,-26,-7]
89±55
5.37±1.37
19/20
RASTS
[55,2,-22]
64±45
5.12±0.8
15/20
LASTS
[-54,3,-25]
40±36
4.5±0.74
15/20
PC
[1,-57,36]
233±80
6.48±1.36
20/20
DMPFC
[3,54,29]
93±86
5.06±1.1
15/20
MMPFC
[6,59,14]
77±70
5.24±0.92
13/20
VMPFC
[2,56,-8]
57±58
4.6±0.96
13/20
Table 4.3: Individual subject ROIs. Individual subject
ROIs are defined as all voxels within a 9mm sphere
around the peak voxel (T of mental-physical from the
localizer) identified in each region. MNI: average
coordinates of the peak voxels, across participants,
reported in Montreal Neurological Institute (MNI) space.
Size: average number ± standard deviation of voxels
included in individual ROIs. Peak: average T value ±
standard deviation of peak voxel for mental>physical. #
Found: number of participants in which an ROI was
identified.
13
Because the distribution of the RMSTS differed significantly from a normal distribution (Lilliefors test for
normality, K=0.26, criterion=0.21, p=0.004), we report results from the relevant parametric and nonparametric tests.
113 of 179
Discussion
Across two experiments, using the pattern of BOLD responses we could decode
whether another person’s belief was based on good versus bad evidence in the RTPJ.
In the first study, we also found evidence for a representation of the quality of others’
evidence in RMSTS. In the same paradigm, we replicated and extended two previously
observed findings: we could decode the perceptual modality of others' evidence (seeing
versus hearing) in the left and right TPJ, and positive versus negative emotional valence
in VMPFC. Overall, these results provide evidence for highly abstract representations of
others’ mental states, with distinct dimensions of mental states represented in different
regions within the “mentalizing network”.
Evidence Quality is explicitly represented in RTPJ and RMSTS
A key result of this paper is that the RTPJ and RMSTS contain a neural code which
tracks the degree to which someone’s evidence supports their beliefs. Using linear
classification, we found that the neural pattern generated by one set of story items
predicted the condition label in independent items, reliably distinguishing situations in
which someone comes to a conclusion based on evidence that strongly supports their
belief (good evidence) from situations in which someone comes to the same conclusion
based on unreliable and ambiguous evidence. Moreover, in both RTPJ and RMSTS, the
classification score of each item — the frequency with which it was labeled as “good”
evidence, based on the neural pattern it elicited — correlated with independent
behavioral ratings on the quality of the evidence. RTPJ and RMSTS each explained
significant and uncorrelated variance in the behavioral ratings, suggesting that the
regions encode different aspects of the stimuli. Together, these data indicate that the
neural populations in RTPJ and RMSTS contain explicit, and possibly complementary,
representations of a key epistemic feature of belief attribution: the quality of the
evidence that other people use to draw their conclusions.
Replicating Experiment 1, we found in a re-analysis of the data from Young et al. (2010)
that the neural pattern of response in the RTPJ reliably distinguished between beliefs
based on good versus insufficient evidence. The replication of this result in the second
experiment suggests that the effect in RTPJ generalizes across participants, stimuli, and
task. The converging results are particularly compelling because the two experiments
used different manipulations of evidence quality. While Experiment 1 manipulated
mainly perceptual aspects of evidence quality (e.g. distance, distraction, darkness,
clutter, brevity), the replication described relevant evidence as either present or absent,
or attended or ignored. Rather than being driven by superficial stimulus features or task
demands, the distinct neural patterns in the RTPJ thus likely reflect an underlying
distinction in the representation of the quality of others’ evidence for their beliefs.
By contrast, we found no evidence of a distinction between good and insufficient
evidence in the RMSTS in the replication. Although null results in MVPA must be
interpreted with caution (e.g., due to limits in spatial scale), this result is consistent with
the results in Experiment 1 showing that the RTPJ and RMSTS represent different
aspects of the quality of evidence. It is possible that the stimuli in the replication did not
114 of 179
include a manipulation of the key aspects of evidence quality represented in the
RMSTS. One interesting hypothesis is that RMSTS is sensitive to information about
perceptual ambiguity, while RTPJ tracks information about conceptual ambiguity. Or, the
RMSTS may be sensitive to features of the explicitly stated bad evidence or observers,
while the RTPJ more explicitly represents the link between the quality of the evidence
and the belief.
The modality of others’ evidence is explicitly represented in right and
left TPJ
Our second key finding is that the neural patterns in RTPJ and LTPJ code information
about the perceptual source of other people’s beliefs (did they see or hear the
evidence?), independent of quality or directness. These results replicate and clarify the
findings in Chapter 3 (Koster-Hale et al; 2014), who originally found that the neural code
in bilateral TPJ reliably distinguished the perceptual source (seeing versus hearing) of
other people’s beliefs. Here, we simultaneously manipulated modality, quality, and
directness. Finding converging results across experiments, despite differences in
stimuli, participants, and manipulation, suggests that the neural data are being driven by
the intended manipulation (modality), rather than other correlated but unintended
features of the stimuli. Interestingly (and unlike RTPJ and RMSTS), the information
represented in RTPJ and LTPJ seems very similar, across items and participants:
participants with high decoding accuracy in RTPJ had high decoding accuracy in LTPJ;
and items that had a high frequency of being classified as “seeing” in one region were
more likely to be frequently categorized as “seeing” in the other.
While RTPJ is implicated in the processing of others’ beliefs, relative to other types of
representations (such as maps or signs), LTPJ has been implicated in more general
meta-representation (tracking differences between two types of representation; for
example a person’s own memory of the geography, and a map of the area), and
possibly in processing misleading information and perspective taking in particular
(Aichhorn et al., 2009; Perner et al., 2006). Future research should test whether there is
a division of labor between these regions, and how they work together to represent
information about other people’s mental states. An intriguing possibility is that the LTPJ
may be implicated in general tracking of information sources, as part of encoding where
our information comes from, and keeping different pieces of information separate.
No region could decode directness: first-person experience versus
hearsay
The third epistemic dimension that we tested in our experiment was directness: whether
someone came to a conclusion based on their own, first person experience, or via
hearsay: information told to them by someone else. Directness plays an important role
in evaluating the experiences of others. Adults are more likely to trust another person if
they acquired their knowledge through first person experience than from hearsay (i.e.
other people’s report; Bull Kovera et al., 1992; Peter Miene et al., 1992; 1993; Robinson
et al., 1999) and children are less likely to be misled by adults reporting information they
learned via hearsay than via first person experience (Aydin & Ceci, 2009). Eyewitness
115 of 179
but not hearsay testimony is admitted as evidence in courts of law (Park, 1987).
Interestingly, however, we found no region that differentiated between direct evidence
and hearsay. This could be due to basic limits in the data (e.g. the representations are
not at a spatial scale measurable by our analysis). More interestingly, first person
experience versus hearsay may not be at opposite ends of a scale. If, for example, first
person experiences are represented as a heterogenous class (encoded via features
such as quality, modality, unexpectedness), there may not be a common neural
representation among items, and would not be detectable using pattern-based SVMs.
The valence of others’ emotions is represented in VMPFC
The final key result of this study is that the neural pattern in VMPFC codes information
about the valence of other people’s experiences — whether the protagonist felt positive
or negative at the end of the story. Moreover, each item’s valence classification score in
VMPFC correlated with independent behavioral ratings on the valence of each story.
These results suggest that VMPFC encodes continuous differences in emotional
valence.
Our results converge with and extend previous results on high-level emotion
representation in MPFC. Most relevantly, Peelen et al. (2010) found cross-modal
representations of emotion across facial expressions, body postures, and vocal
expressions in MPFC; and in the same region Skerry and Saxe (under review) found
evidence of a common neural code for others’ valence that generalizes across both
perceptual cues (facial expressions) and context-based inferences (short cartoons
without any perceptual cues to emotional valence).
Our results support the conclusion that VMPFC contains abstract representations of
valence. Our stories contain no perceptual cues (such as facial expressions), and, in
fact, very few emotion words, leaving the participants to infer valence from the context.
Moreover, to our knowledge, this is the first study to find distinct neural patterns for
others’ positive versus negative valence using text-based stimuli, and, more importantly,
the first study to find that these representations encode a continuous, rather than
categorical, representations of valence.
Neural representations of abstract features
Taken together, these results are particularly striking because they find evidence of
neural representations of highly abstract conceptual features. Information about other
people’s mind cannot be directly observed. Rather, we successfully reason about
others’ minds and actions by positing an internal causal structure of intentions, beliefs
and desires. The present results suggest that cortical regions implicated in social
cognition may represent features of this abstract causal model: the reasons why people
believe what they believe.
At least two of the features we tested (evidence quality and valence) are coded along a
continuous dimension. Previous work has found that, across participants, classification
accuracy can predict task performance (Kok, Jehee, & de Lange, 2012; Chapter 2,
Koster-Hale, Saxe, Dungan, & Young, 2013; Walther, Beck, & Fei-Fei, 2012) and clinical
symptoms (Coutanche, Thompson-Schill, & Schultz, 2011; Dosenbach et al., 2010),
116 of 179
suggesting that differences in neural patterns may reflect individual differences in
behavior. Here, we show that the probability of an item being categorized as one type of
stimuli is predicted by independent behavioral ratings, suggesting that the brain may
represent these stimuli along continuous, abstract feature dimensions (Kriegeskorte,
2008; Kriegeskorte & Kievit, 2013). A key next step will be to be to further characterize
the space that these dimensions are part of: what other features are represented, and
how are they coded?
Finally, these data suggest a division of labor across the brain. We find that information
about the source of others’ beliefs is represented in regions reliably implicated in social
cognition. Disassociations in functional and spatial profile in single individuals (Scholz,
Triantafyllou, Whitfield-Gabrieli, Brown, & Saxe, 2009) and meta-analyses (Schurz et al,
2014; Decety & Lamm, 2007; Kubit & Jack, 2013, Krall et al., 2014), functional
connectivity (Bzdok et al., 2012; Krall et al., 2014), and structural connectivity (Mars et
al., 2012) all suggest that both TPJ and MPFC contain neural populations that activate
selectively to mental state reasoning tasks, relative to a variety of other tasks. However,
a key open question is what kinds of computations these neural populations support.
MVPA has shown that RTPJ contains information about future social actions (Carter,
Bowling, Reeck, & Huettel, 2012), social groups versus individuals (Contreras,
Schirmer, Banaji, & Mitchell, 2013), and intent when causing harm (Chapter 2, KosterHale et al., 2013), and the reliability of others’ cooperation (Behrens, Hunt, Woolrich, &
Rushworth, 2008). In contrast, MPFC has been shown to track information about longterm personality traits (Hassabis et al., 2013) and emotional states (Kassam, Markey,
Cherkassky, Loewenstein, & Just, 2013; Peelen et al., 2010). The data here argue that
the RTPJ represents features of beliefs; and in particular the epistemic features of
beliefs, including the quality and modality of the evidence that gives rise to other
people’s beliefs.
This work adds to the growing body of literature that has found neural representations of
a variety of high-level cognitive domains. For example, person identity, a representation
that is highly abstract (we can determine the identity of someone by hearing their voice,
seeing their face, noticing their distinctive gait from afar, or seeing a corner of their
vintage purse), yet specific (each person we encounter is represented differently), is
encoded in anterior temporal lobe, generalizing across voice identification and facial
identification in a variety of environments, such as visual angle, lighting and partial
occlusion (Anzellotti et al., 2013). Information about personality traits, such as
extraversion, is tracked in MPFC (Hassabis et al., 2013). Value information, in both
monetary and social contexts, drives activity in VMPFC and ventral striatum (Behrens,
Hunt, & Rushworth, 2009; Izuma, Saito, & Sadato, 2008; Poore et al., 2012). Semantic
category is tracked in ATL (Correia et al., 2014). In combination, these types of results
suggest that the brain represents complex and abstract information in neural
populations of relatively circumscribed cortical patches — including in areas associated
with social cognition.
117 of 179
Why track epistemic features of beliefs?
Distinguishing reliable sources of information from unreliable ones informs key social
inferences: deciding what information we believe to be true, what actions to take, and
who we will trust. Critically, tracking the reliability of others’ beliefs plays an important
role in successfully learning from others. Much of our information comes from other
people; having information about the source of their beliefs can be key to knowing what
to keep and what to discard, while retaining the source of the evidence can allow us to
reevaluate our beliefs in the face of additional information — for example, learning that
an informant is nearly deaf allows us to retroactively update our confidence in the
gossip she shares (Robinson, Haigh, & Nurmsoo, 2008). Moreover, source information
allows the learner to probe for new knowledge by making inferences about what else
their informant is likely to know. For example a person who heard a glass smash, and a
person who saw the glass fall, both have good reason to believe that the glass is
broken, but only the person with visual access is likely to know that there is red wine on
the white carpet. Similarly, by age five years, children anticipate that a puppet who
looked at a toy will be more accurate about its color than a puppet who touched the toy,
or a puppet who is just guessing (O'neill & Chong, 2001; Pillow, Hill, Boyce, & Stein,
2000), and both adults and children selectively choose to learn new information from
reliable sources (Robinson & Whitcombe, 2003; Smith & Ellsworth, 1987; Toglia et al.,
1992; Whitcombe & Robinson, 2000), even when the speaker’s evidence contradicts
their own first person experience.
As well as informing our judgments about the reliability of the belief itself, source
knowledge can impact the kinds of inferences we draw about the long-term personality
traits of the belief-holder. Learning about the kinds of inferences that someone is willing
to draw, based on the evidence they observed, can be telling about their
trustworthiness, credibility, and belief system. When someone makes sensible
conclusions based on clear evidence, we are more likely to believe them in the future,
consider them trustworthy, and rely on them as a future source of evidence. In contrast,
when someone comes to a strong conclusion in the face of weak or unreliable evidence,
we are likely to conclude that they themselves are unreliable sources of information,
and possibly crazy, stupid, or untrustworthy (Clément, Koenig, & Harris, 2004; Hardwig,
1991; Koenig & Harris, 2005b; Lucas, Lewis, Pala, Wong, & Berridge, 2013; Olson,
2003). In fact, information about how someone acquired their beliefs can affect how
much we blame (and punish) them when things go wrong. When people make moral
judgements, causing harm by accident is typically judged to be much less blameworthy
than causing harm intentionally (e.g Cushman, 2008). If Grace puts a white powder in
her friend’s coffee, believing that the powder is sugar, she is less blameworthy than if
she knows the powder is poison (Young & Saxe, 2009). However, exoneration for the
accident depends substantially on why Grace thought the powder was sugar. If the
powder was in a jar labeled “sugar”, next to the coffee machine, then her belief was
reasonable and she is not to blame for the accident. If the white powder came from an
unlabeled jar in a chemical factory, then Grace is blamed much more for poisoning her
friend (Young et al., 2010).
118 of 179
An interesting open question is how sensitivity to source information develops. Tracking
the source of others’ beliefs develops along with other aspects of mental state
understanding well through the age of 10: children’s performance on theory of mind
tasks predicts their ability to report the sources of their own knowledge (Bright-Paul,
Jarrold, & Wright, 2008), their resistance to misleading information from uninformed
adults (Welch-Ross, Diecidue, & Miller, 1997), their ability to distinguish between
information from an informed source versus an informed source (Welch-Ross, 1999),
their ability to predict whether someone else will accept or reject conflicting information
(Robinson, 1994). Thus, finding representations of source in the adult brain provides a
key opportunity to probe the developmental trajectory of potentially late-developing
neural mechanisms of mental state inference.
Finally, an interesting open question is what role, if any, these same representations
play in representing one’s own experiences of good and bad evidence. Recalling how
one acquired one’s own information plays an important role in evaluating and justifying
our own beliefs, and deciding how readily they should be discarded (O'neill & Gopnik,
1991), and can improve memory fidelity (Dodson & Johnson, 1993; Lindsay & Johnson,
1989; Morton, 1994; Ross, Ceci, Dunning, & Toglia, 1994; Zaragoza & Lane, 1994).
Summary and Conclusion
In summary, using multi-voxel pattern analysis in two experiments, we find (i) that RTPJ
and RMSTS contain neural codes which track the degree to which someone’s evidence
supports their beliefs, reliably distinguishing between good, reliable evidence and
obscured and ambiguous evidence; (ii) that in both RTPJ and RMSTS, the frequency
that an item was labeled as “good” evidence was predicted by independent behavioral
ratings on the quality of the evidence, and that each brain region explained independent
variance in behavioral ratings; (iii), that RTPJ and LTPJ encode information about the
modality of other people’s evidence, (iv) that the neural pattern in VMPFC codes
information about the valence of other people’s experiences, and (v) that the frequency
of an item being classified as “positive” was predicted by independent behavioral ratings
of emotional valence.
Together, these data suggest that the brain regions implicated in social cognition contain
precise neural representations of abstract features, including epistemic and emotional
features of others’ beliefs, and that these features are represented in a continuous,
abstract feature space.
Acknowledgments
We thank the Athinoula A. Martinos Imaging Center at the McGovern Institute for Brain
Research at MIT, and the Saxelab, especially Amy Skerry and Stefano Anzellotti for
helpful discussion. We also gratefully acknowledge support of this project by an NSF
Graduate Research Fellowship (#0645960 to JKH) and an NSF CAREER award
(#095518), NIH (1R01 MH096914-01A1), and the Packard Foundation (to RS).
119 of 179
120 of 179
Chapter 5.
Theory of Mind: a neural prediction problem 14
Predictive coding posits that neural systems make forward-looking predictions about
incoming information. Neural signals contain information not about the currently
perceived stimulus, but about the difference between the observed and the predicted
stimulus. I propose to extend the predictive coding framework from high-level
sensory processing to the more abstract domain of theory of mind: that is, to
inferences about others’ goals, thoughts, and personalities. I review evidence that,
across brain regions, neural responses to depictions of human behavior, from
biological motion to trait descriptions, exhibit a key signature of predictive coding:
reduced activity to predictable stimuli. I discuss how future experiments could
distinguish predictive coding from alternative explanations of this response profile.
This framework may provide an important new window on the neural computations
underlying theory of mind.
14A
version of this chapter appeared as Koster-Hale, J., & Saxe, R. R. (2013). Theory of mind: a neural
prediction problem. Neuron, 79(5), 836–848. doi:10.1016/j.neuron.2013.08.020
121 of 179
Introduction
Social life depends on developing an understanding of other people’s behavior: why
they do the things they do, and what they are likely to do next. Critically, though, the
externally observable actions are just observable consequences of an unobservable,
internal causal structure: the person’s goals and intentions, beliefs and desires,
preferences and personality traits. Thus, a cornerstone of the human capacity for social
cognition is the ability to reason about these invisible causes: if a person checks her
watch, is she uncertain about the time or bored with the conversation? And is she
chronically rude or just unusually frazzled? The ability to reason about these questions
is sometimes called having a “theory of mind”.
Remarkably, theory of mind seems to depend on a distinct and reliable group of brain
regions, sometimes called the “mentalizing network” (e.g., Aichhorn et al., 2009; Saxe
and Kanwisher 2003), which includes regions in human superior temporal sulcus (STS),
temporo-parietal junction (TPJ), medial precuneus (PC), and medial prefrontal cortex
(MPFC). Indeed, the identity of these regions has been known since the very first
neuroimaging studies were conducted by 2000, based on four empirical studies, Frith
and Frith concluded that “studies in which volunteers have to make inferences about the
mental states of others activate a number of brain areas, most notable the medial
[pre]frontal cortex [(MPFC)] and temporo-parietal junction [(TPJ)]” (Frith and Frith 2000).
Since then, more than 400 studies of these regions have been published. However,
although there is widespread agreement on where to look for neural correlates of theory
of mind, much less is known about the neural representations and computations that
are implemented in these regions. The problem is exacerbated because these brain
regions, and functions, may be uniquely human (Saxe 2006; Santos et al., 2013).
Recent evidence suggests that there is no unique homologue of the TPJ or MPFC
(Rushworth et al., 2013; Mars et al., 2013), making it even harder to directly investigate
the neural responses in these regions.
In the current review, we import a theoretical framework, predictive coding, from other
areas of cognitive neuroscience and explore its application to theory of mind. There has
recently been increasing interest in the idea of predictive coding as a unifying
framework for understanding neural computations across many domains (e.g., Clark
2013). In this review, we adapt a version of the predictive coding framework that has
been developed for mid- and high-level vision. Like vision, theory of mind can be
understood as an inverse problem (Baker et al., 2011; Baker et al., 2009); the challenge
is to use the observable evidence (in this case human behaviors and states) to infer an
invisible causal structure that gave rise to the evidence (the goals, thoughts, and
personality of the individual; Seo and Lee 2012). Also like vision, theory of mind is a
complex cognitive process that depends on many different brain regions with likely
distinct computational roles (DiCarlo et al., 2012). We suggest that a predictive coding
framework can be used both to shed light on existing data about these brain regions,
and to suggest productive new lines of research.
First, we briefly review predictive coding, and sketch a model we believe can serve as
an integrative framework for the neuroscience of theory of mind. Second, we provide a
122 of 179
selective review of existing neuroimaging studies of theory of mind. Across different
stimuli and designs, with correspondingly different social information and predictive
contexts, we find a classic signature of a predictive error code: reduced neural response
to more predictable inputs. Third, we discuss how to distinguish predictive coding from
alternative explanations of this response profile, including differences in attention or
processing time. Based on recent neuroimaging experiments in visual neuroscience, we
suggest strategies for future experiments to test specific predictions of predictive
coding. Finally, we discuss the implications of predictive coding for our understanding of
the neural basis of theory of mind.
A predictive coding framework
The central idea of “predictive coding” is that (some) neural responses contain
information not about the value of a currently perceived stimulus, but about the
difference between the stimulus value and the expected value (Fiorillo et al., 2003;
Schultz et al., 1997; Schultz, 2010). This general idea is most familiar from studies of
“reward prediction error” in dopaminergic neurons in the striatum. Famously, these
neurons initially fire when the animal receives a valued reward, like a drop of juice, and
do not respond above baseline to neutral stimuli, such as aural tones. After the animal
has learned that a particular tone predicts the arrival of a drop of juice two seconds
later, the same neurons fire at the time of the tone. Tellingly, the firing rate of these
neurons no longer rises above baseline at the time the juice drop actually arrives.
Nevertheless, the neurons still respond to juice. If the tone that typically predicts a
single drop of juice is unexpectedly followed by two drops of juice, the neurons will
increase their firing; and if the tone is unexpectedly followed by no drops of juice, the
neurons decrease their firing rate below baseline (Fiorillo et al., 2003; Schultz et al.,
1997). These dopaminergic neurons exhibit the simplest and best known example of a
neural “error” code: the rate of firing corresponds to any currently “new” (i.e. previously
unpredictable) information about the value of coming rewards, not to the actual value of
any currently perceived stimulus (Bayer and Glimcher, 2005; Nakahara et al., 2004;
Tobler et al., 2005)
Predictive coding addresses a general challenge that an animal faces: developing an
accurate model of the expected value of all incoming inputs. Thus, predictive coding
models can be applied beyond the context of reward prediction to cortical processing
more generally. In fact, predictive coding was initially suggested as a model for visual
perception (Barlow, 1961; Gregory, 1980; Mumford, 1992), using a visual “error” code
that preferentially encodes unexpected visual information. The key benefit of such a
code, proponents suggest, is to increase neural efficiency, by devoting more neural
resources to new, unpredictable information.
By contrast to the single population of reward prediction error neurons, predictive coding
in the massively hierarchical structure of cortical processing poses a series of
challenges, however. If sensory neurons respond to errors, there must exist other
neurons to provide the prediction. Thus predictive coding models require at least two
classes of neurons: neurons that formulate predictions for sensory inputs (“predictor”
neurons, also called “representation” neurons; Summerfield et al., 2008; Clark 2013),
123 of 179
and neurons that respond to deviations
from the predictions (“error” neurons).
Because sensory input passes through
many hierarchically organized levels of
processing (DiCarlo et al., 2012;
F e l l e m a n a n d Va n E s s e n 1 9 9 1 ;
Logothetis and Sheinberg 1996;
Desimone et al., 1984; Maunsell and
Newsome 1987), a predictive model of
sensory processing requires an account
of the interactions between prediction
and error signals, both within a single
level and across levels.
To illustrate the idea, we provide our own
sketch of a hierarchical predictive coding
model. This proposal is a hybrid of
multiple approaches (Friston 2010; Clark
2013; Wacongne et al., 2012; de Wit et
al., 2010; Spratling 2010), seems to
capture the essential common ideas, and
is reasonably consistent with existing
data. The key structural idea is that
Figure 5.1. A sensory-coding based model of
social predictive coding. Predictor neurons (P) predictor neurons code expectations
code expectations about the identity of incoming about the identity of incoming sensory
input and pass down the prediction to lower level input and pass down the prediction to
predictor neurons (green arrows) and lower level both lower level predictor neurons and
error neurons (blue arrows). Error neurons (E) act lower level error neurons. Error neurons
as gated comparators, comparing sensory input act like gated comparators: they compare
from lower levels (red arrows) with the information
from predictor neurons (blue arrows). The sensory input from lower levels with the
difference between the predicted input and the information from predictor neurons. When
actual input is passed up higher level error neurons, the information that is being passed up
propagating up the processing hierarchy (red from lower levels matches the information
arrows). Error neurons also modulate the response carried by the predictor neurons, the error
of predictor neurons (purple arrows), likely both by neurons’ response to the input is
inhibiting the predictor neurons making incorrect
predictions, and enhancing predictor neurons reduced. This type of inhibition is the
making correct predictions. When the information classic signature of predictive coding,
that is being passed up from lower levels matches “explaining away” predictable input (Rao
the information carried by the predictor neurons, the and Ballard 1999). However, when
error neurons’ response to the input is reduced, predictor neurons at a higher level fail to
“explaining away” the predictable input (Rao and predict the input (or lack of input), there is
Ballard 1999).
a mismatch between the top-down
information from the predictor neurons
and the bottom-up information from lower levels, and error neurons respond robustly.
This error response propagates up the processing hierarchy. The consequence is a
sparse, efficient representation (mostly in predictor neurons) of predictable input, and a
124 of 179
robust, distributed response (mostly in error neurons) to unpredictable input, both
coordinated across multiple levels of the processing hierarchy (Figure 5.1).
Within a cortical region, population activity reflects a mixture of responses in the
predictor neurons (passing information about predicted inputs down the hierarchy) and
the error neurons (passing information about unpredicted inputs up the hierarchy). In
principle, predictive coding models need make no assumption about the distribution of
these two kinds of neurons within a population; in practice, aggregate population activity
is often dominated by error neurons (Friston 2009; Wacongne et al., 2012; Egner et al.,
2010; Keller et al., 2012; Meyer and Sauerland, 2009). The result is that the classic
signature of predictive coding, reduced activity to predictable stimuli, is typically
observed when averaging across large samples of neurons within a region (Meyer and
Olson 2011; Egner et al., 2010; de Gardelle et al., 2012). However, (as described in
more detail below) signatures of the predictor neurons can also be observed; for
example, the predictor neurons would likely show increased response when the input
matches their predictions (e.g., de Gardelle et al., 2012).
Following work in sensory processing (e.g., Wacongne et al., 2012), in our proposal
both error neurons and predictor neurons convey “representational” information, and
both are likely tuned to specific stimuli or stimulus features. Predictor neurons, present
at each level of the cortical hierarchy, do not code a “complete” representation of the
expected stimulus, but only some features or dimensions of the stimulus, at a relevant
level of processing. Each set of predictor neurons can explain only those particular
features or dimensions of the input, and correspondingly modulates the response in a
highly specific subset of error neurons. Error neurons are similarly distributed
throughout corticex, and respond to specific stimulus features (Meyer and Olson 2011;
den Ouden et al., 2012), rather than, for example, a single “error region” signaling the
overall amount of error or degree to which the observed stimulus is unpredicted (e.g.,
Hayden et al., 2011). Thus, for example, in the early visual cortex, predictor neurons
code information about the predicted orientation and contrast at a certain point in the
visual field, and error neurons signal mismatches between the observed orientation and
contrast and the predicted orientation and contrast. In IT cortex, predictor neurons code
information about object category; error neurons signal mismatches in predicted and
observed object category (den Ouden et al., 2012; Peelen and Kastner, 2011).
One consequence of this model is that, typically, the effects of predictions are limited to
relatively few levels of the processing hierarchy. To illustrate, expecting to see John walk
in the room would lead to predictions of biological motion, a body and a face (and
possibly to specific predictions within each of these domains), thus reducing error
responses in neural populations that respond at this level of abstraction. However, these
predictions often cannot be effectively or specifically translated into predictions at the
level of early visual receptive fields. Thus, while prediction signals may be passed down
the entire cortical hierarchy (Clark 2013; Rao and Ballard 1999), in many cases the
downstream transformation will make the signal too widespread to be informative. For
example, differential responses to predictable complex images have been observed in
monkey IT (Meyer and Olson 2011), and in human ventral temporal cortex (den Ouden
et al., 2010; Egner et al., 2010), without corresponding effects in lower visual areas.
Only when the environment supports specific, low-level predictions (on the scale of e.g.
125 of 179
orientation and contrast at specific points in retinotopy) should error signals be observed
at lower levels of processing (Alink et al., 2010; e.g., Murray et al., 2002; Weiner et al.,
2010).
When there is no relevant prediction available, error neurons act largely as “feature
detection” or “probabilistic belief accumulation” neurons (Drugowitsch and Pouget,
2012). This pattern highlights a key difference between predictive coding models
developed for sensory versus reward systems (den Ouden et al., 2012). Reward errors
are “signed”: the presence of an unexpected reward and the absence of an expected
reward are signaled by the same neurons changing their firing rates in opposite
directions. By contrast, sensory predictor neurons are likely “unsigned”: firing rates
increase in the presence of unexplained input.
Finally, our approach contrasts with other recent attempts to integrate social cognitive
neuroscience and predictive coding. Because predictive coding is most familiar from the
context of reward learning, there has been considerable interest in linking predictive
coding to social reward learning (Behrens et al., 2009; Jones et al., 2011; Fehr and
Camerer 2007). Social reward learning can mean either using social stimuli (e.g. smiling
faces) as rewards, or learning about rewards based on observation or consideration of
others’ experiences (Lin et al., 2012; Zhu et al., 2012; Zaki and Mitchell 2011; Poore et
al., 2012; Jones et al., 2011; Izuma et al., 2008; Chang and Sanfey, 2013; see Dunne
and O'Doherty 2013 for a review). Predictive coding may be an important mechanism
for motor control (i.e. anticipating, and explaining away, the consequences of one’s own
motor actions). Therefore some authors have linked motor predictions to social
predictions via the idea of “mirror neurons,” or shared motor representations for one’s
own and others’ actions (Brown and Brüne 2012; Kilner and Frith 2008; Patel et al.,
2012). The current proposal differs from both of these previous approaches by starting
with a hierarchical predictive coding framework developed for cortical visual processing
and by focusing on theory of mind, and specifically the attribution of internal states like
goals, beliefs, and personality traits.
This proposal is of course too general, and leaves many aspects of the model
unspecified (some of which we address below). Nevertheless, the basic features of
predictive coding described here provide an integrative framework for many findings in
the social cognitive neuroscience of theory of mind.
The sources of predictions
The social environment — the actions and reactions of other human beings — can be
predicted at a range of temporal scales, from milliseconds (where will she look when
the door slams?) to minutes (when she comes back, where will she search for her
glasses?) to months (will she provide trustworthy testimony in a court-case?). All of
these contexts afford predictions of a person's actions in terms of her internal states, but
the sources and timescales of the predictions are different. As we describe in the next
three sections, many experiments find that neural responses to predictable actions and
internal states are reduced, compared to unpredictable actions and states. This
common pattern can provide telling clues about the different types, and sources, of
126 of 179
predictions. We find that, while all regions show a higher response to unexpected
stimuli, what count as unexpected varies across regions and experiments, suggesting
that, at different levels of processing, neural error responses are sensitive to distinct
sources of social prediction.
To help clarify the sources of social prediction, we first review three sources of neural
predictions typically manipulated in visual cognitive neuroscience experiments. First,
given an assumption that the external world is relatively stable, neurons may predict
that sensory stimuli will remain similar over short timescales. Predictions based on very
recent sensory history can account for increased responses to stimuli that deviate from
very recent experience (Wacongne et al., 2012), and reduced responses to stimulus
repetition (Summerfield et al., 2008). Predictive coding may therefore offer an account
of widespread findings of repetition suppression in neural populations (Grill-Spector et
al., 2006). Predictive coding error is consistent with evidence that predictable repetitions
elicit more repetition suppression than unpredictable repetitions (Todorovic et al., 2011,
Todorovic and de Lange 2012).
Second, predictable sequences of sensory inputs can be created arbitrarily, through
training. For example, Meyer and Olson (2011) created associations between pairs of
images; for hundreds of training trials, image A was always presented before image B.
After training, the response in IT neurons to image B was significantly reduced when it
followed image A. This reduction was highly specific: the response remained high when
image B was presented alone, or following some other image, and there was no
reduction in the response to image A presented after image B. Other experiments show
that a tone can be used as a cue to predict the orientation of an upcoming grating (Kok
et al., 2012a), and either a colored frame or an auditory tone can predict whether an
upcoming image will be a face or a house (den Ouden et al., 2010; Egner et al., 2010).
The reliability of the cue can be stable over the experiment (Egner et al., 2010), or can
vary continuously across trials (den Ouden et al., 2010). In all cases, the magnitude of
neural responses tracks with the unpredictability of the stimulus, given the cue.
Perhaps most interesting, however, is the third source of predictions: an internal model
of the causal structure of the world that generated the observed input (Clark 2013;
Tenenbaum et al., 2011). For example, when two visual bars are presented in
alternating positions creating an illusion of motion, the visual system appears to
generate an internal model of a single object moving smoothly from one position to the
other across the intervening space. As a consequence, the addition of a third bar
presented at the right intervening space and time is treated as “predicted”, even though
that stimulus is otherwise unpredictable within the context of the experiment (Alink et al.,
2010).
In principle, all of three these sources of predictions can be applied to social prediction
and human actions. In practice, most of the experiments on theory of mind depend on
predictions based on prior expectations and an internal model of human behavior
(though we do find some evidence of predictions based on temporal proximity). Based
on the patterns of findings, we argue that these internal models must be quite abstract,
and include expectations that actions will be rational and efficient, and consistent with,
for example, the individual’s beliefs, personality traits, and social norms.
127 of 179
To reduce the complexity of this literature review, we focus here on three examples of
neural responses to actions at three conceptual levels: responses to biological motion
and goal-directed action in the superior temporal sulcus (STS), to other people’s beliefs
and desires in the temporo-parietal junction (TPJ) and to people’s stable personality
traits in the medial prefrontal cortex (MPFC; Figure 5.2). We find that, across all three
regions, with respect to the region’s preference and level of abstraction, expected
stimuli systematically elicit lower activation than unexpected stimuli.
Figure 5.2. Three brain regions involved in different aspects of theory of mind: examples of
individual regions of interest (ROIs). A region in posterior superior temporal sulcus (STS, green, peak
voxel [66, -36, 12]) involved in action perception (localized using biomotion relative to scrambled
biological motion, Pelphrey et al., 2003); a region in temporo-parietal junction (TPJ, blue, peak voxel
[62, -52, 18]) involved in thinking about beliefs and desires (localized using stories about mental states
relative to stories about physical events, Saxe and Kanwisher, 2003), and a region in medial prefrontal
cortex (MPFC, red, peak voxel [-2, 56, -4]) involved in thinking about people’s stable preferences and
personalities (localized using attribution of traits to self relative to attribution of traits to someone else,
Mitchell et al., 2006). All three ROIs localized using single subject data, p<0.001
Predicting goal directed action
The most immediate dimension of the social environment is the visibly observable
movements of other people’s bodies (e.g. grasping an item, running away) and faces
(e.g. gaze shifts, emotional expressions). Brain regions in the superior temporal sulcus
(STS) are implicated in many aspects of social action perception, showing robust
responses to face and body action in both humans (Jellema et al., 2000; Puce et al.,
1998; Bonda et al., 1996; Allison et al., 2002) and monkeys (Perrett et al., 1985; Jellema
and Perrett 2003b; Jellema and Perrett 2003a). Patients with lesions to the STS have
difficulty recognizing actions (Battelli et al., 2003; Pavlova et al., 2003), an effect that is
reproduced by creating reversible “lesions” in the STS through repetitive TMS
(Grossman et al., 2005). Consistent with a prediction error code, STS response to
observed actions is reduced when the observed action can be predicted, and enhanced
when the observed action is less predictable. These predictions appear to arise from a
variety of sources, ranging from experimental statistics, to constraints on biological
motion, to assumption about rational action, suggesting that rather than representing
low-level sensory-based statistics, this region represents (and makes predictions about)
coherent, rational actions.
128 of 179
First, like many sensory regions, the STS response is sensitive to the recent history of
the experiment and is reduced by repetition of a stimulus relevant to human action
perception. If two successive images of faces have the same gaze direction (i.e. both
gazing right) or the same facial expression (e.g. fearful), the STS response is reduced
compared both to a non-repeated presentation and to a repeated presentation of an
irrelevant stimulus, such as a house or object (Calder et al., 2007, Ishai et al., 2004, Furl
et al., 2007). Similarly, presenting the same action twice in row, from different viewing
angles, positions, sizes, and actors leads to reduced STS response relative to a
different action (Grossman et al., 2010; Kable and Chatterjee, 2006).
Human action can also be predicted based on internal models at many levels of
abstraction, from biomechanics to a principle of rational action. The most basic (and
most temporally fine-grained) predictions are constrained by the structure of bones and
joints and the forces exerted by muscles. Observers can thus predict the spatiotemporal
trajectory of human movements, especially for ballistic motions (Blake and Shiffrar
2007). Human movements that violate these biomechanical predictions (for example, a
finger bending sideways) elicit a higher response than more predictable movements in
the STS and related areas (e.g. Costantini 2005). Watching a human-like figure make
robot-like, mechanical movements elicits more activity than either a human-like figure
making human-like movements or a robot making mechanical movements (Saygin et
al., 2012).
Even when they do not violate biomechanical laws, human actions have a typical spatial
and temporal structure. Thus, if a person is walking rapidly across the room, we predict
that they will continue in the same trajectory, even if they are temporarily occluded. The
posterior STS responds more when the person reappears later than expected than
when the person emerges at the predicted time; when the person is replaced with a
passively gliding object, there is no effect of the time lag (Saxe et al., 2004).
In addition to intrinsic aspects of the action, observers expect others’ actions to be
temporally and spatially contingent on the structure of the environment. If a bright object
flashes near a woman’s head, she is very likely to immediately shift her gaze toward the
object. Seeing the woman immediately shift her gaze away from the bright object elicits
a higher response in the STS than the predicted gaze shift towards the object (Pelphrey
et al., 2003; Pelphrey and Vander Wyk 2011). This difference is reduced if the woman
first waits a few seconds before shifting her gaze, breaking the perception that that flash
caused the gaze shift. Similar effects are observed in infants as young as 9 months,
using EEG (Senju et al., 2006). In a more extreme mismatch between behavior and
environment, watching an agent twisting empty space next to a gear drives a stronger
STS response than the agent twisting the gear (Pelphrey et al., 2004).
Finally, the STS internal model of human behavior includes something like a principle of
rational action: the expectation that people will tend to choose the most efficient
available action to achieve their goal. The same action may therefore be predicted, or
unpredicted, depending on the individual’s goals and the environmental constraints
(Gergely and Csibra 2003). Correspondingly, the STS response is higher when the
same biomechanical action is unpredicted either because it is inefficient, or because it is
not a means to achieve the individual’s goal.
129 of 179
For example, action efficiency can be manipulated by having a person take a short or
long path to the same goal (Csibra and Gergely 2007), e.g. reaching for a ball efficiently,
arching her arm just enough to avoid a barrier, or inefficiently, arching her arm far above
the barrier. Across differences in barrier height and arm trajectory, activity in a region of
the MTG/STS is correlated with the perceived inefficiency of the action (Jastorff et al.,
2011). In a related experiment, observers watch someone performing an unusual action,
e.g. a girl pressing an elevator button with her knee. The context renders her action
more or less efficient: either her hands are empty, she is carrying a single book, or her
arms are completely occupied with a large stack of books. Activity in STS is highest
when the action appears least efficient, and lowest when the action appears most
efficient (Brass et al., 2007). The STS also responds more to failed actions (e.g., failing
to drop a ring onto a peg), an extreme form of inefficiency, than to successful ones
(getting the ring onto the peg, Shultz et al., 2011). Predictions for efficient action can
even be completely removed from the familiar biomechanics of human body parts: the
same inefficient action (going around a non-existent barrier) elicits stronger responses
in STS than the efficient version of the same action, when executed by a “worm” (a
string of moving dots, Deen and Saxe 2012).
In other experiments, the STS shows enhanced responses to actions that are
unpredicted given the individual’s specific goals, even if the action is not inherently
inefficient or irrational. For example, an individual who likes (and smiles at) a mug and
dislikes (and frowns at) a teddy bear can be predicted to reach for the mug and not the
bear. The goal-inconsistent action (reaching for the mug) elicits a higher response in the
STS (Vander Wyk et al., 2009). Similarly, when two people are cooperating on a joint
action, the STS shows increased responses when one person fails to follow the other’s
instructions: e.g. when asked to select one specific object (e.g., a red ball), the actor
takes the other object (e.g., the white ball; Shibata et al., 2011; see also Bortoletto et al.,
2011).
In sum, observers expect human movements to reflect “actions” which are sensitive to
the environment, and efficient means to achieve the individual’s goals. These
expectations can generate predictions for sequences of movements on the timescale of
seconds. All of these sources of predictions can modulate the neural response in the
STS, which is reduced when the stimulus fits the prediction.
Predicting beliefs and desires
Moving from the scale of seconds to the scale of minutes, the more general version of
the principle of rational action is that people will act efficiently to achieve their desires,
given their beliefs (Baker et al., 2011). Unlike specific motor intentions, beliefs and
desires last from minutes (e.g. the belief that your keys are in your purse) to years (e.g.
the desire to become a neurosurgeon). These beliefs and desires can be used to
predict aspects of a person’s actions, emotions, and other mental states, especially
when the person’s beliefs and desires differ from those of the observer (Wellman et al.,
2001; Wimmer and Perner 1983). Among other regions, a brain region posterior to the
superior temporal sulcus, in the temporo-parietal junction (TPJ), shows a robust
responses while thinking about an individual’s beliefs and desires (Saxe and Kanwisher
130 of 179
2003; Young and Saxe 2009a; Aichhorn et al., 2009; Perner et al., 2006). If the TPJ
includes a prediction error code, it should respond more strongly to beliefs and desires
that are unexpected, given the context. Indeed, there is evidence that the TPJ response
is reduced when a person’s beliefs and desires are predictable (though note that the
results reviewed in this section were generally not interpreted in terms of prediction
error coding by the original authors). In all of these experiments, the source of prediction
is not recent experimental history or trained associations, but rather a high level
generative model of human thoughts and behaviors.
One source of predictions about a person’s beliefs and desires is their actions (Patel et
al., 2012). Observers expect other people to be self-consistent and coherent (e.g.
Hamilton and Sherman 1996). This sensitivity to inconsistencies in belief and action is
reflected in the TPJ. For example, in one group of studies, participants read about an
act of violent harm or murder. Under the assumption that people usually act in
accordance with their beliefs (Malle 1999), the prediction is that the perpetrator intended
the harm; most assaults and murders are not accidental. Next, the participants read
about the perpetrator’s actual beliefs and desires. Responses in the right TPJ are higher
for “unpredicted” innocent or benevolent intentions that exculpate the harm (e.g. she
believed the poison was sugar; he only wanted to end the patient’s misery from an
incurable disease) compared to the “predicted” intention (to kill the person; Buckholtz et
al., 2008; Chapter 2, Koster-Hale et al., 2013; Yamada et al., 2012; Young and Saxe,
2009b).
Not all actions imply the corresponding intention, however: for example, violation of
social norms (e.g., spitting out a friend's cooking back on your plate) are more likely to
be committed accidentally than intentionally. Consistent with a prediction error code, the
TPJ response is higher for violations of norms performed intentionally ("because you
hated the food") versus unintentionally ("because you choked"; Berthoz et al., 2002).
In addition to these general principles, an individual’s beliefs and desires can
sometimes be predicted based on other information you have about his or her specific
group membership and social background. For example, Saxe and Wexler (2005)
introduce characters with different social backgrounds, ranging from the mundane (e.g.
New Jersey) to the exotic (e.g. a polyamorous cult). Participants then read about that
character’s beliefs and desires (e.g. a husband who believed it would be either fun or
awful if his wife had an affair). The response in right TPJ is reduced for the belief that
was predictable, given the character’s social background: the person from New Jersey
thinking his wife having an affair would be awful, and the person from the polyamorous
cult thinking his wife having an affair would be fun. Similarly, when reading about a
political partisan, political beliefs that are unexpected, given the individual’s affiliation
(e.g. a Republican wanting liberal Supreme Court judges) elicits a higher response in
right TPJ (Cloutier et al., 2011).
On the other hand, the general plausibility of a belief, in the absence of specific
background information about the individual, does not seem to be sufficient to generate
a prediction (or a prediction error) in the right TPJ. Without specific background
information about the believer, there is no difference in the right TPJ response to absurd
versus commonsense beliefs (e.g. “If the eggs are dropped on the table, Will thinks
131 of 179
they’ll bounce / break,” (Young et al., 2010), although the participants themselves rated
the absurd beliefs significantly more “unexpected.” A possible interpretation of these
results is that the internal model of the RTPJ predicts another person’s beliefs based on
expectations of a coherent individual mind, using information about that individual’s
specific actions and history, but not based on expectation that others will share one’s
own beliefs (see below for a contrast with MPFC) or common knowledge. However,
since these are null results, they should be interpreted with caution.
In sum, the response in the TPJ to other people’s beliefs and desires can be modulated
by how predictable those beliefs and desires are, relative to the current environment,
the individual’s actions, broader social norms, and the individual’s specific social
background.
Predicting preferences and personalities
At even longer timescales, successful prediction of the social environment depends on
building distinct models of each of the individual humans who compose one’s social
group. While some general rules, like the principle of rational action, apply to all people,
predicting a specific person’s action often depends on knowing the history and traits of
that individual. Brain regions on the medial surface of cortex, in both medial prefrontal
(MPFC) and medial parietal (PC) cortex, show robust responses while thinking about
people’s stable personalities and preferences (Mitchell et al., 2006; Schiller et al., 2009;
Cloutier et al., 2011). Consistent with a predictive error code, these responses are
reduced when new information about a person can be better predicted. Again these
predictions appear to be derived from relatively high level expectations that people’s
traits will be consistent across time and contexts, rather than from local experimental
statistics.
Prior knowledge of a person can be acquired through direct interaction. First person
experience of another person’s traits (e.g. trust-worthiness, reliability), can be
manipulated when participants play a series of simples ‘games’ with one or a few other
players. By gradually changing the other players’ behaviors, it is possible to create
parametric “prediction errors”. In one experiment, for example, the other player provided
“advice” to the participant; this advice shifted over the experiments, so that it was
reliable in some phases, and unreliable in others. The response in MPFC tracks with
trial-by-trial error in expectations about the informant’s reliability (Behrens et al., 2008).
Expectations about other people’s traits can also be based on verbal reports and
descriptions. For example, the initial behaviors of a (fictional) stranger can create an
impression of a certain kind of personality (e.g. ‘Tolvan gave her brother a compliment’).
The MPFC response is enhanced when later actions by the same person are
inconsistent with (i.e. unpredicted by) this trait (e.g. ‘Tolvan gave her sister a slap’)
compared to when they are predictable (e.g. ‘Tolvan gave her sister a hug‘; Ma et al.,
2012; Mende-Siedlecki et al., 2012).
When specific information about a person’s reputation or traits is unavailable, we may
predict others’ preferences by assuming that they will share our own preferences
(Krueger & Clement 1994; Ross, Greene, & House 1977). In one series of studies
(Tamir and Mitchell 2010), participants judged the likely preferences of strangers (e.g. is
132 of 179
this person likely to “fear speaking in public” or “enjoy winter sports”?) about whom they
had almost no background information. Under those circumstances, the response of the
MPFC was predicted by the discrepancy between the attributions to the target and the
participant’s own preference for the same items: the more another person was
perceived as different from the self, for a specific item, the larger the response in MPFC.
In all, human observers appear to formulate predictions for other people’s movements,
actions, beliefs, preferences and behaviors, based on relatively abstract internal models
of people’s bodies, minds, and personalities. These predictions are reflected in multiple
brain regions, including STS, TPJ, and MPFC, where responses to more predictable
inputs are reduced, and to less predictable inputs are enhanced.
Consistent with our general proposal for prediction error coding, reduced responses to
predicted stimuli in these experiments are typically restricted to relatively few brain
regions, and by implication, to relatively few levels of the processing hierarchy. Beliefs
or actions that are unpredicted, based on high level expectations, do not elicit enhanced
responses at every level of stimulus processing (e.g. early visual cortex, word form
areas, etc). Nor are prediction errors signaled by a single centralized domain general
“error detector”. Instead, relatively domain- and content-specific predictions appear to
influence just the error response at the relevant level of abstraction.
Beyond error
In sum, human thoughts and actions can be rendered unexpected in many ways, and
across many such variations a common pattern emerges: brain regions that respond to
these stimuli also show enhanced responses to “unexpected” inputs. This profile is the
classic signature of error neurons, and therefore consistent with a predictive coding
model of action understanding.
While consistent with predictive coding, however, these results provide only weak
evidence in favor of predictive coding. Increased responses to unexpected stimuli can
be explained by many different mechanisms, including increased ‘effort’ required,
increased attention, or longer evidence accumulation under uncertainty. The predictive
coding framework will therefore be most useful if it can make more specific predictions
and suggest new experiments.
Distinguishing prediction from attention
A salient alternative explanation for enhanced responses to unpredicted stimuli relies on
attention. Unexpected stimuli may garner more attention, and increased attention can
lead to more processing and higher activation (e.g., Bradley et al., 2003; Lane et al.,
1999). Similarly, increased processing effort or longer processing time can predict
higher activation (e.g. Cohen et al., 1997). Thus, higher activation to unexpected stimuli
could reflect greater attention or longer processing, rather than prediction coding errors.
However, relative to these accounts, predictive coding has a distinctive signature.
By hypothesis, predictions codes are more precise, more computationally efficient, and
less noisy than error codes (Friston 2005; Jehee and Ballard 2009; Rao and Ballard
1999; Spratling 2008). As a result, in a predictive coding model, better speed and
133 of 179
accuracy of perception are associated with reduced overall neural responses to
predicted stimuli (Kok et al., 2012a; den Ouden et al., 2008). By contrast, attention may
cause better speed and accuracy of performance by increasing overall neural
responses to attended stimuli (Feldman and Friston, 2010; Friston, 2010; Herrmann et
al., 2010; Hillyard et al., 1998; Kok et al., 2012b; Martinez-Trujillo and Treue, 2004;
Reynolds and Heeger, 2009; Treue and Martínez-Trujillo, 1999). That is, whereas
attention may increase gain in neural responses to the attended stimulus, predictions
improve perception by decreasing noise (or increasing sparseness) in neural responses
to the predicted stimulus.
If the neural responses described in the previous sections reflect prediction error,
reduced neural responses should be accompanied by improvements in behavioral
performance: people should make judgments more quickly, with less error, and with
more sensitivity on expected stimuli. Indeed, behavioral evidence suggests that
observers make faster and more accurate judgments about people who behave as
expected in social contexts. After watching two people engage in part of a cooperative
action or conversation, participants are faster and more accurate when both agents are
behaving as expected (e.g., responding aggressively or cooperatively, responding
communicatively or non-communicatively, or right away, instead of too early or late;
Manera et al., 2011; Neri et al., 2006; Graf et al., 2007). Important next questions will be
to look for these signatures in other aspects of social cognition, such as goal inference
or belief attribution.
Figure 5.3. Predicted stimuli elicit reduced activity with sharpened representations. Kok et al.
(2012a) report that an accurate prediction led to reduced response amplitude in early visual cortex, but
also simultaneously to an “improved” stimulus representation, as measured by multi-voxel pattern
analysis. Consistent with this suggestion, we find that in the right temporo-parietal junction (RTPJ; right
box) the response amplitude to a predicted belief was lower, but the spatial pattern associated with that
belief category was more reliable. Left: a sample stimulus. All stories described first a harmful action,
and then the agent’s belief. The “predicted” belief (solid arrow) was consistent with the action (i.e.
making the act an intentional harm). The “unpredicted” belief (dotted arrow) was inconsistent and
rendered the harm an accident. Middle: The amplitude of response in the TPJ was lower for the
intentional than accidental condition. Right: The spatial pattern of response in the TPJ was most robust
and reliable across trials for intentional harms, and somewhat less reliable for accidental harms. Data
from Chapter 2 and Koster-Hale et al. (2013).
134 of 179
An interesting extension of this idea is the proposal that the sparser prediction signal
should also be easier to decode from a neural population than the more distributed error
signal, within a single region and task (Kok et al., 2012a; Sapountzis, Schluppeck,
Bowtell, & Peirce, 2010). In an elegant study, Kok et al. (2012a) asked participants to
make fine perceptual discriminations between oriented gratings. They hypothesized that
when the orientation of the gratings was accurately predicted by a cue, the
representation of the grating would be largely in the sparser predictor neurons, whereas
when the orientation was not accurately predicted (i.e. on the relatively rare invalidly
cued trials), then the representation of the orientation would be largely in the more
distributed error neurons. Three predictions of their model were confirmed in the
responses of early visual cortex. First, the overall response to the gratings was lower
when the orientation was predicted than when it was unpredicted (the classic pattern of
“explaining away” the error signal). Second, behavioral discriminations on the gratings
were more accurate when the orientation was predicted than when it was unpredicted,
consistent with the hypothesis of a more efficient code. Third, and critically, the
orientation of the gratings could be more easily decoded from the spatial pattern of
neural responses in early visual cortex when the orientation was predicted than when it
was unpredicted, consistent with the hypothesis that the prediction signal is spatially
sparser than the error signal. Finally, Kok et al. (2012a) distinguished the effects of
prediction from effects of attention, by manipulating the participants' task. Directing
attention to the gratings’ orientation (versus contrast) improved decoding of orientation
in V1, but the effects of attending to orientation, and of seeing the unpredicted
orientation, were independent and additive.
A corresponding hypothesis should be easy to test with respect to the neural
representation of human behaviors, thoughts and personalities. The lower responses to
expected stimuli should be accompanied by better decoding of relevant stimulus
dimensions. Indeed, our own results from the TPJ are consistent with this hypothesis.
As described above, when reading about harmful actions (e.g. putting poison powder in
someone’s coffee), the TPJ response is higher to “unpredicted” innocent beliefs (e.g.
that the powder was sugar) than to “predicted” consistent beliefs (e.g. that the powder
was poison; Young and Saxe 2009b). We also found that using spatial pattern analysis
in the TPJ, we could decode the difference between innocent and guilty beliefs (Chapter
2, Koster-Hale et al., 2013). Based on Kok et al. (2012a), a further prediction is that the
decoding should be driven by a sparser and more efficient response to the predicted
category; and indeed, re-analysis of our data suggests that the guilty beliefs elicit a
more distinctive (i.e. more correlated across trials) spatial pattern than the “unpredicted”
innocent beliefs (Figure 5.3).
Interestingly, the benefits of an accurate prediction may be quite specific to the aspects
of the stimulus that are accurately predicted. As we suggest earlier, most predictions are
limited to a particular level of abstraction; given a high-level prediction, the probability of
lower-level features appearing will be too widely distributed to be informative. As a
result, accurate predictions may improve behavioral performance (and neural decoding)
at the representational level of the prediction (e.g. which object a person wanted) but fail
135 of 179
to improve, or even degrade, these measures for lower-level features (e.g. where in
space someone looked; He et al., 2012).
Finding the prediction signal
An important direction for future research will be to focus on signatures of the predictor
neurons, in addition to the error neurons. At least four different strategies may help to
identify prediction signals, and distinguish them from the often more dominant error
signals.
First, unlike error neurons, predictor neurons should show robust activity when the
stimulus fits prior predictions. Consistent with this suggestion, a recent study found that
although the majority of voxels in the fusiform face area (FFA, Kanwisher et al., 1997;
Kanwisher 2010) was suppressed for a repeated face, a subset of voxels reliably
showed the reverse pattern (de Gardelle et al., 2012), termed repetition enhancement
(see also Turk-Browne et al., 2006; Müller et al., 2013). Intriguingly, these two
populations of voxels also showed different patterns of functional connectivity. It will be
intriguing to test whether the STS, TPJ, PC, or MPFC similarly contain subsets of voxels
with enhanced responses to predicted actions or beliefs, and whether these voxels have
distinctive patterns of functional connectivity with other regions, especially because
unlike face processing, the direction of information flow among regions involved in
theory of mind is largely unknown.
Second, because both predictor neurons and error neurons may have preferred stimuli
(or stimulus features), it may be possible to identify the content of the prediction
independent from the response to the subsequent stimulus. For example, the response
of the FFA seems to increase when a face stimulus is predicted, as well as (and partially
independent from) when a face stimulus is observed (den Ouden et al., 2010; Egner et
al., 2010). Note though that neither of the existing studies could fully independently
identify the response to predicting a face, because in both cases, the probability of a
face was exactly reciprocal to the probability of the only other possible stimulus, a
house. By including a third category of stimulus, or a third possible cue, or by
independently varying the predictive value of the two cues, it should be possible to
independently measure category-specific responses to the prediction of a category,
versus the response to that category when observed.
Third, and relatedly, predictor neurons can signal the expectation of a stimulus that
never occurs. In some cases, the absence of an expected stimulus should generate
error activity as (den Ouden et al., 2010; Todorovic et al., 2011; Wacongne et al., 2011).
For example, the activity pattern in IT generated by the surprising absence of an object
contains information about the identity of the absent stimulus (Peelen and Kastner
2011). Unlike the “signed” (i.e. below baseline) error response in reward systems,
sensory neurons thus seem to show an increased response to an unexpectedly absent
stimulus (though note that there is some disagreement as to whether this activity is
driven only by the prediction signal before the stimulus is expected to appear, or by a
combination of the prediction signal with a subsequent error signal when the stimulus
fails to appear, e.g., den Ouden et al., 2010).
136 of 179
Fourth, the prediction and the error signals could be separable in time. Specifically,
under some circumstances prediction signals can begin before the stimulus, whereas
error signals are typically triggered by the stimulus itself (e.g., Hesselmann et al., 2010).
The low temporal resolution of fMRI may make it hard to test this hypothesis directly.
However one pattern of results is consistent with the idea that the STS contains
predictions of upcoming biological motion: still photographs of a person in mid-motion
(such as a discus thrower in the middle of throwing a disc) elicited more activity in the
STS than images that do not imply or predict motion (the same discus thrower at rest;
Kourtzi and Kanwisher 2000; Senior et al., 2000).
Fifth, error responses in a single region may be influenced by predictions from different
sources, and these different sources may be spatially separable. For example, FFA
shows repetition suppression for both repetition of one identical face image (plausibly a
very low-level prediction) and for repetition of a face across different sizes (requiring a
higher-level prediction). These error signals were related to different patterns of
functional connectivity between FFA and lower level regions (Ewbank et al., 2013). By
analogy, there may be different patterns of functional correlations related to different
sources of prediction for human actions. In one experiment, for example, the STS
response was enhanced for actions that were unpredicted for two different reasons:
reaching for empty space next to a target (which is an inefficient or failed action), or
reaching for a previously nonpreferred object (which is unpredicted relative to an
inferred goal; Carter et al., 2011; see also Bubic et al., 2009). It would be interesting to
test whether these two kinds of errors are associated with spatially distinct sources of
functional connectivity to the STS.
Using predictive coding to test the neural computations underlying
theory of mind
The framework of predictive coding offers a new opportunity to study the neural
representations of others’ actions and thoughts, using new experimental designs. The
necessary logic has been developed in repetition suppression experiments (GrillSpector et al., 2006). Complex stimuli elicit responses in many different brain regions
simultaneously, making it hard to dissociate the representational and computational
contributions of different brain regions. Consequently, in higher level vision, repetition
suppression has been used to differentiate the stimulus dimensions or features
represented in multiple co-activated regions. For example, although both the FFA and
the STS face area show repetition suppression when the identity of a face is repeated,
only a more anterior STS region shows a reduced response when the emotional
expression is repeated across different faces (Winston et al., 2004).
Looking for prediction error offers a generalized, and more flexible, version of repetition
suppression studies; critically, it only requires that a stimulus be surprising along some
dimension, without having to repeat the stimulus. This flexibility is particularly
advantageous for studies using naturalistic or social stimuli, which can be hard to repeat
without also invoking repetition of a number of confounding, but unrelated
representations — for example, words, syntactic structure, or faces. Exact repetitions of
complex stimuli can be unnatural or pragmatically odd, which may especially limit the
137 of 179
ability to study repetition suppression in young or special populations. By contrast, the
distribution of observed error signals could reveal both which neural populations or
regions are coding the relevant dimensions and features, and what the sources of
predictions are.
Finally, and perhaps most importantly, this framework may enrich theorizing about
neuroimaging results in social cognitive neuroscience. One of the key challenges facing
social cognitive neuroscience is that the richness of the data often surpasses the
precision of the theories. This proves to be a problem both for interpreting the data —
inverse inferences are very rarely well-constrained enough to be compelling, despite
their role in theory building — and for designing new hypotheses and experiments.
Increased response in a brain region has been argued to indicate both that the stimulus
carries many relevant features to a region and that the stimulus was harder to process
or a less good “fit” to the region; this problem is exacerbated when trying to interpret
different neural patterns across groups (i.e. special populations). If we can begin to
break down (a) what kinds of predictions a region makes, (b) what kind of information
directs those predictions, and (c) what constitutes an error, it may be possible to
formulate much more specific hypotheses about the computations, and information flow,
that underly human theory of mind.
In sum, we find a predictive coding approach to theory of mind promising. There is
extensive evidence of a key signature of predictive coding, in fMRI studies of theory of
mind: reduced responses to expected stimuli. Existing data also provide hints of other,
more distinctive signatures of predictive coding. Future experiments designed to more
directly test the predictions and errors represented in different brain regions may
provide an important new window on the neural computations underlying theory of mind.
Acknowledgements
The authors thank Amy Skerry, Hilary Richardson, Todd Thompson, and Nancy
Kanwisher for comments and discussion. The authors gratefully acknowledge support of
this project by an NSF Graduate Research Fellowship (#0645960 to JKH) and an NSF
CAREER award (#095518), NIH (1R01 MH096914-01A1), and the Packard Foundation
(to RS).
138 of 179
Chapter 6.
General Discussion
Summary
Thinking about another’s thoughts increases metabolic activity in a reliable set of
cortical regions, often called the theory of mind network, which includes bilateral
temporo-parietal junction (right and left TPJ), right superior temporal sulcus (right STS),
medial precuneus (PC), and medial prefrontal cortex (MPFC). These regions show
robust and systematic responses to a variety of stimuli that invoke a mental state
attribution (Fletcher et al., 1995; Goel et al., 1995; Gallagher et al., 2000, 2002; Mitchell
et al., 2002; Saxe & Kanwisher, 2003; Perner, Aichhorn, Kronbichler, Wolfgang, &
Laddurner, 2006; Gobbini, Koralek, Bryan, Montgomery & Haxby, 2007; Van Overwalle,
2009; Walter et al., 2010), and respond robustly to mental state information in
comparison both to other social information, including appearance, cultural background,
and subjective sensation such as exhaustion or hunger (Saxe & Powell, 2006; Saxe &
Wexler, 2005), and to non-mental representations, including misleading signs, out-ofdate maps, and temporary changes in reality (Saxe & Kanwisher, 2003; Perner, et al.,
2006). While many studies have investigated the selectivity and domain specificity of
the brain regions for theory of mind (e.g., Aichhorn et al., 2009; Saxe & Kanwisher,
2003), a key challenge is to go beyond basic involvement to determine the
computational roles of these regions: which aspects of people’s beliefs and intentions
are represented, or made explicit, in these brain regions, and how are they coded?
The goal of this thesis is to lay the theoretical and empirical groundwork for
systematically characterizing the representations and computations in mental state
inference. Building on prior work, which has for the most part focused on where in the
brain mental state reasoning occurs, the research here investigates how neural
populations encode concepts underlying mental state inference. In Chapter 1, I
reviewed existing literature on the neuroscience of mental state inference, to argue that
an important next step in understanding the neural basis of social cognition is to
understand the neural representations supported by “social” brain regions. In Chapters
2, 3, and 4, I presented seven empirical studies that demonstrate the successful use of
modern neuroimaging techniques to find abstract and behaviorally relevant neural
representations inside ToM brain regions, and begin to characterize the content of these
representations. In Chapter 5, I offered a novel interpretation of the existing literature
based on a predicative coding model of the brain. Together, this work takes a first step
in characterizing the content and format of abstract, behaviorally relevant features of
mental state inferences inside cortical regions that support social cognition.
139 of 179
Where We Are
Content in theory of mind brain regions
One of the major accomplishments of this thesis is to provide a first characterization of
the content represented by brain regions that support mental state inference. In three
chapters, and seven empirical studies, I present evidence that the neural code in brain
regions supporting social cognition is sensitive to difference types of mental state
representations, rather than simply distinguishing mental state atrributions from other
types of information.
In Chapter 2, I show that brain regions supporting social cognition distinguish between
mental state representations, that this code is detectable using fMRI, and that these
neural representations are behaviorally relevant. With different stimuli, paradigms, and
participants, I find converging results across three experiments: in neurotypical (NT)
adults, stories about intentional versus accidental harms elicited spatially distinct
patterns of response within the right temporo-parietal junction (RTPJ). Moreover, this
neural response mirrored behavioral judgments: individuals whose patterns in RTPJ
were more distinct also made a larger distinction between intentional and accidental
harms in their moral judgments. Notably, in the fourth experiment, this pattern was
absent in adults with autism spectrum disorders. Adults with ASD are characterized by
persistent social difficulty, including in using mental state information to reason about
the actions of others. Mirroring group-level differences in behavior, we found no neural
distinction between accidental and intentional harms in pattern or mean signal in adults
with ASD. These data suggest that one aspect of the role of RTPJ is to make explicit, in
the population response of its neurons, aspects of beliefs that are most relevant for
inference and decision-making — for example, encoding the intent behind a harmful act.
In Chapter 3, I find that information about the epistemic history of other’s beliefs is
represented in RTPJ, and that these neural representations are robust to profound
differences in developmental history. I find that the TPJ encodes information about the
source (specifically, the sensory modality) of others’ mental states, and that these
representations are preserved in the RTPJ of congenitally blind adults. The existence of
these representations in congenitally blind adults, who have no first-person experience
with sight, provides evidence that representations of other’s epistemic history emerge
even in the absence of relevant first-person perceptual experiences. Moreover, it
suggests that the coding of this information must be relatively abstract, rather than tied
to a single type of input or grounded in a specific sensory domain. Together, these
results suggest that the temporo-parietal junction contains explicit, abstract
representations of another feature mental states: the perceptual source.
In Chapter 4, I show that the neural representations of source modality found in
Chapter 3 are part of a more complex feature space of mental state representation, and
are coded along independent, continuous dimensions. I probed three epistemic features
of others’ beliefs (distinctions relevant to the source of another person’s belief), and one
salient but non-epistemic distinction (the valence of the person's resulting emotion). In
two experiments, I find that RTPJ and RMSTS contain complementary, continuous
neural codes which track the degree to which someone’s evidence supports their
140 of 179
beliefs, reliably distinguishing between good, reliable evidence and obscured,
ambiguous evidence; that RTPJ and LTPJ encode information about the modality of
other people’s evidence; and that the neural pattern in VMPFC codes continuous
information about the valence of other people’s experiences. These results suggest that
the cortical regions implicated in mental state inference contain neural representations
of epistemic and emotional features of others’s beliefs, and that these features are
represented along continuous, abstract dimensions.
The data in these chapters suggest that the brain regions implicated in social cognition
contain explicit neural representations of abstract aspects of others’ beliefs, including
epistemic and emotional features, and that these features are represented in a
continuous, abstract feature space. The neural code in RTPJ (a region implicated in all
seven studies) in particular seems to contain information about the beliefs of others,
tracking information about the relationship of someone’s beliefs to their actions (intent),
the reliability of the evidence they based their beliefs on (evidence quality), and the
source of their information (modality). Right STS was also found to contain
representations that support tracking the epistemic features of others’ beliefs,
specifically the quality of others’ evidence, whereas Left TPJ contains information about
the source of someone’s belief — in particular, the modality through which the evidence
was acquired. Finally, VMPFC (unlike RTPJ), contains continuous representations about
how people feel, based on the beliefs they hold.
Finally, in Chapter 5, I take a step back to consider how we might interpret the
activation observed in each of these regions. I propose that an existing framework of
neural coding, predictive coding, can provide a unifying framework for interpreting —
and generating new hypotheses about — the kinds of representations we find in social
brain regions. Predictive coding models posit that the brain generates continuous
predictions of upcoming events, and then adjusts these predictions by computing an
error signal which tracks deviations between predicted and observed events. Human
observers appear to formulate predictions for other people’s movements, actions,
beliefs, preferences and behaviors, based on relatively abstract internal models of
people’s bodies, minds, and personalities. I argue that these predictions are reflected in
multiple brain regions, including RSTS, RTPJ, and MPFC. I observe that across a
variety of tasks and studies, the RSTS responds more to observed actions that are
unpredicted; the RTPJ appears to exhibit a heightened response to unpredicted
information about others’ mental states; and MPFC responds more to unpredicted
information about a person’s preferences or traits. In sum, there is extensive evidence
of a key signature of predictive coding, in fMRI studies of theory of mind: reduced
responses to expected stimuli. I propose that future experiments designed to more
directly test the predictions and errors represented in different brain regions may
provide an important new window on the neural computations underlying theory of mind.
A feature-based approach to mental state representations
This work has largely looked for “features” of mental states — is a particular mental
state representation “intentional” or “accidental”, “reliable” or “unreliable”, “happy” or
“sad? I find that theory of mind brain regions, and in particular, RTPJ, represent features
141 of 179
of others’ beliefs — in particular, features of the relationship between the content of
someone’s beliefs and other relevant information, such as the agent’s actions (intent),
and the evidence available in the world (source modality and evidence quality).
Interestingly, the one feature that I find outside of RTPJ (valence), is not a feature of a
relationship between a belief and another piece of information, but rather an outcome of
an agent having had a belief. Based on these data, an intriguing hypothesis is that one
of the main functions of lateral ToM brain regions, and perhaps the RTPJ specifically, is
to track relationships between beliefs and other concepts, while medial brain regions
track information about agents and individuals. If this is the case, we might expect that
information qualifying the relationship between an agent and their belief, such as
confidence (guesses versus believes), factivity (knows versus thinks), negation
(believes versus doubts), or qualifying relationships between multiple agents and beliefs
(agrees versus disagrees) would also be found in RTPJ, while information about agents,
such as reliability and trustworthiness, would be found in MPFC. If this is correct, a key
role of reasoning about the minds of the others may the ability to track the relationships
between pieces of information, such agents, attitudes, actions, and real-world evidence,
and one role of theory of mind brain may be to track these relationships.
However, a limitation to this work is that, so far, I have looked only for specific types of
features: the relationships between belief content and other components of mental state
attributions, rather than features of, for example, specific agents or belief content. How
can we move from testing two or three features to genuinely looking for a compositional,
flexible, feature space?
First, we can look for features of other parts of mental state attributions, for example
features of the agents holding the beliefs, and features of the belief content. Previous
work suggests that features of the agent might be represented in MPFC (WhitfieldGabrieli et al., 2011; Moran et al., 2011a; Saxe, Moran, Scholz, & Gabrieli, 2006a;
Krienen, Tu, & Buckner, 2010) and anterior temporal lobe (Zanh et al., 2007; Anzellotti,
et al., 2013). Interestingly, however, representations about the content of others’ belief
may not be represented by regions classically associated with mental state inference.
For example, we can reason about many different properties of the proposition The Red
Sox will win the World Series, (e.g. our confidence and emotional response, or its
grammaticality, tense, and likelihood of being true) without having to consider the
information in the context of someone else’s mind. It is possible therefore, is that the
content of mental state representations is not represented in regions of the brain that
specifically support mental state inference, but rather in regions, such as language
regions, that more generally support reasoning about propositions (e.g. Friederici,
2002). If that is the case, we would expect that features of belief content would appear
in the same regions that support other semantic distinctions and processing of
propositional content (Fairhall et al, 2013; Mahon & Caramazza, 2010; Binney et al,
2010; Visser et al. 2012).
A second approach is to use “data-driven” analyses to find potential neural feature
spaces. This approach, which has been applied recently in a number of domains, such
as object recognition (Kriegeskorte, et al., 2008), audition (Norman-Haignere et al.,
2014) and action observation (Deen & Saxe, in prep), takes the neural response to a
142 of 179
diverse and relatively uncontrolled set of stimuli, and asks which classes of stimuli elicit
similar responses, using a variety of dimensionality reduction techniques. The
advantage of this approach is that it can find both a small number of dimensions that
explain a good portion of the variance in their stimuli set, and possibly find new, unhypothesized dimensions in the data. These dimensions can then be used to generate
new hypotheses, to be tested in follow-up experiments. This technique may reveal
important dimensions in the neural representations of social cognition. However, it also
has clear drawbacks. For example, when a dimension emerges that is not interpretable,
it is difficult to know if the researcher picked the wrong clustering methods, or if the
brain is grouping information along dimensions that human minds haven’t yet thought
of. Moreover, it can be difficult to measure whether the chosen stimulus set exhaustively
covers the cognitive or neural domain (and thus whether the features discovered truly
explain most of the variance in a given domain). Nevertheless, a combination of datadrive approaches and hypothesis-driven approaches, may be the key next steps in
characterizing the broader feature space of mental state representation.
Beyond Features
Feature-based approaches have proven invaluable in a number of cognitive and neural
domains, including vision (e.g. Shepard, 1980; Serre et al, 2005; Harrison et al, 2009),
memory, (e.g. Eichenbaum, 2004) and language (e.g. Naselaris et al, 2011), concept
representation (Mahon et al., 2010; Peelen et al., 2012). However, this approach also
has clear limitations. It is unlikely that mental state attributions are best described as
simply a list of features; rather, mental state attributions are likely representations with
internal structure (Baker at al, 2009, 2011; Davidson, 1963; Fodor 1981, 1987; Chen
2009), composed, for example, of an agent (the belief holder, e.g. John), the content of
the belief (e.g. it is raining outside), an attitude (the relationship between the agent and
the belief content, e.g. suspects), and perhaps in some cases, a representation of the
world outside of the agent’s beliefs (e.g., But it is actually sunny). These structures show
the hallmarks of being the product of a compositional, generative process (Chomsky,
1995): agents, attitudes, and beliefs can be combined in novel ways to form coherent
mental state representations (Snuffleupagus doubts that Red Sox will win the World
Series), and mental states can be embedded in one another (Snuffleupagus doubts that
Elmo believes that the Red Sox will win the World Series). Purely feature-based
approaches will fail to capture many of these dimensions, in particular hierarchical and
causal relationships.
Nevertheless, a feature-based approach has proven to be a fruitful first step in
characterizing the representations inside ToM brain regions. Recent computational work
has worked on combining basic “bags of features” approaches with structured
representations, which have increased predictiveness and explanatory power in a
number of domains (e.g. Kemp & Tenebaum 2008; Lazebnik et al, 2006). Given the
success of a feature-based approach here, and in much of higher-level cognitive
neuroscience, exploring models that use features to infer underlying structure may
prove invaluable to answering some of the harder questions about the cognitive and
neural representations of abstract concepts.
143 of 179
Continuous, abstract representations
One way to leverage information about features to infer more about the underlying
structure is to look past simple binary tags. Here, I ask whether the dimensions we
detect using feature-based approaches are binary or gradient. In Chapter 4, I show that
the probability of an item being categorized as one type of stimulus, based on neural
data, is predicted by independent behavioral ratings, suggesting that the brain may
represent these stimuli along continuous, abstract feature dimensions (Kriegeskorte,
2008; Kriegeskorte & Kievit, 2013), rather than with binary tags. Interesting open
questions are whether these regions contain both continuous and discrete dimensions
(for example, perhaps modality is coded discretely), and to what extent they are fully
combinatorial: does every possibly combination of features get represented?
Though these results just scratch the surface of what can be learned about the format of
representations in ToM regions, taken together, they are particularly striking because
they find evidence of neural representations of highly abstract conceptual features, and
thus add to the growing body of literature that has found neural representations of a
variety of high-level cognitive domains (e.g., Anzellotti et al., 2013; Hassabis et al.,
2013; Behrens, Hunt, & Rushworth, 2009; Izuma, Saito, & Sadato, 2008; Poore et al.,
2012; Correia et al., 2014).
Inputs to representations
A second way to leverage feature-based approaches to understand underlying structure
is to examine the information that serves as input to these features, and ask to what
extent the features are invariant to changes in that input. Finding that a distinction
between two features is represented in a neural code fails to tell us about its underlying
structure. For example, evidence that there is a neural distinction between knowing that
someone learned a fact by seeing it as opposed to by hearing it could indicate a
distinction between types of sensory simulations — a “re-experience” of that person’s
hearing or seeing. It could also map onto a more abstract, conceptual distinction, e.g., a
model of perceptual access that encodes that some information can be acquired
through one type of source, and not another. We can begin to distinguish between these
two different types of underlying structure by asking about the kind of information that
feeds these representations: is the input necessarily first-person sensory motor cues,
or, does it come from a broader range of sources? Here, I ask whether the features
found in ToM brain regions are tied to specific sources of information (for example,
sensory-motor input) by asking whether the existence of these representations depends
on having specific machinery or experiences.
In Chapter 3, I find that the RTPJ in blind adults contains distinct neural representations
of seeing and hearing. The developmental history that supports (or limits) a neural
representation can tell us information about its source. By asking whether a neural
representation has been changed in individuals who have different basic building
blocks, such as those with different sensory experiences, we can place important limits
on the possible input of the representation. By asking how blind individuals represent
others’ experiences of sight, for example, we can learn the extent to which theory of
mind depends on the observer having had similar sensations, experiences, beliefs, and
feelings as the target, and constrain our theories of mental state inference accordingly.
144 of 179
Finding robust representations of others’ experience of sight in the RTPJ of blind adults
suggests that, much like representations of value in pre-frontal cortex (e.g. O'Doherty et
al., 2001; Wallis et al., 2001), representation of emotion in MPFC (Skerry & Saxe, in
prep; Peelen et al., 2013), or representations of semantic concepts in language regions
(e.g. Simanova et al., 2014), the neural representation of modality found in RTPJ is not
tied to a single type of input, or grounded in a specific sensory domain. Instead, we
have learned that these regions must be able to take a wide variety of information as
their input (sensory cues, linguistics information, inference), and that the information
inside these regions must be represented in a feature space general enough to support
information across these sources.
An important limitation to this finding, however, is that although we find a difference
between seeing and hearing stories in the same neural area in both blind and sighted
adults, this does not guarantee that we are finding the same difference. In fact, though
the stimuli in Chapter 3 were written to manipulate only modality, they did not explicitly
code for quality of evidence — a dimension which, we discovered in Chapter 4, is also
represented in RTPJ. Therefore it is possible that while the RTPJ of sighted adults
independently represents both modality and quality, the RTPJ of blind adults represents
only quality, though this interpretation seems less likely in light of the specific stimuli
used in Chapter 3. The stimuli were matched across conditions to be minimal pairs
(e.g., he saw her face versus he heard her voice). Importantly, the stimuli often sidestepped the issue of evidence quality by explicitly stating the final outcome of the event,
rather than the evidence (e.g., John heard/saw the glass shatter, rather than John saw
glass shards or John heard a breaking sound), allowing little room to make inferences
about how well the evidence supported the conclusion. Moreover, the inference that
blind and sighted people have similar representations of sight is strengthened both by
blind people’s extensive knowledge of vision (Landau & Gleitman, 1985; Lenci, Baroni,
Cazzolli, & Marotta, 2013, Koster-Hale et al., in prep), and by our use of a priori (and
rather small) regions of interest. Finding a distinction between seeing and hearing
stories in both groups anywhere in the brain would be weak evidence that both groups
make the same distinction. However, finding this distinction in a constrained ROI,
chosen for its known role in mental state processing, makes it less likely that, by
coincidence, this region is involved in distinguishing e.g. between sensory simulations of
seeing and hearing events, in sighted people, and between familiar and unfamiliar
events, in blind people. Thus, though it is more parsimonious to infer that both groups
distinguish between these stimuli in the same region for the same reason, import future
work will need to characterize to what extent these representations are truly shared, and
where the limitations in blind people’s representations in sight are.
Replication and generalization
One the challenges of trying to characterize abstract and complex cognitive processes
using fMRI, is that the data are often too limited to draw meaningful inferences about
true conceptual structure. Even ignoring fundamental limitations in the technique (the
resolution of the data acquired through human neuroimaging is far too low to ask
questions about how single neurons (or even populations of cells) code information —
each data point blurs two seconds of activity in hundreds of thousands of neurons)
145 of 179
finding that a distinction exists in a neural population often tells us little about what part
of the distinction is being represented. Such ambiguity is due in part to a vast array of
potential dimensions that any empirically observable distinction could map on to. There
may be hidden dimensions along which complex stimuli grouped into separate
conditions differ and that it is these differences, rather than the intended manipulation,
that lead to differential brain activity.
One way to tackle that problem is by asking whether the distinction generalizes across
different manipulations of the distinctions. If a distinction is observed across multiple
ways of drawing the conceptual distinction, the observed difference in stimuli is more
likely to map onto the intended manipulation, rather than superficial properties of the
stimulus or task. Here, I find that the three key findings observed in this thesis replicate
across a diverse range of stimuli and tasks. In Chapter 2, I replicate the first key result
in three different experiments, showing that RTPJ distinguishes between accidental and
intentional harm across stimuli, task, and participants. Chapter 3, I replicate the second
key result across participant populations, showing that bilateral TPJ codes information
about source, and in Chapter 4 offer a third replication, using a new design, stimuli, and
participant pool. Chapter 4 also contains a replication the third key result, showing that
RTPJ codes information about the quality of others’ evidence, across task, stimuli,
participants, and critically, across both specific perceptual of good versus bad evidence,
and more abstract ones. (Interestingly, the representations in RMSTS do not generalize,
suggesting that the distinction in RMSTS is not precisely good versus bad evidence, but
rather a more specific dimension that was manipulated, such as quality of sensory
information.) Taken together, these data demonstrate that it is possible to find abstract,
behaviorally relevant features of mental state inferences inside cortical regions that
support social cognition, and have taken a first step in characterizing their content and
structure.
What next?
Information between regions, and between networks
Though this thesis largely focus on the content inside ToM brain regions, the results hint
at differences in the information available in each ToM region. The data here suggest
that lateral regions, including RTPJ, RMSTS, and LTPJ, represents features of beliefs;
and in particular the epistemic features of beliefs, including the quality and modality of
the evidence that gives rise to other people’s beliefs, while MPFC may track features of
individual agents, such as emotional response. These results converge with recent work
suggesting a division of labor between regions. Lateral and medial regions show
differences in overall activation profiles (e.g. Mitchell et al., 2006; Jacques. Conway,
Lowder, & Cabeza, 2011; Krienen et al., 2010), and contain different information in their
neural code. MVPA has shown that RTPJ contains information about future social
actions (Carter, Bowling, Reeck, & Huettel, 2012), social groups versus individuals
(Contreras, Schirmer, Banaji, & Mitchell, 2013), and intent when causing harm (Chapter
2, Koster-Hale et al., 2013), and the reliability of others’ cooperation (Behrens, Hunt,
Woolrich, & Rushworth, 2008). In contrast, MPFC has been shown to track information
about long-term personality traits (Hassabis et al., 2013) and emotional states (Kassam,
146 of 179
Markey, Cherkassky, Loewenstein, & Just, 2013; Peelen et al., 2010). Knowing how
regions are different is just one step in understanding how the brain works to give rise to
many complexities of social cognition. Future work needs to more systematically
characterize not only the division of labor between these regions, but also their
interactions with each other, and with regions in other networks.
Development
This thesis adds to a growing body of work showing that parts of the human brain are
selective for social cognition. A key question, then, is what triggers the specialization of
these brain regions: hard-wired maturation or experience during development? One
possibility is that some number of cognitive functions are implemented by innately
specified brain regions (Fodor 1983). These (potentially) modular brain regions would
allow rapid and efficient computation of evolutionarily significant cognitive processes, for
example, vision. On this view, the association between a selective functional profile,
such as increased response to mental state inference, and a specific brain region, such
as the RTPJ, would provide evidence that mental state inference is subserved by an
innate cognitive module. An alternative possibility is that specific brain regions acquire
highly specialized functions through experience. In particular, extensive early
experience may drive the organization of cortex, so that populations of neurons adopt
specific functions. For example, behavioral training in monkeys increases neuronal
selectivity in cortex recruited for object recognition (Baker et al. 2002; Sigala &
Logothetis 2002; Freedman et al. 2006; Cox & DiCarlo 2008). If these populations of
neurons are spatially contiguous, functionally coherent patches of cortex may be visible
using techniques like fMRI. One such example is the visual word form area, a region in
human temporal cortex that responds selectively to words in familiar (but not unfamiliar)
alphabets (Baker et al. 2007). If specialized brain regions can be organized by intense
early experience, then RTPJ may acquire its highly specific response to other minds
over the course of childhood experiences. Another important line of research will be to
test these hypotheses.
One way to test this hypothesis is to study the neural mechanism of ToM in individuals
who have had atypical developmental histories with regard to theory of mind. For
example, prior evidence suggests that early development of ToM depends on exposure
to language and to conversation about the mind (e.g. Milligan et al. 2007; Moeller &
Schick 2006; de Villiers 2005). D/deaf 15 children born to non-signing parents often have
delayed access to language, including language about mental states compared, to D/
deaf children born to signing parents. As a consequence, they exhibit delayed
performance on standard measures of ToM reasoning (Figueras-Costa & Harris 2001;
Meristo et al. 2007; Peterson & Wellman 2010; Peterson & Siegal 1999; Schick et al.
2007; Woolfe & Want 2002). However, later in childhood, these children appear to catch
up in ToM performance (Peterson & Wellman 2010; Schick et al. 2007). Since most d/
Deaf children are otherwise typically developing, but vary in their early language
15 Deaf
with an uppercase "D" usually refers to adults and children who share the use of American Sign
Language and Deaf culture, while a lowercase "d" is usually an audiological description of a person's
hearing level.
147 of 179
acquisition, they provide a unique window into the effects of early language experience
on theory of mind. In ongoing work, I am asking whether this difference in language
experience, and subsequent delay in ToM development, leave traces on the neural
basis of ToM.
Links to behavior: Content and structure of mental state inference
Finally, a core goal of understanding the neural basis of social cognition is to
understand how it gives rise to cognition. Perhaps not surprisingly, the questions that
have been asked about the neural basis of theory of mind reflect many of the questions
that have been asked behaviorally. Researchers have examined the innateness and
domain specificity of theory of mind (e.g. Carey & Spelke, 1996; Apperley et al, 2005),
and explored how mental state inference interacts with, and relies on, other cognitive
domains, such as language (e.g. Apperly et al., 2004, 2005; Astington & Baird (2005);
Milligan et al. 2007; Moeller & Schick 2006; de Villiers 2005) and executive function
(Ozonoff et al., 1991; Perner & Lang, 1999; Schneider et al., 2014). Others have tried to
characterize the prerequists and development of theory of mind (e.g. Wellman et al.,
2001; Saxe, 2013, Baillargeon et al., 2010), the cues we use when we infer that other
minds exist (e.g. Stich and Nichols, 2003), and the types of domain-general input that
feed mental state inferences (e.g. Apperly, 2014). Thus, while much work has been
done to ask what is (and isn't) in the domain of theory of mind, how it develops, and
where it sits in relation to the rest of the cognitive system, few behavioral approaches
ask about the content inside theory of mind: what the representational or computational
aspects of the mental representation itself are; what the input data, transformation of
that data, and subsequent outputs might be. Here, cognitive neuroscience may be able
to inform theories of cognition. Understanding how the brain represents complex and
abstract information to support mental state inference, and social cognition more
generally, can lay the groundwork for studying the cognitive structure of mental state
inference.
Beyond neuroimaging
However, neuroimaging is cumbersome, expensive, and limited. Many basic questions
about ToM cannot be addressed with neuroimaging. For example, a scientific theory of
how humans understand other minds should address questions like: “When and why do
we (spontaneously) seek to understand another’s thoughts?”, “How do we figure out the
actual content of someone else’s thoughts (i.e. what they are thinking) from specific
cues?", “How do we choose whether or not to incorporate others’ thoughts into our own
decisions and actions?”, and “Why do we care emotionally about others’ thoughts and
feelings?” None of these questions have yet been approached using neuroimaging, and
may pose much bigger challenges than the questions I have addressed so far.
Contemporary neuroimaging technology does not even allow us to address many
fundamental questions about the neural mechanisms of ToM. Existing tools are
extremely slow and blurry, by comparison to the speed and precision of neural
computation: they cannot decipher what is the input of a region, how that input is
transformed, or where the output from that region is sent, during mental state inference.
148 of 179
If our final horizon is a complete theory of how the brain allows us to understand other
minds, we will need to make dramatic progress on (1) the “psychophysics” of ToM in
adulthood, to allow precise quantitative measurements of people’s use of ToM; (2) a
computational model of ToM that is sufficiently explicit to make quantitative predictions
about adult judgments (e.g. Baker, Saxe, & Tenenbaum, 2011); and (3) a mechanism of
how neurons and networks of neurons might implement that computational model, by
sequentially transforming patterns of input into patterns of output that make different
information explicit. That horizon is still far away. However, the fact that we can give a
characterization of some of the boundary conditions that a successful account of ToM
needs to meet is part of what makes this such an exciting time to participate in the
cognitive neuroscience of understanding other minds.
149 of 179
150 of 179
References
Adams, R., Jr., Rule, N.O., Franklin Jr, R.G., Wang, E., Stevenson, M.T., Yoshikawa, S. (2010). Crosscultural reading the mind in the eyes: An fMRI investigation. Journal of Cognitive Neuroscience 22(1),
97-108.
Adolphs, R.R. (2001). The neurobiology of social cognition. Current Opinion in Neurobiology.
Adolphs, R.R. (2009). The social brain: neural basis of social knowledge. Annual review of psychology 60,
693.
Adolphs, R.R. (2010). Conceptual Challenges and Directions for Social Neuroscience. Neuron, 65(6),
752-767. doi:10.1016/j.neuron.2010.03.006
Adolphs, R.R. (2010). Conceptual challenges and directions for social neuroscience. Neuron, 65(6):
752-67.
Aichhorn, M. et al. (2006). Do visual perspective tasks need theory of mind? NeuroImage, 30(3): 1059-68.
Aichhorn, M. M., Perner, J., Weiss, B., Kronbichler, M., Staffen, W., & Ladurner, G. (2009). Temporoparietal junction activity in theory-of-mind tasks: falseness, beliefs, or attention. Journal of Cognitive
Neuroscience, 21(6), 1179-1192. doi:10.1162/jocn.2009.21082
Aichhorn, M., Perner, J., Kronbichler, M., Staffen, W., & Ladurner, G. (2009). Temporo-parietal junction
activity in theory-of-mind tasks: Falseness, beliefs, or attention. Journal of Cognitive Neuroscience
21(6): 1179-92
Aichhorn, M., Perner, J., Weiss, B., Kronbichler, M., Staffen, W., and Ladurner, G. (2009). Temporoparietal junction activity in theory-of-mind tasks: falseness, beliefs, or attention. J Cogn Neurosci,
21(6), 1179-1192.
Aikhenvald, A. Y. (2004). Evidentiality. Oxford University Press.
Alink, A., Schwiedrzik, C. M., Kohler, A., Singer, W., and Muckli, L. (2010). Stimulus predictability reduces
responses in primary visual cortex. J Neurosci, 30(8), 2960-2966.
Allison, T., Puce, A., and McCarthy, G. (2002). Category-sensitive excitatory and inhibitory processes in
human extrastriate cortex. J Neurophysiol, 88(5), 2864-2868.
American Psychiatric Association, & American Psychiatric Association Task Force on DSM-IV. (2000).
Diagnostic and statistical manual of mental disorders : DSM-IV-TR. Washington, DC: American
Psychiatric Association.
Ames, D.L., Jenkins, A.C., Banaji, M.R., & Mitchell, J.P. (2008). Taking another person's perspective
increases self-referential neural processing. Psychological Science 19(7): 642-4.
Anzellotti, S., Fairhall, S. L., & Caramazza, A. (2013). Decoding Representations of Face Identity That are
Tolerant to Rotation. Cerebral Cortex. doi:10.1093/cercor/bht046
Apperly, I. & Butterfill, S. (2009). Do humans have two systems to track beliefs and belief-like states?.
Psychological Review 116(4): 953-70.
Apperly, I. A. (2008). Beyond Simulation-Theory and Theory-Theory: why social cognitive neuroscience
should use its own concepts to study "theory of mind". Cognition, 107(1), 266-283. doi:10.1016/
j.cognition.2007.07.019
Apperly, I. A., Samson, D., Chiavarino, C., Bickerton, W. L., & Humphreys, G. W. (2007). Testing the
domain-specificity of a theory of mind deficit in brain-injured patients: Evidence for consistent
performance on non-verbal,'reality-unknown' false belief and false photograph tasks. Cognition,
103(2), 300-321.
151 of 179
Apperly, I.A., Samson, D., & Humphreys, G.W. (2005). Domain-specificity and theory of mind: evaluating
neuropsychological evidence. Trends in Cognitive Sciences 9(12): 572-7.
Arzy, S., Thut, G., Mohr, C., Michel, C.M., & Blanke, O. (2006). Neural basis of embodiment: distinct
contributions of temporoparietal junction and extrastriate body area. Journal of Neuroscience 26(31):
8074-81.
Averbeck, B. B., Latham, P. E., & Pouget, A. (2006). Neural correlations, population coding and
computation. Nature Reviews Neuroscience, 7(5), 358-366. doi:10.1038/nrn1888
Aydin, â., & Ceci, S. J. (2009). Evidentiality and suggestibility: A new research venue - Aydin - 2009 - New
Directions for Child and Adolescent Development - Wiley Online Library. New Directions for Child and
Adolescent ƒ.
Baez, S., Rattazzi, A., Gonzalez-Gadea, M. L., Torralva, T., Vigliecca, N. S., Decety, J., et al. (2012).
Integrating intention and context: assessing social cognition in adults with Asperger syndrome.
Frontiers in Human Neuroscience, 6, 302. doi:10.3389/fnhum.2012.00302
Baird, J., & Astington, J. W. (2004). The role of mental state understanding in the development of moral
cognition and moral action. New Directions for Child and Adolescent Development, 2004(103), 37-49.
Baillargeon, R., Scott, R. M., & He, Z. (2010). False-belief understanding in infants. Trends in cognitive
sciences, 14(3), 110-118.
Baker, C. L., Saxe, R.R., and Tenenbaum, J. B. (2009). Action understanding as inverse planning.
Cognition, 113(3), 329-349.
Baker, C. L., Saxe, R.R., & Tenenbaum, J. B. (2011). Bayesian theory of mind: Modeling joint belief-desire
attribution. In Proceedings of the thirty-second annual conference of the cognitive science society (pp.
2469-2474).
Bargh, J.A., & Thein, R.D. (1985). Individual construct accessibility, person memory, and the recalljudgment link: The case of information overload. Journal of Personality and Social Psychology 49(5):
1129.
Barlow, H. B. (1961). Possible principles underlying the transformation of sensory messages. Sensory
communication, 217-234.
Baron-Cohen, S. (1995). Mindblindness: As essay on autism and theory of mind. Recherche.
Baron-Cohen, S. (2001). Theory of mind in normal development and autism. Prisme.
Baron-Cohen, S. L., MO'Riordan. (1999). A new test of social sensitivity: Detection of faux pas in normal
children and children with Asperger syndrome. Journal of Autism and Developmental Disorders.
Baron-Cohen, S., & Wheelwright, S. (2004). The empathy quotient: an investigation of adults with
Asperger syndrome or high functioning autism, and normal sex differences. Journal of Autism and
Developmental Disorders 34(2): 163-75.
Baron-Cohen, S., O'Riordan, M., Jones, R., Stone, V., & Plaisted, K. (1999). A new test of social
sensitivity: Detection of faux pas in normal children and children with Asperger syndrome. Journal of
Autism and Developmental Disorders 29, 407-18.
Baron-Cohen, S., Wheelwright, S., Hill, J., Raste, Y., & Plumb, I. (2001). The 'Reading the mind in the
eyes' test revised version: A study with normal adults, and adults with asperger syndrome or highfunctioning autism. Journal of Child Psychology and Psychiatry 42(2): 241-51.
Battelli, L., Cavanagh, P., & Thornton, I. M. (2003). Perception of biological motion in parietal patients.
Neuropsychologia, 41(13), 1808-1816.
Bayer, H.M., and Glimcher, P.W. (2005). Midbrain Dopamine Neurons Encode a Quantitative Reward
Prediction Error Signal. Neuron 47, 13-13.
152 of 179
Bedny, M., & Saxe, R.R. (2012). Insights into the origins of knowledge from the cognitive neuroscience of
blindness. Cognitive Neuropsychology, 29(1-2), 56-84. doi:10.1080/02643294.2012.713342
Bedny, M., McGill, M., & Thompson-Schill, S. L. (2008). Semantic Adaptation and Competition during
Word Comprehension. Cerebral Cortex, 18(11), 2574-2585. doi:10.1093/cercor/bhn018
Bedny, M., Pascual-Leone, A., & Saxe, R.R. (2009). Growing up blind does not change the neural bases
of Theory of Mind. Proceedings of the National Academy of Sciences of the United States of America,
106(27), 11312-11317. doi:10.1073/pnas.0900010106
Behrens, T. E. J., Hunt, L. T., & Rushworth, M. F. S. (2009). The Computation of Social Behavior. Science,
324(5931), 1160-1164. doi:10.1126/science.1169694
Behrens, T. E. J., Hunt, L. T., Woolrich, M. W., & Rushworth, M. F. S. (2008). Associative learning of social
value. Nature, 456(7219), 245-249. doi:10.1038/nature07538
Berthoz, S., Armony, J. L., Blair, R. J. R., and Dolan, R. J. (2002). An fMRI study of intentional and
unintentional (embarrassing) violations of social norms. Brain, 125(8), 1696-1708.
Bhatt, M., Lohrenz, T., Camerer, C.F., & Montague, P.R. (2010). Neural signatures of strategic types in a
two-person bargaining game. In Proceedings of the National Academy of Sciences 107(46): 19720-5
Bialek, W., Rieke, F., Van Steveninck, R.R. R., & Warland, D. (1991). Reading a neural code. Science,
252(5014), 1854-1857.
Bigelow, A. (1991). Hiding in blind and sighted children. Development and Psychopathology, (3), 301-310.
Bigelow, A. (1992). The development of blind children's ability to predict what another sees, (86),
181-184.
Binney, R. J., Embleton, K. V., Jefferies, E., Parker, G. J., & Ralph, M. A. L. (2010). The ventral and
inferolateral aspects of the anterior temporal lobe are crucial in semantic memory: evidence from a
novel direct comparison of distortion-corrected fMRI, rTMS, and semantic dementia. Cerebral Cortex,
20(11), 2728-2738.
Blair, J. (1996, October 30). Brief Report: Morality in the Autistic Child. Journal of Autism and
Developmental Disorders. Retrieved May 30, 2011, from
Blake, R., and Shiffrar, M. (2007). Perception of human motion. Annu Rev Psychol 58, 47-73.
Blakemore, S. J. (2005). Somatosensory activations during the observation of touch and a case of visiontouch synaesthesia. Brain, 128(7), 1571-1583. doi:10.1093/brain/awh500
Blanke, O., & Arzy, S. (2005). The out-of-body experience: disturbed self-processing at the temporoparietal junction. Neuroscientist 11(1): 16-24.
Blanke, O., Mohr, C., Michel, C.M., Pascual-Leone, A., Brugger, P., Seeck, M., et al. (2005). Linking outof-body experience and self processing to mental own-body imagery at the temporoparietal junction.
Journal of Neuroscience 25(3): 550-7.
Bloom, P., & German, T. (2000). Two reasons to abandon the false belief task as a test of theory of mind.
Cognition, 77(1), B25-B31.
Bonda, E., Petrides, M., Ostry, D., & Evans, A. (1996). Specific involvement of human parietal systems
and the amygdala in the perception of biological motion. J Neurosci, 16(11), 3737-3744.
Bortoletto, M., Mattingley, J. B., & Cunnington, R. (2011). Action intentions modulate visual processing
during action perception. Neuropsychologia, 49(7), 2097-2104.
Botvinick, M., Jha, A. P., Bylsma, L. M., Fabian, S. A., Solomon, P. E., & Prkachin, K. M. (2005). Viewing
facial expressions of pain engages cortical areas involved in the direct experience of pain.
NeuroImage, 25(1), 312-319. doi:10.1016/j.neuroimage.2004.11.043
153 of 179
Bradley, M.M., Sabatinelli, D., Lang, P.J., Fitzsimmons, J.R., King, W., and Desai, P. (2003). Activation of
the visual cortex in motivated attention. Behav Neurosci 117, 369-380.
Brambring, M., & Asbrock, D. (2010). Validity of false belief tasks in blind children. Journal of Autism and
Developmental Disorders, 40(12), 1471-1484. doi:10.1007/s10803-010-1002-2
Brass, M., Schmitt, R.M., Spengler, S., and Gergely, G. (2007). Investigating Action Understanding:
Inferential Processes versus Action Simulation. Curr Biol 17, 2117-2121
Bray, S., Chang, C., & Hoeft, F. (2009). Applications of multivariate pattern classification analyses in
developmental neuroimaging of healthy and clinical populations. Frontiers in Human Neuroscience, 3.
doi:10.3389/neuro.09.032.2009
Brieber, S., Herpertz-Dahlmann, B., Fink, G. R., Kamp-Becker, I., Remschmidt, H., & Konrad, K. (2010).
Coherent motion processing in autism spectrum disorder (ASD): An fMRI study. Neuropsychologia,
48(6), 1644-1651. doi:10.1016/j.neuropsychologia.2010.02.007
Bright-Paul, A., Jarrold, C., & Wright, D. B. (2008). Theory-of-mind development influences suggestibility
and source monitoring. Developmental Psychology, 44(4), 1055-1068.
doi:10.1037/0012-1649.44.4.1055
Brown, E. C., & Brüne, M. (2012). The role of prediction in social neuroscience. Front Hum Neurosci, 6.
Bruneau E., Dufour N., & Saxe R. (2012) Social cognition in members of conflict groups: behavioural and
neural responses in Arabs, Israelis and South Americans to each other's misfortunes. Philosophical
Transactions of the Royal Society, London, B Biological Sciences 367(1589) 717-30.
Bruneau, E. G., & Saxe, R.R. (2012). Distinct roles of the 'shared pain' and 'theory of mind' networks in
processing others' emotional suffering. Neuropsychologia, 50(2), 219-231. doi:10.1016/
j.neuropsychologia.2011.11.008
Bruneau, E.G., & Saxe, R.R. (2010). Attitudes toward the outgroup are predicted by activity in the
precuneus in Arabs and Israelis. NeuroImage 52(4): 1704-11.
Bruneau, E.G., Pluta, A., & Saxe, R.R. (2011). Distinct roles of the 'shared pain' and 'theory of mind'
networks in processing others' emotional suffering. Neuropsychologia pp.1-13.
Bubic, A., Cramon, von, D.Y., Jacobsen, T., Schröger, E., and Schubotz, R.I. (2009). Violation of
expectation: neural correlates reflect bases of prediction. J Cogn Neurosci 21, 155-168.
Buckholtz, J. W., Asplund, C. L., Dux, P. E., Zald, D. H., Gore, J. C., Jones, O. D., & Marois, R. (2008).
The Neural Correlates of Third-Party Punishment. Neuron, 60(5), 930-940. doi:10.1016/j.neuron.
2008.10.016
Buhrmester, M., Kwang, T., & Gosling, S. D. (2011). Amazon's Mechanical Turk A New Source of
Inexpensive, Yet High-Quality, Data? Perspectives on Psychological Science, 6(1), 3-5. doi:
10.1177/1745691610393980
Bull Kovera, M., Park, R. C., & Penrod, S. D. (1992). Jurors' Perceptions of Eyewitness and Hearsay
Evidence. Minn L Rev, 76, 703.
Butts, D. A., Weng, C., Jin, J., Yeh, C.-I., Lesica, N. A., Alonso, J.-M., & Stanley, G. B. (2007). Temporal
precision in the neural code and the timescales of natural vision. Nature, 449(7158), 92-95. doi:
10.1038/nature06105
Bzdok, D., Schilbach, L., Vogeley, K., Schneider, K., Laird, A. R., Langner, R., & Eickhoff, S. B. (2012).
Parsing the neural correlates of moral cognition: ALE meta-analysis on morality, theory of mind, and
empathy. Brain Structure and Function. doi:10.1007/s00429-012-0380-y
Calder, A. J., Beaver, J. D., Winston, J. S., Dolan, R. J., Jenkins, R., Eger, E., & Henson, R. N. (2007).
Separate coding of different gaze directions in the superior temporal sulcus and inferior parietal
lobule. Curr Biol, 17(1), 20-25.
154 of 179
Calvo-Merino, B., Glaser, D. E., Grèzes, J., Passingham, R. E., & Haggard, P. (2005). Action observation
and acquired motor skills: an FMRI study with expert dancers. Cerebral Cortex, 15(8), 1243-1249.
doi:10.1093/cercor/bhi007
Calvo-Merino, B., Grèzes, J., Glaser, D. E., Passingham, R. E., & Haggard, P. (2006). Seeing or Doing?
Influence of Visual and Motor Familiarity in Action Observation. Current Biology, 16(19), 1905-1910.
doi:10.1016/j.cub.2006.07.065
Carey, S., & Spelke, E. (1996). Science and core knowledge. Philosophy of Science, 515-533.
Carrington, S. (2009). Are there theory of mind regions in the brain? A review of the neuroimaging
literature - Carrington - 2008 - Human Brain Mapping - Wiley Online Library. Human Brain Mapping.
Carrington, S.J., & Bailey, A.J. (2009). Are there theory of mind regions in the brain? A review of the
neuroimaging literature. Human Brain Mapping 30(8): 2313-35.
Carter, E.J., Hodgins, J.K., and Rakison, D.H. (2011). Exploring the neural correlates of goal-directed
action and intention understanding. NeuroImage 54, 1634-1642.
Carter, R. M., Bowling, D. L., Reeck, C., & Huettel, S. A. (2012). A Distinct Role of the Temporal-Parietal
Junction in Predicting Socially Guided Decisions. Science, 337(6090), 109-111. doi:10.1126/science.
1219681
Castelli, F., Happé, F. G. E., Frith, U., & Frith, C. (2000). Movement and mind: a functional imaging study
of perception and interpretation of complex intentional movement patterns. NeuroImage, 12(3),
314-325.
Castelli, F., Happé, F., Frith, U., & Frith, C. (2000). Movement and mind: a functional imaging study of
perception and interpretation of complex intentional movement patterns. NeuroImage 12(3): 314-25.
Chang, C.-C., & Lin, C.-J. (2011). LIBSVM: A library for support vector machines. ACM Transactions on
Intelligent Systems and Technology (TIST), 2(3), 27-27. doi:10.1145/1961189.1961199
Chang, L.J., and Sanfey, A.G. (2013). Great expectations: neural computations underlying the use of
social norms in decision-making. Soc Cogn Affect Neurosci 8, 277-284.
Chen, W. (2009). Understanding mental states in natural language. In Proceedings of the Eighth
International Conference on Computational Semantics (pp. 61-72). Association for Computational
Linguistics.
Chikazoe, J., Lee, D. H., Kriegeskorte, N., & Anderson, A. K. (2014). Population coding of affect across
stimuli, modalities and individuals. Nature Publishing Group, 1-11. doi:10.1038/nn.3749
Chomsky, N. (1995). The minimalist program (Vol. 28). Cambridge, MA: MIT press.
Clark, A. (2013). Whatever next? Predictive brains, situated agents, and the future of cognitive science.
Behav Brain Sci 36, 181-204.
Clément, F., Koenig, M., & Harris, P. (2004). The Ontogenesis of Trust. Mind & Language, 19(4), 360-379.
doi:10.1111/j.0268-1064.2004.00263.x
Cloutier, J., Gabrieli, J.D.E., O'Young, D., & Ambady, N. (2011). An fMRI study of violations of social
expectations: When people are not who we expect them to be. NeuroImage 57(2): 583-8.
Cloutier, J., Gabrieli, J.D.E., O'Young, D., and Ambady, N. (2011). An fMRI study of violations of social
expectations: When people are not who we expect them to be. NeuroImage 57, 583-588.
Cohen, J.D., Perlstein, W.M., Braver, T.S., Nystrom, L.E., Noll, D.C., Jonides, J., and Smith, E.E. (1997).
Temporal dynamics of brain activation during a working memory task. Nature 386, 604-608.
Cohen, S.B., Wheelwright, S., & Hill, J. (2001). The 'reading the mind in the eyes' test revised version: A
study with normal adults, and adults with Asperger syndrome or high-functioning autism. Journal of
Child Psychology and Psychiatry. 42(2), 241-251.
155 of 179
Contreras, J. M., Schirmer, J., Banaji, M., & Mitchell, J. P. (2013). Common Brain Regions with Distinct
Patterns of Neural Responses during Mentalizing about Groups and Individuals. Journal of Cognitive
Neuroscience, 25(9), 1406-1417. doi:10.1162/jocn-a-00403
Corbetta, M., & Shulman, G.L. (2002). Control of goal-directed and stimulus-driven attention in the brain.
Nature Review Neuroscience 3: 201-15.
Corbetta, M., Kincade, J.M., Ollinger, J.M., McAvoy, M.P., & Shulman, G.L. (2000). Voluntary orienting is
dissociated from target detection in human posterior parietal cortex. Nature Neuroscience 3: 292-7.
Corbetta, M., Patel, G., & Shulman, G.L. (2008). The reorienting system of the human brain: from
environment to theory of mind. Neuron 58(3): 306-24.
Coricelli, G., & Nagel, R. (2009). Neural correlates of depth of strategic reasoning in medial prefrontal
cortex. Proceedings of the National Academy of Sciences 106(23): 9163-8.
Correia, J., Formisano, E., Valente, G., Hausfeld, L., Jansma, B., & Bonte, M. (2014). Brain-based
translation: fMRI decoding of spoken words in bilinguals reveals language-independent semantic
representations in anterior temporal lobe. The Journal of Neuroscience : the Official Journal of the
Society for Neuroscience, 34(1), 332-338. doi:10.1523/JNEUROSCI.1302-13.2014
Costa, A., Torriero, S., & Oliveri, M. (2008). Prefrontal and temporo-parietal involvement in taking others'
perspective: TMS evidence. Behavioural Neurology 19(1-2): 71-4.
Costafreda, S.G., Brammer, M.J., David, A.S., & Fu, C.H. (2008). Predictors of amygdala activation during
the processing of emotional stimuli: A meta-analysis of 385 PET and fMRI studies. Brain Research
Reviews 58(1): 57-70.
Costantini, M. (2005). Neural Systems Underlying Observation of Humanly Impossible Movements: An
fMRI Study. Cereb Cortex 15, 1761-1767.
Coutanche, M. N., Thompson-Schill, S. L., & Schultz, R. T. (2011). Multi-voxel pattern analysis of fMRI
data predicts clinical symptom severity. NeuroImage, 57(1), 113-123. doi:10.1016/j.neuroimage.
2011.04.016
Cox, D. D., & Savoy, R. L. (2003). Functional magnetic resonance imaging (fMRI) 'brain reading':
detecting and classifying distributed patterns of fMRI activity in human visual cortex. NeuroImage,
19(2), 261-270. doi:10.1016/S1053-8119(03)00049-1
Csibra, G., and Gergely, G. (2007). Obsessed with goals': Functions and mechanisms of teleological
interpretation of actions in humans. Acta Psychol 124, 60-78.
Cule, E. (2012). ridge: Ridge Regression with automatic selection of the penalty parameter. R package
version.
Cule, E., & De Iorio, M. (2013). Ridge Regression in Prediction Problems: Automatic Choice of the Ridge
Parameter. Genetic Epidemiology, 37(7), 704-714. doi:10.1002/gepi.21750
Cushman, F. (2008). Crime and punishment: Distinguishing the roles of causal and intentional analyses in
moral judgment. Cognition, 108(2), 353-380. doi:10.1016/j.cognition.2008.03.006
Cushman, F. A. (2008). Crime and punishment: Distinguishing the roles of causal and intentional analyses
in moral judgment. Cognition, 108(2), 353-380. doi:10.1016/j.cognition.2008.03.006
Cushman, F., Young, L.L., & Hauser, M. (2006). The role of conscious reasoning and intuition in moral
judgment testing three principles of harm. Psychological Science, 17(12), 1082-1089.
D'Argembeau, A., Ruby, P., Collette, F., Degueldre, C., Balteau, E., Luxen, A., et al. (2007). Distinct
regions of the medial prefrontal cortex are associated with self-referential processing and perspective
taking. Journal of Cognitive Neuroscience 19(6): 935-944.
Davidson, D. (1963). Actions, reasons, and causes. The Journal of Philosophy, 685-700.
156 of 179
de Gardelle, V., Waszczuk, M., Egner, T., and Summerfield, C. (2012). Concurrent Repetition
Enhancement and Suppression Responses in Extrastriate Visual Cortex. Cereb Cortex. doi: 10.1093/
cercor/bhs211
de Martino, F., Valente, G., Staeren, N., Ashburner, J., Goebel, R., & Formisano, E. (2008). Combining
multivariate voxel selection and support vector machines for mapping and classification of fMRI
spatial patterns. NeuroImage, 43(1), 44-58. doi:10.1016/j.neuroimage.2008.06.037
de Villiers, J. G., Garfield, J., Gernet-Girard, H., Roeper, T., & Speas, M. (2009). Evidentials in Tibetan:
Acquisition, semantics, and cognitive development. New Directions for Child and Adolescent
Development, 2009(125), 29-47. doi:10.1002/cd.248
de Villiers, P.A. (2005). The Role of Language in Theory-of-Mind Development: What Deaf Children Tell
Us A. JW & B. JA, eds. Why language matters for theory of mind.
de Wit, L., Machilsen, B., and Putzeys, T. (2010). Predictive Coding and the Neural Response to
Predictable Stimuli. J Neurosci 30, 8702-8703.
Decety, J., & Lamm, C. (2007). The role of the right temporoparietal junction in social interaction: how
low-level computational processes contribute to meta-cognition. The Neuroscientist : a Review
J o u r n a l B r i n g i n g N e u r o b i o l o g y, N e u r o l o g y a n d P s y c h i a t r y, 1 3 ( 6 ) , 5 8 0 - 5 9 3 . d o i :
10.1177/1073858407304654
Decety, J., Michalska, K. J., & Akitsuki, Y. (2008). Who caused the pain? An fMRI investigation of empathy
and intentionality in children. Neuropsychologia, 46(11), 2607-2614. doi:10.1016/j.neuropsychologia.
2008.05.026
Deen, B., and Saxe, R.R. (2012). Neural correlates of social perception: The posterior superior temporal
sulcus is modulated by action rationality, but not animacy. Proceedings of the 33rd Annual Cognitive
Science Society Conference 276-81.
Dehaene, S., & Cohen, L. (2007). Cultural Recycling of Cortical Maps. Neuron, 56(2), 384-398. doi:
10.1016/j.neuron.2007.10.004
den Ouden, H.E.M., Daunizeau, J., Roiser, J., Friston, K.J., and Stephan, K.E. (2010). Striatal Prediction
Error Modulates Cortical Coupling. J Neurosci 30, 3210-3219.
den Ouden, H.E.M., Friston, K.J., Daw, N.D., McIntosh, A.R., and Stephan, K.E. (2008). A Dual Role for
Prediction Error in Associative Learning. Cereb Cortex 19, 1175-1185.
den Ouden, H.E.M., Kok, P.P., and de Lange, F.P.F. (2012). How prediction errors shape perception,
attention, and motivation. Front Psychol 3, 548-548.
Desimone, R.R., Albright, T.D.T., Gross, C.G.C., and Bruce, C.C. (1984). Stimulus-selective properties of
inferior temporal neurons in the macaque. J Neurosci 4, 2051-2062.
DiCarlo, J. J., & Cox, D. D. (2007). Untangling invariant object recognition. Trends in Cognitive Sciences,
11(8), 333-341. doi:10.1016/j.tics.2007.06.010
DiCarlo, J. J., Zoccolan, D., & Rust, N. C. (2012). How Does the Brain Solve Visual Object Recognition?
Neuron, 73(3), 415-434. doi:10.1016/j.neuron.2012.01.010
Dinstein, I., Thomas, C., Humphreys, K., Minshew, N., Behrmann, M., & Heeger, D. J. (2010). Normal
Movement Selectivity in Autism. Neuron, 66(3), 461-469. doi:10.1016/j.neuron.2010.03.034
Dodell-Feder, D., Koster-Hale, J., Bedny, M., & Saxe, R.R. (2011). fMRI item analysis in a theory of mind
task. NeuroImage, 55(2), 705-712. doi:10.1016/j.neuroimage.2010.12.040
Dodson, C. S., & Johnson, M. K. (1993). Rate of false source attributions depends on how questions are
asked. American Journal of Psychology, 106(4), 541-557.
157 of 179
Döhnel, K., Schuwerk, T., Meinhardt, J., Sodian, B., Hajak, G., & Sommer, M. (2012). Functional activity
of the right temporo-parietal junction and of the medial prefrontal cortex associated with true and false
belief reasoning. NeuroImage, 1-10. doi:10.1016/j.neuroimage.2012.01.073
Dosenbach, N. U. F., Nardos, B., Cohen, A. L., Fair, D. A., Power, J. D., Church, J. A., et al. (2010).
Prediction of Individual Brain Maturity Using fMRI. Science, 329(5997), 1358-1361. doi:10.1126/
science.1194144
Drugowitsch, J.J., and Pouget, A.A. (2012). Probabilistic vs. non-probabilistic approaches to the
neurobiology of perceptual decision-making. Curr Opin Neurobiol 22, 963-969.
Dufour N., Redcay, E, Young, L.L., Mavros, P., Moran, J., Triantafyllou, C., Gabrieli, J., & Saxe, R.R.
(2012). What explains variability in brain regions associated with Theory of Mind in a large sample of
neurotypical adults and adults with ASD? In N. Miyake, D. Peebles, & R. P. Cooper (Eds),
Proceedings of the 34th Annual Conference of the Cognitive Science Society (pp. 312-17). Austin:
Cognitive Science Society.
Dufour, N., Redcay, E., Young, L. L., Mavros, P. L., Moran, J. M., Triantafyllou, C., et al. (2013). Similar
Brain Activation during False Belief Tasks in a Large Sample of Adults with and without Autism. PLoS
ONE, 8(9), e75468. doi:10.1371/journal.pone.0075468
Dunne, S., and O'Doherty, J.P. (2013). Insights from the application of computational neuroimaging to Soc
Neurosci. Curr Opin Neurobiol. 23, 387-392.
Egner, T., Monti, J.M., and Summerfield, C. (2010). Expectation and Surprise Determine Neural
Population Responses in the Ventral Visual Stream. J Neurosci 30, 16601-16608.
Eichenbaum, H. (2004). Hippocampus: cognitive processes and neural representations that underlie
declarative memory. Neuron, 44(1), 109-120.
Ewbank, M.P., Henson, R.N., Rowe, J.B., Stoyanova, R.S., and Calder, A.J. (2013). Different Neural
Mechanisms within Occipitotemporal Cortex Underlie Repetition Suppression across Same and
Different-Size Faces. Cereb Cortex 23, 1073-1084.
Fairhall, S. L., Anzellotti, S., Ubaldi, S., & Caramazza, A. (2013). Person-and place-selective neural
substrates for entity-specific semantic access. Cerebral Cortex, bht039.
Faller, M. (2002). Semantics and Pragmatics of Evidentials in Cuzco Quechua. Doctoral Dissertation,
Stanford University.
Fedorenko, E., & Kanwisher, N. (2009). Neuroimaging of language: Why hasn't a clearer picture
emerged? Language and Linguistics Compass 3(4): 839-65.
Fedorenko, E., Nieto-Casta-Ñn, A., & Kanwisher, N. (2012). Lexical and syntactic representations in the
brain: An fMRI investigation with multi-voxel pattern analyses. Neuropsychologia, 50(4), 499-513. doi:
10.1016/j.neuropsychologia.2011.09.014
Fehr, E., and Camerer, C.F. (2007). Social neuroeconomics: the neural circuitry of social preferences.
Trends Cogn Sci 11, 419-427.
Feldman, H., and Friston, K.J. (2010). Attention, uncertainty, and free-energy. Front Hum Neurosci. 4,
215-215.
Felleman, D.J., and Van Essen, D.C. (1991). Distributed hierarchical processing in the primate Cereb
Cortex. Cereb Cortex 1, 1-47.
Figueras-Costa, B., & Harris, P. (2001). Theory of mind development in deaf children: A nonverbal test of
false-belief understanding. Journal of Deaf Studies and Deaf Education, 6(2), 92.
Fiorillo, C.D., Tobler, P.N., and Schultz, W. (2003). Discrete coding of reward probability and uncertainty
by dopamine neurons. Science 299, 1898-1902.
158 of 179
Fitneva, S. A. (2001). Epistemic marking and reliability judgments: Evidence from Bulgarian. Journal of
Pragmatics, 33, 401-420.
Flesch, R.R. (1948). A new readability yardstick. Journal of Applied Psychology, 32(3), 221-233. doi:
10.1037/h0057532
Fletcher, P., Happe, F., Frith, U., Baker, S.C., Dolan, R.J., Frackowiak, R.S., & Frith, C. D. (1995). Other
minds in the brain: a functional imaging study of 'theory of mind' in story comprehension. Cognition
57(2): 109-28.
Fodor, J. A. (1981). Representations: Philosophical essays on the foundations of cognitive science.
Fodor, J. A. (1987). Modules, frames, fridgeons, sleeping dogs, and the music of the spheres.
Formisano, E., Kim, D. S., Di Salle, F., van de Moortele, P. F., Ugurbil, K., & Goebel, R. (2003). Mirrorsymmetric tonotopic maps in human primary auditory cortex. Neuron, 40(4), 859-869.
Freiwald, W.A., Tsao, D.Y., & Livingstone, M.S. (2009). A face feature space in the macaque temporal
lobe. Nature 12(9): 1187-96.
Friederici, A. D. (2002). Towards a neural basis of auditory sentence processing. Trends in cognitive
sciences, 6(2), 78-84.
Friston, K. (2005). A theory of cortical responses. Philos Trans R Soc Lond B Biol Sci 360, 815-836.
Friston, K. (2009). Modalities, Modes, and Models in Functional Neuroimaging. Science 326, 399-403.
Friston, K. (2010). The free-energy principle: a unified brain theory? Nat Rev Neurosci, 11(2), 127-138.
Friston, K., Holmes, A.P., Worsley, K.J., Poline, J.P., Frith, C.D., & Frackowiak, R.S.J. (1995). Statistical
parametric maps in functional imaging: a general linear approach. Human Brain Mapping 2: 189-210.
Frith, C. D., & Frith, U. (2000). The Physiological Basis of Theory of Mind. In S. L. Baron-Cohen, H. TagerFlusberg, & D. Cohen, Understanding Other Minds: Perspective from Developmental Social
Neuroscience (pp. 335-356). Oxford: Oxford University Press.
Frith, C., & Frith, U. (2012). Social neuroscience. Annual Review of Psychology 63: 287-313.
Furl, N., van Rijsbergen, N. J., Treves, A., Friston, K. J., & Dolan, R. J. (2007). Experience-dependent
coding of facial expression in superior temporal sulcus. Proc Natl Acad Sci U S A, 104(33),
13485-13489.
Gallagher, H. L., Happé, F. G. E., Brunswick, N., Fletcher, P. C., Frith, U., & Frith, C. D. (2000). Reading
the mind in cartoons and stories: an fMRI study of 'theory of mind' in verbal and nonverbal tasks.
Neuropsychologia, 38(1), 11-21.
Gallagher, H., Happé, F., Brunswick, N., Fletcher, P. C., Frith, U., & Frith, C. D. (2000). Reading the mind
in cartoons and stories: an fMRI study of 'theory of mind' in verbal and nonverbal tasks.
Neuropsychologia 38(1): 11-21.
Gallagher, H.L., & Frith, C.D. (2003). Functional imaging of 'theory of mind'. Trends in Cognitive Sciences
7, 77-83.
Gallese, V., & Goldman, A. (1998). Mirror neurons and the simulation theory of mind-reading. Trends in
Cognitive Sciences, 2(12), 493-501.
Gallese, V., & Sinigaglia, C. (2011). What is so special about embodied simulation? Trends in Cognitive
Sciences, 15(11), 512-519. doi:10.1016/j.tics.2011.09.003
Gazzola, V., & Keysers, C. (2009). The Observation and Execution of Actions Share Motor and
Somatosensory Voxels in all Tested Subjects: Single-Subject Analyses of Unsmoothed fMRI Data.
Cerebral Cortex, 19(6), 1239-1255. doi:10.1093/cercor/bhn181
159 of 179
Georgopoulos, A., Schwartz, A., & Kettner, R. (1986). Neuronal population coding of movement direction.
Science 233(4771): 1416-19.
Gergely, G., & Csibra, G. (2003). Teleological reasoning in infancy: The na've theory of rational action.
Trends in Cogn Sci, 7(7), 287-292.
Gilbert, S. J., Meuwese, J. D. I., Towgood, K. J., Frith, C. D., & Burgess, P. W. (2008). Abnormal functional
specialization within medial prefrontal cortex in high-functioning autism: a multi-voxel similarity
analysis. Brain, 132(4), 869-878. doi:10.1093/brain/awn365
Gobbini, M., Koralek, A.C., Bryan, R.E., Montgomery, K.J., & Haxby, J.V. (2007). Two takes on the social
brain: A comparison of theory of mind tasks. Journal of Cognitive Neuroscience 19(11): 1803-14.
Goel, V., Grafman, J., Sadato, N., & Hallett, M. (1995). Modeling other minds. Neuroreport, 6(13),
1741-1746.
Goldman, A. I. (1989). Interpretation Psychologized. Mind & Language, 4(3), 161-185.
Goldman, A. I. (2006). Simulating Minds:The Philosophy, Psychology, and Neuroscience of Mindreading.
Oxford University Press, USA.
Gopnik, A. M., & Meltzoff, A. N. (1997). Words, Thoughts, and Theories. Cambridge, MA: MIT Press.
Gopnik, A. M., & Wellman, H. M. (1992). Why the Child's Theory of Mind Really Is a Theory. Mind &
Language, 7(1-2), 145-171. doi:10.1111/j.1468-0017.1992.tb00202.x
Gopnik, A., & Meltzoff, A.N. (1997). Words, Thoughts, and Theories. Cambridge: MIT Press. 55 Hayward
Street, Cambridge, MA 02142.
Gordon, R. M. (1989). Folk Psychology as Simulation. Mind & Language, 1(2), 158-171. doi:10.1111/j.
1468-0017.1986.tb00324.x
Goris, R. L. T., & Op de Beeck, H. P. (2009). Neural representations that support invariant object
recognition. Frontiers in Computational Neuroscience, 3, 3. doi:10.3389/neuro.10.003.2009
Graesser, A. C., & McNamara, D. S. (2011). Coh-Metrix providing multilevel analyses of text
characteristics. Educational ƒ.
Graf, M., Reitzner, B., Corves, C., Casile, A., Giese, M., and Prinz, W. (2007). Predicting point-light
actions in real-time. NeuroImage 36 Suppl 2, T22-T32.
Grant, C. M. (2005). Moral understanding in children with autism. Autism, 9(3), 317-331. doi:
10.1177/1362361305055418
Greene, J. D., Nystrom, L. E., Engell, A. D., Darley, J. M., & Cohen, J. D. (2004). The Neural Bases of
Cognitive Conflict and Control in Moral Judgment. Neuron, 44(2), 389-400. doi:10.1016/j.neuron.
2004.09.027
Greene, J. D., Sommerville, R. B., Nystrom, L. E., Darley, J. M., & Cohen, J. D. (2001). An fMRI
investigation of emotional engagement in moral judgment. Science, 293(5537), 2105-2108. doi:
10.1126/science.1062872
Gregory, R.L. (1980). Perceptions as hypotheses. Philos Trans R Soc Lond B Biol Sci 290, 181-197.
Grill-Spector, K., Henson, R., & Martin, A. (2006). Repetition and the brain: neural models of stimulusspecific effects. Trends in Cognitive Sciences 10(1): 14-23.
Groen, W. B., Tesink, C., Petersson, K. M., van Berkum, J., van der Gaag, R. J., Hagoort, P., & Buitelaar,
J. K. (2010). Semantic, factual, and social language comprehension in adolescents with autism: an
FMRI study. Cerebral Cortex, 20(8), 1937-1945. doi:10.1093/cercor/bhp264
Grossman, E.D., Battelli, L.L., and Pascual-Leone, A.A. (2005). Repetitive TMS over posterior STS
disrupts perception of biological motion. Vision Res 45, 2847-2853.
160 of 179
Grossman, E.D., Jardine, N. L., & Pyles, J. A. (2010). fMR-adaptation reveals invariant coding of
biological motion on the human STS. Front Hum Neurosci, 4. 15
Gweon, H., Dodell-Feder, D., Bedny, M., & Saxe, R.R. (2012). Theory of Mind Performance in Children
Correlates With Functional Specialization of a Brain Region for Thinking About Thoughts. Child
Development, no-no. doi:10.1111/j.1467-8624.2012.01829.x
Hamilton, D. L., and Sherman, S. J. (1996). Perceiving persons and groups. Psychol Rev, 103(2), 336.
Hamilton, D.L. (1988). Causal attribution viewed from an information-processing perspective. In: D. BarTal & A. W. Kruglanski (Eds), The Social Psychology of Knowledge (pp. 359-85). Cambridge:
Cambridge University Press.
Hamilton, D.L., & Sherman, S.J. (1996). Perceiving persons and groups. Psychological Review 103(2):
336.
Hampton, A. N., & O'Doherty, J. P. (2007). Decoding the neural substrates of reward-related decision
making with functional MRI. Proceedings of the National Academy of Sciences of the United States of
America, 104(4), 1377-1382. doi:10.1073/pnas.0606297104
Hardwig, J. (1991). The role of trust in knowledge. The Journal of Philosophy, 88(12), 693. doi:
10.2307/2027007
Harris, P. (2002a). Checking our sources: the origins of trust in testimony. -Studies in History and
Philosphy of Science, (33), 315-333.
Harris, P. (2002b). What do children learn from testimony? In P. Carruthers, S. Stich, & M. Siegal (Eds.),
The Cognitive Basis of Science. Cambridge University Press.
Harris, P. L. (1992). From Simulation to Folk Psychology: The Case for Development. Mind & Language,
7(1-2), 120-144. doi:10.1111/j.1468-0017.1992.tb00201.x
Harris, P. L., & Koenig, M. A. (2006). Trust in testimony: how children learn about science and religion.
Child Development, 77(3), 505-524. doi:10.1111/j.1467-8624.2006.00886.x
Harrison, S. A., & Tong, F. (2009). Decoding reveals the contents of visual working memory in early visual
areas. Nature, 458(7238), 632-635.
Hassabis, D., Spreng, R. N., Rusu, A. A., Robbins, C. A., Mar, R. A., & Schacter, D. L. (2013). Imagine All
the People: How the Brain Creates and Uses Personality Models to Predict Behavior. Cerebral Cortex
(New York, N.Y. : 1991), 24(8), bht042-1987. doi:10.1093/cercor/bht042
Hastie, T. J., Tibshirani, R. J., & Friedman, J. H. (2011). The elements of statistical learning: data mining,
inference, and prediction.
Haxby, J. V. (2001). Distributed and Overlapping Representations of Faces and Objects in Ventral
Temporal Cortex. Science, 293(5539), 2425-2430. doi:10.1126/science.1063736
Hayasaka, S., & Nichols, T. E. (2004). Combining voxel intensity and cluster extent with permutation test
framework. NeuroImage, 23(1), 54-63. doi:10.1016/j.neuroimage.2004.04.035
Hayashi, A., Abe, N., Ueno, A., Shigemune, Y., Mori, E., Tashiro, M., & Fujii, T. (2010). Neural correlates of
forgiveness for moral transgressions involving deception. Brain Research 1332: 90-9.
Hayden, B.Y., Heilbronner, S.R., Pearson, J.M., and Platt, M.L. (2011). Surprise Signals in Anterior
Cingulate Cortex: Neuronal Encoding of Unsigned Reward Prediction Errors Driving Adjustment in
Behavior. J Neurosci. 31, 4178-4187.
Haynes, J.-D., & Rees, G. (2006). Decoding mental states from brain activity in humans. Nature Reviews
Neuroscience, 7(7), 523-534. doi:10.1038/nrn1931
He, D., Kersten, D., and Fang, F. (2012). Opposite Modulation of High- and Low-Level Visual Aftereffects
by Perceptual Grouping. Curr Biol 22, 1040-1045.
161 of 179
Hearst, M. A., Dumais, S. T., Osman, E., Platt, J., & Scholkopf, B. (1998). Support vector machines.
Intelligent Systems and Their Applications, IEEE, 13(4), 18-28. doi:10.1109/5254.708428
Heider, F., & Simmel, M. (1944). An experimental study of apparent behavior. American Journal of
Psychology 57(2): 243-59.
Hein, G., & Knight, R.T. (2008). Superior temporal sulcus-it's my area: or is it? Journal of Cognitive
Neuroscience 20(12): 2125-36.
Herrmann, K., Montaser-Kouhsari, L., Carrasco, M., and Heeger, D.J. (2010). When size matters:
attention affects performance by contrast or response gain. Nature 13, 1554-1559.
Hesselmann, G.G., Sadaghiani, S.S., Friston, K.J.K., and Kleinschmidt, A.A. (2010). Predictive coding or
evidence accumulation? False inference and neuronal fluctuations. PLoS ONE 5, e9926-e9926.
Heyes, C. (2014). False belief in infancy: a fresh look. Developmental Science, n/a-n/a. doi:10.1111/desc.
12148
Higgins, E.T., & Bargh, J.A. (1987). Social cognition and social perception. Annual Review of Psychology
38(1): 369-425.
Hillyard, S.A., Vogel, E.K., and Luck, S.J. (1998). Sensory gain control (amplification) as a mechanism of
selective attention: electrophysiological and neuroimaging evidence. Philos Trans R Soc Lond B Biol
Sci 353, 1257-1270.
Hoerl, A. E., & Kennard, R. W. (1970). Ridge regression: Biased estimation for nonorthogonal problems.
Technometrics, 12(1), 55-67.
Horgan, T. (1992). Review Essay: From Cognitive Science to Folk Psychology: Computation, Mental
Representation, and Belief. Philosophy and Phenomenological Research, 449-484.
Hung, C. P., Kreiman, G., Poggio, T., & DiCarlo, J. J. (2005). Fast Readout of Object Identity from
Macaque Inferior Temporal Cortex. Science, 310(5749), 863-866. doi:10.1126/science.1117593
Immordino-Yang, M. H., McColl, A., Damasio, H., & Damasio, A. (2009). Neural correlates of admiration
and compassion. Proceedings of the National Academy of Sciences of the United States of America,
106(19), 8021-8026. doi:10.1073/pnas.0810363106
Ishai, A., Pessoa, L., Bikle, P. C., & Ungerleider, L. G. (2004). Repetition suppression of faces is
modulated by emotion. Proc Natl Acad Sci U S A, 101(26), 9827-9832.
Izuma, K., Saito, D. N., & Sadato, N. (2008). Processing of Social and Monetary Rewards in the Human
Striatum. Neuron, 58(2), 284-294. doi:10.1016/j.neuron.2008.03.020
Izuma, K., Saito, D.N., and Sadato, N. (2008). Processing of Social and Monetary Rewards in the Human
Striatum. Neuron 58, 284-294.
Jacques, P., Conway, M.A., Lowder, M.W., & Cabeza, R. (2011). Watching my mind unfold versus yours:
An fMRI study using a novel camera technology to examine neural differences in self-projection of
self vs. other perspectives. Journal of Cognitive Neuroscience 23(6): 1275-84
Jastorff, J., Clavagnier, S., Gergely, G., and Orban, G.A. (2011). Neural Mechanisms of Understanding
Rational Actions: Middle Temporal Gyrus Activation by Contextual Violation. Cereb Cortex 21,
318-329.
Jaswal, V. K., & Neely, L. A. (2006). Adults don't always know best: preschoolers use past reliability over
age when learning new words. Psychological Science, 17(9), 757-758. doi:10.1111/j.
1467-9280.2006.01778.x
Jehee, J. F., & Ballard, D. H. (2009). Predictive feedback can account for biphasic responses in the lateral
geniculate nucleus. PLoS computational biology, 5(5), e1000373.
162 of 179
Jellema, T., and Perrett, D.I. (2003a). Cells in monkey STS responsive to articulated body motions and
consequent static posture: a case of implied motion? Neuropsychologia 41, 1728-1737.
Jellema, T., and Perrett, D.I. (2003b). Perceptual history influences neural responses to face and body
postures. J Cogn Neurosci 15, 961-971.
Jellema, T., Baker, C.I., Wicker, B., and Perrett, D.I. (2000). Neural Representation for the Perception of
the Intentionality of Actions. Brain Cogn 44, 23-23.
Jenkins, A. C., & Mitchell, J. P. (2010). Mentalizing under uncertainty: dissociated neural responses to
ambiguous and unambiguous mental state inferences. Cerebral Cortex, 20(2), 404-410. doi:10.1093/
cercor/bhp109
Jenkins, A., Macrae, C., & Mitchell, J. (2008). Repetition suppression of ventromedial prefrontal activity
during judgments of self and others. Proceedings of the National Academy of Sciences 105(11):
4507.
Johanson, L., & Utas, B. (2000). Evidentials. Walter de Gruyter.
Jones, R.M., Somerville, L.H., Li, J., Ruberry, E.J., Libby, V., Glover, G., Voss, H.U., Ballon, D.J., and
Casey, B.J. (2011). Behavioral and Neural Properties of Social Reinforcement Learning. J Neurosci
31, 13039-13045.
Joseph, R. M., Tager-Flusberg, H., & Lord, C. (2002). Cognitive profiles and social-communicative
functioning in children with autism spectrum disorder. Journal of Child Psychology and Psychiatry,
43(6), 807-821.
Kable, J. W., & Chatterjee, A. (2006). Specificity of action representations in the lateral occipitotemporal
cortex. J Cogn Neurosci, 18(9), 1498-1517.
Kalbe, E., Schlegel, M., Sack, A.T., Nowak, D.A., Dafotakis, M., Bangard, C., et al. (2010). Dissociating
cognitive from affective theory of mind: a TMS study. Cortex, 46(6): 769-80.
Kamitani, Y., & Tong, F. (2005). Decoding the visual and subjective contents of the human brain. Nature
Neuroscience, 8(5), 679-685. doi:10.1038/nn1444
Kamitani, Y., & Tong, F. (2006). Decoding Seen and Attended Motion Directions from Activity in the
Human Visual Cortex. Current Biology, 16(11), 1096-1102. doi:10.1016/j.cub.2006.04.003
Kanwisher, N. (2010). Functional specificity in the human brain: a window into the functional architecture
of the mind. Proceedings of the National Academy of Sciences 107(25): 11163-70.
Kanwisher, N., McDermott, J., and Chun, M. (1997). The fusiform face area: a module in human
extrastriate cortex specialized for face perception. J Neurosci 17, 4302-4311.
Kassam, K. S., Markey, A. R., Cherkassky, V. L., Loewenstein, G., & Just, M. A. (2013). Identifying
Emotions on the Basis of Neural Activation. PLoS ONE, 8(6), e66032. doi:10.1371/journal.pone.
0066032
Keller, G.B., Bonhoeffer, T., and Hübener, M. (2012). Sensorimotor Mismatch Signals in Primary Visual
Cortex of the Behaving Mouse. Neuron 74, 809-815.
Kemp, C., & Tenenbaum, J. B. (2008). The discovery of structural form.Proceedings of the National
Academy of Sciences, 105(31), 10687-10692.
Kennedy, D. P., & Courchesne, E. (2008). Functional abnormalities of the default network during self- and
other-reflection in autism. Social Cognitive and Affective Neuroscience, 3(2), 177-190. doi:10.1093/
scan/nsn011
Keysers, C., & Gazzola, V. (2007). Integrating simulation and theory of mind: from self to social cognition.
Trends in Cognitive Sciences, 11(5), 194-196. doi:10.1016/j.tics.2007.02.002
163 of 179
Kilner, J. M., & Frith, C. D. (2007). Action observation: inferring intentions without mirror neurons. Current
Biology, 18(1), R32-3. doi:10.1016/j.cub.2007.11.008
Kilner, J. M., Paulignan, Y., & Blakemore, S. J. (2003). An Interference Effect of Observed Biological
Movement on Action. Current Biology, 13(6), 522-525. doi:10.1016/S0960-9822(03)00165-9
Kincaid, J. P., Fishburne, R. P., Jr, Rogers, R. L., & Chissom, B. S. (1975). Derivation of new readability
formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted
personnel.
Knight, R.T. (2007). Neural networks debunk phrenology. Science 316(5831): 1578-9.
Kobayashi, C., Glover, G. H., & Temple, E. (2007). Children's and adults' neural bases of verbal and
nonverbal 'theory of mind'. Neuropsychologia, 45(7), 1522-1532. doi:10.1016/j.neuropsychologia.
2006.11.017
Kobayashi, C., Glover, G.H., & Temple, E. (2007). Children's and adults'neural bases of verbal and
nonverbal 'theory of mind'. Neuropsychologia 45: 1522-32.
Koenig, M. A., & Harris, P. L. (2005a). Preschoolers mistrust ignorant and inaccurate speakers. Child
Development, 76(6), 1261-1277. doi:10.1111/j.1467-8624.2005.00849.x
Koenig, M. A., & Harris, P. L. (2005b). The role of social cognition in early trust. Trends in Cognitive
Sciences, 9(10), 457-459. doi:10.1016/j.tics.2005.08.006
Koenig, M. A., Clément, F. & Harris, P. L. (2004). Trust in testimony: children's use of true and false
statements. Psychological Science, 15(10), 694-698. doi:10.1111/j.0956-7976.2004.00742.x
Koenigs, M., Young, L.L., Adolphs, R., Tranel, D., Cushman, F., Hauser, M., et al. (2007). Damage to the
prefrontal cortex increases utilitarian moral judgements. Nature 446(7138): 908-11.
Kok, P. P., Jehee, J. F. M. J., & de Lange, F. P. F. (2012). Less is more: expectation sharpens
representations in the primary visual cortex. Neuron, 75(2), 265-270. doi:10.1016/j.neuron.
2012.04.034
Kok, P.P., Jehee, J.F.M.J., and de Lange, F.P.F. (2012a). Expectation Sharpens Representations in the
Primary Visual Cortex. Neuron 75, 265-270.
Kok, P.P., Rahnev, D.D., Jehee, J.F.M.J., Lau, H.C.H., and de Lange, F.P.F. (2012b). Attention reverses
the effect of prediction in silencing sensory signals. Cereb Cortex 22, 2197-2206.
Konkle, T., & Caramazza, A. (2013). Tripartite Organization of the Ventral Stream by Animacy and Object
Size. The Journal of Neuroscience : the Official Journal of the Society for Neuroscience, 33(25),
10235-10242. doi:10.1523/JNEUROSCI.0983-13.2013
Konkle, T., & Oliva, A. (2012). A real-world size organization of object responses in occipto-temporal
cortex. Neuron 74(6): 1114-24.
Kornblith, H. (1983). Justified belief and epistemically responsible action. The Philosophical Review.
Koshino, H., Kana, R. K., Keller, T. A., Cherkassky, V. L., Minshew, N. J., & Just, M. A. (2007). fMRI
Investigation of Working Memory for Faces in Autism: Visual Coding and Underconnectivity with
Frontal Areas. Cerebral Cortex, 18(2), 289-300. doi:10.1093/cercor/bhm054
Koster-Hale, J., & Saxe, R.R. (2011). Theory of Mind brain regions are sensitive to the content, not the
structural complexity, of belief attributions. In: L. Carlson, C. Hoelscher, & T. F. Shipley (Eds),
Proceedings of the 33rd Annual Cognitive Science Society Conference, pp. 3356-61.
Koster-Hale, J., & Saxe, R.R. (2013). Functional neuroimaging of theory of mind. In S. Baron-Cohen, H.
Tager-Flusberg, & M. V. Lombardo, Understanding Other Minds Perspectives from developmental
social neuroscience (pp. 132-163). Oxford University Press.
164 of 179
Koster-Hale, J., & Saxe, R.R. (2013). Theory of mind: a neural prediction problem. Neuron, 79(5), 836–
848. doi:10.1016/j.neuron.2013.08.020
Koster-Hale, J., Bedny, M., & Saxe, R.R. (2014). Thinking about seeing: Perceptual sources of knowledge
are encoded in the theory of mind brain regions of sighted and blind adults. Cognition, 133(1), 65-78.
doi:10.1016/j.cognition.2014.04.006
Koster-Hale, J., Saxe, R.R., Dungan, J., & Young, L.L. (2013). Decoding moral judgments from neural
representations of intentions. Proceedings of the National Academy of Sciences of the United States
of America, 110(14), 5648-5653. doi:10.1073/pnas.1207992110
Kourtzi, Z. & Kanwisher, N. (2001). Representation of perceived object shape by the human lateral
occipital complex. Science 293(5534): 1506.
Kourtzi, Z., and Kanwisher, N. (2000). Activation in human MT/MST by static images with implied motion.
J Cogn Neurosci 12, 48-55.
Krall, S. C., Rottschy, C., Oberwelland, E., Bzdok, D., Fox, P. T., Eickhoff, S. B., et al. (2014). The role of
the right temporoparietal junction in attention and social interaction as revealed by ALE metaanalysis. Brain Structure and Function. doi:10.1007/s00429-014-0803-z
Kriegeskorte, N. (2008). Representational similarity analysis - connecting the branches of systems
neuroscience. Frontiers in Systems Neuroscience. doi:10.3389/neuro.06.004.2008
Kriegeskorte, N., & Bandettini, P. (2007). Analyzing for information, not activation, to exploit highresolution fMRI. NeuroImage, 38(4), 649-662. doi:10.1016/j.neuroimage.2007.02.022
Kriegeskorte, N., & Kievit, R. A. (2013). Representational geometry: integrating cognition, computation,
and the brain. Trends in Cognitive Sciences, 17(8), 401-412. doi:10.1016/j.tics.2013.06.007
Kriegeskorte, N., Bandettini, P., & Bandettini, P. (2007). Analyzing for information, not activation, to exploit
high-resolution fMRI. NeuroImage, 38(4), 649-662. doi:10.1016/j.neuroimage.2007.02.022
Kriegeskorte, N., Goebel, R., & Bandettini, P.A. (2006). Information-based functional brain mapping.
Proceedings of the National Academy of Sciences USA 103(10), 3863-8.
Kriegeskorte, N., Mur, M., Ruff, D. A., Kiani, R., Bodurka, J., Esteky, H., ... & Bandettini, P. A. (2008).
Matching categorical object representations in inferior temporal cortex of man and monkey. Neuron,
60(6), 1126-1141.
Krienen, F.M., Tu, P.C., & Buckner, R.L. (2010). Clan mentality: Evidence that the medial prefrontal cortex
responds to close others. Journal of Neuroscience 30(41): 13906-15.
Krömer, N., Schöfer, J., & Boulesteix, A.-L. (2009). Regularized estimation of large-scale gene association
networks using graphical Gaussian models. BMC Bioinformatics, 10(1), 384. doi:
10.1186/1471-2105-10-384
Krueger, J., and Clement, R. W. (1994). The truly false consensus effect: An ineradicable and edocentric
bias in social perception. J Pers Soc Psychol, 67, 596 - 610.
Kubit, B., & Jack, A. I. (2013). Rethinking the role of the rTPJ in attention and social cognition in light of
the opposing domains hypothesis: findings from an ALE-based meta-analysis and resting-state
functional connectivity. Frontiers in Human Neuroscience, 7, 323. doi:10.3389/fnhum.2013.00323
Lamm, C., Batson, C.D., & Decety, J. (2007). The neural substrate of human empathy: effects of
perspective-taking and cognitive appraisal. Journal of Cognitive Neuroscience 19(1): 42-58.
Landau, B., & Gleitman, L. R. (1985). Language and Experience. Harvard University Press.
Lane, R.D., Chua, P.M., and Dolan, R.J. (1999). Common effects of emotional valence, arousal and
attention on neural activation during visual processing of pictures. Neuropsychologia 37, 989-997.
165 of 179
Lawrence, E.J., Shaw, P., Baker, D., Baron-Cohen, S., & David, A.S. (2004). Measuring empathy:
reliability and validity of the empathy quotient. Psychological Medicine, 34(5): 911-24.
Lazebnik, S., Schmid, C., & Ponce, J. (2006). Beyond bags of features: Spatial pyramid matching for
recognizing natural scene categories. In Computer Vision and Pattern Recognition, 2006 IEEE
Computer Society Conference on (Vol. 2, pp. 2169-2178). IEEE.
Lenci, A., Baroni, M., Cazzolli, G., & Marotta, G. (2013). BLIND: a set of semantic feature norms from the
congenitally blind. Behavior Research Methods, 1-16. doi:10.3758/s13428-013-0323-4
Leshinskaya, A., & Caramazza, A. (2014). Nonmotor Aspects of Action Concepts. Journal of Cognitive
Neuroscience, 15, 1-17. doi:10.1162/jocn-a-00679
Leslie, A., Mallon, R., & DiCorcia, J. (2006). Transgressors, victims, and cry babies: Is basic moral
judgment spared in autism? Social Neuroscience, 1(3-4), 270-283.
Leslie, A.M., & Frith, U. (1990). Prospects for a cognitive neuropsychology of autism: Hobson's choice.
Psychological Review 97(1): 122-31.
Lin, A., Adolphs, R., and Rangel, A. (2012). Social and monetary reward learning engage overlapping
neural substrates. Soc Cogn Affect Neurosci 7, 274-281.
Lindsay, D. S., & Johnson, M. K. (1989). The eyewitness suggestibility effect and memory for source.
Memory and Cognition, 17(3), 349-358. doi:10.1016/j.jml.2003.08.007
Liu, D., Sabbagh, M., A., Gehring, W.J., & Wellman, H.M. (2004). Decoupling beliefs from reality in the
brain: an ERP study of theory of mind. Neuroreport 15(6): 991-5.
Logothetis, N. K., & Sheinberg, D. L. (1996). Visual object recognition. Ann Rev Neurosci, 19(1), 577-621.
Logothetis, N.K. (2008). What we can do and what we cannot do with fMRI. Nature, 453(7197): 869-78.
Lombardo, M. V., Chakrabarti, B., Bullmore, E. T., Baron-Cohen, S., & Consortium, M. A. (2011).
Specialization of right temporo-parietal junction for mentalizing and its relation to social impairments
in autism. NeuroImage, 56(3), 1832-1838. doi:10.1016/j.neuroimage.2011.02.067
Lombardo, M. V., Chakrabarti, B., Bullmore, E. T., Wheelwright, S. J., Sadek, S. A., Suckling, J., & BaronCohen, S. L. (2010). Shared neural circuits for mentalizing about the self and others. Journal of
Cognitive Neuroscience, 22(7), 1623-1635. doi:10.1162/jocn.2009.21287
Lombardo, M.V., Chakrabarti, B., Bullmore, E.T., Wheelwright, S.J., Sadek, S.A., Suckling, J., et al.
(2010). Shared neural circuits for mentalizing about the self and others. Journal of Cognitive
Neuroscience, 22(7): 1623-35.
Lord, C., Risi, S., Lambrecht, L., Cook, E. H., Leventhal, B. L., DiLavore, P.C., et al. (2000). The Autism
Diagnostic Observation Schedule„Generic: A standard measure of social and communication deficits
associated with the spectrum of autism. Journal of Autism and Developmental Disorders, 30(3),
205-223. doi:10.1023/A:1005592401947
Lucas, A. J., Lewis, C., Pala, F. C., Wong, K., & Berridge, D. (2013). Social-cognitive processes in
preschoolers' selective trust: Three cultures compared. Developmental Psychology, 49(3), 579-590.
doi:10.1037/a0029864
Ma, N., Baetens, K., Vandekerckhove, M., Kestemont, J., Fias, W., & Van Overwalle, F. (2013a). Traits
are represented in the medial prefrontal cortex: an fMRI adaptation study. Social Cognitive and
Affective Neuroscience. doi:10.1093/scan/nst098
Ma, N., Baetens, K., Vandekerckhove, M., Van der Cruyssen, L., & Van Overwalle, F. (2013b).
Dissociation of a trait and a valence representation in the mPFC. Social Cognitive and Affective
Neuroscience. doi:10.1093/scan/nst143
Ma, N., Vandekerckhove, M., Baetens, K., Van Overwalle, F., Seurinck, R., and Fias, W. (2012).
Inconsistencies in spontaneous and intentional trait inferences. Soc Cogn Affect Neurosci 7, 937-950.
166 of 179
Mahon, B. Z., & Caramazza, A. (2010). Judging semantic similarity: an event-related fMRI study with
auditory word stimuli. Nsc, 169(1), 279-286. doi:10.1016/j.neuroscience.2010.04.029
Malle, B. F. (1999). How people explain behavior: A new theoretical framework. Pers Soc Psychol Rev
3(1), 23-48.
Manera, V., Becchio, C., Ben Schouten, Bara, B.G., and Verfaillie, K. (2011). Communicative interactions
improve visual detection of biological motion. PLoS ONE 6, e14594-e14594.
Marmor, G. S. (n.d.). Age at onset of blindness and the development of the semantics of color names.
Journal of Experimental Child Psychology, 25(2), 267-278.
Mars, R. B., Sallet, J., Schüffelgen, U., Jbabdi, S., Toni, I., & Rushworth, M. F. S. (2012). Connectivitybased subdivisions of the human right 'temporoparietal junction area': evidence for different areas
participating in different cortical networks. Cerebral Cortex (New York, N.Y. : 1991), 22(8), 1894-1903.
doi:10.1093/cercor/bhr268
Mars, R.B., Sallet, J., Neubert, F.-X., and Rushworth, M.F.S. (2013). Connectivity profiles reveal the
relationship between brain areas for social cognition in human and monkey temporoparietal cortex.
Proc Natl Acad Sci U S A. doi:10.1073/pnas.1302956110
Martinez-Trujillo, J.C.J., and Treue, S.S. (2004). Feature-Based Attention Increases the Selectivity of
Population Responses in Primate Visual Cortex. Curr Biol 14, 8-8.
Mason, R. A., & Just, M. A. (2011). Differentiable cortical networks for inferences concerning people's
intentions versus physical causality. Human Brain Mapping, 32(2), 313-329. doi:10.1002/hbm.21021
Mason, R. A., Williams, D. L., Kana, R. K., Minshew, N., & Just, M. A. (2008). Theory of Mind disruption
and recruitment of the right hemisphere during narrative comprehension in autism. Neuropsychologia,
46(1), 269-280. doi:10.1016/j.neuropsychologia.2007.07.018
Mason, R.A., & Just, M.A. (2010). Differentiable cortical networks for inferences concerning people's
intentions vs. physical causality. Human Brain Mapping, 32(2): 313-29.
Maunsell, J. H., and Newsome, W. T. (1987). Visual processing in monkey extrastriate cortex. Ann Rev
Neurosci 10(1), 363-401.
McAlpine, L., & Moore, C. D. (1995). The development of social understanding in children with visual
impairments. Journal of Visual Impairment and Blindness, (89), 349-358.
McNamara, D. S., Graesser, A. C., McCarthy, P. M., & Cai, Z. (2014). Automated Evaluation of Text and
Discourse with Coh-Metrix. Cambridge University Press.
Meltzoff, A. N. (2004). The case for a developmental cognitive science: Theories of people and things. In
G. Brenmner & A. Slater (Eds.), Theories of infant development (pp. 145-173). Oxford: Blackwell.
Meltzoff, A. N. (2007). 'Like me': a foundation for social cognition. Developmental Science, 10(1),
126-134. doi:10.1111/j.1467-7687.2007.00574.x
Meltzoff, A. N., & Brooks, R. (2007). Eyes wide shut: The importance of eyes in infant gaze following and
understanding of other minds. In R. Flom, K. Lee, & D. Muir (Eds.), Gaze Following: its development
and signficance (pp. 217-241). Eyes wide shut: The importance of eyes in infant gaze following and
understanding of other minds.: Eyes wide shut: The importance of eyes in infant gaze following and
understanding of other minds.
Mende-Siedlecki, P., Cai, Y., and Todorov, A. (2012). The neural dynamics of updating person
impressions. Soc Cogn Affect Neurosci. doi: 10.1093/scan/nss040
Meristo, M., Falkman, K. W., Hjelmquist, E., Tedoldi, M., Surian, L., & Siegal, M. (2007). Language access
and theory of mind reasoning: Evidence from deaf children in bilingual and oralist environments.
Developmental Psychology, 43(5), 1156-1169. doi:10.1037/0012-1649.43.5.1156
167 of 179
Meyer, M., and Sauerland, U. (2009). A pragmatic constraint on ambiguity detection. Nat Lang Linguist
Theory 27, 139-150.
Meyer, T., and Olson, C. R. (2011). Statistical learning of visual transitions in monkey inferotemporal
cortex. Proc Natl Acad Sci U S A, 108(48), 19401-19406.
Miene, P., Bordiga, E., & Park, R. (1993). The Evaluation of Hearsay Evidence: A Social Psychological
Approach. In N. J. Castellan (Ed.), Individual and Group Decision Making. Psychology Press.
Miene, P., Park, R. C., & Bordiga, E. (1992). Juror Decision Making and the Evaluation of Hearsay
Evidence. 76 Minn . L. Rev, (683).
Millar, S. (1976). Spatial representation by blind and sighted children. Journal of Experimental Child
Psychology, 21(3), 460-479.
Miller, E.K., & Cohen, J.D. (2001). An integrative theory of prefrontal cortex function. Annual Review of
Neuroscience, 24: 167-202.
Minter, M., Hobson, R. P., & B, M. (1998). Congenital visual impairment and 'theory of mind.' British
Journal of Developmental Psychology, (16), 183-196.
Mitchell, J., & Banaji, M. (2005). The link between social cognition and self-referential thought in the
medial prefrontal cortex. Journal of Cognitive Neuroscience.
Mitchell, J., Macrae, C.N., & Banaji, M.R. (2006). Dissociable medial prefrontal contributions to judgments
of similar and dissimilar others. Neuron, 50(4): 655-63.
Mitchell, J.P. (2008). Activity in right temporo-parietal junction is not selective for theory-of-mind. Cerebral
Cortex, 18(2): 262-71.
Mitchell, J.P., Banaji, M.R., & MacRae, C.N. (2005). The link between social cognition and self-referential
thought in the medial prefrontal cortex. Journal of Cognitive Neuroscience 17(8): 1306-15.
Mitchell, J.P., Macrae, C.N., and Banaji, M.R. (2006). Dissociable Medial Prefrontal Contributions to
Judgments of Similar and Dissimilar Others. Neuron 50, 655-663.
Mitchell, T. M., Hutchinson, R., Niculescu, R. S., Pereira, F., Wang, X., Just, M., & Newman, S. (2004).
Learning to Decode Cognitive States from Brain Images. Machine Learning, 57(1/2), 145-175. doi:
10.1023/B:MACH.0000035475.85309.1b
Moeller, M., & Schick, B. (2006). Relations between maternal input and theory of mind understanding in
deaf children. Child Development, 77(3), 751-766.
Moran, J. M., Young, L.L., Saxe, R.R., Lee, S. M., O'Young, D., Mavros, P. L., & Gabrieli, J. D. (2011).
Impaired theory of mind for moral judgment in high-functioning autism. Proceedings of the National
Academy of Sciences of the United States of America, 108(7), 2688-2692. doi:10.1073/pnas.
1011734108
Moran, J.M., Lee, S.M., & Gabrieli, J.D.E. (2011a). Dissociable neural systems supporting knowledge
about human character and appearance in ourselves and others. Journal of Cognitive Neuroscience,
23(9): 2222-30.
Moran, J.M., Young, L.L., Saxe, R.R., Lee, S.M., O'Young, D., Mavros, P.L., & Gabrieli, J.D. (2011b).
Impaired theory of mind for moral judgment in high-functioning autism. Proceedings of the National
Academy of Sciences, 108(7): 2688-92.
Moriguchi, Y., Ohnishi, T., Lane, R. D., Maeda, M., Mori, T., Nemoto, K., et al. (2006). Impaired selfawareness and theory of mind: An fMRI study of mentalizing in alexithymia. NeuroImage, 32(3),
1472-1482. doi:10.1016/j.neuroimage.2006.04.186
Morton, J. (1994). Cognitive perspectives on memory recovery. Applied Cognitive Psychology, 8, 389-398.
168 of 179
Müller, N.G., Strumpf, H., Scholz, M., Baier, B., and Melloni, L. (2013). Repetition suppression versus
enhancement—it's quantity that matters. Cereb Cortex 23, 315-322.
Mumford, D. (1992). On the computational architecture of the neocortex. II. The role of cortico-cortical
loops. Biol Cybern 66, 241-251.
Murray, S.O., Kersten, D., Olshausen, B.A., Schrater, P., and Woods, D.L. (2002). Shape perception
reduces activity in human primary visual cortex. Proc Natl Acad Sci U S A 99, 15164-15169.
Nakahara, H., Itoh, H., Kawagoe, R., Takikawa, Y., & Hikosaka, O. (2004). Dopamine neurons can
represent context-dependent prediction error. Neuron,41(2), 269-280.
Naselaris, T., Kay, K. N., Nishimoto, S., & Gallant, J. L. (2011). Encoding and decoding in fMRI.
NeuroImage, 56(2), 400-410. doi:10.1016/j.neuroimage.2010.07.073
Neri, P.P., Luu, J.Y.J., and Levi, D.M.D. (2006). Meaningful interactions can enhance visual discrimination
of human agents. Nat Neurosci 9, 1186-1192.
Nichols, S., & Ulatowski, J. (2007). Intuitions and Individual Differences: The Knobe Effect Revisited. Mind
& Language, 22(4), 346-365. doi:10.1111/j.1468-0017.2007.00312.x
Nichols, S., Stich, S., Leslie, A., & Klein, D. (1996). Varieties of Off-Line Simulation. Theories of Theories
of Mind.
Nichols, T. (2002). Nonparametric permutation tests for functional neuroimaging: a primer with examples.
Human Brain Mapping.
Nieminen-von Wendt, T., Mets honkala, L., Kulom ki, T., Alto, S., Autti, T., Vanhala, R., & Wendt, von, L.
(2003). Changes in cerebral blood flow in Asperger syndrome during theory of mind tasks presented
by the auditory route. European Child & Adolescent Psychiatry, 12(4), 178-189. doi:10.1007/
s00787-003-0337-z
Norman, K., Polyn, S., Detre, G., & Haxby, J. (2006). Beyond mind-reading: multi-voxel pattern analysis of
fMRI data. Trends in Cognitive Sciences, 10(9), 424-430.
Norman-Haignere S, McDermott JH, Kanwisher N (February 2014). Data-driven clustering of cortical
responses to natural sounds reveals organization for pitch and speech. Poster presented at the
Computational and Systems Neuroscience meeting (COSYNE), Salt Lake City, UT.
O'Doherty, J., Kringelbach, M. L., Rolls, E. T., Hornak, J., & Andrews, C. (2001). Abstract reward and
punishment representations in the human orbitofrontal cortex. Nature neuroscience, 4(1), 95-102.
O'Neill, D. K., & Chong, S. C. (2001). Preschool children's difficulty understanding the types of information
obtained through the five senses. Child Development, 72(3), 803-815. doi:10.2307/1132456
O'Neill, D. K., & Gopnik, A. (1991). Young children's ability to identify the sources of their beliefs.
Developmental Psychology, 27(3), 390-397.
Olson, G. (2003). Reconsidering Unreliability: Fallible and Untrustworthy Narrators. Narrative, 11(1),
93-109. doi:10.1353/nar.2003.0001
Onishi, K., & Baillargeon, R. (2005). Do 15-month-old infants understand false beliefs? Science,
308(5719): 255.
Onishi, K., Baillargeon, R., & Baillargeon, R. (2005). Do 15-month-old infants understand false beliefs?
Science, 308(5719), 255.
Ozonoff, S., Pennington, B. F., & Rogers, S. J. (1991). Executive function deficits in high‐functioning
autistic individuals: relationship to theory of mind.Journal of child Psychology and Psychiatry, 32(7),
1081-1105.
169 of 179
Papafragou, A., & Li, P. (2001). Evidential morphology and theory of mind. Presented at the Proceedings
from the 26th Annual Boston University Conference on Language Development, Cascadilla Press,
Somerville, MA.
Park, R. C. (1987). A Subject Matter Approach to Hearsay Reform. 86 Mich. L. Rev. (51).
Patel, D., Fleming, S.M., and Kilner, J.M. (2012). Inferring subjective states through the observation of
actions. Philos Trans R Soc Lond B Biol Sci 279, 4853-4860.
Pavlova, M., Staudt, M., Sokolov, A., Birbaumer, N., & Krögeloh-Mann, I. (2003). Perception and
production of biological movement in patients with early periventricular brain lesions. Brain, 126(3),
692-701.
Peelen, M. V., Atkinson, A. P., & Vuilleumier, P. (2010). Supramodal Representations of Perceived
Emotions in the Human Brain. Journal of Neuroscience, 30(30), 10127-10134. doi:10.1523/
JNEUROSCI.2161-10.2010
Peelen, M. V., Wiggett, A. J., & Downing, P. E. (2006). Patterns of fMRI Activity Dissociate Overlapping
Functional Brain Areas that Respond to Biological Motion. Neuron, 49(6), 815-822. doi:10.1016/
j.neuron.2006.02.004
Peelen, M.V.M., and Kastner, S.S. (2011). A neural basis for real-world visual search in human
occipitotemporal cortex. Proc Natl Acad Sci U S A 108, 12125-12130.
Pelphrey, K.A., & Morris, J.P. (2006). Brain mechanisms for interpreting the actions of others from
biological-motion cues. Current Directions in Psychological Science, 15(3): 136.
Pelphrey, K.A., and Vander Wyk, B.C. (2011). Functional and neural mechanisms for eye gaze
processing. In OUP Handbook of Face Perception, A Calder, G Rhodes, M Johnson, J Haxby, eds.
(Oxford University Press), pp:591-604.
Pelphrey, K.A., Mitchell, T.V., McKeown, M.J., Goldstein, J., Allison, T., & McCarthy, G. (2003). Brain
activity evoked by the perception of human walking: Controlling for meaningful coherent motion.
Journal of Neuroscience, 23, 6819-25.
Pelphrey, K.A., Morris, J.P., and McCarthy, G. (2004). Grasping the intentions of others: the perceived
intentionality of an action influences activity in the superior temporal sulcus during social perception. J
Cogn Neurosci 16, 1706-1716.
Pelphrey, K.A., Morris, J.P., Michelich, C.R., Allison, T., & McCarthy, G. (2005). Functional anatomy of
biological motion perception in posterior temporal cortex: An fMRI study of eye, mouth and hand
movements. Cerebral Cortex, 15(12): 1866-76.
Pereira, F., & Mitchell, T. (2009). Machine learning classifiers and fMRI: A tutorial overview. NeuroImage.
Perner, J. (1991). Understanding the Representational Mind. Cambridge: MIT Press.
Perner, J. J., Aichhorn, M. M., Kronbichler, M. M., Staffen, W. W., & Ladurner, G. G. (2006). Thinking of
mental and other representations: the roles of left and right temporo-parietal junction. Social
Neuroscience, 1(3-4), 245-258. doi:10.1080/17470910600989896
Perner, J., & Lang, B. (1999). Development of theory of mind and executive control. Trends in cognitive
sciences, 3(9), 337-344.
Perner, J., & Leekam, S. (2008). The curious incident of the photo that was accused of being false: issues
of domain specificity in development, autism, and brain imaging. Quarterly Journal of Experimental
Psychology, 61(1), 76-89. doi:10.1080/17470210701508756
Perner, J., & Roessler, J. (2012). From infantsÒ to childrenÓs appreciation of belief. Trends in Cognitive
Sciences, 16(10), 519-525. doi:10.1016/j.tics.2012.08.004
170 of 179
Perner, J., Aichhorn, M., Kronbichler, M., Staffen, W., & Ladurner, G. (2006). Thinking of mental and other
representations: The roles of left and right temporo-parietal junction. Social Neuroscience 1(3-4):
245-58.
Perrett, D.I.D., Smith, P.A.P., Mistlin, A.J.A., Chitty, A.J.A., Head, A.S.A., Potter, D.D.D., Broennimann,
R.R., Milner, A.D.A., and Jeeves, M.A.M. (1985). Visual analysis of body movements by neurones in
the temporal cortex of the macaque monkey: a preliminary report. Behav Brain Res 16, 153-170.
Peter Miene, Bordiga, E., & Park, R. (1993). The Evaluation of Hearsay Evidence: A Social Psychological
Approach. In N. J. Castellan, Individual and Group Decision Making. Psychology Press.
Peter Miene, Park, R. C., & Bordiga, E. (1992). Juror Decision Making and the Evaluation of Hearsay
Evidence. 76 Minn . L. Rev, (683).
Peterson C., & Wellman, H. M. (2010). From fancy to reason: scaling deaf and hearing children's
understanding of theory of mind and pretence. British Journal of Developmental Psychology, 27(Pt 2),
297-310. doi:10.1348/026151008X299728
Peterson, & Siegal, M. (1999). Representing Inner Worlds: Theory of Mind in Autistic, Deaf, and Normal
Hearing Children. Psychological Science, 10(2), 126-129. doi:10.1111/1467-9280.00119
Peterson, C., Wellman, H. M., & Liu, D. (2005). Steps in theory-of-mind development for children with
deafness or autism. Child Development, 76(2), 502-517.
Peterson, J. L., & Webb, J. (2000). Factors influencing the development of a theory of mind in blind
c h i l d r e n . B r i t i s h J o u r n a l o f D e v e l o p m e n t a l P s y c h o l o g y, 1 8 ( 3 ) , 4 3 1 - 4 4 7 . d o i :
10.1348/026151000165788
Philip, R. C. M., Whalley, H. C., Stanfield, A. C., Sprengelmeyer, R., Santos, I. M., Young, A. W., et al.
(2010). Deficits in facial, body movement and vocal emotional processing in autism spectrum
disorders. Psychological Medicine, 40(11), 1919-1929. doi:10.1017/S0033291709992364
Pillow, B. H., Hill, V., Boyce, A., & Stein, C. (2000). Understanding inference as a source of knowledge:
children's ability to evaluate the certainty of deduction, perception, and guessing. Developmental
Psychology, 36(2), 169-179.
Platek, S. (2004). Where am I? The neurological correlates of self and other. Cognitive Brain Research,
19(2), 114-122. doi:10.1016/j.cogbrainres.2003.11.014
Platek, S., Keenan, J.P., Gallup Jr, G.G., & Mohamed, F.B. (2004). Where am I? The neurological
correlates of self and other. Cognitive Brain Research 19(2): 114-22.
Poldrack, R. (2006). Can cognitive processes be inferred from neuroimaging data? Trends in Cognitive
Sciences, 10(2), 59-63. doi:10.1016/j.tics.2005.12.004
Poldrack, R.A. (2006b). Region of interest analysis for fMRI. Social Cognitive and Affective Neuroscience
2(1): 67-70.
Poore, J. C., Pfeifer, J. H., Berkman, E. T., Inagaki, T. K., Welborn, B. L., & Lieberman, M. D. (2012).
Prediction-error in the context of real social relationships modulates reward system activity. Frontiers
in Human Neuroscience, 6, 1-11. doi:10.3389/fnhum.2012.00218
Puce, A.A., Allison, T.T., Bentin, S.S., Gore, J.C.J., and McCarthy, G.G. (1998). Temporal cortex activation
in humans viewing eye and mouth movements. J Neurosci 18, 2188-2199.
Pyers, J., & Senghas, A. (2009). Language Promotes False-Belief Understanding. Psychological Science,
20(7), 805.
Quian Quiroga, R., & Panzeri, S. (2009). Extracting information from neuronal populations: information
theory and decoding approaches. Nature Reviews Neuroscience, 10(3), 173-185. doi:10.1038/
nrn2578
171 of 179
R Core Team. (2013). R: a language and environment for statistical computing. Vienna: R Foundation for
Statistical Computing.
Rabin, J. S., Gilboa, A., Stuss, D. T., Mar, R. A., & Rosenbaum, R. S. (2010). Common and unique neural
correlates of autobiographical memory and theory of mind. Journal of Cognitive Neuroscience, 22(6),
1095-1111. doi:10.1162/jocn.2009.21344
Raizada, R. D. S., Tsao, F.-M., Liu, H.-M., Holloway, I. D., Ansari, D., & Kuhl, P. K. (2010). Linking brainwide multivoxel activation patterns to behaviour: Examples from language and math. NeuroImage,
51(1), 462-471. doi:10.1016/j.neuroimage.2010.01.080
Rao, R.P., and Ballard, D.H. (1999). Predictive coding in the visual cortex: a functional interpretation of
some extra-classical receptive-field effects. Nat Neurosci 2, 79-87.
Rayner, K., & Duffy, S. A. (1986). Lexical complexity and fixation times in reading: effects of word
frequency, verb complexity, and lexical ambiguity. Memory and Cognition, 14(3), 191-201.
Repacholi, B. M., & Gopnik, A. (1997). Early reasoning about desires: evidence from 14- and 18-montholds. Developmental Psychology, 33(1), 12-21.
Reynolds, J.H., and Heeger, D.J. (2009). The Normalization Model of Attention. Neuron 61, 168-185.
Robinson, E. J. (1994). What people say, what they think, and what is really the case: Children's
understanding of utterances as sources of knowledge. In C. Lewis & P. Mitchell, (pp. 355-381).
Children's early understanding of mind: Origins ƒ.
Robinson, E. J., & Whitcombe, E. L. (2003). Children's Suggestibility in Relation to their Understanding
about Sources of Knowledge. Child Development, 74(1), 48-62. doi:10.1111/1467-8624.t01-1-00520
Robinson, E. J., Champion, H., & Mitchell, P. (1999). Children's ability to infer utterance veracity from
speaker informedness. Developmental Psychology, 35(2), 535-546.
Robinson, E. J., Haigh, S. N., & Nurmsoo, E. (2008). Children's working understanding of knowledge
sources: Confidence in knowledge gained from testimony. Cognitive Development, 23(1), 105-118.
doi:10.1016/j.cogdev.2007.05.001
Rolls, E. T., & Treves, A. (2011). The neuronal encoding of information in the brain. Progress in
Neurobiology, 95(3), 448-490. doi:10.1016/j.pneurobio.2011.08.002
Rosel, J., Caballer, A., Jara, P., & Oliver, J. C. (2005). Verbalism in the Narrative Language of Children
Who Are Blind and Sighted. Journal of Visual Impairment and Blindness, 99, 413-425.
Ross, D. R., Ceci, S. J., Dunning, D., & Toglia, M. P. (1994). Unconscious transference and mistake
identity: When a witness misidentifies a familiar with innocent person. Journal of Applied Psychology,
79(6), 918-930.
Ross, L., Greene, D., & House, P. (1977). The false consensus effect: An egocentric bias in social
perception and attribution processes. J Exp Soc Psychol, 13, 279 -301.
Rushworth, M. F., Mars, R. B., & Sallet, J. (2013). Are there specialized circuits for social cognition and
are they unique to humans?. Curr Opin Neurobiol 23(3) 436-442
Samson, D., Apperly, I. A., Braithwaite, J. J., Andrews, B. J., & Bodley Scott, S. E. (2010). Seeing it their
way: Evidence for rapid and involuntary computation of what other people see. Journal of
Experimental Psychology: Human Perception and Performance, 36(5), 1255-1266. doi:10.1037/
a0018729
Samson, D., Apperly, I.A., & Humphreys, G.W. (2007). Error analyses reveal contrasting deficits in 'theory
of mind': Neuropsychological evidence from a 3-option false belief task. Neuropsychologia 45(11):
2561-9.
Santos, L., Flombaum, J., and Phillips, W. (2013). The Evolution of Human Mindreading. Evol Cogn
Neurosci 362, 659-669.
172 of 179
Sapountzis, P., Schluppeck, D., Bowtell, R., & Peirce, J. W. (2010). A comparison of fMRI adaptation and
multivariate pattern classification analysis in visual cortex. Neuroimage, 49(2), 1632-1640.
Saxe, R.R. (2005). Against simulation: the argument from error. Trends in Cognitive Sciences, 9(4),
174-179. doi:10.1016/j.tics.2005.01.012
Saxe, R.R. (2006). Uniquely human social cognition. Curr Opin Neurobiol 16, 235-239.
Saxe, R.R. (2013). The new puzzle of theory of mind development. In M. Banaji & S. A. Gelman,
Navigating the Social World What Infants, Children, and Other Species Can Teach Us. Navigating the
Social World: What Infants.
Saxe, R.R., & Kanwisher, N. (2003). People thinking about thinking people:: The role of the temporoparietal junction in. NeuroImage, 19(4), 1835-1842.
Saxe, R.R., & Offen, S. (2010). Seeing ourselves: What vision can teach us about metacognition. In: G.
Dimaggio, & P.H. Lysaker (Eds), Metacognition and Severe Adult Mental Disorders: From basic
research to treatment (pp. 13-29). New York: Taylor & Francis.
Saxe, R.R., & Powell, L. J. (2006). It's the thought that counts: specific brain regions for one component
of theory of mind. Psychological Science, 17(8), 692-699. doi:10.1111/j.1467-9280.2006.01768.x
Saxe, R.R., & Wexler, A. (2005). Making sense of another mind: The role of the right temporo-parietal
junction. Neuropsychologia, 43(10), 1391-1399. doi:10.1016/j.neuropsychologia.2005.02.013
Saxe, R.R., Moran, J.M., Scholz, J., & Gabrieli, J. (2006a). Overlapping and non-overlapping brain
regions for theory of mind and self reflection in individual subjects. Social Cognitive and Affective
Neuroscience 1(3): 229-34.
Saxe, R.R., Schulz, L. E., & Jiang, Y. V. (2006). Reading minds versus following rules: Dissociating theory
of mind and executive control in the brain. Social Neuroscience, 1(3-4), 284-298. doi:
10.1080/17470910601000446
Saxe, R.R., Whitfield-Gabrieli, S., Scholz, J., & Pelphrey, K.A. (2009). Brain regions for perceiving and
reasoning about other people in school-aged children. Child Development, 80(4): 1197-209.
Saxe, R.R., Xiao, D.-K., Kovacs, G., Perrett, D.I., and Kanwisher, N. (2004). A region of right posterior
superior temporal sulcus responds to observed intentional actions. Neuropsychologia 42, 12-12.
Saygin, A.P., Chaminade, T., Ishiguro, H., Driver, J., and Frith, C. (2012). The thing that should not be:
predictive coding and the uncanny valley in perceiving human and humanoid robot actions. Soc Cogn
Affect Neurosci 7, 413-422.
Schick, B., de Villiers, P., de Villiers, J., & Hoffmeister, R. (2007). Language and theory of mind: A study of
deaf children. Child Development, 78(2), 376-396.
Schiller, D., Freeman, J.B., Mitchell, J.P., Uleman, J.S., and Phelps, E.A. (2009). A neural mechanism of
first impressions. Nat Neurosci 12, 508-514.
Schneider, W., Schumann-Hengsteler, R., & Sodian, B. (Eds.). (2014). Young children's cognitive
development: Interrelationships among executive functioning, working memory, verbal ability, and
theory of mind. Psychology Press.
Schnell, K., Bluschke, S., Konradt, B., & Walter, H. (2010). Functional relations of empathy and
mentalizing: An fMRI study on the neural basis of cognitive empathy. NeuroImage, 1-41. doi:10.1016/
j.neuroimage.2010.08.024
Scholz, J., Triantafyllou, C., Whitfield-Gabrieli, S., Brown, E. N., & Saxe, R.R. (2009). Distinct Regions of
Right Temporo-Parietal Junction Are Selective for Theory of Mind and Exogenous Attention. PLoS
ONE, 4(3), e4869. doi:10.1371/journal.pone.0004869
Schultz, W. (2010). Dopamine signals for reward value and risk: basic and recent data. Behav Brain
Funct 6, 24.
173 of 179
Schultz, W., Dayan, P., and Montague, P.R. (1997). A neural substrate of prediction and reward. Science
275, 1593-1599.
Schurz, M., Radua, J., Aichhorn, M. M., Richlan, F., & Perner, J. (2014). Fractionating theory of mind: a
meta-analysis of functional brain imaging studies. Neuroscience & Biobehavioral Reviews, 42, 9-34.
doi:10.1016/j.neubiorev.2014.01.009
Scofield, J., & Behrend, D. A. (2008). Learning words from reliable and unreliable speakers. Cognitive
Development, 23(2), 278-290. doi:10.1016/j.cogdev.2008.01.003
Senior, C.C., Barnes, J.J., Giampietro, V.V., Simmons, A.A., Bullmore, E.T.E., Brammer, M.M., and David,
A.S.A. (2000). The functional neuroanatomy of implicit-motion perception or 'representational
momentum.' Curr Biol 10, 7-7.
Senju, A., Johnson, M.H., and Csibra, G. (2006). The development and neural basis of referential gaze
perception. Soc Neurosci 1, 220-234.
Seo, H.H., and Lee, D.D. (2012). Neural basis of learning and preference during social decision-making.
Curr Opin Neurobiol 22, 990-995.
Serences, J. T., & Boynton, G. M. (2007). Feature-Based Attentional Modulations in the Absence of Direct
Visual Stimulation. Neuron, 55(2), 301-312. doi:10.1016/j.neuron.2007.06.015
Serences, J.T., Shomstein, S., Leber, A.B., Golay, X., Egeth, H.E., & Yantis, S. (2005). Coordination of
voluntary and stimulus-driven attentional control in human cortex. Psychological Science 16:114-22.
Serre, T., Wolf, L., & Poggio, T. (2005, June). Object recognition with features inspired by visual cortex. In
Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE Computer Society Conference on
(Vol. 2, pp. 994-1000). IEEE.
Seung, H. S., & Sompolinsky, H. (1993). Simple models for reading neuronal population codes, 90(22),
10749-10753. doi:10.1073/pnas.90.22.10749
Shamay-Tsoory, S., Tomer, R., Berger, B.D., Goldsher, D., & Aharon-Peretz, J. (2005). Impaired affective
theory of mind is associated with right ventromedial prefrontal damage. Cognitive and Behavioral
Neurology 18(1): 55-67.
Shamay-Tsoory, S.G., & Aharon-Peretz, J. (2007). Dissociable prefrontal networks for cognitive and
affective theory of mind: a lesion study. Neuropsychologia 45(13): 3054-67.
Shamir, M., & Sompolinsky, H. (2006). Implications of neuronal diversity on population coding. Neural
Computation, 18(8), 1951-1986. doi:10.1162/neco.2006.18.8.1951
Shepard, R. N. (1980). Multidimensional scaling, tree-fitting, and clustering.Science, 210(4468), 390-398.
Shepard, R. N., & Cooper, L. A. (1992). Representation of Colors in the Blind, Color-Blind, and Normally
Sighted. Psychological Science, 3(2), 97-104. doi:10.1111/j.1467-9280.1992.tb00006.x
Shibata, H., Inui, T., & Ogawa, K. (2011). Understanding interpersonal action coordination: an fMRI study.
Exp Brain Res, 211(3-4), 569-579.
Shultz, S., Lee, S.M., Pelphrey, K., and McCarthy, G. (2011). The posterior superior temporal sulcus is
sensitive to the outcome of human and non-human goal-directed actions. Soc Cogn Affect Neurosci
6, 602-611.
Simanova, I., Hagoort, P., Oostenveld, R., & van Gerven, M. A. (2014). Modality-independent decoding of
semantic information from the human brain.Cerebral Cortex, 24(2), 426-434.
Singer, T., Ben Seymour, O'Doherty, J., Kaube, H., Dolan, R. J., & Frith, C. D. (2004). Empathy for pain
involves the affective but not sensory components of pain. Science, 303(5661), 1157-1162. doi:
10.1126/science.1093535
174 of 179
Slessor, G., Phillips, L.H., & Bull, R. (2007). Exploring the specificity of age-related differences in theory of
mind tasks. Psychology and Aging, 22(3): 639-43.
Smirnova, A. (2012). Evidentiality in Bulgarian: Temporality, Epistemic Modality, and Information Source.
Journal of Semantics. doi:10.1093/jos/ffs017
Smith, A. T., Kosillo, P., & Williams, A. L. (2011). The confounding effect of response amplitude on MVPA
performance measures. NeuroImage, 56(2), 525-530. doi:10.1016/j.neuroimage.2010.05.079
Smith, V. L., & Ellsworth, P. C. (1987). The social psychology of eyewitness accuracy: Misleading
questions and communicator expertise. Journal of Applied Psychology, 72(2), 294-300. doi:
10.1037/0021-9010.72.2.294
Sommer, M., Döhnel, K., Sodian, B., Meinhardt, J., Thoermer, C., & Hajak, G. (2007). Neural correlates of
true and false belief reasoning. NeuroImage, 35(3), 1378-1384. doi:10.1016/j.neuroimage.
2007.01.042
Sommer, M., Rothmayr, C., Döhnel, K., Meinhardt, J., Schwerdtner, J., Sodian, B., & Hajak, G. (2010).
How should I decide? The neural correlates of everyday moral reasoning. Neuropsychologia, 48(7):
2018-26.
Southgate, V., Senju, A., & Csibra, G. (2007). Action anticipation through attribution of false belief by 2year-olds. Psychological Science 18(7): 587-92.
Spiers, H. J., & Maguire, E. A. (2007). Decoding human brain activity during real-world experiences.
Trends in Cognitive Sciences, 11(8), 356-365. doi:10.1016/j.tics.2007.06.002
Spiers, H.J., & Maguire, E.A. (2006). Thoughts, behaviour, and brain dynamics during navigation in the
real world. NeuroImage, 31(4): 1826-40.
Spratling, M.W. (2008). Reconciling Predictive Coding and Biased Competition Models of Cortical
Function. Front Comp Neurosci 2(4), 1-8.
Spratling, M.W. (2010). Predictive Coding as a Model of Response Properties in Cortical Area V1. J
Neurosci 30, 3531-3543.
Spunt, R. P., & Lieberman, M. D. (2012). An integrative model of the neural systems supporting the
comprehension of observed emotional behavior. NeuroImage, 59(3), 3050-3059. doi:10.1016/
j.neuroimage.2011.10.005
Spunt, R., Satpute, A.B., & Lieberman, M. (2011). Identifying the what, why, and how of an observed
action: an fMRI study of mentalizing and mechanizing during action observation. Journal of Cognitive
Neuroscience 23(1): 63-74.
Spunt, R.P., & Lieberman, M.D. (2012). An integrative model of the neural systems supporting the
comprehension of observed emotional behavior. NeuroImage 59(3), 3050.
Steinwart, I., & Christmann, A. (2008). Support Vector Machines. New York, NY: Springer Science &
Business Media. doi:10.1007/978-0-387-77242-4
Stelzer, J., Chen, Y., & Turner, R. (2013). Statistical inference and multiple testing correction in
classification-based multi-voxel pattern analysis (MVPA): Random permutations and cluster size
control. NeuroImage, 65, 69-82. doi:10.1016/j.neuroimage.2012.09.063
Stich, S., & Nichols, S. (1992). Folk Psychology: Simulation or Tacit Theory? Mind & Language, 7(1-2),
35-71.
Summerfield, C., Trittschuh, E.H., Monti, J.M., Mesulam, M.M., and Egner, T. (2008). Neural repetition
suppression reflects fulfilled perceptual expectations. Nat Neurosci 11, 1004-1006.
Tamir, D.I. & Mitchell, J.P. (2010). Neural correlates of anchoring-and-adjustment during mentalizing.
Proceedings of the National Academy of Sciences, 107(24): 10827.
175 of 179
Tenenbaum, J.B., Kemp, C., Griffiths, T.L., and Goodman, N.D. (2011). How to Grow a Mind: Statistics,
Structure, and Abstraction. Science 331, 1279-1285.
Tesink, C. M. J. Y., Buitelaar, J. K., Petersson, K. M., van der Gaag, R. J., Kan, C. C., Tendolkar, I., &
Hagoort, P. (2009). Neural correlates of pragmatic language comprehension in autism spectrum
disorders. Brain, 132(7), 1941-1952. doi:10.1093/brain/awp103
Tobler, P.N., Fiorillo, C.D., and Schultz, W. (2005). Adaptive coding of reward value by dopamine neurons.
Science 307, 1642-1645.
Todorovic, A., & de Lange, F. P. (2012). Repetition suppression and expectation suppression are
dissociable in time in early auditory evoked fields. J Neurosci, 32(39), 13389-13395.
Todorovic, A., van Ede, F., Maris, E., and de Lange, F.P. (2011). Prior Expectation Mediates Neural
Adaptation to Repeated Sounds in the Auditory Cortex: An MEG Study. J Neurosci 31, 9118-9123.
Toglia, M. P., Ross, D. F., Ceci, S. J., & Hembrooke, H. (1992). The Suggestibility of Children's Memory:
A Social-Psychological and Cognitive Interpretation. In Development of Long-Term Retention (pp.
217-241). New York, NY: Springer New York. doi:10.1007/978-1-4612-2868-4-7
Treue, S., and MartÕnez-Trujillo, J.C. (1999). Feature-based attention influences motion processing gain
in macaque visual cortex. Nature 399, 575-579.
Tsakiris, M., Costantini, M., & Haggard, P. (2008). The role of the right temporo-parietal junction in
maintaining a coherent sense of one's body. Neuropsychologia 46(12): 3014-18.
Tudusciuc, O., & Nieder, A. (2007). Neuronal population coding of continuous and discrete quantity in the
primate posterior parietal cortex. Proceedings of the National Academy of Sciences of the United
States of America, 104(36), 14513-14518. doi:10.1073/pnas.0705495104
Turk-Browne, N.B., Yi, D.-J., and Chun, M.M. (2006). Linking implicit and explicit memory: common
encoding factors and shared representations. Neuron 49, 917-927.
Tusche, A., Bode, S., & Haynes, J. D. (2010). Neural Responses to Unattended Products Predict Later
Consumer Choices. Journal of Neuroscience, 30(23), 8024-8031. doi:10.1523/JNEUROSCI.
0064-10.2010
Urwin, C. (1983). Dialogue and cognitive functioning in the early language development of three blind
children. In A. E. Mills (Ed.), Language acquisition in the blind child: Normal and deficient (pp.
142-161). San Diego, CA: College-Hill Press.
Uttal, W.R. (2011). Mind and Brain: A Critical Appraisal of Cognitive Neuroscience. Cambridge: MIT Press.
Van Overwalle, F. (2008). Social cognition and the brain: A meta-analysis. Human Brain Mapping 30(3):
829-58.
Van Overwalle, F., & Van Overwalle, F. (2009). Social cognition and the brain: a meta-analysis. Human
Brain Mapping, 30(3), 829-858. doi:10.1002/hbm.20547
Vander Wyk, B.C., Hudac, C.M., Carter, E.J., Sobel, D.M., and Pelphrey, K.A. (2009). Action
understanding in the superior temporal sulcus region. Psychol Sci 20, 771-777.
Vapnik, V. N., & Vapnik, V. (1998). Statistical learning theory, 2.
Visser, M., Jefferies, E., Embleton, K. V., & Ralph, M. A. L. (2012). Both the middle temporal gyrus and
the ventral anterior temporal area are crucial for multimodal semantic processing: distortion-corrected
fMRI evidence for a double gradient of information convergence in the temporal lobes. Journal of
Cognitive Neuroscience, 24(8), 1766-1778.
Vogeley, K., Bussfeld, P., Newen, A., Herrmann, S., Happe, F., Falkai, P., ... & Zilles, K. (2001). Mind
reading: neural mechanisms of theory of mind and self-perspective. Neuroimage, 14(1), 170-181.
176 of 179
Vogeley, K., May, M., Ritzl, A., Falkai, P., Zilles, K., & Fink, G.R. (2004). Neural correlates of first-person
perspective as one constituent of human self-consciousness. Journal of Cognitive Neuroscience 16,
817-27.
Vuilleumier, P., Armony, J.L., Driver, J., & Dolan, R.J. (2001). Effects of attention and emotion on face
processing in the human brain:: an event-related fMRI study. Neuron, 30(3): 829-41.
Wacongne, C., Changeux, J.P., and Dehaene, S. (2012). A Neuronal Model of Predictive Coding
Accounting for the Mismatch Negativity. J Neurosci 32, 3665-3678.
Wagner, D.D., Kelley, W.M., & Heatherton, T.F. (2011). Individual differences in the spontaneous
recruitment of brain regions supporting mental state understanding when viewing natural social
scenes. Cerebral Cortex, 21(12): 2788-96.
Wallis, J. D., Anderson, K. C., & Miller, E. K. (2001). Single neurons in prefrontal cortex encode abstract
rules. Nature, 411(6840), 953-956.
Walter, H., Schnell, K., Erk, S., Arnold, C., Kirsch, P., Esslinger, C., et al. (2010). Effects of a genome-wide
supported psychosis risk variant on neural activation during a theory-of-mind task. Molecular
Psychiatry, 16(4), 462-470. doi:doi:10.1038/mp.2010.18
Walther, D. B., Beck, D. M., & Fei-Fei, L. (2012). To err is human: Correlating fMRI decoding and
behavioral errors to probe the neural representation of natural scene categories. In N. Kriegeskorte &
G. Kreiman, (pp. 391-415). Visual population codes-Toward a common multivariate framework for cell
recording and functional imaging.
Wang, A. T. (2006). Neural basis of irony comprehension in children with autism: the role of prosody and
context. Brain, 129(4), 932-943. doi:10.1093/brain/awl032
Weiner, K.S., Sayres, R., Vinberg, J., and Grill-Spector, K. (2010). fMRI-Adaptation and Category
Selectivity in Human Ventral Temporal Cortex: Regional Differences Across Time Scales. J
Neurophysiol 103, 3349-3365.
Welch-Ross, M. K. (1999). Interviewer Knowledge and Preschoolers' Reasoning About Knowledge States
Moderate Suggestibility. Cognitive Development.
Welch-Ross, M. K., Diecidue, K., & Miller, S. A. (1997). Young children's understanding of conflicting
mental representation predicts suggestibility. Developmental Psychology, 33(1), 43-53.
Wellman, H. M., Cross, D., & Watson, J. (2001). Meta-analysis of theory-of-mind development: The truth
about false belief. Child Development, 655-684.
Wellman, H., Cross, D., & Watson, J. (2001). Meta-analysis of theory-of-mind development: The truth
about false belief. Child Development, 72(3). 655-84.
Whitcombe, E. L., & Robinson, E. J. (2000). Children's decisions about what to believe and their ability to
report the source of their belief. Cognitive Development, 15(3), 329-346. doi:10.1016/
S0885-2014(00)00033-2
Whitfield-Gabrieli, S., Moran, J.M., Nieto-Castañón, A., Triantafyllou, C., Saxe, R.R., & Gabrieli, J.D.
(2011). Associations and dissociations between default and self-reference networks in the human
brain. NeuroImage, 55(1): 225-32.
Wimmer, H., & Perner, J. (1983). Beliefs about beliefs: Representation and constraining function of wrong
beliefs in young children's understanding of deception. Cognition, 13(1), 103-128.
Winston, J.S., Henson, R.N., Fine-Goulden, M.R., and Dolan, R.J. (2004). fMRI-adaptation reveals
dissociable neural representations of identity and expression in face perception. J Neurophysiol 92,
1830-1839.
Woolfe, T., & Want, S. (2002). Signposts to development: Theory of mind in deaf children. Child
Development, 73(3), 768-778.
177 of 179
Woolfolk, R. L., Doris, J. M., & Darley, J. M. (2006). Identification, situational constraint, and social
cognition: studies in the attribution of moral responsibility. Cognition, 100(2), 283-301. doi:10.1016/
j.cognition.2005.05.002
Worsley, K.J., Evans, A.C., Marrett, S., & Neelin, P. (1992). A three- dimensional statistical analysis for
CBF activation studies in human brain. Journal of Cerebral Blood Flow & Metabolism 12(6): 900-18.
Yamada, M., Camerer, C. F., Fujie, S., Kato, M., Matsuda, T., Takano, H., et al. (2012). Neural circuits in
the brain that are activated when mitigating criminal sentences. Nature Communications, 3, 1-6. doi:
10.1038/ncomms1757
Yamada, Makiko, Colin F. Camerer, Saori Fujie, Motoichiro Kato, Tetsuya Matsuda, Harumasa Takano,
Hiroshi Ito, Tetsuya Suhara, and Hidehiko Takahashi. (2013) Neural circuits in the brain that are
activated when mitigating criminal sentences. Nature 3, 759.
Yang, D., Baillargeon, R., & Baillargeon, R. (2013). Difficulty in Understanding Social Acting (but not False
Beliefs) Mediates the Link Between Autistic Traits and Ingroup Relationships. Journal of Autism and
Developmental Disorders, 1-24.
Yoshida, W., Seymour, B., Friston, K.J., & Dolan, R.J. (2010). Neural mechanisms of belief inference
during cooperative games. Journal of Neuroscience 30(32): 10744-51.
Young, L.L., & Phillips, J. (2011). The paradox of moral focus. Cognition, 119(2), 166-178. doi:10.1016/
j.cognition.2011.01.004
Young, L.L., & Saxe, R.R. (2008). The neural basis of belief encoding and integration in moral judgment.
NeuroImage, 40(4), 1912-1920. doi:10.1016/j.neuroimage.2008.01.057
Young, L.L., & Saxe, R.R. (2009a). An FMRI investigation of spontaneous mental state inference for
moral judgment. Journal of Cognitive Neuroscience, 21(7), 1396-1405. doi:10.1086/591811
Young, L.L., & Saxe, R.R. (2009b). Innocent intentions: A correlation between forgiveness for accidental
harm and neural activity. Neuropsychologia, 47(10), 2065-2072. doi:10.1016/j.neuropsychologia.
2009.03.020
Young, L.L., & Waytz, A. (2013). Mind attribution is for morality. In S. L. Baron-Cohen, H. Tager-Flusberg,
& M . V. L o m b a r d o , U n d e r s t a n d i n g O t h e r M i n d s . O x f o r d U n i v e r s i t y P r e s s . d o i :
10.1177/0956797612472343
Young, L.L., Camprodon, J. A., Hauser, M., Pascual-Leone, A., & Saxe, R.R. (2010a). Disruption of the
right temporoparietal junction with transcranial magnetic stimulation reduces the role of beliefs in
moral judgments. Proceedings of the National Academy of Sciences of the United States of America,
107(15), 6753-6758. doi:10.1073/pnas.0914826107
Young, L.L., Cushman, F., Hauser, M., & Saxe, R.R. (2007). The neural basis of the interaction between
theory of mind and moral judgment. Proceedings of the National Academy of Sciences 104(20):
8235-40.
Young, L.L., Dodell-Feder, D., & Saxe, R.R. (2010b). What gets the attention of the temporo-parietal
junction? An fMRI investigation of attention and theory of mind. Neuropsychologia, 48(9), 2658-2664.
doi:10.1016/j.neuropsychologia.2010.05.012
Young, L.L., Nichols, S., & Saxe, R.R. (2010c). Investigating the Neural and Cognitive Basis of Moral
Luck: It's Not What You Do but What You Know. Review of Philosophy and Psychology, 1(3),
333-349. doi:10.1007/s13164-010-0027-y
Zahn, R., Moll, J., Krueger, F., Huey, E. D., Garrido, G., & Grafman, J. (2007). Social concepts are
represented in the superior anterior temporal cortex.Proceedings of the National Academy of
Sciences, 104(15), 6430-6435.
Zaitchik, D. (1990). When representations conflict with reality: The preschooler's problem with false
beliefs and. Cognition 35: 41-68.
178 of 179
Zaitchik, D., Koff, E., Brownell, H., Winner, E., & Albert, M. (2006). Inference of beliefs and emotions in
patients with Alzheimer. Neuropsychology 20, 11-20.
Zaitchik, D., Walker, C., Miller, S., LaViolette, P., Feczko, E., & Dickerson, B. C. (2010). Mental state
attribution and the temporoparietal junction: An fMRI study comparing belief, emotion, and perception.
Neuropsychologia, 48(9), 2528-2536. doi:10.1016/j.neuropsychologia.2010.04.031
Zaki, J., and Mitchell, J. P. (2011). Equitable decision making is associated with neural markers of intrinsic
value. Proc Natl Acad Sci U S A, 108(49), 19761-19766.
Zaragoza, M. S., & Lane, S. M. (1994). Source misattributions and the suggestibility of eyewitness
memory. Journal of Experimental Psychology-Learning Memory and Cognition, 20(4), 934-945.
Zhu, L., Mathewson, K.E., and Hsu, M. (2012). Dissociable neural representations of reinforcement and
belief prediction errors underlie strategic learning. Proc Natl Acad Sci U S A 109, 1419-1424.
179 of 179
Download