Uploaded by Grace Gardner

Cognition Textbook

advertisement
Cognition
seventh edition
Cognition
exploring the science of the mind
Daniel Reisberg
reed college
n
W. W. Norton & Company
New York • London
7e
W. W. Norton & Company has been independent since its founding in 1923, when William Warder
Norton and Mary D. Herter Norton first published lectures delivered at the People’s Institute,
the adult education division of New York City’s Cooper Union. The firm soon expanded its program beyond the Institute, publishing books by celebrated academics from America and abroad.
By midcentury, the two major pillars of Norton’s publishing program—trade books and college
texts—were firmly established. In the 1950s, the Norton family transferred control of the company to its employees, and today—with a staff of four hundred and a comparable number of
trade, college, and professional titles published each year—W. W. Norton & Company stands as
the largest and oldest publishing house owned wholly by its employees.
Copyright © 2019, 2016, 2013, 2010, 2006, 2001, 1997 by W. W. Norton & Company, Inc.
All rights reserved
Printed in the United States of America
Editor: Ken Barton
Project Editor: David Bradley
Assistant Editor: Eve Sanoussi
Editorial Assistant: Katie Pak
Manuscript Editor: Alice Vigliani
Managing Editor, College: Marian Johnson
Managing Editor, College Digital Media: Kim Yi
Production Managers: Ashley Horna and Jane Searle
Media Editors: Scott Sugarman and Kaitlin Coats
Associate Media Editor: Victoria Reuter
Assistant Media Editors: Alex Trivilino and Allison Nicole Smith
Marketing Manager: Ashley Sherwood
Design Director: Rubina Yeh
Art Director: Jillian Burr
Designer: Lisa Buckley
Photo Editor: Ted Szczepanski
Photo Researcher: Lynn Gadson
Permissions Manager: Megan Schindel
Composition/Illustrations: GraphicWorld
Manufacturing: LCS Communications—Kendallville
Permission to use copyrighted material is included in the credits section of this book,
which begins on page C1.
Library of Congress Cataloging-in-Publication Data
Names: Reisberg, Daniel.
Title: Cognition : exploring the science of the mind / Daniel Reisberg, Reed
College.
Description: Seventh Edition. | New York : W. W. Norton & Company, [2018] |
Revised edition of the author’s Cognition, [2016]o | Includes
bibliographical references and index.
Identifiers: LCCN 2018022174 | ISBN 9780393665017 (hardcover)
Subjects: LCSH: Cognitive psychology.
Classification: LCC BF201 .R45 2018 | DDC 153—dc23 LC record available at https://lccn.loc.
gov/2018022174
W. W. Norton & Company, Inc., 500 Fifth Avenue, New York, NY 10110
wwnorton.com
W. W. Norton & Company Ltd., 15 Carlisle Street, London W1D 3BS
1234567890
With love
— always —
for the family that
enriches every
aspect of my life.
Brief Contents
CONTENTS ix
PREFACE xiii
PART 1 THE FOUNDATIONS OF COGNITIVE PSYCHOLOGY
1 The Science of the Mind 2
2 The Neural Basis for Cognition
1
24
PART 2 LEARNING ABOUT THE WORLD AROUND US
61
3 Visual Perception 62
4 Recognizing Objects 106
5 Paying Attention 148
PART 3 MEMORY
193
6 The Acquisition of Memories and the Working-Memory System
7 Interconnections between Acquisition and Retrieval 238
8 Remembering Complex Events 278
PART 4 KNOWLEDGE
323
9 Concepts and Generic Knowledge
10 Language 364
11 Visual Knowledge 410
PART 5 THINKING
194
324
453
12 Judgment and Reasoning 454
13 Problem Solving and Intelligence 498
14 Conscious Thought, Unconscious Thought
Appendix: Research Methods
Glossary G-1
References R-1
Credits C-1
Author Index I-1
Subject Index I-13
546
A-1
vii
Contents
PREFACE
xiii
PART 1THE FOUNDATIONS OF COGNITIVE
PSYCHOLOGY 1
1 The Science of the Mind
2
The Scope of Cognitive Psychology 3 • The Cognitive Revolution 8
• Research in Cognitive Psychology: The Diversity of Methods 17
• Applying Cognitive Psychology 19 • Chapter Review 21
2 The Neural Basis for Cognition
24
Explaining Capgras Syndrome 26 • The Study of the Brain 31
• Sources of Evidence about the Brain 37 • The Cerebral Cortex 44
• Brain Cells 49 • Moving On 55 • Cognitive Psychology and
Education: Food Supplements and Cognition 55 • Chapter Review 58
PART 2
LEARNING ABOUT THE WORLD AROUND US 61
3 Visual Perception
62
The Visual System 64 • Visual Coding 70 • Form Perception 80
• Constancy 87 • The Perception of Depth 92 • Cognitive Psychology
and Education: An “Educated Eye” 99 • Chapter Review 103
4 Recognizing Objects
106
Recognition: Some Early Considerations 110 • Word Recognition 112
• Feature Nets and Word Recognition 116 • Descendants of the
Feature Net 127 • Face Recognition 133 • Top-Down Influences on
Object Recognition 140 • Cognitive Psychology and Education:
Speed-Reading 142 • Chapter Review 145
ix
5 Paying Attention
148
Selective Attention 150 • Selection via Priming 158 • Spatial
Attention 164 • Divided Attention 177 • Practice 183 • Cognitive
Psychology and Education: ADHD 188 • Chapter Review 190
PART 3 MEMORY 193
6The Acquisition of Memories and
the Working-Memory System 194
Acquisition, Storage, and Retrieval 197 • The Route into Memory 198
• A Closer Look at Working Memory 205 • Entering Long-Term
Storage: The Need for Engagement 214 • The Role of Meaning
and Memory Connections 221 • Organizing and Memorizing 224
• The Study of Memory Acquisition 230 • Cognitive Psychology
and Education: How Should I Study? 232 • Chapter Review 235
7Interconnections between Acquisition
and Retrieval 238
Learning as Preparation for Retrieval 241 • Encoding Specificity 244
• The Memory Network 246 • Different Forms of Memory Testing 250
• Implicit Memory 254 • Theoretical Treatments of Implicit Memory 261
• Amnesia 267 • Cognitive Psychology and Education: Familiarity Can
Be Treacherous 273 • Chapter Review 275
8 Remembering Complex Events
278
Memory Errors, Memory Gaps 280 • Memory Errors: A Hypothesis 282
• The Cost of Memory Errors 288 • Avoiding Memory Errors 296
• Forgetting 297 • Memory: An Overall Assessment 302 • Autobiographical
Memory 304 • How General Are the Principles of Memory? 315
• Cognitive Psychology and Education: Remembering for the Long
Term 317 • Chapter Review 320
PART 4 KNOWLEDGE 323
9 Concepts and Generic Knowledge
324
Understanding Concepts 326 • Prototypes and Typicality Effects 329
• Exemplars 334 • The Difficulties with Categorizing via Resemblance
337 • Concepts as Theories 343 • The Knowledge Network 350
• Concepts: Putting the Pieces Together 358 • Cognitive Psychology
and Education: Learning New Concepts 358 • Chapter Review 361
10 Language
364
The Organization of Language 366 • Phonology 368 • Morphemes
and Words 377 • Syntax 378 • Sentence Parsing 382 • Prosody 390
• Pragmatics 391 • The Biological Roots of Language 392 • Language
and Thought 399 • Cognitive Psychology and Education: Writing 404
• Chapter Review 407
x
•
Contents
11 Visual Knowledge
410
Visual Imagery 412 • Chronometric Studies of Imagery 415 • Imagery
and Perception 422 • Visual Imagery and the Brain 424 • Individual
Differences in Imagery 430 • Images Are Not Pictures 435 • Long-Term
Visual Memory 439 • The Diversity of Knowledge 447 • Cognitive
Psychology and Education: Using Imagery 448 • Chapter Review 450
PART 5 THINKING 453
12 Judgment and Reasoning
454
Judgment 456 • Detecting Covariation 463 • Dual-Process Models 466
• Confirmation and Disconfirmation 471 • Logic 476 • Decision
Making 480 • Cognitive Psychology and Education: Making People
Smarter 491 • Chapter Review 494
13 Problem Solving and Intelligence
498
General Problem-Solving Methods 500 • Drawing on Experience 504
• Defining the Problem 509 • Creativity 514 • Intelligence 522
• Intelligence beyond the IQ Test 530 • The Roots of Intelligence 533
• Cognitive Psychology and Education: The Goals of “Education” 539
• Chapter Review 542
14 Conscious Thought, Unconscious Thought
546
The Study of Consciousness 548 • The Cognitive Unconscious 549
• Disruptions of Consciousness 557 • Consciousness and Executive
Control 560 • The Cognitive Neuroscience of Consciousness 566
• The Role of Phenomenal Experience 572 • Consciousness: What Is
Left Unsaid 579 • Cognitive Psychology and Education: Mindfulness
580 • Chapter Review 583
Appendix: Research Methods
A-1
Glossary
G-1
References
R-1
Credits
C-1
Author Index
I-1
Subject Index
I-13
Contents
•
xi
Preface
I
was a college sophomore when I took my first course in cognitive psychology. I was excited about the material then, and, many years later, the
excitement hasn’t faded. Part of the reason lies in the fact that cognitive
psychologists are pursuing fabulous questions, questions that have intrigued
humanity for thousands of years: Why do we think the things we think? Why
do we believe the things we believe? What is “knowledge,” and how secure
(how complete, how accurate) is our knowledge of the world around us?
Other questions asked by cognitive psychologists concern more immediate,
personal, issues: How can I help myself to remember more of the material that
I’m studying in my classes? Is there some better way to solve the problems I
encounter? Why is it that my roommate can study with music on, but I can’t?
And sometimes the questions have important consequences for our social or
political institutions: If an eyewitness reports what he saw at a crime, should we
trust him? If a newspaper raises questions about a candidate’s integrity, how will
voters react?
Of course, we want more than interesting questions—we also want answers
to these questions, and this is another reason I find cognitive psychology so exciting. In the last half-century or so, the field has made extraordinary progress on
many fronts, providing us with a rich understanding of the nature of memory,
the processes of thought, and the content of knowledge. There are many things
still to be discovered—that’s part of the fun. Even so, we already have a lot to say
about all of the questions just posed and many more as well. We can speak to the
specific questions and to the general, to the theoretical issues and to the practical.
Our research has uncovered principles useful for improving the process of education, and we have made discoveries of considerable importance for the criminal
justice system. What I’ve learned as a cognitive psychologist has changed how I
think about my own memory; it’s changed how I make decisions; it’s changed
how I draw conclusions when I’m thinking about events in my life.
On top of all this, I’m also excited about the connections that cognitive
psychology makes possible. In the academic world, intellectual disciplines are
often isolated from one another, sometimes working on closely related problems
xiii
without even realizing it. In the last decades, though, cognitive psychology has
forged rich connections with its neighboring disciplines, and in this book we’ll
touch on topics in philosophy, neuroscience, law and criminal justice, economics, linguistics, politics, computer science, and medicine. These connections bring
obvious benefits, since insights and information can be traded back and forth
between the domains. But these connections also highlight the importance of the
material we’ll be examining, since the connections make it clear that the issues
before us are of interest to a wide range of scholars. This provides a strong signal
that we’re working on questions of considerable power and scope.
I’ve tried in this text to convey all this excitement. I’ve done my best to describe
the questions being asked within my field, the substantial answers we can provide
for these questions, and, finally, some indications of how cognitive psychology is
(and has to be) interwoven with other intellectual endeavors.
I’ve also had other goals in writing this text. In my own teaching, I try to
maintain a balance among many different elements: the nuts and bolts of how
our science proceeds, the data provided by the science, the practical implications
of our research findings, and the theoretical framework that holds all of these
pieces together. I’ve tried to find the same balance in this text. Perhaps most
important, though, I try, both in my teaching and throughout this book, to “tell a
good story,” one that conveys how the various pieces of our field fit together into
a coherent package. Of course, I want the evidence for our claims to be in view,
so that readers can see how our field tests its hypotheses and why our claims must
be taken seriously. But I’ve also put a strong emphasis on the flow of ideas—how
new theories lead to new experiments, and how those experiments can lead to
new theory.
The notion of “telling a good story” also emerges in another way: I’ve always
been impressed by the ways in which the different parts of cognitive psychology
are interlocked. Our claims about attention, for example, have immediate implications for how we can theorize about memory; our theories of object recognition
are linked to our proposals for how knowledge is stored in the mind. Linkages
like these are intellectually satisfying, because they ensure that the pieces of the
puzzle really do fit together. But, in addition, these linkages make the material
within cognitive psychology easier to learn, and easier to remember. Indeed, if I
were to emphasize one crucial fact about memory, it would be that memory is
best when the memorizer perceives the organization and interconnections within
the material being learned. (We’ll discuss this point further in Chapter 6.) With
an eye on this point, I’ve therefore made sure to highlight the interconnections
among various topics, so that readers can appreciate the beauty of our field and
can also be helped in their learning by the orderly nature of our theorizing.
I’ve tried to help readers in other ways, too. First, I’ve tried throughout the
book to make the prose approachable. I want my audience to gain a sophisticated understanding of the material in this text, but I don’t want readers to
struggle with the ideas.
Second, I’ve taken various steps that I hope will foster an “alliance” with
readers. My strategy here grows out of the fact that, like most teachers, I value
the questions I receive from students and the discussions I have with them. In
xiv •
Preface
the classroom, this allows a two-way flow of information that unmistakably
improves the educational process. Of course, a two-way flow isn’t possible in
a textbook, but I’ve offered what I hope is a good approximation: Often, the
questions I hear from students, and the discussions I have with them, focus
on the relevance of the course material to students’ own lives, or relevance to
the world outside of academics. I’ve tried to capture that dynamic, and to present
my answers to these student questions, in the essay at the end of each chapter
(I’ll say more about these essays in a moment). These essays appear under the
banner Cognitive Psychology and Education, and—as the label suggests—
the essays will help readers understand how the materials covered in that chapter
matter for (and might change!) the readers’ own learning. In addition, I’ve written a separate series of essays (available online), titled Cognitive Psychology and
the Law, to explore how each chapter’s materials matter in another arena—the
enormously important domain of the justice system. I hope that both types of
essays—Education and Law—help readers see that all of this material is indeed
relevant to their lives, and perhaps as exciting for them as it is for me.
Have I met all of these goals? You, the readers, will need to be the judges of
this. I would love to hear from you about what I’ve done well in the book and
what I could have done better; what I’ve covered (but should have omitted) and
what I’ve left out. I’ll do my best to respond to every comment. You can reach me
via email (reisberg@reed.edu); I’ve been delighted to get comments from readers
about previous editions, and I hope for more emails with this edition.
An Outline of the Seventh Edition
The book’s 14 chapters are designed to cover the major topics within cognitive psychology. The chapters in Part 1 lay the foundation. Chapter 1 provides the conceptual and historical background for the subsequent chapters.
In addition, this chapter seeks to convey the extraordinary scope of the field
and why, therefore, research on cognition is so important. The chapter also
highlights the relationship between theory and evidence in cognitive psychology, and it discusses the logic on which this field is built.
Chapter 2 then offers a brief introduction to the study of the brain. Most of
cognitive psychology is concerned with the functions that our brains make possible, and not the brain itself. Nonetheless, our understanding of cognition has
certainly been enhanced by the study of the brain, and throughout this book
we’ll use biological evidence as one means of evaluating our theories. Chapter 2
is designed to make this evidence fully accessible to the reader—by providing
a quick survey of the research tools used in studying the brain, an overview of
the brain’s anatomy, and also an example of how we can use brain evidence as a
source of insight into cognitive phenomena.
Part 2 of the book considers the broad issue of how we gain information from
the world. Chapter 3 covers visual perception. At the outset, this chapter links to
the previous (neuroscience) chapter with descriptions of the eyeball and the basic
mechanisms of early visual processing. In this context, the chapter introduces
the crucial concept of parallel processing and the prospect of mutual influence
Preface
•
xv
among separate neural mechanisms. From this base, the chapter builds toward a
discussion of the perceiver’s activity in shaping and organizing the visual world,
and explores this point by discussing the rich topics of perceptual constancy and
perceptual illusions.
Chapter 4 discusses how we recognize the objects that surround us. This
seems a straightforward matter—what could be easier than recognizing a telephone, or a coffee cup, or the letter Q? As we’ll see, however, recognition is
surprisingly complex, and discussion of this complexity allows me to amplify key
themes introduced in earlier chapters: how active people are in organizing and
interpreting the information they receive from the world; the degree to which
people supplement the information by relying on prior experience; and the ways
in which this knowledge can be built into a network.
Chapter 5 then considers what it means to “pay attention.” The first half of
the chapter is concerned largely with selective attention—cases in which you seek
to focus on a target while ignoring distractors. The second half of the chapter is
concerned with divided attention (“multi-tasking”)—that is, cases in which you
seek to focus on more than one target, or more than one task, at the same time.
Here, too, we’ll see that seemingly simple processes turn out to be more complicated than one might suppose.
Part 3 turns to the broad topic of memory. Chapters 6, 7, and 8 start with a
discussion of how information is “entered’’ into long-term storage, but then turn
to the complex interdependence between how information is first learned and
how that same information is subsequently retrieved. A recurrent theme in this
section is that learning that’s effective for one sort of task, one sort of use, may
be quite ineffective for other uses. This theme is examined in several contexts,
and leads to a discussion of research on unconscious memories—so-called memory without awareness. These chapters also offer a broad assessment of human
memory: How accurate are our memories? How complete? How long-lasting?
These issues are pursued both with regard to theoretical treatments of memory
and also with regard to the practical consequences of memory research, including
the application of this research to the assessment, in the courtroom, of eyewitness
testimony.
The book’s Part 4 is about knowledge. Earlier chapters show over and over
that humans are, in many ways, guided in their thinking and experiences by what
they already know—that is, the broad pattern of knowledge they bring to each
new experience. This invites the questions posed by Chapters 9, 10, and 11: What
is knowledge? How is it represented in the mind? Chapter 9 tackles the question
of how “concepts,” the building blocks of our knowledge, are represented in the
mind. Chapters 10 and 11 focus on two special types of knowledge. Chapter 10
examines our knowledge about language; Chapter 11 considers visual knowledge and examines what is known about mental imagery.
The chapters in Part 5 are concerned with the topic of thinking. Chapter 12
examines how each of us draws conclusions from evidence—including cases
in which we are trying to be careful and deliberate in our judgments, and also
cases of informal judgments of the sort we often make in our everyday lives. The
chapter then turns to the question of how we reason from our beliefs—how we
xvi •
Preface
check on whether our beliefs are correct, and how we draw conclusions, based on
things we already believe. The chapter also considers the practical issue of how
errors in thinking can be diminished through education.
Chapter 13 is also about thinking, but with a different perspective: This
chapter considers some of the ways people differ from one another in their ability to solve problems, in their creativity, and in their intelligence. The chapter
also addresses the often heated, often misunderstood debate about how different
groups—especially American Whites and African Americans—might (or might
not) differ in their intellectual capacities.
The final chapter in the book does double service. First, it pulls together
many of the strands of contemporary research relevant to the topic of
consciousness—what consciousness is, and what consciousness is for. In addition, most readers will reach this chapter at the end of a full semester’s work,
a point at which they are well served by a review of the topics already covered
and ill served by the introduction of much new material. Therefore, this chapter
draws many of its themes and evidence from previous chapters, and in that
fashion it serves as a review of points that appear earlier in the book. Chapter
14 also highlights the fact that we’re using these materials to approach some of
the greatest questions ever asked about the mind, and, in that way, this chapter
should help to convey some of the power of the material we’ve been discussing
throughout the book.
New in the Seventh Edition
What’s new in this edition? Every chapter contains new material, in most
cases because readers specifically requested the new content! Chapter 1, for
example, now includes discussion of how the field of cognitive psychology
emerged in the 1950s and 1960s. Chapter 4 includes coverage of recent work
on how people differ from one another in their level of face-recognition skill.
Chapter 5 discusses what it is that people pay attention to, with a quick
summary of research on how men and women differ in what they focus on,
and how different cultures seem to differ in what they focus on. Chapter 8
discusses a somewhat controversial and certainly dramatic study showing
that college students can be led to a false memory of a time they committed a
felony (an armed assault) while in high school; this chapter also now includes
coverage of the social nature of remembering. Chapter 10 now discusses the
topics of prosody and pragmatics. Chapter 12 discusses the important difference between “opt-in” and “opt-out” procedures for social policy, and
Chapter 14 now includes discussion of (both the myths and the reality of)
subliminal perception.
In this edition, I’ve also added three entirely new features. First, my students
are always curious to learn how cognitive psychology research can be applied to
issues and concerns that arise in everyday life. I’ve therefore added a Cognition
Outside the Lab essay to every chapter. For example, in Chapter 4, in discussing how word recognition proceeds, I’ve tackled the question of how the choice
of font can influence readers (sometimes in good ways and sometimes not). In
Preface
•
xvii
Chapter 7, I’ve written about cryptoplagiarism, a pattern in which you can steal
another person’s ideas without realizing it!
Second, I have always believed that, as someone teaching cognitive psychology, I need to respect the practical lessons of my field. As one example, research
suggests that students’ understanding and memory are improved if they pause
and reflect on materials they’ve just heard in a lecture or just read in a book.
“What did I just hear? What were the main points? Which bits were new, and
which bits had I thought about before?” Guided by that research, I’ve added Test
Yourself questions throughout the book. These questions are then echoed at the
end of the chapter, with the aim of encouraging readers to do another round of
reflection. All these questions are designed to be easy and straightforward—but
should, our research tells us, be genuinely helpful for readers.
Third, the topics covered in this book have many implications, and I hope
readers will find it both fun and useful to think about some of these implications.
On this basis, every chapter also ends with a couple of Think About It questions,
inviting readers to extend the chapter’s lessons into new territory. For example,
at the end of Chapter 3, I invite readers to think about how research on attention
might help us understand what happens in the focused exercise of meditation
(including Buddhist meditation). The question at the end of Chapter 7 invites
readers to think through how we might explain the eerie sensation of déjà vu.
A question at the end of Chapter 8 explores how your memory is worse than a
video recorder, and also how it’s better than a video recorder.
Other Special Features
In addition, I have (of course) held on to features that were newly added in
the previous edition—including an art program that showcases the many
points of contact between cognitive psychology and cognitive neuroscience,
and the “What if . . .” section that launches each chapter. The “What if . . .”
material serves several aims. First, the mental capacities described in each
chapter (the ability to recognize objects, the ability to pay attention, and so
on) are crucial for our day-to-day functioning, and to help readers understand this point, most of the “What if . . .” sections explore what someone’s
life is like if they lack the relevant capacity. Second, the “What if . . .” sections
are rooted in concrete, human stories; they talk about specific individuals
who lack these capacities. I hope these stories will be inviting and thoughtprovoking for readers, motivating them to engage the material in a richer
way. And, third, most of the “What if . . .” sections involve people who have
lost the relevant capacity through some sort of brain damage. These sections
therefore provide another avenue through which to highlight the linkage
between cognitive psychology and cognitive neuroscience.
This edition also includes explicit coverage of Research Methods. As in the
previous edition, this material is covered in an appendix, so that it’s easily accessible to all readers, but set to the side to accommodate readers (or instructors)
who prefer to focus on the book’s substantive content. The appendix is divided
into separate essays for each chapter, so that the appendix can be used on a
xviii •
Preface
chapter-by-chapter basis. This organization will help readers see, for each
chapter, how the research described in the chapter unfolds, and it will simultaneously provide a context for each methods essay so that readers can see why the
methods are so important.
The appendix is surely no substitute for a research methods course, but nonetheless it’s sequenced in a manner that builds toward a broad understanding of
how the scientific method plays out in our field. An early essay, for example,
works through the question of what a “testable hypothesis” is, and why this is
so important; another essay works through the power of random assignment;
another discusses how we deal with confounds. In all cases, my hope is that the
appendix will guide readers toward a sophisticated understanding of why our
research is as it is, and why, therefore, our research is so persuasive.
I have already mentioned the end-of-chapter essays on Cognitive Psychology
and Education, which show students how cognitive psychology is connected to
their own learning. Readers often seek “take-home messages” from the material
that will, in a direct way, benefit them. We are, after all, talking about memory, and
students obviously are engaged in an endeavor of putting lots of new information—
information they’re learning in their courses—into their memories. We’re talking
about attention, and students often struggle with the chore of keeping themselves
“on task” and “on target.” In light of these points, the end-of-chapter essays build
a bridge between the content in the chapter and the concerns that fill students’
lives. This will, I hope, make the material more useful for students, and also make
it clear just how important an enterprise cognitive psychology is.
There are also essays in the ebook on Cognitive Psychology and the Law.
Here, I’ve drawn on my own experience in working with law enforcement and
the criminal justice system. In this work, I’m called on to help juries understand
how an eyewitness might be certain in his recollection, but mistaken. I also work
with police officers to help them elicit as much information from a witness as
possible, without leading the witness in any way. Based on this experience, the
online essays discuss how the material in each chapter might be useful for the
legal system. These essays will, I hope, be immediately interesting for readers, and
will also make it obvious why it’s crucial that the science be done carefully and
well—so that we bring only high-quality information into the legal system.
I’ve also included Demonstrations in the ebook to accompany the book’s
description of key concepts in the field. Many of these demos are miniature versions of experimental procedures, allowing students to see for themselves what
these experiments involve, and also allowing them to see just how powerful
many of our effects are. Readers who want to run the demos for themselves as
they read along certainly can; instructors who want to run the demos within
their classrooms (as I sometimes do) are certainly encouraged to do so. Instructors who want to use the demos in discussion sections, aside from the main
course, can do that as well. In truth, I suspect that some demos will work better
in one of these venues, and that other demos will work better in others; but,
in all cases, I hope the Demonstrations help bring the material to life—putting
students directly in contact with both our experimental methods and our experimental results.
Preface
•
xix
As in previous editions, this version of Cognition also comes with various
supplementary materials, some aimed at students, and some aimed at instructors.
For Students
ZAPS Cognition Labs. Every copy of the text comes packaged with free
access to ZAPS Cognition Labs, an updated revision of Norton’s popular
online psychology labs. Crafted specifically to support cognitive psychology
courses, this version helps students learn about core psychological phenomena. Each lab (one or two per chapter) begins with a brief introduction that
relates the topic to students’ lives. Students then engage in a hands-on experience that, for most labs, produces data based on their individual responses.
The theories behind the concepts are then explained alongside the data the
student has generated. Also, an assessment component lets students confirm
that they understand the concepts central to each lab. Finally, this edition
of Cognition is accompanied by five new ZAPS labs: Encoding Specificity,
Mental Rotation 3D, Memory Span, Operation Span, and Selective Attention.
Ebook. Every print copy of the text comes packaged with free access to the
ebook. The ebook can also be purchased separately at a fraction of the price of
the printed version. The ebook has several advantages over the print text. First,
the ebook includes Demonstrations—quick, pen-and-paper mini experiments—
designed to show students what key experiments involve and how powerful
many of the effects are. Second, the ebook includes the Cognitive Psychology and
the Law essays, described above. In addition, the ebook can be viewed on any
device—laptop, tablet, phone, public computer—and will stay synced between
devices. The ebook is therefore a perfect solution for students who want to learn
in more convenient settings—and pay less for doing so.
For Instructors
All instructor resources for this edition of Cognition can be accessed via the
“Instructor Resources” tile at the following URL: https://digital.wwnorton
.com/cognition7.
Interactive Instructor’s Guide (IIG). This online repository of teaching assets
offers material for every chapter that both veteran and novice instructors of the
course will find helpful. Searchable by chapter or asset class, the IIG provides
multiple resources for teaching: links to online video clips (selected and annotated by the author), teaching suggestions, and other class activities and exercises.
It also includes all of the Education, Law, and Research Methods essays described
above, as well as discussion questions to support the Education and Law essays.
The demonstrations from the ebook can also be found here. This repository of
lecture and teaching materials functions both as a course prep tool and as a
means of tracking the latest ideas in teaching the cognitive psychology course.
I’m especially excited about the online video clips. Students love videos and
probably spend more time than they should surfing the Internet (and YouTube
in particular) for fun clips. As it turns out, though, YouTube contains far more
xx
•
Preface
than cute-kittens movies; it also contains intriguing, powerful material directly
relevant to the topics in this text. The IIG therefore provides a listing of carefully selected online videos to accompany each of the chapters. (A dozen of these
videos are newly added for the seventh edition!) The annotated list describes each
clip, and gives information about timing, in ways that should make these videos
easy to use in the classroom. I use them in my own teaching, and my students
love them. But let me also make a request: I’m sure there are other videos available that I haven’t seen yet. I’ll therefore be grateful to any readers who help me
broaden this set, so that we can make this resource even better.
Test Bank. The test bank features over 900 questions, including multiplechoice and short-answer questions for each chapter. I have personally vetted each
question, and all questions have been updated according to Norton’s assessment
guidelines to make it easy for instructors to construct quizzes and exams that are
meaningful and diagnostic. All questions are classified according to learning objective, text section, difficulty, and question type. This Norton test bank is available
with ExamView Test Generator software, allowing instructors to create, administer, and manage assessments. The intuitive test-making wizard makes it easy to
create customized exams. Other features include the ability to create paper exams
with algorithmically generated variables and to export files directly to your LMS.
Lecture PowerPoints. These text-focused PowerPoints follow the chapter
outlines, include figures from the text, and feature instructor-only notes.
Art Slides. All the figures, photos, and tables from the text are offered as
JPEGs, both separately and embedded in a PowerPoint for each chapter. All text
art is enhanced for optimal viewing when projected in large classrooms.
Coursepack (Blackboard, Canvas, Angel, Moodle, and other LMS systems).
Available at no cost to professors or students, Norton coursepacks for online,
hybrid, or lecture courses are available in a variety of formats. With a simple
download from the instructor’s website, an adopter can bring high-quality Norton
digital media into a new or existing online course (no extra student passwords
required), and it’s theirs to keep. Instructors can edit assignments at the question level and set up custom grading policies to assess student understanding. In
addition to the instructor resources listed above, the coursepack includes additional chapter quizzes, flashcards, chapter outlines, chapter summaries, all of the
Education, Law, and Research Methods essays described above, and additional
questions on the essays.
Acknowledgments
Finally, let me turn to the happiest of chores—thanking all of those who have
contributed to this book. I begin with those who helped with the previous editions: Bob Crowder (Yale University) and Bob Logie (University of Aberdeen)
both read the entire text of the first edition, and the book was unmistakably improved by their insights. Other colleagues read, and helped me enormously with, specific chapters: Enriqueta Canseco-Gonzalez (Reed College),
Rich Carlson (Pennsylvania State University), Henry Gleitman (University
of Pennsylvania), Lila Gleitman (University of Pennsylvania), Peter Graf
Preface
•
xxi
(University of British Columbia), John Henderson (Michigan State University),
Jim Hoffman (University of Delaware), Frank Keil (Cornell University), Mike
McCloskey (Johns Hopkins University), Hal Pashler (UCSD), Steve Pinker
(MIT), and Paul Rozin (University of Pennsylvania).
The second edition was markedly strengthened by the input and commentary
provided by Martin Conway (University of Bristol), Kathleen Eberhard (Notre
Dame University), Howard Egeth (Johns Hopkins University), Bill Gehring
(University of Michigan), Steve Palmer (University of California, Berkeley), Henry
Roediger (Washington University), and Eldar Shafir (Princeton University).
In the third edition, I was again fortunate to have the advice, criticism, and
insights provided by a number of colleagues who, together, made the book
better than it otherwise could have been, and I’d like to thank Rich Carlson (Penn
State), Richard Catrambone (Georgia Tech), Randall Engle (Georgia Tech), Bill
Gehring and Ellen Hamilton (University of Michigan), Nancy Kim (Rochester
Institute of Technology), Steve Luck (University of Iowa), Michael Miller
(University of California, Santa Barbara), Evan Palmer, Melinda Kunar, and
Jeremy Wolfe (Harvard University), Chris Shunn (University of Pittsburgh), and
Daniel Simons (University of Illinois).
A number of colleagues also provided their insights and counsel for the fourth
edition. I’m therefore delighted to thank Ed Awh (University of Oregon), Glen
Bodner (University of Calgary), William Gehring (University of Michigan),
Katherine Gibbs (University of California, Davis), Eliot Hazeltine (University of
Iowa), William Hockley (Wilfrid Laurier University), James Hoffman (University
of Delaware), Helene Intraub (University of Delaware), Vikram Jaswal (University
of Virginia), Karsten Loepelmann (University of Alberta), Penny Pexman
(University of Calgary), and Christy Porter (College of William and Mary).
Then, even more people to thank for their help with the fifth edition: Karin M.
Butler (University of New Mexico), Mark A. Casteel (Penn State University, York),
Alan Castel (University of California, Los Angeles), Robert Crutcher (University
of Dayton), Kara D. Federmeier (University of Illinois, Urbana-Champaign),
Jonathan Flombaum (Johns Hopkins University), Katherine Gibbs (University of
California, Davis), Arturo E. Hernandez (University of Houston), James Hoeffner
(University of Michigan), Timothy Jay (Massachusetts College of Liberal Arts),
Timothy Justus (Pitzer College), Janet Nicol (University of Arizona), Robyn T.
Oliver (Roosevelt University), Raymond Phinney (Wheaton College, and his
comments were especially thoughtful!), Brad Postle (University of Wisconsin,
Madison), Erik D. Reichle (University of Pittsburgh), Eric Ruthruff (University of
New Mexico), Dave Sobel (Brown University), Martin van den Berg (California
State University, Chico), and Daniel R. VanHorn (North Central College).
For the sixth edition: Michael Dodd (University of Nebraska, Lincoln),
James Enns (University of British Columbia), E. Christina Ford (Penn State
University), Danielle Gagne (Alfred University), Marc Howard (Boston University),
B. Brian Kuhlman (Boise State University), Guy Lacroix (Carleton University), Ken
Manktelow (University of Wolverhampton), Aidan Moran (University College
Dublin, Ireland), Joshua New (Barnard College), Janet Nicol (University of
xxii •
Preface
Arizona), Mohammed K. Shakeel (Kent State University), David Somers (Boston
University), and Stefan Van der Stigchel (Utrecht University).
And now, happily, for the current edition: Alan Castel (UCLA), Jim Hoelzle
(Marquette University), Nate Kornell (Williams College), Catherine Middlebrooks (UCLA), Stefanie Sharman (Deakin University), Erin Sparck (UCLA), and
Cara Laney Thede (The College of Idaho).
I also want to thank the people at Norton. I’ve had a succession of terrific
editors, and I’m grateful to Ken Barton, Sheri Snavely, Aaron Javsicas, and Jon
Durbin for their support and fabulous guidance over the years. There’s no question that the book is stronger, clearer, and better because of their input and advice.
I also want to thank David Bradley for doing a fabulous job of keeping this
project on track, and also Eve Sanoussi and Katie Pak for their extraordinary
work in helping me bring out a book of the highest quality. (I should also thank
them for putting up with my gruff impatience when things don’t work quite
as we all hoped.) I’m also delighted with the design that Rubina Yeh, Jillian
Burr, and Lisa Buckley created; Ted Szczepanski has been great in dealing with
my sometimes-zany requests for photographs. Thanks in advance to Ashley
Sherwood and the Norton sales team; I am, of course, deeply grateful for all you
do. Thanks also to the crew that has produced the ZAPS, the Internet presence,
and the various supplements for this book.
And I once again get to celebrate the pleasure of working with Alice Vigliani
as manuscript editor. Alice is, of course, a terrific editor, but, in addition, she’s efficient and fun to work with. As I’ve said before: If she tends her peonies with the
same skill that she devotes to her editing, her garden must be marvelous indeed.
Finally, it brings me joy to reiterate with love the words I said in the previous
edition: In countless ways, Friderike makes all of this possible and worthwhile.
She forgives me the endless hours at the computer, tolerates the tension when
I’m feeling overwhelmed by deadlines, and is always ready to read my pages and
offer thoughtful, careful, instructive insights. My gratitude to, and love for, her
are boundless.
Daniel Reisberg
Portland, Oregon
Preface
•
xxiii
The Foundations
of Cognitive
Psychology
part
1
W
hat is cognitive psychology? In Chapter 1, we’ll define this discipline and offer a sketch of what this field can teach us — through
its theories and its practical applications. We’ll also provide
a brief history to explain why cognitive psychology has taken the form that
it has.
Chapter 2 has a different focus. At many points in this book, we’ll draw
insights from the field of cognitive neuroscience — the effort toward understanding our mental functioning through close study of the brain and nervous
system. To make sure this biological evidence is useful, though, we need to
provide some background, and that’s the main purpose of Chapter 2. There,
we’ll offer a rough mapping of what’s where in the brain, and we’ll describe
the functioning of many of the brain’s parts. We’ll also discuss what it means
to describe the functioning of this or that brain region, because, as we will see,
each of the brain’s parts is highly specialized in what it does. As a result, mental
achievements such as reading, remembering, or deciding depend on the coordinated functioning of many different brain regions, with each contributing its
own small bit to the overall achievement.
1
1
chapter
The Science of
the Mind
Almost everything you do, and everything you feel or say, depends on
your cognition — what you know, what you remember, and what you think.
As a result, the book you’re now reading — a textbook on cognition —
describes the foundation for virtually every aspect of who you are.
As illustrations of this theme, in a few pages we’ll consider the way
in which your ability to cope with grief depends on how your memory
functions. We’ll also discuss the role that memory plays in shaping your
self-image — and, therefore, your self-esteem. As another example, we’ll
discuss a case in which your understanding of a simple story depends on
the background knowledge that you supply. Examples like these make it
clear that cognition matters in an extraordinary range of circumstances,
and it’s on this basis that our focus in this book is on the intellectual
foundation of almost every aspect of human experience.
The Scope of Cognitive Psychology
When the field of cognitive psychology was first launched, it was broadly
focused on the scientific study of knowledge, and this focus led immediately
to a series of questions: How is knowledge acquired? How is knowledge
retained so that it’s available when needed? How is knowledge used — whether
as a basis for making decisions or as a means of solving problems?
These are great questions, and it’s easy to see that answering them might
be quite useful. For example, imagine that you’re studying for next Wednesday’s exam, but for some reason the material just won’t “stick” in your
memory. You find yourself wishing, therefore, for a better strategy to use in
studying and memorizing. What would that strategy be? Is it possible to have
a “better memory”?
As a different case, let’s say that while you’re studying, your friend is moving around in the room, and you find this quite distracting. Why can’t you
just shut out your friend’s motion? Why don’t you have better control over
your attention and your ability to concentrate?
Here’s one more example: You’re looking at your favorite Internet news
site, and you’re horrified to learn how many people have decided to vote for
candidate X. How do people decide whom to vote for? For that matter, how
do people decide what college to attend, or which car to buy, or even what
to have for dinner? And how can we help people make better decisions — so
that, for example, they choose healthier foods, or vote for the candidate who
(in your view) is preferable?
3
preview of chapter themes
•
•
he chapter begins with a description of the scope of
T
cognitive psychology. The domain of this field includes
activities that are obviously “intellectual” (such as remembering, paying attention, or making judgments) but also
a much broader range of activities that depend on these
intellectual achievements.
hat form should a “science of the mind” take? We disW
cuss the difficulties in trying to study the mind by means
of direct observation. But we also explore why we must
study the mental world if we’re to understand behavior;
the reason is that our behavior depends in crucial ways on
how we perceive and understand the world around us.
•
ombining these themes, we come to the view that we
C
must study the mental world indirectly. But as we will see,
the method for doing this is the method used by most
sciences.
Before we’re through, we’ll consider evidence pertinent to all of these
questions. Let’s note, though, that in these examples, things aren’t going
as you might have wished: You remember less than you want to; you can’t
ignore a distraction; the voters make a choice you don’t like. What about
the other side of the picture? What about the remarkable intellectual feats
that humans achieve — brilliant deductions or creative solutions to complex
problems? In this text, we’ll also discuss these cases and explore how people
manage to accomplish the great things they do.
CELEBRATING HUMAN ACHIEVEMENTS
Many of the text’s examples involve failures or limitations in our cognition. But we
also need to explain the incredible intellectual achievements of our species — the
complex problems we’ve solved and the extraordinary devices we’ve invented.
4
•
C H A P T E R O N E The Science of the Mind
The Broad Role for Memory
The questions we’ve mentioned so far might make it sound like cognitive
psychology is concerned just with your functioning as an intellectual —
your ability to remember, or to pay attention, or to think through options
when making a choice. As we’ve said, though, the relevance of cognitive
psychology is much broader — thanks to the fact that a huge range of your
actions, thoughts, and feelings depend on your cognition. As one way to
convey this point, let’s ask: When we investigate how memory functions,
what’s at stake? Or, to turn this around, what aspects of your life depend
on memory?
You obviously rely on memory when you’re taking an exam — memory
for what you learned during the term. Likewise, you rely on memory when
you’re at the supermarket and trying to remember the cheesecake recipe so
that you can buy the right ingredients. You also rely on memory when you’re
reminiscing about childhood. But what else draws on memory?
Consider this simple story (adapted from Charniak, 1972):
Betsy wanted to bring Jacob a present. She shook her piggy bank. It
made no sound. She went to look for her mother.
This four-sentence tale is easy to understand, but only because you provided
important bits of background. For example, you weren’t at all puzzled about
why Betsy was interested in her piggy bank; you weren’t puzzled, specifically,
about why the story’s first sentence led naturally to the second. This is
because you already knew (a) that the things one gives as presents are often
things bought for the occasion (rather than things already owned), (b) that
buying things requires money, and (c) that money is sometimes stored in
piggy banks. Without these facts, you would have wondered why a desire to
give a gift would lead someone to her piggy bank. (Surely you didn’t think
Betsy intended to give the piggy bank itself as the present!) Likewise, you
immediately understood why Betsy shook her piggy bank. You didn’t suppose
that she was shaking it in frustration or trying to find out if it would make a
good percussion instrument. Instead, you understood that she was trying to
determine its contents. But you knew this fact only because you already knew
(d) that Betsy was a child (because few adults keep their money in piggy
banks), (e) that children don’t keep track of how much money is in their
banks, and (f) that piggy banks are made out of opaque material (and so
a child can’t simply look into the bank to see what’s inside). Without these
facts, Betsy’s shaking of the bank would make no sense. Similarly, you understood what it meant that the bank made no sound. That’s because you know
(g) that it’s usually coins (not bills) that are kept in piggy banks, and (h) that
coins make noise when they’re shaken. If you didn’t know these facts, you
might have interpreted the bank’s silence, when it was shaken, as good news,
indicating perhaps that the bank was jammed full of $20 bills — an inference
that would have led you to a very different expectation for how the story
would unfold from there.
A SIMPLE STORY
What is involved in your
understanding of this simple
story? Betsy wanted to bring
Jacob a present. She shook
her piggy bank. It made no
sound. She went to look for
her mother.
The Scope of Cognitive Psychology
•
5
TRYING TO FOCUS
Often, you want to focus on
just one thing, and you want
to shut out the other sights
and sounds that are making
it hard for you to concentrate. What steps should you
take to promote this focus
and to avoid distraction?
6 •
Of course, there’s nothing special about the “Betsy and Jacob” story,
and we’d uncover a similar reliance on background knowledge if we
explored how you understand some other narrative, or follow a conversation, or comprehend a TV show. Our suggestion, in other words, is that
many (perhaps all) of your encounters with the world depend on your
supplementing your experience with knowledge that you bring to the situation.
And perhaps this has to be true. After all, if you didn’t supply the relevant
bits of background, then anyone telling the “Betsy and Jacob” story would
need to spell out all the connections and all the assumptions. That is, the
story would have to include all the facts that, with memory, are supplied
by you. As a result, the story would have to be much longer, and the telling of it much slower. The same would be true for every story you hear,
every conversation you participate in. Memory is thus crucial for each of
these activities.
Amnesia and Memory Loss
Here is a different sort of example: In Chapter 7, we will consider cases of
clinical amnesia — cases in which someone, because of brain damage, has lost
the ability to remember certain materials. These cases are fascinating at many
levels and provide key insights into what memory is for. Without memory,
what is disrupted?
H.M. was in his mid-20s when he had brain surgery intended to control
his severe epilepsy. The surgery was, in a narrow sense, a success, and H.M.’s
epilepsy was brought under control. But this gain came at an enormous cost,
because H.M. essentially lost the ability to form new memories. He survived
for more than 50 years after the operation, and for all those years he had
little trouble remembering events prior to the surgery. But H.M. seemed completely unable to recall any event that occurred after his operation. If asked
who the president is, or about recent events, he reported facts and events that
were current at the time of the surgery. If asked questions about last week, or
even an hour ago, he recalled nothing.
This memory loss had massive consequences for H.M.’s life, and some
of the consequences are surprising. For example, he had an uncle he was
very fond of, and he occasionally asked his hospital visitors how his uncle
was doing. Unfortunately, the uncle died sometime after H.M.’s surgery, and
H.M. was told this sad news. The information came as a horrible shock, but
because of his amnesia, H.M. soon forgot about it.
Sometime later, because he’d forgotten about his uncle’s death, H.M. again
asked how his uncle was doing and was again told of the death. But with no
memory of having heard this news before, he was once more hearing it “for
the first time,” with the shock and grief every bit as strong as it was initially.
Indeed, each time he heard this news, he was hearing it “for the first time.”
With no memory, he had no opportunity to live with the news, to adjust to it.
As a result, his grief could not subside. Without memory, H.M. had no way
to come to terms with his uncle’s death.
C H A P T E R O N E The Science of the Mind
H.M.’S BRAIN
When H.M. died in 2008, the world learned his full name — Henry Molaison. Throughout his life, H.M. had cooperated with researchers in many studies of his memory
loss. Even after his death, H.M. is contributing to science: His brain (shown here) was
frozen and has now been sliced into sections for detailed anatomical study. Unfortunately, though, there has been debate over who “owns” H.M.’s brain and how we
might interpret some observations about his brain (see, for example, Dittrich, 2016).
A different glimpse of memory function comes from some of H.M.’s
comments about what it felt like to be in his situation. Let’s start here with
the notion that for those of us without amnesia, numerous memories support our conception of who we are: We know whether we deserve praise for
our good deeds or blame for our transgressions because we remember those
good deeds and transgressions. We know whether we’ve kept our promises
or achieved our goals because, again, we have the relevant memories. None
of this is true for people who suffer from amnesia, and H.M. sometimes commented that in important ways, he didn’t know who he was. He didn’t know
if he should be proud of his accomplishments or ashamed of his crimes; he
didn’t know if he’d been clever or stupid, honorable or dishonest, industrious or lazy. In a sense, then, without a memory, there is no self. (For broader
discussion, see Conway & Pleydell-Pearce, 2000; Hilts, 1995.)
What, then, is the scope of cognitive psychology? As we mentioned earlier,
this field is sometimes defined as the scientific study of the acquisition, retention, and use of knowledge. We’ve now seen, though, that “knowledge” (and
hence the study of how we gain and use knowledge) is relevant to a huge
range of concerns. Our self-concept, it seems, depends on our knowledge
(and, in particular, on our memory for various episodes in our past). Our
The Scope of Cognitive Psychology
•
7
TEST YOURSELF
1.Why is memory crucial
for behaviors and
mental operations that
don’t in any direct or
explicit way ask you
“to remember”?
2. What aspects of H.M.’s
life were disrupted as
a result of his amnesia?
emotional adjustments to the world rely on our memories. Even our ability to understand a simple story — or, presumably, our ability to understand
any experience — depends on our supplementing that experience with some
knowledge.
The suggestion, then, is that cognitive psychology can help us understand
capacities relevant to virtually every moment of our lives. Activities that don’t
appear to be intellectual would collapse without the support of our cognitive functioning. The same is true whether we’re considering our physical
movements through the world, our social lives, our emotions, or any other
domain. This is the scope of cognitive psychology and, in a real sense, the
scope of this book.
The Cognitive Revolution
The enterprise that we now call “cognitive psychology” is a bit more than
50 years old, and the emergence of this field was in some ways dramatic.
Indeed, the science of psychology went through a succession of changes in
the 1950s and 1960s that are often referred to as psychology’s “cognitive revo­
lution.” This “revolution” involved a new style of research, aimed initially
at questions we’ve already met: questions about memory, decision making,
and so on. But this new type of research, and its new approach to theorizing,
soon influenced other domains, with the result that the cognitive revolution
dramatically changed the intellectual map of our field.
The cognitive revolution centered on two key ideas. One idea is that
the science of psychology cannot study the mental world directly. A second
idea is that the science of psychology must study the mental world if we’re
going to understand behavior. As a path toward understanding these ideas,
let’s look at two earlier traditions in psychology that offered a rather different perspective. Let’s emphasize, though, that our purpose here is not
to describe the full history of modern cognitive psychology. That history
is rich and interesting, but our goal is a narrow one — to explain why the
cognitive revolution’s themes were as they were. (For readers interested in
the history, see Bartlett, 1932; Benjamin, 2008; Broadbent, 1958; Malone,
2009; Mandler, 2011.)
The Limits of Introspection
In the late 19th century, Wilhelm Wundt (1832–1920) and his student
Edward Bradford Titchener (1867–1927) launched a new research enterprise,
and according to many scholars it was their work that eventually led to the
modern field of experimental psychology. In Wundt’s and Titchener’s view,
psychology needed to focus largely on the study of conscious mental events —
feelings, thoughts, perceptions, and recollections. But how should these events
be studied? These early researchers started with the fact that there is no way
for you to experience my thoughts, or I yours. The only person who can experience or observe your thoughts is you. Wundt, Titchener, and their colleagues
8 •
C H A P T E R O N E The Science of the Mind
WILHELM WUNDT
Wilhelm Wundt (1832–1920)
is shown here sitting and
surrounded by his colleagues
and students. Wundt is often
regarded as the “father of
experimental psychology.”
concluded, therefore, that the only way to study thoughts is through introspection, or “looking within,” to observe and record the content of our own
mental lives and the sequence of our own experiences.
Wundt and Titchener insisted, though, that this introspection could not
be casual. Instead, introspectors had to be meticulously trained: They were
given a vocabulary to describe what they observed; they were taught to be as
careful and as complete as possible; and above all, they were trained simply
to report on their experiences, with a minimum of interpretation.
This style of research was enormously influential for several years, but
psychologists gradually became disenchanted with it, and it’s easy to see
why. As one concern, these investigators soon had to acknowledge that
some thoughts are unconscious, which meant that introspection was limi­ted
as a research tool. After all, by its very nature introspection is the study
of conscious experiences, so of course it can tell us nothing about unconscious events.
Indeed, we now know that unconscious thought plays a huge part in
our mental lives. For example, what is your middle name? Most likely, the
moment you read this question, the name “popped” into your thoughts without any effort. But, in fact, there’s good reason to think that this simple bit of
remembering requires a complex series of steps. These steps take place outside
of awareness; and so, if we rely on introspection as our means of studying
mental events, we have no way of examining these processes.
The Cognitive Revolution
•
9
But there’s another, deeper problem with introspection. In order for any
science to proceed, there must be some way to test its claims; otherwise, we
have no means of separating correct assertions from false ones, accurate
descriptions of the world from fictions. Along with this requirement, science needs some way of resolving disagreements. If you claim that Earth
has one moon and I insist that it has two, we need some way of determining who is right. Otherwise, our “science” will become a matter of opinion,
not fact.
With introspection, this testability of claims is often unattainable. To see
why, imagine that I insist my headaches are worse than yours. How could
we ever test my claim? It might be true that I describe my headaches in
extreme terms: I talk about them being “agonizing” and “excruciating.” But
that might indicate only that I like to use extravagant descriptions; those
words might reveal my tendency to exaggerate (or to complain), not the
actual severity of my headaches. Similarly, it might be true that I need bed
rest whenever one of my headaches strikes. Does that mean my headaches are
truly intolerable? It might mean instead that I’m self-indulgent and rest even
when I feel mild pain. Perhaps our headaches are identical, but you’re stoic
about yours and I’m not.
How, therefore, should we test my claim about my headaches? What we
need is some way of directly comparing my headaches to yours, and that
would require transplanting one of my headaches into your experience,
or vice versa. Then one of us could make the appropriate comparison. But
(setting aside science fiction or fantasy) there’s no way to do this, leaving us,
in the end, unable to determine whether my headache reports are distorted or
accurate. We’re left, in other words, with the brute fact that our only information about my headaches is what comes through the filter of my description, and we have no way to know how (or whether) that filter is coloring
the evidence.
For purposes of science, this is unacceptable. Ultimately, we do want
to understand conscious experience, and so, in later chapters, we will consider introspective reports. For example, we’ll talk about the subjective
feeling of “familiarity” and the conscious experience of mental imagery; in
Chapter 14, we’ll talk about consciousness itself. In these settings, though,
we’ll rely on introspection as a source of observations that need to be
explained. We won’t rely on introspective data as a means of evaluating
our hypotheses — because, usually, we can’t. If we want to test hypotheses,
we need data we can rely on, and, among other requirements, this means
data that aren’t dependent on a particular point of view or a particular
descriptive style. Scientists generally achieve this objectivity by making
sure the raw data are out in plain view, so that you can inspect my evidence,
and I can inspect yours. In that way, we can be certain that neither of us is
distorting or misreporting the facts. And that is precisely what we cannot
do with introspection.
10 •
C H A P T E R O N E The Science of the Mind
The Years of Behaviorism
Historically, the concerns just described led many psychologists to abandon
introspection as a research tool. Psychology couldn’t be a science, they
argued, if it relied on this method. Instead, psychology needed objective data,
and that meant data out in the open for all to observe.
What sorts of data does this allow? First, an organism’s behaviors are
observable in the right way: You can watch my actions, and so can anyone
else who is appropriately positioned. Therefore, data concerned with behavior are objective data and thus grist for the scientific mill. Likewise, stimuli
in the world are in the same “objective” category: These are measurable,
recordable, physical events.
In addition, you can arrange to record the stimuli I experience day after
day after day and also the behaviors I produce each day. This means that
you can record how the pattern of my behavior changes over time and with
the accumulation of experience. In other words, my learning history can be
objectively recorded and scientifically studied.
In contrast, my beliefs, wishes, goals, preferences, hopes, and expectations
cannot be directly observed, cannot be objectively recorded. These “mentalistic” notions can be observed only via introspection; and introspection,
we’ve suggested, has little value as a scientific tool. Therefore, a scientific
psychology needs to avoid these invisible internal entities.
This perspective led to the behaviorist movement, a movement that dominated psychology in America for the first half of the 20th century. The movement
JOHN B. WATSON
John B. Watson (1878–1958) was a prominent
and persuasive advocate for the behaviorist movement. Given his focus on learning
and learning histories, it’s not surprising that
Watson was intrigued by babies’ behavior
and learning. Here, he tests the grasp reflex
displayed by human infants.
The Cognitive Revolution
•
11
was in many ways successful and uncovered a range of principles concerned
with how behavior changes in response to various stimuli (including the
stimuli we call “rewards” and “punishments”). By the late 1950s, however, psychologists were convinced that a lot of our behavior could not be explained
in these terms. The reason, basically, is that the ways people act, and the ways
they feel, are guided by how they understand or interpret the situation, and
not by the objective situation itself. Therefore, if we follow the behaviorists’
instruction and focus only on the objective situation, we will often misunderstand why people are doing what they’re doing and make the wrong predictions about how they’ll behave in the future. To put this point another way,
the behaviorist perspective demands that we not talk about mental entities
such as beliefs, memories, and so on, because these things cannot be studied
directly and so cannot be studied scientifically. Yet it seems that these subjective entities play a pivotal role in guiding behavior, and so we must consider
them if we want to understand behavior.
Evidence pertinent to these assertions is threaded throughout the chapters
of this book. Over and over, we’ll find it necessary to mention people’s perceptions and strategies and understanding, as we explain why (and how)
they perform various tasks and accomplish various goals. Indeed, we’ve
already seen an example of this pattern. Imagine that we present the “Betsy and
Jacob” story to people and then ask various questions: Why did Betsy shake
her piggy bank? Why did she go to look for her mother? People’s responses
will surely reflect their understanding of the story, which in turn depends on
far more than the physical stimulus — that is, the 29 syllables of the story
itself. If we want to predict someone’s responses to these questions, therefore,
we’ll need to refer to the stimulus (the story itself) and also to the person’s
knowledge and understanding of this stimulus.
Here’s a different example that makes the same general point. Imagine
you’re sitting in the dining hall. A friend produces this physical stimulus:
“Pass the salt, please,” and you immediately produce a bit of salt-passing
behavior. In this exchange, there is a physical stimulus (the words your friend
uttered) and an easily defined response (your passing of the salt), and so
this simple event seems fine from the behaviorists’ perspective — the elements
are out in the open, for all to observe, and can be objectively recorded. But
note that the event would have unfolded in the same way if your friend had
offered a different stimulus. “Could I have the salt?” would have done the
trick. Ditto for “Salt, please!” or “Hmm, this sure needs salt!” If your friend is
both loquacious and obnoxious, the utterance might have been: “Excuse me,
but after briefly contemplating the gustatory qualities of these comestibles, I
have discerned that their sensory qualities would be enhanced by the addition
of a number of sodium and chloride ions, delivered in roughly equal proportions and in crystalline form; could you aid me in this endeavor?” You might
giggle (or snarl) at your friend, but you would still pass the salt.
Now let’s work on the science of salt-passing behavior. When is this
behavior produced? We’ve just seen that the behavior is evoked by a number
of different stimuli, and so we would surely want to ask: What do these
12 •
C H A P T E R O N E The Science of the Mind
stimuli have in common? If we can answer that question, we’re on our way
to understanding why these stimuli all have the same effect.
The problem, though, is that if we focus on the observable, objective
aspects of these stimuli, they actually have little in common. After all, the
sounds being produced in that long statement about sodium and chloride
ions are rather different from the sounds in the utterance “Salt, please!” And
in many circumstances, similar sounds would not lead to salt-passing behavior. Imagine that your friend says, “Salt the pass” or “Sass the palt.” These
are acoustically similar to “Pass the salt” but wouldn’t have the same impact.
Or imagine that your friend says, “She has only a small part in the play. All
she gets to say is ‘Pass the salt, please.’” In this case, the right syllables were
uttered, but you wouldn’t pass the salt in response.
It seems, then, that our science of salt passing won’t get very far if we
insist on talking only about the physical stimulus. Stimuli that are physically
different from each other (“Salt, please” and the bit about ions) have similar
effects. Stimuli that are physically similar to each other (“Pass the salt” and
“Sass the palt”) have different effects. Physical similarity, therefore, is not
what unites the various stimuli that evoke salt passing.
It’s clear, though, that the various stimuli that evoke salt passing do have
something in common: They all mean the same thing. Sometimes this meaning derives from the words themselves (“Please pass the salt”). In other cases,
the meaning depends on certain pragmatic rules. (For example, you understand that the question “Could you pass the salt?” isn’t a question about arm
strength, although, interpreted literally, it might be understood that way.) In
all cases, though, it seems plain that to predict your behavior in the dining
hall, we need to ask what these stimuli mean to you. This seems an extraordinarily simple point, but it is a point, echoed by countless other examples, that
indicates the impossibility of a complete behaviorist psychology.1
PASSING THE SALT
If a friend requests the salt,
your response will depend
on how you understand your
friend’s words. This is a simple point, echoed in example
after example, but it is the
reason why a rigid behaviorist perspective cannot explain
your behavior.
The Intellectual Foundations of the
Cognitive Revolution
One might think, then, that we’re caught in a trap. On one side, it seems that the
way people act is shaped by how they perceive the situation, how they understand
the stimuli, and so on. If we want to explain behavior, then, we have no choice.
We need to talk about the mental world. But, on the other side, the only direct
means of studying the mental world is introspection, and introspection is scientifically unworkable. Therefore: We need to study the mental world, but we can’t.
There is, however, a solution to this impasse, and it was suggested years ago
by the philosopher Immanuel Kant (1724–1804). To use Kant’s transcendental
method, you begin with the observable facts and then work backward from
1. The behaviorists themselves quickly realized this point. As a result, modern behaviorism has
abandoned the radical rejection of mentalistic terms; indeed, it’s hard to draw a line between
modern behaviorism and a field called “animal cognition,” a field that often uses mentalistic
language! The behaviorism being criticized here is a historically defined behaviorism, and it’s
this perspective that, in large measure, gave birth to modern cognitive psychology.
The Cognitive Revolution
•
13
IMMANUEL KANT
Philosopher Immanuel Kant
(1724–1804) made major contributions to many fields,
and his transcendental method
enabled him to ask what
qualities of the mind make
experience possible.
14 •
these observations. In essence, you ask: How could these observations have
come about? What must be the underlying causes that led to these effects?
This method, sometimes called “inference to best explanation,” is at the
heart of most modern science. Physicists, for example, routinely use this
method to study objects or events that cannot be observed directly. To take just
one case, no physicist has ever observed an electron, but this hasn’t stopped
physicists from learning a great deal about electrons. How do the physicists
proceed? Even though electrons themselves aren’t observable, their presence
often leads to observable results — in essence, visible effects from an invisible
cause. For example, electrons leave observable tracks in cloud chambers, and
they can produce momentary fluctuations in a magnetic field. Physicists can
then use these observations in the same way a police detective uses clues —
asking what the “crime” must have been like if it left this and that clue.
(A size 11 footprint? That probably tells us what size feet the criminal has,
even though no one saw his feet. A smell of tobacco smoke? That suggests the
criminal was a smoker. And so on.) In the same way, physicists observe the clues
that electrons leave behind, and from this information they form hypotheses
about what electrons must be like in order to have produced those effects.
Of course, physicists (and other scientists) have a huge advantage over a
police detective. If the detective has insufficient evidence, she can’t arrange
for the crime to happen again in order to produce more evidence. (She can’t
say to the robber, “Please visit the bank again, but this time don’t wear
a mask.”) Scientists, in contrast, can arrange for a repeat of the “crime”
they’re seeking to explain — they can arrange for new experiments, with
new measures. Better still, they can set the stage in advance, to maximize
the likelihood that the “culprit” (in our example, the electron) will leave
useful clues behind. They can, for example, add new recording devices to the
situation, or they can place various obstacles in the electron’s path. In this
way, scientists can gather more and more data, including data crucial for
testing the predictions of a particular theory. This prospect — of reproducing
experiments and varying the experiments to test hypotheses — is what gives
science its power. It’s what enables scientists to assert that their hypotheses
have been rigorously tested, and it’s what gives scientists assurance that their
theories are correct.
Psychologists work in the same fashion — and the notion that we could
work in this fashion was one of the great contributions of the cognitive revolution. The idea is this: We know that we need to study mental processes;
that’s what we learned from the limitations of classical behaviorism. But we
also know that mental processes cannot be observed directly; we learned that
from the downfall of introspection. Our path forward, therefore, is to study
mental processes indirectly, relying on the fact that these processes, themselves invisible, have visible consequences: measurable delays in producing a
response, performances that can be assessed for accuracy, errors that can be
scrutinized and categorized. By examining these (and other) effects produced
by mental processes, we can develop — and test — hypotheses about what the
mental processes must have been. In this way, we use Kant’s method, just as
C H A P T E R O N E The Science of the Mind
physicists (or biologists or chemists or astronomers) do, to develop a science
that does not rest on direct observation.
The Path from Behaviorism to the
Cognitive Revolution
In setting after setting, cognitive psychologists have applied the Kantian logic
to explain how people remember, make decisions, pay attention, or solve
problems. In each case, we begin with a particular performance — say, a problem that someone solved — and then hypothesize a series of unseen mental
events that made the performance possible. But we don’t stop there. We also
ask whether some other, perhaps simpler, sequence of events might explain
the data. In other words, we do more than ask how the data came about; we
seek the best way to think about the data.
This pattern of theorizing has become the norm in psychology — a powerful indication that the cognitive revolution did indeed change the entire field.
But what triggered the revolution? What happened in the 1950s and 1960s
that propelled psychology forward in this way? It turns out that multiple
forces were in play.
One contribution came from within the behaviorist movement itself. We’ve
discussed concerns about classical behaviorism, and some of those concerns
were voiced early on by Edward Tolman (1886-1959) — a researcher who can
be counted both as a behaviorist and as one of the forerunners of cognitive
psychology. Prior to Tolman, most behaviorists argued that learning could
be understood simply as a change in behavior. Tolman argued, however, that
learning involved something more abstract: the acquisition of new knowledge.
In one of Tolman’s studies, rats were placed in a maze day after day. For
the initial 10 days, no food was available anywhere in the maze, and the rats
wandered around with no pattern to their behavior. Across these days, therefore, there was no change in behavior — and so, according to the conventional
view, no learning. But, in fact, there was learning, because the rats were learning the layout of the maze. That became clear on the 11th day of testing, when
food was introduced into the maze in a particular location. The next day, the
rats, placed back in the maze, ran immediately to that location. Indeed, their
behavior was essentially identical to the behavior of rats who had had many
days of training with food in the maze (Tolman, 1948; Gleitman, 1963).
What happened here? Across the initial 10 days, rats were acquiring what
Tolman called a “cognitive map” of the maze. In the early days of the procedure, however, the rats had no motivation to use this knowledge. On Days 11
and 12, though, the rats gained a reason to use what they knew, and at that
point they revealed their knowledge. The key point, though, is that — even for
rats — we need to talk about (invisible) mental processes (e.g., the formation
of cognitive maps) if we want to explain behavior.
A different spur to the cognitive revolution also arose out of behaviorism —
but this time from a strong critique of behaviorism. B.F. Skinner (1904–1990)
was an influential American behaviorist, and in 1957 he applied his style of
ULRIC NEISSER
Many intellectual developments led to the cognitive
revolution. A huge boost,
though, came from Ulric
Neisser’s book, Cognitive
Psychology (1967). Neisser’s
influence was so large that
many scholars refer to him
as the “father of cognitive
psychology.”
The Cognitive Revolution
•
15
analysis to humans’ ability to learn and use language, arguing that language
use could be understood in terms of behaviors and rewards (Skinner, 1957).
Two years later, the linguist Noam Chomsky (1928– ) published a ferocious
rebuttal to Skinner’s proposal, and convinced many psychologists that an
entirely different approach was needed for explaining language learning and
language use, and perhaps for other achievements as well.
European Roots of the Cognitive Revolution
Research psychology in the United States was, we’ve said, dominated by the
behaviorist movement for many years. The influence of behaviorism was not
as strong, however, in Europe, and several strands of European research fed
into and strengthened the cognitive revolution. In Chapter 3, we will describe
some of the theorizing that grew out of the Gestalt psychology movement, an
important movement based in Berlin in the early decades of the 20th century.
(Many of the Gestaltists fled to the United States in the years leading up to
World War II and became influential figures in their new home.) Overall, the
Gestalt psychologists argued that behaviors, ideas, and perceptions are organized in a way that could not be understood through a part-by-part, elementby-element, analysis of the world. Instead, they claimed, the elements take
on meaning only as part of the whole — and therefore psychology needed to
understand the nature of the “whole.” This position had many implications,
including an emphasis on the role of the perceiver in organizing his or her
experience. As we will see, this notion — that perceivers shape their own
experience — is a central theme for modern cognitive psychology.
Another crucial figure was British psychologist Frederic Bartlett (1886–1969).
Although he was working in a very different tradition from the Gestalt
psychologists, Bartlett also emphasized the ways in which each of us shapes
and organizes our experience. Bartlett claimed that people spontaneously fit
their experiences into a mental framework, or “schema,” and rely on this
schema both to interpret the experience as it happens and to aid memory
later on. We’ll say more about Bartlett’s work (found primarily in his book
Remembering, published in 1932) in Chapter 8.
FREDERIC BARTLETT
Frederic Bartlett was the
first professor of experimental psychology at the University of Cambridge. He is
best known for his studies
of memory and the notion
that people spontaneously
fit their experiences into a
“schema,” and they rely on
the schema both to guide
their understanding and (later)
to guide their memory.
16 •
Computers and the Cognitive Revolution
Tolman, Chomsky, the Gestaltists, and Bartlett disagreed on many points.
Even so, a common theme ran through their theorizing: These scholars all
agreed that we could not explain humans’ (or even rats’) behavior unless we
explain what is going on within the mind — whether our emphasis is on cognitive maps, schemata, or some other form of knowledge. But, in explaining
this knowledge and how the knowledge is put to use, where should we begin?
What sorts of processes or mechanisms might we propose?
Here we meet another crucial stream that fed into the cognitive revolution,
because in the 1950s a new approach to psychological explanation became
available and turned out to be immensely fruitful. This new approach was
C H A P T E R O N E The Science of the Mind
suggested by the rapid developments in electronic information processing,
including developments in computer technology. It soon became clear that computers were capable of immensely efficient information storage and retrieval
(“memory”), as well as performance that seemed to involve decision making
and problem solving. Indeed, some computer scientists proposed that computers would soon be genuinely intelligent — and the field of “artificial intelligence”
was launched and made rapid progress (e.g., Newell & Simon, 1959).
Psychologists were intrigued by these proposals and began to explore the
possibility that the human mind followed processes and procedures similar
to those used in computers. As a result, psychological data were soon being
explained in terms of “buffers” and “gates” and “central processors,” terms
borrowed from computer technology (e.g., Miller, 1956; Miller, Galanter,
& Pribram, 1960). This approach was evident, for example, in the work of
another British psychologist, Donald Broadbent (1926–1993). He was one of
the earliest researchers to use the language of computer science in explaining human cognition. His work emphasized a succession of practical issues,
including the mechanisms through which people focus their attention when
working in complex environments, and his book Perception and Communication (1958) framed discussions of attention for many years.
This computer-based vocabulary allowed a new style of theorizing. Given
a particular performance, say, in paying attention or on some memory task,
one could hypothesize a series of information-processing events that made
the performance possible. As we will see, hypotheses cast in these terms led
psychologists to predict a broad range of new observations, and in this way
both organized the available information and led to many new discoveries.
TEST YOURSELF
3. W
hy is introspection
limited as a source of
scientific evidence?
4. W
hy do modern
psychologists agree
that we have to refer
to mental states (what
you believe, what you
perceive, what you
understand) in order
to explain behavior?
5. D
escribe at least one
historical development
that laid the groundwork for the cognitive
revolution.
Research in Cognitive Psychology:
The Diversity of Methods
Over the last half-century, cognitive psychologists have continued to frame
many hypotheses in these computer-based terms. But we’ve also developed
other options for theorizing. For example, before we’re done in this book,
we’ll also discuss hypotheses framed in terms of the strategies a person is
relying on, or the inferences she is making. No matter what the form of the
hypothesis, though, the next steps are crucial. First, we derive new predictions from the hypothesis, along the lines of “If this is the mechanism behind
the original findings, then things should work differently in this circumstance
or that one.” Then, we gather new data to test those predictions. If the data fit
with the predictions, this outcome confirms the hypothesis. If the data don’t
line up with the predictions, a new hypothesis is needed.
But what methods do we use, and what sorts of data do we collect? The
answer, in brief, is that we use diverse methods and collect many types of
data. In other words, what unites cognitive psychology is not an allegiance
to any particular procedure in the laboratory. Instead, what unites the
field is the logic that underlies our research, no matter what method we use.
Research in Cognitive Psychology: The Diversity of Methods
•
17
(We discuss this logic more fully in the appendix for this textbook. The
appendix contains a series of modules, with each module exploring an aspect
of research methodology directly related to one of the book’s chapters.)
What sorts of data do we use? In some settings, we ask how well people
perform a particular task. For example, in tests of memory we might ask
how complete someone’s memory is (does the person remember all of the
objects in view in a picture?) and also how accurate the memory is (does the
person perhaps remember seeing a banana when, in truth, no banana was in
view?). We can also ask how performance changes if we change the “input”
(how well does the person remember a story, rather than a picture?), and we
can change the person’s circumstances (how is memory changed if the person
is happy, or afraid, when hearing the story?). We can also manipulate the
person’s plans or strategies (what happens if we teach the person some sort
of memorization technique?), and we can compare different people (children
vs. adults; novices at a task vs. experts; people with normal vision vs. people
who have been blind since birth).
A different approach relies on measurements of speed. The idea here is
that mental operations are fast but do take a measurable amount of time, and
by examining the response time (RT) — that is, how long someone needs to
make a particular response — we can often gain important insights into what’s
going on in the mind. For example, imagine that we ask you: “Yes or no: Do
cats have whiskers?” And then: “Yes or no: Do cats have heads?” Both questions are absurdly easy, so there’s no point in asking whether you’re accurate
in your responses — it’s a sure bet that you will be. We can, however, measure
your response times to questions like these, often with intriguing results. For
example, if you’re forming a mental picture of a cat when you’re asked these
questions, you’ll be faster for the “heads” question than the “whiskers” question. If you think about cats without forming a mental picture, the pattern
reverses — you’ll be faster for the “whiskers” question. In Chapter 11, we’ll
use results like these to test hypotheses about how information — and mental
pictures in particular — are represented and analyzed in your mind.
We can also gain insights from observations focused on the brain and
nervous system. Over the last few decades, cognitive psychology has formed
a productive partnership with the field of cognitive neuroscience, the effort
toward understanding humans’ mental functioning through close study of
the brain and nervous system. But here, too, numerous forms of evidence are
available. We’ll say more about these points in the next chapter, but for now
let’s note that we can learn a lot by studying people with damaged brains and
also people with healthy brains. Information about damaged brains comes
from the field of clinical neuropsychology, the study of brain function that
uses, as its main data source, cases in which damage or illness has disrupted
the working of some brain structure. We’ve already mentioned H.M., a man
whose memory was massively disrupted as an unexpected consequence of
surgery. As a different example, in Chapter 12 we’ll consider cases in which
someone’s ability to make ordinary decisions (Coke or Pepsi? Wear the blue
sweater or the green one?) is disrupted if brain centers involved in emotion
18 •
C H A P T E R O N E The Science of the Mind
are disrupted; observations like these provide crucial information about the
role of emotion in decision making.
Information about healthy brains comes from neuroimaging techniques,
which enable us, with some methods, to scrutinize the precise structure of the
brain and, with other methods, to track the moment-by-moment pattern of
activation within someone’s brain. We’ll see in Chapter 7, for example, that
different patterns of brain activation during learning lead to different types
of memory, and we’ll use this fact as we ask what the types of memory are.
There’s no reason for you, as a reader, to memorize this catalogue of different types of evidence. That’s because we’ll encounter each of these forms
of data again and again in this text. Our point for now is simply to highlight
the fact that there are multiple tools with which we can test, and eventually
confirm, various claims. Indeed, relying on these tools, cognitive psychology
has learned a tremendous amount about the mind. Our research has brought
us powerful new theories and enormously useful results. Let’s dive in and
start exploring what the science of the mind has taught us.
TEST YOURSELF
6. D
escribe at least three
types of evidence that
cognitive psychologists
routinely rely on.
APPLYING COGNITIVE PSYCHOLOGY
Research in cognitive psychology can help us understand deep theoretical
issues, such as what it means to be rational or what the function of consciousness might be. But our research also has broad practical implications,
and so our studies often provide lessons for how we should conduct our
daily lives.
Some of the practical lessons from cognitive psychology are obvious. For
example, research on memory can help students who are trying to learn new
materials in the classroom; studies of how people draw conclusions can help
people to draw smarter, more defensible conclusions. Following these leads,
each chapter in this text ends with an essay that explores how the material in
that chapter can be applied to an issue that’s important for education. This
emphasis is rooted, in part, in the fact that most readers of this book will be
college students, using the book in the context of one of their courses. I hope,
therefore, that the Cognitive Psychology and Education essays are directly
useful for these readers! Concretely, the essay at the end of Chapter 4, for
example, will teach you how to speed-read (but will also explain the limitations of speed-reading). The essay at the end of Chapter 6 will offer suggestions for how to study and retain the material you’re hoping to learn.
Let’s emphasize, though, that research in cognitive psychology also has
implications for other domains. For example, think about the criminal justice
system and what happens in a criminal investigation. Eyewitnesses provide
evidence, based on what they paid attention to during a crime and what they
remember. Police officers question the witnesses, trying to get the most out
of what each witness recalls — but without leading the witness in any way.
Then, the police try to deduce, from the evidence, who the perpetrator was.
Applying Cognitive Psychology
•
19
Later, during the trial, jurors listen to evidence and make a judgment about
the defendant’s innocence or guilt.
Cast in these terms, it should be obvious that an understanding of
attention, memory, reasoning, and judgment (to name just a few processes) is
directly relevant to what happens in the legal system. On this basis, therefore,
I’ve also written essays that focus on the interplay between cognitive psychology and the law. The essay for Chapter 3, for example, uses what we know
about visual perception to ask what we can expect witnesses to see. The essay
for Chapter 7 explores a research-based procedure for helping witnesses to
recall more of what they’ve observed. If you’re curious to see these Cognitive
Psychology and the Law essays, you can find them online in the ebook, available at http://digital.wwnorton.com/cognition7.
COGNITIVE PSYCHOLOGY AND THE CRIMINAL JUSTICE SYSTEM
Eyewitnesses in the courtroom rely on what they remember about key events, and
what they remember depends crucially on what they perceived and paid attention
to. Therefore, our understanding of memory, perception, and attention can help the
justice system in its evaluation of witness evidence.
20 •
C H A P T E R O N E The Science of the Mind
chapter review
SUMMARY
• Cognitive psychology is concerned with how
people remember, pay attention, and think. The
importance of all these issues arises partly from
the fact that most of what we do, say, and feel is
guided by things we already know. One example
is our comprehension of a simple story, which turns
out to be heavily influenced by the knowledge
we supply.
• Cognitive psychology emerged as a separate discipline in the late 1950s, and its powerful impact
on the wider field of psychology has led many academics to speak of this emergence as the cognitive
revolution. One predecessor of cognitive psychology was the 19th-century movement that emphasized introspection as the main research tool for
psychology. But psychologists soon became disenchanted with this movement for several reasons:
Introspection cannot inform us about unconscious
mental events; and even with conscious events,
claims rooted in introspection are often untestable because there is no way for an independent
observer to check the accuracy or completeness of
an introspective report.
• The behaviorist movement rejected introspection
as a method, insisting instead that psychology speak
only of mechanisms and processes that are objective and out in the open for all to observe. However, evidence suggests that our thinking, behavior,
and feelings are often shaped by our perception or
understanding of the events we experience. This is
problematic for the behaviorists: Perception and
understanding are exactly the sorts of mental processes that the behaviorists regarded as subjective
and not open to scientific study.
• In order to study mental events, psychologists
have turned to a method in which one focuses on
observable events but then asks what (invisible)
events must have taken place in order to make these
(visible) effects possible.
• Many factors contributed to the emergence
of cognitive psychology in the 1950s and 1960s.
Tolman’s research demonstrated that even in rats,
learning involved the acquisition of new knowledge
and not just a change in behavior. Chomsky argued
powerfully that a behaviorist analysis was inadequate as an explanation for language learning and
language use. Gestalt psychologists emphasized
the role of the perceiver in organizing his or her
experience. Bartlett’s research showed that people
spontaneously fit their experiences into a mental
framework, or schema.
• Early theorizing in cognitive psychology often
borrowed ideas from computer science, including
early work on artificial intelligence.
• Cognitive psychologists rely on a diverse set of
methods and collect many types of data. Included
are measures of the quality of someone’s performance, measures of response speed, and, in some
cases, methods that allow us to probe the under­
lying biology.
21
KEY TERMS
introspection (p. 9)
behaviorist theory (p. 11)
transcendental method (p. 13)
response time (RT) (p. 18)
cognitive neuroscience (p. 18)
clinical neuropsychology (p. 18)
neuroimaging techniques (p. 19)
TEST YOURSELF AGAIN
1.Why is memory crucial for behaviors and
mental operations that don’t in any direct or
explicit way ask you “to remember”?
5.Describe at least one historical development
that laid the groundwork for the cognitive
revolution.
2.What aspects of H.M.’s life were disrupted as a
result of his amnesia?
6.Describe at least three types of evidence that
cognitive psychologists routinely rely on.
3.Why is introspection limited as a source of
scientific evidence?
4.Why do modern psychologists agree that we
have to refer to mental states (what you believe,
what you perceive, what you understand) in
order to explain behavior?
THINK ABOUT IT
1.The chapter argues that in a wide range of settings, our behaviors and our emotions depend
on what we know, believe, and remember. Can
you come up with examples of your own that
illustrate this reliance on cognition in a circumstance that doesn’t seem, on the surface, to be
one that involves “intellectual activity”?
22
2.Some critics of Darwin’s theory of evolution
via natural selection argue this way: “Darwin’s
claims can never be tested, because of course no
one was around to observe directly the processes
of evolution that Darwin proposed.” Why
is this assertion misguided, resting on a false
notion of how science proceeds?
E eBOOK DEMONSTRATIONS & ESSAYS
Go to http://digital.wwnorton.com/cognition7 for the online demonstrations and essays
relevant to this chapter. These can be found in the ebook.
Online Applying Cognitive Psychology and the
Law Essays
• Cognitive Psychology and the Law: Improving
the Criminal Justice System
COGNITION LABS
Go to http://digital.wwnorton.com/cognition7 to experience interactive online psychology
labs relevant to this chapter.
23
2
chapter
The Neural Basis
for Cognition
what if…
Throughout this text, we’ll be examining ordinary
achievements. A friend asks: “Where’d you grow up?”
and you immediately answer. You’re meeting a friend at the airport,
and you instantly recognize her the moment she steps into view. An
instructor says, “Listen carefully,” and you have no trouble focusing
your attention.
Ordinary or not, achievements like these are crucial for you, and
your life would be massively disrupted if you couldn’t draw information from memory, or recognize the objects you encounter, or choose
where you’ll point your attention. As a way of dramatizing this point,
we’ll begin each chapter by asking: What would happen to someone if
one of these fundamental capacities didn’t work as it normally does?
What if . . . ?
The disorder known as Capgras syndrome (Capgras & ReboulLachaux, 1923) is relatively rare, but it can result from various injuries
to the brain (Ellis & De Pauw, 1994) and is sometimes found in people
with Alzheimer’s syndrome (Harwood, Barker, Ownby, & Duara, 1999).
Someone with this syndrome is fully able to recognize the people in her
world — her husband, her parents, her friends — but is utterly convinced
that these people are not who they appear to be. The real husband or
the real son, the afflicted person insists, has been kidnapped (or worse).
The person now in view, therefore, must be a fraud of some sort, impersonating the (allegedly) absent person.
Imagine what it’s like to have this disorder. You turn to your father
and exclaim, “You look like my father, sound like him, and act like him.
But I can tell that you’re not my father! Who are you?”
Often, a person with Capgras syndrome insists that there are slight
differences between the “impostor” and the person he (or she) has
supposedly replaced — subtle changes in personality or appearance.
Of course, no one else detects these (nonexistent) differences, which
can lead to paranoid suspicions about why a loved one has been
taken away and why no one else will acknowledge the replacement.
In the extreme, these suspicions can lead a Capgras sufferer to desperate steps. In some cases, patients suffering from this syndrome
have murdered the supposed impostor in an attempt to end the
charade and relocate the “genuine” character. In one case, a Capgras
patient was convinced his father had been replaced by a robot and so
25
preview of chapter themes
•
e begin by exploring the example of Capgras syndrome
W
to illustrate how seemingly simple achievements actually
depend on many parts of the brain. We also highlight the
ways that the study of the brain can illuminate questions
about the mind.
•
e then survey the brain’s anatomy, emphasizing the
W
function carried out by each region. Identification of these
functions is supported by neuroimaging data, which can
assess the activity levels in different areas, and by studies
of the effects of brain damage.
•
e then take a closer look at the various parts of the cereW
bral cortex — the most important part of the brain for cognitive functioning. These parts include the motor areas,
the sensory areas, and the so-called association cortex.
•
inally, we turn to the individual cells that make up the
F
brain — the neurons and glia — and discuss the basic principles of how these cells function.
decapitated him in order to look for the batteries and microfilm in his
head (Blount, 1986).
What is going on here? The answer lies in the fact that facial recognition involves two separate systems in the brain. One system leads to
a cognitive appraisal (“I know what my father looks like, and I can perceive that you closely resemble him”), and the other to a more global,
emotional appraisal (“You look familiar to me and also trigger a warm
response in me”). When these two appraisals agree, the result is a confident recognition (“You obviously are my father”). In Capgras syndrome,
though, the emotional processing is disrupted, leading to an intellectual
identification without a familiarity response (Ellis & Lewis, 2001; Ellis &
Young, 1990; Ramachandran & Blakeslee, 1998): “You resemble my father
but trigger no sense of familiarity, so you must be someone else.” The
result? Confusion and, at times, bizarre speculation about why a loved
one has been kidnapped and replaced — and a level of paranoia that can,
as we have seen, lead to homicide.
Explaining Capgras Syndrome
We began this chapter with a description of Capgras syndrome, and we’ve
offered an account of the mental processes that characterize this disorder.
Specifically, we’ve suggested that someone with this syndrome is able to recognize a loved one’s face, but with no feeling of familiarity. Is this the right
way to think about Capgras syndrome?
One line of evidence comes from neuroimaging techniques that enable
researchers to take high-quality, three-dimensional “pictures” of living brains
without in any way disturbing the brains’ owners. We’ll have more to say
about neuroimaging later; but first, what do these techniques tell us about
Capgras syndrome?
26 •
C H A P T E R T WO The Neural Basis for Cognition
The Neural Basis for Capgras Syndrome
Some types of neuroimaging provide portraits of the physical makeup of
the brain: What’s where? How are structures shaped or connected to each
other? Are there structures present (such as tumors) that shouldn’t be there,
or structures that are missing (because of disease or birth defects)? This information about structure was gained in older studies from positron emission tomography (more commonly referred to as a PET scan). More recent
studies usually rely on magnetic resonance imaging (MRI; see Figure 2.1).
These scans suggest a link between Capgras syndrome and abnormalities in
several brain areas, indicating that our account of the syndrome will need to
consider several elements (Edelstyn & Oyebode, 1999; also see O’Connor,
Walbridge, Sandson, & Alexander, 1996).
FIGURE 2.1
NEUROIMAGING
Scanners like this one are used for both MRI and fMRI scans. MRI scans tell us
about the structure of the brain; fMRI scans tell us which portions of the brain
are especially active during the scan. An fMRI scan usually results in color
images, with each hue indicating a particular activity level.
Explaining Capgras Syndrome
•
27
FIGURE 2.2
THE LOBES OF THE HUMAN BRAIN
Central fissure
Parietal lobe
Frontal lobe
Occipital
lobe
Lateral fissure
Temporal lobe
Cerebellum
A
B
Panel A identifies the various lobes and some of the brain’s prominent features. Actual brains, however, are uniformly colored, as shown in the photograph in Panel B. The four lobes of the forebrain surround (and hide from
view) the midbrain and most of the hindbrain. (The cerebellum is the only part of the hindbrain that is visible in the
figure, and, in fact, the temporal lobe has been pushed upward a bit in the left panel to make the cerebellum more
visible.) This side view shows the left cerebral hemisphere; the structures on the right side of the brain are similar.
However the two halves of the brain have somewhat different functions, and so the results of brain injury depend
on which half is damaged. The symptoms of Capgras syndrome, for example, result from damage to specific sites
on the right side of the frontal and temporal lobes.
One site of damage in Capgras patients is in the temporal lobe (see
Figure 2.2), particularly on the right side of the head. This damage probably disrupts circuits involving the amygdala, an almond-shaped structure
that — in the intact brain — seems to serve as an “emotional evaluator,”
helping an organism detect stimuli associated with threat or danger (see
Figure 2.3). The amygdala is also important for detecting positive stimuli —
indicators of safety or of available rewards. With damaged amygdalae,
therefore, people with Capgras syndrome won’t experience the warm sense
of feeling good (and safe and secure) when looking at a loved one’s familiar
28 •
C H A P T E R T WO The Neural Basis for Cognition
FIGURE 2.3 THE
AMYGDALA AS AN
“EMOTIONAL EVALUATOR”
The area shown in yellow marks
the location of the amygdala. In
this image, the yellow is a reflection of increased activity created
by a fear memory — the memory
of receiving an electric shock.
face. This lack of an emotional response is probably why these faces don’t
feel familiar to them, and is fully in line with the two-systems hypothesis
we’ve already sketched.
Patients with Capgras syndrome also have brain abnormalities in the
frontal lobe, specifically in the right prefrontal cortex. What is this area’s
normal function? To find out, we turn to a different neuroimaging technique,
functional magnetic resonance imaging (fMRI), which enables us to track
moment-by-moment activity levels in different sites in a living brain. (We’ll
say more about fMRI in a later section.) This technique allows us to answer
such questions as: When a person is reading, which brain regions are particularly active? How about when a person is listening to music? With data like
these, we can ask which tasks make heavy use of a brain area, and from that
base we can draw conclusions about that brain area’s function.
Studies make it clear that the prefrontal cortex is especially active when
a person is doing tasks that require planning or careful analysis. Conversely,
this area is less active when someone is dreaming. Plausibly, this latter pattern
Explaining Capgras Syndrome
•
29
reflects the absence of careful analysis of the dream material, which helps
explain why dreams are often illogical or bizarre.
Related, consider fMRI scans of patients suffering from schizophrenia
(e.g., Silbersweig et al., 1995). Neuroimaging reveals diminished activity in
the frontal lobes whenever these patients are experiencing hallucinations. One
interpretation is that the diminished activity reflects a decreased ability to distinguish internal events (thoughts) from external ones (voices) or to distinguish
imagined events from real ones (cf. Glisky, Polster, & Routhieaux, 1995).
How is all of this relevant to Capgras syndrome? With damage to the
frontal lobe, Capgras patients may be less able to keep track of what is real
and what is not, what is sensible and what is not. As a result, weird beliefs
can emerge unchecked, including delusions (about robots and the like) that
you or I would find totally bizarre.
What Do We Learn from Capgras Syndrome?
Other lines of evidence add to our understanding of Capgras syndrome
(e.g., Ellis & Lewis, 2001; Ramachandran & Blakeslee, 1998). Some of the
evidence comes from the psychology laboratory and confirms the suggestion that recognition of all stimuli (not just faces) involves two separate
mechanisms — one that hinges on factual knowledge, and one that’s more
“emotional” and tied to the warm sense of familiarity (see Chapter 7).
Note, then, that our understanding of Capgras syndrome depends on a
combination of evidence drawn from cognitive psychology and from cognitive neuroscience. We use both perspectives to test (and, ultimately, to confirm) the hypothesis we’ve offered. In addition, just as both perspectives can
illuminate Capgras syndrome, both can be illuminated by the syndrome. That
is, we can use Capgras syndrome (and other biological evidence) to illuminate broader issues about the nature of the brain and of the mind.
For example, Capgras syndrome suggests that the amygdala plays a crucial
role in supporting the feeling of familiarity. Other evidence suggests that the
amygdala also helps people remember the emotional events of their lives (e.g.,
Buchanan & Adolphs, 2004). Still other evidence indicates that the amygdala plays a role in decision making (e.g., Bechara, Damasio, & Damasio,
2003), especially for decisions that rest on emotional evaluations of one’s
options. Facts like these tell us a lot about the various functions that make
cognition possible and, more specifically, tell us that our theorizing needs to
include a broadly useful “emotional evaluator,” involved in many cognitive
processes. Moreover, Capgras syndrome tells us that this emotional evaluator works in a fashion separate from the evaluation of factual information,
and this observation gives us a way to think about occasions in which your
evaluation of the facts points toward one conclusion, while an emotional
evaluation points toward a different conclusion. These are valuable clues
as we try to understand the processes that support ordinary remembering
or decision making. (For more on the role of emotion in decision making,
see Chapter 12.)
30 •
C H A P T E R T WO The Neural Basis for Cognition
What does Capgras syndrome teach us about the brain itself? One lesson
involves the fact that many different parts of the brain are needed for even
the simplest achievement. In order to recognize your father, for example, one
part of your brain needs to store the factual memory of what he looks like.
Another part of the brain is responsible for analyzing the visual input you
receive when looking at a face. Yet another brain area has the job of comparing this now-analyzed input to the factual information provided from
memory, to determine whether there’s a match. Another site provides the
emotional evaluation of the input. A different site presumably assembles the
data from all these other sites — and registers the fact that the face being
inspected does match the factual recollection of your father’s face, and also
produces a warm sense of familiarity.
Usually, all these brain areas work together, allowing the recognition of
your father’s face to go smoothly forward. If they don’t work together — that
is, if coordination among these areas is disrupted — yet another area works
to make sure you offer reasonable hypotheses about this disconnect, and not
zany ones. (In other words, if your father looks less familiar to you on some
occasion, you’re likely to explain this by saying, “I guess he must have gotten
new glasses” rather than “I bet he’s been replaced by a robot.”)
Unmistakably, this apparently easy task — seeing your father and recognizing
who he is — requires multiple brain areas. The same is true of most tasks, and
in this way Capgras syndrome illustrates this crucial aspect of brain function.
TEST YOURSELF
1. W
hat are the symptoms of Capgras
syndrome, and why
do they suggest a
two-part explanation
for how you recognize
faces?
The Study of the Brain
In order to discuss Capgras syndrome, we needed to refer to different brain
areas and had to rely on several different research techniques. In this way, the
syndrome also illustrates another point — that this is a domain in which we
need some technical foundations before we can develop our theories. Let’s
start building those foundations.
The human brain weighs (on average) a bit more than 3 pounds (roughly
1.4 kg), with male brains weighing about 10% more than female brains
(Hartmann, Ramseier, Gudat, Mihatsch, & Polasek, 1994). The brain is roughly
the size of a small melon, yet this compact structure has been estimated to
contain 86 billion nerve cells (Azevedo et al., 2009). Each of these cells is connected to 10,000 or so others — for a total of roughly 860 trillion connections.
The brain also contains a huge number of glial cells, and we’ll have more to say
about all of these individual cells later on in the chapter. For now, though, how
should we begin our study of this densely packed, incredibly complex organ?
One place to start is with a simple fact we’ve already met: that different
parts of the brain perform different jobs. Scientists have known this fact
about the brain for many years, thanks to clinical evidence showing that
the symptoms produced by brain damage depend heavily on the location
of the damage. In 1848, for example, a horrible construction accident
caused Phineas Gage to suffer damage in the frontmost part of his brain
The Study of the Brain
•
31
FIGURE 2.4
A
PHINEAS GAGE
B
C
Phineas Gage was working as a construction foreman when some blasting powder misfired and launched a piece
of iron into his cheek and through the front part of his brain. Remarkably, Gage survived and continued to live a
fairly normal life, but his pattern of intellectual and emotional impairments provide valuable cues about the function of the brain’s frontal lobes. Panel A is a photo of Gage’s skull; the drawing in Panel B depicts the iron bar’s
path as it blasted through his head. Panel C is an actual photograph of Gage, and he’s holding the bar that went
through his brain!
(see Figure 2.4), and this damage led to severe personality and emotional
problems. In 1861, physician Paul Broca noted that damage in a different
location, on the left side of the brain, led to a disruption of language skills.
In 1911, Édouard Claparède (1911/1951) reported his observations with
patients who suffered from profound memory loss produced by damage in
still another part of the brain.
Clearly, therefore, we need to understand brain functioning with reference
to brain anatomy. Where was the damage that Gage suffered? Where was
the damage in Broca’s patients or Claparède’s? In this section, we fill in some
basics of brain anatomy.
Hindbrain, Midbrain, Forebrain
The human brain is divided into three main structures: the hindbrain,
the midbrain, and the forebrain. The hindbrain is located at the very top
of the spinal cord and includes structures crucial for controlling key life
32 •
C H A P T E R T WO The Neural Basis for Cognition
Corpus callosum
Midbrain
Pons
Medulla
Spinal cord
Cerebellum
GROSS ANATOMY OF A BRAIN SHOWING BRAIN STEM
The pons and medulla are part of the hindbrain. The medulla controls vital functions
such as breathing and heart rate. The pons (Latin for “bridge”) is the main connection between the cerebellum and the rest of the brain.
functions. It’s here, for example, that the rhythm of heartbeats and the
rhythm of breathing are regulated. The hindbrain also plays an essential
role in maintaining the body’s overall tone. Specifically, the hindbrain helps
maintain the body’s posture and balance; it also helps control the brain’s level
of alertness.
The largest area of the hindbrain is the cerebellum. For many years,
investigators believed this structure’s main role was in the coordination
of bodily movements and balance. Research indicates, however, that
the cerebellum plays various other roles and that damage to this organ
can cause problems in spatial reasoning, in discriminating sounds, and
in integrating the input received from various sensory systems (Bower &
Parsons, 2003).
The midbrain has several functions. It plays an important part in coordinating movements, including the precise movements of the eyes as they
explore the visual world. Also in the midbrain are circuits that relay auditory
information from the ears to the areas in the forebrain where this information is processed and interpreted. Still other structures in the midbrain help
to regulate the experience of pain.
The Study of the Brain
•
33
For our purposes, though, the most interesting brain region (and, in
humans, the largest region) is the forebrain. Drawings of the brain (like the
one shown in Figure 2.2) show little other than the forebrain, because this
structure surrounds (and so hides from view) the entire midbrain and most
of the hindbrain. Of course, only the outer surface of the forebrain — the
cortex — is visible in such pictures. In general, the word “cortex” (from
the Latin word for “tree bark”) refers to an organ’s outer surface, and many
organs each have their own cortex; what’s visible in the drawing, then, is the
cerebral cortex.
The cortex is just a thin covering on the outer surface of the forebrain; on average, it’s a mere 3 mm thick. Nonetheless, there’s a great
deal of cortical tissue; by some estimates, the cortex makes up 80% of the
human brain. This considerable volume is made possible by the fact
that the cerebral cortex, thin as it is, consists of a large sheet of tissue. If
stretched out flat, it would cover more than 300 square inches, or roughly
2,000 cm2. (For comparison, this is an area roughly 20% greater than
the area covered by an extra-large — 18 inch, or 46 cm — pizza.) But the
cortex isn’t stretched flat; instead, it’s crumpled up and jammed into the
limited space inside the skull. It’s this crumpling that produces the brain’s
most obvious visual feature — the wrinkles, or convolutions, that cover the
brain’s outer surface.
Some of the “valleys” between the wrinkles are actually deep grooves that
divide the brain into different sections. The deepest groove is the longitudinal
fissure, running from the front of the brain to the back, which separates the
left cerebral hemisphere from the right. Other fissures divide the cortex in
each hemisphere into four lobes (again, look back at Figure 2.2), and these
are named after the bones that cover them — bones that, as a group, make
up the skull. The frontal lobes form the front of the brain, right behind the
forehead. The central fissure divides the frontal lobes on each side of the
brain from the parietal lobes, the brain’s topmost part. The bottom edge of
the frontal lobes is marked by the lateral fissure, and below it are the temporal lobes. Finally, at the very back of the brain, connected to the parietal and
temporal lobes, are the occipital lobes.
Subcortical Structures
Hidden from view, underneath the cortex, are several subcortical structures.
One of these structures, the thalamus, acts as a relay station for nearly all
the sensory information going to the cortex. Directly underneath the thalamus is the hypothalamus, a structure that plays a crucial role in controlling
behaviors that serve specific biological needs — behaviors that include eating,
drinking, and sexual activity.
Surrounding the thalamus and hypothalamus is another set of structures
that form the limbic system. Included here is the amygdala, and close by
is the hippocampus, both located underneath the cortex in the temporal
lobe (plurals: amygdalae and hippocampi; see Figure 2.5). These structures
34 •
C H A P T E R T WO The Neural Basis for Cognition
FIGURE 2.5
T HE LIMBIC SYSTEM AND THE
HIPPOCAMPUS
Cingulate cortex
Fornix
Thalamus
Mamillary
body
Hypothalamus
Amygdala
Hippocampus
Color is used in this drawing to help you visualize the arrangement of these
brain structures. Imagine that the cortex is semitransparent, allowing you to
look into the brain to see the (subcortical) structures highlighted here. The
limbic system includes a number of subcortical structures that play a crucial
role in learning and memory and in emotional processing.
are essential for learning and memory, and the patient H.M., discussed in
Chapter 1, developed his profound amnesia after surgeons removed large
portions of these structures — strong confirmation of their role in the formation of new memories.
We mentioned earlier that the amygdala plays a key role in emotional
processing, and this role is reflected in many findings. For example, presentation of frightful faces causes high levels of activity in the amygdala
(Williams et al., 2006). Likewise, people ordinarily show more complete,
longer-lasting memories for emotional events, compared to similar but
emotionally flat events. This memory advantage for emotional events is
especially pronounced in people who showed greater activation in the
amygdala while they were witnessing the event in the first place. Conversely,
the memory advantage for emotional events is diminished (and may not be
observed at all) in people who (through sickness or injury) have suffered
damage to the amygdalae.
The Study of the Brain
•
35
Lateralization
TEST YOURSELF
2.What is the cerebral
cortex?
3.What are the four
major lobes of the
forebrain?
4.Identify some of the
functions of the hippocampus, the amygdala, and the corpus
callosum.
36 •
Virtually all parts of the brain come in pairs, and so there is a hippocampus
on the left side of the brain and another on the right, a left-side amygdala
and a right-side one. The same is true for the cerebral cortex itself: There
is a temporal cortex (i.e., a cortex of the temporal lobe) in the left hemisphere and another in the right, a left occipital cortex and a right one, and
so on. In all cases, cortical and subcortical, the left and right structures
in each pair have roughly the same shape and the same pattern of connections to other brain areas. Even so, there are differences in function
between the left-side and right-side structures, with each left-hemisphere
structure playing a somewhat different role from the corresponding righthemisphere structure.
Let’s remember, though, that the two halves of the brain work together —
the functioning of one side is closely integrated with that of the other side.
This integration is made possible by the commissures, thick bundles of
fibers that carry information back and forth between the two hemispheres.
The largest commissure is the corpus callosum, but several other structures
also make sure that the two brain halves work as partners in almost all
mental tasks.
In certain cases, though, there are medical reasons to sever the corpus
callosum and some of the other commissures. (For many years, this surgery was a last resort for extreme cases of epilepsy.) The person is then
said to be a “split-brain patient” — still having both brain halves, but with
communication between the halves severely limited. Research with these
patients has taught us a great deal about the specialized function of the
brain’s two hemispheres. It has provided evidence, for example, that many
aspects of language processing are lodged in the left hemisphere, while
the right hemisphere seems crucial for a number of tasks involving spatial
judgment (see Figure 2.6).
However, it’s important not to overstate the contrast between the two
brain halves, and it’s misleading to claim (as some people do) that we need
to silence our “left-brain thinking” in order to be more creative, or that
intuitions grow out of “right-brain thinking.” These claims do begin with a
kernel of truth, because some elements of creativity depend on specialized
processing in the right hemisphere (see, e.g., Kounios & Beeman, 2015).
Even so, whether we’re examining creativity or any other capacity, the two
halves of the brain have to work together, with each hemisphere making its
own contribution to the overall performance. Therefore, “shutting down” or
“silencing” one hemisphere, even if that were biologically possible, wouldn’t
allow you new achievements, because the many complex, sophisticated skills
we each display (including creativity, intuition, and more) depend on the
whole brain. In other words, our hemispheres are not cerebral competitors,
each trying to impose its style of thinking on the other. Instead, the hemispheres pool their specialized capacities to produce a seamlessly integrated,
single mental self.
C H A P T E R T WO The Neural Basis for Cognition
FIGURE 2.6
STUDYING SPLIT-BRAIN PATIENTS
“A fork”
When a split-brain patient is asked what he
sees, the left hemisphere sees the fork on the
right side of the screen and can verbalize that.
A
The right hemisphere sees the spoon on the
screen’s left side, but it cannot verbalize that.
However, if the patient reaches with his left hand
to pick up the object, he does select the spoon.
B
In this experiment, the patient is shown two pictures, one of a spoon and one of a fork (Panel A). If asked what
he sees, his verbal response is controlled by the left hemisphere, which has seen only the fork (because it’s in the
right visual field. However, if asked to pick up the object shown in the picture, the patient — reaching with his left
hand — picks up the spoon (Panel B). That happens because the left hand is controlled by the right hemisphere,
and this hemisphere receives visual information from the left-hand side of the visual world.
Sources of Evidence about the Brain
How can we learn about these various structures — and many others that we
haven’t named? Cognitive neuroscience relies on many types of evidence to
study the brain and nervous system. Let’s look at some of the options.
Data from Neuropsychology
We’ve already encountered one form of evidence — the study of individuals
who have suffered brain damage through accident, disease, or birth defect.
The study of these cases generally falls within the domain of neuropsychology: the study of the brain’s structures and how they relate to brain function.
Within neuropsychology, the specialty of clinical neuropsychology seeks
(among other goals) to understand the functioning of intact, undamaged
brains by means of careful scrutiny of cases involving brain damage.
Data drawn from clinical neuropsychology will be important throughout
this text. For now, though, we’ll emphasize that the symptoms resulting from
Sources of Evidence about the Brain
•
37
brain damage depend on the site of the damage. A lesion (a specific area of
damage) in the hippocampus produces memory problems but not language
disorders; a lesion in the occipital cortex produces problems in vision but
spares the other sensory modalities. Likewise, the consequences of brain
lesions depend on which hemisphere is damaged. Damage to the left side of
the frontal lobe, for example, is likely to produce a disruption of language use;
damage to the right side of the frontal lobe generally doesn’t have this effect.
In obvious ways, then, these patterns confirm the claim that different brain
areas perform different functions. In addition, these patterns provide a rich
source of data that help us develop and test hypotheses about those functions.
Data from Neuroimaging
Further insights come from neuroimaging techniques. There are several types
of neuroimaging, but they all produce precise, three-dimensional pictures of
a living brain. Some neuroimaging procedures provide structural imaging,
generating a detailed portrait of the shapes, sizes, and positions of the brain’s
components. Other procedures provide functional imaging, which tells us
about activity levels throughout the brain.
For many years, computerized axial tomography (CT scans) was the primary
tool for structural imaging, and positron emission tomography (PET scans)
was used to study the brain’s activity. CT scans rely on X-rays and so — in
essence — provide a three-dimensional X-ray picture of the brain. PET scans, in
contrast, start by introducing a tracer substance such as glucose into the patient’s
body; the molecules of this tracer have been tagged with a low dose of radioactivity, and the scan keeps track of this radioactivity, allowing us to tell which tissues
are using more of the glucose (the body’s main fuel) and which ones are using less.
For each type of scan, the primary data (X-rays or radioactive emissions)
are collected by a bank of detectors placed around the person’s head. A computer then compares the signals received by each of the detectors and uses this
information to construct a three-dimensional map of the brain — a map of
structures from a CT scan, and a map showing activity levels from a PET scan.
More recent studies have turned to two newer techniques, introduced earlier
in the chapter. Magnetic resonance imaging (MRI scans) relies on the magnetic
properties of the atoms that make up the brain tissue, and it yields fabulously
detailed pictures of the brain. MRI scans provide structural images, but a
closely related technique, functional magnetic resonance imaging (fMRI scans),
provides functional imaging. The fMRI scans measure the oxygen content in
blood flowing through each region of the brain; this turns out to be an accurate
index of the level of neural activity in that region. In this way, fMRI scans offer
an incredibly precise picture of the brain’s moment-by-moment activities.
The results of structural imaging (CT or MRI scans) are relatively stable,
changing only if the person’s brain structure changes (because of an injury,
perhaps, or the growth of a tumor). The results of PET or fMRI scans, in contrast, are highly variable, because the results depend on what task the person
is performing. We can therefore use these latter techniques to explore brain
function — using fMRI scans, for example, to determine which brain sites are
38 •
C H A P T E R T WO The Neural Basis for Cognition
PET SCANS
PET scans measure how much
glucose (the brain’s fuel) is
being used at specific locations
within the brain; this provides
a measurement of each location’s activity level at a certain
moment in time. In the figure,
the brain is viewed from above,
with the front of the head at the
top and the back of the head at
the bottom. The various colors
indicate relative activity levels
(an actual brain is uniformly
colored), using the palette
shown on the right side of the
figure. Dark blue indicates a low
level of activity; red indicates
a high level. And as the figure
shows, visual processing involves increased activity in the
occipital lobe.
especially activated when someone is making a moral judgment or trying to
solve a logic problem. In this way, the neuroimaging data can provide crucial
information about how these activities are made possible by specific patterns
of functioning within the brain.
Data from Electrical Recording
Neuroscientists have another technique in their toolkit: electrical recording of
the brain’s activity. To explain this point, though, we need to say a bit about
how the brain functions. As mentioned earlier, the brain contains billions of
nerve cells — called “neurons” — and it is the neurons that do the brain’s main
work. (We’ll say more about these cells later in the chapter.) Neurons vary in
their functioning, but for the most part they communicate with one another
via chemical signals called “neurotransmitters.” Once a neuron is “activated,”
it releases the transmitter, and this chemical can then activate (or, in some cases,
de-activate) other, adjacent neurons. The adjacent neurons “receive” this
chemical signal and, in turn, send their own signal onward to other neurons.
Let’s be clear, though, that the process we just described is communication
between neurons: One neuron releases the transmitter substance, and this activates (or de-activates) another neuron. But there’s also communication within
each neuron. The reason, basically, is that neurons have an “input” end and an
“output” end. The “input” end is the portion of the neuron that’s most sensitive
to neurotransmitters; this is where the signal from other neurons is received.
The “output” end is the portion that releases neurotransmitters, sending the
Sources of Evidence about the Brain
•
39
A
B
C
MAGNETIC RESONANCE IMAGING
Magnetic resonance imaging produces magnificently detailed pictures of the brain. Panel A shows an “axial
view” — a “slice” of the brain viewed from the top of the head (the front of the head is at the top of the image).
Clearly visible is the longitudinal fissure, which divides the left cerebral hemisphere from the right. Panel B, a “coronal view,” shows a slice of the brain viewed from the front. Again, the separation of the two hemispheres is clearly
visible, as are some of the commissures linking the two brain halves. Panel C, a “sagittal view,” shows a slice of
the brain viewed from the side. Here, many of the structures in the limbic system (see Figure 2.5) are easily seen.
signal on to other neurons. These two ends can sometimes be far apart. (For
example, some neurons in the body run from the base of the spine down to the
toes; for these cells, the input and output ends might be a full meter apart.) The
question, then, is how neurons get a signal from one end of the cell to the other.
The answer involves an electrical pulse, made possible by a flow of charged
atoms (ions) in and out of the neuron (again, we’ll say more about this process later in the chapter). The amount of electrical current involved in this ion
flow is tiny; but many millions of neurons are active at the same time, and
the current generated by all of them together is strong enough to be detected
by sensitive electrodes placed on the surface of the scalp. This is the basis
for electroencephalography — a recording of voltage changes occurring at the
scalp that reflect activity in the brain underneath. This procedure generates
an electroencephalogram (EEG) — a recording of the brain’s electrical activity.
Often, EEGs are used to study broad rhythms in the brain’s activity. For
example, an alpha rhythm (with the activity level rising and falling seven to ten
times per second) can usually be detected in the brain of someone who is awake
but calm and relaxed; a delta rhythm (with the activity rising and falling roughly
one to four times per second) is observed when someone is deeply asleep. A
much faster gamma rhythm (between 30 and 80 cycles per second) has received
a lot of research attention, with a suggestion that this rhythm plays a key role
in creating conscious awareness (e.g., Crick & Koch, 1990; Dehaene, 2014).
Sometimes, though, we want to know about the electrical activity in the
brain over a shorter period — for example, when the brain is responding to
a specific input or a particular stimulus. In this case, we measure changes in
the EEG in the brief periods just before, during, and after the event. These
changes are referred to as event-related potentials (see Figure 2.7).
40 •
C H A P T E R T WO The Neural Basis for Cognition
FIGURE 2.7
RECORDING THE BRAIN’S ELECTRICAL ACTIVITY
Alert wakefulness
Beta waves
Just before sleep
Alpha waves
Stage 1
Theta waves
Stage 2
Sleep spindle
Stage 3
K complex
Delta waves
Stage 4
A
B
EEG
–
Amplifier
+
Prestimulus
period
20 µV
Stimulus
onset
Repeat and combine
for 100 trials
ERP
–
+
Sound
generator
2 µV
Stimulus
onset
700 ms
C
To record the brain’s electrical signals, researchers generally use a cap that has electrodes attached to it. The procedure is easy and entirely safe — it can even be used to measure brain signals in a baby (Panel A). In some procedures, researchers measure recurrent rhythms in the brain’s activity, including rhythms that distinguish the stages
of sleep (Panel B). In other procedures, they measure brain activity produced in response to a single event — such
as the presentation of a well-defined stimulus (Panel C).
The Power of Combining Techniques
Each of the research tools we’ve described has strengths and weaknesses. CT
scans and MRI data tell us about the shape and size of brain structures, but
they tell nothing about the activity levels within these structures. PET scans
and fMRI studies do tell us about brain activity, and they can locate the activity rather precisely (within a millimeter or two). But these techniques are less
precise about when the activity took place. For example, fMRI data summarize the brain’s activity over a period of several seconds and cannot indicate
when exactly, within this time window, the activity took place. EEG data give
more precise information about timing but are much weaker in indicating
where the activity took place.
Researchers deal with these limitations by means of a strategy commonly
used in science: We seek data from multiple sources, so that the strengths of
one technique can make up for the shortcomings of another. As a result, some
studies combine EEG recordings with fMRI scans, with the EEGs telling us
when certain events took place in the brain, and the scans telling us where
the activity took place. Likewise, some studies combine fMRI scans with
CT data, so that findings about brain activation can be linked to a detailed
portrait of the person’s brain anatomy.
Researchers also face another complication: the fact that many of the
techniques described so far provide correlational data. To understand
the concern here, let’s look at an example. A brain area called the fusiform face area (FFA) is especially active whenever a face is being perceived
(see Figure 2.8) — and so there is a correlation between a mental activity
(perceiving a face) and a pattern of brain activity. Does this mean the FFA
is needed for face perception? A different possibility is that the FFA activation may just be a by-product of face perception and doesn’t play a crucial
role. As an analogy, think about the fact that a car’s speedometer becomes
“more activated” (i.e., shows a higher value) whenever the car goes faster.
That doesn’t mean that the speedometer causes the speed or is necessary
for the speed. The car would go just as fast and would, for many purposes,
perform just as well if the speedometer were removed. The speedometer’s
state, in other words, is correlated with the car’s speed but in no sense
causes (or promotes, or is needed for) the car’s speed.
In the same way, neuroimaging data can tell us that a brain area’s activity
is correlated with a particular function, but we need other data to determine
whether the brain site plays a role in causing (or supporting, or allowing)
that function. In many cases, those other data come from the study of brain
lesions. If damage to a brain site disrupts a particular function, it’s an indication that the site does play some role in supporting that function. (And, in
fact, the FFA does play an important role in face recognition.)
Also helpful here is a technique called transcranial magnetic stimulation
(TMS). This technique creates a series of strong magnetic pulses at a specific
location on the scalp, and these pulses activate the neurons directly underneath this scalp area (Helmuth, 2001). TMS can thus be used as a means of
42 •
C H A P T E R T WO The Neural Basis for Cognition
FIGURE 2.8
BRAIN ACTIVITY AND AWARENESS
FFA
A
Percentage of fMRI signal
Face
Face
1.0
FFA
0.8
House
0.8
PPA
0.6
0.6
PPA
0.4
0.4
FFA
0.2
0.2
0.0
C
B
House
1.0
PPA
PPA
–8
–4
0
4
8
12
0.0
–8
–4
0
4
8
12
Time from reported perceptual switch (s)
Panel A shows an fMRI scan of a subject looking at faces. Activation levels are high in the fusiform face area (FFA),
an area that is apparently more responsive to faces than to other visual stimuli. Panel B shows a scan of the same
subject looking at pictures of places; now, activity levels are high in the parahippocampal place area (PPA). Panel
C compares the activity in these two areas when the subject has a picture of a face in front of one eye and a picture of a house in front of the other eye. When the viewer’s perception shifts from the house to the face, activation
increases in the FFA. When the viewer’s perception shifts from the face to the house, PPA activation increases. In
this way, the activation level reflects what the subject is aware of, and not just the pattern of incoming stimulation.
( after tong , nakayama , vaughan , & kanwisher , 1998)
asking what happens if we stimulate certain neurons. In addition, because
this stimulation disrupts the ordinary function of these neurons, it produces
a (temporary) lesion — allowing us to identify, in essence, what functions are
compromised when a particular bit of brain tissue is briefly “turned off.” In
these ways, the results of a TMS procedure can provide crucial information
about the functional role of that brain area.
Sources of Evidence about the Brain
•
43
Localization of Function
TEST YOURSELF
5.What is the difference
between structural imaging of the brain and
functional imaging?
What techniques are
used for each?
6.What do we gain from
combining different
methods in studying
the brain?
7.What is meant by the
phrase “localization
of function”?
Drawing on the techniques we have described, neuroscientists have learned
a great deal about the function of specific brain structures. This type of
research effort is referred to as the localization of function, an effort (to put
it crudely) aimed at figuring out what’s happening where within the brain.
Localization data are useful in many ways. For example, think back to
the discussion of Capgras syndrome earlier in this chapter. Brain scans told
us that people with this syndrome have damaged amygdalae, but how is this
damage related to the symptoms of the syndrome? More broadly, what problems does a damaged amygdala create? To tackle these questions, we rely on
localization of function — in particular, on data showing that the amygdala
is involved in many tasks involving emotional appraisal. This combination
of points helped us to build (and test) our claims about this syndrome and,
in general, claims about the role of emotion within the ordinary experience
of “familiarity.”
As a different illustration, consider the experience of calling up a “mental
picture” before the “mind’s eye.” We’ll have more to say about this experience
in Chapter 11, but we can already ask: How much does this experience have
in common with ordinary seeing — that is, the processes that unfold when
we place a real picture before someone’s eyes? As it turns out, localization
data reveal enormous overlap between the brain structures needed for these
two activities (visualizing and actual vision), telling us immediately that these
activities do have a great deal in common (see Figure 2.9). So, again, we build
on localization — this time to identify how exactly two mental activities are
related to each other.
The Cerebral Cortex
As we’ve noted, the largest portion of the human brain is the cerebral cortex —
the thin layer of tissue covering the cerebrum. This is the region in which an
enormous amount of information processing takes place, and so, for many
topics, it is the brain region of greatest interest for cognitive psychologists.
The cortex includes many distinct regions, each with its own function,
but these regions are traditionally divided into three categories. Motor areas
contain brain tissue crucial for organizing and controlling bodily movements. Sensory areas contain tissue essential for organizing and analyzing
the information received from the senses. Association areas support many
functions, including the essential (but not well-defined) human activity we
call “thinking.”
Motor Areas
Certain regions of the cerebral cortex serve as the “departure points” for
signals leaving the cortex and controlling muscle movement. Other areas
are the “arrival points” for information coming from the eyes, ears, and
44 •
C H A P T E R T WO The Neural Basis for Cognition
FIGURE 2.9
A PORTRAIT OF THE BRAIN AT WORK
Activity while
looking at
pictures
Activity while
visualizing “mental
pictures”
These fMRI images show different “slices” through the living brain, revealing levels of activity in different brain sites. Regions that are more active are
shown in yellow, orange, and red; lower activity levels are indicated in blue.
The first column shows brain activity while a person is making judgments
about simple pictures. The second column shows brain activity while the person is making the same sorts of judgments about “mental pictures,” visualized
before the “mind’s eye.”
The Cerebral Cortex
•
45
other sense organs. In both cases, these areas are called “primary projection
areas,” with the departure points known as the primary motor projection
areas and the arrival points contained in regions known as the primary sensory
projection areas.
Evidence for the motor projection area comes from studies in which
investigators apply mild electrical current to this area in anesthetized animals.
This stimulation often produces specific movements, so that current applied
to one site causes a movement of the left front leg, while current applied to
a different site causes the ears to prick up. These movements show a pattern
of contralateral control, with stimulation to the left hemisphere leading to
movements on the right side of the body, and vice versa.
Why are these areas called “projection areas”? The term is borrowed from
mathematics and from the discipline of map making, because these areas
seem to form “maps” of the external world, with particular positions on the
cortex corresponding to particular parts of the body or particular locations
in space. In the human brain, the map that constitutes the motor projection
area is located on a strip of tissue toward the rear of the frontal lobe, and the
pattern of mapping is illustrated in Figure 2.10. In this illustration, a drawing
of a person has been overlaid on a depiction of the brain, with each part of
the little person positioned on top of the brain area that controls its movement. The figure shows that areas of the body that we can move with great
precision (e.g., fingers and lips) have a lot of cortical area devoted to them;
areas of the body over which we have less control (e.g., the shoulder and the
back) receive less cortical coverage.
Sensory Areas
Information arriving from the skin senses (your sense of touch or your sense
of temperature) is projected to a region in the parietal lobe, just behind the
motor projection area. This is labeled the “somatosensory” area in Figure 2.10.
If a patient’s brain is stimulated in this region (with electrical current or touch),
the patient will typically report a tingling sensation in a specific part of the
body. Figure 2.10 also shows the region (in the temporal lobes) that functions
as the primary projection area for hearing (the “auditory” area). If the brain is
directly stimulated here, the patient will hear clicks, buzzes, and hums. An area
in the occipital lobes is the primary projection area for vision; stimulation here
causes the patient to see flashes of light or visual patterns.
The sensory projection areas differ from each other in important ways,
but they also have features in common — and they’re features that parallel
the attributes of the motor projection area. First, each of these areas provides
a “map” of the sensory environment. In the somatosensory area, each part
of the body’s surface is represented by its own region on the cortex; areas of
the body that are near to each other are typically represented by similarly
nearby areas in the brain. In the visual area, each region of visual space has
its own cortical representation, and adjacent areas of visual space are usually
represented by adjacent brain sites. In the auditory projection area, different
46 •
C H A P T E R T WO The Neural Basis for Cognition
Tongue
Primary auditory
cortex
Th
u
Ey mb
e
N
Fa os
ce e
Hand
Fingers
Genitals
Lips
Gums
Teeth
Jaw
Hip
Trunk
Neck
Head
Shoulder
Arm
Arm
Shoulder
Neck
Trunk
Hip
Fingers
Hand
bk
umec
Th N rowe
B Ey
e
os ce
Fa
N
Knee Knee
Leg
Leg
Ankle Foot
Toes Toes
FIGURE 2.10 THE PRIMARY
PROJECTION AREAS
Primary motor
projection area
Lips
Gums
Teeth
Jaw
Primary
somatosensory
projection area
Tongue
Primary visual
cortex
The primary motor projection area is
located at the rearmost edge of the
frontal lobe, and each region within
this projection area controls the
motion of a specific body part, as
illustrated on the top left. The primary
somatosensory projection area, receiving information from the skin, is
at the forward edge of the parietal
lobe; each region within this area
receives input from a specific body
part. The primary projection areas
for vision and hearing are located
in the occipital and temporal lobes,
respectively. These two areas are
also organized systematically. For
example, in the visual projection
area, adjacent areas of the brain
receive visual inputs that come from
adjacent areas in visual space.
frequencies of sound have their own cortical sites, and adjacent brain sites
are responsive to adjacent frequencies.
Second, in each of these sensory maps, the assignment of cortical space is
governed by function, not by anatomical proportions. In the parietal lobes,
parts of the body that aren’t very discriminating with regard to touch — even
if they’re physically large — get relatively little cortical area. Other, more
sensitive areas of the body (the lips, tongue, and fingers) get much more
space. In the occipital lobes, more cortical surface is devoted to the fovea,
the part of the eyeball that is most sensitive to detail. (For more on the
fovea, see Chapter 3.) And in the auditory areas, some frequencies of sound
get more cerebral coverage than others. It’s surely no coincidence that these
“advantaged” frequencies are those essential for the perception of speech.
Finally, we also find evidence here of contralateral connections. The
somatosensory area in the left hemisphere, for example, receives its main
input from the right side of the body; the corresponding area in the right
hemisphere receives its input from the left side of the body. Likewise for
the visual projection areas, although here the projection is not contralateral
with regard to body parts. Instead, it’s contralateral with regard to physical
space. Specifically, the visual projection area in the right hemisphere receives
information from both the left eye and the right, but the information it
receives corresponds to the left half of visual space (i.e., all of the things
The Cerebral Cortex
•
47
visible to your left when you’re looking straight ahead). The reverse is true
for the visual area in the left hemisphere. It receives information from both
eyes, but from only the right half of visual space. The pattern of contralateral
organization is also evident — although not as clear-cut — for the auditory
cortex, with roughly 60% of the nerve fibers from each ear sending their
information to the opposite side of the brain.
Association Areas
THE SENSORY
HOMUNCULUS
An artist’s rendition of what
a man would look like if his
appearance were proportional
to the area allotted by the
somatosensory cortex to his
various body parts.
TEST YOURSELF
8.What is a projection
area in the brain?
What’s the role of the
motor projection area?
The sensory projection
area?
9.What does it mean
to say that the brain
relies on “contralateral”
connections?
48 •
The areas described so far, both motor and sensory, make up only a small
part of the human cerebral cortex — roughly 25%. The remaining cortical areas are traditionally referred to as the association cortex. This terminology is falling out of use, however, partly because this large volume of
brain tissue can be subdivided further on both functional and anatomical
grounds. These subdivisions are perhaps best revealed by the diversity of
symptoms that result if the cortex is damaged in one or another specific
location. For example, some lesions in the frontal lobe produce apraxias,
disturbances in the initiation or organization of voluntary action. Other
lesions (generally in the occipital cortex, or in the rearmost part of the
parietal lobe) lead to agnosias, disruptions in the ability to identify familiar
objects. Agnosias usually affect one modality only — so a patient with visual
agnosia, for example, can recognize a fork by touching it but not by looking
at it. A patient with auditory agnosia, by contrast, might be unable to identify familiar voices but might still recognize the face of the person speaking.
Still other lesions (usually in the parietal lobe) produce neglect syndrome, in
which the individual seems to ignore half of the visual world. A patient afflicted
with this syndrome will shave only half of his face and eat food from only half
of his plate. If asked to read the word “parties,” he will read “ties,” and so on.
Damage in other areas causes still other symptoms. We mentioned earlier
that lesions in areas near the lateral fissure (the deep groove that separates the
frontal and temporal lobes) can result in disruption to language capacities, a
problem referred to as aphasia.
Finally, damage to the frontmost part of the frontal lobe, the prefrontal area, causes problems in planning and implementing strategies. In some
cases, patients with damage here show problems in inhibiting their own
behaviors, relying on habit even in situations for which habit is inappropriate.
Frontal lobe damage can also (as we mentioned in our discussion of Capgras
syndrome) lead to a variety of confusions, such as whether a remembered
episode actually happened or was simply imagined.
We’ll discuss more about these diagnostic categories — aphasia, agnosia,
neglect, and more — in upcoming chapters, where we’ll consider these disorders in the context of other things that are known about object recognition,
attention, and so on. Our point for the moment, though, is simple: These
clinical patterns make it clear that the so-called association cortex contains
many subregions, each specialized for a particular function, but with all of
the subregions working together in virtually all aspects of our daily lives.
C H A P T E R T WO The Neural Basis for Cognition
Brain Cells
Our brief tour so far has described some of the large-scale structures in the
brain. For many purposes, though, we need to zoom in for a closer look, in
order to see how the brain’s functions are actually carried out.
Neurons and Glia
We’ve already mentioned that the human brain contains many billions of
neurons and a comparable number of glia. The glia perform many functions.
They help to guide the development of the nervous system in the fetus and
young infant; they support repairs if the nervous system is damaged; they
also control the flow of nutrients to the neurons. Specialized glial cells also
provide a layer of electrical insulation surrounding parts of some neurons;
this insulation dramatically increases the speed with which neurons can send
their signals. (We’ll return to this point in a moment.) Finally, some research
suggests the glia may also constitute their own signaling system within
the brain, separate from the information flow provided by the neurons
(e.g., Bullock et al, 2005; Gallo & Chitajullu, 2001).
There is no question, though, that the main flow of information through
the brain — from the sense organs inward, from one part of the brain to
the others, and then from the brain outward — is made possible by the
neurons. Neurons come in many shapes and sizes (see Figure 2.11), but
in general, neurons have three major parts. The cell body is the portion of
the cell that contains the neuron’s nucleus and all the elements needed for
the normal metabolic activities of the cell. The dendrites are usually the
FIGURE 2.11
A
NEURONS
B
C
Panel A shows neurons from the spinal cord (stained in red); Panel B shows neurons from the cerebellum; Panel C
shows neurons from the cerebral cortex.
Brain Cells
•
49
“input” side of the neuron, receiving signals from many other neurons. In
most neurons, the dendrites are heavily branched, like a thick and tangled
bush. The axon is the “output” side of the neuron; it sends neural impulses to other neurons (see Figure 2.12). Axons can vary enormously in
length — the giraffe, for example, has neurons with axons that run the full
length of its neck.
FIGURE 2.12
REGIONS OF THE NEURON
Dendrites
Nucleus
Cell body
Myelin sheath
Axon
Neural
impulse
Nodes of Ranvier
Axon terminals
Most neurons have three identifiable regions. The dendrites are the part of
the neuron that usually detects incoming signals. The cell body contains the
metabolic machinery that sustains the cell. The axon is the part of the neuron
that transmits a signal to another location. When the cell fires, neurotransmitters are released from the terminal endings at the tip of the axon. The myelin
sheath is created by glial cells that wrap around the axons of many neurons.
The gaps in between the myelin cells are called the nodes of Ranvier.
50 •
C H A P T E R T WO The Neural Basis for Cognition
COGNITION
outside the lab
Alcohol
Of the many drugs that influence the brain, one
factors also matter. For example, blackouts are
is readily available and often consumed: alcohol.
more common if you become drunk rapidly — as
Alcohol influences the entire brain, and even at
when you drink on an empty stomach, or when
low levels of intoxication we can detect alcohol’s
you gulp alcohol rather than sipping it.
effects — for example, with measures of motor
skills or response time.
Let’s combine these points about blackouts,
though, with our earlier observation about alco-
Alcohol’s effects are more visible, though, in
hol’s uneven effects. It’s possible for someone to
some functions than in others, and so someone
be quite drunk, and therefore suffer an alcoholic
who’s quite intoxicated can perform many activi-
blackout, even if the person seemed alert and
ties at a fairly normal level. However, alcohol has
coherent during the drunken episode. To see
a strong impact on activities that depend on the
some of the serious problems this can cause,
brain’s prefrontal cortex. This is the brain region
consider a pattern that often emerges in cases
that’s essential for the mind’s executive function —
involving allegations of sexual assault. Victims of
the system that allows you to control your thoughts
assault sometimes report that they have little or
and behaviors. (We’ll say more about executive
no memory of the sexual encounter; they there-
function in upcoming chapters.) As a result, alco-
fore assume they were barely conscious during
hol undercuts your ability to resist temptation or
the event and surely incapable of giving consent.
to overcome habit. Impairments in executive func-
But is this assumption correct?
tion also erode your ability to make thoughtful
decisions and draw sensible conclusions.
The answer is complex. If someone was drunk
enough to end up with a blackout, then that per-
In addition, alcohol can produce impair-
son was probably impaired to a degree that would
ments in memory, including “alcoholic blackouts.”
interfere with decision making — and so the per-
So-called fragmentary blackouts, in which the
son could not have given legitimate, meaningful
person remembers some bits of an experience
consent. But, even so, the person might have been
but not others, are actually quite common. In one
functioning in a way that seemed mostly normal
study, college students were asked: “Have you
(able to converse, to move around) and may even
ever awoken after a night of drinking not able to
have expressed consent in words or actions.
remember things that you did or places where you
In this situation, then, the complainant is cor-
went?” More than half of the students indicated that,
rect in saying that he or she couldn’t have given
yes, this had happened to them at some point;
(and therefore didn’t give) meaningful consent,
40% reported they’d had a blackout within the
but the accused person can legitimately say that
previous year.
he or she perceived that there was consent. We
How drunk do you have to be in order to
can debate how best to judge these situations, but
experience a blackout? Many authorities point to
surely the best path forward is to avoid this sort of
a blood alcohol level of 0.25 (roughly nine or ten
circumstance — by drinking only in safe settings or
drinks for someone of average weight), but other
by keeping a strict limit on your drinking.
Brain Cells
•
51
The Synapse
We’ve mentioned that communication from one neuron to the next is generally made possible by a chemical signal: When a neuron has been sufficiently
stimulated, it releases a minute quantity of a neurotransmitter. The molecules
of this substance drift across the tiny gap between neurons and latch on
to the dendrites of the adjacent cell. If the dendrites receive enough of this
substance, the next neuron will “fire,” and so the signal will be sent along to
other neurons.
Notice, then, that neurons usually don’t touch each other directly. Instead,
at the end of the axon there is a gap separating each neuron from the next.
This entire site — the end of the axon, plus the gap, plus the receiving membrane of the next neuron — is called a synapse. The space between the neurons
is the synaptic gap. The bit of the neuron that releases the transmitter into this
gap is the presynaptic membrane, and the bit of the neuron on the other side
of the gap, affected by the transmitters, is the postsynaptic membrane.
When the neurotransmitters arrive at the postsynaptic membrane, they
cause changes in this membrane that enable certain ions to flow into and out
of the postsynaptic cell (see Figure 2.13). If these ionic flows are relatively
small, then the postsynaptic cell quickly recovers and the ions are transported
back to where they were initially. But if the ionic flows are large enough, they
trigger a response in the postsynaptic cell. In formal terms, if the incoming
signal reaches the postsynaptic cell’s threshold, then the cell fires. That is, it
produces an action potential — a signal that moves down its axon, which in
turn causes the release of neurotransmitters at the next synapse, potentially
causing the next cell to fire.
In some neurons, the action potential moves down the axon at a relatively
slow speed. For other neurons, specialized glial cells are wrapped around the
axon, creating a layer of insulation called the myelin sheath (see Figure 2.12).
Because of the myelin, ions can flow in or out of the axon only at the gaps
between the myelin cells. As a result, the signal traveling down the axon has
to “jump” from gap to gap, and this greatly increases the speed at which the
signal is transmitted. For neurons without myelin, the signal travels at speeds
below 10 m/s; for “myelinated” neurons, the speed can be ten times faster.
Overall, let’s emphasize four points about this sequence of events. First,
let’s note once again that neurons depend on two different forms of information flow. Communication from one neuron to the next is (for most neurons)
mediated by a chemical signal. In contrast, communication from one end of
the neuron to the other (usually from the dendrites down the length of the
axon) is made possible by an electrical signal, created by the flow of ions in
and out of the cell.
Second, the postsynaptic neuron’s initial response can vary in size; the
incoming signal can cause a small ionic flow or a large one. Crucially, though,
once these inputs reach the postsynaptic neuron’s firing threshold, there’s no
variability in the response — either a signal is sent down the axon or it is not.
If the signal is sent, it is always of the same magnitude, a fact referred to as
52 •
C H A P T E R T WO The Neural Basis for Cognition
FIGURE 2.13
SCHEMATIC VIEW OF SYNAPTIC TRANSMISSION
Neuron 1
A
Action
potential
Neuron 2
Axon
Synaptic
vesicle
Synaptic
vesicle
Synaptic
gap
Na+
Ion channel
C
Receptor site
Neurotransmitter
Postsynaptic
membrane
Dendrite
B
Presynaptic
membrane
Na+
D
(Panel A) Neuron 1 transmits a message across the synaptic gap to Neuron 2. The neurotransmitters are initially
stored in structures called “synaptic vesicles” (Panel B). When a signal travels down the axon, the vesicles are
stimulated and some of them burst (Panel C), ejecting neurotransmitter molecules into the synaptic gap and
toward the postsynaptic membrane (Panel D). Neurotransmitter molecules settle on receptor sites, ion channels
open, and sodium (Na+) floods in.
Brain Cells
•
53
TEST YOURSELF
10.What are glia? What
are dendrites? What
is an axon? What is a
synapse?
11.What does it mean
to say that neurons
rely on two different
forms of information
flow, one chemical
and one electrical?
the all-or-none law. Just as pounding on a car horn won’t make the horn
any louder, a stronger stimulus won’t produce a stronger action potential. A
neuron either fires or it doesn’t; there’s no in-between.
This does not mean, however, that neurons always send exactly the same
information. A neuron can fire many times per second or only occasionally. A
neuron can fire just once and then stop, or it can keep firing for an extended span.
But, even so, each individual response by the neuron is always the same size.
Third, we should also note that the brain relies on many different neurotransmitters. By some counts, a hundred transmitters have been catalogued
so far, and this diversity enables the brain to send a variety of different messages. Some transmitters have the effect of stimulating subsequent neurons;
some do the opposite and inhibit other neurons. Some transmitters play an
essential role in learning and memory; others play a key role in regulating the
level of arousal in the brain; still others influence motivation and emotion.
Fourth, let’s be clear about the central role of the synapse. The synaptic gap is actually quite small — roughly 20 to 30 nanometers across. (For
contrast’s sake, the diameter of a human hair is roughly 80,000 nano­
meters.) Even so, transmission across this gap slows down the neuronal
signal, but this is a tiny price to pay for the advantages created by this mode
of signaling: Each neuron receives information from (i.e., has synapses with)
many other neurons, and this allows the “receiving” neuron to integrate
information from many sources. This pattern of many neurons feeding into
one also makes it possible for a neuron to “compare” signals and to adjust its
response to one input according to the signal arriving from a different input.
In addition, communication at the synapse is adjustable. This means that
the strength of a synaptic connection can be altered by experience, and this
adjustment is crucial for the process of learning — the storage of new knowledge and new skills within the nervous system.
Coding
This discussion of individual neurons leads to a further question: How do
these microscopic nerve cells manage to represent a specific idea or a specific
content? Let’s say that right now you’re thinking about your favorite song.
How is this information represented by neurons? The issue here is referred
to as coding, and there are many options for what the neurons’ “code” might
be (Gallistel, 2017). As one option, we might imagine that a specific group of
neurons somehow represents “favorite song,” so that whenever you’re thinking about the song, it’s precisely these neurons that are activated. Or, as a different option, the song might be represented by a broad pattern of neuronal
activity. If so, “favorite song” might be represented in the brain by something
like “Neuron X firing strongly while Neuron Y is firing weakly and Neuron
Z is not firing at all” (and so on for thousands of other neurons). Note that
within this scheme the same neurons might be involved in the representation
of other sounds, but with different patterns. So — to continue our example —
Neuron X might also be involved in the representation of the sound of a car
54 •
C H A P T E R T WO The Neural Basis for Cognition
engine, but for this sound it might be part of a pattern that includes Neurons
Q, R, and S also firing strongly, and Neuron Y not firing at all.
As it turns out, the brain uses both forms of coding. For example, in
Chapter 4 we’ll see that some neurons really are associated with a particular
content. In fact, researchers documented a cell in one of the people they tested
that fired whenever a picture of Jennifer Aniston was in view, and didn’t fire in
response to pictures of other faces. Another cell (in a different person’s brain)
fired whenever a picture of the Sydney Opera House was shown, but didn’t
fire when other buildings were in view (Quiroga, Reddy, Kreiman, Koch, &
Fried, 2005)! These do seem to be cases in which an idea (in particular, a
certain visual image) is represented by specific neurons in the brain.
In other cases, evidence suggests that ideas and memories are represented
in the brain through widespread patterns of activity. This sort of “pattern
coding” is, for example, certainly involved in the neural mechanisms through
which you plan, and then carry out, particular motions — like reaching out to
turn a book page or lifting your foot to step over an obstacle (Georgopoulos,
1990, 1995). We’ll return to pattern coding in Chapter 9, when we discuss
the notion of a distributed representation.
Moving On
We have now described the brain’s basic anatomy and have also taken a brief
look at the brain’s microscopic parts — the individual neurons. But how do
all of these elements, large and small, function in ways that enable us to think,
remember, learn, speak, or feel? As a step toward tackling this issue, the next
chapter takes a closer look at the portions of the nervous system that allow us
to see. We’ll use the visual system as our example for two important reasons.
First, vision is the modality through which humans acquire a huge amount of
information, whether by reading or simply by viewing the world around us.
If we understand vision, therefore, we understand the processes that bring us
much of our knowledge. Second, investigators have made enormous progress
in mapping out the neural “wiring” of the visual system, offering a detailed
and sophisticated portrait of how this system operates. As a result, an
examination of vision provides an excellent illustration of how the study of
the brain can proceed and what it can teach us.
TEST YOURSELF
12. H
ow is information
coded, or represented,
in the brain?
COGNITIVE PSYCHOLOGY AND EDUCATION
food supplements and cognition
Various businesses try to sell you training programs or food supplements that
(they claim) will improve your memory, help you think more clearly, and so
on. Evidence suggests, though, that the currently offered training programs
Cognitive Psychology and Education
•
55
GINKGO BILOBA
A variety of food supplements
derived from the ginkgo tree
are alleged to improve cognitive functioning. Current understanding, though, suggests
that the benefits of Ginkgo
biloba are indirect: This supplement improves functioning
because it can improve blood
circulation and can help the
body to fight some forms of
inflammation.
56 •
may provide little benefit. These programs do improve performance on the
specific exercises contained within the training itself, but they have no impact
on any tasks beyond these exercises. In other words, the programs don’t seem
to help with the sorts of mental challenges you encounter in day-to-day functioning (Simons et al., 2016).
What about food supplements? Most of these supplements have not been
tested in any systematic way, and so there’s little (and often no) solid evidence
to support the claims sometimes made for these products. One supplement,
though, has been rigorously tested: Ginkgo biloba, an extract derived from
a tree of the same name and advertised as capable of enhancing memory. Is
Ginkgo biloba effective? To answer that question, let’s begin with the fact
that for its normal functioning, the brain requires an excellent blood flow
and, with that, a lot of oxygen and a lot of nutrients. Indeed, it’s estimated
that the brain, constituting roughly 2% of your body weight, consumes
15% percent of your body’s energy supply.
It’s not surprising, therefore, that the brain’s operations are impaired if
some change in your health interferes with the flow of oxygen or nutrients.
If (for example) you’re ill, or not eating enough, or not getting enough sleep,
these conditions affect virtually all aspects of your biological functioning.
However, since the brain is so demanding of nutrients and oxygen, it’s one
of the first organs to suffer if the supply of these necessities is compromised.
This is why poor nutrition or poor health almost inevitably undermines your
ability to think, to remember, or to pay attention.
Within this context, it’s important that Ginkgo biloba can improve blood
circulation and reduce some sorts of bodily inflammation. Because of these
effects, Ginkgo can be helpful for people who have circulatory problems or
who are at risk for nerve damage, and one group that may benefit is patients
with Alzheimer’s disease. Evidence suggests that Ginkgo helps these patients
remember more and think more clearly, but this isn’t because Ginkgo is making these patients “smarter” in any direct way. Instead, the Ginkgo is broadly
improving the patients’ blood circulation and the health status of their nerve
cells, allowing these cells to do their work.
What about healthy people — those not suffering from bodily inflammations or damage to their brain cells? Here, the evidence is mixed, but most
studies have observed no benefit from this food supplement. Apparently,
Ginkgo’s effects, if they exist at all in healthy adults, are so small that they’re
difficult to detect.
Are there other steps that will improve the mental functioning of healthy
young adults? Answers here have to be tentative, because new “smart pills”
and “smart foods” are being proposed all the time, and each one has to be
tested before we can know its effects. For now, though, we’ve already indicated part of a positive answer: Good nutrition, plenty of sleep, and adequate
exercise will keep your blood supply in good condition, and this will help
your brain to do its job. In addition, there may be something else you can do.
The brain needs “fuel” to do its work, and the body’s fuel comes from the
sugar glucose. You can protect yourself, therefore, by making sure that your
C H A P T E R T WO The Neural Basis for Cognition
brain has all the glucose it needs. This isn’t a recommendation to jettison all
other aspects of your diet and eat nothing but chocolate bars. In fact, most
of the glucose your body needs doesn’t come from sugary foods; instead,
most comes from the breakdown of carbohydrates — from the grains, dairy
products, fruits, and vegetables you eat. For this reason, it might be a good
idea to have a slice of bread and a glass of milk just before taking an exam or
walking into a particularly challenging class. These steps will help make sure
that you’re not caught by a glucose shortfall that could interfere with your
brain’s functioning.
Also, be careful not to ingest too much sugar. If you eat a big candy bar
just before an exam, you might get an upward spike in your blood glucose
followed by a sudden drop, and these abrupt changes can produce problems
of their own.
Overall, then, it seems that food supplements tested so far offer no “fast
track” toward better cognition. Ginkgo biloba is helpful, but mostly for special populations. A high-carb snack may help, but it will be of little value if
you’re already adequately nourished. Therefore, on all these grounds, the
best path toward better cognition seems to be the one that common sense
would already recommend — eating a balanced diet, getting a good night’s
sleep, and paying careful attention during your studies.
For more on this topic . . .
Allen, A. L., & Strand, N. K. (2015). Cognitive enhancement and beyond:
Recommendations from the Bioethics Commission. Trends in Cognitive
Sciences, 19, 549–555.
Gold, P. E., Cahill, L., & Wenk, G. L. (2002). Ginkgo biloba: A cognitive
enhancer? Psychological Science in the Public Interest, 3, 2–11.
Husain, M., & Mehta, M. A. (2011). Cognitive enhancement by drugs in health
and disease. Trends in Cognitive Sciences, 15, 28–36.
Masicampo, E., & Baumeister, R. (2008). Toward a physiology of dual-process
reasoning and judgment: Lemonade, willpower, and expensive rule-based
analysis. Psychological Science, 19, 255–260.
McDaniel, M. A., Maier, S. F., & Einstein, G. O. (2002). “Brain-specific” nutrients:
A memory cure? Psychological Science in the Public Interest, 3, 12–38.
Stough, C., & Pase, M. P. (2015). Improving cognition in the elderly with
nutritional supplements. Current Directions in Psychological Sciences, 24,
177–183.
Cognitive Psychology and Education
•
57
chapter review
SUMMARY
• The brain is divided into several different structures, but of particular importance for cognitive
psychology is the forebrain. In the forebrain, each
cerebral hemisphere is divided into the frontal lobe,
parietal lobe, temporal lobe, and occipital lobe. In
understanding these brain areas, one important
source of evidence comes from studies of brain damage, enabling us to examine what sorts of symptoms
result from lesions in specific brain locations. This
has allowed a localization of function, an effort that
is also supported by neuroimaging research, which
shows that the pattern of activation in the brain
depends on the particular task being performed.
• Different parts of the brain perform different jobs;
but for virtually any mental process, different brain
areas must work together in a closely integrated way.
When this integration is lost (as it is, for example,
in Capgras syndrome), bizarre symptoms can result.
• The primary motor projection areas are the
departure points in the brain for nerve cells that
initiate muscle movement. The primary sensory projection areas are the main points of arrival in the
brain for information from the eyes, ears, and other
sense organs. These projection areas generally show
a pattern of contralateral control, with tissue in the
left hemisphere sending or receiving its main signals
from the right side of the body, and vice versa. Each
projection area provides a map of the environment
or the relevant body part, but the assignment of
space in this map is governed by function, not by
anatomical proportions.
• Most of the forebrain’s cortex has traditionally
been referred to as the association cortex, but this
area is subdivided into specialized regions. This
subdivision is reflected in the varying consequences
of brain damage, with lesions in the occipital lobes
leading to visual agnosia, damage in the temporal
lobes leading to aphasia, and so on. Damage to the
prefrontal area causes many different problems, but
these are generally in the forming and implementing
of strategies.
• The brain’s functioning depends on neurons and
glia. The glia perform many functions, but the main
flow of information is carried by the neurons. Communication from one end of the neuron to the other
is electrical and is governed by the flow of ions in
and out of the cell. Communication from one neuron to the next is generally chemical, with a neuron
releasing neurotransmitters that affect neurons on
the other side of the synapse.
KEY TERMS
amygdala (p. 28)
prefrontal cortex (p. 29)
hindbrain (p. 32)
cerebellum (p. 33)
midbrain (p. 33)
forebrain (p. 34)
cortex (p. 34)
convolutions (p. 34)
58
longitudinal fissure (p. 34)
cerebral hemisphere (p. 34)
frontal lobes (p. 34)
central fissure (p. 34)
parietal lobes p. 34)
lateral fissure (p. 34)
temporal lobes (p. 34)
occipital lobes (p. 34)
subcortical structures (p. 34)
thalamus (p. 34)
hypothalamus (p. 34)
limbic system (p. 34)
hippocampus (p. 34)
commisures (p. 36)
corpus callosum (p. 36)
lesion (p. 38)
neuroimaging techniques (p. 38)
computerized axial tomography
(CT scans) (p. 38)
positron emission tomography
(PET scans) (p. 38)
magnetic resonance imaging
(MRI scans) (p. 38)
functional magnetic resonance imaging
(fMRI scans) (p. 38)
electroencephalogram (EEG) (p. 40)
event-related potentials (p. 40)
fusiform face area (FFA) (p. 42)
transcranial magnetic stimulation (TMS) (p. 42)
localization of function (p. 44)
primary motor projection areas (p. 46)
primary sensory projection areas (p. 46)
contralateral control (p. 46)
association cortex (p. 48)
apraxias (p. 48)
agnosias (p. 48)
neglect syndrome (p. 48)
aphasia (p. 48)
neurons (p. 49)
glia (p. 49)
cell body (p. 49)
dendrites (p. 49)
axon (p. 50)
neurotransmitter (p. 52)
synapse (p. 52)
presynaptic membrane (p. 52)
postsynaptic membrane (p. 52)
threshold (p. 52)
action potential (p. 52)
myelin sheath (p. 52)
all-or-none law (p. 54)
coding (p. 54)
TEST YOURSELF AGAIN
1.What are the symptoms of Capgras syndrome,
and why do they suggest a two-part explanation
for how people recognize faces?
8.What is a projection area in the brain? What’s
the role of the motor projection area? The
sensory projection area?
2. What is the cerebral cortex?
3. What are the four major lobes of the forebrain?
9.What does it mean to say that the brain relies
on “contralateral” connections?
4.Identify some of the functions of the hippocampus, the amygdala, and the corpus callosum.
10.What are glia? What are dendrites? What is an
axon? What is a synapse?
5.What is the difference between structural
imaging of the brain and functional imaging?
What techniques are used for each?
11.What does it mean to say that neurons rely on
two different forms of information flow, one
chemical and one electrical?
6.What do we gain from combining different
methods in studying the brain?
12.How is information coded, or represented, in
the brain?
7.What is meant by the phrase “localization of
function”?
59
THINK ABOUT IT
1.People often claim that humans only use 10%
of their brains. Does anything in this chapter
help us in evaluating this claim?
2.People claim that you need to “liberate your
right brain” in order to be creative. What’s true
about this claim? What’s false about this claim?
E eBOOK DEMONSTRATIONS & ESSAYS
Go to http://digital.wwnorton.com/cognition7 for the online demonstrations and essays
relevant to this chapter. These can be found in the ebook.
Online Demonstrations
• Demonstration 2.1: Brain Anatomy
• Demonstration 2.2: The Speed of Neural
Online Applying Cognitive Psychology and the
Law Essays
• Cognitive Psychology and the Law: Improving
the Criminal Justice System
Transmission
• Demonstration 2.3: “Acuity” in the
Somatosensory System
COGNITION LABS
Go to http://digital.wwnorton.com/cognition7 to experience interactive online psychology
labs relevant to this chapter.
60
Learning about
the World
around Us
2
part
I
n setting after setting, you rely on your knowledge and beliefs. But where does
knowledge come from? The answer, usually, is experience, but this invites
another question: What is it that makes experience possible? Tackling this issue
will force us to examine mental processes that turn out to be surprisingly complex.
In Chapter 3, we’ll ask how visual perception operates — and, therefore, how
you manage to perceive the world around you. We’ll start with events in the
eyeball and then move to how you organize and interpret the visual information you receive. We’ll also consider the ways in which you can misinterpret this
information, so that you’re vulnerable to illusions.
Chapter 4 takes a further step and asks how you manage to recognize and
categorize the objects that you see. We’ll start with a simple case: how you
recognize printed letters. We’ll then turn to the recognition of more complex
(three-dimensional) objects.
In both chapters, we’ll discuss the active role that you play in shaping your
experience. We’ll see, for example, that your perceptual apparatus doesn’t just
“pick up” the information that’s available to you. You don’t, in other words, just
open your eyes and let the information “flow in.” Instead, we’ll discuss the ways in
which you supplement and interpret the information you receive. In Chapter 3, we’ll
see that this activity begins very early in the sequence of biological events that
support visual perception. In Chapter 4, these ideas will lead us to a mechanism
made up of very simple components, but shaped by a broad pattern of knowledge.
Chapter 5 then turns to the study of attention. As we’ll see, paying attention is a complex achievement involving many elements. We’ll discuss how the
mechanisms of attention can sometimes limit what people achieve — and so
part of what’s at stake in Chapter 5 is the question of what people ultimately
can or cannot accomplish, and whether there may be ways to escape these
apparent limits on human performance.
61
3
chapter
Visual Perception
what if…
You look around the world and instantly, effortlessly,
recognize the objects that surround you — words on this
page, objects in the room where you’re sitting, things you can view out
the window. Perception, in other words, seems fast, easy, and automatic.
But even so, there is complexity here, and your ability to perceive the
world depends on many separate and individually complicated processes.
Consider the disorder akinetopsia (Zeki, 1991). This condition is rare,
and much of what we know comes from a single patient — L.M. — who
developed this disorder because of a blood clot in her brain, at age 43.
L.M. was completely unable to perceive motion — even though other
aspects of her vision (e.g., her ability to recognize objects, to see color,
or to discern detail in a visual pattern) seemed normal.
Because of her akinetopsia, L.M. can detect that an object now is in a
position different from its position a moment ago, but she reports seeing “nothing in between.” As a way of capturing this experience, think
about what you see when you’re looking at really slow movement. If, for
example, you stare at the hour hand on a clock as it creeps around the
clock face, you cannot discern its motion. But you can easily see the
hand is now pointing, say, at the 4, and if you come back a while later,
you can see that it’s closer to the 5. In this way, you can infer motion from
the change in position, but you can’t perceive the motion. This is your
experience with very slow movement; L.M., suffering from akinetopsia,
has the same experience with all movement.
What’s it like to have this disorder? L.M. complained, as one concern,
that it was hard to cross the street because she couldn’t tell which of the
cars in view were moving and which ones were parked. (She eventually
learned to estimate the position and movement of traffic by listening
to cars’ sounds as they approached, even though she couldn’t see their
movement.)
Other problems caused by akinetopsia are more surprising. For
example, L.M. complained about difficulties in following conversations,
because she was essentially blind to the speaker’s lip movement or
changing facial expressions. She also felt insecure in social settings. If
more than two people were moving around in a room, she felt anxious
because “people were suddenly here or there, but [she had] not seen
them moving” (Zihl, von Cramon, & Mai, 1983, p. 315). Or, as a different
example: She had trouble in everyday activities like pouring a cup of
63
preview of chapter themes
•
•
e explore vision — humans’ dominant sensory modality.
W
We discuss the mechanisms through which the visual
system detects patterns in the incoming light, but we also
showcase the activity of the visual system in interpreting
and shaping the incoming information.
e also highlight the ways in which perception of
W
one aspect of the input is shaped by perception of other
aspects — so that the detection of simple features depends
on how the overall form is organized, and the perception of
size depends on the perceived distance of the target object.
•
e emphasize that the interpretation of the visual input
W
is usually accurate — but the same mechanisms can lead to
illusions, and the study of those illusions can often illuminate the processes through which perception functions.
coffee. She couldn’t see the fluid level’s gradual rise as she poured, so
she didn’t know when to stop pouring. For her, “the fluid appeared to be
frozen, like a glacier” (Zihl et al., 1983, p. 315; also Schenk, Ellison, Rice, &
Milner, 2005; Zihl, von Cramon, Mai, & Schmid, 1991).
We will have more to say about cases of disordered perception later in
the chapter. For now, though, let’s note the specificity of this disorder — a
disruption of movement perception, with other aspects of perception
still intact. Let’s also highlight the important point that each of us is, in
countless ways, dependent on our perceptual contact with the world.
That point demands that we ask: What makes this perception possible?
The Visual System
You receive information about the world through various sensory modalities:
You hear the sound of the approaching train, you smell the freshly baked
bread, you feel the tap on your shoulder. Researchers have made impressive
progress in studying all of these modalities, and students interested in, say,
hearing or the sense of smell will find a course in (or a book about) sensation
and perception to be fascinating.
There’s no question, though, that for humans vision is the dominant sense.
This is reflected in how much brain area is devoted to vision compared to
any of the other senses. It’s also reflected in many aspects of our behavior. For
example, if visual information conflicts with information received from other
senses, you usually place your trust in vision. This is the basis for ventriloquism,
in which you see the dummy’s mouth moving while the sounds themselves are
coming from the dummy’s master. Vision wins out in this contest, and so you
experience the illusion that the voice is coming from the dummy.
The Photoreceptors
How does vision operate? The process begins, of course, with light. Light is
produced by many objects in our surroundings — the sun, lamps, candles — and
then reflects off other objects. In most cases, it’s this reflected light — reflected
64 •
C H A P T E R T H R E E Visual Perception
from this book page or from a friend’s face — that launches the processes of
visual perception. Some of this light hits the front surface of the eyeball, passes
through the cornea and the lens, and then hits the retina, the light-sensitive
tissue that lines the back of the eyeball (see Figure 3.1). The cornea and lens
focus the incoming light, just as a camera lens might, so that a sharp image
is cast onto the retina. Adjustments in this process can take place because the
lens is surrounded by a band of muscle. When the muscle tightens, the lens
bulges somewhat, creating the proper shape for focusing the images cast by
nearby objects. When the muscle relaxes, the lens returns to a flatter shape,
allowing the proper focus for objects farther away.
On the retina, there are two types of photoreceptors — specialized neural
cells that respond directly to the incoming light. One type, the rods, are
sensitive to very low levels of light and so play an essential role whenever you’re moving around in semidarkness or trying to view a fairly dim
FIGURE 3.1
THE HUMAN EYE
Retina
Fovea
Pupil
Cornea
Lens
Iris
Optic nerve
(to brain)
Light enters the eye through the cornea, and the cornea and lens refract the
light rays to produce a sharply focused image on the retina. The iris can open or
close to control the amount of light that reaches the retina. The retina is made
up of three main layers: the rods and cones, which are the photoreceptors; the
bipolar cells; and the ganglion cells, whose axons make up the optic nerve.
The Visual System
•
65
RODS AND CONES
Cone
Rod
A
Thousands of photoreceptors
per square millimeter
FIGURE 3.2
Blind spot
180
Cones
Rods
140
Fovea
100
60
20
0
60
40
20
0
20
40
60
Distance on retina from fovea (degrees)
B
(Panel A) Rods and cones are the light-sensitive cells at the back of the retina that launch the neural process of
vision. In this (colorized) photo, cones appear green; rods appear brown. (Panel B) Distribution of photoreceptors.
Cones are most frequent at the fovea, and the number of cones drops off sharply as we move away from the fovea.
In contrast, there are no rods at all on the fovea. There are neither rods nor cones at the retina’s blind spot—the
position at which the neural fibers that make up the optic nerve exit the eyeball. Because this position is filled with
these fibers, there’s no space for any rods or cones.
stimulus. But the rods are also color-blind: They can distinguish different
intensities of light (and in that way contribute to your perception of brightness), but they provide no means of discriminating one hue from another
(see Figure 3.2).
Cones, in contrast, are less sensitive than rods and so need more incoming light to operate at all. But cones are sensitive to color differences. More
precisely, there are three different types of cones, each having its own pattern
of sensitivities to different wavelengths (see Figure 3.3). You perceive color,
therefore, by comparing the outputs from these three cone types. Strong
firing from only the cones that prefer short wavelengths, for example,
accompanied by weak (or no) firing from the other cone types, signals
purple. Blue is signaled by equally strong firing from the cones that prefer
short wavelengths and those that prefer medium wavelengths, with only
modest firing by cones that prefer long wavelengths. And so on, with other
patterns of firing, across the three cone types, corresponding to different
perceived hues.
Cones have another function: They enable you to discern fine detail. The
ability to see fine detail is referred to as acuity, and acuity is much higher for
66 •
C H A P T E R T H R E E Visual Perception
FIGURE 3.3
WAVELENGTHS OF LIGHT
Wavelength
Pressure
Amplitude
Time
A
White light
Prism
400
500
600
700 Nanometers
Visible light
Gamma
rays
X-rays
Ultraviolet
rays
Infrared
rays
Radar
Broadcast
radio
B
The physics of light are complex, but for many purposes light can be thought of as a wave (Panel A), and the
shape of the wave can be described in terms of its amplitude and its wavelength (i.e., the distance from “crest”
to “crest”). The wavelengths our visual system can sense are only a tiny part of the broader electromagnetic
spectrum (Panel B). Light with a wavelength longer than 750 nanometers is invisible to us, although we feel these
longer infrared waves as heat. Ultraviolet light, which has a wavelength shorter than 360 nanometers, is also
invisible to us. That leaves the narrow band of wavelengths between 750 and 360 nanometers — the so-called
visible spectrum. Within this spectrum, we usually see wavelengths close to 400 nanometers as violet, those
close to 700 nanometers as red, and those in between as the rest of the colors in the rainbow.
the cones than it is for the rods. This explains why you point your eyes toward
a target whenever you want to perceive it in detail. What you’re actually
doing is positioning your eyes so that the image of the target falls onto the
fovea, the very center of the retina. Here, cones far outnumber rods (and, in
fact, the center of the fovea has no rods at all). As a result, this is the region
of the retina with the greatest acuity.
The Visual System
•
67
In portions of the retina more distant from the fovea (i.e., portions of
the retina in the so-called visual periphery), the rods predominate; well out
into the periphery, there are no cones at all. This distribution of photo­
receptors explains why you’re better able to see very dim lights out of the
corner of your eyes. Psychologists have understood this point for at least a
century, but the key observation here has a much longer history. Sailors and
astronomers have known for hundreds of years that when looking at a
barely visible star, it’s best not to look directly at the star’s location. By
looking slightly away from the star, they ensured that the star’s image
would fall outside of the fovea and onto a region of the retina dense with
the more light-sensitive rods.
Lateral Inhibition
Rods and cones do not report directly to the cortex. Instead, the photoreceptors stimulate bipolar cells, which in turn excite ganglion cells. The
ganglion cells are spread uniformly across the entire retina, but all of their
axons converge to form the bundle of nerve fibers that we call the optic
nerve. This is the nerve tract that leaves the eyeball and carries information
to various sites in the brain. The information is sent first to a way station
in the thalamus called the lateral geniculate nucleus (LGN); from there,
information is transmitted to the primary projection area for vision, in the
occipital lobe.
Let’s be clear, though, that the optic nerve is not just a cable that conducts
signals from one site to another. Instead, the cells that link retina to brain are
already analyzing the visual input. One example lies in the phenomenon of
lateral inhibition, a pattern in which cells, when stimulated, inhibit the activity of neighboring cells. To see why this is important, consider two cells, each
receiving stimulation from a brightly lit area (see Figure 3.4). One cell (Cell B
in the figure) is receiving its stimulation from the middle of the lit area. It is
intensely stimulated, but so are its neighbors (including Cell A and Cell C).
As a result, all of these cells are active, and therefore each one is trying to inhibit its neighbors. The upshot is that the activity level of Cell B is increased
by the stimulation but decreased by the lateral inhibition it’s receiving from
Cells A and C. This combination leads to only a moderate level of activity
in Cell B.
In contrast, another cell (Cell C in the figure) is receiving its stimulation from the edge of the lit area. It is intensely stimulated, and so are its
neighbors on one side. Therefore, this cell will receive inhibition from one
side but not from the other (in the figure: inhibition from Cell B but
not from Cell D), so it will be less inhibited than Cell B (which is receiving
inhibition from both sides). Thus, Cells B and C initially receive the same
input, but C is less inhibited than B and so will end up firing more strongly
than B.
68 •
C H A P T E R T H R E E Visual Perception
FIGURE 3.4
LATERAL INHIBITION
Bright physical stimulus
Gray physical stimulus
Intense stimulation
Moderate stimulation
Cell
A
Cell
B
Cell
C
Cell
D
Cell
E
Cell
F
To brain
80 spikes
per second
To brain
90 spikes
per second
To brain
90 spikes
per second
Sideways
connection
providing
inhibition
To brain
1,000 spikes
per second
To brain
1,000 spikes
per second
To brain
1,100 spikes
per second
Stimulus as perceived
Stimulus as perceived
Cell B receives strong inhibition from all its neighbors, because its neighbors are intensely stimulated.
Cell C, in contrast, receives inhibition only from one side (because its neighbor on the other side, Cell D, is
only moderately stimulated). As a result, Cells B and C start with the same input, but Cell C, receiving less
inhibition, sends a stronger signal to the brain, emphasizing the edge in the stimulus. The same logic applies
to Cells D and E, and it explains why Cell D sends a weaker signal to the brain. Note, by the way, that the
spikes per second numbers, shown in the figure, are hypothetical and intended only to illustrate lateral
inhibition’s effects.
Notice that the pattern of lateral inhibition highlights a surface’s edges,
because the response of cells detecting the edge of the surface (such as Cell C)
will be stronger than that of cells detecting the middle of the surface (such as
Cell B). For that matter, by increasing the response by Cell C and decreasing
the response by Cell D, lateral inhibition actually exaggerates the contrast at
the edge — a process called edge enhancement. This process is of enormous
The Visual System
•
69
FIGURE 3.5
MACH BANDS
Edge enhancement, produced by lateral inhibition, helps us to perceive
the outline that defines an object’s shape. But the same process can
produce illusions — including the Mach bands. Each vertical strip in this figure is of uniform light intensity, but the strips don’t appear uniform. For
each strip, contrast makes the left edge (next to its darker neighbor) look
brighter than the rest, while the right edge (next to its lighter neighbor)
looks darker. To see that the differences are illusions, try placing a thin
object (such as a toothpick or a straightened paper clip) on top of the
boundary between strips. With the strips separated in this manner, the
illusion disappears.
TEST YOURSELF
1. What are the differences between rods
and cones? What traits
do these cells share?
2. What is lateral inhibition?
How does it contribute
to edge perception?
importance, because it’s obviously highlighting the information that defines an
object’s shape — information essential for figuring out what the object is. And
let’s emphasize that this edge enhancement occurs at a very early stage of
the visual processing. In other words, the information sent to the brain isn’t
a mere copy of the incoming stimulation; instead, the steps of interpretation
and analysis begin immediately, in the eyeball. (For a demonstration of an
illusion caused by this edge enhancement — the so-called Mach bands — see
Figure 3.5.)
Visual Coding
In Chapter 2, we introduced the idea of coding in the nervous system. This
term refers to the relationship between activity in the nervous system and the
stimulus (or idea or operation) that is somehow represented by that activity. In the study of perception, we can ask: What’s the code through which
neurons (or groups of neurons) manage to represent the shapes, colors, sizes,
and movements that you perceive?
70 •
C H A P T E R T H R E E Visual Perception
Single Neurons and Single-Cell Recording
Part of what we know about the visual system — actually, part of what we
know about the entire brain — comes from a technique called single-cell
recording. As the name implies, this is a procedure through which investigators can record, moment by moment, the pattern of electrical changes within
a single neuron.
We mentioned in Chapter 2 that when a neuron fires, each response is
the same size; this is the all-or-none law. But neurons can vary in how often
they fire, and when investigators record the activity of a single neuron, what
they’re usually interested in is the cell’s firing rate, measured in “spikes per
second.” The investigator can then vary the circumstances (either in the
external world or elsewhere in the nervous system) in order to learn what
makes the cell fire more and what makes it fire less. In this way, we can
figure out what job the neuron does within the broad context of the entire
nervous system.
The technique of single-cell recording has been used with enormous success in the study of vision. In a typical procedure, the animal being studied
is first immobilized. Then, electrodes are placed just outside a neuron in the
animal’s optic nerve or brain. Next, a computer screen is placed in front of
the animal’s eyes, and various patterns are flashed on the screen: circles, lines
at various angles, or squares of various sizes at various positions. Researchers
can then ask: Which patterns cause that neuron to fire? To what visual inputs
does that cell respond?
By analogy, we know that a smoke detector is a smoke detector because
it “fires” (i.e., makes noise) when smoke is on the scene. We know that a
motion detector is a motion detector because it “fires” when something
moves nearby. But what kind of detector is a given neuron? Is it responsive
to any light in any position within the field of view? In that case, we might
call it a “light detector.” Or is it perhaps responsive only to certain shapes
at certain positions (and therefore is a “shape detector”)? With this logic,
we can map out precisely what the cell responds to — what kind of detector
it is. More formally, this procedure allows us to define the cell’s receptive
field — that is, the size and shape of the area in the visual world to which
that cell responds.
Multiple Types of Receptive Fields
In 1981, the neurophysiologists David Hubel and Torsten Wiesel were
awarded the Nobel Prize for their exploration of the mammalian visual
system (e.g., Hubel & Wiesel, 1959, 1968). They documented the existence
of specialized neurons within the brain, each of which has a different type
of receptive field, a different kind of visual trigger. For example, some neurons seem to function as “dot detectors.” These cells fire at their maximum
TORSTEN WIESEL
AND DAVID HUBEL
Much of what we know about
the visual system is based on
the pioneering work done
by David Hubel and Torsten
Wiesel. This pair of researchers won the 1981 Nobel Prize
for their discoveries. (They
shared the Nobel with Roger
Sperry for his independent
research on the cerebral
hemispheres.)
Visual Coding
•
71
Receptive field
Center
Neural firing frequency
Time
A
Surround
Time
B
FIGURE 3.6
CELLS
CENTER-SURROUND
Some neurons in the visual system have receptive
fields with a “center-surround” organization. Panels A through D show the firing frequency for one
of those cells. (A) This graph shows the cell’s firing
rate when no stimulus is presented. (B) The cell’s
firing rate goes up when a stimulus is presented in
the middle of the cell’s receptive field. (C) In contrast, the cell’s firing rate goes down if a stimulus is
presented at the edge of the cell’s receptive field.
(D) If a stimulus is presented both to the center of
the receptive field and to the edge, the cell’s firing
rate does not change from its baseline level.
Time
C
Time
D
rate when light is presented in a small, roughly circular area in a specific
position within the field of view. Presentations of light just outside of this
area cause the cell to fire at less than its usual “resting” rate, so the input
must be precisely positioned to make this cell fire. Figure 3.6 depicts such
a receptive field.
These cells are often called center-surround cells, to mark the fact that light
presented to the central region of the receptive field has one influence, while light
presented to the surrounding ring has the opposite influence. If both the center and
the surround are strongly stimulated, the cell will fire neither more nor less than
usual. For this cell, a strong uniform stimulus is equivalent to no stimulus at all.
72 •
C H A P T E R T H R E E Visual Perception
FIGURE 3.7 ORIENTATION-SPECIFIC
VISUAL FIELDS
Some cells in the visual system fire only when the input contains a
line segment at a certain orientation. For example, one cell might
fire very little in response to a horizontal line, fire only occasionally
in response to a diagonal, and fire at its maximum rate only when
a vertical line is present. In this figure, the circles show the stimulus
that was presented. The right side shows records of neural firing.
Each vertical stroke represents a firing by the cell; the left–right
position reflects the passage of time.
( after hubel , 1963)
Other cells fire at their maximum only when a stimulus containing an
edge of just the right orientation appears within their receptive fields. These
cells, therefore, can be thought of as “edge detectors.” Some of these cells
fire at their maximum rate when a horizontal edge is presented; others,
when a vertical edge is in view; still others fire at their maximum to orientations in between horizontal and vertical. Note, though, that in each case,
these orientations merely define the cells’ “preference,” because these cells
are not oblivious to edges of other orientations. If a cell’s preference is for,
say, horizontal edges, then the cell will still respond to other orientations —
but less strongly than it does for horizontals. Specifically, the farther the
edge is from the cell’s preferred orientation, the weaker the firing will
be, and edges sharply different from the cell’s preferred orientation (e.g.,
a vertical edge for a cell that prefers horizontal) will elicit virtually no
response (see Figure 3.7).
Other cells, elsewhere in the visual cortex, have receptive fields that are
more specific. Some cells fire maximally only if an angle of a particular size
appears in their receptive fields; others fire maximally in response to corners
Visual Coding
•
73
and notches. Still other cells appear to be “movement detectors” and fire
strongly if a stimulus moves, say, from right to left across the cell’s receptive
field. Other cells favor left-to-right movement, and so on through the various
possible directions of movement.
Parallel Processing in the Visual System
This proliferation of cell types highlights another important principle —
namely, that the visual system relies on a “divide and conquer” strategy, with
different types of cells, located in different areas of the cortex, each specializing
in a particular kind of analysis. This pattern is plainly evident in Area V1, the
site on the occipital lobe where axons from the LGN first reach the cortex
(see Figure 3.8). In this brain area, some cells fire to (say) horizontals in this
position in the visual world, others to horizontals in that position, others to
verticals in specific positions, and so on. The full ensemble of cells in this area
FIGURE 3.8 AREA V1 IN
THE HUMAN BRAIN
Area V1 is the site on the occipital lobe where axons from the LGN
first reach the cortex. The top panel
shows the brain as if sliced vertically down the middle, revealing
the “inside” surface of the brain’s
right hemisphere. The bottom panel
shows the left hemisphere of the
brain viewed from the side. As the
two panels show, most of Area V1 is
located on the cortical surface bet­
ween the two cerebral hemispheres.
74 •
C H A P T E R T H R E E Visual Perception
Area V1, primary visual
projection area
provides a detector for every possible stimulus, making certain that no matter
what the input is or where it’s located, some cell will respond to it.
The pattern of specialization is also evident when we consider other
brain areas. Figure 3.9, for example, reflects one summary of the brain areas
known to be involved in vision. The details of the figure aren’t crucial, but it
is noteworthy that some of these areas (V1, V2, V3, V4, PO, and MT) are in
the occipital cortex; other areas are in the parietal cortex; others are in the
temporal cortex. (We’ll have more to say in a moment about these areas outside of the occipital cortex.) Most important, each area seems to have its own
function. Neurons in Area MT, for example, are acutely sensitive to direction
and speed of movement. (This area is the brain region that has suffered damage in cases involving akinetopsia.) Cells in Area V4 fire most strongly when
the input is of a certain color and a certain shape.
Let’s also emphasize that all of these specialized areas are active at the
same time, so that (for example) cells in Area MT are detecting movement in
the visual input at the same time that cells in Area V4 are detecting shapes.
In other words, the visual system relies on parallel processing — a system in
FIGURE 3.9
THE VISUAL PROCESSING PATHWAYS
VIP
Parietal
cortex
PO
MST
Occipital
cortex
7a
MT
LIP
Retina
LGN
V1
V3
V2
V4
TEO
Inferotemporal
cortex
TE
Each box in this figure refers to a specific location within the visual system. Notice that vision depends on many
brain sites, each performing a specialized type of analysis. Note also that the flow of information is complex, so
there’s no strict sequence of “this step” of analysis followed by “that step.” Instead, everything happens at once,
with a great deal of back-and-forth communication among the various elements.
Visual Coding
•
75
which many different steps (in this case, different kinds of analysis) are going
on simultaneously. (Parallel processing is usually contrasted with serial
processing, in which steps are carried out one at a time — i.e., in a series.)
One advantage of this simultaneous processing is speed: Brain areas
trying to discern the shape of the incoming stimulus don’t need to wait until
the motion analysis or the color analysis is complete. Instead, all of the analyses go forward immediately when the input appears before the eyes, with no
waiting time.
Another advantage of parallel processing is the possibility of mutual influence among multiple systems. To see why this matters, consider the fact that
sometimes your interpretation of an object’s motion depends on your understanding of the object’s three-dimensional shape. This suggests that it might
be best if the perception of shape happened first. That way, you could use
the results of this processing step as a guide to later analyses. In other cases,
though, the relationship between shape and motion is reversed. In these cases,
your interpretation of an object’s three-dimensional shape depends on your
understanding of its motion. To allow for this possibility, it might be best if
the perception of motion happened first, so that it could guide the subsequent
analysis of shape.
How does the brain deal with these contradictory demands? Parallel processing provides the answer. Since both sorts of analysis go on simultaneously,
each type of analysis can be informed by the other. Put differently, neither
the shape-analyzing system nor the motion-analyzing system gets priority.
Instead, the two systems work concurrently and “negotiate” a solution that
satisfies both systems (Van Essen & DeYoe, 1995).
Parallel processing is easy to document throughout the visual system. As
we’ve seen, the retina contains two types of specialized receptors (rods and
cones) each doing its own job (e.g., the rods detecting stimuli in the periphery
of your vision and stimuli presented at low light levels, and the cones detecting hues and detail at the center of your vision). Both types of receptors function at the same time — another case of parallel processing.
Likewise, within the optic nerve itself, there are two types of cells, P cells
and M cells. The P cells provide the main input for the LGN’s parvocellular
cells and appear to be specialized for spatial analysis and the detailed analysis
of form. M cells provide the input for the LGN’s magnocellular cells and are
specialized for the detection of motion and the perception of depth.1 And,
again, both of these systems are functioning at the same time — more parallel
processing.
Parallel processing remains in evidence when we move beyond the occipital cortex. As Figure 3.10 shows, some of the activation from the occipital
1. The names here refer to the relative sizes of the relevant cells: parvo derives from the Latin
word for “small,” and magno from the word for “large.” To remember the function of these two
types of cells, many students think of the P cells as specialized roughly for the perception of
pattern and M cells as specialized for the perception of motion. These descriptions are crude,
but they’re easy to remember.
76 •
C H A P T E R T H R E E Visual Perception
FIGURE 3.10
THE WHAT AND WHERE PATHWAYS
Parietal
lobe
Posterior
parietal
cortex
Temporal
lobe
Occipital
lobe
Inferotemporal
cortex
Information from the primary visual cortex at the back of the head is transmitted to the inferotemporal cortex (the so-called what system) and to the posterior parietal cortex (the where system). The term “inferotemporal” refers to
the lower part of the temporal lobe. The term “posterior parietal cortex” refers
to the rearmost portion of this cortex.
lobe is passed along to the cortex of the temporal lobe. This pathway, often
called the what system, plays a major role in the identification of visual
objects, telling you whether the object is a cat, an apple, or whatever. At the
same time, activation from the occipital lobe is also passed along a second
pathway, leading to the parietal cortex, in what is often called the where
system. This system seems to guide your action based on your perception of
where an object is located — above or below you, to your right or to your left.
(See Goodale & Milner, 2004; Humphreys & Riddoch, 2014; Ungerleider
& Haxby, 1994; Ungerleider & Mishkin, 1982. For some complications,
though, see Borst, Thompson, & Kosslyn, 2011; de Haan & Cowey, 2011.)
The contrasting roles of these two systems can be revealed in many ways,
including through studies of brain damage. Patients with lesions in the what
system show visual agnosia — an inability to recognize visually presented
objects, including such common things as a cup or a pencil. However, these
patients show little disorder in recognizing visual orientation or in reaching.
The reverse pattern occurs with patients who have suffered lesions in the
where system: They have difficulty in reaching, but no problem in object
identification (Damasio, Tranel, & Damasio, 1989; Farah, 1990; Goodale,
1995; Newcombe, Ratcliff, & Damasio, 1987).
Still other data echo this broad theme of parallel processing among separate systems. For example, we noted earlier that different brain areas are
Visual Coding
•
77
critical for the perception of color, motion, and form. If this is right, then
someone who has suffered damage in just one of these areas might show
problems in the perception of color but not the perception of motion or form,
or problems in the perception of motion but not the perception of form or
color. These predictions are correct. As we mentioned at the chapter’s start,
some patients suffer damage to the motion system and so develop akinetopsia (Zihl et al., 1983). For such patients, the world is described as a succession of static photographs. They’re unable to report the speed or direction of
a moving object; as one patient put it, “When I’m looking at the car first, it
seems far away. But then when I want to cross the road, suddenly the car is
very near” (Zihl et al., 1983, p. 315).
Other patients suffer a specific loss of color vision through damage to the
central nervous system, even though their perception of form and motion
remains normal (Damasio, 1985; Gazzaniga, Ivry, & Mangun, 2014; Meadows,
1974). To them, the entire world is clothed only in “dirty shades of gray.”2
Cases like these provide dramatic confirmation of the separateness of our
visual system’s various elements and the ways in which the visual system is
vulnerable to very specific forms of damage. (For further evidence with neurologically intact participants, see Bundesen, Kyllingsbaek, & Larsen, 2003.)
Putting the Pieces Back Together
Let’s emphasize once again, therefore, that even the simplest of our intellectual achievements depends on an array of different, highly specialized brain
areas all working together in parallel. This was evident in Chapter 2 in our
consideration of Capgras syndrome, and the same pattern has emerged in
our description of the visual system. Here, too, many brain areas must work
together: the what system and the where system, areas specialized for the detection of movement and areas specialized for the identification of simple forms.
We have identified the advantages that come from this division of labor
and the parallel processing it allows. But the division of labor also creates a
problem: If multiple brain areas contribute to an overall task, how is their
functioning coordinated? When you see an athlete make an astonishing jump,
the jump itself is registered by motion-sensitive neurons, but your recognition
of the athlete depends on shape-sensitive neurons. How are the pieces put
back together? When you reach for a coffee cup but stop midway because
you see that the cup is empty, the reach itself is guided by the where system;
the fact that the cup is empty is registered by the what system. How are these
two streams of processing coordinated?
Investigators refer to this broad issue as the binding problem — the
task of reuniting the various elements of a scene, elements that are initially
addressed by different systems in different parts of the brain. And obviously
2. This is different from ordinary color blindness, which is usually present from birth and results from
abnormalities that are outside the brain itself — for example, abnormalities in the photoreceptors.
78 •
C H A P T E R T H R E E Visual Perception
this problem is solved. What you perceive is not an unordered catalogue
of sensory elements. Instead, you perceive a coherent, integrated perceptual
world. Apparently, this is a case in which the various pieces of Humpty
Dumpty are reassembled to form an organized whole.
Visual Maps and Firing Synchrony
Look around you. Your visual system registers whiteness and blueness and
brownness; it also registers a small cylindrical shape (your coffee cup), a mediumsized rectangle (this book page), and a much larger rectangle (your desk). How
do you put these pieces together so that you see that it’s the coffee cup, and not
the book page, that’s blue; the desktop, and not the cup, that’s brown?
There is debate about how the visual system solves this problem, but we
can identify three elements that contribute to the solution. One element is
spatial position. The part of the brain registering the cup’s shape is separate
from the parts registering its color or its motion; nonetheless, these various
brain areas all have something in common. They each keep track of where
the target is — where the cylindrical shape was located, and where the blueness was; where the motion was detected, and where things were still. As a
result, the reassembling of these pieces can be done with reference to position. In essence, you can overlay the map of which forms are where on top
of the map of which colors are where to get the right colors with the right
forms, and likewise for the map showing which motion patterns are where.
Information about spatial position is, of course, useful for its own sake:
You have a compelling reason to care whether the tiger is close to you or
far away, or whether the bus is on your side of the street or the other. But in
addition, location information apparently provides a frame of reference used
to solve the binding problem. Given this double function, we shouldn’t be
surprised that spatial position is a major organizing theme in all the various
brain areas concerned with vision, with each area seeming to provide its own
map of the visual world.
Spatial position, however, is not the whole story. Evidence also suggests
that the brain uses special rhythms to identify which sensory elements belong
with which. Imagine two groups of neurons in the visual cortex. One group
of neurons fires maximally whenever a vertical line is in view; another group
fires maximally whenever a stimulus is in view moving from a high position
to a low one. Let’s also imagine that right now a vertical line is presented
and it is moving downward; as a result, both groups of neurons are firing
strongly. How does the brain encode the fact that these attributes are bound
together, different aspects of a single object? There is evidence that the visual
system marks this fact by means of neural synchrony: If the neurons detecting
a vertical line are firing in synchrony with those signaling movement, then
these attributes are registered as belonging to the same object. If they aren’t
in synchrony, then the features aren’t bound together (Buzsáki & Draguhn,
2004; Csibra, Davis, Spratling, & Johnson, 2000; Elliott & Müller, 2000;
Fries, Reynolds, Rorie, & Desimone, 2001).
Visual Coding
•
79
TEST YOURSELF
3.How do researchers
use single-cell recording to reveal a cell’s
receptive field?
4.What are the advantages of parallel processing in the visual
system? What are the
disadvantages?
5. How is firing synchrony
relevant to the solution of the binding
problem?
What causes this synchrony? How do the neurons become synchronized
in the first place? Here, another factor appears to be important: attention.
We’ll have more to say about attention in Chapter 5, but for now let’s note
that attention plays a key role in binding together the separate features of a
stimulus. (For a classic statement of this argument, see Treisman & Gelade,
1980; Treisman, Sykes, & Gelade, 1977. For more recent views, see Quinlan,
2003; Rensink, 2012; and also Chapter 5.)
Evidence for attention’s role comes from many sources, including the fact
that when we overload someone’s attention, she is likely to make conjunction
errors. This means that she’s likely to correctly detect the features present in
a visual display, but then to make mistakes about how the features are bound
together (or conjoined). Thus, for example, someone shown a blue H and a
red T might report seeing a blue T and a red H — an error in binding.
Similarly, individuals who suffer from severe attention deficits (because
of brain damage in the parietal cortex) are particularly impaired in tasks
that require them to judge how features are conjoined to form complex
objects (e.g., Robertson, Treisman, Friedman-Hill, & Grabowecky, 1997).
Finally, studies suggest that synchronized neural firing occurs in an animal’s
brain when the animal is attending to a specific stimulus but does not occur
in neurons activated by an unattended stimulus (e.g., Buschman & Miller,
2007; Saalmann, Pigarev, & Vidyasagar, 2007; Womelsdorf et al., 2007). All
of these results point toward the claim that attention is crucial for the binding
problem and, moreover, that attention is linked to the neural synchrony that
seems to unite a stimulus’s features.
Notice, then, that there are several ways in which information is represented
in the brain. In Chapter 2, we noted that the brain uses different chemical signals
(i.e., different neurotransmitters) to transmit different types of information.
We now see that there is information reflected in which cells are firing, how often
they are firing, whether the cells are firing in synchrony with other cells, and the
rhythm in which they are firing. Plainly, this is a system of considerable complexity!
Form Perception
So far in this chapter, we’ve been discussing how visual perception begins:
with the detection of simple attributes in the stimulus — its color, its motion,
and its catalogue of features. But this detection is just the start of the process,
because the visual system still has to assemble these features into recognizable
wholes. We’ve mentioned the binding problem as part of this “assembly” — but
binding isn’t the whole story. This point is reflected in the fact that our perception of the visual world is organized in ways that the stimulus input is
not — a point documented early in the 20th century by a group called the
“Gestalt psychologists.”3 The Gestaltists argued that the organization is
3. Gestalt is the German word for “shape” or “form.” The Gestalt psychology movement was
committed to the view that theories about perception and thought need to emphasize the
organization of patterns, not just focus on a pattern’s elements.
80 •
C H A P T E R T H R E E Visual Perception
contributed by the perceiver; this is why, they claimed, the perceptual whole
is often different from the sum of its parts. Some years later, Jerome Bruner
(1973) voiced related claims and coined the phrase “beyond the information
given” to describe some of the ways our perception of a stimulus differs from
(and goes beyond) the stimulus itself.
For example, consider the form shown in the top of Figure 3.11: the
Necker cube. This drawing is an example of a reversible (or ambiguous)
figure — so-called because people perceive it first one way and then another.
Specifically, this form can be perceived as a drawing of a cube viewed from
above (in which case it’s similar to the cube marked A in the figure); it can also
be perceived as a cube viewed from below (in which case it’s similar to the
cube marked B). Let’s be clear, though, that this isn’t an “illusion,” because neither of these interpretations is “wrong,” and the drawing itself (and, therefore,
the information reaching your eyes) is fully compatible with either interpretation. Put differently, the drawing shown in Figure 3.11 is entirely neutral
with regard to the shape’s configuration in depth; the lines on the page don’t
specify which is the “proper” interpretation. Your perception of the cube, however, is not neutral. Instead, you perceive the cube as having one configuration
or the other — similar either to Cube A or to Cube B. Your perception goes
beyond the information given in the drawing, by specifying an arrangement
in depth.
FIGURE 3.11
A
THE NECKER CUBE
B
The top cube can be perceived as if viewed from above (in which case it is a
transparent version of Cube A) or as if viewed from below (in which case it is
a transparent version of Cube B).
Form Perception
•
81
FIGURE 3.12
A
AMBIGUOUS FIGURES
B
Some stimuli easily lend themselves to reinterpretation. The figure in Panel A,
for example, is perceived by many to be a white vase or candlestick on a
black background; others see it as two black faces shown in profile. A similar
bistable form is visible in the Canadian flag (Panel B).
The same point can be made for many other stimuli. Figure 3.12A (after
Rubin, 1915, 1921) can be perceived either as a vase centered in the picture
or as two profiles facing each other. The drawing by itself is compatible
with either of these perceptions, and so, once again, the drawing is neutral
with regard to perceptual organization. In particular, it is neutral with regard to figure/ground organization, the determination of what is the figure
(the depicted object, displayed against a background) and what is the
ground. Your perception of this drawing, however, isn’t neutral about this
point. Instead, your perception somehow specifies that you’re looking at
the vase and not the profiles, or that you’re looking at the profiles and not
the vase.
Figure/ground ambiguity is also detectable in the Canadian flag (Figure 3.12B).
Since 1965, the centerpiece of Canada’s flag has been a red maple leaf. Many
observers, however, note that a different organization is possible, at least for
part of the flag. On their view, the flag depicts two profiles, shown in white
against a red backdrop. Each profile has a large nose, an open mouth, and
a prominent brow ridge, and the profiles are looking downward, toward the
flag’s center.
In all these examples, then, your perception contains information — about
how the form is arranged in depth, or about which part of the form is figure
and which is ground — that is not contained within the stimulus itself. Apparently, this is information contributed by you, the perceiver.
82 •
C H A P T E R T H R E E Visual Perception
The Gestalt Principles
With figures like the Necker cube or the vase/profiles, your role in shaping
the perception seems undeniable. In fact, if you stare at either of these figures,
your perception flips back and forth — first you see the figure one way, then
another, then back to the first way. But the stimulus itself isn’t changing, and
so the information that’s reaching your eyes is constant. Any changes in perception, therefore, are caused by you and not by some change in the stimulus.
One might argue, though, that reversible figures are special — carefully
designed to support multiple interpretations. On this basis, perhaps you play
a smaller role when perceiving other, more “natural” stimuli.
This position is plausible — but wrong, because many stimuli (and not just
the reversible figures) are ambiguous and in need of interpretation. We often
don’t detect this ambiguity, but that’s because the interpretation happens so
quickly that we don’t notice it. Consider, for example, the scene shown in
Figure 3.13. It’s almost certain that you perceive segments B and E as being
FIGURE 3.13
HE ROLE OF INTERPRETATION IN
T
PERCEIVING AN ORDINARY SCENE
C
A
E
B
D
A
B
Consider the still life (Panel A) and an overlay designating five different
segments of the scene (Panel B). For this picture to be perceived correctly,
the perceptual system must first decide what goes with what — for example,
that Segment B and Segment E are different bits of the same object (even
though they’re separated by Segment D) and that Segment B and Segment A
are different objects (even though they’re adjacent and the same color).
Form Perception
•
83
united, forming a complete apple, but notice that this information isn’t provided by the stimulus; instead, it’s your interpretation. (If we simply go with the
information in the figure, it’s possible that segments B and E are parts of
entirely different fruits, with the “gap” between the two fruits hidden from view
by the banana.) It’s also likely that you perceive the banana as entirely bananashaped and therefore continuing downward out of your view, into the bowl,
where it eventually ends with the sort of point that’s normal for a banana. In
the same way, surely you perceive the horizontal stripes in the background as
continuous and merely hidden from view by the pitcher. (You’d be surprised if
we removed the pitcher and revealed a pitcher-shaped gap in the stripes.) But,
of course, the stimulus doesn’t in any way “guarantee” the banana’s shape or
the continuity of the stripes; these points are, again, just your interpretation.
Even with this ordinary scene, therefore, your perception goes “beyond
the information given” — and so the unity of the two apple slices and the
continuity of the stripes is “in the eye of the beholder,” not in the stimulus itself. Of course, you don’t feel like you’re “interpreting” this picture or
extrapolating beyond what’s on the page. But your role becomes clear the
moment we start cataloguing the differences between your perception and
the information that’s truly present in the photograph.
Let’s emphasize, though, that your interpretation of the stimulus isn’t careless or capricious. Instead, you’re guided by a few straightforward principles
that the Gestalt psychologists catalogued many years ago — and so they’re
routinely referred to as the Gestalt principles. For example, your perception
is guided by proximity and similarity: If, within the visual scene, you see
elements that are close to each other, or elements that resemble each other,
you assume these elements are parts of the same object (Figure 3.14). You
also tend to assume that contours are smooth, not jagged, and you avoid
FIGURE 3.14
Similarity
We tend to group
these dots into
columns rather
than rows,
grouping dots of
similar colors.
GESTALT PRINCIPLES OF ORGANIZATION
Proximity
We tend to
perceive groups,
linking dots that
are close together.
Good continuation
We tend to see a
continuous green
bar rather than two
smaller rectangles.
Closure
We tend to perceive
an intact triangle,
reflecting our bias
toward perceiving
closed figures rather
than incomplete ones.
Simplicity
We tend to interpret a form
in the simplest way possible.
We would see the form on
the left as two intersecting
rectangles (as shown on the
right) rather than as a single
12-sided irregular polygon.
As Figure 3.13 illustrated, your ordinary perception of the world requires you to make decisions about what goes
with what — which elements are part of the same object, and which elements belong to different objects. Your
decisions are guided by a few simple principles, catalogued many years ago by the Gestalt psychologists.
84 •
C H A P T E R T H R E E Visual Perception
interpretations that involve coincidences. (For a modern perspective on these
principles and Gestalt psychology in general, see Wagemans, Elder, Kubovy
et al., 2012; Wagemans, Feldman, Gephstein et al., 2010.)
These perceptual principles are quite straightforward, but they’re essential if
your perceptual apparatus is going to make sense of the often ambiguous, often
incomplete information provided by your senses. In addition, it’s worth mentioning that everyone’s perceptions are guided by the same principles, and that’s why
you generally perceive the world in the same way that other people do. Each of
us imposes our own interpretation on the perceptual input, but we all tend to
impose the same interpretation because we’re all governed by the same rules.
Organization and Features
We’ve now considered two broad topics — the detection of simple attributes
in the stimulus, and then the ways in which you organize those attributes. In
thinking about these topics, you might want to think about them as separate
steps. First, you collect information about the stimulus, so that you know
(for example) what corners or angles or curves are in view — the visual features contained within the input. Then, once you’ve gathered the “raw data,”
you interpret this information. That’s when you “go beyond the information
given” — deciding how the form is laid out in depth (as in Figure 3.11), deciding what is figure and what is ground (Figure 3.12A or B), and so on.
The idea, then, is that perception might be divided (roughly) into an
“information gathering” step followed by an “interpretation” step. This view,
however, is wrong, and, in fact, it’s easy to show that in many settings, your
interpretation of the input happens before you start cataloguing the input’s
basic features, not after. Consider Figure 3.15. Initially, these shapes seem to
FIGURE 3.15
A HIDDEN FIGURE
Initially, these dark shapes have no meaning, but after a moment the hidden
figure becomes clearly visible. Notice, therefore, that at the start the figure
seems not to contain the features needed to identify the various letters. Once
the figure is reorganized, with the white parts (not the dark parts) making up
the figure, the features are easily detected. Apparently, the analysis of features depends how the figure is first organized by the viewer.
Form Perception
•
85
have no meaning, but after a moment most people discover the word hidden
in the figure. That is, people find a way to reorganize the figure so that the
familiar letters come into view. But let’s be clear about what this means. At
the start, the form seems not to contain the features needed to identify the L,
the I, and so on. Once the form is reorganized, though, it does contain these
features, and the letters are immediately recognized. In other words, with one
organization, the features are absent; with another, they’re plainly present. It
would seem, then, that the features themselves depend on how the form is
organized by the viewer — and so the features are as much “in the eye of the
beholder” as they are in the figure itself.
As a different example, you have no difficulty reading the word printed
in Figure 3.16, although most of the features needed for this recognition are
absent. You easily “provide” the missing features, though, thanks to the fact
that you interpret the black marks in the figure as shadows cast by solid
letters. Given this interpretation and the extrapolation it involves, you can
easily “fill in” the missing features and read the word.
How should we think about all of this? On one hand, your perception of a
form surely has to start with the stimulus itself and must in some ways be governed by what’s in that stimulus. (After all, no matter how you try to interpret
Figure 3.16, it won’t look to you like a photograph of Queen Elizabeth — the
basic features of the queen are just not present, and your perception respects
this obvious fact.) This suggests that the features must be in place before an
interpretation is offered, because the features govern the interpretation. But,
on the other hand, Figures 3.15 and 3.16 suggest that the opposite is the case:
that the features you find in an input depend on how the figure is interpreted.
Therefore, it’s the interpretation, not the features, that must be first.
The solution to this puzzle, however, is easy, and builds on ideas that we’ve
already met: Many aspects of the brain’s functioning depend on parallel processing, with different brain areas all doing their work at the same time. In
addition, the various brain areas all influence one another, so that what’s
going on in one brain region is shaped by what’s going on elsewhere. In this
FIGURE 3.16
MISSING FEATURES
PERCEPTION
People have no trouble reading this word, even though most of the features
needed for recognition are absent from the stimulus. People easily “supply”
the missing features, illustrating once again that the analysis of features
depends on how the overall figure has been interpreted and organized.
86 •
C H A P T E R T H R E E Visual Perception
way, the brain areas that analyze a pattern’s basic features do their work at
the same time as the brain areas that analyze the pattern’s large-scale configuration, and these brain areas interact so that the perception of the features is
guided by the configuration, and analysis of the configuration is guided by
the features. In other words, neither type of processing “goes first.” Neither
has priority. Instead, they work together, with the result that the perception
that is achieved makes sense at both the large-scale and fine-grained levels.
Constancy
We’ve now seen many indications of the perceiver’s role in “going beyond the
information given” in the stimulus itself. This theme is also evident in another
aspect of perception: the achievement of perceptual constancy. This term
refers to the fact that we perceive the constant properties of objects in the world
(their sizes, shapes, and so on) even though the sensory information we receive
about these attributes changes whenever our viewing circumstances change.
To illustrate this point, consider the perception of size. If you happen to
be far away from the object you’re viewing, then the image cast onto your
retinas by that object will be relatively small. If you approach the object, then
the image size will increase. This change in image size is a simple consequence
of physics, but you’re not fooled by this variation. Instead, you manage to
achieve size constancy — you correctly perceive the sizes of objects despite the
changes in retinal-image size created by changes in viewing distance.
Similarly, if you view a door straight on, the retinal image will be rectangular; but if you view the same door from an angle, the retinal image will
have a different shape (see Figure 3.17). Still, you achieve shape constancy —
that is, you correctly perceive the shapes of objects despite changes in the
retinal image created by shifts in your viewing angle. You also achieve brightness constancy — you correctly perceive the brightness of objects whether
they’re illuminated by dim light or strong sun.
FIGURE 3.17
TEST YOURSELF
6. W
hat evidence tells us
that perception goes
beyond (includes more
information than) the
stimulus input?
7. W
hat are the Gestalt
principles, and how do
they influence visual
perception?
8.What evidence is there
that the perception
of an overall form
depends on the detection of features? What
evidence is there that
the detection of features depends on the
overall form?
SHAPE CONSTANCY
If you change your viewing angle, the shape of the retinal image cast by a target changes. In this figure, the
door viewed straight on casts a rectangular image on
your retina; the door viewed from an angle casts a trape­
zoidal image. Nonetheless, you generally achieve shape
constancy.
Constancy
•
87
Unconscious Inference
How do you achieve each of these forms of constancy? One hypothesis
focuses on relationships within the retinal image. In judging size, for example, you generally see objects against some background, and this can provide
a basis for comparison with the target object. To see how this works, imagine
that you’re looking at a dog sitting on the kitchen floor. Let’s say the dog is
half as tall as the nearby chair and hides eight of the kitchen’s floor tiles from
view. If you take several steps back from the dog, none of these relationships
change, even though the sizes of all the retinal images are reduced. Size constancy, therefore, might be achieved by focusing not on the images themselves
but on these unchanging relationships (see Figure 3.18).
Relationships do contribute to size constancy, and that’s why you’re better
able to judge size when comparison objects are in view or when the target
you’re judging sits on a surface that has a uniform visual texture (like the
floor tiles in the example). But these relationships don’t tell the whole story.
Size constancy is achieved even when the visual scene offers no basis for comparison (if, for example, the object to be judged is the only object in view),
provided that other cues signal the distance of the target object (Harvey &
Leibowitz, 1967; Holway & Boring, 1947).
How does your visual system use this distance information? More than
a century ago, the German physicist Hermann von Helmholtz developed an
influential hypothesis regarding this question. Helmholtz started with the
fact that there’s a simple inverse relationship between distance and retinal image size: If an object doubles its distance from the viewer, the size of
its image is reduced by half. If an object triples its distance, the size of its
FIGURE 3.18 AN INVARIANT
RELATIONSHIP THAT
PROVIDES INFORMATION
ABOUT SIZE
One proposal is that you achieve size
constancy by focusing on relation­
ships in the visual scene. For example,
the dog sitting nearby on the kitchen
floor (Panel A) is half as tall as the
chair and hides eight of the kitchen’s
floor tiles from view. If you take several
steps back from the dog (Panel B),
none of these relationships change,
even though the sizes of all the retinal images are reduced. By focusing
on the relationships, then, you can see
that the dog’s size hasn’t changed.
88 •
A
C H A P T E R T H R E E Visual Perception
B
image is reduced to a third of its initial size. This relationship is guaranteed
to hold true because of the principles of optics, and the relationship makes
it possible for perceivers to achieve size constancy by means of a simple
calculation. Of course, Helmholtz knew that we don’t run through a conscious calculation every time we perceive an object’s size, but he believed
we’re calculating nonetheless — and so he referred to the process as an
unconscious inference (Helmholtz, 1909).
What is the calculation that enables someone to perceive size correctly?
It’s multiplication: the size of the image on the retina, multiplied by the distance between you and the object. (We’ll have more to say about how you
know this distance in a later section.) As an example, imagine an object that,
at a distance of 10 ft, casts an image on the retina that’s 4 mm across. Because
of straightforward principles of optics, the same object, at a distance of 20 ft,
casts an image of 2 mm. In both cases, the product — 10 3 4 or 20 3 2 — is
the same. If, therefore, your size estimate depends on that product, your size
estimate won’t be thrown off by viewing distance — and that’s exactly what
we want (see Figure 3.19).
FIGURE 3.19
THE RELATIONSHIP BETWEEN IMAGE SIZE AND DISTANCE
Closer objects cast larger retinal images
d
Retinal
image
Farther objects cast smaller retinal images
2d
Retinal
image
If you view an object from a greater distance, the object casts a smaller image on your retina. Nonetheless, you
generally achieve size constancy — perceiving the object’s actual size. Helmholtz proposed that you achieve constancy through an unconscious inference—essentially multiplying the image size by the distance.
Constancy
•
89
What’s the evidence that size constancy does depend on this sort of inference? In many experiments, researchers have shown participants an object
and, without changing the object’s retinal image, have changed the apparent
distance of the object. (There are many ways to do this — lenses that change
how the eye has to focus to bring the object into sharp view, or mirrors that
change how the two eyes have to angle inward so that the object’s image
is centered on both foveas.) If people are — as Helmholtz proposed — using
distance information to judge size, then these manipulations should affect
size perception. Any manipulation that makes an object seem farther away
(without changing retinal image size) should make that object seem bigger
(because, in essence, the perceiver would be “multiplying” by a larger number). Any manipulation that makes the object seem closer should make it
look smaller. And, in fact, these predictions are correct — a powerful confirmation that people do use distance to judge size.
A similar proposal explains how people achieve shape constancy. Here,
you take the slant of the surface into account and make appropriate
adjustments — again, an unconscious inference — in your interpretation of
the retinal image’s shape. Likewise for brightness constancy: Perceivers are
sensitive to how a surface is oriented relative to the available light sources,
and they take this information into account in estimating how much light
is reaching the surface. Then, they use this assessment of lighting to judge
the surface’s brightness (e.g., whether it’s black or gray or white). In all
these cases, therefore, it appears that the perceptual system does draw
some sort of unconscious inference, taking viewing circumstances into
account in a way that enables you to perceive the constant properties of
the visual world.
Illusions
This process of taking information into account — whether it’s distance (in
order to judge size), viewing angle (to judge shape), or illumination (to judge
brightness) — is crucial for achieving constancy. More than that, it’s another
indication that you don’t just “receive” visual information; instead, you interpret it. The interpretation is an essential part of your perception and generally helps you perceive the world correctly.
The role of the interpretation becomes especially clear, however, in circumstances in which you misinterpret the information available to you and end
up misperceiving the world. Consider the two tabletops shown in Figure 3.20.
The table on the left looks quite a bit longer and thinner than the one on the
right; a tablecloth that fits one table surely won’t fit the other. Objectively,
though, the parallelogram depicting the left tabletop is exactly the same shape
as the one depicting the right tabletop. If you were to cut out the shape on the
page depicting the left tabletop, rotate it, and slide it onto the right tabletop,
they’d be an exact match. (Not convinced? Just lay another piece of paper on
top of the page, trace the left tabletop, and then move your tracing onto the
right tabletop.)
90 •
C H A P T E R T H R E E Visual Perception
FIGURE 3.20
TWO TABLETOPS
These two tabletops seem to have very different shapes and sizes.
However, this contrast is an illusion — and the shapes drawn here
(the two parallelograms depicting the tabletops) are identical in
shape and size. The illusion is caused by the same mechanisms that,
in most circumstances, allow you to achieve constancy.
Why do people misperceive these shapes? The answer involves the normal
mechanisms of shape constancy. Cues to depth in this figure cause you to
perceive the figure as a drawing of three-dimensional objects, each viewed
from a particular angle. This leads you — quite automatically — to adjust for
the (apparent) viewing angles in order to perceive the two tabletops, and it’s
this adjustment that causes the illusion. Notice, then, that this illusion about
shape is caused by a misperception of depth: You misperceive the depth relationships in the drawing and then take this faulty information into account in
interpreting the shapes. (For a related illusion, see Figure 3.21.)
FIGURE 3.21 THE
MONSTER ILLUSION
The two monsters appear rather
different in size. But, again, this
is an illusion, because the two
drawings are exactly the same
size. The illusion is created by
the distance cues in the picture,
which make the monster on
the right appear to be farther
away. This (mis)perception of distance leads to a (mis)perception of size.
Constancy
•
91
FIGURE 3.22
A BRIGHTNESS ILLUSION
The central square (third row, third column) appears
much brighter than the square marked by the arrow.
Once again, though, this is an illusion. If you don’t
believe it, use your fingers or pieces of paper to cover
everything in the figure except for these two squares.
TEST YOURSELF
9.What does it mean to
say that size constancy may depend on an
unconscious inference? An inference
about what?
10. How do the ordinary
mechanisms of constancy lead to visual
illusions?
A different example is shown in Figure 3.22. It seems obvious to most
viewers that the center square in this checkerboard (third row, third column)
is a brighter shade than the square indicated by the arrow. But, in truth, the
shade of gray shown on the page is identical for these two squares. What has
happened here? The answer again involves the normal processes of perception.
First, the mechanisms of lateral inhibition (described earlier) play a role here
in producing a contrast effect: The central square in this figure is surrounded
by dark squares, and the contrast makes the central square look brighter. The
square marked at the edge of the checkerboard, however, is surrounded by
white squares; here, contrast makes the marked square look darker.
But, in addition, the visual system also detects that the central square is in
the shadow cast by the cylinder. Your vision compensates for this fact — again,
an example of unconscious inference that takes the shadow into account in
judging brightness — and therefore powerfully magnifies the illusion.
The Perception of Depth
In discussing constancy, we said that perceivers take distance, slant, and
illumination into account in judging size, shape, and brightness. But to do this,
they need to know what the distance is (how far away is the target object?),
what the viewing angle is (“Am I looking at the shape straight on or at an
angle?”), and what the illumination is. Otherwise, they’d have no way to take
these factors into account and, therefore, no way to achieve constancy.
Let’s pursue this issue by asking how people judge distance. We’ve just
said that distance perception is crucial for size constancy, but, of course,
information about where things are in your world is also valuable for its own
sake. If you want to walk down a hallway without bumping into obstacles,
you need to know which obstacles are close to you and which ones are far
off. If you wish to caress a loved one, you need to know where he or she is;
92 •
C H A P T E R T H R E E Visual Perception
otherwise, you’re likely to swat empty space when you reach out with your
caress or (worse) poke him or her in the eye. Plainly, then, you need to know
where objects in your world are located.
Binocular Cues
The perception of distance depends on various distance cues — features of the
stimulus that indicate an object’s position. One cue comes from the fact that
your eyes look out on the world from slightly different positions; as a result,
each eye has a slightly different view. This difference between the two eyes’
views is called binocular disparity, and it provides important information
about distance relationships in the world.
Binocular disparity can lead to the perception of depth even when no other
distance cues are present. For example, the bottom panels of Figure 3.23
show the views that each eye would receive while looking at a pair of nearby
objects. If we present each of these views to the appropriate eye (e.g., by drawing the views on two cards and placing one card in front of each eye), we can
obtain a striking impression of depth.
Monocular Cues
Binocular disparity is a powerful determinant of perceived depth. But we can
also perceive depth with one eye closed; plainly, then, there are also depth cues
that depend only on what each eye sees by itself. These are the monocular
distance cues.
B
A
B
A
A
B
Left eye's view
A B
Right eye's view
A
B
FIGURE 3.23
BINOCULAR DISPARITY
Your two eyes look out on the world from slightly different positions, and therefore they get slightly different views. The visual
system uses this difference in views as a cue to distance. This
figure shows what the left eye’s and right eye’s views would be
in looking at objects A and B.
The Perception of Depth
•
93
One monocular cue depends on the adjustment that the eye must make in
order to see the world clearly. We mentioned earlier that in each eye, muscles
adjust the shape of the lens to produce a sharply focused image on the retina.
The amount of adjustment depends on how far away the viewed object
is — there’s a lot of adjustment for nearby objects, less for those a few steps
away, and virtually no adjustment at all for objects more than a few meters
away. It turns out that perceivers are sensitive to the amount of adjustment
and use it as a cue indicating how far away the object is.
Other monocular cues have been exploited by artists for centuries to
create an impression of depth on a flat surface — that is, within a picture — and
that’s why these cues are called pictorial cues. In each case, these cues rely
on straightforward principles of physics. For example, imagine a situation
in which a man is trying to admire a sports car, but a mailbox is in the way
(see Figure 3.24A). In this case, the mailbox will inevitably block the view
simply because light can’t travel through an opaque object. This fact about
the physical world provides a cue you can use in judging distance. The cue is
known as interposition — the blocking of your view of one object by some
other object. In Figure 3.24B, interposition tells the man that the mailbox is
closer than the car.
In the same way, distant objects produce a smaller retinal image than do
nearby objects of the same size; this is a fact about optics. But this physical
fact again provides perceptual information you can use. In particular, it’s the
FIGURE 3.24
A
INTERPOSITION AS A DEPTH CUE
B
This man is looking at the sports car, but the mailbox blocks part of his view
(Panel A). Here’s how the scene looks from the man’s point of view (Panel B).
Because the mailbox blocks the view, the man gets a simple but powerful cue
that the mailbox must be closer to him than the sports car is.
94 •
C H A P T E R T H R E E Visual Perception
FIGURE 3.25
A
E FFECT OF CHANGES IN TEXTURE
GRADIENT
B
Changes in texture provide important information about spatial arrangements
in the world. Examples here show (Panel A) an upward tilt and (Panel B) a
sudden drop.
basis for the cue of linear perspective, the name for the pattern in which parallel lines seem to converge as they get farther and farther from the viewer.
A related cue is provided by texture gradients. Consider what meets your
eye when you look at cobblestones on a street or patterns of sand on a beach.
The retinal projection of the sand or cobblestones shows a pattern of continuous change in which the elements of the texture grow smaller and smaller
as they become more distant. This pattern of change by itself can reveal the
spatial layout of the relevant surfaces. If, in addition, there are discontinuities
in these textures, they can tell you even more about how the surfaces are laid
out (see Figure 3.25; Gibson, 1950, 1966).
The Perception of Depth through Motion
Whenever you move your head, the images projected by objects in your view
move across your retinas. For reasons of geometry, the projected images of
nearby objects move more than those of distant ones, and this pattern of
motion in the retinal images gives you another distance cue, called motion
parallax (Helmholtz, 1909).
A different cue relies on the fact that the pattern of stimulation across the
entire visual field changes as you move forward. This change in the visual
input — termed optic flow — provides another type of information about
depth and plays a large role in the coordination of bodily movements (Gibson,
1950, 1979).
The Perception of Depth
•
95
COGNITION
outside the lab
Virtual Reality
We obviously move around in a three-dimensional
(invented by a man whose son went on to be a
world. For centuries, though, people have been try-
Supreme Court Justice for thirty years!). This wooden
ing to create an illusion of 3-D with displays that are
device (Panel A in the figure below) allows the pre-
actually flat. Painters during the Renaissance, for ex-
sentation of a pair of pictures, one to each eye. The
ample, developed great skill in the use of the “picto-
two pictures show the same scene but viewed from
rial cues” (including visual perspective) to create a
slightly different vantage points, and these “ste-
sense of depth on a flat canvas. In the extreme, the
reoviews” produce a compelling sense of depth.
art technique of trompe l’oeil (French for “deceive
The same principle — and your capacity for “ste-
the eye”) could leave people truly puzzled about
reovision” — is used with the “virtual reality” (VR)
whether an object was painted or actually present.
accessory that works with many smartphones. The
Panel C in the figure on the following page shows
accessory, often made of cardboard, places a lens in
a modern version — created by a talented sidewalk
front of each eye so that you’ll be comfortable point-
artist.
ing your eyes straight ahead (as if you were looking
A different technique relies on binocular (“two-
at something far away), even though you’re actually
eyed”) vision. Consider the Holmes stereoscope
looking at an image just an inch or so away. With this
A
B
CLASSICAL USES OF BINOCULAR DISPARITY
Binocular disparity was the principle behind the stereoscope (Panel A), a device popular in the 19th century that
presented a slightly different photograph to each eye, creating a vivid sense of depth. The ViewMaster (Panel B),
a popular children’s toy, works in exactly the same way. The photos on the wheel are actually in pairs — and so, at
any rotation, the left eye views one photo in the pair (the one at 9 o’clock on the wheel) and the right eye views a
slightly different photo (the one at 3 o’clock), one that shows the same scene from a slightly different angle. Again,
the result is a powerful sense of depth.
96 •
C H A P T E R T H R E E Visual Perception
C
D
MODERN SIMULATIONS OF 3-D
Panel C shows a chalk drawing on a flat (and entirely undamaged) sidewalk. By manipulating pictorial cues, though,
the artist creates a compelling illusion of depth—with a car collapsed into a pit that in truth isn’t there at all.
Panel D shows one of the devices used to turn a smartphone into a “virtual reality” viewer.
setup, your phone displays two views of the same
viewers wear eyeglasses that contain corresponding
scene (one view to each eye), viewed from slightly
filters. The eyeglass filters “pass” light that’s polar-
different angles. Your eyes “fuse” these inputs into a
ized in a way that matches the filter, and block light
single image, but that doesn’t mean you ignore the
that’s polarized differently. As a result, each eye sees
differences between the inputs. Instead, your visual
only one of the projected movies — and, again, view-
system is solving the geometric puzzle posed by the
ers fuse the images but use the binocular disparity
two inputs. In other words, your brain manages to
to produce the experience of depth.
figure out how the scene must have been arranged
If you’ve enjoyed a 3-D movie or a smartphone
in order to produce these two different views, and
VR system, you’ve seen that the sense of depth is
it’s the end product of this computation that you
quite compelling. But these systems don’t work for
experience as a three-dimensional scene.
everyone. Some people have a strong pattern of
Three-D movies work the same way. There are
“ocular dominance,” which means that they rely on
actually two separate movies projected onto the
one eye far more than on the other. For these people,
theater’s screen. In some cases, the movies were
binocular disparity (which depends on combining
recorded from slightly different positions; in other
the inputs from both eyes) loses its force. However,
cases, the two perspectives were computer gen-
these people can still draw depth information from
erated. In either situation, the separate movies are
other (monocular or motion-based) cues, and so
projected through filters that polarize the light and
they can enjoy the same movies as anyone else.
The Perception of Depth
•
97
LINEAR PERSPECTIVE
AS A CUE FOR DEPTH
The Role of Redundancy
TEST YOURSELF
11. W
hat are the mono­
cular cues to distance?
12. Why is it helpful that
people rely on several
different cues in judging distance?
98 •
One might think that the various distance cues all end up providing the same
information — each one tells you which objects are close by and which ones
are distant. On that basis, it might be efficient for the visual system to focus
on just one or two cues and ignore the others. The fact is, however, that you
use all these cues, as well as several others we haven’t described (e.g., see
Figure 3.26).
Why is our visual system influenced by so many cues, especially since
these cues do, in fact, often provide redundant information? It’s because
different distance cues become important in different circumstances. For
example, binocular disparity is a powerful cue, but it’s informative only
when objects are relatively close by. (For targets farther than 30 ft away, the
two eyes receive virtually the same image.) Likewise, motion parallax tells
you a great deal about the spatial layout of your world, but only if you’re
moving. Texture gradients are informative only if there’s a suitably uniform
texture in view. So while these various cues are often redundant, each type
of cue can provide information when the others cannot. By being sensitive to them all, you’re able to judge distance in nearly any situation you
encounter. This turns out to be a consistent theme of perception — with
multiple cues to distance, multiple cues to illumination, multiple paths
through which to detect motion, and so on. The result is a system that
sometimes seems inelegant and inefficient, but it’s one that guarantees flexibility and versatility.
C H A P T E R T H R E E Visual Perception
FIGURE 3.26
M
ONOCULAR CLUES TO DEPTH:
LIGHT AND SHADOW
A
B
In this chapter, we’ve covered only a subset of the cues to distance that are
used by our visual system. Another cue is provided by the shadows “attached”
to an object. In Panel A, most viewers will say that the figure contains
six “bulges” in a smiley-face configuration (two eyes, a nose, a mouth).
In Panel B, the same figure has been turned upside-down. Now, the bulges
appear to be “dents,” and the other circles that appeared concave in the
Panel A view now look like bulges. The reason is the location of the shadows.
When the shadow is at the bottom, the object looks convex — a point that
makes sense because in our day-to-day lives light almost always comes from
above us, not below.
COGNITIVE PSYCHOLOGY AND EDUCATION
an “educated eye”
In the courtroom, eyewitnesses are often asked to describe what they saw
at a crime scene, and asked if they can identify the person who committed
the crime. Judges and juries generally rely on this testimony and accept the
witness’s report as an accurate description of how things unfolded. Judges
and juries are, however, especially likely to accept the witness’s report as
accurate if the witness is a police officer. In support of this position, some
attorneys argue that police officers have “educated eyes,” with the result
that police can (for example) recognize faces that they viewed only briefly or
at a considerable distance. In one trial, a police officer even claimed that
thanks to years of working a late-night shift, he’d improved his ability to
see in the dark.
Cognitive Psychology and Education
•
99
Related ideas arise in other settings. In Chapter 4, we’ll discuss programs
that teach you how to “speed-read,” but for now let’s just note that some of
these programs make a strong claim in their advertising: They claim that they
train your eyes so that you can “see more in a single glance.”
At one level, these claims are nonsensical. How much you can see “in
a single glance” is limited by your visual acuity, and acuity is limited by the
optical properties of the eyeball and the functional properties of the photoreceptors. To see more “in a single glance,” we’d need to give you a new cornea,
a new lens, and a new retina — and, of course, no speed-reading program
offers that sort of transplant surgery. Likewise, your ability to see in the dark
is constrained by the biological properties of the eye (including the structure
of the photoreceptors and the chemical principles that govern the photoreceptors’ response to light). No experience, and no training, is going to change
these properties.
At a different level, though, it is possible to have an “educated eye”— or,
more precisely, to be more observant and more discerning than other people.
For example, when looking at a complex, fast-moving crime scene, police
officers are more likely to focus their attention on details that will matter for the
investigation — and so will likely see (and remember) more of the perpetrator’s
actions (although, ironically, this means they’ll see less of what’s happening elsewhere in the scene). In the same way, referees and umpires in professional sports
know exactly what to focus on during a game. (Did the player’s knee touch the
ground before he fumbled the ball? Did the basketball player drag her pivot
foot or not?) As a result, they’ll see things that ordinary observers would miss.
These advantages (for police officers or for referees) may seem obvious,
but the advantages are closely tied to points raised in the chapter. You are
able to see detail only for visual inputs landing on your foveas; what lands
on your foveas depends on where exactly you’re pointing your eyes; and
movements of the eyes (pointing them first here and then there) turn out to be
relatively slow. As a result, knowledge about where to look has an immense
impact on what you’ll be able to see.
It’s also true that experience can help you to see certain patterns that
you’d otherwise miss. In some cases, the experience helps you to stop looking at a visual input on a feature-by-feature basis, but instead to take a more
“global” perspective so that you look at the pattern overall. Expert chess
players, for example, seem to perceive a chess board in terms of the patterns
in place (patterns indicating an upcoming attack, or patterns revealing the
opponent’s overall strategy), and this perspective helps them to plan their
own moves. (For more on chess experts, see Chapter 13.) Or, as a very different example, consider the dog experts who serve as judges at the Westminster
Kennel Club Dog Show. Evidence suggests that these experts are sensitive to
each dog’s overall form, and not just the shape of the front legs, the chest, the
ears, and so on, with the result that they can make more discerning assessments than an ordinary dog-lover could.
100 •
C H A P T E R T H R E E Visual Perception
KNOWING WHERE TO LOOK
Referees in football games know exactly where to look in order to pick up the
information they need in making their judgments. Did the player tap both feet
on the ground before going out of bounds? Did the player’s knee touch the
ground before he fumbled the ball? Referees seem to have an “educated eye,”
but, in reality, their advantage comes from how (and where) they focus their
attention.
Experience can also help you to see (or hear or feel) certain combinations that are especially important or informative. One prominent example
involves experienced firefighters who sometimes have an eerie ability to
judge when a floor is about to collapse — allowing these professionals to
evacuate a building in time, saving their lives and others’. What explains
this perception? The answer may be a combination of feeling an especially
high temperature and hearing relative quiet — a combination that signals a
ferocious fire burning underneath them, hidden under the floor that they’re
standing on.
In short, then, people can have “educated eyes” (or ears or noses or
palates). This “education” can’t change the basic biological properties of
your sense organs. But knowledge and experience can certainly help you
to see things that others overlook, to detect patterns that are largely invisible to other people, and to pick up on combinations that can — in some
settings — save your life.
Cognitive Psychology and Education
•
101
For more on this topic . . .
Biederman, I., & Shiffrar, M. M. (1987). Sexing day-old chicks: A case study and
expert systems analysis of a difficult perceptual learning task. Journal of
Experimental Psychology: Learning, Memory & Cognition, 13, 640–645.
Diamond, R., & Carey, S. (1986). Why faces are and are not special: An effect
of expertise. Journal of Experimental Psychology: General, 115, 107–117.
Klein, R. (2013). Seeing what others don’t. New York, NY: Public Affairs.
Vredeveldt, A., Knol, J. W., & van Kopen, P. J. (2017). Observing offenders:
Incident reports by surveillance detectives, uniformed police, and civilians.
Legal and Criminological Psychology, 22, 150–163.
102 •
C H A P T E R T H R E E Visual Perception
chapter review
SUMMARY
• One brain area that has been mapped in considerable detail is the visual system. This system takes
its main input from the rods and cones on the retina.
Then, information is sent via the optic nerve to the
brain. An important point is that cells in the optic
nerve do much more than transmit information;
they also begin the analysis of the visual input. This
is reflected in the phenomenon of lateral inhibition,
which leads to edge enhancement.
• Part of what we know about the brain comes
from single-cell recording, which can record the
electrical activity of an individual neuron. In the
visual system, this recording has allowed researchers to map the receptive fields for many cells. The
mapping has provided evidence for a high degree of
specialization among the various parts of the visual
system, with some parts specialized for the perception of motion, others for the perception of color,
and so on. The various areas function in parallel, and
this parallel processing allows great speed. It also
allows mutual influence among multiple systems.
• Parallel processing begins in the optic nerve and
continues throughout the visual system. For example, the what system (in the temporal lobe) appears
to be specialized for the identification of visual
objects; the where system (in the parietal lobe)
seems to identify where an object is located.
• The reliance on parallel processing creates
a problem of reuniting the various elements of a
scene so that these elements are perceived in an
integrated way. This is the binding problem. One
key in solving this problem lies in the fact that
different brain systems are organized in terms of
maps, so that spatial position can be used as a
framework for reuniting the separately analyzed
aspects of the visual scene.
• Visual perception requires more than the “pickup” of features. Those features must be organized
into wholes — a process apparently governed by the
so-called Gestalt principles. The visual system also
must interpret the input, a point that is especially
evident with reversible figures. Crucially, though,
these interpretive steps aren’t separate from, and
occurring after, the pickup of elementary features,
because the features themselves are shaped by the
perceiver’s organization of the input.
• The active nature of perception is also evident in
perceptual constancy. We achieve constancy through
a process of unconscious inference, taking one
aspect of the input (e.g., the distance to the target)
into account in interpreting another aspect (e.g., the
target’s size). This process is usually quite accurate,
but it can produce illusions.
• The perception of distance relies on many
cues — some dependent on binocular vision,
and some on monocular vision. The diversity of
cues lets us perceive distance in a wide range of
circumstances.
KEY TERMS
cornea (p. 65)
lens (p. 65)
retina (p. 65)
photoreceptors (p. 65)
rods (p. 65)
cones (p. 66)
acuity (p. 66)
fovea (p. 67)
103
bipolar cells (p. 68)
ganglion cells (p. 68)
optic nerve (p. 68)
lateral geniculate nucleus (LGN) (p. 68)
lateral inhibition (p. 68)
edge enhancement (p. 69)
Mach band (p. 70)
single-cell recording (p. 71)
receptive field (p. 71)
center-surround cells (p. 72)
Area V1 (p. 74)
parallel processing (p. 75)
serial processing (p. 76)
P cells (p. 76)
M cells (p. 76)
parvocellular cells (p. 76)
magnocellular cells (p. 76)
what system (p. 77)
where system (p. 77)
binding problem (p. 78)
neural synchrony (p. 79)
conjunction errors (p. 80)
Necker cube (p. 81)
reversible figure (p. 81)
figure/ground organization (p. 82)
Gestalt principles (p. 84)
visual features (p. 85)
perceptual constancy (p. 87)
size constancy (p. 87)
shape constancy (p. 87)
brightness constancy (p. 87)
unconscious inference (p. 89)
distance cues (p. 93)
binocular disparity (p. 93)
monocular distance cues (p. 93)
pictorial cues (p. 94)
interposition (p. 94)
linear perspective (p. 95)
motion parallax (p. 95)
optic flow (p. 95)
TEST YOURSELF AGAIN
1.What are the differences between rods and
cones? What traits do these cells share?
7.What are the Gestalt principles, and how do
they influence visual perception?
2.What is lateral inhibition? How does it
contribute to edge perception?
3.How do researchers use single-cell recording to
reveal a cell’s receptive field?
8.What evidence is there that the perception of
an overall form depends on the detection of
features? What evidence is there that the detection of features depends on the overall form?
4.What are the advantages of parallel
processing in the visual system? What are
the disadvantages?
9.What does it mean to say that size constancy
may depend on an unconscious inference? An
inference about what?
5.How is firing synchrony relevant to the solution
of the binding problem?
10.How do the ordinary mechanisms of constancy
lead to visual illusions?
6.What evidence tells us that perception goes
beyond (i.e., includes more information than)
the stimulus input?
11. What are the monocular cues to distance?
104
12.Why is it helpful that people rely on several
different cues in judging distance?
THINK ABOUT IT
1.The chapter emphasizes the active nature of
perception — and the idea that we don’t just
“pick up” information from the environment;
instead, we interpret and supplement that information. What examples of this pattern can you
think of — either from the chapter or from your
own experience?
2.Chapter 2 argued that the functioning of the
brain depends on the coordination of many
specialized operations. How does that claim,
about the brain in general, fit with the discussion of visual perception in this chapter?
E eBOOK DEMONSTRATIONS & ESSAYS
Go to http://digital.wwnorton.com/cognition7 for the online demonstrations and essays
relevant to this chapter. These can be found in the ebook.
Online Demonstrations
• Demonstration 3.1: Foveation
• Demonstration 3.2: Eye Movements
• Demonstration 3.3: The Blind Spot and the
• Demonstration 3.4: A Brightness Illusion
• Demonstration 3.5: A Size Illusion and a Motion
Illusion
Active Nature of Vision
COGNITION LABS
Go to http://digital.wwnorton.com/cognition7 to experience interactive online psychology
labs relevant to this chapter.
105
4
chapter
Recognizing
Objects
what if…
In Chapter 3, we discussed some of the steps involved
in visual perception — steps allowing you to see that
the object in front of you is, let’s say, brown, large, and moving. But
you don’t leave things there; you also recognize objects and can identify
what they are (perhaps: a UPS truck). This sort of recognition is usually
easy for you, so you have no difficulty in recognizing the vast array of
objects in your world — trucks, squirrels, shoes, frying pans, and more.
But, easy or not, recognition relies on processes that are surprisingly
sophisticated, and your life would be massively disrupted if you couldn’t
manage this (seemingly simple) achievement.
We mentioned in Chapter 2 that certain types of brain damage
produce a disorder called “agnosia.” In some cases, patients suffer from
apperceptive agnosia — they seem able to see an object’s shape and color
and position, but they can’t put these elements together to perceive
the entire object. For example, one patient — identified as D.F. — suffered
from brain damage in the sites shown in Figure 4.1. D.F. was asked
to copy drawings that were in plain view (Figure 4.2A). The resulting
attempts are shown in Figure 4.2B. The limit here is not some problem
in drawing ability. Figure 4.2C shows what happened when D.F. was
asked to draw various forms from memory. Plainly, D.F. can draw; the
problem instead is in her ability to see and assemble the various
elements that she sees.
Other patients suffer from associative agnosia. They can see but cannot link what they see to their basic visual knowledge. One remarkable
example comes from a case described by neurologist Oliver Sacks:
“What is this?” I asked, holding up a glove.
“May I examine it?” he asked, and, taking it from me, he proceeded to
examine it. “A continuous surface,” he announced at last, “infolded
in itself. It appears to have” — he hesitated — ”five outpouchings, if
this is the word.”
“Yes,” I said cautiously. “. . .Now tell me what it is.”
“A container of some sort?”
“Yes,” I said, “and what would it contain?”
“It would contain its contents!” said Dr. P., with a laugh. “There are
many possibilities. It could be a change purse, for example, for
coins of five sizes. It could . . .” (Sacks, 1985, p. 14)
107
preview of chapter themes
•
ecognition of visual inputs begins with features, but it’s not
R
just the features that matter. How easily people recognize a
pattern also depends on how frequently or recently they have
viewed the pattern and on whether the pattern is well formed
(such as letter sequences with “normal” spelling patterns).
•
e explain these findings in terms of a feature net — a netW
work of detectors, each of which is “primed” according to
how often or how recently it has fired. The network relies
on distributed knowledge to make inferences, and this
process gives up some accuracy in order to gain efficiency.
FIGURE 4.1
•
he feature net can be extended to other domains, includT
ing the recognition of three-dimensional objects. However,
the recognition of faces requires a different sort of model,
sensitive to configurations rather than to parts.
•
inally, we consider top-down influences on recognition.
F
The existence of these influences tells us that object recognition is not a self-contained process. Instead, knowledge external to object recognition is imported into and
clearly shapes the process.
D.F.’S LESIONS
A Lesions in subject D.F.
B Location of LOC in neurologically intact subjects
Panel A shows the location of the brain damage in D.F. Panel B shows the areas in the lateral occipital complex
(LOC) that are especially activated when neurologically healthy people are recognizing objects.
108 •
C H A P T E R F O U R Recognizing Objects
FIGURE 4.2
DRAWINGS FROM PATIENT D.F.
Line-drawing models
A
Drawn from the models
B
Drawn from memory
C
Patients who suffer from apperceptive agnosia can see, but they can’t organize the elements they see in order
to perceive an entire object. This deficit was evident when patient D.F. was asked to copy the drawings shown in
Panel A. Her attempts are shown in Panel B. The problem is not in her drawing ability, because D.F.’s performance
was much better (as shown in Panel C) when she was asked to draw the same forms from memory, rather than
from a model.
Dr. P. obviously can see, and he uses his (considerable) intelligence
to figure out what he is seeing. Nonetheless, his agnosia profoundly
disrupts his life. Sacks describes one incident in which Dr. P. failed to
put on his shoe, because he didn’t recognize it as a shoe. (In fact, Sacks
notes that at one point Dr. P. was confused about which object was his
shoe and which was his foot.) Then, at the end of their time together,
Sacks reports that Dr. P. “reached out his hand and took hold of his wife’s
head, tried to lift it off, to put it on. He had apparently mistaken his wife
for a hat!” (p. 11).
As these examples make clear, object recognition may not be a glamorous skill, but it is one that we all rely on for even our most ordinary
interactions with the world. What are the processes that make object
recognition possible?
Recognizing Objects
•
109
Recognition: Some Early Considerations
You’re obviously able to recognize a huge number of different patterns —
different objects (cats, cups, coats), various actions (crawling, climbing,
clapping), and different sorts of situations (crises, comedies). You can
also recognize many variations of each of these things. You recognize
cats standing up and cats sitting down, cats running and cats asleep.
And the same is true for recognition of most other patterns in your recog­
nition repertoire.
You also recognize objects even when the available information is
incomplete. For example, you can still recognize a cat if only its head
and one paw are visible behind a tree. You recognize a chair even when
someone is sitting on it, even though the person blocks much of the chair
from view.
All of this is true for print as well. You can recognize tens of thousands of
words, and you can recognize them whether the words are printed in large
type or small, italics or straight letters, UPPER CASE or lower. You can even
recognize handwritten words, for which the variation from one to the next
is huge.
These variations in the “stimulus input” provide our first indication
that object recognition involves some complexity. Another indication
comes from the fact that your recognition of various objects, print or other­
wise, is influenced by the context in which you encounter those objects.
Consider Figure 4.3. The middle character is the same in both words, but
the character looks more like an H in the word on the left and more like
an A in the word on the right. With this, you easily read the word on
the left as “THE” and not “TAE” and the word on the right as “CAT” and
not “CHT.”
Of course, object recognition is powerfully influenced by the stimulus
itself — that is, by the features that are in view. Processes directly shaped by
the stimulus are sometimes called “data driven” but are more commonly said
FIGURE 4.3
CONTEXT INFLUENCES PERCEPTION
You are likely to easily read this sequence as “THE CAT,” recognizing the middle
symbol as an H in one case and as an A in the other.
( after selfridge , 1955)
110 •
C H A P T E R F O U R Recognizing Objects
to involve bottom-up processing. The effect of context, however, reminds us
that recognition is also influenced by one’s knowledge and expectations. As a
result, your reading of Figure 4.3 is guided by your knowledge that “THE” and
“CAT” are common words but that “TAE” and “CHT” are not. This sort of
influence — relying on your knowledge — is sometimes called “concept-driven,”
and processes shaped by knowledge are said to involve top-down processing.
What mechanism underlies both the top-down and bottom-up influences?
In the next section, we’ll consider a classic proposal for what the mechanism
might be. We’ll then build on this base as we discuss more recent elaborations
of this proposal.
The Importance of Features
Common sense suggests that many objects can be recognized by virtue of their
parts. You recognize an elephant because you see the trunk, the thick legs, the
large body. You know a lollipop is a lollipop because you see the circle shape
on top of the straight stick. But how do you recognize the parts themselves?
How, for example, do you recognize the trunk on the elephant or the circle in
the lollipop? The answer may be simple: Perhaps you recognize the parts by
looking at their parts — such as the arcs that make up the circle in the lollipop,
or the (roughly) parallel lines that identify the elephant’s trunk.
To put this more generally, recognition might begin with the identification
of visual features in the input pattern — the vertical lines, curves, diagonals,
and so on. With these features appropriately catalogued, you can start assem­
bling the larger units. If you detect a horizontal together with a vertical, you
know you’re looking at a right angle; if you’ve detected four right angles,
you know you’re looking at a square.
This broad proposal lines up well with the neuroscience evidence we dis­
cussed in Chapter 3. There, we saw that specialized cells in the visual system
do seem to act as feature detectors, firing (producing an action potential)
whenever the relevant input (i.e., the appropriate feature) is in view. Also,
we’ve already noted that people can recognize many variations on the objects
they encounter — cats in different positions, A’s in different fonts or different
handwritings. An emphasis on features, though, might help with this point.
The various A’s, for example, differ from one another in overall shape, but
they do have certain things in common: two inwardly sloping lines and a
horizontal crossbar. Focusing on features, therefore, might allow us to con­
centrate on elements shared by the various A’s and so might allow us to
recognize A’s despite their apparent diversity.
The importance of features is also evident in data from visual search
tasks — tasks in which participants are asked to examine a display and to
judge whether a particular target is present in the display or not. This search
is remarkably efficient when someone is searching for a target defined by a
simple feature — for example, finding a vertical segment in a field of horizon­
tals or a green shape in a field of red shapes. But people are generally slower
THE VARIABILITY
OF STIMULI WE
RECOGNIZE
We recognize cats from the
side or the front, whether we
see them close up or far away.
Recognition: Some Early Considerations
•
111
FIGURE 4.4
VISUAL SEARCH
A
B
C
In Panel A, you can immediately spot the vertical, distinguished from the other shapes by just one feature. Likewise, in Panel B, you can immediately spot the lone green bar in the field of reds. But in Panel C, it takes longer
to find the one red vertical, because now you need to search for a combination of features — not just for red or
vertical, but for the one form that has both of these attributes.
TEST YOURSELF
1.What is the difference
between “bottomup” and “top-down”
processing?
2.What is the evidence
that features play a
special role in object
recognition?
in searching for a target defined as a combination of features (see Figure 4.4).
This is just what we would expect if feature analysis is an early step in your
analysis of the visual world — and separate from the step in which you com­
bine the features you’ve detected.
Further support for these claims comes from studies of brain damage.
At the start of the chapter, we mentioned apperceptive agnosia — a disorder
that involves an inability to assemble the various aspects of an input into an
organized whole. A related disorder, integrative agnosia, derives from damage
to the parietal lobe. Patients with this disorder appear relatively normal in
tasks requiring them simply to detect features in a display, but they are mark­
edly impaired in tasks that require them to judge how the features are bound
together to form complex objects. (See, for example, Behrmann, Peterson,
Moscovitch, & Suzuki, 2006; Humphreys & Riddoch, 2014; Robertson, Tre­
isman, Friedman-Hill, & Grabowecky, 1997. For related results, in which
transcranial magnetic stimulation was used to disrupt portions of the brain
in healthy individuals, see Ashbridge, Walsh, & Cowey, 1997.)
Word Recognition
Several lines of evidence, therefore, indicate that object recognition does
begin with the detection of simple features. Then, once this detection has
occurred, separate mechanisms are needed to put the features together,
assembling them into complete objects. But how does this assembly proceed,
so that we end up seeing not just the features but whole words — or Chihuahuas,
or fire hydrants? In tackling this question, it will be helpful to fill in some
more facts that we can then use as a guide to our theory building.
112 •
C H A P T E R F O U R Recognizing Objects
Factors Influencing Recognition
In many studies, participants have been shown stimuli for just a brief duration —
perhaps 20 or 30 ms (milliseconds). Older research did this by means of a
tachistoscope, a device designed to present stimuli for precisely controlled
amounts of time. More modern research uses computers, but the brief displays
are still called “tachistoscopic presentations.”
Each stimulus is followed by a post-stimulus mask — often, a random pat­
tern of lines and curves, or a random jumble of letters such as “XJDKEL.” The
mask interrupts any continued processing that participants might try to do
for the stimulus just presented. In this way, researchers can be certain that a
stimulus presented for (say) 20 ms is visible for exactly 20 ms and no longer.
Can people recognize these briefly visible stimuli? The answer depends on
many factors, including how familiar a stimulus is. If the stimulus is a word, we
can measure familiarity by counting how often that word appears in print, and
these counts are an excellent predictor of tachistoscopic recognition. In one early
experiment, Jacoby and Dallas (1981) showed participants words that were
either very frequent (appearing at least 50 times in every million printed words)
or infrequent (occurring only 1 to 5 times per million words of print). Participants
viewed these words for 35 ms, followed by a mask. Under these circumstances,
they recognized almost twice as many of the frequent words (see Figure 4.5A).
WORD FREQUENCY’S EFFECT ON WORD RECOGNITION
100
100
90
90
80
80
70
70
Percent Recognition
Percent Recognition
FIGURE 4.5
60
50
40
30
50
40
30
20
20
10
10
High-frequency
words
A
60
Low-frequency
words
B
Unprimed Primed
High-frequency
words
Unprimed Primed
Low-frequency
words
In one study, recognition was much more likely for words appearing often in print, in comparison to words appearing only rarely — an effect of frequency (Panel A). Similarly, words that had been viewed recently were more often
recognized, an effect of recency that in this case creates a benefit called “repetition priming” (Panel B).
( after jacoby & dallas , 1981)
Word Recognition
•
113
Another factor influencing recognition is recency of view. If partici­
pants view a word and then, a little later, view it again, they will recognize
the word more readily the second time around. The first exposure primes
the participant for the second exposure; more specifically, this is a case of
repetition priming.
As an example, participants in one study read a list of words aloud. The
participants were then shown a series of words in a tachistoscope. Some
of these words were from the earlier list and so had been primed; others
were unprimed. For words that were high in frequency, 68% of the unprimed
words were recognized, compared to 84% of the primed words. For words
low in frequency, 37% of the unprimed words were recognized, compared to
73% of the primed words (see Figure 4.5B; Jacoby & Dallas, 1981).
The Word-Superiority Effect
Figure 4.3 suggests that the recognition of a letter depends on its context —
and so an ambiguous letter is read as an A in one setting but an H in another
setting. But context also has another effect: Even when a letter is properly
printed and quite unambiguous, it’s easier to recognize if it appears within a
word than if it appears in isolation.
This result might seem paradoxical, because here we have a setting in
which it seems easier to do “more work” rather than “less” — and so you’re
more accurate in recognizing all the letters that make up a word (maybe
a total of five or six letters) than you are in recognizing just one letter on
its own. Paradoxical or not, this pattern is easy to demonstrate, and the
advantage for perceiving letters-in-context is called the word-superiority
effect (WSE).
The WSE is demonstrated with a “two-alternative, forced-choice” pro­
cedure. For example, in some trials we might present a single letter — let’s
say K — followed by a post-stimulus mask, and follow that with a question:
“Which of these was in the display: an E or a K?” In other trials, we might
present a word — let’s say “DARK” — followed by a mask, followed by a
question: “Which of these was in the display: an E or a K?”
Note that participants have a 50-50 chance of guessing correctly in either
of these situations, and so any contribution from guessing is the same for the
letters as it is for the words. Also, for the word stimulus, both of the letters
we’ve asked about are plausible endings for the stimulus; either ending would
create a common word (“DARE” or “DARK”). Therefore, participants who
saw only part of the display (perhaps “DAR”) couldn’t use their knowledge
of the language to figure out the display’s final letter. In order to choose
between E and K, therefore, participants really need to have seen the relevant
letter — and that is exactly what we want.
In this procedure, accuracy rates are reliably higher in the word condi­
tion. Apparently, recognizing an entire word is easier than recognizing iso­
lated letters (see Figure 4.6; Johnston & McClelland, 1973; Reicher, 1969;
Rumelhart & Siple, 1974; Wheeler, 1970).
114 •
C H A P T E R F O U R Recognizing Objects
Percent correct
80
70
FIGURE 4.6
60
50
Single letters Entire words
Stimulus type
THE WORD-SUPERIORITY EFFECT
The word-superiority effect is usually demonstrated with a twoalternative forced-choice procedure (which means that a participant
can get a score of 50% just by guessing randomly). Performance is
much better if the target letter is shown in context — within an entire
word — than if it is shown on its own.
( after johnston & m c clelland , 1973)
Degree of Well-Formedness
As it turns out, though, the term “word-superiority effect” may be mis­
leading, because we don’t need words to produce the pattern evident in
Figure 4.6. We get a similar effect if we present participants with letter
strings like “FIKE” or “LAFE.” These letter strings are not English words
and they’re not familiar, but they look like English strings and (related)
are easy to pronounce. And, crucially, strings like these produce a context
effect, with the result that letters in these contexts are easier to identify than
letters alone.
This effect occurs, though, only if the context is of the right sort. There’s
no context benefit if we present a string like “HZYE” or “SBNE.” An E pre­
sented within these strings will not show the word-superiority effect — that
is, it won’t be recognized more readily than an E presented in isolation.
A parallel set of findings emerge if, instead of asking participants to detect
specific letters, we ask them to report all of what they have seen. A letter
string like “HZYE” is extremely hard to recognize if presented briefly. With
a stimulus like this and, say, a 30-ms exposure, participants may report that
they only saw a flash and no letters at all; at best, they may report a letter or
two. But with the same 30-ms exposure, participants will generally recognize
(and be able to report) strings like “FIKE” or “LAFE,” although they do even
better if the stimuli presented are actual, familiar words.
How should we think about these findings? One approach emphasizes the
statistically defined regularities in English spelling. Specifically, we can work
through a dictionary, counting how often (for example) the letter combination
“FI” occurs, or the combination “LA,” or “HZ.” We can do the same for threeletter sequences (“FIK,” “LAF,” and so on). These counts will give us a tally that
reveals which letter combinations are more probable in English spelling and
Word Recognition
•
115
which are less probable. We can then use this tally to evaluate new strings —
asking, for any string, whether its letter sequences are high-probability ones
(occurring often) or low-probability (occurring rarely).
These statistical measures allow us to evaluate how “well formed” a letter
string is — that is, how well the letter sequence conforms to the usual spelling
patterns of English — and well-formedness is a good predictor of word rec­
ognition: The more English-like the string is, the easier it will be to recognize
that string, and also the greater the context benefit the string will produce.
This well-documented pattern has been known for more than a century (see,
e.g., Cattell, 1885) and has been replicated in many studies (Gibson, Bishop,
Schiff, & Smith, 1964; Miller, Bruner, & Postman, 1954).
Making Errors
TEST YOURSELF
3.What is repetition
priming, and how
is it demonstrated?
4. What procedure demonstrates the wordsuperiority effect?
5. What’s the evidence
that word perception
is somehow governed
by the rules of ordinary
spelling?
Let’s recap some important points. First, it seems that a letter will be easier
to recognize if it appears in a well-formed sequence, but not if it appears in
a random sequence. Second, well-formed strings are, overall, easier to per­
ceive than ill-formed strings; this advantage remains even if the well-formed
strings are made-up ones that you’ve never seen before (strings like “HAKE”
or “COTER”). All of these facts suggest that you somehow are using your
knowledge of spelling patterns when you look at, and recognize, the words
you encounter — and so you have an easier time with letter strings that con­
form to these patterns, compared to strings that do not.
The influence of spelling patterns is also evident in the mistakes you make.
With brief exposures, word recognition is good but not perfect, and the errors
that occur are systematic: There’s a strong tendency to misread less-common
letter sequences as if they were more-common patterns. So, for example,
“TPUM” is likely to be misread as “TRUM” or even “DRUM.” But the reverse
errors are rare: “DRUM” is unlikely to be misread as “TRUM” or “TPUM.”
These errors can sometimes be quite large — so that someone shown
“TPUM” might instead perceive “TRUMPET.” But, large or small, the errors
show the pattern described: Misspelled words, partial words, or nonwords
are read in a way that brings them into line with normal spelling. In effect,
people perceive the input as being more regular than it actually is. Once
again, therefore, our recognition seems to be guided by (or, in this case, mis­
guided by) some knowledge of spelling patterns.
Feature Nets and Word Recognition
What lies behind this broad pattern of evidence? What are the processes in­
side of us that lead to the findings we’ve described? Psychologists’ under­
standing of these points grows out of a theory published many years ago
(Selfridge, 1959). Let’s start with that theory, and then use it as our base
as we look at more modern work. (For a glimpse of some of the modern
research, including work that links theorizing to neuroscience, see Carreiras,
Armstrong, Perea, & Frost, 2014.)
116 •
C H A P T E R F O U R Recognizing Objects
COGNITION
outside the lab
Font
You encounter printed material in a variety of
the students read these facts in a clear font (Arial
formats and in a wide range of fonts. You also
printed in pure black), and half read the facts in less
come across hand-written material, and of course
clear font (e.g., Bodoni MT, printed in 60% grayscale).
people differ enormously in their handwriting.
When tested later, participants who’d seen the
Despite all this variety, you’re able to read almost
fluent print remembered 73% of the facts; par-
everything you see — somehow rising above the
ticipants who’d seen the less fluent print recalled
variations from one bit to the next.
86% of the facts. What was going on here? We’ll
These variations do matter, however. Some
see in Chapter 6 that memory is promoted by
people’s handwriting is an almost impenetrable
active engagement with the to-be-remembered
scrawl; some fonts are difficult to decipher. Even
materials, and it seems that the somewhat
if we step away from these extremes and only
obscure font promoted that sort of engagement —
consider cases in which you can figure out what’s
and so created what (in Chapter 6) we’ll refer to
on the page, poor handwriting or an obscure
as “desirable difficulty” in the learning process.
font can make your reading less fluent. How this
What about other aspects of formatting?
drop in fluency matters, though, depends on the
We’ve discussed the individual features that you
circumstances.
use in recognizing letters, but it turns out that
In one study, college students read a pas-
you’re also sensitive to a word’s overall shape. This
sage printed either in a clear font (Times New
is one of the reasons WHY IT IS MORE DIFFICULT
Roman) or in a difficult font (italicized Juice ITC).
TO READ CAPITALIZED TEXT. Capitalized words
Students in both groups were then asked to rate
all have the same rectangular shape; gone are the
the intelligence of the author who’d written the
portions of the letter that hang belong the line —
passage (Oppenheimer, 2006). Students who
the so-called descenders, like the bottom tail on a
read the less clear font rated the author as less
g or a j. Also gone are the portions of the letters
intelligent; apparently, they had noticed that
that stick up (ascenders), like the top of an h or
their reading wasn’t fluent but didn’t realize the
an l, or the dot over an i. Your reading slows down
problem was in the font. Instead, they decided
when these features aren’t available, so it’s slower
that the lack of fluency was the author’s fault:
IF YOU READ BLOCK CAPITALS compared to the
They decided that the author hadn’t been clear
normal pattern of print.
enough in composing the passage, and there-
Are there practical lessons here? In some
cases, you might prefer the look of block capitals;
fore was less intelligent!
an
but if so, be aware that this format slows reading a
advantage for a (slightly) obscure font (Diemond-
bit. In choosing a font, you should probably avoid
Yauman, Oppenheimer, & Vaughan, 2011). Col-
the obscure styles (unless you want less-fluent
lege students were asked to read made-up facts
reading!), but notice that a moderately challeng-
about space aliens — for example, that the Nor-
ing font can actually help readers to process and
gletti are 2 ft tall and eat flower petals. Half of
remember what you’ve written.
Another
experiment,
though,
showed
Feature Nets and Word Recognition
•
117
The Design of a Feature Net
Imagine that we want to design a system that will recognize the word
“CLOCK” whenever it is in view. How might our “CLOCK” detector work?
One option is to “wire” this detector to a C-detector, an L-detector, an
O-detector, and so on. Then, whenever these letter detectors are activated,
this would activate the word detector. But what activates the letter detectors?
Maybe the L-detector is “wired” to a horizontal-line detector and also a
vertical-line detector, as shown in Figure 4.7. When these feature detectors
are activated, this activates the letter detector.
The idea is that there could be a network of detectors, organized in layers.
The “bottom” layer is concerned with features, and that is why networks of
this sort are often called feature nets. As we move “upward” in the network,
each subsequent layer is concerned with larger-scale objects; using the term
we introduced earlier, the flow of information would be bottom-up — from
the lower levels toward the upper levels.
But what does it mean to “activate” a detector? At any point in time, each
detector in the network has a particular activation level, which reflects the
status of the detector at that moment — roughly, how energized the detec­
tor is. When a detector receives some input, its activation level increases. A
strong input will increase the activation level by a lot, and so will a series of
weaker inputs. In either case, the activation level will eventually reach the
detector’s response threshold, and at that point the detector will fire — that is,
send its signal to the other detectors to which it is connected.
FIGURE 4.7
A SIMPLE FEATURE NET
Word
detector
CLOCK
C
L
O
C
K
Letter
detectors
Feature
detectors
An example of a feature net. Here, the feature detectors respond to simple elements in the visual input. When the appropriate feature detectors are activated,
they trigger a response in the letter detectors. When these are activated, in
turn, they can trigger a response in a higher-level detector, such as a detector
for an entire word.
118 •
C H A P T E R F O U R Recognizing Objects
These points parallel our description of neurons in Chapter 2, and that’s
no accident. If the feature net is to be a serious candidate for how humans
recognize patterns, then it has to use the same sorts of building blocks that
the brain does. However, let’s be careful not to overstate this point: No one
is suggesting that detectors are neurons or even large groups of neurons.
Instead, detectors probably involve complex assemblies of neural tissue.
Nonetheless, it’s plainly attractive that the hypothesized detectors in the fea­
ture net function in a way that’s biologically sensible.
Within the net, some detectors will be easier to activate than others —
that is, some will require a strong input to make them fire, while oth­
ers will fire even with a weak input. This difference is created in part by
how activated each detector is to begin with. If the detector is moderately
activated at the start, then only a little input is needed to raise the activa­
tion level to threshold, and so it will be easy to make this detector fire. If
a detector is not at all activated at the start, then a strong input is needed
to bring the detector to threshold, and so it will be more difficult to make
this detector fire.
What determines a detector’s starting activation level? As one factor,
detectors that have fired recently will have a higher activation level (think
of it as a “warm-up” effect). In addition, detectors that have fired frequently
in the past will have a higher activation level (think of it as an “exercise”
effect). Overall, then, activation level is dependent on principles of recency
and frequency.
We now can put these mechanisms to work. Why are frequent words in
the language easier to recognize than rare words? Frequent words, by defini­
tion, appear often in the things you read. Therefore, the detectors needed for
recognizing these words have been frequently used, so they have relatively
high levels of activation. Thus, even a weak signal (e.g., a brief or dim presen­
tation of the word) will bring these detectors to their response threshold and
will be enough to make them fire. As a result, the word will be recognized
even with a degraded input.
Repetition priming is explained in similar terms. Presenting a word once
will cause the relevant detectors to fire. Once they’ve fired, activation levels
will be temporarily lifted (because of recency of use). Therefore, only a weak
signal will be needed to make the detectors fire again. As a result, the word
will be more easily recognized the second time around.
The Feature Net and Well-Formedness
The net we’ve described so far cannot, however, explain all of the data.
Consider the effects of well-formedness — for instance, the fact that people
are able to read letter strings like “PIRT” or “HICE” even when those strings
are presented very briefly (or dimly or in low contrast), but not strings like
“ITPR” or “HCEI.” How can we explain this finding? One option is to add
another layer to the net, a layer filled with detectors for letter combinations.
Feature Nets and Word Recognition
•
119
FIGURE 4.8
BIGRAM DETECTORS
Word
detector
Bigram
detectors
Letter
detectors
Feature
detectors
It seems plausible that the network includes a layer of bigram detectors
between the letter detectors and word detectors.
Thus, in Figure 4.8, we’ve added a layer of bigram detectors — detectors of
letter pairs. These detectors, like all the rest, will be triggered by lower-level
detectors and send their output to higher-level detectors. And just like any
other detector, each bigram detector will start out with a certain activation
level, influenced by the frequency with which the detector has fired in the past
and by the recency with which it has fired.
This turns out to be all the theory we need. You have never seen the
sequence “HICE” before, but you have seen the letter pair HI (in “HIT,”
“HIGH,” or “HILL”) and the pair CE (“FACE,” “MICE,” “JUICE”).
The detectors for these letter pairs, therefore, have high activation levels at
the start, so they don’t need much additional input to reach their thresh­
old. As a result, these detectors will fire with only weak input. That will
make the corresponding letter combinations easy to recognize, facilitating
the recognition of strings like “HICE.” None of this is true for “IJPV”
or “RSFK.” Because none of these letter combinations are familiar, these
strings will receive no benefits from priming. As a result, a strong input will
be needed to bring the relevant detectors to threshold, and so these strings
will be recognized only with difficulty. (For more on bigram detectors
and how they work, see Grainger, Rey, & Dufau, 2008; Grainger &
Whitney, 2004; Whitney, 2001. For some complications, see Rayner &
Pollatsek, 2011.)
120 •
C H A P T E R F O U R Recognizing Objects
Recovery from Confusion
Imagine that we present the word “CORN” for just 20 ms. In this setting, the
visual system has only a limited opportunity to analyze the input, so it’s possible that
you’ll miss some of the input’s features. For example, let’s imagine that the second
letter in this word — the O — is hard to see, so that only the bottom curve is detected.
This partial information invites confusion. If all you know is “the second
letter had a bottom curve,” then perhaps this letter was an O, or perhaps it
was a U, or a Q, or maybe an S. Figure 4.9 shows how this would play out
in terms of the network. We’ve already said that you detected the bottom
curve, and that means the “bottom-curve detector” is activated. This detec­
tor, in turn, provides input to the O-detector and also to the detectors for U,
FIGURE 4.9
THE VISUAL PROCESSING PATHWAYS
Word
detectors
Bigram
detectors
Letter
detectors
Feature
detectors
Stimulus
input
If “CORN” is presented briefly, not all of its features will be detected. Imagine
that only the bottom curve of the O is detected, not the O’s top or sides. This will
(weakly) activate the O-detector, but it will also activate the detectors of various
other letters having a bottom curve, including U, Q, and S. This will, in turn, send
weak activation to the appropriate bigram detectors. The CO-detector, however,
is well primed and so is likely to respond even though it is receiving only a weak
input. The other bigram detectors (for CQ or CS) are less well primed and so will
not respond to this weak input. Therefore, “CORN” will be correctly perceived,
despite the confusion at the letter level caused by the weak signal.
Feature Nets and Word Recognition
•
121
Q, and S, and so activation in this feature detector causes activation in all of
these letter detectors.
Of course, each of these letter detectors is wired so that it can also
receive input from other feature detectors. (And so usually the O-detector also
gets input from detectors for left curves, right curves, and top curves.) We’ve
already said, though, that with this brief input these other features weren’t
detected this time around. As a result, the O-detector will only be weakly
activated (because it’s not getting its usual full input), and the same is true for
the detectors for U, Q, and S.
In this situation, therefore, the network has partial information at the
feature level (because only one of the O’s features was detected), and this
leads to confusion at the letter level: Too many letter detectors are firing
(because the now-activated bottom-curve detector is wired to all of them).
And, roughly speaking, all of these letter detectors are firing in a fashion that
signals uncertainty, because they’re each receiving input from only one of
their usual feature detectors.
The confusion continues in the information sent upward from the let­
ter level to the bigram level. The detector for the CO bigram will receive a
strong signal from the C-detector (because the C was clearly visible) but only
a weak signal from the O-detector (because the O wasn’t clearly visible).
The CU-detector will get roughly the same input — a strong signal from the
C-detector and a weak signal from the U-detector. Likewise for the CQ- and
CS-detectors. In other words, we can imagine that the signal being sent from
the letter detectors is “maybe CO or maybe CU or maybe CQ or maybe CS.”
The confusion is, however, sorted out at the bigram level. All four bigram
detectors in this situation are receiving the same input — a strong signal from
one of their letters and a weak signal from the other. But the four detectors don’t
all respond in the same way. The CO-detector is well primed (because this is
a frequent pattern), so the activation it’s receiving will probably be enough to
fire this (primed) detector. The CU-detector is less primed (because this is a less
frequent pattern); the CQ- and CS-detectors, if they even exist, are not primed
at all. The input to these latter detectors is therefore unlikely to activate them —
because, again, they’re less well primed and so won’t respond to this weak input.
What will be the result of all this? The network was “under-stimulated”
at the feature level (with only a subset of the input’s features detected) and
therefore confused at the letter level (with too many detectors firing). But
then, at the bigram level, it’s only the CO-detector that fires, because at this
level it is the detector (because of priming) most likely to respond to the weak
input. Thus, in a totally automatic fashion, the network recovers from its
own confusion and, in this case, avoids an error.
Ambiguous Inputs
Look again at Figure 4.3. The second character is exactly the same as the fifth,
but the left-hand string is perceived as “THE” (and the character is identified as
an H) and the right-hand string is perceived as “CAT” (and the character as an A).
122 •
C H A P T E R F O U R Recognizing Objects
What’s going on here? In the string on the left, the initial T is clearly in
view, and so presumably the T-detector will fire strongly in response. The
next character in the display will probably trigger some of the features nor­
mally associated with an A and some normally associated with an H. This
will cause the A-detector to fire, but only weakly (because only some of the
A’s features are present), and likewise for the H-detector. At the letter level,
then, there will be uncertainty about what this character is.
What happens next, though, follows a by-now familiar logic: With only
weak activation of the A- and H-detectors, only a moderate signal will be
sent upward to the TH- and TA-detectors. Likewise, it seems plausible that
only a moderate signal will be sent to the THE- and TAE-detectors at the
word level. But, of course, the THE-detector is enormously well primed;
if there is a TAE-detector, it would be barely primed, since this is a string
that’s rarely encountered. Thus, the THE- and TAE-detectors might be
receiving similar input, but this input is sufficient only for the (well-primed)
THE-detector, so only it will respond. In this way, the net will recognize
the ambiguous pattern as “THE,” not “TAE.” (The same logic applies,
of course, to the ambiguous pattern on the right, perceived as “CAT,”
not “CHT.”)
A similar explanation will handle the word-superiority effect (see, e.g.,
Rumelhart & Siple, 1974). To take a simple case, imagine that we present
the letter A in the context “AT.” If the presentation is brief enough, par­
ticipants may see very little of the A, perhaps just the horizontal crossbar.
This wouldn’t be enough to distinguish among A, F, or H, and so all these
letter detectors would fire weakly. If this were all the information the par­
ticipants had, they’d be stuck. But let’s imagine that the participants did
perceive the second letter in the display, the T. It seems likely that the AT
bigram is much better primed than the FT or HT bigrams. (That’s because
you often encounter words like “CAT” or “BOAT”; words like “SOFT”
or “HEFT” are used less frequently.) Therefore, the weak firing of the
A-detector would be enough to fire the AT bigram detector, while the weak
firing for the F and H might not trigger their bigram detectors. In this way,
a “choice” would be made at the bigram level that the input was “AT” and
not something else. Once this bigram has been detected, answering the
question “Was there an A or an F in the display?” is easy. In this way, the
letter will be better detected in context than in isolation. This isn’t because
context enables you to see more; instead, context allows you to make better
use of what you see.
Recognition Errors
There is, however, a downside to all this. Imagine that we present the string
“CQRN” to participants. If the presentation is brief enough, the partici­
pants will register only a subset of the string’s features. Let’s imagine that
they register only the bottom bit of the string’s second letter. This detec­
tion of the bottom curve will weakly activate the Q-detector and also the
Feature Nets and Word Recognition
•
123
FIGURE 4.10
RECOGNITION ERRORS
Word
detectors
Bigram
detectors
Letter
detectors
Feature
detectors
Stimulus
input
If “CQRN” is presented briefly, not all of its features will be detected. Perhaps
only the bottom curve of the Q is detected, and this will weakly activate various other letters having a bottom curve, including O, U, and S. However, the
same situation would result from a brief presentation of “CORN” (as shown in
Figure 4.9); therefore, by the logic we have already discussed, this stimulus is
likely to be misperceived as “CORN.”
U-detector and the O-detector. The resulting pattern of network activation
is shown in Figure 4.10.
Of course, the pattern of activation here is exactly the same as it was in
Figure 4.9. In both cases, perceivers have seen the features for the C, R, and
N and have only seen the second letter’s bottom curve. And we’ve already
walked through the network’s response to this feature pattern: This con­
figuration will lead to confusion at the letter level, but the confusion will
get sorted out at the bigram level, with the (primed) CO-detector respond­
ing to this input and other (less well primed) detectors not responding.
As a result, the stimulus will be (mis)identified as “CORN.” In the situa­
tion described in Figure 4.9, the stimulus actually was “CORN,” and so
the dynamic built into the net aids performance, allowing the network to
recover from its initial confusion. In the case we’re considering now (with
124 •
C H A P T E R F O U R Recognizing Objects
“CQRN” as the stimulus), the exact same dynamic causes the network to
misread the stimulus.
This example helps us understand how recognition errors occur and why
those errors tend to make the input look more regular than it really is. The
basic idea is that the network is biased, favoring frequent letter combinations
over infrequent ones. In effect, the network operates on the basis of “when
in doubt, assume that the input falls into the frequent pattern.” The reason,
of course, is simply that the detectors for the frequent pattern are well primed —
and therefore easier to trigger.
Let’s emphasize, though, that the bias built into the network facilitates perception if the input is, in fact, a frequent word, and these (by
definition) are the words you encounter most of the time. The bias will
pull the network toward errors if the input happens to have an unusual
spelling pattern, but (by definition) these inputs are less common in your
experience. Hence, the network’s bias helps perception more often than
it hurts.
Distributed Knowledge
We’ve now seen many indications that the network’s functioning is guided by
knowledge of spelling patterns. This is evident in the fact that letter strings
are easier to recognize if they conform to normal spelling. The same point
is evident in the fact that letter strings provide a context benefit (the WSE)
only if they conform to normal spelling. Even more evidence comes from the
fact that when errors occur, they “shift” the perception toward patterns of
normal spelling.
To explain these results, we’ve suggested that the network “knows” (for
example) that CO is a common bigram in English, while CF is not, and
also “knows” that THE is a common sequence but TAE is not. The net­
work seems to rely on this “knowledge” in “choosing” its “interpretation” of
unclear or ambiguous inputs. Similarly, the network seems to “expect”
certain patterns and not others, and is more efficient when the input lines up
with those “expectations.”
Obviously, we’ve wrapped quotations around several of these words to
emphasize that the sense in which the net “knows” facts about spelling,
or the sense in which it “expects” things or makes “interpretations,” is a
little peculiar. In reality, knowledge about spelling patterns isn’t explicitly
stored anywhere in the network. Nowhere within the net is there a sentence
like “CO is a common bigram in English; CF is not.” Instead, this memory
(if we even want to call it that) is manifest only in the fact that the COdetector happens to be more primed than the CF-detector. The CO-detector
doesn’t “know” anything about this advantage, nor does the CF-detector
know anything about its disadvantage. Each one simply does its job, and
in the course of doing their jobs, sometimes a “competition” will take
place between these detectors. (This sort of competition was illustrated in
Feature Nets and Word Recognition
•
125
Figures 4.9 and 4.10.) When these competitions occur, they’ll be “decided” by
activation levels: The better-primed detector will be more likely to respond
and therefore will be more likely to influence subsequent events. That’s the
entire mechanism through which these “knowledge effects” arise. That’s
how “expectations” or “inferences” emerge — as a direct consequence of
the activation levels.
To put this into technical terms, the network’s “knowledge” is not locally
represented anywhere; it isn’t stored in a particular location or built into
a specific process. As a result, we cannot look just at the level of priming
in the CO-detector and conclude that this detector represents a frequently
seen bigram. Nor can we look at the CF-detector and conclude that it
represents a rarely seen bigram. Instead, we need to look at the relationship between these priming levels, and we also need to look at how this
relationship will lead to one detector being more influential than the
other. In this way, the knowledge about bigram frequencies is contained
within the network via a distributed representation; it’s knowledge, in
other words, that’s represented by a pattern of activations that’s distrib­
uted across the network and detectable only if we consider how the entire
network functions.
What may be most remarkable about the feature net, then, lies in how
much can be accomplished with a distributed representation, and thus with
simple, mechanical elements correctly connected to one another. The net
appears to make inferences and to know the rules of English spelling. But
the actual mechanics of the net involve neither inferences nor knowledge (at
least, not in any conventional sense). You and I can see how the inferences
unfold by taking a bird’s-eye view and considering how all the detectors work
together as a system. But nothing in the net’s functioning depends on the
bird’s-eye view. Instead, the activity of each detector is locally determined —
influenced by just those detectors feeding into it. When all these detectors
work together, though, the result is a process that acts as if it knows the rules.
But the rules themselves play no role in guiding the network’s moment-bymoment activities.
Efficiency versus Accuracy
One other point about the network needs emphasis: The network does
make mistakes, misreading some inputs and misinterpreting some pat­
terns. As we’ve seen, though, these errors are produced by exactly
the same mechanisms that are responsible for the network’s main
advantages — its ability to deal with ambiguous inputs, for example, or to
recover from confusion. Perhaps, therefore, we should view the errors as
the price you pay in order to gain the benefits associated with the net:
If you want a mechanism that’s able to deal with unclear or partial inputs,
you have to live with the fact that sometimes the mechanism will make
errors.
126 •
C H A P T E R F O U R Recognizing Objects
But do you really need to pay this price? After all, outside of the lab
you’re unlikely to encounter fast-paced tachistoscopic inputs. Instead, you
see stimuli that are out in view for long periods of time, stimuli that you can
inspect at your leisure. Why, therefore, don’t you take the moment to scruti­
nize these inputs so that you can rely on fewer inferences and assumptions,
and in that way gain a higher level of accuracy in recognizing the objects
you encounter?
The answer is straightforward. To maximize accuracy, you could, in
principle, scrutinize every character on the page. That way, if a character
were missing or misprinted, you would be sure to detect it. But the cost
associated with this strategy would be intolerable. Reading would be un­
speakably slow (partly because the speed with which you move your eyes
is relatively slow — no more than four or five eye movements per second).
In contrast, it’s possible to make inferences about a page with remark­
able speed, and this leads readers to adopt the obvious strategy: They read
some of the letters and make inferences about the rest. And for the most
part, those inferences are safe — thanks to the simple fact that our language
(like most aspects of our world) contains some redundncies, so that one
doesn’t need every lettr to identify what a wrd is; oftn the missng letter is
perfctly predctable from the contxt, virtually guaranteeing that inferences
will be correct.
TEST YOURSELF
6. H
ow does a feature
net explain the wordfrequency effect?
7. How does a feature
net explain the types
of errors people make
in recognizing words?
8. What are the benefits,
and what are the
costs, associated
with the feature net’s
functioning?
Descendants of the Feature Net
We mentioned early on that we were discussing the “classic” version of the
feature net. This discussion has enabled us to bring a number of themes into
view — including the trade-off between efficiency and accuracy and the idea
of distributed knowledge built into a network’s functioning.
Over the years, though, researchers have offered improvements on
this basic conceptualization, and in the next sections we’ll consider three
of their proposals. All three preserve the idea of a network of intercon­
nected detectors, but all three extend this idea in important ways. We’ll
look first at a proposal that highlights the role of inhibitory connections
among detectors. Then we’ll turn to a proposal that applies the network
idea to the recognition of complex three-dimensional objects. Finally, we’ll
consider a proposal that rests on the idea that your ability to recognize
objects may depend on your viewing perspective when you encounter
those objects.
The McClelland and Rumelhart Model
In the network proposal we’ve considered so far, activation of one detec­
tor serves to activate other detectors. Other models involve a mechanism
through which detectors can inhibit one another, so that the activation of one
detector can decrease the activation in other detectors.
Descendants of the Feature Net
•
127
FIGURE 4.11
N ALTERNATIVE CONCEPTION OF THE
A
FEATURE NETWORK
The McClelland and Rumelhart (1981) pattern-recognition model includes
both excitatory connections (indicated by red arrows) and inhibitory connections (indicated by connections with dots). Connections within a specific level
are also possible — so that, for example, activation of the “TRIP” detector will
inhibit the detectors for “TRAP,” “TAKE,” and “TIME.”
One highly influential model of this sort was proposed by McClelland and
Rumelhart (1981); a portion of their model is illustrated in Figure 4.11. This net­
work, like the one we’ve been discussing, is better able to identify well-formed
strings than irregular strings; this net is also more efficient in identifying charac­
ters in context as opposed to characters in isolation. However, several attributes
of this net make it possible to accomplish all this without bigram detectors.
In Figure 4.11, excitatory connections — connections that allow one detec­
tor to activate its neighbors — are shown as red arrows; for example, detec­
tion of a T serves to “excite” the “TRIP” detector. Other connections are
inhibitory, and so (for example) detection of a G deactivates, or inhibits,
the “TRIP” detector. These inhibitory connections are shown in the figure
with dots. In addition, this model allows for more complicated signaling
than we’ve used so far. In our discussion, we have assumed that lower-level
detectors trigger upper-level detectors, but not the reverse. The flow of infor­
mation, it seemed, was a one-way street. In the McClelland and Rumelhart
128 •
C H A P T E R F O U R Recognizing Objects
model, though, higher-level detectors (word detectors) can influence lowerlevel detectors, and detectors at any level can also influence other detectors
at the same level (e.g., letter detectors can inhibit other letter detectors; word
detectors can inhibit other word detectors).
To see how this would work, let’s say that the word “TRIP” is briefly shown,
allowing a viewer to see enough features to identify only the R, I, and P.
Detectors for these letters will therefore fire, in turn activating the detector for
“TRIP.” Activation of this word detector will inhibit the firing of other word
detectors (e.g., detectors for “TRAP” and “TAKE”), so that these other words
are less likely to arise as distractions or competitors with the target word.
At the same time, activation of the “TRIP” detector will also excite
the detectors for its component letters — that is, detectors for T, R, I,
and P. The R-, I-, and P-detectors, we’ve assumed, were already firing, so this
extra activation “from above” has little impact. But the T-detector wasn’t
firing before. The relevant features were on the scene but in a degraded form
(thanks to the brief presentation), and this weak input was insufficient to
trigger an unprimed detector. But once the excitation from the “TRIP” detec­
tor primes the T-detector, it’s more likely to fire, even with a weak input.
In effect, then, activation of the word detector for “TRIP” implies that
this is a context in which a T is quite likely. The network therefore responds
to this suggestion by “preparing itself” for a T. Once the network is suitably
prepared (by the appropriate priming), detection of this letter is facilitated. In
this way, the detection of a letter sequence (the word “TRIP”) makes the net­
work more sensitive to elements that are likely to occur within that sequence.
That is exactly what we need in order for the network to be responsive to the
regularities of spelling patterns.
Let’s also note that the two-way communication that’s in play here fits
well with how the nervous system operates: Neurons in the eyeballs send
activation to the brain but also receive activation from the brain; neurons
in the lateral geniculate nucleus (LGN) send activation to the visual cortex
but also receive activation from the cortex. Facts like these make it clear that
visual processing is not a one-way process, with information flowing simply
from the eyes toward the brain. Instead, signaling occurs in both an ascend­
ing (toward the brain) and a descending (away from the brain) direction, just
as the McClelland and Rumelhart model claims.
Recognition by Components
The McClelland and Rumelhart model — like the feature net we started
with — was designed initially as an account of how people recognize printed
language. But, of course, we recognize many objects other than print, includ­
ing the three-dimensional objects that fill our world — chairs and lamps and
cars and trees. Can these objects also be recognized by a feature network?
The answer turns out to be yes.
Consider a network theory known as the recognition by components
(RBC) model (Hummel & Biederman, 1992; Hummel, 2013). This model
Descendants of the Feature Net
•
129
FIGURE 4.12
GEONS
Geons
Objects
2
4
2
1
3
3
3
5
5
3
4
3
2
5
5
5
3
A
1
3
B
Panel A shows five different geons; Panel B shows how these geons can
be assembled into objects. The numbers in Panel B identify the specific
geons — for example, a bucket contains Geon 5 top-connected to Geon 3.
includes several important innovations, one of which is the inclusion of an
intermediate level of detectors, sensitive to geons (short for “geometric
ions”). The idea is that geons might serve as the basic building blocks of all
the objects we recognize — geons are, in essence, the alphabet from which
all objects are constructed.
Geons are simple shapes, such as cylinders, cones, and blocks (see
Figure 4.12A), and according to Biederman (1987, 1990), we only need 30
or so different geons to describe every object in the world, just as 26 letters
are all we need to spell all the words of English. These geons can be com­
bined in various ways — in a top-of relation, or a side-connected relation,
and so on — to create all the objects we perceive (see Figure 4.12B).
The RBC model, like the other networks we’ve been discussing, uses
a hierarchy of detectors. The lowest-level detectors are feature detectors,
which respond to edges, curves, angles, and so on. These detectors in turn
activate the geon detectors. Higher levels of detectors are then sensitive
to combinations of geons. More precisely, geons are assembled into com­
plex arrangements called “geon assemblies,” which explicitly represent
the relations between geons (e.g., top-of or side-connected). These assemblies,
130 •
C H A P T E R F O U R Recognizing Objects
finally, activate the object model, a representation of the complete, recog­
nized object.
The presence of the geon and geon-assembly levels within this hierarchy of­
fers several advantages. For one, geons can be identified from virtually any angle
of view. As a result, recognition based on geons is viewpoint-independent. Thus,
no matter what your position is relative to a cat, you’ll be able to identify its
geons and identify the cat. Moreover, it seems that most objects can be recog­
nized from just a few geons. As a consequence, geon-based models like RBC
can recognize an object even if many of the object’s geons are hidden from view.
Recognition via Multiple Views
A number of researchers have offered a different approach to object recogni­
tion (Hayward & Williams, 2000; Tarr, 1995; Tarr & Bülthoff, 1998; Vuong
& Tarr, 2004; Wallis & Bülthoff, 1999). They propose that people have stored
in memory a number of different views of each object they can recognize: an
image of what a cat looks like when viewed head-on, an image of what it looks
like from the left, and so on. According to this perspective, you’ll recognize
Felix as a cat only if you can match your current view of Felix with one of these
remembered views. But the number of views in memory is limited — maybe a
half dozen or so — and so, in many cases, your current view won’t line up with
any of the available images. In that situation, you’ll need to “rotate” the cur­
rent view to bring it into alignment with one of the views in memory, and this
mental rotation will cause a slight delay in the recognition.
The key, then, is that recognition sometimes requires mental rotation, and
as a result it will be slower from some viewpoints than from others. In other
words, the speed of recognition will be viewpoint-dependent, and a growing
body of data confirms this claim. We’ve already noted that you can recognize
objects from many different angles, and your recognition is generally fast.
However, data indicate that recognition is faster from some angles than oth­
ers, in a way that’s consistent with this multiple-views proposal.
According to this perspective, how exactly does viewpoint-dependent rec­
ognition proceed? One proposal resembles the network models we’ve been
discussing (Riesenhuber & Poggio, 1999, 2002; Tarr, 1999). In this proposal,
there is a hierarchy of detectors, with each successive layer within the network
concerned with more complex aspects of the whole. Thus, low-level detectors
respond to lines at certain orientations; higher-level detectors respond to cor­
ners and notches. At the top of the hierarchy are detectors that respond to
the sight of whole objects. It is important, though, that these detectors each
represent what the object looks like from a particular vantage point, and so the
detectors fire when there is a match to one of these view-tuned representations.
These representations are probably supported by tissue in the infero­
temporal cortex, near the terminus of the what pathway (see Figure 3.10).
Recording from cells in this area has shown that many neurons here seem
object-specific — that is, they fire preferentially when a certain type of object
is on the scene. (For an example of just how specific these cells can be in their
Descendants of the Feature Net
•
131
FIGURE 4.13
THE JENNIFER ANISTON CELL
Researchers in one study were able to do single-cell recording within the brains of people who were undergoing
surgical treatment for epilepsy. The researchers located cells that fired strongly whenever a picture of Jennifer
Aniston was in view — whether the picture showed her close up (picture 32) or far away (picture 29), with long
hair (picture 32) or shorter (picture 5). These cells are largely viewpoint-independent; other cells, though, are
viewpoint-dependent.
TEST YOURSELF
9. H
ow does the
McClelland and
Rumelhart model
differ from the older,
“classical” version of
the feature net?
10. On what issues is
there disagreement
between the recognition by components
(RBC) proposal and
the recognition via
multiple views proposal? On what issues
is there agreement?
132 •
“preferred” target, see Figure 4.13.) Crucially, though, many of these neurons
are view-tuned: They fire most strongly to a particular view of the target
object. This is just what one might expect with the multiple-views proposal
(Peissig & Tarr, 2007).
However, there has been lively debate between advocates of the RBC
approach (with its claim that recognition is largely viewpoint-independent)
and the multiple-views approach (with its argument that recognition is
viewpoint-dependent). And this may be a case in which both sides are right —
with some brain tissue being sensitive to viewpoint, and some brain tissue
not being sensitive (see Figure 4.14). Moreover, the perceiver’s task may be
crucial. Some neuroscience data suggest that categorization tasks (“Is this
a cup?”) may rely on viewpoint-independent processing in the brain, while
identification tasks (“Is this the cup I showed you before?”) may rely on
viewpoint-dependent processing (Milivojevic, 2012).
In addition, other approaches to object recognition are being explored (e.g.,
Hayward, 2012; Hummel, 2013; Peissig & Tarr, 2007; Ullman, 2007). Obviously,
there is disagreement in this domain. Even so, let’s be clear that all of the avail­
able proposals involve the sort of hierarchical network we’ve been discussing.
In other words, no matter how the debate about object recognition turns out, it
looks like we’re going to need a network model along the lines we’ve considered.
C H A P T E R F O U R Recognizing Objects
1.4
1.4
1.2
1.2
1.0
0.8
0.6
0.4
0.2
0
A
Right fusiform area
BOLD signal change (%)
BOLD signal change (%)
Left fusiform area
Repeated objects
0.8
0.6
0.4
0.2
0
New Same Different
objects view
view
B
FIGURE 4.14 VIEWPOINT
INDEPENDENCE
1.0
New Same Different
objects view
view
Repeated objects
Is object recognition viewpoint-dependent? Some
aspects of object recognition may be viewpointdependent while other aspects are not. Here, researchers documented viewpoint independence in
the left occipital cortex (Panel A), and so the activity in the fusiform area was the same even when
an object was viewed from a novel perspective.
However, as we see in Panel B, other data show
viewpoint dependence in the right occipital cortex.
Face Recognition
We began our discussion of network models with a focus on how people rec­
ognize letters and words. We’ve now extended our reach and considered how
a network might support the recognition of three-dimensional objects. But
there’s one type of recognition that seems to demand a different approach:
the recognition of faces.
Faces Are Special
As we described at the start of this chapter, damage to the visual system
can produce a disorder known as agnosia — an inability to recognize cer­
tain stimuli — and one type of agnosia specifically involves the perception
of faces. People who suffer from prosopagnosia generally have normal
vision. Indeed, they can look at a photograph and correctly say whether
the photo shows a face or something else; they can generally say whether
a face is a man’s or a woman’s, and whether it belongs to someone young
or someone old. But they can’t recognize individual faces — not even of
their own parents or children, whether from photographs or “live.” They
can’t recognize the faces of famous performers or politicians. In fact, they
can’t recognize themselves (and so they sometimes think they’re looking
through a window at a stranger when they’re actually looking at them­
selves in a mirror).
Often, this condition is the result of brain damage, but in some people
it appears to be present from birth, without any detectable brain damage
(e.g., Duchaine & Nakayama, 2006). Whatever its origin, prosopagnosia
seems to imply the existence of special neural structures involved almost
exclusively in the recognition and discrimination of faces. Presumably,
Face Recognition
•
133
prosopagnosia results from some problem or limitation in the functioning
of this brain tissue. (See Behrman & Avidan, 2005; Burton, Young, Bruce,
Johnston, & Ellis, 1991; Busigny, Graf, Mayer, & Rossion, 2010; Damasio,
Tranel, & Damasio, 1990; De Renzi, Faglioni, Grossi, & Nichelli, 1991. For
a related condition, involving an inability to recognize voices, see Shilowich
& Biederman, 2016.)
The special nature of face recognition is also suggested by a pattern that is
the opposite of prosopagnosia. Some people seem to be “super-recognizers”
and are magnificently accurate in face recognition, even though they have
no special advantage in other perceptual or memory tasks (e.g., Bobak,
Hancock, & Bate, 2015; Davis, Lander, Evans, & Jansari, 2016; Russell,
Duchaine, & Nakayama, 2009; Tree, Horry, Riley, & Wilmer, 2017). These
people are consistently able to remember (and recognize) faces that they
viewed only briefly at some distant point in the past, and they’re also more
successful in tasks that require “face matching” — that is, judging whether
two different views of a face actually show the same person.
There are certainly advantages to being a super-recognizer, but also some
disadvantages. On the plus side, being able to remember faces is obviously
a benefit for a politician or a sales person; super-recognizers also seem to
be much more accurate as eyewitnesses (e.g., in selecting a culprit from a
police lineup). In fact, London’s police force now has a special unit of superrecognizers involved in many aspects of crime investigation (Keefe, 2016).
On the downside, being a super-recognizer can produce some social awk­
wardness. Imagine approaching someone and cheerfully announcing, “I
know you! You used to work at the grocery store on Main Street.” The other
person (who, let’s say, did work in that grocery eight years earlier) might find
this puzzling, perhaps creepy, and maybe even alarming.
What about the rest of us — people who are neither prosopagnosic nor
super-recognizers? It turns out that people differ widely in their ability to
remember and recognize faces (Bindemann, Brown, Koyas, & Russ, 2012;
DeGutis, Wilmer, Mercado, & Cohan, 2013; Wilmer, 2017). These differ­
ences, from person to person, are easy to measure, and there are online
face memory tests that can help you find out whether you’re someone who
has trouble recognizing faces. (If you’re curious, point your browser at the
Cambridge Face Memory Test.)
In all people, though, face recognition seems to involve processes dif­
ferent from those used for other forms of recognition. For example, we’ve
mentioned the debate about whether recognition of houses, or teacups,
or automobiles is viewpoint-dependent. There is no question about this
issue, however, when we’re considering faces: Face recognition is strongly
dependent on orientation, and so it shows a powerful inversion effect. In
one study, four categories of stimuli were considered — right-side-up faces,
upside-down faces, right-side-up pictures of common objects other than
faces, and upside-down pictures of common objects. As Figure 4.15 shows,
performance suffered for all of the upside-down (i.e., inverted) stimuli.
However, this effect was much larger for faces than for other kinds of
134 •
C H A P T E R F O U R Recognizing Objects
Number of errors
6
Faces
4
Houses
FIGURE 4.15
2
Upright
Inverted
FACES AND THE INVERSION EFFECT
People’s memory for faces is quite good, when compared with memory for
other pictures (in this case, pictures of houses). However, performance is
very much disrupted when the pictures of faces are inverted. Performance
with houses is also worse with inverted pictures, but the effect of inversion
is much smaller.
( after yin , 1969)
stimuli (Bruyer, 2001; Yin, 1969). Moreover, with non-faces, the (relatively
small) effect of inversion becomes even smaller with practice; with faces,
the effect of inversion remains in place even after practice (McKone,
Kanwisher, & Duchaine, 2007).
The role of orientation in face recognition can also be illustrated infor­
mally. Figure 4.16 shows two upside-down photographs of former British
prime minister Margaret Thatcher (from Thompson, 1980). You can prob­
ably tell that something is odd about them, but now try turning the book
upside down so that the faces are right side up. As you can see, the difference
FIGURE 4.16 PERCEPTION
OF UPSIDE-DOWN FACES
The left-hand picture looks somewhat odd, but the two pictures
still look relatively similar to each
other. Now, try turning the book
upside down (so that the faces are
upright). In this position, the lefthand face (now on the right) looks
ghoulish, and the two pictures look
very different from each other. Our
perception of upside-down faces is
apparently quite different from our
perception of upright faces.
( after thompson , 1980)
Face Recognition
•
135
between the faces is striking, and yet this fiendish contrast is largely lost
when the faces are upside down. (Also see Rhodes, Brake, & Atkinson, 1993;
Valentine, 1988.)
Plainly, then, face recognition is strongly dependent on orientation in
ways that other forms of object recognition are not. Once again, though, we
need to acknowledge an ongoing debate. According to some authors, the rec­
ognition of faces really is in a category by itself, distinct from all other forms
of recognition (e.g., Kanwisher, McDermott, & Chun, 1997). Other authors,
however, offer a different perspective: They agree that face recognition is
special but argue that certain other types of recognition, in addition to faces,
are special in the same way. As one line of evidence, they argue that proso­
pagnosia isn’t just a disorder of face recognition. In one case, for example, a
prosopagnosic bird-watcher lost not only the ability to recognize faces but
also the ability to distinguish the different types of warblers (Bornstein, 1963;
Bornstein, Sroka, & Munitz, 1969). Another patient with prosopagnosia lost
the ability to tell cars apart; she can locate her car in a parking lot only by
reading all the license plates until she finds her own (Damasio, Damasio, &
Van Hoesen, 1982).
Likewise, in Chapter 2, we mentioned neuroimaging data showing that a
particular brain site — the fusiform face area (FFA) — is specifically respon­
sive to faces. (See, e.g., Kanwisher & Yovel, 2006. For a description of other
brain areas involved in face recognition, see Gainotti & Marra, 2011.) One
study, however, suggests that tasks requiring subtle distinctions among birds,
or among cars, can also produce high levels of activation in this brain area
(Gauthier, Skudlarski, Gore, & Anderson, 2000; also Bukach, Gauthier, &
Tarr, 2006). This finding suggests that the neural tissue “specialized” for faces
isn’t used only for faces. (For more on this debate, see, on the one side,
Grill-Spector, Knouf, & Kanwisher, 2004; McKone et al., 2007; Weiner &
Grill-Spector, 2013. On the other side, see McGugin, Gatenby, Gore, &
Gauthier, 2012; Richler & Gauthier, 2014; Stein, Reeder, & Peeler, 2016;
Wallis, 2013; Zhao, Bülthoff, & Bülthoff, 2016.)
What should we make of all this? There’s no question that humans have a
specialized recognition system that’s crucial for face recognition. This system
certainly involves the FFA in the brain, and damage to this system can cause
prosopagnosia. What’s controversial is how exactly we should describe this
system. According to some authors, the system is truly a face recognition
system and will be used for other stimuli only if those stimuli happen to be
“face-like” (see Kanwisher & Yovel, 2006). According to other authors, this
specialized system needs to be defined more broadly: It is used whenever you
are trying to recognize specific individuals within a highly familiar category
(e.g., Gauthier et al., 2000). The recognition of faces certainly has these traits
(e.g., you distinguish Fred from George from Jacob within the familiar category
of “faces”), but other forms of recognition may have the same traits (e.g., if
a bird-watcher is distinguishing different types within the familiar category
of “warblers”).
136 •
C H A P T E R F O U R Recognizing Objects
So far, the data don’t provide a clear resolution of this debate; both sides
of the argument have powerful evidence supporting their view. But let’s focus
on the key point of agreement: Face recognition is achieved by a process
that’s different from the process described earlier in this chapter. We need to
ask, therefore, how face recognition proceeds.
Holistic Recognition
The networks we’ve been considering so far all begin with an analysis of
a pattern’s parts (e.g., features, geons); the networks then assemble those
parts into larger wholes. Face recognition, in contrast, seems not to depend
on an inventory of a face’s parts; instead, this process seems to depend on
holistic perception of the face. In other words, face recognition depends on
the face’s overall configuration — the spacing of the eyes relative to the length
of the nose, the height of the forehead relative to the width of the face, and
so on. (For more on face recognition, see Bruce & Young, 1986; Duchaine &
Nakayama, 2006; Hayward, Crookes, Chu, Favelle, & Rhodes, 2016.)
Of course, a face’s features still matter in this holistic process. The key,
however, is that the features can’t be considered one by one, apart from
the context of the face. Instead, the features matter because of the rela­
tionships they create. It’s the relationships, not the features on their own,
that guide face recognition. (See Fitousi, 2013; Rakover, 2013; Rhodes,
BRAIN AREAS CRUCIAL FOR FACE PERCEPTION
Several brain sites seem to be especially activated when people are looking at faces.
These sites include the fusiform face area (FFA), the occipital face area (OFA), and
the superior temporal sulcus (fSTS).
Face Recognition
•
137
2012; Wang, Li, Fang, Tian, & Liu, 2012, but also see Richler & Gauthier,
2014. For more on holistic perception of facial movement, see Zhao &
Bülthoff, 2017.)
Some of the evidence for this holistic processing comes from the composite effect in face recognition. In an early demonstration of this effect,
Young, Hellawell, and Hay (1987) combined the top half of one face with
the bottom half of another, and participants were asked to identify just
the top half. This task is difficult if the two halves are properly aligned.
In this setting, participants seemed unable to focus only on the top half;
instead, they saw the top of the face as part of the whole (see Figure 4.17A).
Thus, in the figure, it’s difficult to see that the top half of the face is Hugh
Jackman (shown in normal view in Figure 4.17C). This task is relatively easy,
though, if the halves are misaligned (as in Figure 4.17B). Now, the stimulus
itself breaks up the configuration, making it possible to view the top half
on its own. (For related results, see Amishav & Kimchi, 2010; but also see
Murphy, Gray, & Cook, 2017. For evidence that the strength of holistic
processing is predictive of face-recognition accuracy, see Richler, Cheung,
& Gauthier, 2011. For a complication, though, see Rezlescu, Susilo, Wilmer,
& Caramazza, 2017.)
More work is needed to specify how the brain detects and interprets the
relationships that define each face. Also, our theorizing will need to take
some complications into account — including the fact that the recognition
processes used for familiar faces may be different from the processes used
for faces you’ve seen only once or twice (Burton, Jenkins, & Schweinberg,
2011; Burton, Schweinberger, Jenkins, & Kaufmann, 2015; Young & Burton,
2017). Evidence suggests that in recognizing familiar faces, you rely more
heavily on the relationships among the internal features of the face; for unfa­
miliar faces, you may be more influenced by the face’s outer parts such as the
hair and the overall shape of the head (Campbell et al., 1999).
Moreover, psychologists have known for years that people are more
accurate in recognizing faces of people from their own racial background
(e.g., Caucasians looking at other Caucasians, or Asians looking at other
Asians) than they are when trying to recognize people of other races (e.g.,
Meissner & Brigham, 2001). In fact, some people seem entirely prosop­
agnosic when viewing faces of people from other groups, even though
they have no difficulty recognizing faces of people from their own group
(Wan et al., 2017). These points may suggest that people rely on different
mechanisms for, say, “same-race” and “cross-race” face perception, and this
point, too, must be accommodated in our theorizing. (For recent discus­
sions, see Horry, Cheong, & Brewer, 2015; Wan, Crookes, Reynolds, Irons,
& McKone, 2015.)
Obviously, there is still work to do in explaining how we recognize our
friends and family — not to mention how we manage to remember and recog­
nize someone we’ve seen only once before. We know that face recognition re­
lies on processes different from those discussed earlier in the chapter, and we
know that these processes rely on the configuration of the face, rather than
138 •
C H A P T E R F O U R Recognizing Objects
FIGURE 4.17
THE COMPOSITE EFFECT IN FACE RECOGNITION
A
B
C
D
Participants were asked to identify the top half of composite faces like those in
Panels A and B. This task was much harder if the halves were properly aligned
(as in Panel A), and easier if the halves weren’t aligned (as in Panel B). With
the aligned faces, participants have a difficult time focusing on just the face’s
top (and so have a hard time recognizing Hugh Jackman — shown in Panel C).
Instead, they view the face as a whole, and this context changes their perception of Jackman’s features, making it harder to recognize him. (The bottom of
the composite face belongs to Justin Timberlake, shown in Panel D.)
Face Recognition
•
139
TEST YOURSELF
11. W
hat’s the evidence
that face recognition is different from
other forms of object
recognition?
12. What’s the evidence
that face recognition
depends on the face’s
configuration, rather
than the features one
by one?
its individual features. More research is needed, though, to fill in the details of
this holistic processing. (For examples of other research on memory for faces,
see Jones & Bartlett, 2009; Kanwisher, 2006; Michel, Rossion, Han, Chung,
& Caldara, 2006; Rhodes, 2012. For discussion of how these issues play out
in the justice system, with evidence coming from eyewitness identifications,
see Reisberg, 2014.)
Top-Down Influences on Object
Recognition
We’ve now discussed one important limitation of feature nets. These nets
can, as we’ve seen, accomplish a great deal, and they’re crucial for the rec­
ognition of print, three-dimensional objects in the visual environment, and
probably sounds as well. But there are some targets — faces, and perhaps
others — for which recognition depends on configurations rather than indi­
vidual features.
It turns out, though, that there is another limit on feature nets, even if
we’re focusing on the targets for which a feature net is useful — print, com­
mon objects, and so on. Even in this domain, feature nets must be supple­
mented with additional mechanisms. This requirement doesn’t undermine the
importance of the feature net idea; feature nets are definitely needed as part
of our theoretical account. The key word, however, is “part,” because we
need to place feature nets within a larger theoretical frame.
The Benefits of Larger Contexts
Earlier in the chapter, we saw that letter recognition is improved by context.
For example, the letter V is easier to recognize in the context “VASE,” or
even the nonsense context “VIMP,” than it is if presented alone. These are
examples of “top-down” effects — effects driven by your knowledge and ex­
pectations. And these particular top-down effects, based on spelling patterns,
are easily accommodated by the network: As we have discussed, priming
(from recency and frequency of use) guarantees that detectors that have often
been used in the past will be easier to activate in the future. In this way, the
network “learns” which patterns are common and which are not, and it is
more receptive to inputs that follow the usual patterns.
Other top-down effects, however, require a different type of explanation.
Consider the fact that words are easier to recognize if you see them as part
of a sentence than if you see them in isolation. There have been many formal
demonstrations of this effect (e.g., Rueckl & Oden, 1986; Spellman, Holyoak,
& Morrison, 2001; Tulving & Gold, 1963; Tulving, Mandler, & Baumal,
1964), but for our purposes an informal example will work. Imagine that we
tell research participants, “I’m about to show you a word very briefly on a
computer screen; the word is the name of something that you can eat.” If we
forced the participants to guess the word at this point, they would be unlikely
140 •
C H A P T E R F O U R Recognizing Objects
to name the target word. (There are, after all, many things you can eat, so
the chances are slim of guessing just the right one.) But if we briefly show
the word “CELERY,” we’re likely to observe a large priming effect; that is,
participants are more likely to recognize “CELERY” with this cue than they
would have been without the cue.
Think about what this priming involves. First, the person needs to under­
stand each of the words in the instruction. If she didn’t understand the word
“eat” (e.g., if she mistakenly thought we had said, “something that you can
beat”), we wouldn’t get the priming. Second, the person must understand the
relations among the words in the instruction. For example, if she mistakenly
thought we had said, “something that can eat you,” we would expect a very
different sort of priming. Third, the person has to know some facts about the
world — namely, the kinds of things that can be eaten; without this knowl­
edge, we would expect no priming.
Obviously, then, this instance of priming relies on a broad range of
knowledge, and there is nothing special about this example. We could ob­
serve similar priming effects if we tell someone that the word about to be
shown is the name of a historical figure or that the word is related to the Star
Wars movies. In each case, the instruction would facilitate perception, with
the implication that in order to explain these various priming effects, we’ll
need to hook up our object-recognition system to a much broader library of
information.
Here’s a different example, this time involving what you hear. Participants
in one study listened to a low-quality recording of a conversation. Some par­
ticipants were told they were listening to an interview with a job candidate;
others were told they were listening to an interview with a suspect in a crimi­
nal case (Lange, Thomas, Dana, & Dawes, 2011). This difference in context
had a powerful effect on what the participants heard. For example, the audio
contained the sentence “I got scared when I saw what it’d done to him.”
Participants who thought they were listening to a criminal often mis-heard
this statement and were sure they had heard “. . . when I saw what I’d done
to him.”
Where does all of this bring us? Examples like we’re considering here tell
us that we cannot view object recognition as a self-contained process. Instead,
knowledge that is external to object recognition (e.g., knowledge about what
is edible, or about the sorts of things a criminal might say) is imported into
and influences the process. In other words, these examples (unlike the ones
we considered earlier in the chapter) don’t depend just on the specific stimuli
you’ve encountered recently or frequently. Instead, what’s crucial for this
sort of priming is what you know coming into the experiment — knowledge
derived from a wide range of life experiences.
We have, therefore, reached an important juncture. We’ve tried in this
chapter to examine object recognition apart from other cognitive processes,
considering how a separate object-recognition module might function, with
the module then handing its product (the object it had recognized) on to
subsequent processes. We have described how a significant piece of object
TEST YOURSELF
13. W
hat’s the evidence
that word recognition
(or object recognition
in general) is influenced
by processes separate from what has
been seen recently or
frequently?
Top-Down Influences on Object Recognition
•
141
FIGURE 4.18
THE FLOW OF TOP-DOWN PROCESSING
LH Orbitofrontal Cortex
RH Fusiform
LH Fusiform
LH
10–4
RH
p<
10–2
130 ms
180 ms
215 ms
When viewers had only a very brief glimpse of a target object, brain activity
indicating top-down processing was evident in the front part of the brain (the
orbitofrontal cortex) 130 ms after the target came into view. Roughly 50 ms
later (and so 180 ms after the target came into view), brain activity increased
further back in the brain (in the right hemisphere’s fusiform area), indicating
successful recognition. This pattern was not evident when object recognition
was easy (because of a longer presentation of the target). Sensibly, top-down
processing plays a larger role when bottom-up processing is somehow limited
or inadequate.
recognition might proceed, but in the end we have run up against a problem —
namely, top-down priming that draws on knowledge from outside of object
recognition itself. (For neuroscience evidence that word and object recogni­
tion interacts with other sorts of information, see Carreiras et al., 2014; also
Figure 4.18.) This sort of priming depends on what is in memory and on how
that knowledge is accessed and used, and so we can’t tackle this sort of prim­
ing until we’ve said more about memory, knowledge, and thought. We there­
fore must leave object recognition for now in order to fill in other pieces of
the puzzle. We’ll have more to say about object recognition in later chapters,
once we have some additional theoretical machinery in place.
COGNITIVE PSYCHOLOGY AND EDUCATION
speed-reading
Students often wish they could read more quickly, and, in fact, it’s easy to
teach people how to speed-read. It’s important to understand, however, how
speed-reading works, because this will help you see when speed-reading is a
good idea — and when it’s a terrible strategy.
142 •
C H A P T E R F O U R Recognizing Objects
As the chapter describes, in normal reading there’s no need to look at every
word on the page. Printed material (like language in general) follows predictable
patterns, and so, having read a few words, you’re often able to guess what the
next words will be. And without realizing you’re doing it, you’re already exploit­
ing this predictability. In reading this (or any) page, your eyes skip over many of
the words, and you rely on rapid inference to fill in what you’ve skipped.
The same process is central for speed-reading. Courses that teach you
how to speed-read actually encourage you to skip more, as you move down
the page, and to rely more on inference. As a result, speed-reading isn’t really
“reading faster”; it is instead “reading less and inferring more.”
How does this process work? First, before you speed-read some text, you
need to lay the groundwork for the inference process — so that you’ll make
the inferences efficiently and accurately. Therefore, before you speed-read a
text, you should flip through it quickly. Look at the figures and the figure
captions. If there’s a summary at the end or a preview at the beginning, read
them. These steps will give you a broad sense of what the material is about,
preparing you to make rapid — and sensible — inferences about the material.
Second, you need to make sure you do rely on inference, rather than
word-by-word scrutiny of the page. To do this, read for a while holding an
index card just under the line you’re reading, or using your finger to slide
along the line of print to indicate what you’re reading at that moment. These
WHEN SHOULD YOU SPEED-READ?
Students are often assigned an enormous amount of reading, so strategies for
speed-reading can be extremely helpful. But it’s important to understand why speedreading works as it does; knowing this will help you decide when speed-reading is
appropriate and when it’s unwise.
Cognitive Psychology and Education
•
143
procedures establish a physical marker that helps you keep track of where
your eyes are pointing as you move from word to word.
This use of a pointer will become easy and automatic after a little practice,
and once it does, you’re ready for the key step. Rather than using the marker
to follow your eye position, use the marker to lead your eyes. Specifically, try
moving the index card or your finger a bit more quickly than you have so far,
and try to move your eyes to “keep up” with this marker.
Of course, if you suddenly realize that you don’t have a clue what’s on the
page, then you’ve been going too fast. Just move quickly enough so that you
have to hustle along to keep up with your pointer. Don’t move so quickly that
you lose track of what you’re reading.
This procedure will feel awkward at first, but it will become easier with
practice, and you’ll gradually learn to move the pointer faster and faster.
As a result, you’ll increase your reading speed by 30%, 40%, or more. But
let’s be clear about what’s going on here: You’re simply shifting the balance
between how much input you’re taking in and how much you’re filling in the
gaps with sophisticated guesswork. Often, this is a fine strategy. Many of the
things you read are highly predictable, so your inferences about the skipped
words are likely to be correct. In settings like these, you might as well use the
faster process of making inferences, rather than the slower process of looking
at individual words.
But speed-reading is a bad bet if the material is hard to understand.
In that case, you won’t be able to figure out the skipped words via infer­
ence, so speed-reading will hurt you. Speed-reading is also a poor choice if
you’re trying to appreciate an author’s style. If, for example, you speed-read
Shakespeare’s Romeo and Juliet, you probably will be able to make infer­
ences about the plot, but you won’t be able to make inferences about the
specific words you’re skipping over; you won’t be able to make inferences
about the language Shakespeare actually used. And, of course, if you miss the
language of Shakespeare and miss the poetry, you’ve missed the point.
Speed-reading will enable you to zoom through many assignments. But
don’t speed-read material that’s technical, filled with details that you’ll need,
or beautiful for its language. In those cases, you need to pay attention to the
words on the page and not rely on your own inferences.
For more on this topic . . .
Rayner, K., Schotter, E.R., Masson, M.E.J., Potter, M.C., & Treiman, R. (2016). So
much to read, so little time: How do we read, and can speed reading help?
Psychological Science in the Public Interest, 17, 4–34.
144 •
C H A P T E R F O U R Recognizing Objects
chapter review
SUMMARY
• We easily recognize a wide range of objects in
a wide range of circumstances. Our recognition is
significantly influenced by context, which can deter­
mine how or whether we recognize an object. To
study these achievements, investigators have often
focused on the recognition of printed language,
using this case to study how object recognition in
general might proceed.
• Many investigators have proposed that recogni­
tion begins with the identification of features in the
input pattern. Key evidence for this claim comes
from neuroscience studies showing that the detection
of features is separate from the processes needed to
assemble these features into more complex wholes.
• To study word recognition, investigators often
use tachistoscopic presentations. In these studies,
words that appear frequently in the language are
easier to identify than words that don’t appear fre­
quently, and so are words that have been recently
viewed — an effect known as repetition priming.
The data also show a pattern known as the “wordsuperiority effect”; this refers to the fact that let­
ters are more readily perceived if they appear in the
context of a word than if they appear in isolation.
In addition, well-formed nonwords are more readily
perceived than letter strings that do not conform to
the rules of normal spelling. Another reliable pat­
tern is that recognition errors, when they occur, are
quite systematic, with the input typically perceived
as being more regular than it actually is. These find­
ings, taken together, indicate that recognition is
influenced by the regularities that exist in our envi­
ronment (e.g., the regularities of spelling patterns).
• We can understand these results in terms of a
network of detectors. Each detector collects input
and fires when the input reaches a threshold level.
A network of these detectors can accomplish a great
deal; for example, it can interpret ambiguous in­
puts, recover from its own errors, and make infer­
ences about barely viewed stimuli.
• The feature net seems to “know” the rules of
spelling and “expects” the input to conform to these
rules. However, this knowledge is distributed across
the entire network and emerges only through the
network’s parallel processing. This setup leads to
enormous efficiency in our interactions with the
world because it enables us to recognize patterns
and objects with relatively little input and under
diverse circumstances. But these gains come at the
cost of occasional error. This trade-off may be nec­
essary, though, if we are to cope with the informa­
tional complexity of our world.
• A feature net can be implemented in different
ways — with or without inhibitory connections, for
example. With some adjustments (e.g., the addition
of geon detectors), the net can also recognize threedimensional objects. However, some stimuli — for
example, faces — probably are not recognized
through a feature net but, instead, require a differ­
ent sort of recognition system, one that is sensitive
to relationships and configurations within the stim­
ulus input.
• The feature net also needs to be supplemented to
accommodate top-down influences on object recog­
nition. These influences can be detected in the bene­
fits of larger contexts in facilitating recognition and
in forms of priming that are concept-driven rather
than data-driven. These other forms of priming call
for an interactive model that merges bottom-up and
top-down processes.
145
KEY TERMS
bottom-up processing (p. 111)
top-down processing (p. 111)
visual search task (p. 111)
integrative agnosia (p. 112)
tachistoscope (p. 113)
mask (p. 113)
priming (p. 114)
repetition priming (p. 114)
word-superiority effect (WSE) (p. 114)
well-formedness (p. 116)
feature nets (p. 118)
activation level (p. 118)
response threshold (p. 118)
recency (p. 119)
frequency (p. 119)
bigram detectors (p. 120)
local representation (p. 126)
distributed representation (p. 126)
excitatory connections (p. 128)
inhibitory connections (p. 128)
recognition by components (RBC) model (p. 129)
geons (p. 130)
viewpoint-independent recognition (p. 131)
viewpoint-dependent recognition (p. 131)
prosopagnosia (p. 133)
inversion effect (p. 134)
holistic perception (p. 137)
TEST YOURSELF AGAIN
1.What is the difference between “bottom-up”
and “top-down” processing?
2.What is the evidence that features play a special
role in object recognition?
3.What is repetition priming, and how is it
demonstrated?
4.What procedure demonstrates the wordsuperiority effect?
5.What’s the evidence that word perception is
somehow governed by the rules of ordinary
spelling?
6.How does a feature net explain the wordfrequency effect?
7.How does a feature net explain the types of
errors people make in recognizing words?
8.What are the benefits, and what are the costs,
associated with the feature net’s functioning?
146
9.How does the McClelland and Rumelhart
model differ from the older, “classical” version
of the feature net?
10.On what issues is there disagreement between
the recognition by components (RBC) proposal
and the recognition via multiple views proposal?
On what issues is there agreement?
11.What’s the evidence that face recognition
is different from other forms of object
recognition?
12.What’s the evidence that face recognition
depends on the face’s configuration, rather
than the features one by one?
13.What’s the evidence that word recognition
(or object recognition in general) is influenced
by processes separate from what has been seen
recently or frequently?
THINK ABOUT IT
1.Imagine that you were designing a mechanism
that would recognize items of clothing (shirts,
pants, jackets, belts). Would some sort of
feature net be possible? If so, what would the
net involve?
2.Imagine that you were designing a mechanism
that would recognize different smells (roses,
cinnamon, freshly mown grass, car exhaust).
Do you think some sort of feature net would be
possible? If so, what would the net involve?
E eBOOK DEMONSTRATIONS & ESSAYS
Go to http://digital.wwnorton.com/cognition7 for the online demonstrations and essays
relevant to this chapter. These can be found in the ebook.
Online Demonstrations
Online Applying Cognitive Psychology and the
Law Essays
• Demonstration 4.1: Features and Feature
• Cognitive Psychology and the Law: Cross-Race
Identification
Combination
• Demonstration 4.2: The Broad Influence of the
Rules of Spelling
• Demonstration 4.3: Inferences in Reading
COGNITION LABS
Go to http://digital.wwnorton.com/cognition7 to experience interactive online psychology
labs relevant to this chapter.
147
5
chapter
Paying Attention
what if…
Right now, you’re paying attention to this page, reading these words. But you could, if you chose, pay
attention to the other people in the room, or your plans for the weekend,
or even the feel of the floor under your feet.
What would your life be like if you couldn’t control your attention in
this way? Every one of us has, of course, had the maddening experience
of being distracted when we’re trying to concentrate. For example, there
you are on the bus, trying to read your book. You have no interest in the
conversation going on in the seats behind you, but you seem unable to
shut it out, and so you make no progress in your book.
The frustration in this experience is surely fueled by the fact that usually
you can control your attention, so it’s especially irritating when you can’t
focus in the way you want to. The life challenge is much worse, though,
for people who suffer from attention deficit disorder. This disorder is often
associated with hyperactivity — and hence the abbreviation ADHD. People with ADHD are often overwhelmed by the flood of information that’s
available to them, and they’re unable to focus on their chosen target.
The diagnosis of ADHD can range from relatively mild to quite severe,
and we’ll have more to say about ADHD later in the chapter. Nothing
in this range, though, approaches a much more extreme disruption in
attention termed “unilateral neglect syndrome.” This pattern is generally the result of damage to the parietal cortex, and patients with this
syndrome ignore all inputs coming from one side of the body. A patient
with neglect syndrome will, for example, eat food from only one side
of the plate, wash only half of his or her face, and fail to locate soughtfor objects if they’re on the neglected side (see Logie, 2012; Sieroff,
Pollatsek, & Posner, 1988). Someone with this disorder cannot safely
drive a car and, as a pedestrian, is likely to trip over unnoticed obstacles.
This syndrome typically results from damage to the right parietal
lobe, and so the neglect is for the left side of space. (Remember the
brain’s contralateral organization; see Chapter 2.) Neglect patients will
therefore read only the right half of words shown to them — they’ll read
“threat” as “eat,” “parties” as “ties.” If asked to draw a clock, they’ll probably remember that the numbers from 1 to 12 need to be included, but
they’ll jam all the numbers into the clock’s right side.
All these observations remind us just how crucial the ability to pay
attention is — so that you can focus on the things you want to focus
149
preview of chapter themes
•
•
•
ultiple mechanisms are involved in the seemingly
M
simple act of paying attention, because people must take
various steps to facilitate the processing of desired inputs.
Without these steps, their ability to pick up information
from the world is dramatically reduced.
•
any of the steps necessary for perception have a “cost”:
M
They require the commitment of mental resources. These
resources are limited in availability, which is part of the reason you usually can’t pay attention to two inputs at once—
doing so would require more resources than you have.
ome of the mental resources you use are specialized,
S
which means they’re required only for tasks of a certain
sort. Other resources are more general, needed for a wide
range of tasks. However, the resource demand of a task
can be diminished through practice.
•
e emphasize that attention is best understood not as a
W
process or mechanism but as an achievement. Like most
achievements, paying attention involves many elements, all
of which help you to be aware of the stimuli you’re interested in and not be pulled off track by irrelevant distractors.
perform two activities at the same time only if the activities don’t require more resources than you have available.
ivided attention (the attempt to do two things at once)
D
can also be understood in terms of resources. You can
on and not be pulled off track by distraction. But what is “attention”?
As we’ll see in this chapter, the ability to pay attention involves many
independent elements.
Selective Attention
William James (1842–1910) is one of the historical giants of the field of
psychology, and he is often quoted in the modern literature. One of his most
famous quotes provides a starting point for this chapter. Roughly 125 years
ago, James wrote:
Everyone knows what attention is. It is the taking possession by the
mind, in clear and vivid form, of one out of what seem several simultaneously possible objects or trains of thought. Focalization, concentration,
of consciousness are of its essence. It implies withdrawal from some
things in order to deal effectively with others, and is a condition which
has a real opposite in the confused, dazed, scatterbrained state which in
French is called distraction. . . . (James, 1890, pp. 403–404)
In this quote, James is describing what modern psychologists call selective
attention — that is, the skill through which a person focuses on one input or
one task while ignoring other stimuli that are also on the scene. But what
does this skill involve? What steps do you need to take in order to achieve the
focus that James described, and why is it that the focus “implies withdrawal
from some things in order to deal effectively with others”?
Dichotic Listening
Early studies of attention used a setup called dichotic listening: Participants
wore headphones and heard one input in the left ear and a different input
in the right ear. The participants were instructed to pay attention to one of
these inputs — the attended channel — and to ignore the message in the other
ear — the unattended channel.
150 •
C H A P T E R F I V E Paying Attention
FIGURE 5.1
THE INVISIBLE GORILLA
In this procedure, participants are instructed to keep track of the ballplayers in the white shirts. Intent on their
task, participants are oblivious to what the black-shirted players are doing, and—remarkably—they fail to see the
person in the gorilla suit strolling through the scene.
(figure provided by daniel j. simons.)
To make sure participants were paying attention, investigators gave them a
task called shadowing: Participants were required to repeat back what they were
hearing, word for word, so that they were echoing the attended channel. Their
shadowing performance was generally close to perfect, and they were able to
echo almost 100% of what they heard. At the same time, they heard remarkably
little from the unattended channel. If asked, after a minute or so of shadowing,
to report what the unattended message was about, they had no idea (e.g., Cherry,
1953). They couldn’t even tell if the unattended channel contained a coherent
message or random words. In fact, in one study, participants shadowed speech in
the attended channel, while in the unattended channel they heard a text in Czech,
read with English pronunciation. The individual sounds, therefore (the vowels,
the consonants), resembled English, but the message itself was (for an English
speaker) gibberish. After a minute of shadowing, only 4 of 30 participants
detected the peculiar character of the unattended message (Treisman, 1964).
We can observe a similar pattern with visual inputs. Participants in one study
viewed a video that has now gone viral on the Internet and is widely known
as the “invisible gorilla” video. In this video, a team of players in white shirts
is passing a basketball back and forth; people watching the video are urged to
count how many times the ball is passed from one player to another. Interwoven
with these players (and visible in the video) is another team, wearing black shirts,
also passing a ball back and forth; viewers are instructed to ignore these players.
Viewers have no difficulty with this task, but, while doing it, they usually don’t see another event that appears on the screen right in front of their
eyes. Specifically, they fail to notice when someone wearing a gorilla costume
walks through the middle of the game, pausing briefly to thump his chest
before exiting. (See Figure 5.1; Neisser & Becklen, 1975; Simons & Chabris,
1999; also see Jenkins, Lavie, & Driver, 2005.)
Selective Attention
•
151
Even so, people are not altogether oblivious to the unattended channel.
In selective listening experiments, research participants easily and accurately
report whether the unattended channel contained human speech, musical
instruments, or silence. If the unattended channel did contain speech, participants can report whether the speaker was male or female, had a high or
low voice, or was speaking loudly or softly. (For reviews of this early work,
see Broadbent, 1958; Kahneman, 1973.) Apparently, then, physical attributes
of the unattended channel are heard, even though participants are generally
clueless about the unattended channel’s semantic content.
In one study, however, participants were asked to shadow one passage
while ignoring a second passage. Embedded within the unattended channel
was a series of names, and roughly one third of the participants did hear their
own name when it was spoken — even though (just like in other studies) they
heard almost nothing else from the unattended input (Moray, 1959).
And it’s not just names that can “catch” your attention. Mention of a recently
seen movie, or of a favorite restaurant, will often be noticed in the unattended
channel. More broadly, words with some personal importance are often noticed,
even though the rest of the unattended channel is perceived only as an undifferentiated blur (Conway, Cowan, & Bunting, 2001; Wood & Cowan, 1995).
Inhibiting Distractors
THE COCKTAIL
PARTY EFFECT
We have all experienced some
version of the so-called cocktail party effect. There you are
at a party, deep in conversation. Other conversations are
going on, but somehow you’re
able to “tune them out.” All
you hear is the single conversation you’re attending to, plus
a buzz of background noise.
But now imagine that someone a few steps away from
you mentions the name of a
close friend of yours. Your attention is immediately caught,
and you find yourself listening to that other conversation
and (momentarily) oblivious
to the conversation you had
been engaged in. This experience, easily observed outside
the laboratory, matches the
pattern of experimental data.
152 •
How can we put all these research results together? How can we explain
both the general insensitivity to the unattended channel and also the cases in
which the unattended channel “leaks through”?
One option focuses on what you do with the unattended input. The proposal is that you somehow block processing of the inputs you’re not interested in, much as a sentry blocks the path of unwanted guests but stands
back and does nothing when legitimate guests are in view, allowing them
to pass through the gate unimpeded. This sort of proposal was central for
early theories of attention, which suggested that people erect a filter that
shields them from potential distractors. Desired information (the attended
channel) is not filtered out and so goes on to receive further processing
(Broadbent, 1958).
But what does it mean to “filter” something out? The key lies in the nervous system’s ability to inhibit certain responses, and evidence suggests that
you do rely on this ability to avoid certain forms of distraction. This inhibition, however, is rather specific, operating on a distractor-by-distractor basis.
In other words, you might have the ability to inhibit your response to this
distractor and the same for that distractor, but these abilities are of little
value if some new, unexpected distractor comes along. In that case, you need
to develop a new skill aimed at blocking the new intruder. (See Cunningham
& Egeth, 2016; Fenske, Raymond, Kessler, Westoby, & Tipper, 2005; Frings
& Wühr, 2014; Jacoby, Lindsay, & Hessels, 2003; Tsushima, Sasaki, &
Watanabe, 2006; Wyatt & Machado, 2013. For a glimpse of brain mechanisms
that support this inhibition, see Payne & Sekuler, 2014.)
C H A P T E R F I V E Paying Attention
The ability to ignore certain distractors — to shut them out — therefore
needs to be part of our theory. Other evidence, though, indicates that this
isn’t the whole story. That’s because you not only inhibit the processing of
distractors, you also promote the processing of desired stimuli.
Inattentional Blindness
We saw in Chapters 3 and 4 that perception involves a lot of activity, as
you organize and interpret the incoming stimulus information. It seems plausible that this activity would require some initiative and some resources from
you — and evidence suggests that it does.
In one experiment, participants were told that they would see large “+”
shapes on a computer screen, presented for 200 ms (milliseconds), followed by a
pattern mask. (The mask was just a meaningless jumble on the screen, designed
to disrupt any further processing.) If the horizontal bar of the “+” was longer
than the vertical, participants were supposed to press one button; if the vertical
bar was longer, they had to press a different button. As a complication, participants weren’t allowed to look directly at the “+.” Instead, they fixated on (i.e.,
pointed their eyes at) a mark in the center of the computer screen — a fixation
target — and the “+” shapes were shown just off to one side (see Figure 5.2).
FIGURE 5.2
INATTENTIONAL BLINDNESS
100
90
Percent failing
to see the change
80
70
60
50
40
30
20
10
0
No warning
Warning
Condition
Participants were instructed to point their eyes at the dot and to make judgments about the “+” shown just off to
the side. However, the dot itself briefly changed to another shape. If participants weren’t warned about this (and
so weren’t paying attention to the dot), they failed to detect this change — even though they had been pointing
their eyes right at the dot the whole time.
(after mack & rock, 1998)
Selective Attention
•
153
INATTENTIONAL
BLINDNESS OUTSIDE
THE LAB
Inattentional blindness is usually demonstrated in the laboratory, but it has a number of
real-world counterparts. Most
people, for example, have experienced the peculiar situation in which they can’t find
the mayonnaise in the refrigerator (or the ketchup or the
salad dressing) even though
they’re staring right at the
bottle. This happens because
they’re so absorbed in other
thoughts that they become
blind to an otherwise salient
stimulus.
154 •
For the first three trials of the procedure, events proceeded just as the participants
expected, and the task was relatively easy. On Trial 4, though, things were slightly
different: While the target “+” was on the screen, the fixation target disappeared and
was replaced by one of three shapes — a triangle, a rectangle, or a cross. Then, the
entire configuration (the “+” target and this new shape) was replaced by the mask.
Immediately after the trial, participants were asked: Was there anything different on this trial? Was anything present, or anything changed, that wasn’t there
on previous trials? Remarkably, 89% of the participants reported that there was
no change; they had failed to see anything other than the (attended) “+.” To
probe the participants further, the researchers told them (correctly) that during
the previous trial the fixation target had momentarily disappeared and had been
replaced by a shape. The participants were then asked what that shape had been,
and were given the choices of a triangle, a rectangle, or a cross (one of which,
of course, was the right answer). The responses to this question were essentially
random. Even when probed in this way, participants seemed not to have seen the
shape directly in front of their eyes (Mack & Rock, 1998; also see Mack, 2003).
This pattern has been named inattentional blindness (Mack & Rock,
1998; also Mack, 2003) — a pattern in which people fail to see a prominent
stimulus, even though they’re staring straight at it. In a similar effect, called
“inattentional deafness,” participants regularly fail to hear prominent stimuli
if they aren’t expecting them (Dalton & Fraenkel, 2012). In other studies,
participants fail to feel stimuli if the inputs are unexpected; this is “inattentional numbness” (Murphy & Dalton, 2016).
What’s going on here? Are participants truly blind (or deaf or numb) in
response to these various inputs? As an alternative, some researchers propose
that participants in these experiments did see (or hear or feel) the targets but, a
moment later, couldn’t remember what they’d just experienced (e.g., Wolfe, 1999;
also Schnuerch, Kreiz, Gibbons, & Memmert, 2016). For purposes of theory, this
distinction is crucial, but for now let’s emphasize what the two proposals have in
common: By either account, your normal ability to see what’s around you, and to
make use of what you see, is dramatically dim­inished in the absence of attention.
Think about how these effects matter outside of the laboratory. Chabris
and Simons (2010), for example, call attention to reports of traffic accidents
in which a driver says, “I never saw the bicyclist! He came out of nowhere!
But then — suddenly — there he was, right in front of me.” Drew, Võ, and
Wolfe (2013) showed that experienced radiologists often miss obvious ano­
malies in a patient’s CT scan, even when looking right at the anomaly. (For
similar concerns, related to inattentional blindness in eyewitnesses to crimes,
see Jaeger, Levin, & Porter, 2017.) Or, as a more mundane example, you go
to the refrigerator to find the mayonnaise (or the ketchup or the juice) and
don’t see it, even though it’s right in front of you.
In these cases, we lament the neglectful driver and the careless radiologist,
and your inability to find the mayo may cause you to worry that you’re losing
your mind (as well as your condiments). The reality, though, is that these cases
of failing-to-see are entirely normal. Perception requires more than “merely”
having a stimulus in front of your eyes. Perception requires some work.
C H A P T E R F I V E Paying Attention
Change Blindness
The active nature of perception is also evident in studies of change
blindness — observers’ inability to detect changes in scenes they’re looking
directly at. In some experiments, participants are shown pairs of pictures separated by a brief blank interval (e.g., Rensink, O’Regan, & Clark, 1997). The
pictures in each pair are identical except for one aspect — an “extra” engine
shown on the airplane in one picture and not in the other; a man wearing a
hat in one picture but not wearing one in the other; and so on (see Figure 5.3).
FIGURE 5.3
CHANGE BLINDNESS
In some change-blindness demonstrations, participants see one picture, then a second, then the first again, then
the second, and must spot the difference between the two pictures. Here, we’ve displayed the pictures side by
side, rather than putting them in alternation. Can you find the differences? For most people, it takes a surprising
amount of time and effort to locate the differences — even though some of the differences are large. Apparently,
having a stimulus directly in front of your eyes is no guarantee that you will perceive the stimulus.
Selective Attention
•
155
FIGURE 5.4
CHANGE
BLINDNESS
In this video, every time
there was a shift in camera angle, there was a
change in the scene — so
that the woman in the red
sweater abruptly gained
a scarf, the plates that
had been red were suddenly white, and so on.
When viewers watched
the video, though, they
noticed none of these
changes.
Participants know that their task is to detect any changes in the pictures, but
even so, the task is difficult. If the change involves something central to the
scene, participants may need to look back and forth between the pictures as
many as a dozen times before they detect the change. If the change involves
some peripheral aspect of the scene, as many as 25 alternations may be required.
A related pattern can be documented when participants watch videos. In
one study, observers watched a movie of two women having a conversation.
The camera first focused on one woman, then the other, just as it would in an
ordinary TV show or movie. The crucial element of this experiment, though,
was that certain aspects of the scene changed every time the camera angle
changed. For example, from one camera angle, participants could plainly see
the red plates on the table between the women. When the camera shifted to a
different position, though, the plates’ color had changed to white. In another
shift, one of the women gained a prominent scarf that she didn’t have on a
fraction of a second earlier (see Figure 5.4). Most observers, however, noticed
none of these changes (Levin & Simons, 1997; Shore & Klein, 2000; Simons
& Rensink, 2005).
Incredibly, the same pattern can be documented with live (i.e., not
filmed) events. In a remarkable study, an investigator (let’s call him “Leon”)
approached pedestrians on a college campus and asked for directions to a certain building. During the conversation, two men carrying a door approached
and deliberately walked between Leon and the research participant. As a
result, Leon was momentarily hidden (by the door) from the participant’s
view, and in that moment Leon traded places with one of the men carrying
the door. A second later, therefore, Leon was able to walk away, unseen, while
the new fellow (who had been carrying the door) stayed behind and conti­
nued the conversation with the participant.
Roughly half of the participants failed to notice this switch. They conti­
nued the conversation as though nothing had happened — even though Leon
and his replacement were wearing different clothes and had easily distinguishable voices. When asked whether anything odd had happened in this
event, many participants commented only that it was rude that the guys carrying the door had walked right through their conversation. (See Simons
& Ambinder, 2005; Chabris & Simons, 2010; also see Most et al., 2001;
Rensink, 2002; Seegmiller, Watson, & Strayer, 2011. For similar effects with
auditory stimuli, see Gregg & Samuel, 2008; Vitevitch, 2003.)
Early versus Late Selection
It’s clear, then, that people are often oblivious to stimuli directly in front of
their eyes — whether the stimuli are simple displays on a computer screen,
photographs, videos, or real-life events. (Similarly, people are sometimes
oblivious to prominent sounds in the environment.) As we’ve said, though,
there are two ways to think about these results. First, the studies may reveal
genuine limits on perception, so that participants literally don’t see (or hear)
these stimuli; or, second, the studies may reveal limits on memory, so that
156 •
C H A P T E R F I V E Paying Attention
participants do see (or hear) the stimuli but immediately forget what they’ve
just experienced.
Which proposal is correct? One approach to this question hinges on
when the perceiver selects the desired input and (correspondingly) when
the perceiver stops processing the unattended input. According to the early
selection hypothesis, the attended input is privileged from the start, so that
the unattended input receives little analysis and therefore is never perceived.
According to the late selection hypothesis, all inputs receive relatively complete analysis, and selection occurs after the analysis is fini­shed. Perhaps
the selection occurs just before the stimuli reach consciousness, so that we
become aware only of the attended input. Or perhaps the selection occurs
later still — so that all inputs make it (briefly) into consciousness, but then
the selection occurs so that only the attended input is remembered.
Each hypothesis captures part of the truth. On the one side, there are cases in
which people seem unaware of distractors but are influenced by them anyway —
so that the (apparently unnoticed) distractors guide the interpretation of the
attended stimuli (e.g., Moore & Egeth, 1997; see Figure 5.5). This seems to
be a case of late selection: The distractors are perceived (so that they do have
FIGURE 5.5
A
UNCONSCIOUS PERCEPTION
B
C
One study, apparently showing late selection, found that participants perceived (and were influenced) by background stimuli even though the participants did not consciously perceive these stimuli. The participants were
shown a series of images, each containing a pair of horizontal lines; their task was to decide which line was
longer. For the first three trials, the background dots in the display were arranged randomly (Panel A). For the
fourth trial, the dots were arranged as shown in Panel B, roughly reproducing the configuration of the Müller-Lyer
illusion; Panel C shows the standard form of this illusion. Participants in this study didn’t perceive the “fins” consciously, but they were influenced by them — judging the top horizontal line in Panel B to be longer, fully in accord
with the usual misperception of this illusion.
Selective Attention
•
157
TEST YOURSELF
1.What information do
people reliably pick
up from the attended
channel? What do
they pick up from the
unattended channel?
2.How is inattentional
blindness demonstrated? What situations
outside of the laboratory
seem to reflect inattentional blindness?
3.What evidence seems
to confirm early selection? What evidence
seems to confirm late
selection?
an influence) but are selected out before they make it to consciousness. On
the other side, though, we can also find evidence for early selection, with
attended inputs being privileged from the start and distractor stimuli falling
out of the stream of processing at a very early stage. Relevant evidence comes,
for example, from studies that record the brain’s electrical activity in the milli­
seconds after a stimulus has arrived. These studies confirm that the brain
activity for attended inputs is distinguishable from that for unattended inputs
just 80 ms or so after the stimulus presentation — a time interval in which
early sensory processing is still under way (Hillyard, Vogel, & Luck, 1998;
see Figure 5.6).
Other evidence suggests that attention can influence activity levels in the
lateral geniculate nucleus, or LGN (Kastner, Schneider, & Wunderlich, 2006;
McAlonan, Cavanaugh & Wurtz, 2008; Moore & Zirnsak, 2017; Vanduffel,
Tootell, & Orban, 2000). In this case, attention is changing the flow of signals
within the nervous system even before the signals reach the brain. (For more
on how attention influences processing in the visual cortex, see Carrasco,
Ling, & Read, 2004; Carrasco, Penpeci-Talgar, & Eckstein, 2000; McAdams
& Reid, 2005; Reynolds, Pasternak, & Desimone, 2000; also see O’Connor,
Fukui, Pinsk, & Kastner, 2002; Yantis, 2008.)
Selection via Priming
Whether selection is early or late, it’s clear that people often fail to see stimuli
that are directly in front of them, in plain view. But what is the obstacle here?
Why don’t people perceive these stimuli?
In Chapter 4, we proposed that recognition requires a network of
detectors, and we argued that these detectors fire most readily if they’re
suitably primed. In some cases, the priming is produced by your visual
experience — specifically, whether each detector has been used recently or
frequently in the past. But we suggested that priming can also come from
another source: your expectations about what the stimulus will be.
The proposal, then, is that you can literally prepare yourself for perceiving
by priming the relevant detectors. In other words, you somehow reach into
the network and deliberately activate just those detectors that, you believe,
will soon be needed. Then, once primed in this way, those detectors will be
on “high alert” and ready to fire.
Let’s also suppose that this priming isn’t “free.” Instead, you need to spend
some effort or allocate some resources in order to do the priming, and these
resources are in limited supply. As a result, there’s a limit on just how much
priming you can do.
We’ll need to flesh out this proposal in several ways, but even so, we can
already use it to explain some of the findings we’ve already met. Why don’t
participants notice the shapes in the inattentional blindness studies? The
answer lies in the fact that they don’t expect any stimulus to appear, so they
have no reason to prepare for any stimulus. As a result, when the stimulus
is presented, it falls on unprepared (unprimed, unresponsive) detectors. The
158 •
C H A P T E R F I V E Paying Attention
FIGURE 5.6
EVIDENCE FOR EARLY SELECTION
*
*
Right ear
*
Left ear
Time
A
Electrical activity in the brain in the interval just
after stimulus presentation
–2µV
Attended
N1 efect
0
Unattended
+2µV
0
B
80 100
200
Time (ms)
Participants were instructed to pay attention to the targets arriving in one
ear, but to ignore targets in the other ear (Panel A; dots indicate which of the
input signals were actually targets). During this task, researchers monitored
the electrical activity in the participants’ brains, with special focus on a brain
wave termed the “N1” (so-called because the wave reflects a nega­tive voltage
roughly 100 ms after the target). As Panel B shows, the N1 effect was different
for the attended and unattended inputs within 80 ms of the target’s arrival —
indicating that the attended and unattended inputs were processed differently from a very early stage.
(from hillyard et. al. “electric signs of selective
attention in the human brain,” science 182 © 1973 aaas. reprinted with permission.)
detectors therefore don’t respond to the stimulus, so the participants end up
not perceiving it.
What about selective listening? In this case, you’ve been instructed to
ignore the unattended input, so you have no reason to devote any resources
to this input. Hence, the detectors needed for the distractor message
are unprimed, and this makes it difficult to hear the distractor. But why
Selection via Priming
•
159
does attention sometimes “leak,” so that you do hear some aspects of the
unattended input? Think about what will happen if your name is spoken
on the unattended channel. The detectors for this stimulus are already
primed, but this isn’t because at that moment you’re expecting to hear
your name. Instead, the detectors for your name are primed simply because
this is a stimulus you’ve often encountered in the past. Thanks to this
prior exposure, the activation level of these detectors is already high;
you don’t need to prime them further. So they will fire even if your attention is elsewhere.
Two Types of Priming
The idea before us, in short, has three elements. First, perception is vastly
facilitated by the priming of relevant detectors. Second, the priming is
sometimes stimulus-driven — that is, produced by the stimuli you’ve
encountered (recently or frequently) in the past. This is repetition
priming — priming produced by a prior encounter with the stimulus. This
type of priming takes no effort on your part and requires no resources,
and it’s this sort of priming that enables you to hear your name on the
unattended channel. But third, a different sort of priming is also possible.
This priming is expectation-driven and under your control. In this form of
priming, you deliberately prime detectors for inputs you think are upcoming, so that you’re ready for those inputs when they arrive. You don’t do
this priming for inputs you have no interest in, and you can’t do this priming for inputs you can’t anticipate.
Can we test these claims? In a classic series of studies, Posner and Snyder
(1975) gave participants a straightforward task: A pair of letters was shown
on a computer screen, and participants had to decide, as swiftly as they could,
whether the letters were the same or different. So someone might see “AA”
and answer “same” or might see “AB” and answer “different.”
Before each pair, participants saw a warning signal. In the neutral
condition, the warning signal was a plus sign (“+”). This signal notified
participants that the stimuli were about to arrive but provided no other
information. In a different condition, the warning signal was a letter that
actually matched the stimuli to come. So someone might see the warning signal “G” followed by the pair “GG.” In this case, the warning
signal served to prime the participants for the stimuli. In a third condition,
though, the warning signal was misleading. It was again a letter, but a
different letter from the stimuli to come. Participants might see “H”
followed by the pair “GG.” Let’s consider these three conditions neutral,
primed, and misled.
In this simple task, accuracy rates are very high, but Posner and
Snyder also recorded how quickly people responded. By comparing these
response times (RTs) in the primed and neutral conditions, we can ask
what benefit there is from the prime. Likewise, by comparing RTs in the
160 •
C H A P T E R F I V E Paying Attention
misled and neutral conditions, we can ask what cost there is, if any, from
being misled.
Before we turn to the results, there’s a complication: Posner and Snyder
ran this procedure in two different versions. In one version, the warning
signal was an excellent predictor of the upcoming stimuli. For example, if
the warning signal was an A, there was an 80% chance that the upcoming
stimulus pair would contain A’s. In Posner and Snyder’s terms, the warning
signal provided a “high validity” prime. In a different version of the procedure, the warning signal was a poor predictor of the upcoming stimuli. For
example, if the warning signal was an A, there was only a 20% chance that
the upcoming pair would contain A’s. This was the “low validity” condition
(see Table 5.1).
Let’s consider the low-validity condition first, and let’s focus on those
few occasions in which the prime did match the subsequent stimuli. That
is, we’re focusing on 20% of the trials and ignoring the other 80% for the
moment. In this condition, the participant can’t use the prime as a basis
for predicting the stimuli because the prime is a poor indicator of things
to come. Therefore, the prime should not lead to any specific expectations.
Nonetheless, we do expect faster RTs in the primed condition than in the
neutral condition. Why? Thanks to the prime, the relevant detectors have
just fired, so the detectors should still be warmed up. When the target
stimuli arrive, therefore, the detectors should fire more readily, allowing a
faster response.
TABLE 5.1
D ESIGN OF POSNER AND SNYDER’S EXPERIMENT
TYPICAL SEQUENCE
Type of
Trial
Lowvalidity
Condition
Highvalidity
Condition
Warning
Signal
Test
Stimuli
Provides
Repetition
Priming?
Provides
Basis
for Expectation?
Neutral
+
AA
No
No
Primed
G
GG
Yes
No
Misled
H
GG
No
No
Neutral
+
AA
No
No
Primed
G
GG
Yes
Prime leads to
correct expectation
Misled
H
GG
No
Prime leads to
incorrect expectation
In the low-validity condition, misled trials occurred four times as often as primed
trials (80% vs. 20%). Therefore, participants had no reason to trust the primes and,
correspondingly, no reason to generate an expectation based on the primes. In the
high-validity condition, the arrangement was reversed: Now, primed trials occurred
four times as often as misled trials. Therefore, participants had good reason to trust
the primes and good reason to generate an expectation based on the prime.
(after posner & snyder, 1975)
Selection via Priming
•
161
The results bear this out. RTs were reliably faster (by roughly 30 ms) in
the primed condition than in the neutral condition (see Figure 5.7, left side;
the figure shows the differences between conditions). Apparently, detectors
can be primed by mere exposure to a stimulus, even in the absence of expectations, and so this priming is truly stimulus-based.
What about the misled condition? With a low-validity prime, misleading
the participants had no effect: Performance in the misled condition was the
same as performance in the neutral condition. Priming the “wrong” detector,
it seems, takes nothing away from the other detectors — including the detectors actually needed for that trial. This fits with our discussion in Chapter 4:
Each of the various detectors works independently of the others, and so
priming one detector obviously influences the functioning of that specific
detector but neither helps nor hinders the other detectors.
T HE EFFECTS OF PRIMING ON
STIMULUS PROCESSING
30
20
10
High-validity
condition
Benefit of
priming
40
Low-validity
condition
Cost of being
misled
50
Benefit of
priming
0
10
20
30
40
Cost of being
misled
Difference in response times between neutral
condition and experimental conditions (ms)
Cost
Benefit
FIGURE 5.7
50
As one way of assessing the Posner and Snyder (1975) results, we can subtract the response times for the neutral condition from those for the primed
condition; in this way, we measure the benefits of priming. Likewise, we
can subtract the response times for the neutral condition from those for
the misled condition; in this way, we measure the costs of being misled.
In these terms, the low-validity condition shows a small benefit (from repetition priming) but zero cost from being misled. The high-validity condition,
in contrast, shows a larger benefit — but also a substantial cost. The results
shown here reflect trials with a 300 ms interval between the warning signal
and the test stimuli. Results were somewhat different at other intervals.
162 •
C H A P T E R F I V E Paying Attention
Let’s look next at the high-validity primes. In this condition, people
might see, for example, a “J” as the warning signal and then the stimulus
pair “JJ.” Presentation of the prime itself will fire the J-detectors, and this
should, once again, “warm up” these detectors, just as the low-validity
primes did. As a result, we expect a stimulus-driven benefit from the prime.
However, the high-validity primes may also have another influence: Highvalidity primes are excellent predictors of the stimulus to come. Participants are told this at the outset, and they have lots of opportunity to see
that it’s true. High-validity primes will therefore produce a warm-up effect
and also an expectation effect, whereas low-validity primes produce only
the warm-up. On this basis, we should expect the high-validity primes
to help participants more than low-validity primes — and that’s exactly
what the data show (Figure 5.7, right side). The combination of warm-up
and expectations leads to faster responses than warm-up alone. From the
participants’ point of view, it pays to know what the upcoming stimulus
might be.
Explaining the Costs and Benefits
The data make it clear, then, that we need to distinguish two types of primes.
One type is stimulus-based — produced merely by presentation of the priming
stimulus, with no role for expectations. The other type is expectation-based
and is created only when the participant believes the prime allows a prediction of what’s to come.
These types of primes can be distinguished in various ways, including
the biological mechanisms that support them (see Figure 5.8; Corbetta &
Shulman, 2002; Hahn, Ross, & Stein, 2006; but also Moore & Zirnsak,
2017) and also a difference in what they “cost.” Stimulus-based priming
FIGURE 5.8 BIOLOGICAL
MECHANISMS FOR THE TWO
TYPES OF PRIMING
Brain sites shown in black have been
identified in various studies as involved
in expectation-based (sometimes called “goal
directed”) attention; sites shown in blue have
been implicated in “stimulus-driven” attention.
Sites shown in gray have been identified as
involved in both types of attention.
Selection via Priming
•
163
TEST YOURSELF
4.What are the differences between the
way that stimulusbased priming functions and the way that
expectation-based
priming functions?
5.Why is there a “cost”
associated with being
misled by expectationbased priming?
appears to be “free” — we can prime one detector without taking anything
away from other detectors. (We saw this in the low-validity condition, in
the fact that the misled trials led to responses just as fast as those in the neutral trials.) Expectation-based priming, in contrast, does have a cost, and
we see this in an aspect of Figure 5.7 that we’ve not yet mentioned: With
high-validity primes, responses in the misled condition were slower than
responses in the neutral condition. That is, misleading the participants actually hurt their performance. As a concrete example, F-detection was slower
if G was primed, compared to F-detection when the prime was simply the
neutral warning signal (“+”). In broader terms, it seems that priming the
“wrong” detector takes something away from the other detectors, and so
participants are worse off when they’re misled than when they receive no
prime at all.
What produces this cost? As an analogy, let’s say that you have just $50 to
spend on groceries. You can spend more on ice cream if you wish, but if you
do, you’ll have less to spend on other foods. Any increase in the ice cream
allotment, in other words, must be covered by a decrease somewhere else.
This trade-off arises, though, only because of the limited budget. If you had
unlimited funds, you could spend more on ice cream and still have enough
money for everything else.
Expectation-based priming shows the same pattern. If the Q-detector is
primed, this takes something away from the other detectors. Getting prepared for one target seems to make people less prepared for other targets. But
we just said that this sort of pattern implies a limited “budget.” If an unlimi­
ted supply of activation were available, you could prime the Q-detector and
leave the other detectors just as they were. And that is the point: Expectationbased priming, by virtue of revealing costs when misled, reveals the presence
of a limited-capacity system.
We can now put the pieces together. Ultimately, we need to explain the
facts of selective attention, including the fact that while listening to one
message you hear little content from other messages. To explain this, we’ve
proposed that perceiving involves some work, and this work requires some
limited mental resources — some process or capacity needed for performance,
but in limited supply. That’s why you can’t listen to two messages at the same
time; doing so would require more resources than you have. And now, finally,
we’re seeing evidence for those limited resources: The Posner and Snyder
research (and many other results) reveals the workings of a limited-capacity
system, just as our hypothesis demands.
Spatial Attention
The Posner and Snyder study shows that expectations about an upcoming
stimulus can influence the processing of that stimulus. But what exactly is the
nature of these expectations? How precise or vague are they?
As one way of framing this issue, imagine that participants in a study are
told, “The next stimulus will be a T.” In this case, they know exactly what to
164 •
C H A P T E R F I V E Paying Attention
get ready for. But now imagine that participants are told, “The next stimulus
will be a letter” or “The next stimulus will be on the left side of the screen.”
Will these cues allow participants to prepare themselves?
These issues have been examined in studies of spatial attention — that is,
the mechanism through which someone focuses on a particular position in
space. In one early study, Posner, Snyder, and Davidson (1980) required their
participants simply to detect letter presentations; the task was just to press
a button as soon as a letter appeared. Participants kept their eyes pointed at
a central fixation mark, and letters could appear either to the left or to the
right of this mark.
For some trials, a neutral warning signal was presented, so that participants knew a trial was about to start but had no information about stimulus
location. For other trials, an arrow was used as the warning signal. Sometimes the arrow pointed left, sometimes right; and the arrow was generally
an accurate predictor of the location of the stimulus-to-come. If the arrow
pointed right, the stimulus would be on the right side of the computer screen.
(In the terms we used earlier, this was a high-validity cue.) On 20% of the
trials, however, the arrow misled participants about location.
The results show a familiar pattern (Posner et al., 1980). With high-validity
priming, the data show a benefit from cues that correctly signal where the
upcoming target will appear. The differences between conditions aren’t large,
but keep the task in mind: All participants had to do was detect the input.
Even with the simplest of tasks, it pays to be prepared (see Figure 5.9).
What about the trials in which participants were misled? RTs in this condition were about 12% slower than those in the neutral condition. Once
again, therefore, we’re seeing evidence of a limited-capacity system. In order
to devote more attention to (say) the left position, you have to devote less
310
FIGURE 5.9 SPATIAL
ATTENTION
Mean response time (ms)
300
290
280
270
260
250
240
230
220
Expected location
(No expectation/
neutral cue)
Target appears at . . .
Unexpected location
In the Posner et al. (1980) study,
participants simply had to press
a button as soon as they saw the
target. If the target appeared in
the expected location, participants
detected it a bit more quickly. If,
however, participants were misled
about the target’s position (so that
the target appeared in an unexpected location), their responses
were slower than when the participants had no expectations at all.
Spatial Attention
•
165
attention to the right. If the stimulus then shows up on the right, you’re less
prepared for it — which is the cost of being misled.
Attention as a Spotlight
Studies of spatial attention suggest that visual attention can be compared to
a spotlight beam that can “shine” anywhere in the visual field. The “beam”
marks the region of space for which you are prepared, so inputs within the
beam are processed more efficiently. The beam can be wide or narrowly
focused (see Figure 5.10) and can be moved about at will as you explore
(i.e., attend to) various aspects of the visual field.
FIGURE 5.10
ADJUSTING THE “BEAM” OF ATTENTION
Charles Allan Gilbert’s painting All Is Vanity can be perceived either as a
woman at her dressing table or as a human skull. As you shift from one of
these perceptions to the other, you need to adjust the spotlight beam of
attention — to a narrow beam to see details (e.g., to see the woman) or to a
wider beam to see the whole scene (e.g., to see the skull).
166 •
C H A P T E R F I V E Paying Attention
Let’s emphasize, though, that the spotlight idea refers to movements of
attention, not movements of the eyes. Of course, eye movements do play
an important role in your selection of information from the world: If you
want to learn more about something, you generally look at it. (For more on
how you move your eyes to explore a scene, see Henderson, 2013; Moore
& Zirnsak, 2017.) Even so, movements of the eyes can be separated from
movements of attention, and it’s attention, not the eyes, that’s moving around
in the Posner et al. (1980) study. We know this because of the timing of the
effects. Eye movements are surprisingly slow, requiring 180 to 200 ms. But
the benefits of primes can be detected within the first 150 ms after the priming stimulus is presented. Therefore, the benefits of attention occur prior to
any eye movement, so they cannot be a consequence of eye movements.
But what does it mean to “move attention”? The spotlight beam is just a
metaphor, so we need to ask what’s really going on in the brain to produce
these effects. The answer involves a network of sites in the frontal cortex and
the parietal cortex. According to one proposal (Posner & Rothbart, 2007;
see Figure 5.11), one cluster of sites (the orienting system) is needed to disengage attention from one target, shift attention to a new target, and then
FIGURE 5.11
ANY BRAIN SITES ARE CRUCIAL
M
FOR ATTENTION
Frontal eye field
Anterior
cingulate gyrus
Frontal area
Prefrontal area
Superior parietal lobe
Posterior area
Temporoparietal
junction
Thalamus
Alerting
Orienting
Executive
Pulvinar
Superior
colliculus
Many brain sites are important for controlling attention. Some sites play a
pivotal role in alerting the brain, so that it is ready for an upcoming event.
Other sites play a key role in orienting attention, so that you’re focused on
this position or that, on one target or another. Still other sites are crucial
for controlling the brain’s executive function — a function we’ll discuss later in
the chapter.
(after posner & rothbart, 2007)
Spatial Attention
•
167
engage attention on the new target. A second set of sites (the alerting system)
is responsible for maintaining an alert state in the brain. A third set of sites
(the executive system) controls voluntary actions.
These points echo a theme we first met in Chapter 2. There, we argued
that cognitive capacities depend on the coordinated activity of multiple brain
regions, with each region providing a specialized process necessary for the
overall achievement. As a result, a problem in any of these regions can disrupt
the overall capacity, and if there are problems in several regions, the disruption can be substantial.
As an illustration of this interplay between brain sites and symptoms,
consider a disorder we mentioned earlier — ADHD. (We’ll have more to say
about ADHD later in the chapter.) Table 5.2 summarizes one proposal about
this disorder. Symptoms of ADHD are listed in the left column; the right
column identifies brain areas that may be the main source of each symptom.
This proposal is not the only way to think about ADHD, but it illustrates
the complex, many-part relationship between overall function (in this case,
the ability to pay attention) and brain anatomy. (For more on ADHD, see
Barkley, Murphy, & Fischer, 2008; Brown, 2005; Seli, Smallwood, Cheyne,
& Smilek, 2015; Zillmer, Spiers, & Culbertson, 2008.)
In addition, the sites listed in Table 5.2 can be understood roughly as
forming the “control system” for attention. Entirely different sites (including
TABLE 5.2
ATTENTION-DEFICIT/HYPERACTIVITY DISORDER
SYMPTOMS, COGNITIVE PROCESSES, AND
NEURAL NETWORKS
Symptom Domains and Cognitive Processes
Relevant Brain Site
Problems in the “Alerting” system
Has difficulty sustaining attention
Right frontal cortex
Fails to finish
Right posterior parietal
Avoids sustained efforts
Locus ceruleus
Problems in the “Orienting” system
Is distracted by stimuli
Bilateral parietal
Does not appear to listen
Superior colliculus
Fails to pay close attention
Thalamus
Problems in the “Executive” system
Blurts out answers
Anterior cingulate
Interrupts or intrudes
Left lateral frontal
Cannot wait
Basal ganglia
Earlier in the chapter, we mentioned the disorder known as ADHD. The table summarizes one influential proposal about this disorder, linking the symptoms of ADHD to
the three broad processes (alerting, orienting, and executive) described in the text,
and then linking these processes to relevant brain areas (Swanson et al., 2000; for a
somewhat different proposal, though, see Barkley, Murphy, & Fischer, 2008).
168 •
C H A P T E R F I V E Paying Attention
FIGURE 5.12
S ELECTIVE ATTENTION ACTIVATES THE VISUAL CORTEX
Attend left > Attend right
Left
hemisphere
Attend right > Attend left
Right
hemisphere
A
B
The brain sites that control attention are separate from the brain sites that do the actual analysis of the input. Thus,
the intention to attend to, say, stimuli on the left is implemented through the many brain sites shown in Figure 5.11.
However, these sites collectively activate a different set of sites — in the visual cortex — to promote the actual processing of the incoming stimulus. Shown here are activity levels in one participant (measured through fMRI scans)
overlaid on a structural image of the brain (obtained through MRI scans). Keep in mind that because of the brain’s
contralateral organization, the intention to pay attention to the left side of space requires activation in the right
hemisphere (Panel A); the intention to pay attention to the right requires activation in the left (Panel B).
the visual areas in the occipital cortex) do the actual analysis of the incoming
information (see Figure 5.12). In other words, neural connections from the
areas listed in the table carry signals to the brain regions that do the work
of analyzing the input. These control signals can amplify (or, in some cases,
inhibit) the activity in these other areas and, in this way, they can promote
the processing of inputs you’re interested in, and undermine the processing of
distractors. (See Corbetta & Shulman, 2002; Hampshire, Duncan, & Owen,
2007; Hon, Epstein, Owen, & Duncan, 2006; Hung, Driver, & Walsh, 2005;
Miller & Cohen, 2001.)
Thus, there is no spotlight beam. Instead, certain neural mechanisms
enable you to adjust your sensitivity to certain inputs. This is, of course,
entirely in line with the proposal we’re developing — namely, that a large
part of “paying attention” involves priming. For stimuli you don’t care
about, you don’t bother to prime yourself, and so those stimuli fall on
unprepared (and unresponsive) detectors. For stimuli you do care about,
you do your best to anticipate the input, and you use these anticipations
to prime the relevant processing channel. This increases your sensitivity
to the desired input, which is just what you want. (For further discussion
of the “spotlight” idea, see Cave, 2013; Rensink, 2012; Wright & Ward,
2008; but also Awh & Pashler, 2000; Morawetz, Holz, Baudewig, Treue,
& Dechent, 2007.)
Spatial Attention
•
169
Where Do We “Shine” the “Beam”?
So far, we’ve been discussing how people pay attention, but we can also ask
what people pay attention to. Where do people “shine” the “spotlight beam”?
The answer has several parts. As a start, you pay attention to elements of
the input that are visually prominent (Parkhurst, Law, & Niebur, 2002) and
also to elements that you think are interesting or important. Decisions about
what’s important, though, depend on the context. For example, Figure 5.13
shows classic data recording a viewer’s eye movements while inspecting a
picture (Yarbus, 1967). The target picture is shown in the top left. Each of the
other panels shows a three-minute recording of the viewer’s eye movements;
FIGURE 5.13
EYE MOVEMENTS AND VISION
1
2
Free examination.
3
Give the ages of the people.
Estimate material circumstances
of the family.
4
5
Remember the clothes
worn by the people.
Surmise what the family had
been doing before the arrival
of the unexpected visitor.
6
Remember positions of people and
objects in the room.
7
Estimate how long the visitor had
been away from the family.
Participants were shown the picture in the top left. Each of the other panels shows a three-minute recording
of one viewer’s eye movements while inspecting the picture. The labels for each panel summarize the viewer’s
goal while looking at the picture. Plainly, the pattern of the movements depended on what the viewer was trying
to learn.
170 •
C H A P T E R F I V E Paying Attention
plainly, the pattern of movements depended on what the viewer was trying
to learn about the picture.
In addition, your beliefs about the scene play an important role. You’re
unlikely to focus, for example, on elements of a scene that are entirely predictable, because you’ll gain no information from inspecting things that are
already obvious (Brewer & Treyens, 1981; Friedman, 1979; Võ & Henderson,
2009). But you’re also unlikely to focus on aspects of the scene that are
totally unexpected. If, for example, you’re walking through a forest, you
won’t be on the lookout for a stapler sitting on the ground, and so you may
fail to notice the stapler (unless it’s a bright color, or directly in your path, or
some such). This point provides part of the basis for inattentional blindness
(pp. 155–156) and also leads to a pattern called the “ultra-rare item effect”
(Mitroff & Biggs, 2014). The term refers to a pattern in which rare items
are often overlooked; as the authors of one paper put it, “If you don’t find it
often, you often don’t find it” (Evans, Birdwell, & Wolfe, 2013; for a troubling consequence of this pattern, see Figure 5.14).
As another complication, people differ in what they pay attention to
(e.g., Castelhano & Henderson, 2008), although some differences aren’t surprising. For example, in looking at a scene, women are more likely than men
to focus on how the people within the scene are dressed; men are more likely
to focus on what the people look like (including their body shapes; Powers,
Andriks, & Loftus, 1979).
Perhaps more surprising are differences from one culture to the next
in how people pay attention. The underlying idea here is that people in
the West (the United States, Canada, most of Europe) live in “individualistic” cultures that emphasize the achievements and qualities of the single
person; therefore, in thinking about the world, Westerners are likely to
focus on individual people, individual objects, and their attributes. In contrast, people in East Asia have traditionally lived in “collectivist” cultures
that emphasize the ways in which all people are linked to, and shaped
by, the people around them. East Asians are therefore encouraged to
think more holistically, with a focus on the context and how people and
FIGURE 5.14
IF YOU DON’T FIND IT OFTEN . . .
Data both in the laboratory and in real-world settings tell us that
people often overlook targets if the targets happen to be quite
rare. As one group of authors put it, “If you don’t find it often, you
often don’t find it.” This pattern has troubling implications for the
security inspections routinely conducted at airports: The inspectors
will see troubling items only rarely and, as a result, are likely to overlook those troubling items.
Spatial Attention
•
171
objects are related to one another. (See Nisbett, 2003; Nisbett, Peng, Choi, &
Norenzayan, 2001; also Tardif et al., 2017. For a broad review, see
Heine, 2015.)
This linkage between culture and cognition isn’t rigid, and so people in
any culture can stray from these patterns. Even so, researchers have docu­
mented many manifestations of these differences from one culture to the
next — including differences in how people pay attention. In one study,
researchers tracked participants’ eye movements while the participants were
watching animated scenes on a computer screen (Masuda et al., 2008). In the
initial second of viewing, there was little difference between the eye movements of American participants and Japanese participants: Both groups spent
90% of the time looking at the target person, located centrally in the display.
But in the second and third seconds of viewing, the groups differed. The
Americans continued to spend 90% of their time looking directly at the central figure (and so spent only 10% of their time looking at the faces of people
visible in the scene’s background); Japanese participants, in contrast, spent
between 20% and 30% of their time looking at the faces in the background.
This difference in viewing time was reflected in the participants’ judgments
about the scene. When asked to make a judgment about the target person’s
emotional state, American participants weren’t influenced by the emotional
expressions of the people standing in the background, but Japanese participants were.
In another study, American and East Asian students were tested in a
change-blindness procedure. They viewed one picture, then a second, then
the first again, and this alternation continued until the participants spotted the difference between the two pictures. For some pairs of pictures, the
difference involved an attribute of a central object in the scene (e.g., the
color of a truck changed from one picture to the next). For these pictures,
Americans and East Asians performed equivalently, needing a bit more than
9 seconds to spot the change. For other pairs of pictures, the difference
involved the context (e.g., a pattern of the clouds in the background of
the scene). For these pictures, the East Asians were notably faster than the
Americans in detecting the change. (See Masuda & Nisbett, 2006. For other
data, exploring when Westerners have a performance advantage and when
East Asians have the advantage, see Amer, Ngo, & Hasher, 2016; Boduroglu,
Shah, & Nisbett, 2010.)
Finally, let’s acknowledge that sometimes you choose what to pay
attention to — a pattern called endogenous control of attention. But sometimes an element of the scene “seizes” your attention whether you like it or
not, and this pattern is called exogenous control of attention. Exogenous
control is of intense interest to theorists, and it’s also important for
pragmatic reasons. For example, people who design ambulance sirens or
warning signals in an airplane cockpit want to make sure these stimuli
cannot be ignored. In the same way, advertisers do all they can to ensure
that their product name or logo will grab your attention even if you’re
172 •
C H A P T E R F I V E Paying Attention
FIGURE 5.15
EXOGENOUS CONTROL OF ATTENTION
Public health officials would like the health warning shown here to seize your attention, so that you can’t
overlook it. The tobacco industry, however, might have a different preference.
intensely focused on something else (e.g., the competitor’s product; also
see Figure 5.15).
Attending to Objects or Attending to Positions
A related question is also concerned with the “target” of the attention “spotlight.” To understand the issue, think about how an actual spotlight works. If
a spotlight shines on a donut, then part of the beam will fall on the donut’s
hole and will illuminate part of the plate underneath the donut. Similarly, if
the beam isn’t aimed quite accurately, it may also illuminate the plate just to
the left of the donut. The region illuminated by the beam, in other words, is
defined purely in spatial terms: a circle of light at a particular position. That
circle may or may not line up with the boundaries of the object you’re shining
the beam on.
Is this how attention works — so that you pay attention to whatever falls
in a certain region of space? If this is the case, you might at times end up
paying attention to part of this object, part of that. An alternative is that you
pay attention to objects rather than to positions in space. To continue the
example, the target of your attention might be the donut itself rather than
Spatial Attention
•
173
FIGURE 5.16
UNILATERAL NEGLECT SYNDROME
A patient with damage to the right parietal cortex was
asked to draw a typical clock face. In his drawing, the patient
seemed unaware of the left side, but he still recalled that all
12 numbers had to be displayed. The drawing shows how he
resolved this dilemma.
its location. In that case, the plate just to the left and the bit of plate visible
through the donut’s hole might be close to your focus, but they aren’t part of
the attended object and so aren’t attended.
Which is the correct view of attention? Do you pay attention to regions
in space, no matter what objects (or parts of objects) fall in that region? Or
do you pay attention to objects? It turns out that each view captures part of
the truth.
One line of evidence comes from the study of people we mentioned at
the chapter’s start – people who suffer from unilateral neglect syndrome (see
Figure 5.16). Taken at face value, the symptoms shown by these patients seem
to support a space-based account of attention: The afflicted patient seems
insensitive to all objects within a region that’s defined spatially — namely,
everything to the left of his or her current focus. If an object falls half within
the region and half outside of it, then the spatial boundary is what matters, not the object’s boundaries. This is clear, for example, in how these
patients read words (likely to read “BOTHER” as “HER” or “CARROT” as
“ROT”) — responding only to the word’s right half, apparently oblivious to
the word’s overall boundaries.
Other evidence, however, demands further theory. In one study, patients
with neglect syndrome had to respond to targets that appeared within a
barbell-shaped frame (see Figure 5.17). Not surprisingly, they were much
more sensitive to the targets appearing within the red circle (on the right) and
missed many of the targets appearing in the blue circle (on the left); this result
confirms the patients’ diagnosis. What’s crucial, though, is what happened
next. While the patients watched, the barbell frame was slowly spun around,
174 •
C H A P T E R F I V E Paying Attention
FIGURE 5.17
SPACE-BASED OR OBJECT-BASED ATTENTION
Patient initially sees:
As the patient watches:
Patient now sees:
Patients with unilateral neglect syndrome were much more sensitive to
targets appearing within the red circle (on the right) and missed many of
the targets appearing within the blue circle (on the left); this observation
confirms their clinical diagnosis. Then, as the patients watched, the barbellshaped frame rotated, so that now the red circle was on the left and the blue
circle was on the right. After this rotation, participants were still more sensitive to targets in the red circle (now on the left), apparently focusing on this
attended object even though it had moved into their “neglected” side.
so that the red circle, previously on the right, was now on the left and the blue
circle, previously on the left, was now on the right.
If the patients consistently neglect a region of space, they should now be
more sensitive to the (right-side) blue circle. But here’s a different possibility:
Perhaps these patients have a powerful bias to attend to the right side, and so
initially they attend to the red circle. Once they have “locked in” to this circle,
however, it’s the object, not the position in space, that defines their focus of
attention. According to this view, if the barbell rotates, they will continue
attending to the red circle (this is, after all, the focus of their attention), even
though it now appears on their “neglected” side. This prediction turns out to
be correct: When the barbell rotates, the patients’ focus of attention seems to
rotate with it (Behrmann & Tipper, 1999).
To describe these patients, therefore, we need a two-part account. First, the
symptoms of neglect syndrome plainly reveal a spatially defined bias: These
Spatial Attention
•
175
patients neglect half of space. But, second, once attention is directed toward
a target, it’s the target itself that defines the focus of attention; if the target
moves, the focus moves with it. In this way, the focus of attention is objectbased, not space-based. (For more on these issues, see Chen & Cave, 2006;
Logie & Della Salla, 2005; Richard, Lee, & Vecera, 2008.)
And it’s not just people with brain damage who show this complex pattern. People with intact brains also show a mix of space-based and objectbased attention. We’ve already seen evidence for the spatial base: The Posner
et al. (1980) study and many results like it show that participants can focus
on a particular region of space in preparation for a stimulus. In this situation,
the stimulus has not yet appeared; there is no object to focus on. Therefore,
the attention must be spatially defined.
In other cases, though, attention is heavily influenced by object boundaries. For example, in some studies, participants have been shown displays with
visually superimposed stimuli (e.g., Becklen & Cervone, 1983; Neisser &
Becklen, 1975). Participants can usually pay attention to one of these stimuli and ignore the other. This selection cannot be space-based (because both
stimuli are in the same place) and so must be object-based.
This two-part account is also supported by neuroscience evidence. Various studies have examined the pattern of brain activation when participants are attending to a particular position in space, and the pattern of
activation when participants are attending to a particular object. These
data suggest that the tasks involve different brain circuits — with one set
of circuits (the dorsal attention system), near the top of the head, being
primarily concerned with spatial attention, and a different set of circuits
(the ventral attention system) being crucial for nonspatial tasks (Cave,
2013; Cohen, 2012; Corbetta & Shulman, 2011). Once again, therefore,
our description of attention needs to include a mix of object-based and
space-based mechanisms.
Perceiving and the Limits on Cognitive Capacity:
An Interim Summary
Let’s pause to review. In some circumstances, you seem to inhibit the processing of unwanted inputs. This inhibitory process is quite specific (the inhibition
blocks the processing of a particular well-defined input) and certainly benefits
from practice.
More broadly, though, various mechanisms facilitate the processing of
desired inputs. The key here is priming. You’re primed for some stimuli
because you’ve encountered them often in the past, with the result that you’re
more likely to process (and therefore more likely to notice) these stimuli if
you encounter them again. In other cases, the priming depends on your ability to anticipate what the upcoming stimulus will be. If you can predict what
the stimulus will likely be, you can prime the relevant processing pathway
so that you’re ready for the stimulus when it arrives. The priming will make
you more responsive to the anticipated input if it does arrive, and this gives
176 •
C H A P T E R F I V E Paying Attention
the anticipated input an advantage relative to other inputs. That advantage
is what you want — so that you end up perceiving the desired input (the one
you’ve prepared yourself for) but don’t perceive the inputs you’re hoping to
ignore (because they fall on unprimed detectors).
The argument, then, is that your ability to pay attention depends to a large
extent on your ability to anticipate the upcoming stimulus. This anticipation,
in turn, depends on many factors. You’ll have a much easier time anticipating
(and so an easier time paying attention to) materials that you understand as
opposed to materials that you don’t understand. Likewise, when a stimulus
sequence is just beginning, you have little basis for anticipation, so your only
option may be to focus on the position in space that holds the sequence. Once
the sequence begins, though, you get a sense of how it’s progressing, and this
lets you sharpen your anticipations (shifting to object-based attention, rather
than space-based) — which, again, makes you more sensitive to the attended
input and more resistant to distractors.
Putting all these points together, perhaps it’s best not to think of the term
“attention” as referring to a particular process or a particular mechanism.
Instead, it’s better to think of attention (as one research group put it) as an
effect rather than as a cause (Krauzlis, Bollimunta, Arcizet, & Wang, 2014).
In other words, the term “attention” doesn’t refer to some mechanism in the
brain that produces a certain outcome. It’s better to think of attention as itself
an outcome — a byproduct of many other mechanisms.
As a related perspective, it may be helpful to think of paying attention, not as a process or mechanism, but as an achievement — something
that you’re able to do. Like many other achievements (e.g., doing well in
school, staying healthy), paying attention involves many elements, and the
exact set of elements needed will vary from one occasion to the next. In all
cases, though, multiple steps are needed to ensure that you end up being
aware of the stimuli you’re interested in, and not getting pulled off track by
irrelevant inputs.
TEST YOURSELF
6.In what ways does the
notion of a spotlight
beam accurately
reflect how spatial
attention functions?
7.In what ways does the
notion of a spotlight
beam differ from the
way spatial attention
functions?
8.When you first start
paying attention to an
input, your attention
seems to be spacebased. Once you’ve
learned a bit about
the input, though,
your attention seems
to be object-based.
How does this pattern
fit with the idea that
you pay attention by
anticipating the input?
Divided Attention
So far in this chapter, we’ve emphasized situations in which you’re trying to
focus on a single input. If other tasks and other stimuli were on the scene,
they were mere distractors. Sometimes, though, your goal is different: You
want to “multitask” — that is, deal with multiple tasks, or multiple inputs, all
at the same time. What can we say about this sort of situation — a situation
involving divided attention?
Sometimes divided attention is easy. Almost anyone can walk and sing
simultaneously; many people like to knit while they’re holding a conversation or listening to a lecture. It’s much harder, though, to do calculus homework while listening to a lecture; and trying to get your assigned reading
done while watching TV is surely a bad bet. What lies behind this pattern?
Why are some combinations difficult while others are easy?
Divided Attention
•
177
Our first step toward answering these questions is already in view. We’ve
proposed that perceiving requires resources that are in limited supply; the
same is presumably true for other tasks — remembering, reasoning, problem
solving. They, too, require resources, and without these resources the processes cannot go forward. What are the resources? The answer includes a
mix of things: certain mechanisms that do specific jobs, certain types of
memory that hold on to information while you’re working on it, energy supplies to keep the mental machinery going, and more. No matter what the
resources are, though, a task will be possible only if you have the needed
resources — just as a dressmaker can produce a dress only if he has the raw
materials, the tools, the time needed, the energy to run the sewing machine,
and so on.
All of this leads to a proposal: You can perform concurrent tasks only if
you have the resources needed for both. If the two tasks, when combined,
require more resources than you’ve got, then divided attention will fail.
The Specificity of Resources
CAESAR THE
MULTITASKER
Some writers complain about
the hectic pace of life today
and view it as a sad fact
about the pressured reality of the modern world. But
were things different in earlier times? More than 2,000
years ago, Julius Caesar was
praised for his ability to multi­
task. (The term is new, but
the capacity is not.) According to the Roman historian
Suetonius,
Caesar
could
write, dictate letters, and read
at the same time. Even on the
most important subjects, he
could dictate four letters at
once — and if he had nothing
else to do, as many as seven
letters at once. From a modern perspective, though, we
can ask: Is any of this plausible? Perhaps it is — some
people do seem especially
skilled at multitasking (Just &
Buchweitz, 2017; Redick et al.,
2016), and maybe Caesar was
one of those special people!
178 •
Imagine that you’re hoping to read a novel while listening to an academic
lecture. These tasks are different, but both involve the use of language, and
so it seems likely that these tasks will have overlapping resource requirements. As a result, if you try to do the tasks at the same time, they’re likely
to compete for resources — and therefore this sort of multitasking will
be difficult.
Now, think about two very different tasks, such as knitting and listening
to a lecture. These tasks are unlikely to interfere with each other. Even if all
your language-related resources are in use for the lecture, this won’t matter
for knitting, because it’s not a language-based task.
More broadly, the prediction here is that divided attention will be easier
if the simultaneous tasks are very different from each other, because different
tasks are likely to have distinct resource requirements. Resources consumed
by Task 1 won’t be needed for Task 2, so it doesn’t matter for Task 2 that
these resources are tied up in another endeavor.
Is this the pattern found in the research data? In an early study by Allport,
Antonis, and Reynolds (1972), participants heard a list of words presented through headphones into one ear, and their task was to shadow (i.e.,
repeat back) these words. At the same time, they were also presented with
a second list. No immediate response was required to the second list, but
later on, memory was tested for these items. In one condition, the second
list (the memory items) consisted of words presented into the other ear, so
the participants were hearing (and shadowing) a list of words in one ear
while simultaneously hearing the memory list in the other. In another condition, the memory items were presented visually. That is, while the participants were shadowing one list of words, they were also seeing a different list
of words on a screen before them. Finally, in a third condition, the memory
items consisted of pictures, also presented on a screen.
C H A P T E R F I V E Paying Attention
These three conditions had the same requirements — shadowing one list
while memorizing another. But the first condition (hear words + hear words)
involved very similar tasks; the second condition (hear words + see words)
involved less similar tasks; the third condition (hear words + see pictures),
even less similar tasks. On the logic we’ve discussed, we should expect the
most interference in the first condition and the least interference in the third.
And that is what the data showed (see Figure 5.18).
The Generality of Resources
Similarity among tasks, however, is not the whole story. If it were, then we’d
observe less and less interference as we consider tasks further and further apart.
Eventually, we’d find tasks so different from each other that we’d observe no
interference at all between them. But that’s not the pattern of the evidence.
Consider the common practice of talking on a cell phone while driving.
When you’re on the phone, the main stimulus information comes into your
ear, and your primary response is by talking. In driving, the main stimulation comes into your eyes, and your primary response involves control of
your hands on the steering wheel and your feet on the pedals. For the phone
conversation, you’re relying on language skills. For driving, you need spatial
skills. Overall, it looks like there’s little overlap in the specific demands of
these two tasks, and so little chance that the tasks will compete for resources.
Data show, however, that driving and cell-phone use do interfere with
each other. This is reflected, for example, in the fact that phone use has
been implicated in many automobile accidents (Lamble, Kauranen, Laakso,
& Summala, 1999). Even with a hands-free phone, drivers engaged in cellphone conversations are more likely to be involved in accidents, more likely
to overlook traffic signals, and slower to hit the brakes when they need to.
Errors in recognition
50
40
30
FIGURE 5.18 DIVIDED ATTENTION AMONG
DISTINCT TASKS
20
10
0
Words
Words Pictures
heard
seen
Type of remembered items
Participants perform poorly if they are trying to shadow one
list of words while hearing other words. They do somewhat
better if shadowing while seeing other words. They do better still
if shadowing while seeing pictures. In general, the greater the
difference between two tasks, the easier it will be to combine
the tasks.
(after allport, antonis, & reynolds, 1972)
Divided Attention
•
179
(See Kunar, Carter, Cohen, & Horowitz, 2008; Levy & Pashler, 2008;
Sanbonmatsu, Strayer, Biondi, Behrends, & Moore, 2016; Stothart, Mitchum,
& Yehnert, 2015; Strayer & Drews, 2007; Strayer, Drews, & Johnston, 2003.
For some encouraging data, though, on why phone-related accidents don’t
occur even more often, see Garrison & Williams, 2013; Medeiros-Ward,
Watson, & Strayer, 2015.)
As a practical matter, therefore, talking on the phone while driving is a
bad idea. In fact, some people estimate that the danger caused by driving
while on the phone is comparable to (and perhaps greater than) the risk of
driving while drunk. But, on the theoretical side, notice that the interference
observed between driving and talking is interference between two hugely
distinctive activities — a point that provides important information about
the nature of the resource competition involved, and therefore the nature of
mental resources.
Before moving on, we should mention that the data pattern is different
if the driver is talking to a passenger in the car rather than using the phone.
Conversations with passengers seem to cause little interference with driving (Drews, Pasupathi, & Strayer, 2008; see Figure 5.19), and the reason is
simple. If the traffic becomes complicated or the driver has to perform some
FIGURE 5.19
CELL PHONE USE AND DRIVING
.04
.02
A
No cell
phone
500
400
With cell
phone
Success in simple highway
navigation (percentage)
.06
0
100
600
Reaction time (ms)
Fraction of lights missed
.08
B
No cell
phone
90
80
70
60
50
40
30
20
10
0
With cell
phone
C
Drive while
talking with
a passenger
Drive while
conversing
via cell phone
Many studies show that driving performance is impaired when the driver is on the phone (whether hand-held or
hands-free). While on the phone, drivers are more likely to miss a red light (Panel A) and are slower in responding
to a red light (Panel B). Disruption is not observed, however, if the driver is conversing with a passenger rather
than on the phone (Panel C). That’s because the passenger is likely to adjust her conversation to accommodate
changes in driving — such as not speaking while the driver is navigating an obstruction.
(after strayer & johnston, 2001)
180 •
C H A P T E R F I V E Paying Attention
tricky maneuver, the passenger can see this — either by looking out of the
car’s window or by noticing the driver’s tension and focus. In these cases,
passengers helpfully slow down their side of the conversation, which takes
the load off of the driver, enabling the driver to focus on the road (Gaspar
et al., 2014; Hyman, Boss, Wise, McKenzie, & Caggiano, 2010; Nasar, Hecht,
& Wener, 2008).
Executive Control
The evidence is clear, then, that tasks as different as driving and talking compete with each other for some mental resource. But what is this resource,
which is apparently needed for both verbal tasks and spatial ones, tasks with
visual inputs and tasks with auditory inputs?
Evidence suggests that multiple resources may be relevant. Some authors
describe resources that serve (roughly) as an energy supply, drawn on by
all tasks (Eysenck, 1982; Kahneman, 1973; Lavie, 2001, 2005; Lavie, Lin,
Zokaei, & Thoma, 2009; MacDonald & Lavie, 2008; Murphy, Groeger, &
Greene, 2016). According to this perspective, tasks vary in the “load” they
put on you, and the greater the load, the greater the interference with other
tasks. In one study, drivers were asked to estimate whether their vehicle
would fit between two parked cars (Murphy & Greene, 2016). When the
judgment was difficult, participants were less likely to notice an unexpected
pedestrian at the side of the road. In other words, higher perceptual load
(from the driving task) increased inattentional blindness. (Also see Murphy
& Greene, 2017.)
Other authors describe mental resources as “mental tools” rather than
as some sort of mental “energy supply.” (See Allport, 1989; Baddeley, 1986;
Bourke & Duncan, 2005; Dehaene, Sergent, & Changeux, 2003; JohnsonLaird, 1988; Just, Carpenter, & Hemphill, 1996; Norman & Shallice, 1986;
Ruthruff, Johnston, & Remington, 2009; Vergaujwe, Barrouillet, & Camos,
2010. For an alternative perspective on these resources, see Franconeri,
2013; Franconeri, Alvarez, & Cavanagh, 2013.) One of these tools is
especially important, and it involves the mind’s executive control. This term
refers to the mechanisms that allow you to control your own thoughts, and
these mechanisms have multiple functions. Executive control helps keep your
current goals in mind, so that these goals (and not habit) will guide your
actions. The executive also ensures that your mental steps are organized into
the right sequence — one that will move you toward your goals. And if your
current operations aren’t moving you toward your goal, executive control
allows you to shift plans, or change strategy. (For discussion of how the
executive operates and how the brain tissue enables executive function, see
Brown, Reynolds, & Braver, 2007; Duncan et al., 2008; Gilbert & Shallice,
2002; Kane, Conway, Hambrick, & Engle, 2007; Miller & Cohen, 2001;
Miyake & Friedman, 2012; Shipstead, Harrison, & Engle, 2015; Stuss &
Alexander, 2007; Unsworth & Engle, 2007; Vandierendonck, Liefooghe, &
Verbruggen, 2010.)
CELL-PHONE DANGERS
FOR PEDESTRIANS
It’s not just driving that’s
disrupted by cell-phone use.
Compared to pedestrians
who aren’t using a phone,
pedestrians engaged in phone
conversations tend to walk
more slowly and more
erratically, and are less likely
to check traffic before they
cross a street. They’re also
less likely to notice things
along their path. In one
study, researchers observed
pedestrians walking across a
public square (Hyman, Boss,
Wise, McKenzie, & Caggiano,
2010). If the pedestrian was
walking with a friend (and
so engaged in a “live” conversation), there was a 71%
chance the pedestrian would
notice the unicycling clown
just off the pedestrian’s path.
But if the pedestrian was on
the phone (i.e., engaging in a
telephonic conversation), the
person had only a 25% chance
of detecting the clown.
Divided Attention
•
181
Executive control can only handle one task at a time, and this point
obviously puts limits on your ability to multitask — that is, to divide your
attention. But executive control is also important when you’re trying to do
just a single task. Evidence comes from studies of people who have suffered
damage to the prefrontal cortex (PFC), a brain area right behind the eyes
that seems crucial for executive control. People with this damage (including
Phineas Gage, whom we met in Chapter 2) can lead relatively normal lives,
because in their day-to-day behavior they can often rely on habit or can simply respond to prominent cues in their environment. With appropriate tests,
though, we can reveal the disruption that results from frontal lobe damage.
In one commonly used task, patients with frontal lesions are asked to sort a
deck of cards into two piles. At the start, the patients have to sort the cards
according to color; later, they need to switch strategies and sort according
to the shapes shown on the cards. The patients have enormous difficulty in
making this shift and continue to sort by color, even though the experimenter
tells them again and again that they’re placing the cards on the wrong piles
(Goldman-Rakic, 1998). This is referred to as a perseveration error, a tendency to produce the same response over and over even when it’s plain that
the task requires a change in the response.
These patients also show a pattern of goal neglect — failing to organize
their behavior in a way that moves them toward their goals. For example,
when one patient was asked to copy Figure 5.20A, the patient produced
the drawing shown in Figure 5.20B. The copy preserves features of the
FIGURE 5.20
A
GOAL NEGLECT
B
C
Patients who had suffered damage to the prefrontal cortex were asked to copy the drawing in Panel A. One
patient’s attempt is shown in Panel B; the drawing is reasonably accurate but seems to have been drawn with
no overall plan — for example, the large rectangle in the original, and the main diagonals, were created piecemeal rather than being used to organize the drawing. Another patient’s attempt is shown in Panel C; this patient
started to re-create the drawing but then got swept up in her own artistic impulses.
182 •
C H A P T E R F I V E Paying Attention
original, but close inspection reveals that the patient drew the copy with
no particular plan in mind. The large rectangle that defines the shape was
never drawn, and the diagonal lines that organize the figure were drawn in a
piecemeal fashion. Many details are correctly reproduced but weren’t drawn
in any sort of order; instead, these details were added whenever they happened to catch the patient’s attention (Kimberg, D’Esposito, & Farah, 1998).
Another patient, asked to copy the same figure, produced the drawing shown
in Figure 5.20C. This patient started to draw the figure in a normal way,
but then she got swept up in her own artistic impulses, adding stars and a
smiley face (Kimberg et al., 1998). (For more on executive control, see Aron,
2008; Courtney, Petit, Maisog, Ungerleider, & Haxby, 1998; Duncan et al.,
2008; Gilbert & Shallice, 2002; Huey, Krueger, & Grafman, 2006; Kane &
Engle, 2003; Kimberg et al., 1998; Logie & Della Salla, 2005; Ranganath &
Blumenfeld, 2005; Stuss & Levine, 2002; also Chapter 13.)
Divided Attention: An Interim Summary
Our consideration of selective attention drove us toward a several-part
account, with one mechanism serving to block out unwanted distractors and
other mechanisms promoting the processing of interesting stimuli. Now, in
our discussion of divided attention, we again need several elements in our
theory. Interference between tasks is increased if the tasks are similar to each
other, presumably because similar tasks overlap in their processing requirements and make competing demands on mental resources that are specialized
for that sort of task.
But interference can also be observed with tasks that are entirely different
from each other — such as driving and talking on a cell phone. Therefore, our
account needs to include resources that are general enough in their use that
they’re drawn on by almost any task. We’ve identified several of these general
resources: an energy supply needed for mental tasks, executive control, and
others as well. No matter what the resource, though, the key principle will
be the same: Tasks will interfere with each other if their combined demand
for a resource is greater than the amount available — that is, if the demand
exceeds the supply.
TEST YOURSELF
9.Why is it easier to
divide attention
between very different
activities (e.g., knitting while listening
to a lecture) than it
is to divide attention
between more similar
activities?
10.What is executive
control, and why does
it create limits on your
ability to divide your
attention between two
simultaneous tasks?
Practice
For a skilled driver, talking on a cell phone while driving is easy as long as
the driving is straightforward and the conversation is simple. Things fall
apart, though, the moment the conversation becomes complex or the driving becomes challenging. Engaged in deep conversation, the driver misses
a turn; while maneuvering through an intersection, the driver suddenly
stops talking.
The situation is different, though, for a novice driver. For someone who’s
just learning to drive, driving is difficult all by itself, even on a straight
road with no traffic. If we ask the novice to do anything else at the same
Practice
•
183
COGNITION
outside the lab
“I Can’t Ignore . . .”
At the start of this chapter, we refer to common
won’t be especially sensitive to the input when
experiences like this one: You’re trying to read a
it arrives.
book — perhaps an assignment for one of your
What about the conversation you’re overhear-
courses. You’re in a public place, though, and two
ing and hoping to ignore? Maybe it’s unfolding
people nearby are having a conversation. You have
according to a familiar script — for example, the
no interest in their conversation and really need
people behind you on the bus, or on the other side
to get through your reading assignment. Even so,
of the room, are discussing romance or a popular
you find yourself unable to ignore their conver-
movie. In this setting, with almost no thought you’ll
sation, so your reading doesn’t get done. What’s
easily anticipate where this distractor conversation
going on here? Why can’t you control what you’re
is going, and the anticipation will prime the rele-
paying attention to?
vant nodes in your mind, making you more sensi-
Think about the mechanisms that allow you to
tive to the input — the opposite of what you want.
pay attention. When you’re reading or listening
Part of our explanation, then, lies in ease-
to something, you do what you can to anticipate
of-anticipation. That’s why you’ll probably avoid
the upcoming input, and that anticipation lets
distraction if the material you’re trying to read is
you prime the relevant detectors so that they’ll
something you can anticipate (and so prime your-
be ready when the input arrives. As a result, the
self for) and if the irrelevant conversation involves
input falls onto “prepared” detectors and so you’re
content you can’t easily anticipate. In the extreme,
more sensitive to the input — more likely to notice
imagine that the irrelevant conversation is in some
it, more likely to process it.
foreign language that you don’t speak; here, be-
In contrast, you won’t try to anticipate inputs
you don’t care about. With no anticipation, the
cause there’s no basis for anticipation, the distraction will be minimal.
input falls on unprepared detectors — and so the
However, we need another element in our expla-
detectors are less sensitive to the input. In other
nation. Most people aren’t distracted if they try to
words, you’ve done nothing to make yourself
read while music is playing in the room or if there are
sensitive to these inputs, and so you’re relatively
traffic noises in the background. Why don’t these
insensitive to them, just as you wish.
(potential) distractors cause problems? Here, the
Now think about the situation with which
key is resource competition. Reading a book and
we began. Perhaps you’re trying to read some-
hearing a conversation both involve language, so
thing challenging. You’ll therefore have some
these activities draw on the same mental resources
difficulty anticipating how the passage will
and compete for those resources. But reading a
unfold — what words or phrases will be coming
book and hearing music (especially instrumen-
up. Therefore, you’ll have little basis for priming
tal music) draw on different mental resources, so
the soon-to-be-needed detectors, and so you
those activities don’t compete for resources.
184 •
C H A P T E R F I V E Paying Attention
time — whether it’s talking on a cell phone or even listening to the radio — we
put the driver (and other cars) at substantial risk. Why is this? Why are things
so different after practice?
Practice Diminishes Resource Demand
We’ve already said that mental tasks require resources, with the particular
resources required being dependent on the nature of the task. Let’s now add
another claim: As a task becomes more practiced, it requires fewer resources,
or perhaps it requires less frequent use of these resources.
This decrease in a task’s resource demand may be inevitable, given the
function of some resources. Consider executive control. We’ve mentioned
that this control plays little role if you can rely on habit or routine in performing a task. (That’s why Phineas Gage was able to live a mostly normal
life, despite his brain damage.) But early in practice, when a task is new, you
haven’t formed any relevant habits yet, so you have no habits to fall back on.
As a result, executive control is needed all the time. Once you’ve done the
task over and over, though, you do acquire a repertoire of suitable habits, and
so the demand for executive control decreases.
How will this matter? We’ve already said that tasks interfere with each
other if their combined resource demand is greater than the amount of resources available. Interference is less likely, therefore, if the “cognitive cost”
of a task is low. In that case, you’ll have an easier time accommodating the
task within your “resource budget.” And we’ve now added the idea that the
resource demand (the “cost”) will be lower after practice than before. Therefore, it’s no surprise that practice makes divided attention easier — enabling
the skilled driver to continue chatting with her passenger as they cruise down
the highway, even though this combination is hopelessly difficult for the
novice driver.
Automaticity
With practice in a task, then, the need for executive control is diminished, and
we’ve mentioned one benefit of this: Your control mechanisms are available
for other chores, allowing you to divide your attention in ways that would
have been impossible before practice. Let’s be clear, though, that this gain
comes at a price. With sufficient practice, task performance can go forward
with no executive control, and so the performance is essentially not controlled. This can create a setting in which the performance acts as a “mental
reflex,” going forward, once triggered, whether you like it or not.
Psychologists use the term automaticity to describe tasks that are well
practiced and involve little (or no) control. (For a classic statement, see
Shiffrin & Schneider, 1977; also Moors, 2016; Moors & De Houwer, 2006.)
The often-mentioned example is an effect known as Stroop interference. In
the classic demonstration of this effect, study participants are shown a series
of words and asked to name aloud the color of the ink used for each word.
The trick, though, is that the words themselves are color names. So people
Practice
•
185
might see the word “BLUE” printed in green ink and would have to say
“green” out loud, and so on (see Figure 5.21; Stroop, 1935).
This task turns out to be extremely difficult. There’s a strong tendency
to read the printed words themselves rather than to name the ink color, and
people make many mistakes in this task. Presumably, this reflects the fact
that word recognition, especially for college-age adults, is enormously well
practiced and therefore can proceed automatically. (For more on these issues,
including debate about what exactly automaticity involves, see Besner et al.,
2016; Durgin, 2000; Engle & Kane, 2004; Jacoby et al., 2003; Kane & Engle,
2003; Labuschagne & Besner, 2015; Moors, 2016.)
Where Are the Limits?
As we near the end of our discussion of attention, it may be useful again
to summarize where we are. Two simple ideas are key: First, tasks require
resources, and second, you can’t “spend” more resources than you have.
FIGURE 5.21
STROOP INTERFERENCE
Column A
Column B
As rapidly as you can, name out loud the colors of the ink in Column A.
(You’ll say, “black, green” and so on.) Next, do the same for Column B — again,
naming out loud the colors of the ink. You’ll probably find it much easier to
do this for Column A, because in Column B you experience interference from
the automatic habit of reading the words.
186 •
C H A P T E R F I V E Paying Attention
These claims are central for almost everything we’ve said about selective and
divided attention.
As we’ve seen, though, there are different types of resources, and the exact
resource demand of a task depends on several factors. The nature of the
task matters, of course, so that the resources required by a verbal task (e.g.,
reading) are different from those required by a spatial task (e.g., remembering a shape). The novelty of the task and the amount of flexibility the task
requires also matter. Connected to this, practice matters, with well-practiced
tasks requiring fewer resources.
What, then, sets the limits on divided attention? When can you do two
tasks at the same time, and when not? The answer varies, case by case. If two
tasks make competing demands on task-specific resources, the result will be
interference. If two tasks make competing demands on task-general resources
(the energy supply or executive control), again the result will be interference. Also, it will be especially difficult to combine tasks that involve similar stimuli — tasks that both involve printed text, for example, or that both
involve speech. The problem here is that these stimuli can “blur together,”
with a danger that you’ll lose track of which elements belong in which input
(“Was it the man who said ‘yes,’ or was it the woman?”; “Was the red dog in
the top picture or the bottom one?”). This sort of “crosstalk” (leakage of bits
of one input into the other input) can compromise your performance.
In short, it seems again like we need a multipart theory of attention, with
performance being limited by different factors at different times. This perspective draws us back to a claim we’ve made several times in this chapter:
Attention cannot be thought of as a skill or a mechanism. Instead, attention
is an achievement — an achievement of performing multiple activities simultaneously or an achievement of successfully avoiding distraction when you
want to focus on a single task. And this achievement rests on an intricate
base, so that many elements contribute to your ability to attend.
Finally, we have discussed various limits on human performance — that is,
limits on how much you can do at any one time. How rigid are these limits?
We’ve discussed the improvements in divided attention that are made possible by practice, but are there boundaries on what practice can accomplish?
Can you gain new mental resources or find new ways to accomplish a task
in order to avoid the bottleneck created by some limited resource? Some
evidence indicates that the answer may be yes; if so, many of the claims made
in this chapter must be understood as claims about what is usual, not about
what is possible. (See Hirst, Spelke, Reaves, Caharack, & Neisser, 1980;
Spelke, Hirst, & Neisser, 1976. For a neuroscience perspective, see Just &
Buchweitz, 2017.) With this, some traditions in the world — Buddhist meditation traditions, for example — claim it’s possible to train attention so that
one has better control over one’s mental life. How do these claims fit into
the framework we’ve developed in this chapter? These are issues in need of
exploration, and, in truth, what’s at stake here is a question about the boundaries on human potential, making these issues of deep interest for future
researchers to pursue.
TEST YOURSELF
11.Why does practice
decrease the resource
demand of a task?
12.Why does practice
create a situation in
which you can lose
control over your own
mental steps?
Practice
•
187
COGNITIVE PSYCHOLOGY AND EDUCATION
ADHD
When students learn about attention, they often have questions about failures
of attention: “Why can’t I focus when I need to?”; “Why am I so distracted
by my roommate moving around the room when I’m studying?”; “Why can
some people listen to music while they’re reading, but I can’t?”
One question comes up more than any other: “I [or “my friend” or “my
brother] was diagnosed with ADHD. What’s that all about?” This question
refers to a common diagnosis: attention-deficit/hyperactivity disorder. The disorder is characterized by a number of problems, including impulsivity, constant
fidgeting, and difficulty in keeping attention focused on a task. People with
ADHD hop from activity to activity and have trouble organizing or completing
projects. They sometimes have trouble following a conversation and are easily
distracted by an unimportant sight or sound. Even their own thoughts can
distract them — and so they can be pulled off track by their own daydreams.
The causes of ADHD are still unclear. Contributing factors that have been
mentioned include encephalitis, genetic influences, food allergies, high lead
concentrations in the environment, and more. The uncertainty about this
point comes from many sources, including some ambiguity in the diagnosis
of ADHD. There’s no question that there’s a genuine disorder, but diagnosis is
complicated by the fact that the disorder can vary widely in its severity. Some
people have relatively mild symptoms; others are massively disrupted, and
this variation can make diagnosis difficult.
In addition, some critics argue that in many cases the ADHD diagnosis
is just a handy label for children who are particularly active or who don’t
easily adjust to a school routine or a crowded classroom. Indeed, some critics
suggest that ADHD is often just a convenient categorization for physicians
or school counselors who don’t know how else to think about an especially
energetic child.
In cases in which the diagnosis is warranted, though, what does it involve? As we describe in the chapter, there are many steps involved in “paying
attention,” and some of those steps involve inhibition — so that we don’t follow
every stray thought, or every cue in the environment, wherever it may lead. For
most of us, this is no problem, and we easily inhibit our responses to most distractors. We’re thrown off track only by especially intrusive distractors — such
as a loud noise or a stimulus that has special meaning for us.
Some researchers propose, though, that people with ADHD have less
effective inhibitory circuits in their brains, making them more vulnerable to
momentary impulses and chance distractions. This is what leads to their scattered thoughts, their difficulty in schoolwork, and so on.
What can be done to help people with ADHD? One of the common treatments is Ritalin, a drug that is a powerful stimulant. It seems ironic that we’d
give a stimulant to people who are already described as too active and too
188 •
C H A P T E R F I V E Paying Attention
energetic, but the evidence suggests that Ritalin is effective in treating actual
cases of ADHD — plausibly because the drug activates the inhibitory circuits
within the brain, helping the person to guard against wayward impulses.
However, we probably shouldn’t rely on Ritalin as the sole treatment for
ADHD. One reason is the risk of overdiagnosis — it’s worrisome that this
drug may be routinely given to people, including young children, who don’t
actually have ADHD. Also, there are concerns about the long-term effects
and possible side effects of Ritalin, and this certainly motivates us to seek
other forms of treatment. (Common side effects include weight loss, insomnia, anxiety, and slower growth during childhood.) Some of the promising
alternatives involve restructuring of the environment. If people with ADHD
are vulnerable to distraction, we can help them by the simple step of reducing the sources of distraction in their surroundings. Likewise, if people with
ADHD are influenced by whatever cues they detect, we can surround them
with helpful cues — reminders of what they’re supposed to be doing and the
tasks they’re supposed to be working on. These simple interventions do seem
to be helpful, especially with adults diagnosed with ADHD.
Overall, then, our description of ADHD requires multiple parts. Researchers have considered a diverse set of causes, and there may be a diverse set
of psychological mechanisms involved in the disorder. (See pp. 168.) The
diagnosis probably is overused, but the diagnosis is surely genuine in many
cases, and the problems involved in ADHD are real and serious. Medication
can help, but we’ve noted the concern about side effects of the medication.
Environmental interventions can also help and may, in fact, be the best bet
for the long term, especially given the important fact that in most cases the
symptoms of ADHD diminish as the years go by.
For more on this topic . . .
Barkley, R. A. (2004). Adolescents with ADHD: An overview of empirically
based treatments. Journal of Psychiatric Practice, 10, 39–56.
Barkley, R. A. (2008). ADHD in adults: What the science says. New York, NY:
Guilford.
Zillmer, E. A., Spiers, M. V., & Culbertson, W. C. (2008). Principles of neuropsychology. Belmont, CA: Wadsworth.
Cognitive Psychology and Education
•
189
chapter review
SUMMARY
• People are often oblivious to unattended inputs;
they usually cannot tell if an unattended auditory
input was coherent prose or random words, and
they often do not detect unattended visual inputs,
even though such inputs are right in front of their
eyes. However, some aspects of the unattended
inputs are detected. For example, people can report
on the pitch of the unattended sound and whether
it contained human speech or some other sort
of noise. Sometimes they can also detect stimuli
that are especially meaningful; some people, for
example, hear their own name if it is spoken on the
unattended channel.
• These results suggest that perception may require the commitment of mental resources, with
some of these resources helping to prime the detectors needed for perception. This proposal is supported by studies of inattentional blindness — that
is, studies showing that perception is markedly
impaired if the perceiver commits no resources to
the incoming stimulus information. The proposal
is also supported by results showing that participants perceive more efficiently when they can anticipate the upcoming stimulus (and so can prime
the relevant detectors). In many cases, the anticipation is spatial — if, for example, participants
know that a stimulus is about to arrive at a particular location. This priming, however, seems to
draw on a limited-capacity system, with the result
that priming one stimulus or one position takes
away resources that might be spent on priming
some other stimulus.
• The ability to pay attention to certain regions
of space has caused many researchers to compare
190
attention to a spotlight beam, with the idea that
stimuli falling “within the beam” are processed
more efficiently than stimuli that fall “outside the
beam.” However, this spotlight analogy is potentially misleading. In many circumstances, people do
seem to devote attention to identifiable regions of
space, no matter what falls within those regions. In
other circumstances, attention seems to be objectbased, not space-based, and so people pay attention
to specific objects, not specific positions.
• Perceiving seems to require the commitment of
resources, and so do most other mental activities.
This observation suggests an explanation for the
limits on divided attention: It is possible to perform
two tasks simultaneously only if the two tasks do
not in combination demand more resources than
are available. Some of the relevant mental resources,
including executive control, are task-general, being
required in a wide variety of mental activities. Other
mental resources are task-specific, being required
only for tasks of a certain type.
• Divided attention is influenced by practice, with
the result that it is often easier to divide attention
between familiar tasks than between unfamiliar
tasks. In the extreme, practice may produce automaticity, in which a task seems to require virtually
no mental resources but is also difficult to control.
One proposal is that automaticity results from the
fact that decisions are no longer needed for a wellpracticed routine; instead, one can simply run off
the entire routine, doing on this occasion just what
one did on prior occasions.
KEY TERMS
selective attention (p. 150)
dichotic listening (p. 150)
attended channel (p. 150)
unattended channel (p. 150)
shadowing (p. 151)
filter (p. 152)
fixation target (p. 153)
inattentional blindness (p. 154)
change blindness (p. 155)
early selection hypothesis (p. 157)
late selection hypothesis (p. 157)
repetition priming (p. 160)
limited-capacity system (p. 164)
mental resources (p. 164)
spatial attention (p. 165)
endogenous control of attention (p. 172)
exogenous control of attention (p. 172)
unilateral neglect syndrome (p. 174)
divided attention (p. 177)
executive control (p. 181)
perseveration error (p. 182)
goal neglect (p. 182)
automaticity (p. 185)
Stroop interference (p. 185)
TEST YOURSELF AGAIN
1.What information do people reliably pick up
from the attended channel? What do they pick
up from the unattended channel?
2.How is inattentional blindness demonstrated?
What situations outside of the laboratory seem
to reflect inattentional blindness?
3.What evidence seems to confirm early selection?
What evidence seems to confirm late selection?
4.What are the differences between the way that
stimulus-based priming functions and the way
that expectation-based priming functions?
8.When you first start paying attention to an
input, your attention seems to be space-based.
Once you’ve learned a bit about the input,
though, your attention seems to be objectbased. How does this pattern fit with the idea
that you pay attention by anticipating the
input?
9.Why is it easier to divide attention between
very different activities (e.g., knitting while
listening to a lecture) than it is to divide
attention between more similar activities?
5.Why is there a “cost” associated with being
misled by expectation-based priming?
10.What is executive control, and why does it
create limits on your ability to divide your
attention between two simultaneous tasks?
6.In what ways does the notion of a spotlight
beam accurately reflect how spatial attention
functions?
11.Why does practice decrease the resource
demand of a task?
7.In what ways does the notion of a spotlight
beam differ from the way spatial attention
functions?
12.Why does practice create a situation in which
you can lose control over your own mental
steps?
191
THINK ABOUT IT
1. It’s easy to keep your attention focused on
materials that you understand. But if you try to
focus on difficult material, your mind is likely
to wander. Does the chapter help you in understanding why that is? Explain your response.
help those who do it become better at paying
attention — staying focused and not suffering
from distraction. Does the chapter help you
in understanding why that might be? Explain
your response.
2. People claim that some forms of meditation
training (including Buddhist meditation) can
E eBOOK DEMONSTRATIONS & ESSAYS
Go to http://digital.wwnorton.com/cognition7 for the online demonstrations and essays
relevant to this chapter. These can be found in the ebook.
Online Demonstrations
• Demonstration 5.1: Shadowing
• Demonstration 5.2: Color-Changing Card Trick
• Demonstration 5.3: The Control of Eye
Online Applying Cognitive Psychology and the
Law Essays
• Cognitive Psychology and the Law: Guiding the
Formulation of New Laws
Movements
• Demonstration 5.4: Automaticity and the
Stroop Effect
COGNITION LABS
Go to http://digital.wwnorton.com/cognition7 to experience interactive online psychology
labs relevant to this chapter.
192
Memory
3
part
A
s you move through life, you encounter new facts, gain new skills, and
have new experiences. And you’re often changed by all of this, so that
later on you know things and can do things that you couldn’t know or do
before. How do these changes happen? How do you get new information into your
memory, and then how do you retrieve this information when you need it? And
how much trust can you put in this process? Why, for example, do people some-
times not remember things (including important things)? And why are memories
sometimes wrong — so that in some cases you remember an event one way, but a
friend who was present at the same event remembers things differently?
We’ll tackle all these issues in this section, and they will lead us to theoretical claims and practical applications. We’ll offer suggestions, for example,
about how students should study their class materials to maximize retention,
and also what students can do later on so that they’ll retain things they learned
at some earlier point.
In our discussion, several themes will emerge again and again. One theme
concerns the active nature of learning, and we’ll discuss the fact that passive
exposure to information, with no intellectual engagement, leads to poor
memory. From this base, we’ll consider why some forms of engagement with
to-be-learned material lead to especially good memory but other forms do not.
A second theme concerns the role of memory connections. In Chapter 6,
we’ll see that learning involves the creation of connections, and the more connections formed, the better the learning. In Chapter 7, we’ll argue that these
connections can serve as “retrieval paths” — paths that, you hope, will lead you
from your memory search’s starting point to the information you’re trying to
recall. As we’ll see, this idea has clear implications for when you will remember
a previous event and when you won’t.
Chapter 8 then explores a different aspect of the connections idea: Memory
connections can actually be a source of memory errors. We’ll ask what this
means for memory accuracy overall, and we’ll discuss what you can do to minimize error and to improve the completeness and accuracy of your memory.
193
6
chapter
The Acquisition of
Memories and the
Working-Memory
System
what if…
Clive Wearing is an accomplished musician and a
scholar of Renaissance music. When he was 47 years
old, however, his brain was horribly damaged by a Herpes virus, and
he now has profound amnesia. He is still articulate and intelligent, able
to participate in an ongoing conversation, and still able to play music
(beautifully) and conduct. But ever since the viral infection, he’s been
unable to form new memories. He can’t recall any of the experiences
he’s had in the thirty years since he suffered the brain damage. He can’t
even remember events that happened just moments ago, with a bizarre
result: Every few minutes, Wearing realizes he can’t recall anything from
a few seconds back, and so he concludes that he must have just woken
up. He grabs his diary and writes “8:31 a.m. Now I am really, completely
awake.” A short while later, though, he again realizes he can’t recall the
last seconds, so he decides that now he has just woken up. He picks up
his diary to record this event and immediately sees his previous entry.
Puzzled, he crosses it out and replaces it with “9:06 a.m. Now I am perfectly, overwhelmingly awake.” But then the process repeats, and so this
entry, too, gets scribbled out and a new entry reads “9:34 a.m.: Now I am
superlatively, actually awake.”
Despite this massive disruption, Wearing’s love for his wife, Deborah,
has not in any way been diminished by his amnesia. But here, too,
the memory loss has powerful effects. Each time Deborah enters his
room — even if she’s been away just a few minutes — he races to embrace
her as though it’s been countless lonely months since they last met. If
asked directly, he has no recollection of her previous visits — including a
visit that might have happened just minutes earlier.
We met a different case of amnesia in Chapter 1 — the famous patient
H.M. He, too, was unable to recall his immediate past — and the many,
deep problems this produced included an odd sort of disorientation: If
you were smiling at him, was it because you’d just said something funny?
Or because he’d said something embarrassing? Or had you been smiling
all along? As H.M. put it, “Right now, I’m wondering. Have I done or said
anything amiss? You see, at this moment, everything looks clear to me,
but what happened just before? That’s what worries me” (Milner, 1970,
p. 37; also see Corkin, 2013).
Cases like these remind us how profoundly our memories shape our
everyday lives. But these cases also raise many questions: Why is it that
195
A
B
CLIVE WEARING
Clive Wearing (shown here with his wife) developed profound amnesia as a result of
viral encephalitis, and now he seems to have only a moment-to-moment consciousness. With no memory at all of what he was doing just seconds ago, he is often convinced he just woke up, and he repeatedly writes in his diary, “Now perfectly awake
(1st time).” On the diary page shown here, he has recorded his thought, at 5:42 a.m.,
as his “1st act” because he has no memory of any prior activity. Soon after, though,
he seems to realize again that he has no memory of any earlier events, and so he
scribbles out the entry and now records that his bath is his “1st act.” The sequence
repeats over and over, with Wearing never recalling what he did before his current
activity, and so he records act after act as his “first.”
196 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
preview of chapter themes
•
e begin the chapter with a discussion of the broad archiW
tecture of memory. We then turn to a closer examination
of one component of this architecture: working memory.
•
e emphasize the active nature of working memory —
W
activity that is especially evident when we discuss working memory’s “central executive,” a mental resource that
serves to order, organize, and control our mental lives.
•
he active nature of memory is also evident in the proT
cess of rehearsal. Rehearsal is effective only if the person
engages the materials in some way; this is reflected,
for example, in the contrast between deep processing
(which leads to excellent memory) and mere maintenance
rehearsal (which produces basically no memory benefit).
•
ctivity during learning appears to establish memory
A
connections, which can serve as retrieval routes when it
comes time to remember the target material. For complex
material, the best way to establish these connections is
to seek to understand the material; the better the understanding, the better the memory will be.
Wearing still remembers who his wife is? How is it possible that even
with his amnesia, Wearing remains such a talented musician? Why does
H.M. still remember his young adult years, even though he can’t remember what he said just five minutes earlier? We’ll tackle questions like
these in this chapter and the next two.
Acquisition, Storage, and Retrieval
How does new information — whether it’s a friend’s phone number or a fact
you hope to memorize for the bio exam — become established in memory?
Are there ways to learn that are particularly effective? Then, once information is in storage, how do you locate it and “reactivate” it later? And why
does search through memory sometimes fail — so that, for example, you
forget the name of that great restaurant downtown (but then remember the
name when you’re midway through a mediocre dinner someplace else)?
In tackling these questions, there’s a logical way to organize our inquiry.
Before there can be a memory, you need to gain, or “acquire,” some new
information. Therefore, acquisition — the process of gaining information
and placing it into memory — should be our first topic. Then, once you’ve
acquired this information, you need to hold it in memory until the information is needed. We refer to this as the storage phase. Finally, you remember. In
other words, you somehow locate the information in the vast warehouse that
is memory and you bring it into active use; this is called retrieval.
This organization seems logical; it fits, for example, with the way most
“electronic memories” (e.g., computers) work. Information (“input”) is provided to a computer (the acquisition phase). The information then resides
in some dormant form, generally on the hard drive or perhaps in the cloud
(the storage phase). Finally, the information can be brought back from this
dormant form, often via a search process that hunts through the disk (the
retrieval phase). And there’s nothing special about the computer comparison
here; “low-tech” information storage works the same way. Think about a file
Acquisition, Storage, and Retrieval
•
197
TEST YOURSELF
1.Define the terms
“acquisition,” “storage,”
and “retrieval.”
drawer — information is acquired (i.e., filed), rests in this or that folder, and
then is retrieved.
Guided by this framework, we’ll begin our inquiry by focusing on the
acquisition of new memories, leaving discussion of storage and retrieval for
later. As it turns out, though, we’ll soon find reasons for challenging this overall approach to memory. In discussing acquisition, for example, we might
wish to ask: What is good learning? What guarantees that material is firmly
recorded in memory? As we’ll see, evidence indicates that what counts as
“good learning” depends on how the memory is to be used later on, so that
good preparation for one kind of use may be poor preparation for a different kind of use. Claims about acquisition, therefore, must be interwoven
with claims about retrieval. These interconnections between acquisition and
retrieval will be the central theme of Chapter 7.
In the same way, we can’t separate claims about memory acquisition from
claims about memory storage. This is because how you learn (acquisition)
depends on what you already know (information in storage). We’ll explore
this important relationship in both this chapter and Chapter 8.
We begin, though, in this chapter, by describing the acquisition process.
Our approach will be roughly historical. We’ll start with a simple model,
emphasizing data collected largely in the 1970s. We’ll then use this as the
framework for examining more recent research, adding refinements to the
model as we proceed.
The Route into Memory
For many years, theorizing in cognitive psychology focused on the process
through which information was perceived and then moved into memory
storage — that is, on the process of information acquisition. One early proposal was offered by Waugh and Norman (1965). Later refinements were
added by Atkinson and Shiffrin (1968), and their version of the proposal
came to be known as the modal model. Figure 6.1 provides a simplified
depiction of this model.
Updating the Modal Model
According to the modal model, when information first arrives, it is stored
briefly in sensory memory. This form of memory holds on to the input in
“raw” sensory form — an iconic memory for visual inputs and an echoic
memory for auditory inputs. A process of selection and interpretation then
moves the information into short-term memory — the place where you hold
information while you’re working on it. Some of the information is then
transferred into long-term memory, a much larger and more permanent storage place.
This proposal captures some important truths, but it needs to be updated
in several ways. First, the idea of “sensory memory” plays a much smaller
role in modern theorizing, so modern discussions of perception (like our
198 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
FIGURE 6.1
N INFORMATION-PROCESSING VIEW
A
OF MEMORY
Retrieval
Sensory
memory
Incoming
information
Short-term
memory
Maintenance
via rehearsal
Long-term
memory
Lost
Diagrams like this one depict the flow of information hypothesized by the
modal model. The model captures many important truths but must be
updated in important ways. Current theorizing, for example, emphasizes that
short-term memory (now called “working memory”) is not a place serving as
a “loading dock” outside of long-term memory. Instead, working memory is
best understood as an activity, in ways described in the chapter.
ICONIC MEMORY
discussion in Chapters 2 and 3) often make no mention of this memory.
(For a recent assessment of visual sensory memory, though, see Cappiello
& Zhang, 2016.) Second, modern proposals use the term working memory
rather than “short-term memory,” to emphasize the function of this memory.
Ideas or thoughts in this memory are currently activated, currently being
thought about, and so they’re the ideas you’re currently working on. Longterm memory (LTM), in contrast, is the vast repository that contains all of
your knowledge and all of your beliefs — most of which you aren’t thinking
about (i.e., aren’t working on) at this moment.
The modal model also needs updating in another way. Pictures like the
one in Figure 6.1 suggest that working memory is a storage place, sometimes described as the “loading dock” just outside of the long-term memory
“warehouse.” The idea is that information has to “pass through” working
memory on the way into longer-term storage. Likewise, the picture implies
that memory retrieval involves the “movement” of information out of storage
and back into working memory.
In contrast, contemporary theorists don’t think of working memory as a
“place” at all. Instead, working memory is (as we will see) simply the name we
give to a status. Therefore, when we say that ideas are “in working memory,”
we simply mean that these ideas are currently activated and being worked on
by a specific set of operations.
In a classic experiment (Sperling,
1960), participants viewed a
grid like this one for just 50 ms.
If asked to report all of the
letters, participants could report just three or four of them.
In a second condition, participants saw the grid and then
immediately afterward heard
a cue signaling which row
they had to report. No matter which row they were asked
about, participants could recall most of the row’s letters.
It seems, therefore, that participants could remember the
entire display (in iconic memory) for a brief time, and could
“read off” the contents of any
row when appropriately cued.
The limitation in the report-all
condition, then, came from the
fact that iconic memory faded
away before the participants
could report on all of it.
The Route into Memory
•
199
We’ll have more to say about this modern perspective before we’re
through. It’s important to emphasize, though, that contemporary thinking
also preserves some key ideas from the modal model, including its claims
about how working memory and long-term memory differ from each other.
Let’s identify those differences.
First, working memory is limited in size; long-term memory is enormous.
In fact, long-term memory has to be enormous, because it contains all of
your knowledge — including specific knowledge (e.g., how many siblings you
have) and more general themes (e.g., that water is wet, that Dublin is in
Ireland, that unicorns don’t exist). Long-term memory also contains all of
your “episodic” knowledge — that is, your knowledge about events, including events early in your life as well as more recent experiences.
Second, getting information into working memory is easy. If you think
about a particular idea or some other type of content, then you’re “working
on” that idea or content, and so this information — by definition — is now in
your working memory. In contrast, we’ll see later in the chapter that getting
information into long-term memory often involves some work.
Third, getting information out of working memory is also easy. Since (by
definition) this memory holds the ideas you’re thinking about right now, the
information is already available to you. Finding information in long-term
memory, in contrast, can sometimes be difficult and slow — and in some
settings can fail completely.
Fourth, the contents of working memory are quite fragile. Working
memory, we emphasize, contains the ideas you’re thinking about right now.
If your thoughts shift to a new topic, therefore, the new ideas will enter
working memory, pushing out what was there a moment ago. Long-term
memory, in contrast, isn’t linked to your current thoughts, so it’s much less
fragile — information remains in storage whether you’re thinking about it
right now or not.
We can make all these claims more concrete by looking at some classic
research findings. These findings come from a task that’s quite artificial (i.e.,
not the sort of memorizing you do every day) but also quite informative.
Working Memory and Long-Term Memory:
One Memory or Two?
In many studies, researchers have asked participants to listen to a series of
words, such as “bicycle, artichoke, radio, chair, palace.” In a typical experiment, the list might contain 30 words and be presented at a rate of one word
per second. Immediately after the last word is read, the participants must
repeat back as many words as they can. They are free to report the words in
any order they choose, which is why this task is called a free recall procedure.
People usually remember 12 to 15 words in this test, in a consistent pattern.
They’re very likely to remember the first few words on the list, something
known as the primacy effect, and they’re also likely to remember the last
few words on the list, a recency effect. The resulting pattern is a U-shaped
200 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
100
Recency
effect
Percent recall
80
FIGURE 6.2 PRIMACY AND RECENCY
EFFECTS IN FREE RECALL
60
Primacy
effect
40
20
0
1
5
10
15
Serial position
20
Research participants in this study heard a list of 20 common words presented at a rate of one word per second.
Immediately after hearing the list, participants were asked to
write down as many of the words on the list as they could
recall. The results show that position in the series strongly
affected recall—participants had better recall for words at the
beginning of the list (the primacy effect) and for words at
the end of the list (the recency effect), compared to words
in the middle of the list.
curve describing the relation between positions within the series — or serial
position — and the likelihood of recall (see Figure 6.2; Baddeley & Hitch,
1977; Deese & Kaufman, 1957; Glanzer & Cunitz, 1966; Murdock, 1962;
Postman & Phillips, 1965).
Explaining the Recency Effect
What produces this pattern? We’ve already said that working memory contains the material someone is working on at just that moment. In other words,
this memory contains whatever the person is currently thinking about; and
during the list presentation, the participants are thinking about the words
they’re hearing. Therefore, it’s these words that are in working memory. This
memory, however, is limited in size, capable of holding only five or six words.
Consequently, as participants try to keep up with the list presentation, they’ll
be placing the words just heard into working memory, and this action will
bump the previous words out of working memory. As a result, as participants
proceed through the list, their working memories will, at each moment, contain only the half dozen words that arrived most recently. Any words that
arrived earlier than these will have been pushed out by later arrivals.
Of course, the last few words on the list don’t get bumped out of working memory, because no further input arrives to displace them. Therefore, when the list presentation ends, those last few words stay in place.
Moreover, our hypothesis is that materials in working memory are readily
available — easily and quickly retrieved. When the time comes for recall,
then, working memory’s contents (the list’s last few words) are accurately
and completely recalled.
The key idea, then, is that the list’s last few words are still in working
memory when the list ends (because nothing has arrived to push out these
items), and we know that working memory’s contents are easy to retrieve.
This is the source of the recency effect.
The Route into Memory
•
201
Explaining the Primacy Effect
The primacy effect has a different source. We’ve suggested that it takes
some work to get information into long-term memory (LTM), and it seems
likely that this work requires some time and attention. So let’s examine how
participants allocate their attention to the list items. As participants hear
the list, they do their best to be good memorizers, and so when they hear
the first word, they repeat it over and over to themselves (“bicycle, bicycle,
bicycle”) — a process known as memory rehearsal. When the second word
arrives, they rehearse it, too (“bicycle, artichoke, bicycle, artichoke”). Likewise for the third (“bicycle, artichoke, radio, bicycle, artichoke, radio”), and
so on through the list. Note, though, that the first few items on the list are
privileged. For a brief moment, “bicycle” is the only word participants have
to worry about, so it has 100% of their attention; no other word receives
this privilege. When “artichoke” arrives a moment later, participants divide
their attention between the first two words, so “artichoke” gets only 50%
of their attention — less than “bicycle” got, but still a large share of the
participants’ efforts. When “radio” arrives, it has to compete with “bicycle”
and “artichoke” for the participants’ time, and so it receives only 33% of
their attention.
Words arriving later in the list receive even less attention. Once six or
seven words have been presented, the participants need to divide their
attention among all these words, which means that each one receives only
a small fraction of the participants’ focus. As a result, words later in the
list are rehearsed fewer times than words early in the list — a fact that
can be confirmed simply by asking participants to rehearse out loud
(Rundus, 1971).
This view of things leads immediately to our explanation of the primacy
effect — that is, the observed memory advantage for the early list items.
These early words didn’t have to share attention with other words (because
the other words hadn’t arrived yet), so more time and more rehearsal were
devoted to them than to any others. This means that the early words have a
greater chance of being transferred into LTM — and so a greater chance of
being recalled after a delay. That’s what shows up in these classic data as the
primacy effect.
Testing Claims about Primacy and Recency
This account of the serial-position curve leads to many predictions. First, we’re
claiming the recency portion of the curve is coming from working memory,
while other items on the list are being recalled from LTM. Therefore, manipulations of working memory should affect recall of the recency items but not
items earlier in the list. To see how this works, consider a modification of our
procedure. In the standard setup, we allow participants to recite what they
remember immediately after the list’s end. But instead, we can delay recall
by asking participants to perform some other task before they report the list
items — for example, we can ask them to count backward by threes, starting
from 201. They do this for just 30 seconds, and then they try to recall the list.
202 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
We’ve hypothesized that at the end of the list working memory still contains the last few items heard from the list. But the task of counting backward will itself require working memory (e.g., to keep track of where you
are in the counting sequence). Therefore, this chore will displace working
memory’s current contents; that is, it will bump the last few list items out of
working memory. As a result, these items won’t benefit from the swift and
easy retrieval that working memory allows, and, of course, that retrieval was
the presumed source of the recency effect. On this basis, the simple chore of
counting backward, even if only for a few seconds, will eliminate the recency
effect. In contrast, the counting backward should have no impact on recall
of the items earlier in the list: These items are (by hypothesis) being recalled
from long-term memory, not working memory, and there’s no reason to
think the counting task will interfere with LTM. (That’s because LTM, unlike
working memory, isn’t dependent on current activity.)
Figure 6.3 shows that these predictions are correct. An activity interpolated,
or inserted, between the list and recall essentially eliminates the recency effect,
but it has no influence elsewhere in the list (Baddeley & Hitch, 1977; Glanzer
& Cunitz, 1966; Postman & Phillips, 1965). In contrast, merely delaying the
recall for a few seconds after the list’s end, with no interpolated activity, has
no impact. In this case, participants can continue rehearsing the last few items
during the delay and so can maintain them in working memory. With no new
materials coming in, nothing pushes the recency items out of working memory,
and so, even with a delay, a normal recency effect is observed.
100
30-second
unfilled
delay
90
Percentage of words recalled
80
Immediate
recall
70
60
50
30-second
filled
delay
40
30
20
10
0
1
2,3
4,5
6,7
Input position
8,9
10,11
12
FIGURE 6.3 THE IMPACT
OF INTERPOLATED ACTIVITY
ON THE RECENCY EFFECT
With immediate recall (the red line in
the figure), or if recall is delayed by
30 seconds with no activity during
the delay (the purple line), a strong
recency effect is detected. In contrast, if participants spend 30 seconds on some other activity between
hearing the list and the subsequent
memory test (the blue line), the
recency effect is eliminated. This interpolated activity has no impact on
the pre-recency portion of the curve
(i.e., the portion of the curve other
than the last few positions).
The Route into Memory
•
203
FIGURE 6.4 RATE OF LIST
PRESENTATION AND THE
SERIAL-POSITION EFFECT
Presenting the to-be-remembered materials at a slower rate improves pre-recency
performance but has no effect on recency.
The slow presentation rate in this case was
9 seconds per item; the faster rate was
3 seconds per item.
TEST YOURSELF
2.List the four ways in
which (either in the
modal model or in
more recent views)
working memory is
different from longterm storage.
3.How is the primacy
effect usually
explained? How is
the recency effect
usually explained?
204 •
Percentage of words recalled
100
75
Slow presentation
50
25
Fast presentation
1
5
10
15
20
Serial position
We’d expect a different outcome, though, if we manipulate long-term
memory rather than working memory. In this case, the manipulation should
affect all performance except for recency (which, again, is dependent on
working memory, not LTM). For example, what happens if we slow down the
presentation of the list? Now, participants will have more time to spend on
all of the list items, increasing the likelihood of transfer into more permanent
storage. This should improve recall for all items coming from LTM. Working memory, in contrast, is limited by its size, not by ease of entry or ease of
access. Therefore, the slower list presentation should have no influence on
working-memory performance. Research results confirm these claims: Slowing the list presentation improves retention of all the pre-recency items but
does not improve the recency effect (see Figure 6.4).
Other variables that influence long-term memory have similar effects.
Using more familiar or more common words, for example, would be expected to ease entry into long-term memory and does improve pre-recency
retention, but it has no effect on recency (Sumby, 1963).
It seems, therefore, that the recency and pre-recency portions of the
curve are influenced by distinct sets of factors and obey different principles.
Apparently, then, these two portions of the curve are the products of different mechanisms, just as our theory proposed. In addition, fMRI scans suggest
that memory for early items on a list depends on brain areas (in and around
the hippocampus) that are associated with long-term memory; memory
for later items on the list do not show this pattern (Talmi, Grady, GoshenGottstein, & Moscovitch, 2005; also Eichenbaum, 2017; see Figure 6.5). This
provides further confirmation for our memory model.
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
FIGURE 6.5
RAIN REGIONS SUPPORTING WORKING
B
MEMORY AND LONG-TERM MEMORY
Retrieval from long-term
memory specifically
activates the hippocampus.
Retrieval from working
memory specifically activates
the perirhinal cortex.
We can confirm the distinction between working memory and long-term
memory with fMRI scans. These scans suggest that memory for early items
on a list depends on brain areas (in and around the hippocampus) that are
associated with long-term memory; memory for later items on the list do not
show this pattern.
(talmi, grady, goshen-gottstein, & moscovitch, 2005)
A Closer Look at Working Memory
Earlier, we counted four fundamental differences between working memory and LTM — the size of these two stores, the ease of entry, the ease of
retrieval, and the fact that working memory is dependent on current activity
(and therefore fragile) while LTM is not. These are all points proposed by the
modal model and preserved in current thinking. As we’ve said, though, investigators’ understanding of working memory has developed over the years.
Let’s examine the newer conception in more detail.
The Function of Working Memory
Virtually all mental activities require the coordination of several pieces of
information. Sometimes the relevant bits come into view one by one, so that
you need to hold on to the early-arrivers until the rest of the information is
available, and only then weave all the bits together. Alternatively, sometimes
the relevant bits are all in view at the same time — but you still need to hold
on to them together, so that you can think about the relations and combinations. In either case, you’ll end up with multiple ideas in your thoughts, all
A Closer Look at Working Memory
•
205
activated simultaneously, and thus several bits of information in the status we
describe as “in working memory.” (For more on how you manage to focus on
these various bits, see Oberauer & Hein, 2012.)
Framing things in this way makes it clear how important working memory is: You use it whenever you have multiple ideas in your mind, multiple
elements that you’re trying to combine or compare. Let’s now add that
people differ in the “holding capacity” of their working memories. Some
people are able to hold on to (and work with) more elements, and some
with fewer. How does this matter? To find out, we first need a way of
measuring working memory’s capacity, to determine if your memory capacity is above average, below, or somewhere in between. The procedure for
obtaining this measurement, however, has changed over the years; looking
at this change will help clarify what working memory is, and what working
memory is for.
Digit Span
For many years, the holding capacity of working memory was measured with
a digit-span task. In this task, research participants hear a series of digits read
to them (e.g., “8, 3, 4”) and must immediately repeat them back. If they do so
successfully, they’re given a slightly longer list (e.g., “9, 2, 4, 0”). If they can
repeat this one without error, they’re given a still longer list (“3, 1, 2, 8, 5”),
and so on. The procedure continues until the participant starts to make
errors — something that usually happens when the list contains more than
seven or eight items. The number of digits the person can echo back without
errors is referred to as that person’s digit span.
Procedures such as this imply that working memory’s capacity is typically around seven items — at least five and probably not more than nine.
These estimates have traditionally been summarized by the statement that
this memory holds “7 plus-or-minus 2” items (Chi, 1976; Dempster, 1981;
Miller, 1956; Watkins, 1977).
However, we immediately need a refinement of these measurements.
If working memory can hold 7 plus-or-minus 2 items, what exactly is an
“item”? Can people remember seven sentences as easily as seven words?
Seven letters as easily as seven equations? In a classic paper, George Miller
(one of the founders of the field of cognitive psychology) proposed that
working memory holds 7 plus-or-minus 2 chunks (Miller, 1956). The term
“chunk” doesn’t sound scientific or technical, and that’s useful because this
informal terminology reminds us that a chunk doesn’t hold a fixed quantity
of information. Instead, Miller proposed, working memory holds 7 plusor-minus 2 packages, and what those packages contain is largely up to the
individual person.
The flexibility in how people “chunk” input can easily be seen in the span
test. Imagine that we test someone’s “letter span” rather than their “digit
span,” using the procedure already described. So the person might hear
206 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
“R, L” and have to repeat this sequence back, and then “F, C, H,” and so on.
Eventually, let’s imagine that the person hears a much longer list, perhaps
one starting “H, O, P, T, R, A, S, L, U. . . .” If the person thinks of these as
individual letters, she’ll only remember 7 of them, more or less. But she
might reorganize the list into “chunks” and, in particular, think of the letters as forming syllables (“HOP, TRA, SLU, . . .”). In this case, she’ll still
remember 7 plus-or-minus 2 items, but the items are syllables, and by remembering the syllables she’ll be able to report back at least a dozen letters
and probably more.
How far can this process be extended? Chase and Ericsson (1982;
Ericsson, 2003) studied a remarkable individual who happens to be a fan of
track events. When he hears numbers, he thinks of them as finishing times
for races. The sequence “3, 4, 9, 2,” for example, becomes “3 minutes and
49.2 seconds, near world-record mile time.” In this way, four digits become
one chunk of information. This person can then retain 7 finishing times
(7 chunks) in memory, and this can involve 20 or 30 digits! Better still, these
chunks can be grouped into larger chunks, and these into even larger chunks.
For example, finishing times for individual racers can be chunked together
into heats within a track meet, so that, now, 4 or 5 finishing times (more than
a dozen digits) become one chunk. With strategies like this and a lot of practice, this person has increased his apparent memory span from the “normal”
7 digits to 79 digits.
However, let’s be clear that what has changed through practice is merely
this person’s chunking strategy, not the capacity of working memory itself.
This is evident in the fact that when tested with sequences of letters, rather
than numbers, so that he can’t use his chunking strategy, this individual’s
memory span is a normal size — just 6 consonants. Thus, the 7-chunk limit
is still in place for this man, even though (with numbers) he’s able to make
extraordinary use of these 7 slots.
Operation Span
Chunking provides one complication in our measurement of working memory’s capacity. Another — and deeper — complication grows out of the very
nature of working memory. Early theorizing about working memory, as we
said, was guided by the modal model, and this model implies that working
memory is something like a box in which information is stored or a location
in which information can be displayed. The traditional digit-span test fits
well with this idea. If working memory is like a box, then it’s sensible to ask
how much “space” there is in the box: How many slots, or spaces, are there
in it? This is precisely what the digit span measures, on the idea that each
digit (or each chunk) is placed in its own slot.
We’ve suggested, though, that the modern conception of working memory is more dynamic — so that working memory is best thought of as a
status (something like “currently activated”) rather than a place. (See, e.g.,
A Closer Look at Working Memory
•
207
FIGURE 6.6
IS WORKING MEMORY A “PLACE”?
Verbal and numbers
Objects
Spatial
Problem solving
Modern theorists argue that working memory is not a place at all, but is
instead the name we give for a certain set of mental activities. Consistent with
this modern view, there’s no specific location within the brain that serves as
working memory. Instead, working memory is associated with a wide range
of brain sites, as shown here.
(after cabeza & nyberg, 2000)
Christophel, Klink, Spitzer, Roelfsema, & Haynes, 2017; also Figure 6.6.)
On this basis, perhaps we need to rethink how we measure this memory’s
capacity — seeking a measure that reflects working memory’s active
operation.
Modern researchers therefore measure this memory’s capacity in terms
of operation span, a measure of working memory when it is “working.”
There are several ways to measure operation span, with the types differing in what “operation” they use (e.g., Bleckley, Foster, & Engle, 2015;
Chow & Conway, 2015). One type is reading span. To measure this span,
a research participant might be asked to read aloud a series of sentences,
like these:
Due to his gross inadequacies, his position as director was terminated
abruptly.
It is possible, of course, that life did not arise on Earth at all.
Immediately after reading the sentences, the participant is asked to recall
each sentence’s final word — in this case, “abruptly” and “all.” If she can
do this with these two sentences, she’s asked to do the same task with a
group of three sentences, and then with four, and so on, until the limit on
208 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
FIGURE 6.7
DYNAMIC MEASURES OF WORKING MEMORY
(7
7) + 1 = 50; dog
(10/2) + 6 = 10; gas
(4
2) + 1 = 9; nose
(3/1) + 1 = 5; beat
(5/5) + 1 = 2; tree
Operation span can be measured in several different ways. In one procedure,
participants must announce whether each of these “equations” is true or
false, and then recall the words appended to each equation. If participants
can do this with two equations, we ask them to do three; if they can do that,
we ask them to try four. By finding out how far they can go, we measure their
working-memory capacity.
her performance is located. This limit defines the person’s working-memory
capacity, or WMC. (However there are other ways to measure operation
span — see Figure 6.7.)
Let’s think about what this task involves: storing materials (the ending
words) for later use in the recall test, while simultaneously working with
other materials (the full sentences). This juggling of processes, as the participant moves from one part of the task to the next, is exactly what working
memory must do in day-to-day life. Therefore, performance in this test is
likely to reflect the efficiency with which working memory will operate in
more natural settings.
Is operation span a valid measure — that is, does it measure what it’s supposed to? Our hypothesis is that someone with a higher operation span has
a larger working memory. If this is right, then someone with a higher span
should have an advantage in tasks that make heavy use of this memory.
Which tasks are these? They’re tasks that require you to keep multiple ideas
active at the same time, so that you can coordinate and integrate various
bits of information. So here’s our prediction: People with a larger span (i.e.,
a greater WMC) should do better in tasks that require the coordination of
different pieces of information.
Consistent with this claim, people with a greater WMC do have an advantage in many settings — in tests of reasoning, assessments of reading comprehension, standardized academic tests (including the verbal SAT), tasks that
require multitasking, and more. (See, e.g., Ackerman, Beier, & Boyle, 2002;
Butler, Arrington, & Weywadt, 2011; Daneman & Hannon, 2001; Engle &
Kane, 2004; Gathercole & Pickering, 2000; Gray, Chabris, & Braver, 2003;
Redick et al., 2016; Salthouse & Pink, 2008. For some complications, see
A Closer Look at Working Memory
•
209
Chow & Conway, 2015; Harrison, Shipstead, & Engle, 2015; Kanerva &
Kalakoski, 2016; Mella, Fagot, Lecert, & de Ribaupierre, 2015.)
These results convey several messages. First, the correlations between
WMC and performance provide indications about when it’s helpful to have
a larger working memory, which in turn helps us understand when and how
working memory is used. Second, the link between WMC and measures of
intellectual performance provide an intriguing hint about what we’re measuring with tests (like the SAT) that seek to measure “intelligence.” We’ll return
to this issue in Chapter 13 when we discuss the nature of intelligence. Third,
it’s important that the various correlations are observed with the more active
measure of working memory (operation span) but not with the more traditional (and more static) span measure. This point confirms the advantage of
the more dynamic measures and strengthens the idea that we’re now thinking
about working memory in the right way: not as a passive storage box, but
instead as a highly active information processor.
The Rehearsal Loop
Working memory’s active nature is also evident in another way: in the actual
structure of this memory. The key here is that working memory is not a single
entity but is instead, a system built of several components (Baddeley, 1986,
1992, 2012; Baddeley & Hitch, 1974; also see Logie & Cowan, 2015). At
the center of the working-memory system is a set of processes we discussed
in Chapter 5: the executive control processes that govern the selection and
sequence of thoughts. In discussions of working memory, these processes
have been playfully called the “central executive,” as if there were a tiny
agent embedded in your mind, running your mental operations. Of course,
there is no agent, and the central executive is just a name we give to the set of
mechanisms that do run the show.
The central executive is needed for the “work” in working memory; if you
have to plan a response or make a decision, these steps require the executive.
But in many settings, you need less than this from working memory. Specifically, there are settings in which you need to keep ideas in mind, not because
you’re analyzing them right now but because you’re likely to need them soon.
In this case, you don’t need the executive. Instead, you can rely on the executive’s “helpers,” leaving the executive free to work on more difficult matters.
Let’s focus on one of working memory’s most important helpers, the
articulatory rehearsal loop. To see how the loop functions, try reading the
next few sentences while holding on to these numbers: “1, 4, 6, 3.” Got them?
Now read on. You’re probably repeating the numbers over and over to yourself, rehearsing them with your inner voice. But this takes very little effort, so
you can continue reading while doing this rehearsal. Nonetheless, the moment
you need to recall the numbers (what were they?), they’re available to you.
In this setting, the four numbers were maintained by working memory’s
rehearsal loop, and with the numbers thus out of the way, the central executive could focus on the processes needed for reading. That is the advantage of
210 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
this system: With mere storage handled by the helpers, the executive is available for other, more demanding tasks.
To describe this sequence of events, researchers would say that you used
subvocalization — silent speech — to launch the rehearsal loop. This production by the “inner voice” produced a representation of the target numbers in
the phonological buffer, a passive storage system used for holding a representation (essentially an “internal echo”) of recently heard or self-produced
sounds. In other words, you created an auditory image in the “inner ear.” This
image started to fade away after a second or two, but you then subvocalized
the numbers once again to create a new image, sustaining the material in this
buffer. (For a glimpse of the biological basis for the “inner voice” and “inner
ear,” see Figure 6.8.)
FIGURE 6.8
B
RAIN ACTIVITY AND WORKING-MEMORY
REHEARSAL
Verbal
memory
Spatial
memory
Left lateral
Superior
Right lateral
Color is used here as an indication of increased brain activity (measured
in this case by positron emission tomography). When research participants
are doing a verbal memory task (and using the articulatory loop), activation
increases in areas ordinarily used for language production and perception.
A very different pattern is observed when participants are doing a task requiring memory for spatial position. Notice, then, that the “inner voice” and
“inner ear” aren’t casual metaphors; instead, they involve mechanisms that
are ordinarily used for overt speech and actual hearing.
( after jonides , lacey , & nee , 2005; also see jonides et al ., 2008)
A Closer Look at Working Memory
•
211
Many lines of evidence confirm this proposal. For example, when people
are storing information in working memory, they often make “sound-alike”
errors: Having heard “F,” they’ll report back “S.” When trying to remember the name “Tina,” they’ll slip and recall “Deena.” The problem isn’t that
people mis-hear the inputs at the start; similar sound-alike confusions emerge
if the inputs are presented visually. So, having seen “F,” people are likely to
report back “S”; they aren’t likely in this situation to report back the similarlooking “E.”
What produces this pattern? The cause lies in the fact that for this task
people are relying on the rehearsal loop, which involves a mechanism (the
“inner ear”) that stores the memory items as (internal representations of)
sounds. It’s no surprise, therefore, that errors, when they occur, are shaped by
this mode of storage.
As a test of this claim, we can ask people to take the span test while
simultaneously saying “Tah-Tah-Tah” over and over, out loud. This concurrent
articulation task obviously requires the mechanisms for speech production. Therefore, those mechanisms are not available for other use, including subvocalization. (If you’re directing your lips and tongue to produce the
“Tah-Tah-Tah” sequence, you can’t at the same time direct them to produce
the sequence needed for the subvocalized materials.)
How does this constraint matter? First, note that our original span test
measured the combined capacities of the central executive and the loop. That
is, when people take a standard span test (as opposed to the more modern
measure of operation span), they store some of the to-be-remembered items
in the loop and other items via the central executive. (This is a poor use of the
executive, underutilizing its talents, but that’s okay here because the standard
span task doesn’t require anything beyond mere storage.)
With concurrent articulation, though, the loop isn’t available for use, so
we’re now measuring the capacity of working memory without the rehearsal
loop. We should predict, therefore, that concurrent articulation, even though
it’s extremely easy, should cut memory span drastically. This prediction turns
out to be correct. Span is ordinarily about seven items; with concurrent
articulation, it drops by roughly a third — to four or five items (Chincotta &
Underwood, 1997; see Figure 6.9).
Second, with visually presented items, concurrent articulation should
eliminate the sound-alike errors. Repeatedly saying “Tah-Tah-Tah” blocks
use of the articulatory loop, and it’s in this loop, we’ve proposed, that the
sound-alike errors arise. This prediction, too, is correct: With concurrent
articulation and visual presentation of the items, sound-alike errors are
largely eliminated.
The Working-Memory System
As we have mentioned, your working memory contains the thoughts and
ideas you’re working on right now, and often this means you’re trying to
keep multiple ideas in working memory all at the same time. That can cause
212 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
Control
Suppression
10
Digit span (no. of items)
9
8
7
6
5
4
Chinese
English
Finnish
Greek
Language
Spanish
Swedish
FIGURE 6.9 THE EFFECT
OF CONCURRENT
ARTICULATION ON SPAN
In the Control condition, participants
were given a normal digit-span
test. In the Suppression condition, participants were required
to do concurrent articulation
while taking the test. Concurrent
articulation is easy, but it blocks
use of the articulatory loop and
consistently decreases memory
span, from roughly seven items
to five or so. And, plainly, this
use of the articulatory loop is
not an occasional strategy; instead, it can be found in a wide
range of countries and languages.
(after chincotta & underwood, 1997)
difficulties, because working memory only has a small capacity. That’s why
working memory’s helpers are so important, because they substantially
increase working memory’s capacity.
Against this backdrop, it’s not surprising that the working-memory system relies on other helpers in addition to the rehearsal loop. For example,
the system also relies on the visuospatial buffer, used for storing visual
materials such as mental images, in much the same way that the rehearsal
loop stores speech-based materials. (We’ll have more to say about mental
images in Chapter 11.) Baddeley (the researcher who launched the idea of a
working-memory system) has also proposed another component of the system: the episodic buffer. This component is proposed as a mechanism that
helps the executive organize information into a chronological sequence — so
that, for example, you can keep track of a story you’ve just heard or a film
clip you’ve just seen (e.g., Baddeley, 2000, 2012; Baddeley & Wilson, 2002;
Baddeley, Eysenck, & Anderson, 2009). The role of this component is evident
in patients with profound amnesia who seem unable to put new information into long-term storage, but who still can recall the flow of narrative in
a story they just heard. This short-term recall, it seems, relies on the episodic
buffer — an aspect of working memory that’s unaffected by the amnesia.
In addition, other helpers can be documented in some groups of people.
Consider people who have been deaf since birth and communicate via sign
language. We wouldn’t expect these individuals to rely on an “inner voice”
and an “inner ear” — and they don’t. People who have been deaf since birth
A Closer Look at Working Memory
•
213
rely on a different helper for working memory: They use an “inner hand”
(and covert sign language) rather than an “inner voice” (and covert speech).
As a result, they are disrupted if they’re asked to wiggle their fingers during
a memory task (similar to a hearing person saying “Tah-Tah-Tah”), and they
also tend to make “same hand-shape” errors in working memory (similar to
the sound-alike errors made by the hearing population).
The Central Executive
TEST YOURSELF
4.What does it mean
to say that working
memory holds seven
(plus-or-minus two)
“chunks”? What is a
chunk?
5.What evidence suggests that operation span is a better
measure of working
memory than the
more standard digitspan measure?
6.How does the rehearsal loop manage
to hold on to information with only occasional involvement by
the central executive?
What can we say about the main player within the working-memory
system — the central executive? In our discussion of attention (in Chapter 5),
we argued that executive control processes are needed to govern the sequence
of thoughts and actions; these processes enable you to set goals, make plans for
reaching those goals, and select the steps needed for implementing those plans.
Executive control also helps whenever you want to rise above habit or routine,
in order to “tune” your words or deeds to the current circumstances.
For purposes of the current chapter, though, let’s emphasize that the same
processes control the selection of ideas that are active at any moment in time.
And, of course, these active ideas (again, by definition) constitute the contents of working memory. It’s inevitable, then, that we would link executive
control with this type of memory.
With all these points in view, we’re ready to move on. We’ve now updated the modal model (Figure 6.1) in important ways, and in particular
we’ve abandoned the notion of a relatively passive short-term memory serving largely as storage container. We’ve shifted to a dynamic conception of
working memory, with the proposal that this term is merely the name for an
organized set of activities — especially the complex activities of the central
executive together with its various helpers.
But let’s also emphasize that in this modern conception, just as in the
modal model, working memory is quite fragile. Each shift in attention brings
new information into working memory, and the newly arriving material
displaces earlier items. Storage in this memory, therefore, is temporary.
Obviously, then, we also need some sort of enduring memory storage, so that
we can remember things that happened an hour, or a day, or even years ago.
Let’s turn, therefore, to the functioning of long-term memory.
Entering Long-Term Storage:
The Need for Engagement
We’ve already seen an important clue regarding how information gets established in long-term storage: In discussing the primacy effect, we suggested
that the more an item is rehearsed, the more likely you are to remember
that item later. To pursue this point, though, we need to ask what exactly
rehearsal is and how it might work to promote memory.
214 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
Two Types of Rehearsal
The term “rehearsal” doesn’t mean much beyond “thinking about.” In other
words, when a research participant rehearses an item on a memory list, she’s
simply thinking about that item — perhaps once, perhaps over and over; perhaps
mechanically, or perhaps with close attention to what the item means. Therefore, there’s considerable variety within the activities that count as rehearsal,
and psychologists find it useful to sort this variety into two broad types.
As one option, people can engage in maintenance rehearsal, in which they
simply focus on the to-be-remembered items themselves, with little thought
about what the items mean or how they relate to one another. This is a rote,
mechanical process, recycling items in working memory by repeating them
over and over. In contrast, relational, or elaborative, rehearsal involves thinking about what the to-be-remembered items mean and how they’re related to
one another and to other things you already know.
Relational rehearsal is vastly superior to maintenance rehearsal for
establishing information in memory. In fact, in many settings maintenance
rehearsal provides no long-term benefit at all. As an informal demonstration
of this point, consider the following experience (although, for a formal demonstration of this point, see Craik & Watkins, 1973). You’re watching your
favorite reality show on TV. The announcer says, “To vote for Contestant #4,
text 4 to 21523 from your mobile phone!” You reach into your pocket for
your phone but realize you left it in the other room. So you recite the number
to yourself while scurrying for your phone, but then, just before you dial,
you see that you’ve got a text message. You pause, read the message, and
then you’re ready to dial, but . . . you don’t have a clue what the number was.
What went wrong? You certainly heard the number, and you rehearsed it a
couple of times while moving to grab your phone. But despite these rehearsals,
the brief interruption from reading the text message seems to have erased the
number from your memory. However, this isn’t ultra-rapid forgetting. Instead,
you never established the number in memory in the first place, because in
this setting you relied only on maintenance rehearsal. That kept the number
in your thoughts while you were moving across the room, but it did nothing
to establish the number in long-term storage. And when you try to dial the
number after reading the text message, it’s long-term storage that you need.
The idea, then, is that if you think about something only in a mindless and mechanical way, the item won’t be established in your long-term
memory. Similarly, long-lasting memories aren’t created simply by repeated
exposures to the items to be remembered. If you encounter an item over and
over but, at each encounter, barely think about it (or think about it only in
a mechanical way), then this, too, won’t produce a long-term memory. As
a demonstration, consider the ordinary penny. Adults in the United States
have probably seen pennies tens of thousands of times. Adults in other
countries have seen their own coins just as often. If sheer exposure is what
counts for memory, people should remember perfectly what these coins
look like.
WE DON’T REMEMBER
THINGS WE DON’T
PAY ATTENTION TO
To promote public safety,
many buildings have fire
extinguishers and automatic
defibrillators positioned in
obvious and easily accessible
locations. But in a moment of
need, will people in the building remember where this
safety equipment is located?
Will they even remember
that the safety equipment
is conveniently available?
Research suggests they may
not. Occupants of the building have passed by the safety
equipment again and again—
but have had no reason to
notice the equipment. As a
result, they’re unlikely to remember where the equipment is located. (After Castel,
Vendetti, & Holyoak, 2012)
Entering Long-Term Storage: The Need for Engagement
•
215
But, of course, most people have little reason to pay attention to the penny.
Pennies are a different color from the other coins, so they can be identified
at a glance without further scrutiny. And, if it’s scrutiny that matters for
memory — or, more broadly, if we remember what we pay attention to and
think about — then memory for the coin should be quite poor.
The evidence on this point is clear: People’s memory for the penny is
remarkably bad. For example, most people know that Lincoln’s head is on
the “heads” side, but which way is he facing? Is it his right cheek that’s visible
or his left? What other markings are on the coin? Most people do very badly
with these questions; their answers to the “Which way is he facing?” question
are close to random (Nickerson & Adams, 1979). And performance is similar
for people in other countries remembering their own coins. (Also see Bekerian
& Baddeley, 1980; Rinck, 1999, for a much more consequential example.)
As a related example, consider the logo that identifies Apple products — the
iPhone, the iPad, or one of the Apple computers. Odds are good that you’ve
seen this logo hundreds and perhaps thousands of time, but you’ve probably had no reason to pay attention to its appearance. The prediction, then,
is that your memory for the logo will be quite poor — and this prediction is
correct. In one study, only 1 of 85 participants was able to draw the logo
correctly — with the bite on the proper side, the stem tilted the right way, and
the dimple properly placed in the logo’s bottom (Blake, Nazarian, & Castel,
2015; see Figure 6.10). And — surprisingly — people who use an Apple
computer (and therefore see the logo every time they turn on the machine)
perform at a level not much better than people who use a PC.
The Need for Active Encoding
Apparently, it takes some work to get information into long-term memory.
Merely having an item in front of your eyes isn’t enough — even if the item
is there over and over and over. Likewise, repeatedly thinking about an item
doesn’t, by itself, establish a memory. That’s evident in the fact that maintenance rehearsal seems ineffective at promoting memory.
FIGURE 6.10 MEMORY FOR AN
OFTEN-VIEWED LOGO
Most people have seen the Apple logo
countless times, but they’ve had no reason to
pay attention to its features. As a result, they
have poor memories for the features. Test
yourself. Can you find the correct version
among the options displayed here?
(the answer is at the end of the chapter.)
216 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
Further support for these claims comes from studies of brain activity during learning. In several procedures, researchers have used fMRI recording to
keep track of the moment-by-moment brain activity in participants who were
studying a list of words (Brewer, Zhao, Desmond, Glover, & Gabrieli, 1998;
Wagner, Koutstaal, & Schacter, 1999; Wagner et al., 1998; also see Levy,
Kuhl, & Wagner, 2010). Later, the participants were able to remember some
of the words they had learned, but not others, which allowed the investigators to return to their initial recordings and compare brain activity during the
learning process for words that were later remembered and words that were
later forgotten. Figure 6.11 shows the results, with a clear difference, during
the initial encoding, between these two types of words. Greater levels of brain
FIGURE 6.11
BRAIN ACTIVITY DURING LEARNING
Learn a series of words,
and, during learning,
record the neural
response to each word.
Based on what happened
at Time 2, go back and
examine the data from
Time 1, looking separately
at what happened during
learning for words that
were later remembered,
and what happened during
learning for words
that were later forgotten.
Test memory for
the words.
A
Left medial temporal lobe
Left inferior prefrontal cortex
Remembered
Forgotten
3
2
1
0
–1
0
B
Remembered
Forgotten
4
Activity level
Activity level
4
3
2
1
0
–1
4
Time (s)
8
12
0
4
8
Time (s)
12
(Panel A) Participants in this study were given a series of words to memorize, and their brain activity was
recorded during this initial presentation. These brain scans were then divided into two types: those showing
brain activity during the encoding of words that were remembered in the subsequent test, and those showing
activity during encoding of words that were forgotten in the test. (Panel B) As the figure shows, activity levels
during encoding were higher for the later-remembered words than they were for the later-forgotten words. This
finding confirms that whether a word is forgotten or not depends on participants’ mental activity when they
encountered the word in the first place.
Entering Long-Term Storage: The Need for Engagement
•
217
activity (especially in the hippocampus and regions of the prefrontal cortex)
were reliably associated with greater probabilities of retention later on.
These fMRI results are telling us, once again, that learning is not a passive process. Instead, activity is needed to lodge information into long-term
memory, and, apparently, higher levels of this activity lead to better memory. But this raises some new questions: What is this activity? What does it
accomplish? And if — as it seems — maintenance rehearsal is a poor way to
memorize, what type of rehearsal is more effective?
Incidental Learning, Intentional Learning,
and Depth of Processing
Consider a student taking a course in college. The student knows that her
memory for the course materials will be tested later (e.g., in the final exam).
And presumably she’ll take various steps to help herself remember: She may
read through her notes again and again; she may discuss the material with
friends; she may try outlining the material. Will these various techniques
work — so that she’ll have a complete and accurate memory when the exam
takes place? And notice that the student is taking these steps in the context
of wanting to memorize; she wants to do well on the exam! How does this
motivation influence performance? In other words, how does the intention to
memorize influence how or how well material is learned?
In an early experiment, participants in one condition heard a list of
24 words; their task was to remember as many as they could. This is
intentional learning — learning that is deliberate, with an expectation that
memory will be tested later. Other groups of participants heard the same
24 words but had no idea that their memories would be tested. This allows us
to examine the impact of incidental learning — that is, learning in the absence
of any intention to learn. One of the incidental-learning groups was asked
simply, for each word, whether the word contained the letter e. A different
incidental-learning group was asked to look at each word and to report how
many letters it contained. Another group was asked to consider each word
and to rate how pleasant it seemed.
Later, all the participants were tested — and asked to recall as many of the
words as they could. (The test was as expected for the intentional-learning
group, but it was a surprise for the other groups.) The results are shown in
Figure 6.12A (Hyde & Jenkins, 1969). Performance was relatively poor for
the “Find the e” and “Count the letters” groups but appreciably better for
the “How pleasant?” group. What’s striking, though, is that the “How pleasant?” group, with no intention to memorize, performed just as well as the
intentional-learning (“Learn these!”) group. The suggestion, then, is that the
intention to learn doesn’t add very much; memory can be just as good without
this intention, provided that you approach the materials in the right way.
This broad pattern has been reproduced in many other experiments (to
name just a few: Bobrow & Bower, 1969; Craik & Lockhart, 1972; Hyde &
Jenkins, 1973; Jacoby, 1978; Lockhart, Craik, & Jacoby, 1976; Parkin, 1984;
218 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
Slamecka & Graf, 1978). As one example, consider a study by Craik and
Tulving (1975). Their participants were led to do incidental learning (i.e.,
they didn’t know their memories would be tested). For some of the words
shown, the participants did shallow processing — that is, they engaged the
material in a superficial way. Specifically, they had to say whether the word
was printed in CAPITAL letters or not. (Other examples of shallow processing would be decisions about whether the words are printed in red or
in green, high or low on the screen, etc.) For other words, the participants
had to do a moderate level of processing: They had to judge whether each
word shown rhymed with a particular cue word. Finally, for other words,
participants had to do deep processing. This is processing that requires some
thought about what the words mean; specifically, Craik and Tulving asked
whether each word shown would fit into a particular sentence.
The results are shown in Figure 6.12B. Plainly, there is a huge effect of
level of processing, with deeper processing (i.e., more attention to meaning)
leading to better memory. In addition, Craik and Tulving (and many other
researchers) have confirmed the Hyde and Jenkins finding that the intention
FIGURE 6.12
Activity during first exposure
(Craik & Tulving, 1975)
24
25
12
6
10
5
U
pp
or er
lo ca
w se
er
?
es
th
rn
Le
a
ow
15
0
e!
?
nt
sa
pl
ea
le
H
ou
nt
t
C
Fi
n
d
he
th
e
tte
“e
rs
.
.”
0
20
B
rh D
ym oe
e… s it
se F ?
nt it
en in
ce th
… is
?
18
Percent of words recalled
Number of words recalled
7.What is the difference
between maintenance
rehearsal and relational (or elaborative)
rehearsal?
8.What does it mean to
say, “It doesn’t matter
if you intend to memo­
rize; all that matters for
memory is how exactly
you engage the material
you encounter”?
9.What is deep processing, and what
impact does it have on
memory?
THE IMPACT OF DEEPER PROCESSING
Activity during first exposure
(Hyde & Jenkins, 1969)
A
TEST YOURSELF
The two sets of results shown here derive from studies described in the text, but they are part of an avalanche
of data confirming the broad pattern: Shallow processing leads to poor memory. Deeper processing (paying
attention to meaning) leads to much better memory. And what matters seems to be the level of engagement;
the specific intention to learn (because participants know their memory will be tested later on) contributes little.
Entering Long-Term Storage: The Need for Engagement
•
219
COGNITION
outside the lab
Gender Differences?
Most of this book focuses on principles that apply
memory—but only when remembering the faces
to all people—young or old, sociable or shy, smart
of other women. All these differences probably
or slow. But, of course, people differ in many ways,
reflect the “attention priorities” that Western cul-
leading us to ask: Are there differences in how
ture encourages for men and women, priorities
people remember? As one aspect of this issue, do
that derive from the conventional roles assumed
men and women differ in their memories?
(for better or worse) for each gender.
Let’s emphasize at the start that there’s no
Some results also suggest that women may
overall difference between the genders in mem-
have better memory for day-to-day events,
ory accuracy, or quantity of information retained,
especially emotional events; but this, too, might
or susceptibility to outside influences that might
be a difference in attention rather than a true
pull memory off track. (See Chapter 8 for more
difference in memory. Women are, in Western
on the influences on memory.) If we take a closer
culture, encouraged to pay attention to social
look, though, we do find some differences (e.g.,
dynamics and in many settings are encouraged to
Herlitz & Rehnman, 2014)—with some studies
be more emotionally responsive and more emo-
suggesting an advantage for women in remem-
tionally sensitive than men. It may be these points
bering verbal materials, and other studies sug-
that color how women pay attention to and think
gesting an advantage for men in remembering
about an event—and ultimately how they remem-
spatial arrangement.
ber the event.
Other differences in what the genders remem-
It seems likely, then, that most of these dif-
ber are the consequence of cultural factors. Bear
ferences (none of them profound) are a direct
in mind that people tend to remember what they
reflection of cultural bias. Even so, the differences
paid attention to, and don’t remember things they
do underscore some important messages. First,
didn’t attend. From this base, it’s not surprising
with regard to their cognition, men and women
that after viewing an event, women are more likely
are much more similar to each other than they are
than men to recall the clothing people were wear-
different. Second, it’s crucial to bear in mind that
ing or their jewelry. Men, in contrast, are more likely
what you remember now is dependent on what
than women to recall the people’s body shapes.
you paid attention to earlier. Therefore, if people
(Also see Chapter 5, pp. 170–173.) There is even
differ in what they focus on, they’ll remember
some indication that women may have better face
different things later on.
to learn adds little. That is, memory performance is roughly the same in
conditions in which participants do shallow processing with an intention to
memorize, and in conditions in which they do shallow processing without
this intention. Likewise, the outcome is the same whether people do deep
processing with the intention to memorize or without. In study after study,
what matters is how people approach the material they’re seeing or hearing.
220 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
It’s that approach — that manner of engagement — that determines whether
memory will be excellent or poor later on. The intention to learn seems, by
itself, not to matter.
The Role of Meaning and
Memory Connections
The message so far seems clear: If you want to remember the sentences you’re
reading in this text, or the materials you’re learning in the training sessions
at your job, you should pay attention to what these materials mean. That is,
you should try to do deep processing. And if you do deep processing, it won’t
matter if you’re trying hard to memorize the materials (intentional learning)
or merely paying attention to the meaning because you find the material
interesting, with no plan for memorizing (incidental learning).
But what lies behind these effects? Why does attention to meaning lead to
good recall? Let’s start with a broad proposal; we’ll then fill in the evidence
for this proposal.
Connections Promote Retrieval
Perhaps surprisingly, the benefits of deep processing may not lie in the learning process itself. Instead, deep processing may influence subsequent events.
More precisely, attention to meaning may help you by facilitating retrieval
of the memory later on. To understand this point, consider what happens
whenever a library acquires a new book. On its way into the collection, the
new book must be catalogued and shelved appropriately. These steps happen when the book arrives, but the cataloguing doesn’t literally influence
the arrival of the book into the building. The moment the book is delivered,
it’s physically in the library, catalogued or not, and the book doesn’t become
“more firmly” or “more strongly” in the library because of the cataloguing.
Even so, the cataloguing is crucial. If the book were merely tossed on a
random shelf somewhere, with no entry in the catalogue, users might never
be able to find it. Without a catalogue entry, users of the library might not
even realize that the book was in the building. Notice, then, that cataloguing happens at the time of arrival, but the benefit of cataloguing isn’t for
the arrival itself. (If the librarians all went on strike, so that no books were
being catalogued, books would continue to arrive, magazines would still
be delivered, and so on. Again: The arrival doesn’t depend on cataloguing.)
Instead, the benefit of cataloguing is for events that happen after the book’s
arrival — cataloguing makes it possible (and maybe makes it easy) to find the
book later on.
The same is true for the vast library that is your memory (cf. Miller
& Springer, 1973). The task of learning is not merely a matter of placing
information into long-term storage. Learning also needs to establish some
appropriate indexing; it must pave a path to the newly acquired information,
so that this information can be retrieved at some future point. Thus, one of
The Role of Meaning and Memory Connections
•
221
WHY DO MEMORY CONNECTIONS HELP?
When books arrive in a library, the librarians must catalogue them. This doesn’t facilitate the “entry” of books into the
library, because the books are in the building whether they’re catalogued or not. But cataloguing makes the books much
easier to find later on. Memory connections may serve the same function: The connections don’t “bring” material into
memory, but they do make the material “findable” in long-term storage later.
the main chores of memory acquisition is to lay the groundwork for memory
retrieval.
But what is it that facilitates memory retrieval? There are, in fact, several
ways to search through memory, but a great deal depends on memory connections. Connections allow one memory to trigger another, and then that
memory to trigger another, so that you’re “led,” connection by connection,
to the sought-after information. In some cases, the connections link one of
the items you’re trying to remember to some of the other items; if so, finding
the first will lead you to the others. In other settings, the connections might
link some aspect of the context-of-learning to the target information, so
that when you think again about the context (“I recognize this room — this
is where I was last week”), you’ll be led to other ideas (“Oh, yeah, I read
the funny story in this room”). In all cases, though, this triggering will happen only if the relevant connections are in place — and establishing those
connections is a large part of what happens during learning.
This line of reasoning has many implications, and we can use those implications as a basis for testing whether this proposal is correct. But right at the
start, it should be clear why, according to this account, deep processing (i.e.,
attention to meaning) promotes memory. The key is that attention to meaning
involves thinking about relationships: “What words are related in meaning to
the word I’m now considering? What words have contrasting meaning? What
is the relationship between the start of this story and the way the story turned
out?” Points like these are likely to be prominent when you’re thinking about
what some word (or sentence or event) means, and these points will help you
to find (or, perhaps, to create) connections among your various ideas. It’s these
connections, we’re proposing, that really matter for memory.
222 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
Elaborate Encoding Promotes Retrieval
Notice, though, that on this account, attention to meaning is not the only
way to improve memory. Other strategies should also be helpful, provided
that they help you to establish memory connections. As an example, consider
another classic study by Craik and Tulving (1975). Participants were shown
a word and then shown a sentence with one word left out. Their task was to
decide whether the word fit into the sentence. For example, they might see
the word “chicken” and then the sentence “She cooked the _________.” The
appropriate response would be yes, because the word does fit in this sentence.
After a series of these trials, there was a surprise memory test, with participants asked to remember all the words they had seen.
But there was an additional element in this experiment. Some of the sentences shown to participants were simple, while others were more elaborate.
For example, a more complex sentence might be: “The great bird swooped
down and carried off the struggling _________.” Sentences like this one produced a large memory benefit — words were much more likely to be remembered if they appeared with these rich, elaborate sentences than if they had
appeared in the simpler sentences (see Figure 6.13).
Apparently, then, deep and elaborate processing leads to better recall than
deep processing on its own. Why? The answer hinges on memory connections. Maybe the “great bird swooped” sentence calls to mind a barnyard
scene with the hawk carrying away a chicken. Or maybe it calls to mind
FIGURE 6.13
DEEP AND ELABORATE ENCODING
Percent of words recalled
60
50
40
30
20
10
0
Simple
Complex
Sentence type
Deep processing (paying attention to meaning) promotes memory, but it isn’t
the only factor that has this benefit. More elaborate processing (e.g., by thinking about the word in the context of a complex sentence, rich with relationships) also has a powerful effect on memory.
(after craik & tulving, 1975)
The Role of Meaning and Memory Connections
•
223
TEST YOURSELF
10.What does it mean
to say, “The creation
of memory connections often occurs at
the time of learning,
but the main benefit
of those connections
comes later, at the
time of memory
retrieval”?
11.In what ways is deep
and elaborate processing superior to
deep processing on
its own?
thoughts about predator-prey relationships. One way or another, the richness of the sentence offers the potential for many connections as it calls other
thoughts to mind, each of which can be connected to the target sentence.
These connections, in turn, provide potential retrieval paths — paths that can,
in effect, guide your thoughts toward the content to be remembered. All of
this seems less likely for the simpler sentences, which will evoke fewer connections and so establish a narrower set of retrieval paths. Consequently,
words associated with these sentences are less likely to be recalled later on.
Organizing and Memorizing
Sometimes, we’ve said, memory connections link the to-be-remembered
material to other information already in memory. In other cases, the connections link one aspect of the to-be-remembered material to another aspect of
the same material. Such a connection ensures that if any part of the material
is recalled, then all will be recalled.
In all settings, though, the connections are important, and that leads us to
ask how people go about discovering (or creating) these connections. More
than 70 years ago, a psychologist named George Katona argued that the key
lies in organization (Katona, 1940). Katona’s argument was that the processes of organization and memorization are inseparable: You memorize well
when you discover the order within the material. Conversely, if you find (or
impose) an organization on the material, you will easily remember it. These
suggestions are fully compatible with the conception we’re developing here,
since what organization provides is memory connections.
Mnemonics
MNEMOSYNE
Strategies that are used to
improve memory are known
as mnemonic strategies, or
mnemonics. The term derives
from the name of the goddess of memory in Greek
mythology—Mnemosyne.
224 •
For thousands of years, people have longed for “better” memories and,
guided by this desire, people in the ancient world devised various techniques
to improve memory — techniques known as mnemonic strategies. In fact,
many of the mnemonics still in use date back to ancient Greece. (It’s therefore appropriate that these techniques are named in honor of Mnemosyne,
the goddess of memory in Greek mythology.)
How do mnemonics work? In general, these strategies provide some way
of organizing the to-be-remembered material. For example, one broad class of
mnemonic, often used for memorizing sequences of words, links the first letters
of the words into some meaningful structure. Thus, children rely on ROY G.
BIV to memorize the sequence of colors in the rainbow (red, orange, yellow . . .),
and they learn the lines in music’s treble clef via “Every Good Boy Deserves
Fudge” or “. . . Does Fine” (the lines indicate the musical notes E, G, B, D, and F).
Biology students use a sentence like “King Philip Crossed the Ocean to Find
Gold and Silver” (or: “. . . to Find Good Spaghetti”) to memorize the sequence of
taxonomic categories: kingdom, phylum, class, order, family, genus, and species.
Other mnemonics involve the use of mental imagery, relying on “mental pictures” to link the to-be-remembered items to one another. (We’ll have
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
much more to say about “mental pictures” in Chapter 11.) For example,
imagine a student trying to memorize a list of word pairs. For the pair eagletrain, the student might imagine the eagle winging back to its nest with a
locomotive in its beak. Classic research evidence indicates that images like
this can be enormously helpful. It’s important, though, that the images show
the objects in some sort of relationship or interaction — again highlighting
the role of organization. It doesn’t help just to form a picture of an eagle
and a train sitting side-by-side (Wollen, Weber, & Lowry, 1972; for another
example of a mnemonic, see Figure 6.14).
A different type of mnemonic provides an external “skeleton” for the tobe-remembered materials, and mental imagery can be useful here, too. Imagine that you want to remember a list of largely unrelated items — perhaps
FIGURE 6.14
MNEMONIC STRATEGIES
Yukon
British
Columbia
Alberta
Nunavit
New
fou
nd
lan
da
Manitoba
Quebec
Onatrio
nd
La
br
a
r
do
Sas
katc
hew
an
Northwest
Territories
Prince
Edward
Island
Nova
Scotia
New Brunswick
With a bit of creativity, you can make up mnemonics for memorizing all sorts of things. For example, can you
name all ten of the Canadian provinces? Perhaps there is a great mnemonic available, but in the meantime,
this will do. It’s a complicated mnemonic but unified by the theme of the early-morning meal: “Breakfast
Cooks Always Sell More Omelets. Quiche Never Bought; Never Sold. Perhaps Eggs In New Forms?” (You’re
on your own for remembering the three northern territories.)
Organizing and Memorizing
•
225
the entries on your shopping list, or a list of questions you want to ask your
adviser. For this purpose, you might rely on one of the so-called peg-word
systems. These systems begin with a well-organized structure, such as this one:
One is a bun.
Two is a shoe.
Three is a tree.
Four is a door.
Five is a hive.
Six are sticks.
Seven is heaven.
Eight is a gate.
Nine is a line.
Ten is a hen.
This rhyme provides ten “peg words” (“bun,” “shoe,” etc.), and in memorizing something you can “hang” the materials to be remembered on these
“pegs.” Let’s imagine that you want to remember the list of topics you need
to discuss with your adviser. If you want to discuss your unhappiness with
chemistry class, you might form an association between chemistry and the
first peg, “bun.” You might picture a hamburger bun floating in an Erlenmeyer
flask. If you also want to discuss your plans for after graduation, you might
form an association between some aspect of those plans and the next peg,
“shoe.” (You could think about how you plan to pay your way after college
by selling shoes.) Then, when meeting with your adviser, all you have to do is
think through that silly rhyme again. When you think of “one is a bun,” it’s
highly likely that the image of the flask (and therefore of chemistry lab) will
come to mind. With “two is a shoe,” you’ll be reminded of your job plans.
And so on.
Hundreds of variations on these techniques — the first-letter mnemonics,
visualization strategies, peg-word systems — are available. Some variations are
taught in self-help books (you’ve probably seen the ads — ”How to Improve Your
Memory!”); some are taught as part of corporate management training. But all
the variations use the same basic scheme. To remember a list with no apparent
organization, you impose an organization on it by using a tightly organized skele­
ton or scaffold. And, crucially, these systems all work. They help you remember
individual items, and they also help you remember those items in a specific
sequence. Figure 6.15 shows some of the data from one early study; many other
studies confirm this pattern (e.g., Bower, 1970, 1972; Bower & Reitman, 1972;
Christen & Bjork, 1976; Higbee, 1977; Roediger, 1980; Ross & Lawrence, 1968;
Yates, 1966). All of this strengthens our central claim: Mnemonics work because
they impose an organization on the materials you’re trying to memorize. And,
consistently and powerfully, organizing improves recall.
Given the power of mnemonics, students are well advised to use these
strategies in their studies. In fact, for many topics there are online databases
226 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
FIGURE 6.15
THE POWER OF MNEMONICS
Number of items recalled in proper
sequence, after 24-hour delay
5
4
3
2
1
te
im rac
ag tiv
er e
y
In
Pe
g
sy -wo
st rd
em
re Ve
he rb
ar al
sa
l
Is
o
im lat
ag ed
es
0
Type of Learning
Mnemonics can be enormously effective. In this study, students who had
relied on peg words or interactive imagery vastly outperformed students
who’d used other memorizing strategies.
(after roediger, 1980)
containing thousands of useful mnemonics — helping medical students to
memorize symptom lists, chemistry students to memorize the periodic table,
neuroscientists to remember the brain’s anatomy, and more.
Bear in mind, though, that there’s a downside to the use of mnemonics
in educational settings. When using a mnemonic, you typically focus on just
one aspect of the material you’re trying to memorize — for example, just
the first letter of the word to be remembered — and so you may cut short
your effort toward understanding this material, and likewise your effort
toward finding multiple connections between the material and other things
you know.
To put this point differently, mnemonic use involves a trade-off. If you
focus on just one or two memory connections, you’ll spend little time thinking about other possible connections, including those that might help you
understand the material. This trade-off will be fine if you don’t care very much
about the meaning of the material. (Do you care why, in taxonomy, “order”
is a subset of “class,” rather than the other way around?) But the trade-off
is troubling if you’re trying to memorize material that is meaningful. In this
case, you’d be better served by a memory strategy that seeks out multiple connections between the material you’re trying to learn and things you already
Organizing and Memorizing
•
227
know. This effort toward multiple links will help you in two ways. First, it
will foster your understanding of the material to be remembered, and so will
lead to better, richer, deeper learning. Second, it will help you retrieve this
information later. We’ve already suggested that memory connections serve as
retrieval paths, and the more paths there are, the easier it will be to find the
target material later.
For these reasons, mnemonic use may not be the best approach in many
situations. Still, the fact remains that mnemonics are immensely useful in
some settings (What were those rainbow colors?), and this confirms our
initial point: Organization promotes memory.
Understanding and Memorizing
So far, we’ve said a lot about how people memorize simple stimulus
materials — lists of randomly selected words, or colors that have to be learned
in the right sequence. In our day-to-day lives, however, we typically want to
remember more meaningful, more complicated, material. We want to remember the episodes we experience, the details of rich scenes we’ve observed,
or the many-step arguments we’ve read in a book. Do the same memory
principles apply to these cases?
The answer is clearly yes (although we’ll have more to say about this
issue in Chapter 8). In other words, your memory for events, or pictures,
or complex bodies of knowledge is enormously dependent on your being
able to organize the material to be remembered. With these more complicated materials, though, we’ve suggested that your best bet for organization
isn’t some arbitrary skeleton like those used in mnemonics. Instead, the best
organization of these complex materials is generally dependent on understanding. That is, you remember best what you understand best.
There are many ways to show that this is true. For example, we can give
people a sentence or paragraph to read and test their comprehension by asking questions about the material. Sometime later, we can test their memory.
The results are clear: The better the participants’ understanding of a sentence
or a paragraph, if questioned immediately after viewing the material, the
greater the likelihood that they will remember the material after a delay (for
classic data on this topic, see Bransford, 1979).
Likewise, consider the material you’re learning right now in the courses
you’re taking. Will you remember this material 5 years from now, or 10, or 20?
The answer depends on how well you understand the material, and one
measure of understanding is the grade you earn in a course. With full and rich
understanding, you’re likely to earn an A; with poor understanding, your grade
is likely to be lower. This leads to a prediction: If understanding is (as we’ve
proposed) important for memory, then the higher someone’s grade in a course,
the more likely that person is to remember the course contents, even years later.
This is exactly what the data show, with A students remembering the material quite well, and C students remembering much less (Conway, Cohen, &
Stanhope, 1992).
228 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
The relationship between understanding and memory can also be demonstrated in another way: by manipulating whether people understand
the material or not. For example, in an early experiment by Bransford and
Johnson (1972, p. 722), participants read this passage:
The procedure is actually quite simple. First you arrange items into
different groups. Of course one pile may be sufficient depending on
how much there is to do. If you have to go somewhere else due to lack
of facilities that is the next step; otherwise you are pretty well set. It is
important not to overdo things. That is, it is better to do too few things
at once than too many. In the short run, this may not seem important
but complications can easily arise. A mistake can be expensive as well.
At first, the whole procedure will seem complicated. Soon, however, it
will become just another facet of life. It is difficult to foresee any end to
the necessity for this task in the immediate future, but then, one never
can tell. After the procedure is completed one arranges the materials
into different groups again. Then they can be put into their appropriate
places. Eventually they will be used once more and the whole cycle will
then have to be repeated. However, that is part of life.
You’re probably puzzled by the passage, and so are most research participants. The story is easy to understand, though, if we give it a title: “Doing
the Laundry.” In the experiment, some participants were given the title before
reading the passage; others were not. Participants in the first group easily
understood the passage and were able to remember it after a delay. Participants in the second group, reading the same words, weren’t confronting a
meaningful passage and did poorly on the memory test. (For related data, see
Bransford & Franks, 1971; Sulin & Dooling, 1974. For another example, see
Figure 6.16.)
FIGURE 6.16
MEMORY FOR DIGITS
149162536496481
Examine this series of digits for a moment, and then turn away from the
page and try to recall all 15 in their proper sequence. The chances are good
that you will fail in this task—perhaps remembering the first few and the last
few digits, but not the entire list. Things will go differently, though, if you
discover the pattern within the list. Now, you’ll easily be able to remember
the full sequence. What is the pattern? Try thinking of the series this way: 1, 4,
9, 16, 25, 36. . . . Here, as always, organizing and understanding aid memory.
Organizing and Memorizing
•
229
FIGURE 6.17
OMPREHENSION ALSO AIDS MEMORY
C
FOR PICTURES
People who perceive this picture as a pattern of meaningless blotches are
unlikely to remember the picture. People who perceive the “hidden” form do
remember the picture.
(after wiseman & neisser, 1974)
TEST YOURSELF
12.Why do mnemonics
help memory? What
are the limitations
involved in mnemonic
use?
13.What’s the evidence
that there’s a clear
linkage between how
well you understand
material when you
first meet it, and
how fully you’ll recall
that information
later on?
230 •
Similar effects can be documented with nonverbal materials. Consider
the picture shown in Figure 6.17. At first it looks like a bunch of meaningless blotches; with some study, though, you may discover a familiar object.
Wiseman and Neisser (1974) tested people’s memory for this picture. Consistent with what we’ve seen so far, their memory was good if they understood
the picture — and bad otherwise. (Also see Bower, Karlin, & Dueck, 1975;
Mandler & Ritchey, 1977; Rubin & Kontis, 1983.)
The Study of Memory Acquisition
This chapter has largely been about memory acquisition. How do we
acquire new memories? How is new information, new knowledge, established in long-term memory? In more pragmatic terms, what is the best,
most effective way to learn? We now have answers to these questions, but
our discussion has indicated that we need to place these questions into a
broader context — with attention on the substantial contribution from the
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
memorizer, and also a consideration of the interconnections among acquisition, retrieval, and storage.
The Contribution of the Memorizer
Over and over, we’ve seen that memory depends on connections among ideas,
connections fostered by the steps you take in your effort toward organizing
and understanding the materials you encounter. Hand in hand with this, it
appears that memories are not established by sheer contact with the items
you’re hoping to remember. If you’re merely exposed to the items without
giving them any thought, then subsequent recall of those items will be poor.
These points draw attention to the huge role played by the memorizer. If,
for example, we wish to predict whether this or that event will be recalled,
it isn’t enough to know that someone was exposed to the event. Instead, we
need to ask what the person was doing during the event. Did she only do
maintenance rehearsal, or did she engage the material in some other way? If
the latter, how did she think about the material? Did she pay attention to the
appearance of the words or to their meaning? If she thought about meaning,
was she able to understand the material? These considerations are crucial for
predicting the success of memory.
The contribution of the memorizer is also evident in another way. We’ve
argued that learning depends on making connections, but connections to
what? If you want to connect the to-be-remembered material to other knowledge, to other memories, then you need to have that other knowledge — you
need to have other (potentially relevant) memories that you can “hook” the
new material on to.
This point helps us understand why sports fans have an easy time learning new facts about sports, and why car mechanics can easily learn new
facts about cars, and why memory experts easily memorize new information
about memory. In each situation, the person enters the learning situation with
a considerable advantage — a rich framework that the new materials can be
woven into. But, conversely, if someone enters a learning situation with little
relevant background, then there’s no framework, nothing to connect to, and
learning will be more difficult. Plainly, then, if we want to predict someone’s success in memorizing, we need to consider what other knowledge the
individual brings into the situation.
The Links among Acquisition, Retrieval, and Storage
These points lead us to another important theme. The emphasis in this chapter has been on memory acquisition, but we’ve now seen that claims about
acquisition cannot be separated from claims about storage and retrieval.
For example, why is memory acquisition improved by organization? We’ve
suggested that organization provides retrieval paths, making the memories
“findable” later on, and this is a claim about retrieval. Therefore, our claims
about acquisition are intertwined with claims about retrieval.
TEST YOURSELF
14.Explain why memorizing involves a contribution from the memorizer, both in terms of
what the memorizer
does while memorizing, and also in terms
of what the memorizer knows prior to the
memorizing.
The Study of Memory Acquisition
•
231
Likewise, we just noted that your ability to learn new material depends,
in part, on your having a framework of prior knowledge to which the new
materials can be tied. In this way, claims about memory acquisition need to
be coordinated with claims about the nature of what is already in storage.
These interactions among acquisition, knowledge, and retrieval are crucial
for our theorizing. But the interactions also have important implications for
learning, for forgetting, and for memory accuracy. The next two chapters
explore some of those implications.
COGNITIVE PSYCHOLOGY AND EDUCATION
how should i study?
Throughout your life, you encounter information that you hope to remember
later — whether you’re a student taking courses or an employee in training
for a new job. In these and many other settings, what helpful lessons can you
draw from memory research?
For a start, bear in mind that the intention to memorize, on its own, has
no effect. Therefore, you don’t need any special “memorizing steps.” Instead,
you should focus on making sure you understand the material, because if you
do, you’re likely to remember it.
As a specific strategy, it’s useful to spend a moment after a class, or after
you’ve done a reading assignment, to quiz yourself about what you’ve just
learned. You might ask questions like these: “What are the new ideas here?”;
“Do these new ideas fit with other things I know?”; “Do I know what evidence or arguments support the claims here?” Answering questions like these
will help you find meaningful connections within the material you’re learning, and between this material and other information already in your memory. In the same spirit, it’s often useful to rephrase material you encounter,
putting it into your own words. Doing this will force you to think about what
the words mean — again, a good thing for memory.
Surveys suggest, however, that most students rely on study strategies that
are much more passive than this — in fact, far too passive. Most students try
to learn materials by simply rereading the textbook or reading over their
notes several times. The problem with these strategies should be obvious:
As the chapter explains, memories are produced by active engagement with
materials, not by passive exposure.
As a related point, it’s often useful to study with a friend — so that he
or she can explain topics to you, and you can do the same in return. This
step has several advantages. In explaining things, you’re forced into a more
active role. Working with a friend is also likely to enhance your understanding, because each of you can help the other to understand bits you’re having
trouble with. You’ll also benefit from hearing your friend’s perspective on the
materials. This additional perspective offers the possibility of creating new
connections among ideas, making the information easier to recall later on.
232 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
Memory will also be best if you spread your studying out across multiple occasions — using spaced learning (e.g., spreading out your learning
across several days) rather than massed learning (essentially, “cramming” all
at once). It also helps to vary your focus while studying — working on your
history assignment for a while, then shifting to math, then over to the novel
your English professor assigned, and then back to history. There are several
reasons for this, including the fact that spaced learning and a changing focus
will make it likely that you’ll bring a somewhat different perspective to the
material each time you turn to it. This new perspective will let you see connections you didn’t see before; and — again — these new connections provide
retrieval paths that can promote recall.
Spaced learning also has another advantage. With this form of learning,
some time will pass between the episodes of learning. (Imagine, for example,
that you study your sociology text for a while on Tuesday night and then
return to it on Thursday, so that two days go by between these study sessions.)
This situation allows some amount of forgetting to take place, and that’s
actually helpful because now each episode of learning will have to take a bit
more effort, a bit more thought. This stands in contrast to massed learning,
in which your second and third passes through the material may only be
separated by a few minutes. In this setting, the second and third passes may
feel easy enough so that you zoom through them, with little engagement in
the material.
Note an ironic point here: Spaced learning may be more difficult (because
of the forgetting in between sessions), but this difficulty leads to better learning overall. Researchers refer to this as “desirable difficulty” — difficulty that
may feel obnoxious when you’re slogging through the material you hope
to learn but that is nonetheless beneficial, because it leaves you with more
complete, more long-lasting memory.
What about mnemonic strategies, such as a peg-word system? These are
enormously helpful — but often at a cost. When you’re first learning something new, focusing on a mnemonic can divert your time and attention away
from efforts at understanding the material, and so you’ll end up understanding the material less well. You’ll also be left with only the one or two retrieval
paths that the mnemonic provides, not the multiple paths created by comprehension. In some circumstances these drawbacks aren’t serious — and so,
for example, mnemonics are often useful for memorizing dates, place names,
or particular bits of terminology. But for richer, more meaningful material,
mnemonics may hurt you more than they help.
Mnemonics can be more helpful, though, after you’ve understood the new
material. Imagine that you’ve thoughtfully constructed a many-step argument or a complex derivation of a mathematical formula. Now, imagine that
you hope to re-create the argument or the derivation later on — perhaps for
an oral presentation or on an exam. In this situation, you’ve already achieved
a level of mastery, and you don’t want to lose what you’ve gained. Here, a
mnemonic (like the peg-word system) might be quite helpful, allowing you to
remember the full argument or derivation in its proper sequence.
MEANINGFUL
CONNECTIONS
What sort of connections will
help you to remember? The
answer is that almost any
connection can be helpful.
Here’s a silly—but useful—
example. Students learning
about the nervous system
have to learn that efferent
fibers carry information away
from the brain and central
nervous system, while afferent fibers carry information
inward. How to keep these
terms straight? It may be
helpful to notice that efferent
fibers carry information exiting the nervous system, while
afferent fibers provide information arriving in the nervous
system. And, as a bonus, the
same connections will help
you remember that you can
have an effect on the world
(an influence outward, from
you), but that the world can
also affect you (an influence
coming inward, toward you).
Cognitive Psychology and Education
•
233
Finally, let’s emphasize that there’s more to say about these issues.
Our discussion here (like Chapter 6 itself) focuses on the “input” side of
memory — getting information into storage, so that it’s available for use later
on. There are also steps you can take that will help you to locate information in the vast warehouse of your memory, and still other steps that you can
take to avoid forgetting materials you’ve already learned. Discussion of those
steps, however, depends on materials we’ll cover in Chapters 7 and 8.
For more on this topic . . .
Brown, P. C., Roediger, H. L., & McDaniel, M. A. (2014). Make it stick: The science
of successful learning. New York, NY: Belknap Press.
McCabe, J. A., Redick, T. S., & Engle, R. W. (2016). Brain-training pessimism, but
applied memory optimism. Psychological Science in the Public Interest, 17,
187–191.
Putnam, A. L., Sungkhasettee, V. W., & Roediger, H. L. (2016). Optimizing learning in college: Tips from cognitive psychology. Perspective on Psychological
Science, 11(5), 652–660.
234 •
C H A P T E R S I X The Acquisition of Memories and the Working-Memory System
chapter review
SUMMARY
• It is convenient to think of memorizing as having separate stages. First, one acquires new information (acquisition). Next, the information remains in
storage until it is needed. Finally, the information is
retrieved. However, this separation among the stages
may be misleading. For example, in order to memorize new information, you form connections bet­
ween this information and things you already know.
In this way, the acquisition stage is intertwined with
the retrieval of information already in storage.
• Information that is currently being considered is
held in working memory; information that isn’t currently active but is nonetheless in storage is in longterm memory. The distinction between these two
forms of memory has traditionally been described
in terms of the modal model and has been examined
in many studies of the serial-position curve. The primacy portion of this curve reflects items that have
had extra opportunity to reach long-term memory;
the recency portion of this curve reflects the accurate retrieval of items currently in working memory.
• Psychologists’ conception of working memory has
evolved in important ways in the last few decades.
Crucially, psychologists no longer think of working
memory as a “storage container” or even as a “place.”
Instead, working memory is a status—and so we say
items are “in working memory” when they’re being
actively thought about. This activity is governed by
working memory’s central executive. For mere storage, the executive often relies on low-level assistants,
including the articulatory rehearsal loop and the
visuospatial buffer, which work as mental scratch pads.
The activity inherent in this overall system is reflected
in the flexible way material can be chunked in working memory. The activity is also reflected in current
measures of working memory, via operation span.
• Maintenance rehearsal serves to keep information
in working memory and requires little effort, but it
has little impact on subsequent recall. To maximize
your chances of recall, elaborative rehearsal is needed,
in which you seek connections within the material to
be remembered or connections between the material
to be remembered and things you already know.
• In many cases, elaborative processing takes the
form of attention to meaning. This attention to
meaning is called “deep processing,” in contrast to
attention to sounds or visual form, which is considered “shallow processing.” Many studies have
shown that deep processing leads to good memory
performance later on, even if the deep processing
occurred with no intention of memorizing the target
material. In fact, the intention to learn has no direct
effect on performance; what matters instead is how
someone engages or thinks about the material to be
remembered.
• Deep processing has beneficial effects by creating effective retrieval paths that can be used later
on. Retrieval paths depend on connections linking
one memory to another; each connection provides
a path potentially leading to a target memory. Mnemonic strategies rely on the same mechanism and
focus on the creation of specific memory connections, often tying the to-be-remembered material to
a frame (e.g., a strongly structured poem).
• Perhaps the best way to form memory connections
is to understand the material to be remembered. In
understanding, you form many connections within
the material to be remembered, as well as between
this material and other knowledge. With all these retrieval paths, it becomes easy to locate this material
in memory. Consistent with these suggestions, studies have shown a close correspondence between the
ability to understand some material and the ability
to recall that material later on. This pattern has been
demonstrated with stories, visual patterns, number
series, and many other stimuli.
235
KEY TERMS
acquisition (p. 197)
storage (p. 197)
retrieval (p. 197)
modal model (p. 198)
sensory memory (p. 198)
short-term memory (p. 198)
working memory (p. 199)
long-term memory (LTM) (p. 199)
free recall (p. 200)
primacy effect (p. 200)
recency effect (p. 200)
serial position (p. 201)
memory rehearsal (p. 202)
digit-span task (p. 206)
“7 plus-or-minus 2” (p. 206)
chunks (p. 206)
operation span (p. 208)
working-memory capacity (WMC) (p. 209)
working-memory system (p. 210)
articulatory rehearsal loop (p. 210)
subvocalization (p. 211)
phonological buffer (p. 211)
concurrent articulation (p. 212)
maintenance rehearsal (p. 215)
relational (or elaborative) rehearsal (p. 215)
intentional learning (p. 218)
incidental learning (p. 218)
shallow processing (p. 219)
deep processing (p. 219)
level of processing (p. 219)
retrieval paths (p. 224)
mnemonic strategies (p. 224)
peg-word systems (p. 226)
TEST YOURSELF AGAIN
1.Define the terms “acquisition,” “storage” and
“retrieval.”
2.List the four ways in which (either in the
modal model or in more recent views) working
memory is different from long-term storage.
3.How is the primacy effect usually explained?
How is the recency effect usually explained?
4.What does it mean to say that working memory
holds seven (plus-or-minus two) “chunks”?
What is a chunk?
5.What evidence suggests that operation span is
a better measure of working memory than the
more standard digit-span measure?
6.How does the rehearsal loop manage to hold
on to information with only occasional involvement by the central executive?
7.What is the difference between maintenance
rehearsal and relational (or elaborative)
rehearsal?
236
8.What does it mean to say, “It doesn’t matter
if you intend to memorize; all that matters for
memory is how exactly you engage the material you encounter”?
9.What is deep processing, and what impact does
it have on memory?
10.What does it mean to say, “The creation of
memory connections often occurs at the time
of learning, but the main benefit of those connections comes later, at the time of memory
retrieval”?
11.In what ways is deep and elaborate processing
superior to deep processing on its own?
12.Why do mnemonics help memory? What are
the limitations of mnemonic use?
13.What’s the evidence that there’s a clear linkage
between how well you understand material
when you first meet it, and how fully you’ll
recall that information later on?
14.Explain why memorizing involves a contribution from the memorizer, both in terms of what
the memorizer does while memorizing, and
also in terms of what the memorizer knows
prior to the memorizing.
THINK ABOUT IT
1.Imagine that, based on what you’ve read in this
chapter, you were asked to write a “training
pamphlet” advising students how to study more
effectively, so that they would remember what
they studied more fully and more accurately.
What would you write in the pamphlet?
E eBOOK DEMONSTRATIONS & ESSAYS
Go to http://digital.wwnorton.com/cognition7 for the online demonstrations and essays
relevant to this chapter. These can be found in the ebook.
Online Demonstrations
• D
emonstration 6.1: Primacy and Recency Effects
• D
emonstration 6.2: Chunking
• D
emonstration 6.3: The Effects of Unattended
Online Applying Cognitive Psychology and the
Law Essays
• Cognitive Psychology and the Law: The VideoRecorder View
Exposure
• D
emonstration 6.4: Depth of Processing
• Demonstration 6.5: The Articulatory Rehearsal
Loop
• Demonstration 6.6: Sound-Based Coding
COGNITION LABS
Go to http://digital.wwnorton.com/cognition7 to experience interactive online psychology
labs relevant to this chapter.
Answer: Actually, none of the images shown in Figure 6.10 depict the Apple logo. The bottommiddle image has the bite and the dimple in the right positions, but it shows the stem pointing
the wrong way. The bottom-left image shows the stem and bite correctly, but it’s missing the
dimple!
237
7
chapter
Interconnections
between Acquisition
and Retrieval
what if…
The man known as H.M. was in his mid-20s when he
had brain surgery intended to control his epilepsy.
(We first met H.M. in Chapter 1; we also mentioned him briefly in
Chapters 2 and 6.) This surgery did achieve its aim, and H.M.’s seizures
were reduced. But the surgery had an unexpected and horrible consequence: H.M. lost the ability to form new memories. If asked what
he did last week, or yesterday, or even an hour ago, H.M. had no idea. He
couldn’t recognize the faces of medical staff he’d seen day after day.
He could read and reread a book yet never realize he’d read the same
book many times before.
A related pattern of memory loss occurs among patients who suffer from Korsakoff’s syndrome. We’ll say more about this syndrome
later in the chapter, but for now let’s highlight a paradox. These
patients, like H.M., are profoundly amnesic; they’re completely unable
to recall the events of their own lives. But these patients (again, like
H.M.) have “unconscious” memories — memories that they don’t know
they have.
We reveal these unconscious memories if we test Korsakoff’s patients
indirectly. For example, if we ask them, “Which of these melodies did
you hear an hour ago?” they’ll answer randomly — confirming their
amnesia. But if we ask them, “Which of these melodies do you prefer?”
they’re likely to choose the ones that, in fact, they heard an hour ago —
indicating that they do somehow remember (and are influenced by) the
earlier experience. If we ask them, “Have you ever seen a puzzle like this
one before?” they’ll say no. But if we ask them to solve the puzzle, their
speed will be much faster the second time — even though they insist it’s
the first time they’ve seen the puzzle. Their speed will be even faster
the third time they solve the puzzle and the fourth, although again
and again they’ll claim they’re seeing the puzzle for the very first time.
Likewise, they’ll fail if we ask them, “I showed you some words a few
minutes ago; can you tell me which of those words began ‘CHE . . .’?”
But, alternatively, we can ask them, “What’s the first word that comes to
mind that starts ‘CHE . . .’?” With this question, they’re likely to respond
with the word they’d seen earlier — a word that they ostensibly could
not remember.
239
preview of chapter themes
•
earning does not simply place information in memory;
L
instead, learning prepares you to retrieve the information
in a particular way. As a result, learning that is good preparation for one sort of retrieval may be inadequate for other
sorts of retrieval.
•
ome experiences seem to produce unconscious memoS
ries. Consideration of these “implicit memory” effects will
help us understand the various ways in which memory
influences you and will also help us see where the feeling
of familiarity comes from.
•
In general, retrieval is more likely to succeed if your perspective is the same during learning and during retrieval,
just as we would expect if learning establishes retrieval
paths that help you later when you “travel” the same path
in your effort toward locating the target material.
•
inally, an examination of amnesia confirms a central
F
theme of the chapter — namely, that we cannot speak
of “good” or “bad” memory in general. Instead, we need
to evaluate memory by considering how, and for what
purposes, the memory will be used.
These observations strongly suggest that there must be different types
of memory — including a type that’s massively disrupted in these amnesic
patients, yet one that is apparently intact (also see Figure 7.1). But how
many types of memory are there? How does each one function? Is it possible
that processes or strategies that create one type of memory might be less
useful for some other type? These questions will be central in this chapter.
Number of errors in
each attempt
40
Day 1
Day 2
Day 3
30
20
10
0
1
B
FIGURE 7.1
A
240 •
10 1
10 1
10
Attempts each day
MIRROR DRAWING
(Panel A) In a mirror-drawing task, participants must draw a
precisely defined shape — they might be asked, for example,
to trace a line between the inner and outer star. The trick,
though, is that the participants can see the figure (and their
own hand) only in the mirror. (Panel B) Performance is
usually poor at first but gradually gets better. Remarkably,
the same pattern of improvement is observed with amnesic
patients, even though on each attempt they insist that
they’re performing this task for the very first time.
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
Learning as Preparation for Retrieval
Putting information into long-term memory helps you only if you can
retrieve that information later on. Otherwise, it would be like putting
money into a savings account without the option of ever making withdrawals, or writing books that could never be read. But let’s emphasize
that there are different ways to retrieve information from memory. You
can try to recall the information (“What was the name of your tenth-grade
homeroom teacher?”) or to recognize it (“Was the name perhaps Miller?”).
If you try to recall the information, a variety of cues may or may not be
available (you might be told, as a hint, that the name began with an M or
rhymes with “tiller”).
In Chapter 6, we largely ignored these variations in retrieval. We talked as
if material was well established in memory or was not, with little regard for
how the material would be retrieved from memory. There’s reason to believe,
however, that we can’t ignore these variations in retrieval, and in this chapter
we’ll examine the interaction between how a bit of information was learned
and how it is retrieved later.
Crucial Role of Retrieval Paths
In Chapter 6, we argued that when you’re learning, you’re making connections between the newly acquired material and other information already in
your memory. These connections make the new knowledge “findable” later
on. Specifically, the connections serve as retrieval paths: When you want to
locate information in memory, you travel on those paths, moving from one
memory to the next until you reach the target material.
These claims have an important implication. To see this, bear in mind
that retrieval paths — like any paths — have a starting point and an ending
point: The path leads you from Point A to Point B. That’s useful if you want
to move from A to B, but what if you’re trying to reach B from somewhere
else? What if you’re trying to reach Point B, but at the moment you happen
to be nowhere close to Point A? In that case, the path linking A and B may
not help you.
As an analogy, imagine that you’re trying to reach Chicago from somewhere to the west. For this purpose, what you need is some highway coming
in from the west. It won’t help that you’ve constructed a wonderful road
coming into Chicago from the east. That road might be valuable in other
circumstances, but it’s not the path you need to get from where you are right
now to where you’re heading.
Do retrieval paths in memory work the same way? If so, we might find
cases in which your learning is excellent preparation for one sort of retrieval
but useless for other types of retrieval — as if you’ve built a road coming
in from one direction but now need a road from another direction. Do the
research data show this pattern?
Learning as Preparation for Retrieval
•
241
Context-Dependent Learning
Consider classic studies on context-dependent learning (Eich, 1980; Overton,
1985). In one such study, Godden and Baddeley (1975) asked scuba divers to
learn various materials. Some of the divers learned the material while sitting
on dry land; others learned it while underwater, hearing the material via a
special communication set. Within each group, half of the divers were then
tested while above water, and half were tested below (see Figure 7.2).
Underwater, the world has a different look, feel, and sound, and this
context could easily influence what thoughts come to mind for the divers
in the study. Imagine, for example, that a diver is feeling cold while underwater. This context will probably lead him to think “cold-related” thoughts,
so those thoughts will be in his mind during the learning episode. In this
situation, the diver is likely to form memory connections between these
thoughts and the materials he’s trying to learn.
Let’s now imagine that this diver is back underwater at the time of the
memory test. Most likely he’ll again feel cold, which may once more lead
him to “cold-related” thoughts. These thoughts, in turn, are now connected
(we’ve proposed) to the target materials, and that gives us what we want: The
cold triggers certain thoughts, and because of the connections formed during
learning, those thoughts can trigger the target memories.
Of course, if the diver is tested for the same memory materials on land,
he might have other links, other memory connections, that will lead to the
target memories. Even so, on land the diver will be at a disadvantage because
the “cold-related” thoughts aren’t triggered — so there will be no benefit from
the memory connections that are now in place, linking those thoughts to the
sought-after memories.
Test while
FIGURE 7.2 THE DESIGN OF A
CONTEXT-DEPENDENT LEARNING
EXPERIMENT
Half of the participants (deep-sea divers) learned
the test material while underwater; half learned
while on land. Then, within each group, half
were tested while underwater; half were tested
on land. We expect a retrieval advantage if the
learning and test circumstances match. Therefore, we expect better performance in the top
left and bottom right cells.
242 •
On land
On land
Underwater
Learning
and test
circumstances
match
CHANGE of
circumstances
between learning
and test
CHANGE of
circumstances
between learning
and test
Learning
and test
circumstances
match
Learn while
Underwater
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
By this logic, we should expect that divers who learn material while
underwater will remember the material best if they’re again underwater at
the time of the test. This setting will enable them to use the connections they
established earlier. In terms of our previous analogy, they’ve built certain
highways, and we’ve put the divers into a situation in which they can use
what they’ve built. And the opposite is true for divers who learned while on
land; they should do best if tested on land. And that is exactly what the data
show (see Figure 7.3).
Similar results have been obtained in other studies, including those
designed to mimic the learning situation of a college student. In one experiment, research participants read a two-page article similar to the sorts
of readings they might encounter in their college courses. Half the participants read the article in a quiet setting; half read it in noisy circumstances.
When later given a short-answer test, those who read the article in quiet
did best if tested in quiet — 67% correct answers, compared to 54% correct if tested in a noisy environment. Those who read the article in a noisy
environment did better if tested in a noisy environment — 62% correct,
compared to 46%. (See Grant et al., 1998; also see Balch, Bowman, &
Mohler, 1992; Cann & Ross, 1989; Schab, 1990; Smith, 1985; Smith &
Vela, 2001.)
In another study, Smith, Glenberg, and Bjork (1978) reported the same
pattern if learning and testing took place in different rooms — with the rooms
varying in appearance, sounds, and scent. In this study, though, there was an
important twist: In one version of the procedure, the participants learned
materials in one room and were tested in a different room. Just before testing,
however, the participants were urged to think about the room in which they
had learned — what it looked like and how it made them feel. When tested,
these participants performed as well as those for whom there was no room
change (Smith, 1979). What matters, therefore, is not the physical context
Test environment
14
Land
Mean words recalled
12
Underwater
10
FIGURE 7.3
LEARNING
8
6
4
2
0
Studied on land
Studied underwater
CONTEXT-DEPENDENT
Scuba divers learned materials either while on land or
while underwater. Then, they were tested while on land or
underwater. Performance was best if the divers’ circumstances at the time of test were matched to those in place
during learning.
( after godden & baddeley , 1975)
Learning as Preparation for Retrieval
•
243
TEST YOURSELF
1. What does contextdependent learning
tell us about the
nature of retrieval
paths?
2. In what ways is a
retrieval path like an
“ordinary” path (e.g.,
a path or highway
leading to a particular
city)?
244 •
but the psychological context — a result that’s consistent with our account of
this effect. As a result, you can get the benefits of context-dependent learning
through a strategy of context reinstatement — re-creating the thoughts and
feelings of the learning episode even if you’re in a very different place at the
time of recall. That’s because what matters for memory retrieval is the mental
context, not the physical environment itself.
Encoding Specificity
The results we’ve been describing also illuminate a further point: what it is
that’s stored in memory. Let’s go back to the scuba-diving experiment. The
divers in this study didn’t just remember the words they’d learned; apparently, they also remembered something about the context in which the learning took place. Otherwise, the data in Figure 7.3 (and related findings) make
no sense: If the context left no trace in memory, there’d be no way for a
return to the context to influence the divers later.
Here’s one way to think about this point, still relying on our analogy. Your
memory contains both the information you were focusing on during learning, and the highways you’ve now built, leading toward that information.
These highways — the memory connections — can of course influence your
search for the target information; that’s what we’ve been emphasizing so far.
But the connections can do more: They can also change the meaning of what
is remembered, because in many settings “memory plus this set of connections” has a different meaning from “memory plus that set of connections.”
This change in meaning, in turn, can have profound consequences for how
you remember the past.
In one of the early experiments exploring this point, participants read
target words (e.g., “piano”) in one of two contexts: “The man lifted
the piano” or “The man tuned the piano.” In each case, the sentence led
the participants to think about the target word in a particular way, and it
was this thought that was encoded into memory. In other words, what was
placed in memory wasn’t just the word “piano.” Instead, what was recorded
in memory was the idea of “piano as something heavy” or “piano as musical instrument.”
This difference in memory content became clear when participants were
later asked to recall the target words. If they had earlier seen the “lifted” sentence, they were likely to recall the target word if given the cue “something
heavy.” The hint “something with a nice sound” was much less effective.
But if participants had seen the “tuned” sentence, the result reversed: Now,
the “nice sound” hint was effective, but the “heavy” hint wasn’t (Barclay,
Bransford, Franks, McCarrell, & Nitsch, 1974). In both cases, the cue was
effective only if it was congruent with what was stored in memory.
Other experiments show a similar pattern, traditionally called encoding
specificity (Tulving, 1983; also see Hunt & Ellis, 1974; Light & Carter-Sobell,
1970). This label reminds us that what you encode (i.e., place into memory) is
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
indeed specific — not just the physical stimulus as you encountered it, but the
stimulus together with its context. Then, if you later encounter the stimulus
in some other context, you ask yourself, “Does this match anything I learned
previously?” and you correctly answer no. And we emphasize that this “no”
response is indeed correct. It’s as if you had learned the word “other” and
were later asked whether you’d been shown the word “the.” In fact, “the”
does appear as part of “other” — because the letters t h e do appear within
“other.” But it’s the whole that people learn, not the parts. Therefore, if you’ve
seen “other,” it makes sense to deny that you’ve seen “the” — or, for that matter, “he” or “her” — even though all these letter combinations are contained
within “other.”
Learning a list of words works in the same way. The word “piano” was
contained in what the research participants learned, just as “the” is contained
in “other.” What was learned, however, wasn’t just this word. Instead, what
was learned was the broader, integrated experience: the word as the perceiver
understood it. Therefore, “piano as musical instrument” isn’t what participants learned if they saw the “lifted” sentence, so they were correct in asserting that this item wasn’t on the earlier list (also see Figure 7.4).
TEST YOURSELF
3. W
hat is encoding
specificity? How is it
demonstrated?
FIGURE 7.4
REMEMBERING
“RE-CREATES” AN
EARLIER EXPERIENCE
z = –8
z = 44
R
z = 16
A
C
E
B
D
F
The text argues that what goes
into your memory is a record
of the material you’ve encountered and also a record of
the connections you established
during learning. On this basis, it
makes sense that the brain areas
activated when you’re remembering a target overlap considerably with the brain areas that
were activated when you first
encountered the target. Here,
the top panels show brain activation while viewing one picture
(A) or another picture (C) or
while hearing a particular sound
(E). The bottom panels show brain
activation while remembering
the same targets. (after wheeler,
peterson, & buckner, 2000)
Encoding Specificity
•
245
The Memory Network
In Chapter 6, we introduced the idea that memory acquisition — and, more
broadly, learning — involves the creation (or strengthening) of memory
connections. In this chapter, we’ve returned to the idea of memory connections, building on the idea that these connections serve as retrieval
paths guiding you toward the information you seek. But what are these
connections? How do they work? And who (or what?) is traveling on
these “paths”?
According to many theorists, memory is best thought of as a vast network of ideas. In later chapters, we’ll consider how exactly these ideas are
represented (as pictures? as words? in some more abstract format?). For
now, let’s just think of these representations as nodes within the network,
just like the knots in a fisherman’s net. (In fact, the word “node” is derived
from the Latin word for knot, nodus.) These nodes are tied to each other
via connections we’ll call associations or associative links. Some people find
it helpful to think of the nodes as being like light bulbs that can be turned
on by incoming electricity, and to imagine the associative links as wires that
carry the electricity.
Spreading Activation
Theorists speak of a node becoming activated when it has received a strong
enough input signal. Then, once a node has been activated, it can activate other nodes: Energy will spread out from the just-activated node via
its associations, and this will activate the nodes connected to the justactivated node.
To put all of this more precisely, nodes receive activation from
their neighbors, and as more and more activation arrives at a particular
node, the activation level for that node increases. Eventually, the activation level will reach the node’s response threshold. Once this happens, we
say that the node fires. This firing has several effects, including the fact
that the node will now itself be a source of activation, sending energy
to its neighbors and activating them. In addition, firing of the node will
draw attention to that node; this is what it means to “find” a node within
the network.
Activation levels below the response threshold, so-called subthre­
shold activation, also play an important role. Activation is assumed
to accumulate, so that two subthreshold inputs may add together, in
a process of summation, and bring the node to threshold. Likewise,
if a node has been partially activated recently, it is in effect already
“warmed up,” so that even a weak input will now be sufficient to bring it
to threshold.
These claims mesh well with points we raised in Chapter 2, when we
considered how neurons communicate with one another. Neurons receive
246 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
activation from other neurons; once a neuron reaches its threshold, it fires,
sending activation to other neurons. All of this is precisely parallel to the
suggestions we’re describing here.
Our current discussion also parallels claims offered in Chapter 4, when we
described how a network of detectors might function in object recognition.
In other words, the network linking memories to each other will resemble the
networks we’ve described linking detectors to each other (e.g., Figures 4.9
and 4.10). Detectors, like memory nodes, receive their activation from other
detectors; they can accumulate activation from different inputs, and once
activated to threshold levels, they fire.
Returning to long-term storage, however, the key idea is that activation
travels from node to node via associative links. As each node becomes
activated and fires, it serves as a source for further activation, spreading
onward through the network. This process, known as spreading activation,
enables us to deal with a key question: How does one navigate through
the maze of associations? If you start a search at one node, how do you
decide where to go from there? The answer is that in most cases you don’t
“choose” at all. Instead, activation spreads out from its starting point in
all directions simultaneously, flowing through whatever connections are
in place.
Retrieval Cues
This sketch of the memory network leaves a great deal unspecified,
but even so it allows us to explain some well-established results. For
example, why do hints help you to remember? Why, for example, do
you draw a blank if asked, “What’s the capital of South Dakota?” but
then remember if given the cue “Is it perhaps a man’s name?” Here’s one
likely explanation. Mention of South Dakota will activate nodes in
memory that represent your knowledge about this state. Activation will
then spread outward from these nodes, eventually reaching nodes that
represent the capital city’s name. It’s possible, though, that there’s only a
weak connection between the south dakota nodes and the nodes representing pierre. Maybe you’re not very familiar with South Dakota,
or maybe you haven’t thought about this state’s capital for some time.
In either case, this weak connection will do a poor job of carrying the
activation, with the result that only a trickle of activation will flow into
the pierre nodes, and so these nodes won’t reach threshold and won’t
be “found.”
Things will go differently, though, if a hint is available. If you’re told,
“South Dakota’s capital is also a man’s name,” this will activate the man’s
name node. As a result, activation will spread out from this source at the
same time that activation is spreading out from the south dakota nodes.
Therefore, the nodes for pierre will now receive activation from two sources
simultaneously, and this will probably be enough to lift the nodes’ activation
The Memory Network
•
247
FIGURE 7.5
ACTIVATION OF A NODE FROM TWO SOURCES
JACOB
SOLOMON
FRED
Nodes
representing
“South Dakota”
MAN’S
NAME
ONE OF THE
STATES
IN THE
MIDWEST
CLOSE TO
CANADA
PIERRE
TRANH
LUIS
A participant is asked, “What is the capital of South Dakota?” This activates
the south dakota nodes, and activation spreads from there to all of the
associated nodes. However, it’s possible that the connection between south
dakota and pierre is weak, so pierre may not receive enough activation to reach
threshold. Things will go differently, though, if the participant is also given
the hint “South Dakota’s capital is also a man’s name.” Now, the pierre node
will receive activation from two sources: the south dakota nodes and the
man’s name node. With this double input, it’s more likely that the pierre node
will reach threshold. This is why the hint (“man’s name”) makes the memory
search easier.
to threshold levels. In this way, question-plus-hint accomplishes more than
the question by itself (see Figure 7.5).
Semantic Priming
The explanation we’ve just offered rests on a key assumption — namely, the
summation of subthreshold activation. In other words, we relied on the idea
that the insufficient activation received from one source can add to the insufficient activation received from another source. Either source of activation
on its own wouldn’t be enough, but the two can combine to activate the
target nodes.
Can we document this summation more directly? In a lexical-decision task,
research participants are shown a series of letter sequences on a computer
screen. Some of the sequences spell words; other sequences aren’t words (e.g.,
“blar, plome”). The participants’ task is to hit a “yes” button if the sequence
spells a word and a “no” button otherwise. Presumably, they perform this
task by “looking up” these letter strings in their “mental dictionary,” and they
248 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
base their response on whether or not they find the string in the dictionary.
We can therefore use the participants’ speed of response in this task as an
index of how quickly they can locate the word in their memories.
In a series of classic studies, Meyer and Schvaneveldt (1971; Meyer,
Schvaneveldt, & Ruddy, 1974) presented participants with pairs of letter
strings, and participants had to respond “yes” if both strings were words
and “no” otherwise. For example, participants would say “yes” in response
to “chair, bread” but “no” in response to “house, fime.” Also, if both strings
were words, sometimes the words were semantically related in an obvious
way (e.g., “nurse, doctor”) and sometimes they weren’t (“cake, shoe”).
Of interest was how this relationship between the words would influence
performance.
Consider a trial in which participants see a related pair, like “bread,
butter.” To choose a response, they first need to “look up” the word
“bread” in memory. This means they’ll search for, and presumably activate, the relevant node, and in this way they’ll decide that, yes, this string
is a legitimate word. Then, they’re ready for the second word. But in this
sequence, the node for bread (the first word in the pair) has just been
activated. This will, we’ve hypothesized, trigger a spread of activation
outward from this node, bringing activation to other, nearby nodes. These
nearby nodes will surely include butter, since the association between
“bread” and “butter” is a strong one. Therefore, once the bread node
(from the first word) is activated, some activation should also spread to
the butter node.
From this base, think about what happens when a participant turns her
attention to the second word in the pair. To select a response, she must
locate “butter” in memory. If she finds this word (i.e., finds the relevant
node), then she knows that this string, too, is a word, and she can hit the
“yes” button. But the process of activating the butter node has already
begun, thanks to the (subthreshold) activation this node just received from
bread. This should accelerate the process of bringing this node to threshold (since it’s already partway there), and so it will require less time to
activate. As a result, we expect quicker responses to “butter” in this
context, compared to a context in which “butter” was preceded by some
unrelated word.
Our prediction, therefore, is that trials with related words will produce
semantic priming. The term “priming” indicates that a specific prior event
(in this case, presentation of the first word in the pair) will produce a state
of readiness (and, therefore, faster responding) later on. There are various forms of priming (in Chapter 4, we discussed repetition priming). In
the procedure we’re considering here, the priming results from the fact
that the two words in the pair are related in meaning — therefore, this is
semantic priming.
The results confirm these predictions. Participants’ lexical-decision
responses were faster by almost 100 ms if the stimulus words were related
The Memory Network
•
249
SEMANTIC PRIMING
Mean response time for pair
of words (in ms)
FIGURE 7.6
1000
900
800
700
600
First word
primes second
No priming
Condition
Participants were given a lexical-decision task involving pairs of words.
In some pairs, the words were semantically related (and so the first
word in the pair primed the second); in other pairs, the words were
unrelated (and so there was no priming). Responses to the second
word were reliably faster if the word had been primed — providing clear
evidence of the importance of subthreshold activation.
( a f t e r m e y e r & s c h va n e v e l dt , 1971)
TEST YOURSELF
4. What is subthreshold activation of a
memory node? What
role does subthreshold
activation play in explaining why retrieval
hints are often helpful?
5. How does semantic
priming illustrate the
effectiveness of subthreshold activation?
(see Figure 7.6), just as we would expect on the model we’re developing.
(For other relevant studies, including some alternative conceptions of priming, see Hutchison, 2003; Lucas, 2000.)
Before moving on, though, we should mention that this process of spreading activation — with one node activating nearby nodes — is not the whole
story for memory search. As one complication, people have some degree of
control over the starting points for their memory searches, relying on the processes of reasoning (Chapter 12) and the mechanisms of executive control
(Chapters 5 and 6). In addition, evidence suggests that once the spreading
activation has begun, people have the option of “shutting down” some of this
spread if they’re convinced that the wrong nodes are being activated (e.g.,
Anderson & Bell, 2001; Johnson & Anderson, 2004). Even so, spreading
activation is a crucial mechanism. It plays a central role in retrieval, and it
helps us understand why memory connections are so important and so helpful.
Different Forms of Memory Testing
Let’s pause to review. In Chapter 6, we argued that learning involves the
creation or strengthening of connections. This is why memory is promoted
by understanding (because understanding consists, in large part, of seeing
250 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
how new material is connected to other things you know). We also proposed that these connections later serve as retrieval paths, guiding your
search through the vast warehouse that is memory. In this chapter, we’ve
explored an important implication of this idea: that (like all paths) the
paths through memory have both a starting point and an end point. Therefore, retrieval paths will be helpful only if you’re at the appropriate starting point; this, we’ve proposed, is the basis for the advantage produced
by context reinstatement. And, finally, we’ve now started to lay out what
these paths really are: connections that carry activation from one memory
to another.
This theoretical base also helps us with another issue: the impact of
different forms of memory testing. Both in the laboratory and in dayto-day life, you often try to recall information from memory. This means
that you’re presented with a retrieval cue that broadly identifies the information you seek, and then you need to come up with the information on
your own: “What was the name of that great restaurant your parents took
us to?”; “Can you remember the words to that song?”; “Where were you
last Saturday?”
In other circumstances, you draw information from your memory via
recognition. This term refers to cases in which information is presented to
you, and you must decide whether it’s the sought-after information: “Is this
the man who robbed you?”; “I’m sure I’ll recognize the street when we get
there”; “If you let me taste that wine, I’ll tell you if it’s the same one we had
last time.”
These two modes of retrieval — recall and recognition — are fundamentally different from each other. Recall requires memory search because
you have to come up with the sought-after item on your own; you need to
locate that item within memory. As a result, recall depends heavily on the
memory connections we’ve been emphasizing so far. Recognition, in contrast, often depends on a sense of familiarity. Imagine, for example, that
you’re taking a recognition test, and the fifth word on the test is “butler.”
In response to this word, you might find yourself thinking, “I don’t recall
seeing this word on the list, but this word feels really familiar, so I guess I
must have seen it recently. Therefore, it must have been on the list.” In this
case, you don’t have source memory; that is, you don’t have any recollection of the source of your current knowledge. But you do have a strong
sense of familiarity, and you’re willing to make an inference about where
that familiarity came from. In other words, you attribute the familiarity
to the earlier encounter, and thanks to this attribution you’ll probably
respond “yes” on the recognition test.
Familiarity and Source Memory
We need to be clear about our terms here, because source memory is actually a type of recall. Let’s say, for example, that you hear a song on the radio
and say, “I know I’ve heard this song before because it feels familiar and
Different Forms of Memory Testing
•
251
I remember where I heard it.” In this setting, you’re able to remember the
source of your familiarity, and that means you’re recalling when and where
you encountered the song. On this basis, we don’t need any new theory to
talk about source memory, because we can use the same theory that we’d
use for other forms of recall. Hearing the song was the retrieval cue that
launched a search through memory, a search that allowed you to identify
the setting in which you last encountered the song. That search (like any
search) was dependent on memory connections, and would be explained by
the spreading activation process that we’ve already described.
But what about familiarity? What does this sort of remembering
involve? As a start, let’s be clear that familiarity is truly distinct from source
memory. This is evident in the fact that the two types of memory are independent of each other — it’s possible for an event to be familiar without any
source memory, and it’s possible for you to have source memory without
any familiarity. This independence is evident when you’re watching a movie
and realize that one of the actors is familiar, but (sometimes with considerable frustration, and despite a lot of effort) you can’t recall where you’ve
seen that actor before. Or you’re walking down the street, see a familiar
face, and find yourself asking, “Where do I know that woman from? Does
she work at the grocery store I shop in? Is she the driver of the bus I often
take?” You’re at a loss to answer these questions; all you know is that the
face is familiar.
In cases like these, you can’t “place” the memory; you can’t identify the
episode in which the face was last encountered. But you’re certain the face is
familiar, even though you don’t know why — a clear example of familiarity
without source memory.
The inverse case is less common, but it too can be demonstrated. For
example, in Chapter 2 we discussed Capgras syndrome. Someone with this
syndrome might have detailed, accurate memories of what friends and family
members look like, and probably remembers where and when these other
people were last encountered. Even so, when these other people are in view
they seem hauntingly unfamiliar. In this setting, there is source memory without familiarity. (For further evidence — and a patient who, after surgery, has
intact source memory but disrupted familiarity — see Bowles et al., 2007;
also see Yonelinas & Jacoby, 2012.)
We can also document the difference between source memory and familiarity in another way. In many studies, (neurologically intact) participants
have been asked, during a recognition test, to make a “remember/know”
distinction. This involves pressing one button (to indicate “remember”) if
they actually recall the episode of encountering a particular item, and pressing a different button (“know”) if they don’t recall the encounter but just
have a broad feeling that the item must have been on the earlier list. With
one response, participants are indicating that they have a source memory;
with the other, they’re indicating an absence of source memory. Basically, a
participant using the “know” response is saying, “This item seems familiar,
252 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
“FAMILIAR . . . BUT WHERE DO I KNOW HIM FROM?!?”
The photos here show successful TV or film actors. The odds are good that for some of them you’ll immediately know
their faces as familiar but won’t be sure why. You know you’ve seen these actors in some movie or show, but which one?
(We provide the actors’ names at the chapter’s end.)
so I know it was on the earlier list even though I don’t remember the experience of seeing it” (Gardiner, 1988; Hicks & Marsh, 1999; Jacoby, Jones, &
Dolan, 1998).
Researchers can use fMRI scans to monitor participants’ brain activity while they’re taking these memory tests, and the scans indicate that
“remember” and “know” judgments depend on different brain areas. The
scans show heightened activity in the hippocampus when participants indicate that they “remember” a particular test item, suggesting that this brain
structure is crucial for source memory. In contrast, “know” responses
are associated with activity in a different area — the anterior parahippocampus, with the implication that this brain site is crucial for familiarity.
(See Aggleton & Brown, 2006; Diana, Yonelinas, & Ranganath, 2007;
Dobbins, Foley, Wagner, & Schacter, 2002; Eldridge, Knowlton, Furmanski,
Bookheimer, & Engel, 2000; Montaldi, Spencer, Roberts, & Mayes, 2006;
Wagner, Shannon, Kahn, & Buckner, 2005. Also see Rugg & Curran, 2007;
Rugg & Yonelinas, 2003.)
Familiarity and source memory can also be distinguished during learning. If certain brain areas (e.g., the rhinal cortex) are especially active during
learning, then the stimulus is likely to seem familiar later on. In contrast,
if other brain areas (e.g., the hippocampal region) are particularly active
during learning, there’s a high probability that the person will indicate
source memory for that stimulus when tested later (see Figure 7.7). (See, e.g.,
Davachi & Dobbins, 2008; Davachi, Mitchell, & Wagner, 2003; Ranganath
et al., 2003.)
We still need to ask, though, what’s going on in these various brain areas
to create the relevant memories. Activity in the hippocampus is probably
helping to create the memory connections we’ve been discussing all along,
and it’s these connections, we’ve suggested, that promote source memory. But
TEST YOURSELF
6.Define “recognition”
and “recall.”
7. W
hat evidence indicates that source
memory and familiarity are distinct from
each other?
Different Forms of Memory Testing
•
253
FIGURE 7.7
FAMILIARITY VERSUS SOURCE MEMORY
Subsequent recollection effects
Subsequent familiarity effects
If the rhinal cortex
was especially
activated during
encoding, then
the stimulus was
likely to seem
familiar when
viewed later on.
If the hippocampus
was especially
activated during
encoding, then later
on the participant
was likely to
recollect having
seen that stimulus.
0.0000008
0.0006
Posterior hippocampus
Rhinal cortex
0.0005
Level of activation
Level of activation
0.0000006
0.0000004
0.0000002
0.0000000
0.0003
0.0002
0.0001
–0.0000002
0.0000
1
A
0.0004
2
3
4
5
Recognition confidence
6
Source
incorrect
Source
correct
B
In this study, researchers tracked participants’ brain activity during encoding and then analyzed the data according
to what happened later, when the time came for retrieval.
( after ranganath et al ., 2003)
what about familiarity? What “record” does it leave in memory? The answer
to this question leads us to a very different sort of memory.
Implicit Memory
How can we find out if someone remembers a previous event? The obvious
path is to ask her — “How did the job interview go?”; “Have you ever seen
Casablanca?”; “Is this the book you told me about?” But at the start of this
chapter, we talked about a different approach: We can expose someone to an
event, and then later re-expose her to the same event and assess whether her
response on the second encounter is different from the first. Specifically, we
254 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
can ask whether the first encounter somehow primed the person — got her
ready — for the second exposure. If so, it would seem that the person must
retain some record of the first encounter — she must have some sort of memory.
Memory without Awareness
In a number of studies, participants have been asked to read through a list
of words, with no indication that their memories would be tested later on.
(They might be told that they’re merely checking the list for spelling errors.)
Then, sometime later, the participants are given a lexical-decision task: They
are shown a series of letter strings and, for each, must indicate (by pressing
one button or another) whether the string is a word or not. Some of the letter
strings in the lexical-decision task are duplicates of the words seen in the first
part of the experiment (i.e., they were on the list participants had checked for
spelling), enabling us to ask whether the first exposure somehow primed the
participants for the second encounter.
In these experiments, lexical decisions are quicker if the person has
recently seen the test word; that is, lexical decision shows the pattern that in
Chapter 4 we called “repetition priming” (e.g., Oliphant, 1983). Remarkably,
this priming is observed even when participants have no recollection for having encountered the stimulus words before. To demonstrate this, we can show
participants a list of words and then test them in two different ways. One test
assesses memory directly, using a standard recognition procedure: “Which of
these words were on the list I showed you earlier?” The other test is indirect
and relies on lexical decision: “Which of these letter strings form real words?”
In this procedure, the two tests will yield different results. At a sufficient delay,
the direct memory test is likely to show that the participants have completely
forgotten the words presented earlier; their recognition performance is essentially random. According to the lexical-decision results, however, the participants still remember the words — and so they show a strong priming effect.
In this situation, then, participants are influenced by a specific past experience
that they seem (consciously) not to remember at all — a pattern that some
researchers refer to as “memory without awareness.”
A different example draws on a task called word-stem completion. In
this task, participants are given three or four letters and must produce a
word with this beginning. If, for example, they’re given cla-, then “clam” or
“clatter” would be acceptable responses, and the question of interest for us
is which of these responses the participants produce. It turns out that people
are more likely to offer a specific word if they’ve encountered it recently;
once again, this priming effect is observed even if participants, when tested
directly, show no conscious memory of their recent encounter with that word
(Graf, Mandler, & Haden, 1982).
Results like these lead psychologists to distinguish two types of memory.
Explicit memories are those usually revealed by direct memory testing —
testing that urges participants to remember the past. Recall is a direct memory test; so is a standard recognition test. Implicit memories, however, are
Implicit Memory
•
255
typically revealed by indirect memory testing and are often manifested as
priming effects. In this form of testing, participants’ current behavior is
demonstrably influenced by a prior event, but they may be unaware of this.
Lexical decision, word-stem completion, and many other tasks provide
indirect means of assessing memory. (See, for example, Mulligan & Besken,
2013; for a different perspective on these data, though, see Cabeza &
Moscovitch, 2012.)
How exactly is implicit memory different from explicit memory? We’ll say
more about this question before we’re done; but first we need to say more
about how implicit memory feels from the rememberer’s point of view. This
will lead us back into our discussion of familiarity and source memory.
False Fame
In a classic research study, Jacoby, Kelley, Brown, and Jasechko (1989) presented participants with a list of names to read out loud. The participants were
told nothing about a memory test; they thought the experiment was concerned
with how they pronounced the names. Some time later, during the second step
of the procedure, the participants were shown a new list of names and asked
to rate each person on this list according to how famous each one was. The list
included some real, very famous people; some real but not-so-famous people;
and some fictitious names that the experimenters had invented. Crucially, the
fictitious names were of two types: Some had occurred on the prior (“pronunciation”) list, and some were simply new names. A comparison between those
two types will indicate how the prior familiarization (during the pronunciation
task) influenced the participants’ judgments of fame.
For some participants, the “famous” list was presented right after the
“pronunciation” list; for other participants, there was a 24-hour delay
between these two steps. To see how this delay matters, imagine that you’re
a participant in the immediate-testing condition: When you see one of the
fictitious-but-familiar names, you might decide, “This name sounds familiar,
but that’s because I just saw it on the previous list.” In this situation, you have
a feeling that the (familiar) name is distinctive, but you also know why it’s
distinctive — because you remember your earlier encounter with the name.
In other words, you have both a sense of familiarity and a source memory,
so there’s nothing here to persuade you that the name belongs to someone
famous, and you respond accordingly. But now imagine that you’re a participant in the other condition, with the 24-hour delay. Because of the delay, you
may not recall the earlier episode of seeing the name in the pronunciation
task. But the broad sense of familiarity remains anyway, so in this setting you
might say, “This name rings a bell, and I have no idea why. I guess this must
be a famous person.” And this is, in fact, the pattern of the data: When the
two lists are presented one day apart, participants are likely to rate the madeup names as being famous.
Apparently, the participants in this study noted (correctly) that some of
the names did “ring a bell” and so did trigger a certain feeling of familiarity.
256 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
The false judgments of fame, however, come from the way the participants
interpreted this feeling and what conclusions they drew from it. Basically,
participants in the 24-hour-delay condition forgot the real source of the
familiarity (appearance on a recently viewed list) and instead filled in a
bogus source (“Maybe I saw this person in a movie?”). And it’s easy to see
why they made this misattribution. After all, the experiment was described
to them as being about fame, and other names on the list were actually those
of famous people. From the participants’ point of view, therefore, it was
reasonable to infer in this setting that any name that “rings a bell” belongs
to a famous person.
We need to be clear, though, that this misattribution is possible only
because the feeling of familiarity produced by these names was relatively
vague, and therefore open to interpretation. The suggestion, then, is that
implicit memories may leave people with only a broad sense that a stimulus
is somehow distinctive — that it “rings a bell” or “strikes a chord.” What
happens after this depends on how they interpret that feeling.
Implicit Memory and the “Illusion of Truth”
How broad is this potential for misinterpreting an implicit memory?
Participants in one study heard a series of statements and had to judge
how interesting each statement was (Begg, Anas, & Farinacci, 1992). As an
example, one sentence was “The average person in Switzerland eats about
25 pounds of cheese each year.” (This is false; the average in 1992, when the
experiment was done, was closer to 18 pounds.) Another was “Henry Ford
forgot to put a reverse gear in his first automobile.” (This is true.)
After hearing these sentences, the participants were presented with some
more sentences, but now they had to judge the credibility of these sentences,
rating them on a scale from certainly true to certainly false. However, some
of the sentences in this “truth test” were repeats from the earlier presentation, and the question of interest is how sentence credibility is influenced by
sentence familiarity.
The result was a propagandist’s dream: Sentences heard before were
more likely to be accepted as true; that is, familiarity increased credibility.
(See Begg, Armour, & Kerr, 1985; Brown & Halliday, 1990; Fiedler, Walther,
Armbruster, Fay, & Naumann, 1996; Moons, Mackie, & Garcia-Marques,
2009; Unkelbach, 2007.) This effect was found even when participants were
warned in advance not to believe the sentences in the first list. In one procedure, participants were told that half of the statements had been made
by men and half by women. The women’s statements, they were told, were
always true; the men’s, always false. (Half the participants were told the
reverse.) Then, participants rated how interesting the sentences were, with
each sentence attributed to either a man or a woman: for example, “Frank
Foster says that house mice can run an average of 4 miles per hour” or “Gail
Logan says that crocodiles sleep with their eyes open.” Later, participants
were presented with more sentences and had to judge their truth, with these
Implicit Memory
•
257
new sentences including the earlier assertions about mice, crocodiles, and
so forth.
Let’s focus on the sentences initially identified as being false — in our
example, Frank’s claim about mice. If someone explicitly remembers this
sentence (“Oh yes — Frank said such and such”), then he should judge
the assertion to be false (“After all, the experimenter said that the men’s
statements were all lies”). But what about someone who lacks this explicit
memory? This person will have no conscious recall of the episode in which
he last encountered this sentence (i.e., will have no source memory), and
so he won’t know whether the assertion came from a man or a woman.
He therefore can’t use the source as a basis for judging the truthfulness of
the sentence. But he might still have an implicit memory for the sentence
left over from the earlier exposure (“Gee, that statement rings a bell”), and
this might increase his sense of the statement’s credibility (“I’m sure I’ve
heard that somewhere before; I guess it must be true”). This is exactly the
pattern of the data: Statements plainly identified as false when they were
first heard still created the so-called illusion of truth; that is, these statements were subsequently judged to be more credible than sentences never
heard before.
The relevance of this result to the political arena or to advertising should
be clear. A newspaper headline might inquire, “Is Mayor Wilson a crook?”
Or the headline might declare, “Known criminal claims Wilson is a crook!”
In either case, the assertion that Wilson is a crook would become familiar.
The Begg et al. data indicate that this familiarity will, by itself, increase
the likelihood that you’ll later believe in Wilson’s dishonesty. This will be
true even if the paper merely raised the question; it will be true even if the
allegation came from a disreputable source. Malicious innuendo does, in
fact, produce nasty effects. (For related findings, see Ecker, Lewandowsky,
Chang, & Pillai, 2014.)
Attributing Implicit Memory to
the Wrong Source
Apparently, implicit memory can influence us (and, perhaps, bias us) in the
political arena. Other evidence suggests that implicit memory can influence
us in the marketplace — and can, for example, guide our choices when we’re
shopping (e.g., Northup & Mulligan, 2013, 2014). Yet another example
involves the justice system, and it’s an example with troubling implications.
In an early study by Brown, Deffenbacher, and Sturgill (1977), research participants witnessed a staged crime. Two or three days later, they were shown
“mug shots” of individuals who supposedly had participated in the crime.
But as it turns out, the people in these photos were different from the actual
“criminals” — no mug shots were shown for the truly “guilty” individuals.
Finally, after four or five more days, the participants were shown a lineup
and asked to select the individuals seen in Step 1 — namely, the original crime
(see Figure 7.8).
258 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
FIGURE 7.8
A PHOTO LINEUP
On TV, crime victims view a live lineup, but it’s far more common in the United
States for the victim (or witness) to see a “photo lineup” like this one. Unfortunately, victims sometimes pick the wrong person, and this error is more likely
to occur if the suspect is familiar to the victim for some reason other than the
crime. The error is unlikely, though, if the face is very familiar, because in that
case the witness will have both a feeling of familiarity and an accurate source
memory. (“Number Two looks familiar, but that’s because I see him at the gym
all the time.”)
The data in this study show a pattern known as source confusion. The
participants correctly realized that one of the faces in the lineup looked familiar, but they were confused about the source of the familiarity. They falsely
believed they had seen the person’s face in the original “crime,” when, in
truth, they’d seen that face only in a subsequent photograph. In fact, the
likelihood of this error was quite high, with 29% of the participants (falsely)
selecting from the lineup an individual they had seen only in the mug shots.
(Also see Davis, Loftus, Vanous, & Cucciare, 2008; Kersten & Earles, 2017.
For examples of similar errors that interfere with real-life criminal investigations, see Garrett, 2011. For a broader discussion of eyewitness errors, see
Reisberg, 2014.)
TEST YOURSELF
8. W
hat is the difference between implicit
and explicit memory?
Which of these is said
to be “memory without awareness”?
9.What is the role of
implicit memory in
explaining the false
fame effect?
Implicit Memory
•
259
COGNITION
outside the lab
Cryptoplagiarism
In 1970, (former Beatle) George Harrison released
when, in fact, they’d been mentioned by some one
the song “My Sweet Lord.” It turned out that the
else in the initial session.
song is virtually identical to one released years
This pattern fits well with the chapter’s discus-
before that — ”He’s So Fine,” by the Chiffons — and
sion of implicit memory. The participants in this
in 1976 Harrison was found guilty of copyright
study (and others) have lost their explicit memory
infringement. (You can find both recordings on
of the earlier episode in which they encountered
YouTube, and you’ll instantly see that they’re
someone else’s ideas. Even so, an implicit memory
essentially the same song.) In his conclusion to the
remains and emerges as a priming effect. With
court proceedings, the judge wrote, “Did Harrison
certain words primed in memory, participants
deliberately use the music of ‘He’s So Fine’? I do
are more likely to produce those words when
not believe he did so deliberately. Nevertheless,
asked — with no realization that their production
it is clear that ‘My Sweet Lord’ is the very same
has been influenced by priming.
song as ‘He’s So Fine.’ . . . This is, under the law,
Likewise, imagine talking with a friend about
infringement of copyright, and is no less so even
your options for an upcoming writing assignment.
though subconsciously accomplished” (Bright
Your friend suggests a topic, but after a moment’s
Tunes Music Corp. v. Harrisongs Music, Ltd., 420 F.
thought you reject the suggestion, convinced
Supp. 177 — Dist. Court, SD New York 1976).
that the topic is too challenging. A few days later,
How can we understand the judge’s remarks?
though, you’re again trying to choose a topic, and
Can there be “subconscious” plagiarism? The
(thanks to the priming) your friend’s suggestion
answer is yes, and the pattern at issue is some-
comes to mind. As a result of the earlier conversa-
times referred to as “cryptoplagiarism” — inadvertent
tion with your friend, you’ve had some “warm-up”
copying that is entirely unwitting and uncontrolla-
in considering this topic, so your thinking now is a
ble, and usually copying that comes with the strong
bit more fluent — and you decide that the topic isn’t
sense that you’re the inventor of the idea, even
so challenging. As a result, you go forward with the
though you’ve taken the idea from someone else.
topic. You may, however, have no explicit memory
In one early study, participants sat in groups
of the initial conversation with your friend — and you
and were asked to generate words in particu-
may not realize either that the idea “came to you”
lar categories — for example, names of sports or
because of a priming effect or that the idea seemed
musical instruments (Brown & Murphy, 1989; also
workable because of the “warm-up” provided by
Marsh, Ward, & Landau, 1999). Later, the same
the earlier conversation. The outcome: You’ll pre­
participants were asked to recall the words they
sent the idea as though it’s entirely your own, giving
(and not others in the group) had generated, and
your friend none of the credit she deserves.
also to generate new entries in the same category.
We’ll never know if this is what happened with
In this setting, participants often “borrowed” oth-
George Harrison. Even so, the judge’s conclusion
ers’ contributions — sometimes (mis)remembering
in that case seems entirely plausible, and there’s
others’ words as though they had themselves pro-
no question that inadvertent, unconscious plagia-
duced them, sometimes offering words as “new”
rism is a real phenomenon.
260 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
Theoretical Treatments of Implicit Memory
One message coming from these studies is that people are often better at
remembering that something is familiar than they are at remembering why
it is familiar. This explains why it’s possible to have a sense of familiarity
without source memory (“I’ve seen her somewhere before, but I can’t figure
out where!”) and also why it’s possible to be correct in judging familiarity
but mistaken in judging source.
In addition, let’s emphasize that in many of these studies participants are
being influenced by memories they aren’t aware of. In some cases, participants
realize that a stimulus is somehow familiar, but they have no memory of the
encounter that produced the familiarity. In other cases, they don’t even have a
sense of familiarity for the target stimulus; nonetheless, they’re influenced by
their previous encounter with the stimulus. For example, experiments show
that participants often prefer a previously presented stimulus over a novel stimulus, even though they have no sense of familiarity with either stimulus. In such
cases, people have no idea that their preference is being guided by memory
(Murphy, 2001; also Montoy, Horton, Vevea, Citkowicz, & Lauber, 2017).
It does seem, then, that the phrase “memory without awareness” is appropriate, and it does make sense to describe these memories as implicit memories. But how can we explain this form of unconscious “remembering”?
Processing Fluency
Our discussion so far — here and in Chapters 4 and 5 — has laid the foundation for a proposal about implicit memory. Let’s build the argument in steps.
When a stimulus arrives in front of your eyes, it triggers certain detectors,
and these trigger other detectors, and these still others, until you recognize the
object. (“Oh, it’s my stuffed bear, Blueberry.”) We can think of this sequence
as involving a “flow” of activation that moves from detector to detector. We
could, if we wished, keep track of this flow and in this way identify the “path”
that the activation traveled through the network. Let’s refer to this path as a
processing pathway — the sequence of detectors, and the connections between
detectors, that the activation flows through in recognizing a specific stimulus.
In the same way, we’ve proposed in this chapter that remembering often
involves the activation of a node, and this node triggers other, nearby, nodes
so that they become activated; they trigger still other nodes, leading eventually to the information you seek in memory. So here, too, we can speak
of a processing pathway — the sequence of nodes, and connections between
nodes, that the activation flows through during memory retrieval.
We’ve also said the use of a processing pathway strengthens that pathway.
This is because the baseline activation level of nodes or detectors increases if
the nodes or detectors have been used frequently in the past, or if they’ve been
used recently. Likewise, connections (between detectors or nodes) grow stronger
with use. For example, by thinking about the link between, say, “Jacob” and
“Boston,” you can strengthen the connection between the corresponding nodes,
and this will help you remember that your friend Jacob comes from Boston.
Theoretical Treatments of Implicit Memory
•
261
Now, let’s put the pieces together. Use of a processing pathway strengthens
the pathway. As a result, the pathway will be a bit more efficient, a bit faster,
the next time you use it. Theorists describe this fact by saying that use of a
pathway increases the pathway’s processing fluency — that is, the speed and
ease with which the pathway will carry activation.
In many cases, this is all the theory we need to explain implicit memory
effects. Consider implicit memory’s effect on lexical decision. In this procedure, you first are shown a list of words, including the word “bubble.” Then,
we ask you to do the lexical-decision task, and we find that you’re faster
for words (like “bubble”) that had been included in the earlier list. This increase in speed provides evidence for implicit memory, and the explanation
is straightforward. When we show you “bubble” early in the experiment,
you read the word, and this involves activation flowing through the appropriate processing pathway for this word. This warms up the pathway, and
as a result the path’s functioning will be more fluent the next time you use
it. Of course, when “bubble” shows up later as part of the lexical-decision
task, it’s handled by the same (now more fluent) pathway, and so the word is
processed more rapidly — exactly the outcome that we’re trying to explain.
For other implicit-memory effects, though, we need a further assumption —
namely, that people are sensitive to the degree of processing fluency. That is,
just as people can tell whether they’ve lifted a heavy carton or a lightweight
one, or whether they’ve answered an easy question (“What’s 2 + 2?”) or a
harder one (“What’s 17 3 19?”), people also have a broad sense of when they
have perceived easily and when they have perceived only by expending more
effort. They likewise know when a sequence of thoughts was particularly fluent
and when the sequence was labored.
This fluency, however, is perceived in an odd way. For example, when
a stimulus is easy to perceive, you don’t experience something like “That
stimulus sure was easy to recognize!” Instead, you merely register a vague
sense of specialness. You feel that the stimulus “rings a bell.” No matter how
it is described, though, this sense of specialness has a simple cause — namely,
the detection of fluency, created by practice.
There’s one complication, however. What makes a stimulus feel “special”
may not be fluency itself. Instead, people seem sensitive to changes in fluency
(e.g., they notice if it’s a little harder to recognize a face this time than it
was in the past). People also seem to notice discrepancies between how easy
(or hard) it was to carry out some mental step and how easy (or hard) they
expected it to be (Wanke & Hansen, 2015; Whittlesea, 2002). In other words,
a stimulus is registered as distinctive, or “rings a bell,” when people detect a
change or a discrepancy between experience and expectations.
To see how this matters, imagine that a friend unexpectedly gets a haircut (or gets new eyeglasses, or adds or removes some facial hair). When you
see your friend, you realize immediately that something has changed, but
you’re not sure what. You’re likely to ask puzzled questions (“Are those new
glasses?”) and get a scornful answer. (“No, you’ve seen these glasses a hundred times over the last year.”) Eventually your friend tells you what the
262 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
change is — pointing out that you failed to notice that he’d shaved off his
mustache (or some such).
What’s going on here? You obviously can still recognize your friend, but
your recognition is less fluent than in the past because of the change in your
friend’s appearance, and you notice this change — but then are at a loss to
explain it (see Figure 7.9).
On all of these grounds, we need another step in our hypothesis, but it’s
a step we’ve already introduced: When a stimulus feels special (because of
a change in fluency, or a discrepancy between the fluency expected and the
fluency experienced), you often want to know why. Thus the vague feeling of
specialness (again, produced by fluency) can trigger an attribution process, as
you ask, “Why did that stimulus stand out?”
In many circumstances, you’ll answer this question correctly, and so the
specialness will be (accurately) interpreted as familiarity and attributed to
the correct source. (“That woman seems distinctive, and I know why: It’s the
woman I saw yesterday in the dentist’s office.”) Often, you make this attribution because you have the relevant source memory — and this memory guides
you in deciding why a stimulus (a face, a song, a smell) seems to stand out. In
other cases, you make a reasonable inference, perhaps guided by the context.
(“I don’t remember where I heard this joke before, but it’s the sort of joke
that Conor is always telling, so I bet it’s one of his and that’s why the joke
is familiar.”) In other situations, though, things don’t go so smoothly, and
FIGURE 7.9
CHANGES IN APPEARANCE
The text emphasizes our sensitivity to increases in fluency, but we
can also detect decreases. In viewing a picture of a well-known
actor, for example, you might notice immediately that something
is new in his appearance, but you might be unsure about what
exactly the change involves. In this setting, the change in appearance
disrupts your well-practiced steps of perceiving for an otherwise
familiar face, so the perception is less fluent than in the past. This
lack of fluency is what gives you the “something is new” feeling. But
then the attribution step fails: You can’t identify what produced this
feeling (so you end up offering various weak hypotheses, such as
“Is that a new haircut?” when, in fact, it’s the mustache and goatee that are new). This case provides the mirror image of the cases
we’ve been considering, in which familiarity leads to an increase in
fluency, so that something “rings a bell” but you can’t say why. In the
picture shown here, you probably recognize Denzel Washington, and
you probably also realize that something is “off” in the picture. Can
you figure out what it is? (We’ve actually made several changes to
Denzel’s appearance; can you spot them all?)
Theoretical Treatments of Implicit Memory
•
263
so — as we have seen — people sometimes misinterpret their own processing
fluency, falling prey to the errors and illusions we have been discussing.
The Nature of Familiarity
All of these points provide us — at last — with a proposal for what “fami­
liarity” is, and the proposal is surprisingly complex. You might think that
familiarity is simply a feeling that’s produced more or less directly when you
encounter a stimulus you’ve met before. But the research findings described
in the last few sections point toward a different proposal — namely, that
“familiarity” is more like a conclusion that you draw rather than a feeling
triggered by a stimulus. Specifically, the evidence suggests that a stimulus
will seem familiar whenever the following list of requirements is met: First,
you have encountered the stimulus before. Second, because of that prior
encounter (and the “practice” it provided), your processing of that stimulus
is now faster and more efficient; there is, in other words, an increase in processing fluency. Third, you detect that increased fluency, and this leads you
to register the stimulus as somehow distinctive or special. Fourth, you try
to figure out why the stimulus seems special, and you reach a particular
conclusion — namely, that the stimulus has this distinctive quality because it’s
a stimulus you’ve met before in some prior episode (see Figure 7.10).
FIGURE 7.10
THE CHAIN OF EVENTS LEADING TO THE SENSE OF “FAMILIARITY”
The Steps Leading to a Judgment of Familiarity
Exposure
to a
stimulus
Practice
in
perceiving
Fluency
Stimulus
registered
as “special”
Attribution
of fluency,
perhaps
attribution
to a specific
prior event
“Familiarity”
Stimulus
registered
as “special”
Attribution
of fluency,
perhaps
attribution
to a specific
prior event
“Familiarity”
The Creation of an Illusion of Familiarity
Manipulation of stimulus
presentation designed to
make perceiving easier
Fluency
In the top line, practice in perceiving leads to fluency, and if the person attributes the fluency to some specific
prior event, the stimulus will “feel familiar.” The bottom line, however, indicates that fluency can be created in other
ways: by presenting the stimulus more clearly or for a longer exposure. Once this fluency is detected, though, it
can lead to steps identical to those in the top row. In this way, an “illusion of familiarity” can be created.
264 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
Let’s be clear, though, that none of these steps happens consciously — you’re
not aware of seeking an interpretation or trying to explain why a stimulus
feels distinctive. All you experience consciously is the end product of all these
steps: the sense that a stimulus feels familiar. Moreover, this conclusion about
a stimulus isn’t one you draw capriciously; instead, you’re likely to arrive at
this conclusion and decide a stimulus is familiar only when you have supporting information. Thus, imagine that you encounter a stimulus that “rings a
bell.” As we mentioned before, you’re likely to decide the stimulus is familiar
if you also have an (explicit) source memory, so that you can recall where and
when you last encountered that stimulus. You’re also more likely to decide a
stimulus is familiar if the surrounding circumstances support it. For example,
if you’re asked, “Which of these words were on the list you saw earlier?” the
question itself gives you a cue that some of the words were recently encountered, and so you’re more likely to attribute fluency to that encounter.
The fact remains, though, that judgments like these sometimes go astray,
which is why we need this complicated theory. We’ve considered several cases
in which a stimulus is objectively familiar (you’ve seen it recently) but doesn’t
feel familiar — just as our theory predicts. In these cases, you detect the fluency
but attribute it to some other source. (“That melody is lovely” rather than “The
melody is familiar.”) In other words, you go through all of the steps shown in
the top of Figure 7.10 except for the last two: You don’t attribute the fluency to
a specific prior event, and so you don’t experience a sense of familiarity.
We can also find the opposite sort of case — in which a stimulus is not
familiar (i.e., you’ve not seen it recently) but feels familiar anyhow — and
this, too, fits with the theory. This sort of illusion of familiarity can be
produced if the processing of a completely novel stimulus is more fluent than you
expected — perhaps because (without telling you) we’ve sharpened the focus of
a computer display or presented the stimulus for a few milliseconds longer than
other stimuli you’re inspecting (Jacoby & Whitehouse, 1989; Whittlesea, 2002;
Whittlesea, Jacoby, & Girard, 1990). Cases like these can lead to the situation
shown in the bottom half of Figure 7.10. And as our theory predicts, these situations do produce an illusion: Your processing of the stimulus is unexpectedly
fluent; you seek an attribution for this fluency, and you’re fooled into thinking the stimulus is familiar — so you say you’ve seen the stimulus before, when
in fact you haven’t. This illusion is a powerful confirmation that the sense of
familiarity does rest on processes like the ones we’ve described. (For more on
fluency, see Besken & Mulligan, 2014; Griffin, Gonzalez, Koehler, & Gilovich,
2012; Hertwig, Herzog, Schooler, & Reimer, 2008; Lanska, Olds, & Westerman,
2013; Oppenheimer, 2008; Tsai & Thomas, 2011. For a glimpse of what
fluency amounts to in the nervous system, see Knowlton & Foerde, 2008.)
The Hierarchy of Memory Types
Clearly, we’re often influenced by the past without being aware of that influence. We often respond differently to familiar stimuli than we do to novel
stimuli, even if we have no subjective feeling of familiarity. On this basis, it
Theoretical Treatments of Implicit Memory
•
265
seems that our conscious recollection seriously underestimates what’s in our
memories, and research has documented many ways in which unconscious
memories influence what we do, think, and feel.
In addition, the data are telling us that there are two different kinds
of memory: one type (“explicit”) is conscious and deliberate, the other
(“implicit”) is typically unconscious and automatic. These two broad categories can be further subdivided, as shown in Figure 7.11. Explicit memories
can be subdivided into episodic memories (memory for specific events) and
semantic memory (more general knowledge). Implicit memory is often
divided into four subcategories, as shown in the figure. Our emphasis here
has been on one of the subtypes — priming — largely because of its role in
producing the feeling of familiarity. However, the other subtypes of implicit
memory are also important and can be distinguished from priming both in
terms of their functioning (i.e., they follow somewhat different rules) and in
terms of their biological underpinnings.
Some of the best evidence for these distinctions, though, comes from the
clinic, not the laboratory. In other words, we can learn a great deal about
TEST YOURSELF
10. W
hat is processing
fluency, and how
does it influence us?
11. In what sense is
familiarity more like
a conclusion that you
draw, rather than a
feeling triggered by a
stimulus?
FIGURE 7.11
HIERARCHY OF MEMORY TYPES
Memory
Explicit memory
Conscious
Episodic memory
Memory for
specific events
Implicit memory
Revealed by indirect tests
Semantic memory
General knowledge,
not tied to any
time or place
Procedural memory
Knowing how
(i.e., memory
for skills)
Priming
Changes in
perception and
belief caused by
previous experience
Perceptual learning
Recalibration of
perceptual systems
as a result of
experience
Classical conditioning
Learning about
associations
among stimuli
In our discussion, we’ve distinguished two types of memory — explicit and implicit. However, there are reasons to
believe that each of these categories must be subdivided further, as shown here. Evidence for these subdivisions
includes functional evidence (the various types of memory follow different rules) and biological evidence (the
types depend on different aspects of brain functioning).
266 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
these various types of memory by considering individuals who have suffered
different forms of brain damage. Let’s look at some of that evidence.
Amnesia
As we have already mentioned, a variety of injuries or illnesses can lead to a loss
of memory, or amnesia. Some forms of amnesia are retrograde, meaning that
they disrupt memory for things learned prior to the event that initiated the amnesia (see Figure 7.12). Retrograde amnesia is often caused by blows to the head;
the afflicted person is unable to recall events that occurred just before the blow.
Other forms of amnesia have the reverse effect, causing disruption of memory for
experiences after the onset of amnesia; these are cases of anterograde amnesia.
(Many cases of amnesia involve both retrograde and anterograde memory loss.)
Disrupted Episodic Memory, but Spared
Semantic Memory
Studies of amnesia can teach us many things. For example, do we need all
the distinctions shown in Figure 7.11? Consider the case of Clive Wearing,
whom we met in the opening to Chapter 6. (You can find more detail
about Wearing’s case in an extraordinary book by his wife — see Wearing,
2011.) Wearing’s episodic memory is massively disrupted, but his memory
for generic information, as well as his deep love for his wife, seem to be
entirely intact. Other patients show the reverse pattern — disrupted semantic memory but preserved episodic knowledge. One patient, for example,
suffered damage (from encephalitis) to the front portion of her temporal
lobes. As a consequence, she lost her memory of many common words,
important historical events, famous people, and even the fundamental traits
of animate and inanimate objects. “However, when asked about her wedding
and honeymoon, her father’s illness and death, or other specific past episodes,
FIGURE 7.12
RETROGRADE AND ANTEROGRADE AMNESIA
Moment of
brain injury
Time
Period for which
retrograde amnesia
disrupts memory
Period for which
anterograde amnesia
disrupts memory
Retrograde amnesia disrupts memory for experiences before the injury,
accident, or disease that triggered the amnesia. Anterograde amnesia disrupts memory for experiences after the injury or disease. Some patients suffer
from both retrograde and anterograde amnesia.
Amnesia
•
267
she readily produced detailed and accurate recollections” (Schacter, 1996,
p. 152; also see Cabeza & Nyberg, 2000). (For more on amnesia, see
Brown, 2002; Clark & Maguire, 2016; Kopelman & Kapur, 2001; Nadel &
Moscovitch, 2001; Riccio, Millin, & Gisquet-Verrier, 2003.)
These cases (and other evidence too; see Figure 7.13) provide the double
dissociation that demands a distinction between episodic and semantic
memory. It’s observations like these that force us to the various distinctions
shown in Figure 7.11. (For evidence, though, that episodic and semantic
memory are intertwined in important ways, see McRae & Jones, 2012.)
Anterograde Amnesia
We’ve already mentioned the patient known as H.M. His memory loss was the
result of brain surgery in 1953, and over the next 55 years (until his death in
2008) H.M. participated in a vast number of studies. Some people suggest he
FIGURE 7.13
S EMANTIC MEMORY WITHOUT EPISODIC MEMORY
Kent Cochrane — known for years as “Patient K.C.” — died in 2014. In 1981, at age 30, he skidded off the road on his
motorcycle and suffered substantial brain damage. The damage caused severe disruption of Cochrane’s episodic
memory, but it left his semantic memory intact. As a result, he could still report on the events of his life, but these
reports were entirely devoid of autobiographical quality. In other words, he could remember the bare facts of,
say, what happened at his brother’s wedding, but the memory was totally impersonal, with no recall of context
or emotion. He also knew that during his childhood his family had fled their home because a train had derailed
nearby, spilling toxic chemicals. But, again, he simply knew this as factual material — the sort of information you
might pick up from a reference book — and he had no recall of his own experiences during the event.
268 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
was the most-studied individual in the entire history of psychology (which is one
of the reasons we’ve returned to his case several times). In fact, the data gathering
continued after H.M.’s death — with careful postmortem scrutiny of his brain.
(For a review of H.M.’s case, see Corkin, 2013; Milner, 1966, 1970; also O’Kane,
Kensinger, & Corkin, 2004; Skotko et al., 2004; Skotko, Rubin, & Tupler, 2008.)
After his surgery, H.M. was still able to recall events that took place
before the surgery — and so his amnesia was largely anterograde, not retrograde. But the amnesia was severe. Episodes he had experienced after the
surgery, people he had met, stories he had heard — all seemed to leave no
lasting record, as though nothing new could get into his long-term storage.
H.M. could hold a mostly normal conversation (because his working
memory was still intact), but his deficit became instantly clear if the conversation was interrupted. If you spoke with him for a while, then left the room
and came back 3 or 4 minutes later, he seemed to have totally forgotten that
the earlier conversation ever took place. If the earlier conversation was your
first meeting with H.M., he would, after the interruption, be certain he was
now meeting you for the very first time.
A similar amnesia has been found in patients who have been longtime alcoholics. The problem isn’t the alcohol itself; the problem instead is that alcoholics tend
to have inadequate diets, getting most of their nutrition from whatever they’re
drinking. It turns out, though, that most alcoholic beverages are missing several
key nutrients, including vitamin B1 (thiamine). As a result, longtime alcoholics
are vulnerable to problems caused by thiamine deficiency, including the disorder
known as Korsakoff’s syndrome (Rao, Larkin, & Derr, 1986; Ritchie, 1985).
Patients suffering from Korsakoff’s syndrome seem similar to H.M. in many
ways. They typically have no problem remembering events that took place
before the onset of alcoholism. They can also maintain current topics in mind
as long as there’s no interruption. New information, though, if displaced from
the mind, seems to be lost forever. Korsakoff’s patients who have been in the
hospital for decades will casually mention that they arrived only a week ago; if
Hippocampus
missing
Hippocampus
intact
A Anterior
B Posterior
H.M.’S BRAIN
For many years, researchers thought that surgery had left H.M. with no hippocampus
at all. These MRI scans of his brain show that the surgery did destroy the anterior
portion of the hippocampus (the portion closer to the front of the head; Panel A) but
not the posterior portion (closer to the rear of the head; Panel B).
Amnesia
•
269
asked the name of the current president or events in the news, they unhesitatingly give answers appropriate for two or three decades earlier, whenever the
disorder began (Marslen-Wilson & Teuber, 1975; Seltzer & Benson, 1974).
Anterograde Amnesia: What Kind of
Memory Is Disrupted?
At the chapter’s beginning, we alluded to other evidence that complicates this
portrait of anterograde amnesia, and it’s evidence that brings us back to the distinction between implicit and explicit memory. As it turns out, some of this evidence has been available for a long time. In 1911, the Swiss psychologist Édouard
Claparède (1911/1951) reported the following incident. He was introduced to a
young woman suffering from Korsakoff’s amnesia, and he reached out to shake
her hand. However, Claparède had secretly positioned a pin in his own hand so
that when they clasped hands the patient received a painful pinprick. (Modern
investigators would regard this experiment as a cruel violation of a patient’s
rights, but ethical standards were much, much lower in 1911.) The next day,
Claparède returned and reached out to shake hands with the patient. Not surprisingly, she gave no indication that she recognized Claparède or remembered
anything about the prior encounter. (This confirms the diagnosis of amnesia.)
But just before their hands touched, the patient abruptly pulled back and refused
to shake hands with Claparède. He asked her why, and after some confusion the
patient said vaguely, “Sometimes pins are hidden in people’s hands.”
What was going on here? On the one side, this patient seemed to have
no memory of the prior encounter with Claparède. She certainly didn’t mention it in explaining her refusal to shake hands, and when questioned closely
about the earlier encounter, she showed no knowledge of it. But, on the other
side, she obviously remembered something about the painful pinprick she’d
gotten the previous day. We see this clearly in her behavior.
A related pattern occurs with other Korsakoff’s patients. In one of the
early demonstrations of this point, researchers used a deck of cards like those
used in popular trivia games. Each card contained a question and some possible answers, in a multiple-choice format (Schacter, Tulving, & Wang, 1981).
The experimenter showed each card to a Korsakoff’s patient, and if the
patient didn’t know the answer, he was told it. Then, outside of the patient’s
view, the card was replaced in the deck, guaranteeing that the same question
would come up again in a few minutes.
When the question did come up again, the patients in this study were likely
to get it right — and so apparently had learned the answer in the previous
encounter. Consistent with their diagnosis, though, the patients had no recollection of the learning: They were unable to explain why their answers were
correct. They didn’t say, “I know this bit of trivia because the same question
came up just five minutes ago.” Instead, patients were likely to say things like
“I read about it somewhere” or “My sister once told me about it.”
Many studies show similar results. In setting after setting, Korsakoff’s
patients are unable to recall episodes they’ve experienced; they seem to have
270 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
no explicit memory. But if they’re tested indirectly, we see clear indications
of memory — and so these patients seem to have intact implicit memories.
(See, e.g., Cohen & Squire, 1980; Graf & Schacter, 1985; Moscovitch, 1982;
Schacter, 1996; Schacter & Tulving, 1982; Squire & McKee, 1993.) In fact, in
many tests of implicit memory, amnesic patients seem indistinguishable from
ordinary individuals.
Can There Be Explicit Memory without Implicit?
We can also find patients with the reverse pattern — intact explicit memory,
but impaired implicit. One study compared a patient who had suffered brain
damage to the hippocampus but not the amygdala with a second patient who
had the opposite pattern: damage to the amygdala but not the hippocampus
(Bechara et al., 1995). These patients were exposed to a series of trials in
which a particular stimulus (a blue light) was reliably followed by a loud
boat horn, while other stimuli (green, yellow, or red lights) were not followed
by the horn. Later on, the patients were exposed to the blue light on its own,
and their bodily arousal was measured; would they show a fright reaction
in response to this stimulus? In addition, the patients were asked directly,
“Which color was followed by the horn?”
The patient with damage to the hippocampus did show a fear reaction to
the blue light — assessed via the skin conductance response (SCR), a measure
of bodily arousal. As a result, his data on this measure look just like results
for control participants (i.e., people without brain damage; see Figure 7.14).
A
3
DAMAGE TO HIPPOCAMPUS AND AMYGDALA
Fear response
2
1
0
Explicit memory
4
Total score
SCR magnitude
(µS)
FIGURE 7.14
Controls SM046 WC1606
3
2
1
0
B
Controls SM046 WC1606
Panel A shows results for a test probing implicit memory via a fear response;
Panel B shows results for a test probing explicit memory. Patient SM046 had
suffered damage to the amygdala and shows little evidence of implicit memory
(i.e., no fear response — indexed by the skin conductance response, or SCR)
but a normal level of explicit memory. Patient WC1606 had suffered damage
to the hippocampus and shows the opposite pattern: massively disrupted
explicit memory but a normal fear response.
( after bechara et al ., 1995)
Amnesia
•
271
However, when asked directly, this patient couldn’t recall which of the lights
had been associated with the boat horn.
In contrast, the patient with damage to the amygdala showed the opposite
pattern. She was able to report that just one of the lights had been associated
with the horn and that the light’s color had been blue — demonstrating fully
intact explicit memory. When presented with the blue light, however, she
showed no fear response.
Optimal Learning
Before closing this chapter, let’s put these amnesia findings into the broader
context of the chapter’s main themes. Throughout the chapter, we’ve suggested that we cannot make claims about learning or memory acquisition
without some reference to how the learning will be used later on. For example, whether it’s better to learn underwater or on land depends on where you
will be tested. Whether it’s better to learn while listening to jazz or while
sitting in a quiet room depends on the acoustic background of the memory
test environment.
These ideas are echoed in the neuropsychology data. Specifically, it would
be misleading to say that brain damage (whether from Korsakoff’s syndrome or some other source) ruins someone’s ability to create new memories.
Instead, brain damage is likely to disrupt some types of learning but not
others, and how this matters for the person depends on how the newly
learned material will be accessed. Thus, someone who suffers hippocampal
damage will probably appear normal on an indirect memory test but seem
amnesic on a direct test, while someone who suffers amygdala damage will
probably show the reverse pattern.
All these points are enormously important for our theorizing about memory, but they also have a practical implication. Right now, you are reading this material and presumably want to remember it later on. You’re also
encountering new material in other settings (perhaps in other classes you’re
taking), and surely you want to remember that as well. How should you
study all of this information if you want the best chances of retaining it for
later use?
At one level, the message from this chapter might be that the ideal form
of learning would be one that’s “in tune with” the approach to the material
that you’ll need later. If you’re going to be tested explicitly, you want to learn
the material in a way that prepares you for that form of retrieval. If you’ll
be tested underwater or while listening to music, then, again, you want to
learn the material in a way that prepares you for that context and the mental
perspective it produces. If you’ll need source memory, then you want one
type of preparation; if you’ll need familiarity, you might want a different type
of preparation.
The problem, though, is that during learning, you often don’t know how
you’ll be approaching the material later — what the retrieval environment
272 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
will be, whether you’ll need the information implicitly or explicitly, and so on.
As a result, maybe the best strategy in learning would be to use multiple
perspectives. To revisit our earlier analogy, imagine that you know at some
point in the future you’ll want to reach Chicago, but you don’t know yet
whether you’ll be approaching the city from the north, the south, or the west.
In that case, your best bet might be to build multiple highways, so that you
can reach your goal from any direction. Memory works the same way. If you
initially think about a topic in different ways and in relation to many other
ideas, then you’ll establish many paths leading to the target material — and
so you’ll be able to access that material from many different perspectives.
The practical message from this chapter, then, is that this multiperspective
approach may provide the optimal learning strategy.
TEST YOURSELF
12. Define “retrograde”
and “anterograde”
amnesia.
13. What type(s) of
memory are disrupted
in patients suffering
from Korsakoff’s
syndrome?
COGNITIVE PSYCHOLOGY AND EDUCATION
familiarity can be treacherous
Sometimes you see a picture of someone and immediately say, “Gee — she
looks familiar!” This seems like a simple and direct reaction to the picture,
but the chapter describes how complicated familiarity really is. Indeed, the
chapter makes it clear that we can’t think of familiarity just as a “feeling”
somehow triggered by a stimulus. Instead, familiarity seems more like a conclusion that you draw at the end of a many-step process. As a result of these
complexities, errors about familiarity are possible: cases in which a stimulus
feels familiar even though it’s not, or cases in which you correctly realize that
the stimulus is familiar but then make a mistake about why it’s familiar.
These points highlight the dangers, for students, of relying on familiarity. As one illustration, consider the advice that people sometimes give for
taking a multiple-choice test. They tell you, “Go with your first inclination”
or “Choose the answer that feels familiar.” In some cases these strategies will
help, because sometimes the correct answer will indeed feel familiar. But in
other cases these strategies can lead you astray, because the answer you’re
considering may seem familiar for a bad reason. What if your professor once
said, “One of the common mistakes people make is to believe . . .” and then
talked about the claim summarized in the answer you’re now considering?
Alternatively, what if the answer seems familiar because it resembles the correct answer but is, in some crucial way, different from the correct answer
(and therefore mistaken)? In either of these cases, your sense of familiarity
might lead you to a wrong answer.
Even worse, one study familiarized people with phrases like “the record
for tallest pine tree.” Because of this exposure, these people were later more
likely to accept as true a longer phrase, such as “the record for tallest pine tree
is 350 feet.” Why? Because they realized that (at least) part of the sentence
was familiar and therefore drew the reasonable inference that they must have
FAMILIARITY CAN
BE TREACHEROUS
“Option C rings a bell. . . .”
Often, in taking a multiplechoice test, students will rea­lize
they don’t know the answer
to a question but, even so,
one of the answer options
seems somehow familiar. For
reasons described in the
chapter, though, this sense
of familiarity is an unreliable
guide in choosing a response.
Cognitive Psychology and Education
•
273
encountered the entire sentence at some previous point. The danger here
should be obvious: On a multiple-choice test, part of an incorrect option
may be an exact duplicate of some phrase in your reading; if so, relying on
familiarity will get you into trouble! (And, by the way, this claim about pines
is false; the tallest pine tree — a sugar pine — is only about 273 feet tall.)
As a different concern, think back to the end-of-chapter essay for Chapter 6.
There, we noted that one of the most common study strategies used by students is to read and reread their notes, or read and reread the textbook. This
strategy turns out not to help memory very much, and other strategies are
demonstrably better. But, in addition, the rereading strategy can actually hurt
you. Thanks to the rereading, you become more and more familiar with the
materials, which makes it easy to interpret this familiarity as mastery. But
this is a mistake, and because of the mistake, familiarity can sometimes lead
students to think they’ve mastered material when they haven’t, causing them
to end their study efforts too soon.
What can you do to avoid all these dangers? You’ll do a much better job of
assessing your own mastery if, rather than relying on familiarity, you give yourself some sort of quiz (perhaps one you find in the textbook, or one that a friend
creates for you). More broadly, it’s valuable to be alert to the various complexities associated with familiarity. After all, you don’t want to ignore familiarity,
because sometimes it’s all you’ve got. If you really don’t know the answer to a
multiple-choice question but option B seems somehow familiar, then choosing B
may be your only path forward. But given the difficulties we’ve mentioned here,
it may be best to regard familiarity just as a weak clue about the past and not
as a guaranteed indicator. That attitude may encourage the sort of caution that
will allow you to use familiarity without being betrayed by it.
For more on this topic . . .
Brown, P. C., Roediger, H. L., & McDaniel, M. A. (2014). Make it stick: The science
of successful learning. New York, NY: Belknap Press.
Jacoby, L., & Hollingshead, A. (1990). Reading student essays may be hazardous
to your spelling: Effects of reading incorrectly and correctly spelled words.
Canadian Journal of Psychology, 44, 345–358.
Preston, J., & Wegner, D. M. (2007). The Eureka error: Inadvertent plagiarism
by misattributions of effort. Journal of Personality and Social Psychology,
92, 575–584.
Stark, L.-J., Perfect, T., & Newstead, S. (2008). The effects of repeated idea
elaboration on unconscious plagiarism. Memory & Cognition, 36, 65–73.
Swire, B., Ecker, U. K. H., & Lewandowsky, S. (2017). The role of familiarity in
correcting inaccurate information. Journal of Experimental Psychology:
Learning, Memory & Cognition, 43, 1948–1961.
274 •
C H A P T E R S E V E N Interconnections between Acquisition and Retrieval
chapter review
SUMMARY
• In general, the chances that someone will remember
• Activating one node does seem to prime nearby
an earlier event are greatest if the physical and mental circumstances in place during memory retrieval
match those in place during learning. This is reflected
in the phenomenon of context-dependent learning.
nodes through the process of spreading activation.
This is evident in studies of semantic priming in
lexical-decision tasks.
• A similar pattern is reflected in the phenomenon
tion for some sorts of memory tests but ineffective
for other sorts of tests. Some strategies, for example,
are effective at establishing source memory rather
than familiarity; other strategies do the reverse.
of “encoding specificity.” This term refers to the
idea that people usually learn more than the specific
material to be remembered itself; they also learn
that material within its associated context.
• All these results arise from the fact that learning
establishes connections among memories, and these
connections serve as retrieval paths. Like any path,
these lead from some starting point to some target.
To use a given path, therefore, you must return to
the appropriate starting point. In the same way, if
there is a connection between two memories, then
activating the first memory is likely to call the second to mind. But if the first memory isn’t activated,
this connection, no matter how well established,
will not help in locating the second memory — just
as a large highway approaching Chicago from
the south won’t be helpful if you’re trying to reach
Chicago from the north.
• This emphasis on memory connections fits well
with a conceptualization of memory as a vast network, with individual nodes joined to one another
via connections or associations. An individual node
becomes activated when it receives enough of an
input signal to raise its activation level to its
response threshold. Once activated, the node sends
activation out through its connections to all the
nodes connected to it.
• Hints are effective because the target node can
receive activation from two sources simultaneously —
from nodes representing the main cue or question,
and also from nodes representing the hint.
• Some learning strategies are effective as prepara-
• Different forms of learning also play a role in
producing implicit and explicit memories. Implicit
memories are those that influence you even when
you have no awareness that you’re being influenced
by a previous event. In many cases, implicit-memory
effects take the form of priming — for example, in
a lexical decision task or word-stem completion. But
implicit memories can also influence you in other
ways, producing a number of memory-based illusions.
• Implicit memory can be understood as the consequence of increased processing fluency, produced
by experience in a particular task with a particular stimulus. The fluency is sometimes detected and
registered as a sense of “specialness” attached to
a stimulus. Often, this specialness is attributed to
some cause, but this attribution can be inaccurate.
• Implicit memory is also important in understanding the pattern of symptoms in anterograde amnesia.
Amnesic patients perform badly on tests requiring
explicit memory and may not even recall events that
happened just minutes earlier. However, they often
perform at near-normal levels on tests involving
implicit memory. This disparity underscores the fact
that we cannot speak in general about good and bad
memories, good and poor learning. Instead, learning
and memory must be matched to a particular task and
a particular form of test; learning and memory that
are excellent for some tasks may be poor for others.
275
KEY TERMS
context-dependent learning (p. 242)
context reinstatement (p. 244)
encoding specificity (p. 244)
nodes (p. 246)
associations (or associative links) (p. 246)
activation level (p. 246)
response threshold (p. 246)
subthreshold activation (p. 246)
summation (p. 246)
spreading activation (p. 247)
lexical-decision task (p. 248)
semantic priming (p. 249)
recall (p. 251)
recognition (p. 251)
source memory (p. 251)
familiarity (p. 251)
attribution (p. 251)
“remember/know” distinction (p. 252)
word-stem completion (p. 255)
explicit memory (p. 255)
direct memory testing (p. 255)
implicit memory (p. 255)
indirect memory testing (p. 256)
illusion of truth (p. 258)
source confusion (p. 259)
processing pathway (p. 261)
processing fluency (p. 262)
amnesia (p. 267)
retrograde amnesia (p. 267)
anterograde amnesia (p. 267)
Korsakoff’s syndrome (p. 269)
TEST YOURSELF AGAIN
1.What does context-dependent learning tell us
about the nature of retrieval paths?
2.In what ways is a retrieval path like an
“ordinary” path (e.g., a path or highway
leading to a particular city)?
8.What is the difference between implicit and
explicit memory? Which of these is said to
be “memory without awareness”?
9.What is the role of implicit memory in
explaining the false fame effect?
3.What is encoding specificity? How is it
demonstrated?
10.What is processing fluency, and how does it
influence us?
4.What is subthreshold activation of a memory
node? What role does subthreshold activation
play in the explanation of why retrieval hints are
often helpful?
11.In what sense is familiarity more like a
conclusion that you draw, rather than a
feeling triggered by a stimulus?
5.How does semantic priming illustrate the
effectiveness of subthreshold activation?
6.Define “recognition” and “recall.”
7.What evidence indicates that source memory
and familiarity are distinct from each other?
276
12.Define “retrograde” and “anterograde”
amnesia.
13.What type(s) of memory are disrupted
in patients suffering from Korsakoff’s
syndrome?
THINK ABOUT IT
1.Some people describe the eerie sensation of
“déjà vu” — a feeling in which a place or face
seems familiar, even though they’re quite
certain they’ve never been in this place, or
seen this face, before. Can you generate a
hypothesis about the roots of déjà vu,
drawing on the material in the chapter?
E eBOOK DEMONSTRATIONS & ESSAYS
Go to http://digital.wwnorton.com/cognition7 for the online demonstrations and essays
relevant to this chapter. These can be found in the ebook.
Online Demonstrations
Online Applying Cognitive Psychology and the
Law Essays
• Demonstration 7.1: Retrieval Paths and
• Cognitive Psychology and the Law: Unconscious
Connections
Transference
• Demonstration 7.2: Encoding Specificity
• Demonstration 7.3: Spreading Activation in
Memory Search
• Demonstration 7.4: Semantic Priming
• Demonstration 7.5: Priming From Implicit
Memory
COGNITION LABS
Go to http://digital.wwnorton.com/cognition7 to experience interactive online psychology
labs relevant to this chapter.
The performers who appear on p. 253 are (left to right). Michael K. Williams, Margo Martindale, Adam Rodriguez, Judy Greer, and
Adina Porter.
277
8
chapter
Remembering
Complex Events
what if…
What were you doing on March 23, 2009? What did
you have for lunch that day? What was the weather
like? These seem like odd questions; why should you remember these
details from almost a decade ago? But imagine if you could remember
those details — and similar details for every other day in your life. In other
words, what would life be like if you had a “super memory” — so that,
essentially, you never forgot anything?
Researchers have identified a small number of people who have
hyperthymesia, also called “highly superior autobiographical recall”
(HSAM). These people seem able to remember every single day of their
lives, over a span of many years. One of these individuals claims that she
can recall every day of her life over the last four decades. “Starting on
February 5, 1980, I remember everything. That was a Tuesday.”
If asked about a randomly selected date — say, February 10,
2012 — these individuals can recall exactly where they were that day,
what time they woke up, and what shoes they wore. Their performance
is just as good if they’re asked about October 8, 2008, or December 19,
2007. And when checked, their memories turn out to be uniformly
accurate.
These individuals are remarkable in how much they remember, but
they seem quite normal in other ways. For example, their extraordinary
memory capacity hasn’t made them amazing geniuses or incredible
scholars. Even though they have an exceptional capacity for remembering their own lives, they have no advantage in remembering other
sorts of content or performing other mental tasks. This point has been
documented with careful testing, but it is also evident in the fact that
researchers weren’t aware such people existed until just a few years
ago (e.g., Parker, Cahill, & McGaugh, 2006). Apparently, these individuals, even with their incredible memory capacity, are ordinary enough
in other ways so that we didn’t spot them until recently (McGaugh &
LePort, 2014.)
Humans have been trying for centuries to improve their memories,
but it seems that a “perfect” memory may provide less of an advantage
than you might think. We’ll return to this point, and what it implies about
memory functioning, later in the chapter.
279
preview of chapter themes
•
utside the lab, you often try to remember materials that
O
are related in some way to other things you know or have
experienced. Over and over, we will see that this other
knowledge — the knowledge you bring to a situation —
helps you to remember by promoting retrieval, but it can
also promote error.
•
he memory errors produced by prior knowledge tend to be
T
quite systematic: You often recall the past as more “normal,”
more in line with your expectations, than it actually was.
•
ven acknowledging the memory errors, our overall
E
assessment of memory can be quite positive. This is
because memories are accurate most of the time, and
the errors that do occur can be understood as the byproducts of mechanisms that generally serve you well.
•
inally, we will consider three factors that play an imporF
tant role in shaping memory outside of the laboratory:
involvement with an event, emotion, and the passage of
time. These factors require some additional principles as
part of our overall theory, but they also confirm the power
of more general principles — principles hinging, for example,
on the role of memory connections.
Memory Errors, Memory Gaps
Where did you spend last summer? What country did you grow up in? Where
were you five minutes ago? These are easy questions, and you effortlessly
retrieve this information from memory the moment you need it. If we want
to understand how memory functions, therefore, we need to understand how
you locate these bits of information (and thousands of others just like them)
so readily.
But we also need to account for some other observations. Sometimes,
when you try to remember an episode, you draw a blank. On other
occasions, you recall something, but with no certainty that you’re correct:
“I think her nickname was Dink, but I’m not sure.” And sometimes,
when you do recall a past episode, it turns out that your memory is mistaken. Perhaps a few details of the event were different from the way you
recall them. Or perhaps your memory is completely wrong, misrepresenting large elements of the original episode. Worse, in some cases you can
remember entire events that never happened at all! In this chapter, we’ll
consider how, and how often, these errors arise. Let’s start with some
examples.
Memory Errors: Some Initial Examples
In 1992, an El Al cargo plane lost power in two of its engines just after
taking off from Amsterdam’s Schiphol Airport. The pilot attempted to return
the plane to the airport but couldn’t make it. A few minutes later, the plane
crashed into an 11-story apartment building in Amsterdam’s Bijlmermeer
neighborhood. The building collapsed and burst into flames; 43 people were
killed, including the plane’s entire crew.
280 •
C H A P T E R E I G H T Remembering Complex Events
Ten months later, researchers questioned 193 Dutch people about
the crash, asking them in particular, “Did you see the television film of
the moment the plane hit the apartment building?” More than half of the
participants (107 of them) reported seeing the film, even though there
was no such film. No camera had recorded the crash; no film (or any
reenactment) was shown on television. The participants seemed to be
remembering something that never took place (Crombag, Wagenaar, &
van Koppen, 1996).
In a follow-up study, investigators surveyed another 93 people about
the plane crash. These people were also asked whether they’d seen the (nonexistent) TV film, and then they were asked detailed questions about exactly
what they had seen in the film: Was the plane burning when it crashed, or did
it catch fire a moment later? In the film, did they see the plane come down
vertically with no forward speed, or did it hit the building while still moving
horizontally at a considerable speed?
Two thirds of these participants reported seeing the film, and most of them
were able to provide details about what they had seen. When asked about the
plane’s speed, for example, only 23% said that they couldn’t remember.
The others gave various responses, presumably based on their “memory” of
the (nonexistent) film.
Other studies have produced similar results. There was no video footage of
the car crash in which Princess Diana was killed, but 44% of the British participants in one study recalled seeing the footage (Ost, Vrij, Costall, & Bull,
2002). More than a third of the participants questioned about a nightclub
bombing in Bali recalled seeing a (nonexistent) video, and nearly all these
participants reported details about what they’d seen in the video (Wilson &
French, 2006).
It turns out that more persistent questioning can lead some of these people
to admit they actually don’t remember seeing the video. Even with persistent
questioning, though, many participants continue to insist that they did see
the video — and they offer additional information about exactly what they
saw in the film (e.g., Patihis & Loftus, 2015; Smeets et al., 2006). Also, in all
these studies, let’s emphasize that participants are thinking back to an emotional and much-discussed event; the researchers aren’t asking them to recall
a minor occurrence.
Is memory more accurate when the questions come after a shorter delay? In
a study by Brewer and Treyens (1981), participants were asked to wait briefly
in the experimenter’s office prior to the procedure’s start. After 35 seconds,
participants were taken out of this office and told that there actually was no
experimental procedure. Instead, the study was concerned with their memory
for the room in which they’d just been sitting.
Participants’ descriptions of the office were powerfully influenced by their
prior beliefs. Surely, most participants would expect an academic office to
contain shelves filled with books. In this particular office, though, no books
TEST YOURSELF
1. W
hat is the evidence
that in some circumstances many people
will misremember
significant events they
have experienced?
2. W
hat is the evidence
that in some circumstances people will
even misremember
recent events?
Memory Errors, Memory Gaps
•
281
FIGURE 8.1 THE
OFFICE USED IN
THE BREWER AND
TREYENS STUDY
No books were in view in this
office, but many participants,
biased by their expectations
of what should be in an
academic office, remembered
seeing books.
(after brewer & treyens, 1981)
were in view (see Figure 8.1). Even so, almost one third of the participants
(9 of 30) reported seeing books in the office. Their recall, in other words, was
governed by their expectations, not by reality.
How could this happen? How could so many Dutch participants be wrong
in their recall of the plane crash? How could intelligent, alert college students
fail to remember what they’d seen in an office just moments earlier?
Memory Errors: A Hypothesis
In Chapters 6 and 7, we emphasized the importance of memory connections
that link each bit of knowledge in your memory to other bits. Sometimes
these connections tie together similar episodes, so that a trip to the beach
ends up connected in memory to your recollection of other trips. Sometimes
the connections tie an episode to certain ideas — ideas, perhaps, that were
part of your understanding of the episode, or ideas that were triggered by
some element within the episode.
It’s not just separate episodes and ideas that are linked in this way. Even
for a single episode, the elements of the episode are stored separately from
one another and are linked by connections. In fact, the storage is “modalityspecific,” with the bits representing what you saw stored in brain areas
devoted to visual processing, the bits representing what you heard stored
in brain areas specialized for auditory processing, and so on (e.g., Nyberg,
Habib, McIntosh, & Tulving, 2000; Wheeler, Peterson, & Buckner, 2000;
also see Chapter 7, Figure 7.4, p. 245).
282 •
C H A P T E R E I G H T Remembering Complex Events
With all these connections in place — element to element, episode to episode, episode to related ideas — information ends up stored in memory in a
system that resembles a vast spider web, with each bit of information connected by many threads to other bits elsewhere in the web. This was the idea
that in Chapter 7 we described as a huge network of interconnected nodes.
However, within this network there are no boundaries keeping the elements of one episode separate from elements of other episodes. The episodes,
in other words, aren’t stored in separate “files,” each distinct from the others. What is it, therefore, that holds together the various bits within each
episode? To a large extent, it’s simply the density of connections. There are
many connections linking the various aspects of your “trip to the beach” to
one another; there are fewer connections linking this event to other events.
As we’ve discussed, these connections play a crucial role in memory retrieval. Imagine that you’re trying to recall the restaurant you ate at during
your beach trip. You’ll start by activating nodes in memory that represent
some aspect of the trip — perhaps your memory of the rainy weather. Activation will then flow outward from there, through the connections you’ve
established, and this will energize nodes representing other aspects of the trip.
The flow of activation can then continue from there, eventually reaching the
nodes you seek. In this way, the connections serve as retrieval paths, guiding
your search through memory.
Obviously, then, memory connections are a good thing; without them,
you might never locate the information you’re seeking. But the connections
can also create problems. As you add more and more links between the bits
of this episode and the bits of that episode, you’re gradually knitting these
two episodes together. As a result, you may lose track of the “boundary” between the episodes. More precisely, you’re likely to lose track of which bits
of information were contained within which event. In this way, you become
vulnerable to what we might think of as “transplant” errors, in which a bit of
information encountered in one context is transplanted into another context.
In the same way, as your memory for an episode becomes more and more
interwoven with other thoughts you’ve had about the event, it will become
difficult to keep track of which elements are linked to the episode because
they were actually part of the episode itself, and which are linked merely because they were associated with the episode in your thoughts. This, too, can
produce transplant errors, in which elements that were part of your thinking
get misremembered as if they were actually part of the original experience.
Understanding Both Helps and Hurts Memory
It seems, then, that memory connections both help and hurt recollection.
They help because the connections, serving as retrieval paths, enable you to
locate information in memory. But connections can hurt because they sometimes make it difficult to see where the remembered episode stops and other,
related knowledge begins. As a result, the connections encourage intrusion
errors — errors in which other knowledge intrudes into the remembered event.
Memory Errors: A Hypothesis
•
283
To see how these points play out, consider an early study by Owens,
Bower, and Black (1979). In this study, half of the participants read the
following passage:
Nancy arrived at the cocktail party. She looked around the room to see
who was there. She went to talk with her professor. She felt she had
to talk to him but was a little nervous about just what to say. A group
of people started to play charades. Nancy went over and had some
refreshments. The hors d’oeuvres were good, but she wasn’t interested
in talking to the rest of the people at the party. After a while she
decided she’d had enough and left the party.
Other participants read the same passage, but with a prologue that set
the stage:
Nancy woke up feeling sick again, and she wondered if she really was
pregnant. How would she tell the professor she had been seeing? And
the money was another problem.
All participants were then given a recall test in which they were asked
to remember the sentences as exactly as they could. Table 8.1 shows the
results — the participants who had read the prologue (the Theme condition)
recalled much more of the original story (i.e., they remembered the propositions actually contained within the story). This is what we should expect,
based on the claims made in Chapter 6: The prologue provided a meaningful
context for the remainder of the story, and this helped understanding. Understanding, in turn, promoted recall.
At the same time, the story’s prologue also led participants to include elements in their recall that weren’t mentioned in the original episode. In fact,
participants who had seen the prologue made four times as many intrusion
errors as did participants who hadn’t seen the prologue. For example, they
might include in their recall something like “The professor had gotten Nancy
pregnant.” This idea isn’t part of the story but is certainly implied, so will
probably be part of participants’ understanding of the story. It’s then this
understanding (including the imported element) that is remembered.
TABLE 8.1
UMBER OF PROPOSITIONS REMEMBERED
N
BY PARTICIPANTS
STUDIED PROPOSITIONS
(THOSE IN STORY)
Theme Condition
Neutral Condition
29.2
20.2
INFERRED PROPOSITIONS
(THOSE NOT IN STORY)
Theme Condition
15.2
Neutral Condition
3.7
In the Theme condition, a brief prologue set the theme for the passage that was to
(after owens et al., 1979)
be remembered.
284 •
C H A P T E R E I G H T Remembering Complex Events
The DRM Procedure
Similar effects, with memory connections both helping and hurting memory,
can be demonstrated with simple word lists. For example, in many experiments, participants have been presented with lists like this one: “bed, rest,
awake, tired, dream, wake, snooze, blanket, doze, slumber, snore, nap, peace,
yawn, drowsy.” Immediately after hearing this list, participants are asked to
recall as many of the words as they can.
As you surely noticed, the words in this list are all associated with sleep,
and the presence of this theme helps memory: The list words are easy to remember. It turns out, though, that the word “sleep” is not itself included in the
list. Nonetheless, research participants spontaneously make the connection
between the list words and this associated word, and this connection almost
always leads to a memory error. When the time comes for recall, participants
are extremely likely to recall that they heard “sleep.” In fact, they’re just as
likely to recall “sleep” as they are to recall the actual words on the list (see
Figure 8.2). When asked how confident they are in their memories, participants are just as confident in their (false) recall of “sleep” as they are in their
(correct) memory of genuine list words (Gallo, 2010; for earlier and classic
papers in this arena, see Deese, 1957; Roediger & McDermott, 1995, 2000).
This experiment (and many others like it) uses the DRM procedure, a
bit of terminology that honors the investigators who developed it (James
Deese, Henry Roediger III, and Kathleen McDermott). The procedure yields
many errors even if participants are put on their guard before the procedure
begins — that is, told about the nature of the lists and the frequency with
which they produce errors (Gallo, Roberts, & Seamon, 1997; McDermott &
100
90
Percent of recall
80
70
60
50
40
FIGURE 8.2 THE EFFECTS OF THE
DRM PARADIGM
30
20
10
0
Words from
actual list
Unrelated
words
Mistaken “recall”
of theme words
Because of the theme uniting the list, participants can remember almost 90% of the words
they encountered. However, they’re just as likely
to “recall” the list’s theme word — even though it
was not presented.
Memory Errors: A Hypothesis
•
285
Roediger, 1998). Apparently, the mechanisms leading to these errors are so
automatic that people can’t inhibit them.
Schematic Knowledge
Imagine that you go to a restaurant with a friend. This setting is familiar for
you, and you have some commonsense knowledge about what normally happens here. You’ll be seated; someone will bring menus; you’ll order, then eat;
eventually, you’ll pay and leave. Knowledge like this is often referred to with
the Greek word schema (plural: schemata). Schemata summarize the broad
pattern of what’s normal in a situation — and so your kitchen schema tells
you that a kitchen is likely to have a stove but no piano; your dentist’s office
schema tells you that there are likely to be magazines in the waiting room,
that you’ll probably get a new toothbrush when you leave, and so on.
Schemata help you in many ways. In a restaurant, for example, you’re not
puzzled when someone keeps filling your water glass or when someone else
drops by to ask, “How is everything?” Your schema tells you that these are
normal occurrences in a restaurant, and you instantly understand how they
fit into the broader framework.
Schemata also help when the time comes to recall how an event unfolded.
This is because there are often gaps in your recollection — either because you
didn’t notice certain things in the first place, or because you’ve gradually
forgotten some aspects of the experience. (We’ll say more about forgetting
later in the chapter.) In either case, you can rely on your schemata to fill in
these gaps. So, in thinking back to your dinner at Chez Pierre, you might not
remember anything about the menus. Nonetheless, you can be reasonably
sure that there were menus and that they were given to you early on and
taken away after you placed your order. On this basis, you’re likely to include
menus within your “recall” of the dinner, even if you have no memory of
seeing the menus for this particular meal. In other words, you’ll supplement
what you actually remember with a plausible reconstruction based on your
schematic knowledge. And in most cases this after-the-fact reconstruction
will be correct, since schemata do, after all, describe what happens most of
the time.
Evidence for Schematic Knowledge
Clearly, then, schematic knowledge helps you, by guiding your understanding
and enabling you to reconstruct things you can’t remember. But schematic
knowledge can sometimes hurt you, by promoting errors in perception and
memory. Moreover, the types of errors produced by schemata are quite predictable. As an example, imagine that you visit a dentist’s office, and this
one happens not to have any magazines in the waiting room. It’s likely that
you’ll forget this detail after a while, so what will happen when you later
try to recall your trip to the dentist? Odds are good that you’ll rely on schematic knowledge and “remember” that there were magazines (since, after all,
there usually are some scattered around a waiting room). In this way, your
286 •
C H A P T E R E I G H T Remembering Complex Events
recollection will make this dentist’s office seem more typical, more ordinary,
than it actually was.
Here’s the same point in more general terms. We’ve already said that schemata tell you what’s typical in a setting. Therefore, if you rely on schematic
knowledge to fill gaps in your recollection, you’ll fill those gaps with what’s
normally in place in that sort of situation. As a result, any reliance on schemata will make the world seem more “normal” than it really is and will make
the past seem more “regular” than it actually was.
This tendency toward “regularizing” the past has been documented in
many settings. The classic demonstration, however, comes from studies published long ago by British psychologist Frederick Bartlett. Bartlett presented
his participants with a story taken from the folklore of Native Americans
(Bartlett, 1932). When tested later, the participants did reasonably well in
recalling the gist of the story, but they made many errors in recalling the
particulars. The pattern of errors, though, was quite systematic: The details
omitted tended to be ones that made little sense to Bartlett’s British participants. Likewise, aspects of the story that were unfamiliar were often changed
into aspects that were more familiar; steps of the story that seemed inexplicable were supplemented to make the story seem more logical.
Overall, then, the participants’ memories seem to have “cleaned up” the
story they had read — making it more coherent (from their perspective), more
sensible. This is exactly what we would expect if the memory errors derived
from the participants’ attempts to understand the story and, with that,
their efforts toward fitting the story into a schematic frame. Elements that
fit within the frame remained in their memories (or could be reconstructed
later). Elements that didn’t fit dropped out of memory or were changed.
In the same spirit, consider the Brewer and Treyens study mentioned at
the start of this chapter — the study in which participants remembered seeing
shelves full of books, even though there were none. This error was produced
by schematic knowledge. During the event itself (while the participants were
sitting in the office), schematic knowledge told the participants that academic
offices usually contain many books, and this knowledge biased what the participants paid attention to. (If you’re already certain that the shelves contain
books, why should you spend time looking at the shelves? This would only
confirm something you already know — see Vo & Henderson, 2009.) Then,
when the time came to recall the office, participants used their schema to reconstruct what the office must have contained — a desk, a chair, and of course
lots of books. In this way, the memory for the actual office was eclipsed by
generic knowledge about what a “normal” academic office contains.
Likewise, think back to the misremembered plane crash and the related
studies of people remembering videos of other prominent events, even though
there were no videos of these events. Here, too, the memory errors distort
reality by making the past seem more regular, more typical, than it really
was. After all, people often hear about major news events via a television
broadcast or Internet coverage, and these reports usually include vivid video
footage. So here, too, the past as remembered seems to have been assimilated
TEST YOURSELF
3. W
hat is the evidence
that your understanding of an episode can
produce intrusion
errors?
4. What is the DRM
procedure, and what
results does this
procedure produce?
5. What is schematic
knowledge, and what
evidence tells us that
schematic knowledge
can help us — and
also can undermine
memory accuracy?
Memory Errors: A Hypothesis
•
287
into the pattern of the ordinary. The event as it unfolded was unusual, but the
event as remembered becomes typical of its kind — just as we would expect if
understanding and remembering were guided by our knowledge of the way
things generally unfold.
The Cost of Memory Errors
There’s clearly a “good news, bad news” quality to our discussion so far. On
the positive side, memory connections serve as retrieval paths, allowing you
to locate information in storage. The connections also enrich your understanding, because they tie each of your memories into a context provided
by other things you know. In addition, links to schematic knowledge enable
you to supplement your perception and recollection with well-informed (and
usually accurate) inference.
On the negative side, though, the same connections can undermine memory accuracy, and memory errors are troubling. As we’ve discussed in other
contexts, you rely on memory in many aspects of life, and it’s unsettling that
the memories you rely on may be wrong — misrepresenting how the past
unfolded.
Eyewitness Errors
EXONERATION OF
THE INNOCENT
Guy Miles spent more than
18 years in prison for an armed
robbery he did not commit.
He is one of the hundreds of
people who were convicted
in U.S. courts but then proven
innocent by DNA evidence.
Mistaken eyewitness evidence
accounts for more of these
false convictions than all other
causes combined. Note in
addition that these false
convictions typically involve
a “double error” — with someone innocent doing time in
jail, and the guilty person
walking around free.
288 •
In fact, we can easily find circumstances in which memory errors are large
in scale (not just concerned with minor details in the episode) and deeply
consequential. For example, errors in eyewitness testimony (e.g., identifying the wrong person as the culprit or misreporting how an event unfolded)
can potentially send an innocent person to jail and allow a guilty person
to go free.
How often do eyewitnesses make mistakes? One answer comes from U.S.
court cases in which DNA evidence, not available at the time of the trial,
shows that the courts had convicted people who were, in truth, not guilty.
There are now more than 350 of these exonerations, and the exonerees had
(on average) spent more than a dozen years in jail for crimes they didn’t
commit. Many of them were on death row, awaiting execution.
When closely examined, these cases yield a clear message. Some of
these men and women were convicted because of dishonest informants;
some because analyses of forensic evidence had been botched. But by far
the most common concern is eyewitness errors. In fact, according to most
analyses, eyewitness errors account for at least three quarters of these false
convictions — more than all other causes combined (e.g., Garrett, 2011;
Reisberg, 2014).
Cases like these make it plain that memory errors, including misidentifications, are profoundly important. We’re therefore led to ask: Are there ways
to avoid these errors? Or are there ways to detect the errors, so that we can
decide which memories are correct and which ones are not?
C H A P T E R E I G H T Remembering Complex Events
Planting False Memories
An enormous number of studies have examined eyewitness memory — the
sort of memory that police rely on when investigating crimes. In one of the
earliest procedures, Loftus and Palmer (1974) showed participants a series
of pictures depicting an automobile collision. Later, participants were asked
questions about the collision, but the questions were phrased in different ways for different groups. Some participants were asked, for example,
“How fast were the cars going when they hit each other?” A different group
was asked, “How fast were the cars going when they smashed into each
other?” The differences among these questions were slight, but had a substantial influence: Participants in the “hit” group estimated the speed to have
been 34 miles per hour; those in the “smashed” group estimated 41 miles
per hour — 20% higher (see Figure 8.3).
But what is critical comes next: One week later, the participants were
asked in a perfectly neutral way whether they had seen any broken glass
in the pictures. Participants who had initially been asked the “hit” question
tended to remember (correctly) that no glass was visible; participants who
had been asked the “smashed” question, though, often made this error. It
FIGURE 8.3
THE IMPACT OF LEADING QUESTIONS
Estimated speed (mph)
45
40
35
30
25
20
Contacted
Hit
Bumped
Collided
Smashed
Verb used
Witnesses who were asked how fast cars were going when they “hit” each
other reported (on average) a speed of 34 miles per hour. Other witnesses,
asked how fast the cars were going when they “smashed” into each other,
gave estimates 20% higher. When all participants were later asked whether
they’d seen broken glass at the scene, participants who’d been asked the
“smashed” question were more likely to say yes — even though there was no
broken glass.
(after loftus & palmer, 1974)
The Cost of Memory Errors
•
289
seems, therefore, that the change of just one word within the initial question
can have a significant effect — in this case, more than doubling the likelihood
of memory error.
In other studies, participants have been asked questions that contain overt
misinformation about an event. For example, they might be asked, “How
fast was the car going when it raced by the barn?” when, in truth, no barn
was in view. In still other studies, participants are exposed to descriptions of
the target event allegedly written by “other witnesses.” They might be told,
for example, “Here’s how someone else recalled the crime; does this match
what you recall?” Of course, the “other witness” descriptions contained some
misinformation, enabling researchers to determine if participants “pick up”
the false leads (e.g., Paterson & Kemp, 2006; also Edelson, Sharon, Dolan,
& Dudai, 2011). In other studies, researchers ask questions that require the
participants themselves to make up some bit of misinformation. For example,
participants could be asked, “In the video, was the man bleeding from his
knee or from his elbow after the fall?” Even though it was clear in the video
that the man wasn’t bleeding at all, participants are forced to choose one
of the two options (e.g., Chrobak & Zaragoza, 2008; Zaragoza, Payment,
Ackil, Drivdahl, & Beck, 2001).
These procedures differ in important ways, but they are all variations
on the same theme. In each case, the participant experiences an event and
then is exposed to a misleading suggestion about how the event unfolded.
Then some time is allowed to pass. At the end of this interval, the participant’s memory is tested. And in each of these variations, the outcome is the
same: A substantial number of participants — in some studies, more than
one third — end up incorporating the false suggestion into their memory of
the original event.
Of course, some attempts at manipulating memory are more successful, some less so. It’s easier, for example, to plant plausible memories rather
than implausible ones. (However, memories for implausible events can also
be planted — see Hyman, 2000; Mazzoni, Loftus, & Kirsch, 2001; Pezdek,
Blandon-Gitlin, & Gabbay, 2006; Scoboria, Mazzoni, Kirsch, & Jimenez,
2006; Thomas & Loftus, 2002.) Errors are also more likely if the post-event
information supplements what the person remembers, in comparison to contradicting what the person would otherwise remember. It’s apparently easier,
therefore, to “add to” a memory than it is to “replace” a memory (Chrobak &
Zaragoza, 2013). False memories are also more easily planted if the research
participants don’t just hear about the false event but, instead, are urged to
imagine how the suggested event unfolded. In one study, participants were
given a list of possible childhood events (going to the emergency room late at
night; winning a stuffed animal at a carnival; getting in trouble for calling 911)
and were asked to “picture each event as clearly and completely” as they
could. This simple exercise was enough to increase participants’ confidence
that the event had really occurred (Garry, Manning, Loftus, & Serman, 1996;
also Mazzoni & Memon, 2003; Sharman & Barnier, 2008; Shidlovski, Schul,
& Mayo, 2014).
290 •
C H A P T E R E I G H T Remembering Complex Events
Even acknowledging these variations, though, let’s emphasize the consistency of the findings. We can use subtle procedures (with slightly leading
questions) to plant false information in someone’s memory, or we can use
a more blatant procedure (demanding that the person make up the bogus
facts). We can use pictures, movies, or live events as the to-be-remembered
materials. In all cases, it’s remarkably easy to alter someone’s memory, with
the result that the past as the person remembers it can differ markedly from
the past as it really was. This is a widespread pattern, with numerous implications for how we think about the past and how we think about our reliance
on our own memories. (For more on research in this domain, see Carpenter
& Schacter, 2017; Cochran, Greenspan, Bogart, & Loftus, 2016; Frenda,
Nichols, & Loftus, 2011; Laney, 2012; Loftus, 2017; Rich & Zaragoza,
2016. For research documenting similar memory errors in children, see, e.g.,
Bruck & Ceci, 1999, 2009; Reisberg, 2014.)
Are There Limits on the Misinformation Effect?
The studies just described reflect the misinformation effect — a term referring
to memory errors that result from misinformation received after an event was
experienced. What sorts of memory errors can be planted in this way?
We’ve mentioned studies in which participants remember broken glass
when really there was none or remember a barn when there was no barn in
view. Similar procedures have altered how people are remembered — and so,
with just a few “suggestions” from the experimenter, participants remember
clean-shaven men as bearded, young people as old, and fat people as thin
(e.g., Christiaansen, Sweeney, & Ochalek, 1983; Frenda et al., 2011).
It’s remarkably easy to produce these errors — with just one word (“hit”
vs. “smashed”) being enough to alter an individual’s recollection. What happens, though, if we ramp up our efforts to plant false memories? Can we
create larger-scale errors? In one study, college students were told that the
investigators were trying to learn how different people remember the same
experience. The students were then given a list of events that (they were told)
had been reported by their parents; the students were asked to recall these
events as well as they could, so that the investigators could compare the students’ recall with their parents’ (Hyman, Husband, & Billings, 1995).
Some of the events on the list actually had been reported by the participants’ parents. Other events were bogus — made up by the experimenters.
One of the bogus events was an overnight hospitalization for a high fever; in
a different experiment, the bogus event was attending a wedding reception
and accidentally spilling a bowlful of punch on the bride’s family.
The college students were easily able to remember the genuine events (i.e.,
the events actually reported by their parents). In an initial interview, more than
80% of these events were recalled, but none of the students recalled the bogus
events. However, repeated attempts at recall changed this pattern. By a third
interview, 25% of the participants were able to remember the embarrassment
of spilling the punch, and many were able to supply the details of this (entirely
The Cost of Memory Errors
•
291
FIGURE 8.4
A
THE BALLOON RIDE THAT NEVER WAS
B
In this study, participants were shown a faked photo (Panel B) created from a real childhood snapshot
(Panel A). With this prompt, many participants were led to a vivid, detailed recollection of the balloon
ride — even though it never occurred!
fictitious) episode. Other studies have shown similar results. Participants have
been led to recall details of particular birthday parties that, in truth, they never
had (Hyman et al., 1995); or an incident of being lost in a shopping mall
even though this event never took place; or a (fictitious) event in which they
were the victim of a vicious animal attack (Loftus, 2003, 2004; also see, e.g.,
Chrobak & Zaragoza, 2008; Geraerts et al., 2009; Laney & Loftus, 2010).
Errors Encouraged through “Evidence”
Other researchers have taken a further step and provided participants with
“evidence” in support of the bogus memory. In one procedure, researchers
obtained a real childhood snapshot of the participant (see Figure 8.4A for an
example) and, with a few clicks of a computer mouse, created a fictitious picture like the one shown in Figure 8.4B. With this prompt, many participants
were led to a vivid, detailed recollection of the hot-air balloon ride — even
though it never occurred (Wade, Garry, Read, & Lindsay, 2002). Another
study used an unaltered photo showing the participants’ second-grade
class (see Figure 8.5 for an example). This was apparently enough to persuade participants that the experimenters really did have information about
their childhood. Therefore, when the experimenters “reminded” the participants about an episode of their childhood misbehavior, the participants
took this reminder seriously. The result: Almost 80% were able to “recall”
292 •
C H A P T E R E I G H T Remembering Complex Events
FIGURE 8.5
PHOTOGRAPHS CAN ENCOURAGE MEMORY ERRORS
In one study, participants were “reminded” of a (fictitious) stunt they’d pulled while in the second grade. Participants were much more likely to “remember” the stunt (and so more likely to develop a false memory) if the
experimenter showed them a copy of their actual second-grade class photo. Apparently, the photo convinced
the participants that the experimenter really did know what had happened, and this made the experimenter’s
(false) suggestion much more persuasive.
(lindsay et al., 2004)
the episode, often in detail, even though it had never happened (Lindsay,
Hagen, Read, Wade, & Garry, 2004).
False Memories, False Confessions
It is clear that people can sometimes remember entire events that never took
place. They sometimes remember emotional episodes (like being lost in a
shopping mall) that never happened. They can remember their own transgressions (spilling the punch bowl, misbehaving in the second grade), even
though these misdeeds never occurred.
One study pushed things still further, using a broad mix of techniques to
encourage false memories (Shaw & Porter, 2015). The interviewer repeatedly asked participants to recall an event that (supposedly) she had learned
about from their parents. She assured participants that she had detailed
information about the (fictitious) event, and she applied social pressure with
comments like “Most people are able to retrieve lost memories if they try
The Cost of Memory Errors
•
293
REMEMBERING
VISITORS
A substantial number of people
have vivid, elaborate memories
for an episode in which they
were abducted by space aliens.
They report the aliens’ medical examination of their human
captive; in some cases, they
describe being impregnated
by the aliens. Some people
regard these reports as proof
that our planet has been visited
by extraterrestrials. Most scientists, however, regard these
reports as false — as “memories” for an event that never
happened. On this interpretation, the abduction reports illustrate how wrong our memories
can sometimes be.
TEST YOURSELF
6. What is the misinformation effect?
Describe three
different procedures
that can produce
this effect.
7. Some people insist
that our memories
are consistently
accurate in remembering the gist, or overall
content, of an event;
when we make memory errors, they claim,
we make mistakes
only about the details
within an event. What
evidence allows us to
reject this claim?
294 •
hard enough.” She offered smiles and encouraging nods whenever participants showed signs of remembering the (bogus) target events. If participants
couldn’t recall the target events, she showed signs of disappointment and said
things like “That’s ok. Many people can’t recall certain events at first because
they haven’t thought about them for such a long time.” She also encouraged
participants to use a memory retrieval technique (guided imagery) that is
known to foster false memories.
With these (and other) factors in play, Shaw and Porter persuaded many
of their participants that just a few years earlier the participants had committed a crime that led to police contact. In fact, many participants seemed
able to remember an episode in which they had assaulted another person
with a weapon and had then been detained by the police. This felony never
happened, but many participants “recalled” it anyhow. Their memories were
in some cases vivid and rich with detail, and on many measures indistinguishable from memories known to be accurate.
Let’s be clear, though, that this study used many forms of influence and
encouragement. It takes a lot to pull memory this far off track! There has
also been debate over just how many of the participants in this study truly
developed false memories. Even so, the results show that it’s possible for a
large number of people to have memories that are emotionally powerful,
deeply consequential, and utterly false. (For discussion of Shaw and Porter’s
study, see Wade, Garry, & Pezdek, 2017. Also see Brewin & Andrews, 2017,
and then in response, Becker-Blease & Freyd, 2017; Lindsay & Hyman,
2017; McNally, 2017; Nash, Wade, Garry, Loftus, & Ost, 2017; Otgaar,
Merckelbach, Jelicic, & Smeets, 2017; and Scoboria & Mazzoni, 2017.)
C H A P T E R E I G H T Remembering Complex Events
COGNITION
outside the lab
“It’s Common Sense”
Psychology students sometimes get teased by
the risk of someday being kidnapped by space
their peers: “Why are you taking Psych courses?
aliens. In contrast, though, studies make it plain
It’s all a matter of common sense!” The same senti-
that just a word or two of leading can produce
ment can arise when psychologists testify in court
memory errors in roughly one third of the people
cases, with the goal of helping judges and juries
questioned. Surely, the danger of extraterrestrial
understand how memory functions — and how
abduction is much lower than this.
someone’s memory can be mistaken. Some judges,
Other commonsense beliefs are flatly wrong. For
however, refuse this testimony. In support of this
example, some people have the view that certain
refusal, they note that expert testimony is allowed
types of events are essentially immune to forget-
only if it will be helpful in deciding the case, and
ting. They speak about those events as somehow
the testimony won’t be helpful if it simply covers
“burned into the brain” and say things like “I’ll never
points that judge and jury already know. In legal
forget the events of 9/11” or “. . . the day I got mar-
jargon, the testimony is permitted only if it covers
ried” or “. . . what he looked like when he pulled the
topics “beyond the ken of the average juror.”
trigger.” However, the “burned into the brain” idea is
How should we think about these notions? Are
psychology’s claims about memory simply a con-
wrong, and investigators can often document largescale errors in these singular, significant memories.
firmation of common sense? Each of us, of course,
Additional examples are easy to find. These in-
has had a lifetime of experience working with and
clude the widely held view that someone’s degree of
relying on our memories; that experience has surely
certainty is a good index of whether his or her mem-
taught us a lot about how memory functions. Even
ory is accurate (this is true only in a narrow set of cir-
so, it’s easy to find widespread beliefs about mem-
cumstances); the common belief that our memories
ory that are incorrect. Often, these beliefs start
function just as a video recorder functions (not at all
with a kernel of truth but understate the actual
true); or the belief that hypnosis can allow someone
facts. For example, everyone knows that memo-
to recover long-lost memories (utterly false).
ries are sometimes inaccurate; people talk about
In fact, let’s note an irony here. Commonsense
their memories “playing tricks” on them. However,
beliefs about memory (or about psychology in
most people are astonished to learn how common
general) are sometimes sensible and sometimes
memory errors are and how large the errors can
not. If scientific research corrects a mistaken com-
sometimes be. Therefore, in relying on common
monsense belief, then obviously we’ve learned
sense, people (including judges and juries) prob-
something. But if the research turns out to con-
ably trust memory more than they should.
firm common sense, then here too we’ve learned
For example, in one study, college students
something — because we’ve learned that this is
were surveyed about their perceptions of vari-
one of the times when common sense is on track.
ous risks (Wilson & Brekke, 1994). These students
On that basis, we shouldn’t scoff at results that
were largely unconcerned about the risk of some-
“merely” confirm common sense, because these
one biasing their memory with leading questions;
results can be just as informative as results that
they regarded this risk as roughly equivalent to
truly surprise us.
The Cost of Memory Errors
•
295
Avoiding Memory Errors
Evidence is clear that people do make mistakes — at times, large mistakes — in
remembering the past. But people usually don’t make mistakes. In other
words, you generally can trust your memory, because more often than not
your recollection is detailed, long-lasting, and correct.
This mixed pattern, though, demands a question: Is there some way to
figure out when you’ve made a memory mistake and when you haven’t? Is there
a way to decide which memories you can rely on and which ones you can’t?
Memory Confidence
In evaluating memories, people rely heavily on expressions of certainty or
confidence. Specifically, people tend to trust memories that are expressed
with confidence. (“I distinctly remember her yellow jacket; I’m sure of it.”)
They’re more cautious about memories that are hesitant. (“I think she was
wearing yellow, but I’m not certain.”) We can see these patterns when people are evaluating their own memories (e.g., when deciding whether to take
action or not, based on a bit of recollection); we see the same patterns when
people are evaluating memories they hear from someone else (e.g., when
juries are deciding whether they can rely on an eyewitness’s testimony).
Evidence suggests, though, that a person’s degree of certainty is an uneven
indicator of whether a memory is trustworthy. On the positive side, there
are circumstances in which certainty and memory accuracy are highly correlated (e.g., Wixted, Mickes, Clark, Gronlund, & Roediger, 2015; Wixted
& Wells, 2017). On the negative side, though, we can easily find exceptions
to this pattern — including memories that are expressed with total certainty
(“I’ll never forget that day; I remember it as though it were yesterday”) but
that turn out to be entirely mistaken. In fact, we can find circumstances in
which there’s no correspondence at all between how certain someone says
she is, in recalling the past, and how accurate that recollection is likely to be.
As a result, if we try to categorize memories as correct or incorrect based on
someone’s confidence, we’ll often get it wrong. (For some of the evidence,
see Busey, Tunnicliff, Loftus, & Loftus, 2000; Hirst et al., 2009; Neisser &
Harsch, 1992; Reisberg, 2014; Wells & Quinlivan, 2009.)
How can this be? One reason is that a person’s confidence in a memory is
often influenced by factors that have no impact on memory accuracy. When
these factors are present, confidence can shift (sometimes upward, sometimes
downward) with no change in the accuracy level, with the result that any connection between confidence and accuracy can be strained or even shattered.
Participants in one study witnessed a (simulated) crime and later were asked
if they could identify the culprit from a group of pictures. Some of the participants were then given feedback — “Good, you identified the suspect”; others
weren’t. The feedback couldn’t possibly influence the accuracy of the identification, because the feedback arrived only after the identification had occurred.
But the feedback did have a large impact on how confident participants said
296 •
C H A P T E R E I G H T Remembering Complex Events
FIGURE 8.6
CONFIDENCE MALLEABILITY
In one study, participants first tried to identify a culprit
from a police lineup and then indicated (on a scale of
0 to 100) how confident they had been in their selection.
Some participants received no feedback about their choice;
others received feedback after making their selection
but before indicating their confidence level. The feedback
couldn’t possibly influence accuracy (because the selection had already been made), but it dramatically increased
confidence.
(after wells & bradfield, 1998)
Mean reported confidence
80
70
60
50
40
30
No feedback
they’d been when making their lineup selection (see Figure 8.6), and so, with
confidence inflated but accuracy unchanged, the linkage between confidence
and accuracy was essentially eliminated. (Wells & Bradfield, 1998; also see
Douglas, Neuschatz, Imrich, & Wilkinson, 2010; Semmler & Brewer, 2006;
Wells, Olson, & Charman, 2002, 2003; Wright & Skagerberg, 2007.)
Similarly, think about what happens if someone is asked to report on an
event over and over. The repetitions don’t change the memory content — and
so the accuracy of the report won’t change much from one repetition to
the next. However, with each repetition, the recall becomes easier and more
fluent, and this ease of recall seems to make people more confident that their
memory is correct. So here, too, accuracy is unchanged but confidence is
inflated — and thus there’s a gradual erosion, with each repetition, of the
correspondence between accuracy and confidence. (For more on the disconnection between accuracy and confidence, see, e.g., Bradfield Douglas &
Pavletic, 2012; Charman, Wells, & Joy, 2011.)
In many settings, therefore, we cannot count on confidence as a means
of separating accurate memories from inaccurate ones. In addition, other
findings tell us that memory errors can be just as emotional, just as vivid, as
accurate memories (e.g., McNally et al., 2004). In fact, research overall suggests that there simply are no indicators that can reliably guide us in deciding
which memories to trust and which ones not to trust. For now, it seems that
memory errors, when they occur, may often be undetectable.
“Good, you identified
our suspect”
Feedback condition
TEST YOURSELF
8. W
hat factors seem to
undermine the relationship between your
degree of certainty in a
memory and the likelihood that the memory
is accurate?
Forgetting
We’ve been discussing the errors people sometimes make in recalling the past,
but of course there’s another way your memory can let you down: Sometimes
you forget. You try to recall what was on the shopping list, or the name of
Forgetting
•
297
an acquaintance, or what happened last week, and you simply draw a blank.
Why does this happen? Are there things you can do to diminish forgetting?
The Causes of Forgetting
Let’s start with one of the more prominent examples of “forgetting” — which
turns out not to be forgetting at all. Imagine meeting someone at a party,
being told his name, and moments later realizing you don’t have a clue what
his name is — even though you just heard it. This common (and embarrassing)
experience is not the result of ultra-rapid forgetting. Instead, it stems from a
failure in acquisition. You were exposed to the name but barely paid attention to it and, as a result, never learned it in the first place.
What about “real” cases of forgetting — cases in which you once knew
the information but no longer do? For these cases, one of the best predictors
of forgetting (not surprisingly) is the passage of time. Psychologists use the
term retention interval to refer to the amount of time that elapses between
the initial learning and the subsequent retrieval; as this interval grows, you’re
likely to forget more and more of the earlier event (see Figure 8.7).
FIGURE 8.7
FORGETTING CURVE
100
90
Percentage retained
80
70
60
50
40
30
20
31 days
6 days
2 days
1 day
9 hours
60 mins
0 min
0
20 mins
10
Retention interval
The figure shows retention after various intervals since learning. The data
shown here are from classic work by Hermann Ebbinghaus, so the pattern is
often referred to as an “Ebbinghaus forgetting curve.” The actual speed of
forgetting (i.e., how “steep” the “drop-off” is) depends on how well learned
the material was at the start. Across most situations, though, the pattern is
the same — with the forgetting being rapid at first but then slowing down.
Mathematically, this pattern is best described by an equation framed in
terms of “exponential decay.”
298 •
C H A P T E R E I G H T Remembering Complex Events
One explanation for this pattern comes from the decay theory of forgetting,
which proposes rather directly that memories fade or erode with the passage
of time. Maybe this is because the relevant brain cells die off. Or maybe
the connections among memories need to be constantly refreshed — and if
they’re not refreshed, the connections gradually weaken.
A different possibility is that new learning somehow interferes with older
learning. This view is referred to as interference theory. According to this
view, the passage of time isn’t the direct cause of forgetting. Instead, the passage of time creates the opportunity for new learning, and it is the new learning that disrupts the older memories.
A third hypothesis blames retrieval failure. The idea here is that the “forgotten memory” is still in long-term storage, but the person trying to retrieve the
memory simply cannot locate it. This proposal rests on the notion that retrieval
from memory is far from guaranteed, and we argued in Chapter 7 that retrieval
is more likely if your perspective at the time of retrieval matches the perspective in place at the time of learning. If we now assume that your perspective is
likely to change as time goes by, we can make a prediction about forgetting:
The greater the retention interval, the greater the likelihood that your perspective has changed, and therefore the greater the likelihood of retrieval failure.
Which of these hypotheses is correct? It turns out that they all are. Memo­
ries do decay with the passage of time (e.g., Altmann & Schunn, 2012;
Wixted, 2004; also Hardt, Nader, & Nadel, 2013; Sadeh, Ozubko, Winocur,
& Moscovitch, 2016), so any theorizing about forgetting must include this
factor. But there’s also no question that a great deal of “forgetting” is retrieval
failure. This point is evident whenever you’re initially unable to remember
some bit of information, but then, a while later, you do recall that information. Because the information was eventually retrieved, we know that it
wasn’t “erased” from memory through either decay or interference. Your
initial failure to recall the information, then, must be counted as an example
of retrieval failure.
Sometimes retrieval failure is partial: You can recall some aspects of the
desired content, but not all. An example comes from the maddening circumstance in which you’re trying to think of a word but simply can’t come up
with it. The word is, people say, on the “tip of their tongue,” and following
this lead, psychologists refer to this as the TOT phenomenon. People experiencing this state can often recall the starting letter of the sought-after word
and approximately what it sounds like. So, for example, a person might
remember “it’s something like Sanskrit” in trying to remember “scrimshaw”
or “something like secant” in trying to remember “sextant” (Brown, 1991;
Brown & McNeill, 1966; Harley & Brown, 1998; James & Burke, 2000;
Schwartz & Metcalfe, 2011).
What about interference? In one early study, Baddeley and Hitch (1977)
asked rugby players to recall the names of the other teams they had played
against over the course of a season. The key here is that not all players
made it to all games (because of illness, injuries, or schedule conflicts).
This fact allows us to compare players for whom “two games back” means
Forgetting
•
299
two weeks ago, to players for whom “two games back” means four weeks
ago. In this way, we can look at the effects of retention interval (two
weeks vs. four) with the number of intervening games held constant. Likewise, we can compare players for whom the game a month ago was “three
games back” to players for whom a month ago means “one game back.”
Now, we have the retention interval held constant, and we can look at the
effects of intervening events. In this setting, Baddeley and Hitch reported
that the mere passage of time accounts for very little; what really matters
is the number of intervening events (see Figure 8.8). This is just what we
would expect if interference, and not decay, is the major contributor to
forgetting.
But why does memory interference occur? Why can’t the newly acquired
information coexist with older memories? The answer has several parts, but
one element is linked to issues we’ve already discussed: In many cases, newly
arriving information gets interwoven with older information, producing a
risk of confusion about which bits are old (i.e., the event you’re trying to
remember) and which bits are new (i.e., information that you picked up after
the event). In addition, in some cases, new information seems literally to
replace old information — much as you no longer save the rough draft of one
FIGURE 8.8
FORGETTING FROM INTERFERING EVENTS
100
Percentage of team names
recalled correctly
90
80
70
60
50
40
30
20
10
1
2
3
4
5 6 7 8 9 10 11 12 13 14 15 16 17 18
Number of intervening games played
Members of a rugby team were asked to recall the names of teams they had
played against. Overall, the broad pattern of the data shows that memory performance was powerfully influenced by the number of games that intervened
between the game to be recalled and the attempt to remember. This pattern fits
with an interference view of forgetting.
(after baddeley & hitch, 1977)
300 •
C H A P T E R E I G H T Remembering Complex Events
of your papers once the final draft is done. In this situation, the new information isn’t woven into the older memory; instead, it erases it.
Undoing Forgetting
Is there any way to undo forgetting and to recover seemingly lost memories?
One option, often discussed, is hypnosis. The idea is that under hypnosis a
person can “return” to an earlier event and remember virtually everything
about the event, including aspects the person didn’t even notice (much less
think about) at the time.
The reality, however, is otherwise. Hypnotized participants often do give
detailed reports of the target event, but not because they remember more; instead, they’re just willing to say more in order to comply with the hypnotist’s
instructions. As a result, their “memories” are a mix of recollection, guesses,
and inferences — and, of course, the hypnotized individual cannot tell which
of these are which (Lynn, Neuschatz, Fite, & Rhue, 2001; Mazzoni & Lynn,
2007; Spiegel, 1995).
On the positive side, though, there are procedures that do seem to diminish forgetting, including the so-called cognitive interview. This procedure was
A Drawings done by hypnotized adult told
that he was 6 years old
B Drawings done at age 6
HYPNOTIC AGE REGRESSION
In one study, participants were asked to draw a picture while mentally “regressed” to age 6. At first glance, their drawings
(see Panel A for an example) looked remarkably childlike. But when compared to the participants’ own drawings made
at that age (see Panel B for an example), it’s clear that the hypnotized adults’ drawings were much more sophisticated.
They represent an adult’s conception of what a childish drawing is, rather than being the real thing.
Forgetting
•
301
TEST YOURSELF
9. Explain the mechanisms hypothesized
by each of the three
major theories of
forgetting: decay,
interference, and
retrieval failure.
10. W
hat techniques or
procedures seem
ineffective as a means
of “un-doing” forgetting? What techniques
or procedures seem
to diminish or avoid
forgetting?
designed to help police in their investigations and, specifically, is aimed at
maximizing the quantity and accuracy of information obtained from eyewitnesses to crimes (Fisher & Schreiber, 2007; Memon, Meissner, & Fraser,
2010). The cognitive interview has several elements, including an effort
toward context reinstatement — steps that put witnesses back into the mindset they were in at the time of the crime. (For more on context reinstatement,
see Chapter 7.) In addition, the cognitive interview builds on the simple fact
that retrieval of memories from long-term storage is more likely if a suitable
cue is provided. The interview therefore offers a diverse set of retrieval cues
with the idea that the more cues provided, the greater the chance of finding
one that triggers the target memory.
The cognitive interview is quite successful, both in the laboratory and
in real crime investigations, producing more complete recollection without
compromising accuracy. This success adds to the argument that much of
what we call “forgetting” can be attributed to retrieval failure, and can be
undone simply by providing more support for retrieval.
Also, rather than undoing forgetting, perhaps we can avoid forgetting.
The key here is simply to “revisit” a memory periodically. Each “visit” seems
to refresh the memory, with the result that forgetting is much less likely. Researchers have examined this effect in several contexts, including one that’s
pragmatically quite important: Students often have to take exams, and confronting the material on an exam is, of course, an occasion in which students
“revisit” what they’ve learned. These revisits, we’ve just suggested, should
slow forgetting, and on this basis, taking an exam can actually help students
to hang on to the material they’ve learned. Several studies have confirmed
this “testing effect”: Students have better long-term retention for materials they were tested on, compared to materials they weren’t tested on. (See,
e.g., Carpenter, Pashler, & Cepeda, 2009; Halamish & Bjork, 2011; Healy,
Jones, Lalchandani, & Tack, 2017; Karpicke, 2012; Karpicke & Blunt, 2011;
McDaniel, Anderson, Derbish, & Morrisette, 2007; Pashler, Rohrer, Cepeda,
& Carpenter, 2007; Rowland, 2014.)
We might mention that similar effects can be observed if students test
themselves periodically, taking little quizzes that they’ve created on their
own. Related effects emerge if students are occasionally asked questions that
require a brief revisit to materials they’ve encountered (Brown, Roediger, &
McDaniel, 2014). In fact, that’s the reason why this textbook includes Test
Yourself questions; those questions will actually help readers to remember
what they’ve read!
Memory: An Overall Assessment
We’ve now seen that people sometimes recall with confidence events that
never took place, and sometimes forget information they’d hoped to remember. But we’ve also mentioned the positive side of things: how much people
can recall, and the key fact that your memory is accurate far more often than
not. Most of the time, it seems, you do recall the past as it truly was.
302 •
C H A P T E R E I G H T Remembering Complex Events
Perhaps most important, we’ve also suggested that memory’s “failings”
may simply be the price you pay in order to gain crucial advantages. For
example, we’ve argued that memory errors arise because the various episodes in your memory are densely interconnected with one another; it’s these
interconnections that allow elements to be transplanted from one remembered episode to another. But we’ve also noted that these connections have
a purpose: They’re the retrieval paths that make memory search possible.
Therefore, to avoid the errors, you would need to restrict the connections;
but if you did that, you would lose the ability to locate your own memories
within long-term storage.
The memory connections that lead to error also help you in other ways.
Our environment, after all, is in many ways predictable, and it’s enormously useful for you to exploit that predictability. There’s little point,
for example, in scrutinizing a kitchen to make sure there’s a stove in the
room, because in the vast majority of cases there is. So why take the time
to confirm the obvious? Likewise, there’s little point in taking special note
that, yes, this restaurant does have menus and, yes, people in the restaurant are eating and not having their cars repaired. These, too, are obvious
points, and it would be a waste of effort to give them special notice.
On these grounds, reliance on schematic knowledge is a good thing. Schemata guide your attention to what’s informative in a situation, rather than
what’s self-evident (e.g., Gordon, 2006), and they guide your inferences at
the time of recall. If this use of schemata sometimes leads you astray, that’s a
small price to pay for the gain in efficiency that schemata allow. (For similar
points, see Chapter 4.)
In the same way, the blurring together of episodes may be a blessing, not
a problem. Think, for example, about all the times when you’ve been with
a particular friend. These episodes are related to one another in an obvious
way, and so they’re likely to become interconnected in your memory. This
will cause difficulties if you want to remember which episode is which and
whether you had a particular conversation in this episode or in that one.
But rather than lamenting this, maybe we should celebrate what’s going on
here. Because of the “interference,” all the episodes will merge together in
your memory, so that what resides in memory is one integrated package,
containing all of your knowledge about your friend. As a result, rather than
complaining about memory confusion, we should rejoice over the memory
integration and “cross-referencing.”
In all of these ways, then, our overall assessment of memory can be rather
upbeat. We have, to be sure, discussed a range of memory errors, but these
errors are in most cases a side product of mechanisms that otherwise help
you — to locate your memories within storage, to be efficient in your contact
with the world, and to form general knowledge. Thus, even with the errors,
even with forgetting, it seems that human memory functions in a way that
serves us extraordinarily well. (For more on the benefits produced by memory’s apparent limitations, see Howe, 2011; Nørby, 2015; Schacter, Guerin,
& St. Jacques, 2011.)
TEST YOURSELF
11.Explain why the
mechanisms that
produce memory
errors may actually be
mechanisms that help
us in important ways.
Memory: An Overall Assessment
•
303
Autobiographical Memory
Most of the evidence in Chapters 6 and 7 was concerned with memory for
simple stimuli — such as word lists or short sentences. In this chapter, we’ve
considered memories for more complex materials, and this has drawn our
attention to the ways in which your knowledge (whether knowledge of a
general sort or knowledge about related episodes) can both improve memory
and also interfere with it.
In making these points, we’ve considered memories in which the research
participant was actually involved in the remembered episode, and not just an
external witness (e.g., the false memory that he committed a felony). We’ve
also looked at studies that involved memories for emotional events (e.g., the
plane crash discussed at the chapter’s start) and memory over the very long
term (e.g., memories for childhood events “planted” in adult participants).
Do these three factors — involvement in the remembered event, emotion,
and long delay — affect how or how well someone remembers? These factors
are surely relevant to the sorts of remembering people do outside the laboratory, and all three are central for autobiographical memory. This is the memory that each of us has for the episodes and events of our lives, and this sort
of memory plays a central role in shaping how each of us thinks about ourselves and, therefore, how we behave. (For more on the importance of autobiographical memory, see Baddeley, Aggleton, & Conway, 2002; Prebble, Addis,
& Tippett, 2013; Steiner, Thomsen, & Pillemer, 2017. For more on the distinction between the types of memory, including biological differences between
autobiographical memory and “lab memory,” see Cabeza & St. Jacques, 2007;
Hodges & Graham, 2001; Kopelman & Kapur, 2001; Tulving, 1993, 2002.)
Let’s explore how the three factors we’ve mentioned, each seemingly central for autobiographical memory, influence what we remember.
Memory and the Self
Having some involvement in an event (as opposed to passively witnessing it)
turns out to have a large effect on memory, because, overall, information
relevant to the self is better remembered than information that’s not selfrelevant — a pattern known as the “self-reference effect” (e.g., Symons &
Johnson, 1997; Westmacott & Moscovitch, 2003). This effect emerges in many
forms, including an advantage in remembering adjectives that apply to you
relative to adjectives that don’t, better memory for names of places you have
visited relative to names of places you’ve never been, and so on (see Figure 8.9).
But here, too, we can find memory errors, in part because your “memory”
for your own life is (just like other memories) a mix of genuine recall and
some amount of schema-based reconstruction. For example, consider the
fact that most adults believe they’ve been reasonably consistent, reasonably
stable, over their lifetimes. They believe, in other words, that they’ve always
been pretty much the same as they are now. This idea of consistency is part of
their self-schema — the set of interwoven beliefs and memories that constitute
304 •
C H A P T E R E I G H T Remembering Complex Events
FIGURE 8.9
SELF-REFERENCING AND THE BRAIN
Words in relation to another person
Words in relation to its printed format
MPFC activated
during self-referential
condition.
Words in relation to self
0.2
Signal change (%)
0.1
0
–0.1
–0.2
–0.3
–0.4
–5
0
5
10
15
20
Time (s)
You are more likely to remember words that refer to you, in comparison to words in other categories. Here, participants were asked to judge adjectives in three conditions: answering questions like “Does this word describe
the president?” or “Is this word printed in capital letters?” or “Does this word describe you?” Data from fMRI
recordings showed a distinctive pattern of processing when the words were “self-referential.” Specifically, selfreferential processing is associated with activity in the medial prefrontal cortex (MPFC). This extra processing is
part of the reason why self-referential words are better remembered.
(after kelley et al., 2002)
people’s knowledge about themselves. When the time comes to remember
the past, therefore, people will rely to some extent on this belief in their own
consistency, so they’ll reconstruct their history in a biased way — one that
maximizes the (apparent) stability of their lives. As a result, people often misremember their past attitudes and past romantic relationships, unwittingly
distorting their personal history in a way that makes the past look more
like the present than it really was. (See Conway & Ross, 1984; Holmberg &
Homes, 1994. For related results, see Levine, 1997; Marcus, 1986; McFarland
& Buehler, 2012; Ochsner & Schacter, 2000; Ross & Wilson, 2003.)
It’s also true that most of us would like to have a positive view of ourselves, including a positive view of how we’ve acted in the past. This, too, can
shape memory. As one illustration, Bahrick, Hall, and Berger (1996) asked
college students to recall their high school grades as accurately as they could,
and the data showed a clear pattern of self-service. When students forgot a
good grade, their (self-serving) reconstruction led them to the (correct) belief
Autobiographical Memory
•
305
that the grade must have been a good one; consistent with this, 89% of the A’s
were correctly remembered. But when students forgot a poor grade, reconstruction led them to the (false) belief that the grade must have been okay; as
a result, only 29% of the D’s were correctly recalled. (For other mechanisms
through which motivation can color autobiographical recall, see Conway &
Holmes, 2004; Conway & Pleydell-Pearce, 2000; Molden & Higgins, 2012.)
Memory and Emotion
Another factor important for autobiographical memory is emotion. Many
of your life experiences are of course emotional, making you feel happy, or
sad, or angry, or afraid, and in general emotion helps you to remember. One
reason is emotion’s impact on memory consolidation — the process through
which memories are biologically “cemented in place.” (See Hardt, Einarsson,
& Nader, 2010; Wang & Morris, 2010; although also see Dewar, Cowan, &
Della Sala, 2010.)
Whenever you experience an event or gain new knowledge, your memory
for this new content is initially fragile and is likely represented in the brain
via a pattern of neural activation. Over the next few hours, though, various
biological processes stabilize this memory and put it into a more enduring
form. This process — consolidation — takes place “behind the scenes,” without you thinking about it, but it’s crucial. If the consolidation is interrupted
for some reason (e.g., because of extreme fatigue or injury), no memory is established and recall later will be impossible. (That’s because there’s no information in memory for you to retrieve; you can’t read text off a blank page!)
A number of factors can promote consolidation. For example, evidence is increasing that key steps of consolidation take place while you’re
asleep — and so a good night’s rest actually helps you, later on, to remember
things you learned while awake the day before. (See Ackermann & Rasch,
2014; Giuditta, 2014; Rasch & Born, 2013; Tononi & Cirelli, 2013; Zillmer,
Spiers, & Culbertson, 2008.)
Also, there’s no question that emotion enhances consolidation. Specifically, emotional events trigger a response in the amygdala, and the amygdala
in turn increases activity in the hippocampus. The hippocampus is, as we’ve
seen, crucial for getting memories established. (See Chapter 7; for reviews of
emotion’s biological effects on memory, see Buchanan, 2007; Hoschedidt,
Dongaonkar, Payne, & Nadel, 2010; Joels, Fernandez, & Roosendaal, 2011;
Kensinger, 2007; LaBar, 2007; LaBar & Cabeza, 2006; Yonelinas & Ritchey,
2015. For a complication, though, see Figure 8.10.)
Emotion also shapes memory through other mechanisms. An event that’s
emotional is likely to be important to you, virtually guaranteeing that you’ll pay
close attention as the event unfolds, and we know that attention and thoughtful processing help memory. Moreover, you tend to mull over emotional events
in the minutes (or hours) following the event, and this is tantamount to memory rehearsal. For all these reasons, it’s not surprising that emotional events
are well remembered (Reisberg & Heuer, 2004; Talmi, 2013).
306 •
C H A P T E R E I G H T Remembering Complex Events
FIGURE 8.10
I NDIVIDUAL DIFFERENCES IN
EPISODIC MEMORY
BB
EE
KB
BK
HG
NL
CC
JL
SC
Group
Researchers have made enormous progress in explaining the brain mechanisms that support memory. One complication, though, is that the brain
mechanisms may differ from one individual to the next. This figure shows
data from nine different people (and then an average of the nine) engaged
in a task requiring the retrieval of episodic memories. As you can see, the
pattern of brain activation differs somewhat from person to person.
( after miller et al ., 2002)
Let’s note, though, that emotion doesn’t just influence how well you remember; it also influences what you remember. Specifically, in many settings,
emotion seems to produce a “narrowing” of attention, so that all of your
attention will be focused on just a few aspects of the scene (Easterbrook,
1959). This narrowing helps guarantee that these attended aspects will be
firmly placed into memory, but it also implies that the rest of the event,
excluded from the narrowed focus, won’t be remembered later (e.g., Gable
& Harmon-Jones, 2008; Reisberg & Heuer, 2004; Steblay, 1992).
What exactly you’ll focus on, though, may depend on the specific emotion.
Different emotions lead you to set different goals: If you’re afraid, your goal
is to escape; if you’re angry, your goal is to deal with the person or issue that’s
Autobiographical Memory
•
307
made you angry; if you’re happy, your goal may be to relax and enjoy! In each
case, you’re more likely to pay attention to aspects of the scene directly rele­
vant to your goal, and this will color how you remember the emotional event.
(See Fredrickson, 2000; Harmon-Jones, Gable, & Price, 2013; Huntsinger,
2012, 2013; Kaplan, Van Damme, & Levine, 2012; Levine & Edelstein, 2009.)
Flashbulb Memories
One group of emotional memories seems special. These are the so-called
flashbulb memories — memories of extraordinary clarity, typically for highly
emotional events, retained despite the passage of many years. When Brown
and Kulik (1977) introduced the term “flashbulb memory,” they pointed to
the memories people had of the moment in 1963 when they first heard that
President Kennedy had been assassinated. In the Brown and Kulik study,
people interviewed more than a decade after that event remembered it
“as though it were yesterday,” and many participants were certain they’d
never forget that awful day. Moreover, participants’ recollection was quite
detailed — with people remembering where they were at the time, what they
were doing, and whom they were with. Indeed, many participants were able
to recall the clothing worn by people around them, the exact words uttered,
and the like.
Many other events have also produced flashbulb memories. For example,
most Americans can clearly recall where they were when they heard about
the attack on the World Trade Center in 2001; many people vividly remember what they were doing in 2009 when they heard that Michael Jackson had
died; many Italians have clear memories of their country’s victory in the 2006
World Cup; and so on. (See Pillemer, 1984; Rubin & Kozin, 1984; also see
Weaver, 1993; Winograd & Neisser, 1993.)
Remarkably, though, these vivid, high-confidence memories can contain
substantial errors. Thus, when people say, “I’ll never forget that day . . .” they’re
sometimes wrong. For example, Hirst et al. (2009) interviewed more than
3,000 people soon after the September 11 attack on the World Trade Center,
asking how they first heard about the attack; who brought them the news; and
what they were doing at the time. When these individuals were re-interviewed
a year later, however, more than a third (37%) provided a substantially different account. Even so, the participants were strongly confident in their recollection (rating their degree of certainty, on a 1-to-5 scale, at an average of 4.4).
The outcome was the same for participants interviewed three years after the
attack — with 43% offering different accounts from those they had given initially. (For similar data, see Neisser & Harsch, 1992; also Hirst & Phelps,
2016; Rubin & Talarico, 2007; Schmidt, 2012; Talarico & Rubin, 2003.)
Other data, though, tell a different story, suggesting that some flashbulb
memories are entirely accurate. Why should this be? Why are some flashbulb
events remembered well, while others aren’t? The answer involves several
factors, including how, how often, and with whom someone discusses the flashbulb event. In many cases, this discussion may encourage people to “polish”
their reports — so that they’re offering their audience a “better,” more interesting narrative. After a few occasions of telling and re-telling this version of the
308 •
C H A P T E R E I G H T Remembering Complex Events
FLASHBULB MEMORIES
People often have especially clear and long-lasting memories for events like first hearing about Princess Diana’s
death in 1997, the attack on the World Trade Center in September 2001, or the news of Michael Jackson’s death in
2009. These memories — called “flashbulb memories” — are vivid and compelling, but they are not always accurate.
event, the new version may replace the original memory. (For more on these
issues, see Conway et al., 1994; Hirst et al., 2009; Luminet & Curci, 2009;
Neisser, Winograd, & Weldon, 1991; Palmer, Schreiber, & Fox, 1991; Tinti,
Schmidt, Sotgiu, Testa, & Curci, 2009; Tinti, Schmidt, Testa, & Levine, 2014.)
Notice, then, that an understanding of flashbulb memories requires us to
pay attention to the social aspects of remembering. In many cases, people
“share” memories with one another (and so, for example, I tell you about
my vacation, and you tell me about yours). Likewise, in the aftermath of
an important event, people often compare their recollections. (“Did you see
how he ran when the alarm sounded!?”) In all cases, people are likely to alter
their accounts in various ways, to allow for a better conversation. They may,
for example, leave out mundane bits, or add bits to make their account more
interesting or to impress their listeners. These new points about how the event
is described will, in turn, often alter the way the event is later remembered.
In addition, people sometimes “pick up” new information in these conversations — if, for example, someone who was present for the same event noticed
a detail that you missed. Often, this new information will be absorbed into
other witnesses’ memory — a pattern sometimes referred to as “co-witness
contamination.” Let’s note, though, that sometimes another person who witnessed the event will make a mistake in recalling what happened, and, after
conversation, other witnesses may absorb this mistaken bit into their own
recollection (Hope, Gabbert, & Fraser, 2013). In this way, conversations after
Autobiographical Memory
•
309
an event can sometimes have a positive impact on the accuracy and content
of a person’s eventual report, and sometimes a negative impact.
For all these reasons, then, it seems that “remembering” is not an activity
shaped only by the person who holds the memory, and exploring this point will
be an important focus for future research. (For early discussion of this broad
issue, see Bartlett, 1932. For more recent discussion, see Choi, Kensinger, &
Rajaram, 2017; Gabbert & Hope, 2013; Roediger & Abel, 2015.)
Returning to flashbulb memories, though, let’s not lose track of the fact
that the accuracy of these memories is uneven. Some flashbulb memories
are marvelously accurate; others are filled with error. Therefore, the commonsense idea that these memories are somehow “burned into the brain,”
and thus always reliable, is surely mistaken. In addition, let’s emphasize that
from the point of view of the person who has a flashbulb memory, there’s no
detectable difference between an accurate flashbulb memory and an inaccurate
one: Either one will be recalled with great detail and enormous confidence. In
each case, the memory can be intensely emotional. Apparently, memory errors
can occur even in the midst of our strongest, most vivid recollections.
Traumatic Memories
Flashbulb memories usually concern events that were strongly emotional.
Sadly, though, we can also find cases in which people experience truly extreme
emotion, and this leads us to ask: How are traumatic events remembered? If
someone has witnessed wartime atrocities, can we count on the accuracy of
their testimony in a war-crimes trial? If someone suffers through the horrors
of a sexual assault, will the painful memory eventually fade?
Evidence suggests that most traumatic events are well remembered for
many years. In fact, victims of atrocities often seem plagued by a cruel
enhancement of memory, leaving them with extra-vivid and long-lived recollections of the terrible event (e.g., Alexander et al., 2005; Goodman et al.,
2003; Peace & Porter, 2004; Porter & Peace, 2007; Thomsen & Berntsen,
2009). As a result, people who have experienced trauma sometimes complain
about having “too much” memory and wish they remembered less.
This enhanced memory can be understood in terms of a mechanism we’ve
already discussed: consolidation. This process is promoted by the conditions that accompany bodily arousal, including the extreme arousal typically
present in a traumatic event (Buchanan & Adolphs, 2004; Hamann, 2001;
McGaugh, 2015). But this doesn’t mean that traumatic events are always
well remembered. There are, in fact, cases in which people who’ve suffered
through extreme events have little or no recall of their experience (e.g.,
Arrigo & Pezdek, 1997). We can also sometimes document substantial errors
in someone’s recall of a traumatic event (Paz-Alonso & Goodman, 2008).
What factors are producing this mixed pattern? In some cases, traumatic
events are accompanied by sleep deprivation, head injuries, or substance
abuse, each of which can disrupt memory (McNally, 2003). In other cases,
the memory-promoting effects of arousal are offset by the complex memory effects of stress. The key here is that the experience of stress sets off a
310 •
C H A P T E R E I G H T Remembering Complex Events
cascade of biological reactions. These reactions produce changes throughout
the body, and the changes are generally beneficial, helping the organism to
survive the stressful event. However, the stress-produced changes are disruptive to some biological functions, and this can lead to a variety of problems
(including medical problems caused by stress).
How does the mix of stress reactions influence memory? The answer is
complicated. Stress experienced at the time of an event seems to enhance
memory for materials directly relevant to the source of the stress, but has
the opposite effect — undermining memory — for other aspects of the event
(Shields, Sazma, McCullough, & Yonelinas, 2017). Also, stress experienced
during memory retrieval interferes with memory, especially if the target information was itself emotionally charged.
How does all this play out in situations away from the laboratory? One
line of evidence comes from a study of soldiers who were undergoing survival training. As part of their training, the soldiers were deprived of sleep
and food, and they went through a highly realistic simulation of a prisonerof-war interrogation. One day later, the soldiers were asked to identify the
interrogator from a lineup. Despite the extensive (40-minute) face-to-face
encounter with the interrogator and the relatively short (one-day) retention
interval, many soldiers picked the wrong person from the lineup. Soldiers
who had experienced a moderate-stress interrogation picked the wrong
person from a live lineup 38% of the time; soldiers who had experienced a
high-stress interrogation (one that included a physical confrontation) picked
the wrong person 56% of the time if tested with a live lineup, and 68% of the
time if tested with a photographic lineup. (See Morgan et al., 2004; also see
Deffenbacher, Bornstein, Penrod, & McCorty, 2004; Hope, Lewinski, Dixon,
Blocksidge, & Gabbert, 2012; Valentine & Messout, 2008.)
Repression and “Recovered” Memories
Some authors argue in addition that people defend themselves against extremely painful memories by pushing these memories out of awareness. Some
writers suggest that the painful memories are “repressed”; others use the term
“dissociation” to describe this self-protective mechanism. No matter what
terms we use, the idea is that these painful memories (including, in many
cases, memories for childhood abuse) won’t be consciously available but will
still exist in a person’s long-term storage and in suitable circumstances can be
“recovered” — that is, made conscious again. (See, for discussion, Belli, 2012;
Freyd, 1996, 1998; Terr, 1991, 1994.)
Most memory researchers, however, are skeptical about this proposal. As
one consideration, painful events — including events that seem likely candidates
for repression — seem typically to be well remembered, and this is the opposite
of what we would expect if a self-protective mechanism was in place. In addition, some of the abuse memories reported as “recovered” may, in fact, have
been remembered all along, and so they provide no evidence of repression or
dissociation. In these cases, the memories had appeared to be “lost” because the
person refused to discuss these memories for many years; “recovery” of these
Autobiographical Memory
•
311
memories simply reflects the fact that the person is at last willing to talk about
them. This sort of “recovery” can be extremely consequential — emotionally
and legally — but doesn’t tell us anything about how memory works.
Sometimes, though, memories do seem to be genuinely lost for a while
and then recovered. But this pattern may not reveal the operation (and, eventually, the “lifting”) of repression or dissociation. Instead, this pattern may
be the result of retrieval failure — a mechanism that can “hide” memories
for periods of time, only to have them reemerge once a suitable retrieval
cue is available. Here, too, the recovery may be of enormous importance
for the person who is finally remembering the long-lost episodes; but again,
this merely confirms the role of an already-documented memory mechanism,
with no need for theorizing about repression.
In addition, we need to acknowledge the possibility that at least some
recovered memories may, in fact, be false memories. After all, we know that
false memories occur and that they’re more likely when someone is recalling
the distant past than when one is trying to remember recent events. It’s also
relevant that many recovered memories emerge only with the assistance of a
therapist who is genuinely convinced that a client’s psychological problems
stem from long-forgotten episodes of childhood abuse. Even if therapists
scrupulously avoid leading questions, their expectations might still lead them
to shape their clients’ memory in other ways — for example, by giving signs
of interest or concern if the clients hit on the “right” line of exploration, by
spending more time on topics related to the alleged memories, and so on. In
these ways, the climate within a therapeutic session could guide the client
toward finding exactly the “memories” the therapist expects to find.
Overall, then, the idea of a self-protective mechanism “hiding” painful
memories from view is highly controversial. Some psychologists (often, those
working in a mental health specialty) insist that they routinely observe this
sort of self-protection, and other psychologists (generally, memory researchers) reject the idea that memories can be hidden in this way. It does seem
clear, however, that at least some of these now-voiced memories are accurate and provide evidence for terrible crimes. As in all cases, though, the
veracity of recollection cannot be taken for granted. This warning is important in evaluating any memory, but especially so for anyone wrestling with
traumatic recollection. (For discussions of this difficult — and sometimes
angrily debated — issue, see, among others, Belli, 2012; Brewin &Andrews,
2014, 2016; Dalenberg et al., 2012; Geraerts et al., 2009; Giesbrecht,
Lynn, Lilienfeld, & Merckelbach, 2008; Kihlstrom, 2006; Küpper, Benoid,
Dalgleish, & Anderson, 2014; Loftus, 2017; Ost, 2013; Patihis, Lilienfeld,
Ho, & Loftus, 2014; Pezdek & Blandon-Gitlin, 2017.)
Long, Long-Term Remembering
In the laboratory, a researcher might ask you to recall a word list you read just
minutes ago or a film you saw a week ago. Away from the lab, however, people routinely try to remember events from years — perhaps decades — back.
312 •
C H A P T E R E I G H T Remembering Complex Events
We’ve mentioned that these longer retention intervals are generally associated with a greater amount of forgetting. But, impressively, memories from
long ago can sometimes turn out to be entirely accurate.
In an early study, Bahrick, Bahrick, and Wittlinger (1975; also Bahrick,
1984; Bahrick & Hall, 1991) tracked down the graduates of a particular high
school — people who had graduated in the previous year, and the year before,
and the year before that, and ultimately, people who had graduated 50 years
earlier. These alumni were shown photographs from their own year’s high
school yearbook, and for each photo they were given a group of names and
had to choose the name of the person shown in the picture. The data for this
“name-matching” task show remarkably little forgetting; performance was
approximately 90% correct if tested 3 months after graduation, the same
after 7 years, and the same after 14 years. In some versions of the test, performance was still excellent after 34 years (see Figure 8.11).
FIGURE 8.11
MEMORY OVER THE VERY LONG TERM
Percentage of correct answers
100
80
60
40
Name matching
20
25 yr 10 mo
34 yr 1 mo
47 yr 7 mo
14 yr 6 mo
7 yr 5 mo
3 yr 10 mo
1 yr 11 mo
9.3 mo
0
3.3 mo
Picture cueing
Time since graduation
When people were tested for how well they remembered names and faces of
their high school classmates, their memory was remarkably long-lasting. In
the name-matching task, participants were given a group of names and had
to choose the right one. In the picture-cueing task, participants had to come
up with the names on their own. In both tasks, the data show a sharp dropoff after 47 years, but it is unclear whether this reflects an erosion of memory
or a more general drop-off in performance caused by the normal process
of aging.
(after bahrick, bahrick, & wittlinger, 1975)
Autobiographical Memory
•
313
As a different example, what about the material you’re learning right
now? Five years from now, will you still remember what you’ve learned?
How about a decade from now? Conway, Cohen, and Stanhope (1991, 1992)
explored these questions, testing students’ retention of a cognitive psychology
course taken years earlier. The results echo the pattern we’ve already seen.
Some forgetting of names and specific concepts was observed during the first
3 years after the course. After the third year, however, performance stabilized,
so that students tested after 10 years still remembered a fair amount — in
fact, just as much as students tested after 3 years (see Figure 8.12).
In an earlier section, we argued that the retention interval is crucial for
memory and that memory gets worse as times goes by. The data now in front
of us, though, indicate that how much the interval matters — that is, how
quickly memories “fade” — may depend on how well established the memories were in the first place. The high school students in the Bahrick et al. study
had seen their classmates day after day, for (perhaps) several years. They
therefore knew their classmates’ names very, very well — and this is why the
passage of time had only a slight impact on their memories for the names.
Likewise, students in the Conway et al. study had apparently learned their
psychology quite well — and so they retained what they’d learned for a very
long time. In fact, we first met this study in Chapter 6, when we mentioned
that students’ grades in the course were good predictors of how much the
students would still remember many years after the course was done. Here,
too, the better the original learning, the slower the forgetting.
314 •
80
70
Concepts
60
Names
50
Chance
C H A P T E R E I G H T Remembering Complex Events
Retention interval Times
10 yr 5 mo
9 yr 5 mo
8 yr 5 mo
7 yr 5 mo
6 yr 5 mo
5 yr 5 mo
4 yr 5 mo
3 yr 5 mo
3 yr 3 mo
2 yr 3 mo
1 yr 3 mo
40
3 mo
Participants in this study were
quizzed about material they had
learned in a college course taken
as recently as three months ago
or as far back as a decade ago.
The data showed some forgetting,
but then performance leveled
off; memory seemed remarkably
stable from three years onward.
Note that in a recognition task,
memory is probed with “familiaror-not” questions, so someone
with no memory, responding at
random, would get 50% right just
by chance.
(after conway, cohen, & stanhope, 1991)
90
Mean percentage correctly recognized
FIGURE 8.12 LONG-TERM
RETENTION OF COURSE
MATERIALS
We can maintain our claim, therefore, that the passage of time is the
enemy of memory: Longer retention intervals produce lower levels of recall.
However, if the material is very well learned at the start, and also if you
periodically “revisit” the material, you can dramatically diminish the impact
of the passing years.
How General Are the Principles
of Memory?
TEST YOURSELF
12. W
hat is memory
consolidation?
13. What is a flashbulb
memory? Are
flashbulb memories
distinctive in how
accurate they seem
to be?
There is certainly more to be said about autobiographical memory. For
example, it can’t be surprising that people tend to remember significant
turning points in their lives and often use these turning points as a means
of organizing their autobiographical recall (Enz & Talarico, 2015; Rubin
& Umanath, 2015). Perhaps related, there are also memory patterns associated with someone’s age. Specifically, most people recall very little from
the early years of childhood (before age 3 or so; e.g., Akers et al., 2014;
Bauer, 2007; Hayne, 2004; Howe, Courage, & Rooksby, 2009; Morrison &
Conway, 2010). In contrast, people generally have clear and detailed memories of their late adolescence and early adulthood, a pattern known as the
“reminiscence bump.” (See Figure 8.13; Conway & Haque, 1999; Conway,
Wang, Hanyu, & Haque 2005; Dickson, Pillemer, & Bruehl, 2011; Koppel &
Rubin, 2016; Rathbone, Moulin, & Conway, 2008; Rathbone, O’Connor, &
Moulin, 2017.) As a result, for many Americans, the last years of high school
35
Percentage of Memories
30
Japan
China
Bangladesh
US
UK
All
25
20
15
10
5
0
5
10
15
20
25
30
35
40
45
Age at Encoding (in 5-year bins)
50
55
60
FIGURE 8.13 THE
LIFESPAN RETRIEVAL
CURVE
Most people have few memories
of their early childhood (roughly
from birth to age 3 or 4); this
pattern is referred to as “childhood amnesia.” In contrast, the
period from age 10 to 30 is well
remembered, producing a pattern called the “reminiscence
bump.” This “bump” has been
observed in multiple studies and
in diverse cultures; events from
this time in young adulthood are
often remembered in more detail
(although perhaps less accurately)
than more recent events.
How General Are the Principles of Memory?
•
315
and the years they spend in college are likely to be the most memorable
periods of their lives.
But in terms of the broader themes of this chapter, where does our brief
survey of autobiographical memory leave us? In many ways, this form
of memory is similar to other sorts of remembering. Autobiographical
memories can last for years and years, but so can memories that don’t refer
directly to your own life. Autobiographical remembering is far more likely
if the person occasionally revisits the target memories; these rehearsals dramatically reduce forgetting. But the same is true in non-autobiographical
remembering.
Autobiographical memory is also open to error, just as other forms of
remembering are. We saw this in cases of flashbulb memories that turn out
to be false. We’ve also seen that misinformation and leading questions can
plant false autobiographical memories — about birthday parties that never
happened and trips to the hospital that never took place (also see Brown
& Marsh, 2008). Misinformation can even reshape memories for traumatic
events, just as it can alter memories for trivial episodes in the laboratory
(Morgan, Southwick, Steffan, Hazlett, & Loftus, 2013; Paz-Alonso &
Goodman, 2008).
These facts strengthen a claim that has been emerging in our discussion
over the last three chapters: Certain principles seem to apply to memory in
general, no matter what is being remembered. All memories depend on connections. The connections promote retrieval. The connections also facilitate
interference, because they allow one memory to blur into another. The connections can fade with the passage of time, producing memory gaps, and
the gaps are likely to be filled via reconstruction based on generic knowledge. All these things seem to be true whether we’re talking about relatively
recent memories or memories from long ago, emotional memories or memories of calm events, memories for complex episodes or memories for simple
word lists.
But this doesn’t mean that all principles of memory apply to all types of
remembering. As we saw in Chapter 7, the rules that govern implicit memory
may be different from those that govern explicit memory. And as we’ve now
seen, some of the factors that play a large role in shaping autobiographical remembering (e.g., the role of emotion) may be irrelevant to other sorts
of memory.
In the end, therefore, our overall theory of memory is going to need
more than one level of description. We’ll need some principles that apply
to only certain types of memory (e.g., principles specifically aimed at emotional remembering). But we’ll also need broader principles, reflecting the
fact that some themes apply to memory of all sorts (e.g., the importance of
memory connections). As the last three chapters have shown, these more
general principles have moved us forward considerably in our understanding of memory in many different domains and have enabled us to illuminate many aspects of learning, of memory retrieval, and of the sources of
memory error.
316 •
C H A P T E R E I G H T Remembering Complex Events
COGNITIVE PSYCHOLOGY AND EDUCATION
remembering for the long term
Sometimes you need to recall things after a short delay — a friend tells you her
address and you drive to her apartment an hour later, or you study for a quiz
that you’ll take tomorrow morning. Sometimes, however, you want to remember things over a much longer time span — perhaps trying to recall things you
learned months or years ago. This longer-term retention is certainly important
in educational settings. Facts that you learn in high school may be crucial for
your professional work later in life. Likewise, facts that you learn in your first
year at college, or in your first year in a job, may be crucial in your third or
fourth year. How, therefore, can we help people to remember things for the
very long term?
The chapter has suggested a two-part answer to this question. First, you’re
more likely to hang on to material that you learned very well in the first
place. The chapter mentions one study in which people tried to recall the
material they’d learned in a college course a decade earlier. In that study, students’ grades in the course were good predictors of how much the students
would remember years after the course was done — and so, apparently, the
better the original learning, the slower the forgetting.
But long-term retention also depends on another factor — whether you
occasionally “revisit” the material you’ve learned. Even a brief refresher
can help enormously. In one study, students were quizzed on little factoids
they had most likely learned at some prior point in their lives (Berger, Hall,
& Bahrick, 1999) — for example, “Who was the first astronaut to walk
on the moon?”; “Who wrote the fable about the fox and the grapes?” In
many cases, the students knew these little facts but couldn’t recall them at
that moment. In that situation, the students were given a quick reminder.
The correct answer was shown to them for 5 seconds, with the simple
instruction that they should look at the answer because they would need
it later on.
Nine days after this reminder, participants were able to recall roughly
half the answers. This obviously wasn’t perfect performance, but it was an
enormous return (an improvement from 0% to 50%) from a very small
investment (5 seconds of “study time”). And it’s likely that a second reminder
a few days later, again lasting just 5 seconds, would have lifted their performance still further and allowed the participants to recall the items after an
even longer delay.
One suggestion, then, is that testing yourself (perhaps with flashcards — with
a cue on one side and an answer on the other) can be quite useful. Flashcards
are often a poor way to learn material, because (as we’ve seen) learning
requires thoughtful and meaningful engagement with the materials you’re
trying to memorize, and running through a stack of flash cards probably
won’t promote that thoughtful engagement. But using flashcards may be an
Cognitive Psychology and Education
•
317
AIDS FOR STUDENTS?
Memory research provides
power­ful lessons for students
hoping to retain what they
are learning in their courses.
excellent way to review material that is already learned — and so a way to
avoid forgetting this material.
Other, more substantial, forms of testing can also be valuable. Think about
what happens each time you take a vocabulary quiz in your Spanish class.
A question like “What’s the Spanish word for ‘bed’?” gives you practice in
retrieving the word, and that practice promotes fluency in retrieval. In addition,
seeing the word (cama) can itself refresh the memory, promoting retention.
The key idea here is the “testing effect.” This term refers to a consistent
pattern in which students who have taken a test have better retention later
on, in comparison to students who didn’t take the initial test. (See, e.g.,
Carpenter, Pashler, & Cepeda, 2009; Glass & Sinha, 2013; Halamish &
Bjork, 2011; Karpicke, 2012; McDermott, Agarwal, D’Antonio, Roediger, &
McDaniel, 2014; Pyc & Rawson, 2012.) This pattern has been documented
with students of various ages (including high school and college students)
and with different sorts of material.
The implications for students should be clear. It really does pay to go
back periodically and review what you’ve learned — including material
you learned earlier this academic year as well as material from previous
years. The review doesn’t have to be lengthy or intense; in the first study
described here, just a 5-second exposure was enough to decrease forgetting
dramatically.
Finally, you shouldn’t complain if a teacher insists on giving frequent
quizzes. Of course, quizzes can be a nuisance, but they serve two functions. First, they can help you assess your learning, so that you can judge
whether — perhaps — you need to adjust your study strategies. Second, the
quizzes actually help you retain what you’ve learned — for days, and probably months, and perhaps even decades after you’ve learned it.
318 •
C H A P T E R E I G H T Remembering Complex Events
For more on this topic . . .
Brown, P. C., Roediger, H. L., & McDaniel, M. A. (2014). Make it stick: The science
of successful learning. New York, NY: Belknap Press.
Putnam, A. L., Nestojko, J. F., & Roediger, H. L. (2016). Improving student learning: Two strategies to make it stick. In J. C. Horvath, J. Lodge, & J. A. C.
Hattie (Eds.), From the laboratory to the classroom: Translating the science
of learning for teachers (pp. 94–121). Oxford, UK: Routledge.
Putnam, A. L., Sungkhasettee, V., & Roediger, H. L. (2016). Optimizing learning
in college: Tips from cognitive psychology. Perspectives on Psychological
Science, 11(5), 652–660.
Cognitive Psychology and Education
•
319
chapter review
SUMMARY
• Memory is usually accurate, but errors do occur
and can be quite significant. In general, these errors
are produced by the connections that link memories to one another and link memories for specific
episodes to other, more general knowledge. These
connections help you because they serve as retrieval
paths. But the connections can also “knit” separate
memories together, making it difficult to keep track
of which elements belong in which memory.
• Some memory errors arise from your understanding of an episode. The understanding promotes
memory for the episode’s gist but also encourages
memory errors. A similar pattern emerges in the
DRM procedure, in which a word related to other
words on a list is (incorrectly) recalled as being part
of the list. Closely related effects arise from schematic knowledge. This knowledge helps you understand an episode, but at the same time a reliance
on schematic knowledge can lead you to remember
an episode as being more “regular,” more “normal,”
than it actually was.
• Memory errors can also arise through the misinformation effect, in which people are exposed to some
(false) suggestion about a previous event. Such suggestions can easily change the details of how an event
is remembered and can, in some cases, plant memories for entire episodes that never occurred at all.
• People seem genuinely unable to distinguish their
accurate memories from their inaccurate ones. This
is because false memories can be recalled with just as
much detail, emotion, and confidence as historically
accurate memories. The absence of a connection
between memory accuracy and memory confidence contrasts with the commonsense belief that
you should rely on someone’s degree of certainty in
assessing their memory. The problem in this commonsense belief lies in the fact that confidence is
influenced by factors (such as feedback) that have no
320
impact on accuracy, and this influence undermines
the linkage between accuracy and confidence.
• While memory errors are easily documented,
cases of accurate remembering can also be observed,
and they are probably more numerous than cases
involving memory error. Memory errors are more
likely, though, in recalling distant events rather than
recent ones. One reason is decay of the relevant
memories; another reason is retrieval failure. Retrieval failure can be either complete or partial; the
tip-of-the-tongue pattern provides a clear example
of partial retrieval failure. Perhaps the most important source of forgetting, though, is interference.
• People have sought various ways of undoing forgetting, including hypnosis and certain drugs. These
approaches, however, seem ineffective. Forgetting can
be diminished, though, through procedures that provide a rich variety of retrieval cues, and it can be avoided
through occasional revisits to the target material.
• Although memory errors are troubling, they may
be the price you pay in order to obtain other advantages. For example, many errors result from the
dense network of connections that link your various
memories. These connections sometimes make it
difficult to recall which elements occurred in which
setting, but the same connections serve as retrieval
paths — and without those connections, you might
have great difficulty in locating your memories in
long-term storage.
• Autobiographical memory is influenced by the
same principles as any other form of memory, but
it is also shaped by its own set of factors. For example, episodes connected to the self are, in general,
better remembered — a pattern known as the “selfreference effect.”
• Autobiographical memories are often emotional,
and this has multiple effects on memory. Emotion
seems to promote memory consolidation, but it may
also produce a pattern of memory narrowing. Some
emotional events give rise to very clear, long-lasting
flashbulb memories. Despite their subjective clarity, these memories can contain errors and in some
cases can be entirely inaccurate. At the extreme
of emotion, trauma has mixed effects on memory.
Some traumatic events are not remembered, but
most traumatic events seem to be remembered for a
long time and in great detail.
• Some events can be recalled even after many
years have passed. In some cases, this is because the
knowledge was learned very well in the first place.
In other cases, occasional rehearsals preserve a
memory for a very long time.
KEY TERMS
intrusion errors (p. 283)
DRM procedure (p. 285)
schema (plural: schemata) (p. 286)
misinformation effect (p. 291)
retention interval (p. 298)
decay theory of forgetting (p. 299)
interference theory (p. 299)
retrieval failure (p. 299)
TOT phenomenon (p. 299)
autobiographical memory (p. 304)
self-schema (p. 304)
consolidation (p. 306)
flashbulb memories (p. 308)
TEST YOURSELF AGAIN
1. What is the evidence that in some circumstances many people will misremember significant events they have experienced?
2. What is the evidence that in some circumstances
people will even misremember recent events?
3. What is the evidence that your understanding
of an episode can produce intrusion errors?
4. What is the DRM procedure, and what results
does this procedure produce?
5. What is schematic knowledge, and what evidence tells us that schematic knowledge can help
us — and also can undermine memory accuracy?
6. What is the misinformation effect? Describe
three different procedures that can produce this
effect.
7. Some people insist that our memories are
consistently accurate in remembering the gist,
or overall content, of an event; when we make
memory errors, they claim, we make mistakes
only about the details within an event. What
evidence allows us to reject this claim?
8. What factors seem to undermine the relationship
between your degree of certainty in a memory
and the likelihood that the memory is accurate?
9. Explain the mechanisms hypothesized by each
of the three major theories of forgetting: decay,
interference, and retrieval failure.
10. What techniques or procedures seem ineffective as a means of “un-doing” forgetting? What
techniques or procedures seem to diminish or
avoid forgetting?
321
11. Explain why the mechanisms that produce
memory errors may actually be mechanisms
that help us in important ways.
13. What is a flashbulb memory? Are flashbulb
memories distinctive in how accurate they seem
to be?
12. What is memory consolidation?
THINK ABOUT IT
1.People sometimes compare the human eye
to a camera, and compare human memory
to a video recorder (like TiVo or the video
recorder on your smartphone). Ironically,
though, there are important ways in which
your memory is worse than a video recorder,
and also important ways in which it’s far better
than a video recorder. Describe both versions
of this comparison — the ways in which video
recorders are superior, and the ways in which
your memory is superior.
E eBOOK DEMONSTRATIONS & ESSAYS
Go to http://digital.wwnorton.com/cognition7 for the online demonstrations and essays
relevant to this chapter. These can be found in the ebook.
Online Demonstrations
Online Applying Cognitive Psychology and the
Law Essays
• Demonstration 8.1: Associations and Memory
• Cognitive Psychology and the Law: Jurors’
Error
Memory
• Demonstration 8.2: Memory Accuracy and
Confidence
• D
emonstration 8.3: The Tip-of-the-Tongue Effect
• Demonstration 8.4: Childhood Amnesia
COGNITION LABS
Go to http://digital.wwnorton.com/cognition7 to experience interactive online psychology
labs relevant to this chapter.
322
Knowledge
4
part
I
n Parts 2 and 3, we saw case after case in which your interactions with
the world are guided by knowledge. In perceiving, for example, you make
inferences guided by knowledge about the world’s regular patterns. In
attending, you anticipate inputs guided by your knowledge about what’s likely
to occur. In learning, you connect new information to things you already know.
But what is knowledge? How is it represented in your mind? How do you locate
knowledge in memory when you need it? We’ve already taken some steps toward
answering these questions — by arguing that knowledge is represented in the mind
by means of a network of interconnected nodes. In this section, we’ll expand this
proposal in important ways. In Chapter 9, we’ll describe the basic building blocks
of knowledge — individual concepts — and consider several hypotheses about how
concepts are represented in the mind. Because each hypothesis captures a part of
the truth, we’ll be driven toward a several-part theory that combines the various
views. We’ll also see that knowledge about individual concepts depends on linkages
to other, related concepts. For example, you can’t know what a “dog” is without also
understanding what an “animal” is, what a “living thing” is, and so on. As a result,
connections among ideas will be crucial here, just as they were in previous chapters.
Chapters 10 and 11 then focus on two special types of knowledge: knowledge about language and knowledge about visual images. In Chapter 10, we’ll
see that your knowledge of language is highly creative, allowing you to produce
new words and new sentences that no one has ever used before. But at the
same time, the creativity is constrained, so there are some words and sequences
of words that are considered unacceptable by virtually any language-user. In
order to understand this pattern of “constrained creativity,” we’ll consider the
possibility that language knowledge involves abstract rules that are, in some
way, honored by every user of the language.
In Chapter 11, we’ll see that mental images involve representations that are
distinct from those involved in other forms of knowledge, but we’ll also consider some of the ways in which memory for visual appearances is governed by
the same principles as other forms of knowledge.
323
9
chapter
Concepts and
Generic Knowledge
what if…
In Chapter 8, we mentioned people who have superior autobiographical recall. It’s remarkable how much
these individuals can remember — but some people, it turns out, remember
even more. One might say that these people have “perfect memories,”
but this terminology would be misleading.
We begin with a work of fiction. In a wonderful short story titled
“Funes the Memorious,” the Argentine writer Jorge Luis Borges
describes a character — Funes — who never forgets anything. But rather
than being proud of this capacity, Funes is immensely distressed by his
memorial prowess: “My memory, sir, is like a garbage heap” (p. 152).
Among other problems, Funes complains that he’s incapable of thinking in general terms. He remembers so much about how individuals differ
that he has a hard time focusing on what they might have in common:
“Not only was it difficult for him to comprehend that the generic symbol
dog embraces so many unlike individuals of diverse size and form; it bothered him that the dog at 3:14 (seen from the side) should have the same
name as the dog at 3:15 (seen from the front)” (Borges, 1964, p. 153).
Funes is a fictional character, but consider the actual case of Solomon
Shereshevsky (Luria, 1968). Shereshevsky, like Funes, never forgot anything. After hearing a lengthy speech, he could repeat it back word for
word. If shown a complex mathematical formula (even one that had
no meaning for him), he could reproduce it perfectly months later. He
effortlessly memorized poems written in languages he didn’t understand. And Shereshevsky’s flawless retention wasn’t the result of some
deliberate trick or strategy. Just the opposite: Shereshevsky seemed to
have no choice about his level of recall.
Like Funes, Shereshevsky wasn’t well served by his extraordinary
memory. He was so alert to the literal form of his experiences that he
couldn’t remember their deeper implications. Similarly, he had difficulty
recognizing faces because he was so alert to the changes in a face from
one view to the next. And, like Funes, Shereshevsky was often distracted
by the detail of his own recollections, so he found it difficult to think in
abstract terms.
There are, of course, settings in which you do want to remember the
specific episodes of your life. You want to recall what you saw at a crime
scene, holding to the side (as best you can) the information you picked
up later in your conversation with the police. You hope to remember
325
preview of chapter themes
•
asic concepts — like “chair” and “dog” — are the building
B
blocks of all knowledge. However, attempts at defining
these concepts usually fail because we easily find exceptions to any definition that might be proposed.
•
hese beliefs may be represented in the mind as proposiT
tions encoded in a network structure. Alternatively, they
may be represented in a distributed form in a connectionist network.
•
his leads to a suggestion that knowledge of these conT
cepts is cast in terms of probabilities — so that a creature
that has wings and feathers, and that flies and lays eggs, is
probably a bird.
•
•
any results are consistent with this probabilistic idea and
M
show that the more a test case resembles the “prototype”
for a category, the more likely people are to judge the case
as being in that category.
e are driven, therefore, to a multipart theory of conW
cepts. Your conceptual knowledge likely includes a
prototype for each category and also a set of remembered exemplars. But you also seem to have a broad set
of beliefs about each concept — beliefs that provide a
“theory” for why the concept takes the form it does, and
you use this theory in a wide range of judgments about
the concept.
•
ther results, however, indicate that conceptual knowlO
edge includes other beliefs — beliefs that link a concept to
other concepts and also specify why the concept is as it is.
what you read in your textbook, trying to ignore the (possibly bogus)
information you heard from your roommate. Funes and Shereshevsky
obviously excel in this type of memory, but their limitations remind us
that there are also disadvantages for this type of particularized recall. In
many settings, you want to set aside the details of this or that episode
and, instead, weave your experiences together so that you can pool
information received from various sources. This allows you to create a
more complete, more integrated type of knowledge — one that allows
you to think about dogs in general rather than focusing on this view of
that dog; or one that helps you remember what your friend’s face generally looks like rather than what she looked like, say, yesterday at 1:42 in
the afternoon. This more general type of knowledge is surely drawn from
your day-to-day experience, but it is somehow abstracted away from
that experience. What is this more general type of knowledge?
Understanding Concepts
Imagine that a friend approaches you and boasts that he knows what a
“spoon” is or what a “shoe” is. You’d probably be impressed by your
friend’s foolishness, not by his knowledge. After all, concepts like these are so
ordinary, so straightforward, that there seems to be nothing special about
knowing — and being able to think about — these simple ideas.
However, ordinary concepts like these are the building blocks out of
which all knowledge is created, and as we’ve seen in previous chapters, you
depend on your knowledge in many aspects of day-to-day functioning. Thus,
you know what to pay attention to in a restaurant because you understand
326 •
C H A P T E R N I N E Concepts and Generic Knowledge
the basic concept of “restaurant.” You’re able to understand a simple story
about a child checking her piggy bank because you understand the concepts
of “money,” “shopping,” and so on.
The idea, then, is that you need concepts in order to have knowledge, and
you need knowledge in order to function. In this way, your understanding of
ideas like “spoon” and “shoe” might seem commonplace, but it is an ingredient without which cognition cannot proceed.
But what exactly does it mean to understand concepts like these? How
is this knowledge represented in the mind? In this chapter, we’ll begin with
the hypothesis that understanding a concept is like knowing a dictionary
definition — and so, if someone knows what a “house” is, or a “taxi,” he or
she can offer something like a definition for these terms — and likewise for
all the other concepts in each person’s knowledge base. As we’ll see, though,
this hypothesis quickly runs into problems, so we’ll need to turn to a more
complicated proposal.
Definitions: What Is a “Dog”?
You know perfectly well what a dog is. But what is it that you know? One possibility is that your knowledge is somehow akin to a dictionary definition —
that is, what you know is something like: “A dog is a creature that (a) is an
animal, (b) has four legs, (c) barks, (d) wags its tail.” You could then use this
definition in straightforward ways: When asked whether a candidate creature
is a dog, you could use the definition as a checklist, scrutinizing the candidate
for the various defining features. When told that “a dog is an animal,” you
would know that you hadn’t learned anything new, because this information
is already contained within the definition. If you were asked what dogs, cats,
and horses have in common, you could scan your definition of each one looking for common elements.
This proposal is correct in some cases, and so, for example, you certainly
know definitions for concepts like “triangle” and “even number.” But what about
more commonplace concepts? The concern here was brought to light by the
20th-century philosopher Ludwig Wittgenstein, who argued (e.g., Wittgenstein,
1953) that the simple terms we all use every day actually don’t have definitions.
For example, consider the word “game.” You know this word and can use it
sensibly, but what is a game? As an approach to this question, we could ask,
for example, about the game of hide-and-seek. What makes hide-and-seek a
“game”? Hide-and-seek (a) is an activity most often practiced by children, (b) is
engaged in for fun, (c) has certain rules, (d) involves several people, (e) is in some
ways competitive, and (f) is played during periods of leisure. All these are plausible attributes of games, and so we seem well on our way to defining “game.”
But are these attributes really part of the definition of “game”? What about the
Olympic Games? The competitors in these games aren’t children, and runners in
marathon races don’t look like they’re having a lot of fun. Likewise, what about
card games played by one person? These are played alone, without competition.
For that matter, what about the case of professional golfers?
Understanding Concepts
•
327
THE HUNT FOR DEFINITIONS
It is remarkably difficult to define even very familiar terms. For example, what is a “dog”?
Most people include “has fur” in the definition, but what about the hairless Chihuahua?
Many people include “communicates by barking” in the definition, but what about the
Basenji (one of which is shown here) — a breed of dog that doesn’t bark?
TEST YOURSELF
1.Consider the word
“chair.” Name some
attributes that might
plausibly be included
in a definition for this
word. But, then, can
you describe objects
that you would count
as chairs even though
they don’t have one
or more of these
attributes?
2.What does it mean to
say that there is a “family
resemblance” among the
various animals that we
call “dogs”?
328 •
It seems that for each clause of the definition, we can find an exception —
an activity that we call a “game” but that doesn’t have the relevant characteristic. And the same is true for almost any concept. We might define “shoe” as
an item of apparel made out of leather, designed to be worn on the foot.
But what about wooden shoes? What about a shoe designed by a master
shoemaker, intended only for display? What about a shoe filled with cement,
which therefore can’t be worn? Similarly, we might define “dog” in a way
that includes four-leggedness, but what about a dog that has lost a limb in
some accident? We might specify “communicates by barking” as part of the
definition of dog, but what about the African Basenji, which has no bark?
Family Resemblance
It seems, then, we can’t say things like “A dog is a creature that has fur and
four legs and barks.” That’s because we easily find exceptions to this rule
(a hairless Chihuahua; a three-legged dog; the barkless Basenji). But surely
we can say, “Dogs usually are creatures that have fur, four legs, and bark, and
a creature without these features is unlikely to be a dog.” This probabilistic
phrasing preserves what’s good about definitions — the fact that they do name
relevant features, shared by most members of the category. But this phrasing
also allows a degree of uncertainty, some number of exceptions to the rule.
C H A P T E R N I N E Concepts and Generic Knowledge
In a similar spirit, Wittgenstein proposed that members of a category have
a family resemblance to one another. To understand this term, think about
an actual family — your own, perhaps. There are probably no “defining
features” for your family — features that every family member has. Nonetheless, there are features that are common in the family, and so, if we consider
family members two or three at a time, we can usually find shared attributes.
For example, you, your brother, and your mother might all have the family’s
beautiful red hair and the same wide lips; as a result, you three look alike to
some extent. Your sister, however, doesn’t have these features. But she’s still
recognizable as a member of the family because (like you and your father)
she has the family’s typical eye shape and the family’s distinctive chin. In this
way, the common features in the family depend on what “subgroup” you’re
considering — hair color shared for these family members; eye shape shared
by those family members; and so on.
One way to think about this pattern is by imagining the “ideal” for each
family — someone who has all of the family’s features. (In our example,
this would be a wide-lipped redhead with the right eye and chin shapes.) In
many families, this person may not exist, so perhaps there’s nobody who has
every one of the family’s distinctive features — and so no one who looks like
the “perfect Jones” (or the “perfect Martinez” or the “perfect Goldberg”).
Nonetheless, each member of the family shares at least some features with
this ideal — and therefore has some features in common with other family members. This feature overlap is why the family members resemble one
another, and it’s how we manage to recognize these individuals as all belonging to the same family.
Wittgenstein proposed that ordinary categories like “dog” or “game” or
“furniture” work in the same way. There may be no features that are shared
by all dogs or all games. Even so, we can identify “characteristic features” for
each category — features that many (perhaps most) category members have.
These are the features that enable you to recognize that a dog is a dog, a game
is a game, and so on.
There are several ways we might translate all these points into a psychological theory, but one influential translation was proposed by psychologist Eleanor Rosch in the mid-1970s (Rosch, 1973, 1978; Rosch & Mervis,
1975; Rosch, Mervis, Gray, Johnson, & Boyes-Braem, 1976). Let’s look at
her model.
TYPICALITY IN
FAMILIES
In the Smith family, many (but
not all) of the brothers have
dark hair, so dark hair is typical for the family (i.e., is found
in many family members)
but doesn’t define the family
(i.e., is not found in all family
members). Likewise, wearing
glasses is typical for the family but not a defining feature;
so is having a mustache and a
big nose. Many concepts have
the same character — with
many features shared among
the instances of the concept,
but no features shared by all
of the instances.
Prototypes and Typicality Effects
One way to think about definitions is that they set the “boundaries” for a
category. If a test case has certain attributes, then it’s “inside” the boundaries. If a test case doesn’t have the defining attributes, then it’s “outside” the
category. Prototype theory, in contrast, begins with a different tactic: Perhaps
the best way to identify a category is to specify the “center” of the category,
rather than the boundaries. Just as we spoke earlier about the “ideal” family
member, perhaps the concept of “dog” is represented in the mind by some
Prototypes and Typicality Effects
•
329
depiction of the “ideal” dog, and all judgments about dogs are made with
reference to this ideal. Likewise for “bird” or “house” or any other concept
in your repertoire — in each case, the concept is represented by the appropriate prototype.
In most cases, this “ideal” — the prototype — will be an average of the
various category members you’ve encountered. So, for example, the prototype dog will be the average color of the dogs you’ve seen, the average size
of the dogs you’ve seen, and so forth. (Notice, then, that different people,
each with their own experiences, will have slightly different prototypes.)
No matter what the specifics of the prototype, though, you’ll use this
“ideal” as the benchmark for your conceptual knowledge. Thus, whenever
you use your conceptual knowledge, your reasoning is done with reference
to the prototype.
Prototypes and Graded Membership
To make these ideas concrete, imagine that you’re trying to decide whether
a creature currently before your eyes is or is not a dog. In making this
decision, you’ll compare the creature with the prototype in your memory.
If there’s no similarity, the creature standing before you is probably not
in the category; if there’s considerable similarity, you draw the opposite
conclusion.
This sounds plausible enough, but note an important implication.
Membership in a category depends on resemblance to the prototype, and
resemblance is a matter of degree. (After all, some dogs are likely to resemble
the prototype closely, while others will have less in common with this ideal.)
As a result, membership in the category isn’t a simple “yes or no” decision;
CATEGORIES HAVE PROTOTYPES
As the text describes, people seem to have a prototype
in their minds for a category like “dog.” For many people,
the German shepherd shown here is close to that prototype, and the other dogs depicted are more distant from
the prototype.
330 •
C H A P T E R N I N E Concepts and Generic Knowledge
instead, it’s a matter of “more” or “less.” In technical terms, we’d say that
categories, on this view, have a graded membership, such that objects closer
to the prototype are “better” members of the category than objects farther
from the prototype. Basically, the idea is that some dogs are “doggier” than
others, some books “bookier” than others, and so on for all the other categories you can think of.
Testing the Prototype Notion
This proposal — that mental categories have a graded membership — was
tested in a series of experiments conducted years ago. For example, in
classic studies using a sentence verification task, research participants were
presented with a series of sentences, and their job was to indicate (by pressing the appropriate button) whether each sentence was true or false. In this
procedure, participants’ responses were slower for sentences like “A penguin
is a bird” than for sentences like “A robin is a bird”; slower for “An Afghan
hound is a dog” than for “A German shepherd is a dog” (Smith, Rips, &
Shoben, 1974).
Why should this be? According to a prototype perspective, participants
chose their response (“true” or “false”) by comparing the thing mentioned
(e.g., penguin) to their prototype for that category (i.e., their bird prototype).
When there was close similarity between the test case and the prototype,
participants could make their decisions quickly; in contrast, judgments about
items distant from the prototype took more time. And given the results, it
seems that penguins and Afghans are more distant from their respective prototypes than are robins and German shepherds.
Other early results can also be understood in these terms. For example, in a production task we simply ask people to name as many birds or
dogs as they can. According to a prototype view, they’ll do this task by first
locating their bird or dog prototype in memory and then asking themselves
what resembles this prototype. In essence, they’ll start with the center of the
category (the prototype) and work their way outward from there. So birds
close to the prototype should be mentioned first; birds farther from the prototype, later on.
By this logic, the first birds mentioned in the production task should
be the birds that yielded fast response times in the verification task; that’s
because what matters in both tasks is proximity to the prototype. Likewise, the birds mentioned later in production should have yielded slower
response times in verification. This is exactly what happened (Mervis,
Catlin, & Rosch, 1976).
In fact, this outcome sets the pattern of evidence for prototype theory.
Over and over, in category after category, members of a category that are
“privileged” on one task (e.g., they yield the fastest response times) turn out
also to be privileged on other tasks (e.g., they’re most likely to be mentioned).
As another illustration of this pattern, consider the data from rating tasks. In
these tasks, participants are given instructions like these: “We all know that
Prototypes and Typicality Effects
•
331
some birds are ‘birdier’ than others, some dogs are ‘doggier’ than others, and
so on. I’m going to present you with a list of birds or of dogs, and I want you
to rate each one on the basis of how ‘birdy’ or ‘doggy’ it is” (Rosch, 1975;
also Malt & Smith, 1984).
People are easily able to make these judgments, and quite consistently they
rate items as being very “birdy” or “doggy” when these instances are close to
the prototype (as determined in the other tasks). They rate items as being less
“birdy” or “doggy” when these are farther from the prototype. This finding
suggests that once again, people perform the task by comparing the test item
to the prototype (see Table 9.1).
Basic-Level Categories
It does seem, then, that certain category members are “privileged,” just as the
prototype theory proposes. It turns out, also, that certain types of category are
privileged — in their structure and the way they’re used. For example, imagine
TABLE 9.1
ARTICIPANTS’ TYPICALITY RATINGS FOR THE
P
CATEGORY “FRUIT” AND THE CATEGORY “BIRD”
Fruit
Rating
Bird
Rating
Apple
6.25
Robin
6.89
Peach
5.81
Bluebird
6.42
Pear
5.25
Seagull
6.26
Grape
5.13
Swallow
6.16
Strawberry
5.00
Falcon
5.74
Lemon
4.86
Mockingbird
5.47
Blueberry
4.56
Starling
5.16
Watermelon
4.06
Owl
5.00
Raisin
3.75
Vulture
4.84
Fig
3.38
Sandpiper
4.47
Coconut
3.06
Chicken
3.95
Pomegranate
2.50
Flamingo
3.37
Avocado
2.38
Albatross
3.32
Pumpkin
2.31
Penguin
2.63
Olive
2.25
Bat
1.53
Ratings were made on a 7-point scale, with 7 corresponding to the highest typicality.
Note also that the least “birdy” of the birds isn’t (technically speaking) a bird at all!
( after malt & smith , 1984)
332 •
C H A P T E R N I N E Concepts and Generic Knowledge
FIGURE 9.1 BASIC VERSUS
SUPERORDINATE LABELING
What is this? The odds are good that
you would answer, “It’s a chair,” using
the basic-level description rather than
the more general label (“It’s a piece of
furniture”) or a more specific description
(“It’s an upholstered armchair”) — even
though these other descriptions would
certainly be correct.
that we show you a picture like the one in Figure 9.1 and ask, “What is this?”
You’re likely to say “a chair” and unlikely to offer a more specific response
(“upholstered armchair”) or a more general one (“an item of furniture”).
Likewise, we might ask, “How do people get to work?” In responding, you’re
unlikely to say, “Some people drive Fords; some drive Toyotas.” Instead, your
answer is likely to use more general terms, such as “cars,” “trains,” and “buses.”
In keeping with these observations, Rosch and others have argued that
there is a “natural” level of categorization, neither too specific nor too general,
that people tend to use in their conversations and their reasoning. The special
status of this basic-level categorization can be demonstrated in many ways.
Basic-level categories are usually represented in our language via a single
word, while more specific categories are identified with a phrase. Thus,
“chair” is a basic-level category, and so is “apple.” The more specific (subordinate) categories of “lawn chair” or “kitchen chair” aren’t basic level; neither is
“Granny Smith apple” or “Golden Delicious apple.”
We’ve already suggested that if you’re asked to describe an object, you’re
likely to use the basic-level term. In addition, if asked to explain what members of a category have in common with one another, you have an easy time
with basic-level categories (“What do all chairs have in common?”) but some
difficulty with more inclusive (superordinate) categories (“What does all furniture have in common?”). Moreover, children learning to talk often acquire
basic-level terms earlier than either the more specific subcategories or the more
general, more encompassing categories. In these (and other) ways, basic-level
categories do seem to reflect a natural way to categorize the objects in our
world. (For more on basic-level categories, see Corter & Gluck, 1992; Murphy,
2016; Pansky & Koriat, 2004; Rogers & Patterson, 2007; Rosch et al., 1976.)
TEST YOURSELF
3.Why is graded membership a consequence
of representing the
category in terms of a
prototype?
4.What tasks show us that
concept judgments often
rely on prototypes and
typicality?
5.Give an example of a
basic-level category,
and then name some
of the subcategories
within this basic-level
grouping.
Prototypes and Typicality Effects
•
333
Exemplars
Let’s return, though, to our main agenda. As we’ve seen, a broad range of
tasks reflects the graded membership of mental categories. In other words,
some members of the categories are “better” than others, and the better members are recognized more readily, mentioned more often, judged to be more
typical, and so on. (For yet another way you’re influenced by typicality, see
Figure 9.2.) All of this fits well with the idea that conceptual knowledge is
represented via a prototype and that we categorize by making comparisons
to that prototype. It turns out, though, that your knowledge about “birds”
and “fruits” and “shoes” and so on also includes another element.
FIGURE 9.2
TYPICALITY AND ATTRACTIVENESS
Typicality influences many judgments about category members, including attractiveness. Which of these pictures
shows the most attractive-looking fish? Which one shows the least attractive-looking? In several studies, participants’ ratings of attractiveness have been closely related to (other participants’) ratings of typicality — so that
people seem to find more-typical category members to be more attractive (e.g., Halberstadt & Rhodes, 2003).
334 •
C H A P T E R N I N E Concepts and Generic Knowledge
Analogies from Remembered Exemplars
Imagine that we place a wooden object in front of you and ask, “Is this
a chair?” According to the prototype view, you’ll answer by calling up
your chair prototype from memory and then comparing the candidate
to that prototype. If the resemblance is great, you’ll announce, “Yes, this
is a chair.”
But you might make this decision in a different way. You might notice that
the object is very similar to an object in your Uncle Jerry’s living room, and
you know that Uncle Jerry’s object is a chair. (After all, you’ve seen Uncle
Jerry sitting in the thing, reading his newspaper; you’ve heard Jerry referring
to the thing as “my chair,” and so on.) These points allow an easy inference:
If the new object resembles Jerry’s, and if Jerry’s object is a chair, then it’s a
safe bet that the new object is a chair too.
The idea here is that in some cases categorization relies on knowledge
about specific category members (e.g., “Jerry’s chair”) rather than the prototype (e.g., the ideal chair). This process is referred to as exemplar-based
reasoning, with an exemplar defined as a specific remembered instance — in
essence, an example.
The exemplar-based approach is in many ways similar to the prototype view. According to each of these proposals, you categorize objects by
comparing them to a mentally represented “standard.” The difference
between the views lies in what that standard is. For prototype theory,
the standard is the prototype — an average representing the entire category; for exemplar theory, the standard is provided by whatever example of the category comes to mind (and different examples may come to
mind on different occasions). In either case, the process is then the same.
You assess the similarity between a candidate object and the standard. If
the resemblance is great, you judge the candidate as being within the
relevant category; if the resemblance is minimal, you seek some alternative
categorization.
A Combination of Exemplars and Prototypes
There is, in fact, reason for you to rely on prototypes and on exemplars in
your thinking about categories. Prototypes provide an economical representation of what’s typical for a category, and there are many circumstances in
which this quick summary is useful. But exemplars, for their part, provide
information that’s lost from the prototype — including information about the
variability within the category.
To see how this matters, consider the fact that people routinely “tune”
their concepts to match the circumstances. For example, they think about
birds differently when considering Chinese birds than when thinking about
American birds; they think about gifts differently when considering gifts for
a student rather than gifts for a faculty member (Barsalou, 1988; Barsalou
& Sewell, 1985). In fact, people can adjust their categories in fairly precise
Exemplars
•
335
ways: not just “gift,” but “gift for a 4-year-old” or “gift for a 4-year-old
who recently broke her wrist” or “gift for a 4-year-old who likes sports but
recently broke her wrist.” This pliability in concepts is easy to understand if
people are relying on exemplars; after all, different settings, or different perspectives, would trigger different memories and so bring different exemplars
to mind.
It’s useful, then, that conceptual knowledge includes prototypes and
exemplars, because each has its own advantages. However, the mix of
exemplar and prototype knowledge may vary from person to person and
from concept to concept. One person might have extensive knowledge about
individual horses, so she has many exemplars in memory; the same person
might have only general information (a prototype, perhaps) about snowmobiles. Some other person might show the reverse pattern. And for all people,
the pattern of knowledge might depend on the size of the category and how
easily confused the category memories are with one another — with exemplars
being used when the individuals are more distinct. (For further discussion,
see Murphy, 2016; Rips, Smith, & Medin, 2012; Rouder & Ratcliff, 2006;
Smith, Zakrzewski, Johnson, & Valleau, 2016; Vanpaemel & Storms, 2008.
For discussion of the likely neural basis for exemplar storage, see Ashby &
Rosedahl, 2017; also see Figure 9.3.)
Overall, though, it cannot be surprising that people can draw on either
prototypes or exemplars when thinking about concepts. The reason is that
the two types of information are used in essentially the same way. In either
case, an object before your eyes triggers some representation in memory
(either a representation of a specific instance, according to exemplar theory,
or the prototype, according to prototype theory). In either case, you assess the
FIGURE 9.3 DISTINCTIONS
WITHIN CATEGORIES
The chapter suggests that you have
knowledge of both exemplars and
prototypes. As a further complication, you also have special knowledge
about distinctive individuals within a
category. Thus, you know that Kermit
has many frogly properties (he’s
green, he eats flies, he hops) but also
has unusual properties that make him
a rather unusual frog (since, after all,
he can talk, he can sing, and he’s in
love with a pig).
336 •
C H A P T E R N I N E Concepts and Generic Knowledge
resemblance between this conceptual knowledge, supplied by memory, and
the novel object before you: “Does this object resemble my sister’s couch?” If
so, the object is a couch. “Does the object resemble my prototype for a soup
bowl?” If so, it’s probably a soup bowl.
Given these similarities, it seems plausible that we might merge the proto­
type and exemplar proposals, with each of us on any particular occasion
relying on whichever sort of information (exemplar or prototype) comes to
mind more readily.
The Difficulties with Categorizing
via Resemblance
TEST YOURSELF
6.What is similar in the
processes of categorizing via a prototype
and the processes of
categorizing via an
exemplar? What is
different between
these two types of
processes?
We’re moving, it seems, toward a clear-cut set of claims. First, for most concepts, definitions are not available. Second, for many purposes, you don’t
need a definition and can rely instead on a mix of prototypes and exemplars.
Third, typicality — the degree to which a particular object or situation or
event is typical for its kind — plays a large role in people’s thinking, with
more-typical category members being “privileged” in many ways. Fourth,
typicality is exactly what we would expect if category knowledge does, in
fact, hinge on prototypes and exemplars.
This reasoning seems straightforward enough. However, some results
don’t fit into this picture, so the time has come to broaden our conception of
concepts.
The Differences between Typicality
and Categorization
If you decide that Mike is a bully, or that an event is a tragedy, or that a particular plant is a weed, it’s usually because you’ve compared the “candidate”
in each case to the relevant prototype or exemplar. In essence, you’ve asked
yourself how much the candidate person (or event, or object) resembles a
typical member of the target category. If the resemblance is strong, you decide
the candidate is typical for the category, and likely a member of that category.
If the resemblance is poor, you decide the candidate isn’t at all typical, and its
category status is (at best) uncertain.
The essential point, then, is that judgments of category membership
depend on judgments of typicality, and so these two types of judgment will
inevitably go hand in hand. This point certainly fits with the data we’ve
seen so far, but it doesn’t fit with some other results — results that show
no linkage at all between judgments of category membership and judgments
of typicality.
Armstrong, Gleitman, and Gleitman (1983) gave participants this peculiar
instruction: “We all know that some numbers are even-er than others. What
I want you to do is to rate each of the numbers on this list for how good an
example it is for the category ‘even number.’” Participants were then given a
The Difficulties with Categorizing via Resemblance
•
337
list of numbers (4, 16, 32, and so on) and had to rate “how even” each number was. The participants thought this was a strange task but followed the
instruction nonetheless — and, interestingly, were quite consistent with one
another in their judgments (see Table 9.2).
Of course, participants responded differently (and correctly!) if asked in a
direct way which numbers on the list were even and which were not. Apparently, then, participants could judge category membership as easily as they
could judge typicality, but — importantly — these judgments were entirely
independent of each other. Thus, for example, participants responded that
4 is a more typical even number than 7,534, but they knew this has nothing
to do with the fact that both are unmistakably in the category “even number.”
Clearly, therefore, there’s some basis for judging category membership that’s
separate from the assessment of typicality.
One might argue, though, that mathematical concepts like “even number”
are somehow special, and so their status doesn’t tell us much about other,
more “ordinary” concepts. However, this suggestion is quickly rebutted,
because other concepts show a similar distinction between category membership and typicality. For example, robins strike us as being closer to the typical
bird than penguins are; even so, most of us are certain that both robins and
penguins are birds. Likewise, Moby Dick was definitely not a typical whale,
but he certainly was a whale; Abraham Lincoln wasn’t a typical American,
but he was an American. These informal observations, like the even-number
result, drive a wedge between typicality and category membership — a wedge
that doesn’t fit with our theory so far.
TABLE 9.2
ARTICIPANTS’ TYPICALITY RATINGS FOR
P
WELL-DEFINED CATEGORIES
EVEN NUMBER
Stimulus
ODD NUMBER
Typicality Rating
Stimulus
Typicality Rating
4
5.9
3
5.4
8
5.5
7
5.1
10
5.3
23
4.6
18
4.4
57
4.4
34
3.6
501
3.5
106
3.1
447
3.3
Participants rated numbers on how typical they were for the category “even number.”
Ratings were on a scale from 0 to 7, with 7 meaning the item was (in the participants’
view) very typical. Mathematically this is absurd: Either a number is even (divisible
by 2 without a remainder) or it is not. Even so, participants rated some numbers
as “evener” than others, and likewise rated some odd numbers as being “odder”
than others.
( after armstrong et al ., 1983)
338 •
C H A P T E R N I N E Concepts and Generic Knowledge
How are category judgments made when they don’t rely on typicality?
As an approach to this question, let’s think through an example. Consider a
lemon. Paint the lemon with red and white stripes. Is it still a lemon? Most
people say that it is. Now, inject the lemon with sugar water, so it has a sweet
taste. Then, run over the lemon with a truck, so that it’s flat as a pancake.
What have we got at this point? Do we have a striped, artificially sweet, flattened lemon? Or do we have a non-lemon? Most people still accept this poor,
abused fruit as a lemon, but consider what this judgment involves. We’ve
taken steps to make this object more and more distant from the prototype
and also very different from any specific lemon you’ve ever encountered (and
therefore very different from any remembered exemplars). But this seems
not to shake your faith that the object remains a lemon. To be sure, we have
a not-easily-recognized lemon, an exceptional lemon, but it’s still a lemon.
Apparently, something can be a lemon with virtually no resemblance to
other lemons.
Related points emerge in research with children. In one early study,
preschool children were asked what makes something a “coffeepot,” a “raccoon,” and so on (Keil, 1986). As a way of probing their beliefs, the children
were asked whether it would be possible to turn a toaster into a coffeepot.
Children realized that we’d have to widen the holes in the top of the toaster
CATEGORIZATION OUTSIDE
OF TYPICALITY
Moby Dick was not a typical whale, but he unmistakably
was a whale. Clearly, then, in some settings, typicality
can be separated from category membership.
The Difficulties with Categorizing via Resemblance
•
339
and fix things so that the water wouldn’t leak out of the bottom. We’d also
need to design a place to put the coffee grounds. But the children saw no
obstacles to these manipulations and were quite certain that with these
adjustments in place, we would have created a bona fide coffeepot.
Things were different, though, when the children were asked a parallel
question — whether one could, with suitable adjustments, turn a skunk into
a raccoon. The children understood that we could dye the skunk’s fur, teach
it to climb trees, and, in general, teach it to behave in a raccoon-like fashion.
Even with these adjustments, the children steadfastly denied that we would
have created a raccoon. A skunk that looks, sounds, and acts just like a raccoon might be a very peculiar skunk, but it would be a skunk nonetheless.
(For other evidence suggesting that people reason differently about naturally
occurring items like raccoons and manufactured items like coffeepots, see
Caramazza & Shelton, 1998; Estes, 2003; German & Barrett, 2005; Levin,
Takarae, Miner, & Keil, 2001. Also see “Different Profiles for Different
Concepts” section on p. 347.)
What lies behind all these judgments? If people are asked why the abused
lemon still counts as a lemon, they’re likely to mention that it grew on a
lemon tree, is genetically a lemon, and is still made up of (mostly) the “right
stuff.” It’s these “deep” features that matter, not the lemon’s current properties. And so, too, for raccoons: In the children’s view, being a raccoon
isn’t merely a function of having the relevant features; instead, according
to the children, the key to being a raccoon involves (among other things)
having a raccoon mommy and a raccoon daddy. In this way, a raccoon,
just like a lemon, is defined in ways that refer to deep properties and not to
mere appearances.
Notice, though, that these claims about an object’s deep properties
depend on a web of other beliefs — beliefs that are, in each case, “tuned”
to the category being considered. Thus, you’re more likely to think that a
creature is a raccoon if you’re told that it has raccoons as parents, but this is
true only because you have some ideas about how a creature comes to be a
raccoon — ideas that are linked to your broader understanding of biological
categories and inheritance. It’s this understanding that tells you that parentage is relevant here. If this point isn’t clear, consider as a contrasting case the
steps you’d go through in deciding whether Judy really is a doctor. In this
case, you’re unlikely to worry about whether Judy has a doctor mommy and
a doctor daddy, because your beliefs tell you, of course, that for this category
parentage doesn’t matter.
As a different example, think about the category “counterfeit money.” A
counterfeit bill, if skillfully produced, will have a nearly perfect resemblance
to the prototype for legitimate money. Despite this resemblance, you understand that a counterfeit bill isn’t in the category of legitimate money, so here,
too, your categorization doesn’t depend on typicality. Instead, your categorization depends on a web of other beliefs, including beliefs about circumstances
of printing. A $20 bill is legitimate, you believe, only if it was printed
with the approval of, and under the supervision of, the relevant government
340 •
C H A P T E R N I N E Concepts and Generic Knowledge
A SPANIEL, NOT A WOLF,
IN SHEEP’S CLOTHING?
Both of these creatures resemble the prototype for sheep,
and both resemble many sheep
exemplars you’ve seen (or perhaps read about). But are they
really sheep?
agencies. And once again, these beliefs arise only because you have a broader
understanding of what money is and how government regulations apply to
monetary systems. In other words, you consider circumstances of printing
only because your understanding tells you that the circumstances are relevant
here, and you won’t consider circumstances of printing in a wide range of
other cases. If asked, for example, whether a copy of the Lord’s Prayer is
“counterfeit,” your beliefs tell you that the Lord’s Prayer is the Lord’s Prayer
no matter where (or by whom) it was printed. Instead, what’s crucial for the
prayer’s “authenticity” is simply whether the words are the correct words.
The Complexity of Similarity
Let’s pause to review. There’s no question that judgments about categories
are often influenced by typicality, and we’ll need to account for this fact in
our theorizing. Sometimes, though, category judgments are independent of
typicality: You judge some candidates to be category members even though
they don’t resemble the prototype (think about Moby Dick or the abused
lemon). You judge some candidates not to be in the category even though
they do resemble the prototype (think about counterfeit money or the disguised skunk).
We need to ask, therefore, how you think about categories when you’re
not guided by typicality. The answer, it seems, is that you focus on attributes
that you believe are essential for each category. Your judgments about what’s
essential, however, depend, on your beliefs about that category. Therefore,
you consider parentage when thinking about a category (like skunk or
raccoon) for which you believe biological inheritance is important. You
consider circumstances of printing when you’re concerned with a category
The Difficulties with Categorizing via Resemblance
•
341
(like counterfeit money) that’s shaped by your beliefs about economic systems. And so on.
Is it possible, though, that we’re pushed into these complexities only because
we’ve been discussing oddball categories such as abused citrus fruits and transformed forest animals? The answer is no, because similar complexities emerge
in less exotic cases. The reason is that the prototype and exemplar views both
depend on judgments of resemblance (resemblance either to a prototype or to
some remembered instance), and resemblance, in turn, is itself a complex notion.
How do you decide whether two objects resemble each other? The obvious suggestion is that objects resemble each other if they share properties,
and the more properties shared, the greater the resemblance. Therefore, we
can say there’s some resemblance between an apple and a tennis ball because
they share a shape (round) and a size (about 3 or 4 inches in diameter). The
resemblance is limited, though, because there are many properties that these
objects don’t share (color, “furry” surface, and so on).
It turns out, though, that this idea of “resemblance from shared properties” won’t work. To see why, consider plums and lawn mowers; how much
do these two things resemble each other? Common sense says they don’t
resemble each other at all, but we’ll reach the opposite conclusion if we simply
count “shared properties” (Murphy & Medin, 1985). After all, both weigh
less than a ton, both are found on Earth, both have a detectable odor, both
are used by people, both can be dropped, both cost less than a thousand dollars, both are bigger than a grain of sand, both are unlikely birthday presents
for your infant daughter, both contain carbon molecules, both cast a shadow
on a sunny day. And on and on and on. With a little creativity, you could
probably count thousands of properties shared by these two objects — but
that doesn’t change the basic assessment that there’s not a close resemblance here. (For discussion, see Goldstone & Son, 2012; Goodman, 1972;
Markman & Gentner, 2001; Medin, Goldstone, & Gentner, 1993.)
The solution to this puzzle, though, seems easy: Resemblance does depend
on shared properties, but — more precisely — it depends on whether the
objects share important, essential properties. On this basis, you regard plums
and lawn mowers as different from each other because the features they share
are trivial or inconsequential. But this idea leads to a question: How do you
decide which features to ignore when assessing similarity and which features
to consider? How do you decide, in comparing a plum and a lawn mower,
which features are relevant and which ones aren’t?
These questions bring us back to familiar territory, because your decisions
about which features are important depend on your beliefs about the concept in question. Thus, in judging the resemblance between plums and lawn
mowers, you were unimpressed that they share the feature “cost less than a
thousand dollars.” That’s because you believe cost is irrelevant for these categories. (If a super-deluxe lawn mower cost a million dollars, it would still be
a lawn mower, wouldn’t it?) Likewise, you don’t perceive plums to be similar
to lawn mowers even though both weigh less than a ton, because you know
this attribute, too, is irrelevant for these categories.
342 •
C H A P T E R N I N E Concepts and Generic Knowledge
Overall, then, the idea is that prototype use depends on judgments of
resemblance (i.e., resemblance between a candidate object and a prototype). Judgments of resemblance, in turn, depend on your being able to
focus on the features that are essential, so that you’re not misled by trivial
features. And, finally, decisions about what’s essential (cost or weight or
whatever) vary from category to category, and vary in particular according to your beliefs about that category. Thus, cost isn’t essential for plums
and lawn mowers, but it is a central attribute for other categories (e.g., the
category “luxury item”). Likewise, having a particular weight isn’t
essential for plums or lawn mowers, but it is prominent for other categories. (Does a sumo wrestler resemble a hippopotamus? Here you might be
swayed by weight.)
The bottom line is that you’re influenced by your background beliefs
when considering oddball cases like the mutilated lemon. But you’re also
influenced by your beliefs in ordinary cases, including, we now see, any case
in which you’re relying on a judgment of resemblance.
TEST YOURSELF
7.Give an example in
which something is
definitely a category
member even though
it has little resemblance to the prototype for the category.
8.In judging similarity,
why is it not enough
simply to count all of
the properties that
two objects have in
common?
Concepts as Theories
It seems clear, then, that our theorizing needs to include more than prototypes
and exemplars. Several pieces of evidence point this way, including the fact
that whenever you use a prototype or exemplar, you’re relying on a judgment
of resemblance, and resemblance, we’ve argued, depends on other knowledge —
knowledge about which attributes to pay attention to and which ones to
regard as trivial. But what is this other knowledge?
BLUE GNU
In judging resemblance or in categorizing an object,
you focus on the features that you believe are
important for an object of that type, and you ignore
nonessential features. Imagine that you encounter a
creature and wonder what it is. Perhaps you reason,
“This creature reminds me of the animal I saw in the
zoo yesterday. The sign at the zoo indicated that
the animal was a gnu, so this must be one, too. Of
course, the gnu in the zoo was a different color and
slightly smaller. But I bet that doesn’t matter. Despite
the new blue hue, this is a zoo gnu, too.” Notice that
in drawing this conclusion you’ve decided that color
isn’t a critical feature, so you categorize despite
the contrast on this dimension. But you know that
color does matter for other categories — and so, for
example, you know that something’s off if a jeweler
tries to sell you a green ruby or a red emerald. Thus,
in case after case, the features that you consider
depend on the specific category.
Concepts as Theories
•
343
Explanatory Theories
In the cases we’ve discussed, your understanding of a concept seems to
involve a network of beliefs linking the target concept to other concepts. To
understand what counterfeit is, you need to know what money is, and probably what a government is, and what crime is. To understand what a raccoon
is, you need to understand what parents are, and with that, you need to know
some facts about life cycles, heredity, and the like.
Perhaps, therefore, we need to change our overall approach. We’ve been
trying throughout this chapter to characterize concepts one by one, as
though each concept could be characterized independently of other concepts.
We talked about the prototype for bird, for example, without considering
how this prototype is related to the animal prototype or the egg prototype.
Maybe, though, we need a more holistic approach, one in which we put more
emphasis on the interrelationships among concepts. This would enable us to
include in our accounts the wide network of beliefs in which concepts seem
to be embedded.
To see how this might play out, let’s again consider the concept “raccoon.” Your knowledge about this concept probably includes a raccoon
prototype and some exemplars, and you rely on these representations in
many settings. But your knowledge also includes your belief that raccoons
are biological creatures (and therefore the offspring of adult raccoons) and
your belief that raccoons are wild animals (and therefore usually not pets,
usually living in the woods). These various beliefs may not be sophisticated,
and they may sometimes be inaccurate, but nonetheless they provide you
with a broad cause-and-effect understanding of why raccoons are as they
are. (Various authors have suggested different proposals for how we should
conceptualize this web of beliefs. See, e.g., Bang, Medin, & Atran, 2007;
Keil, 1989, 2003; Lakoff, 1987; Markman & Gentner, 2001; Murphy, 2003;
Rips et al., 2012.)
Guided by these considerations, many authors suggest that each of us has
something we can think of as a “theory” about raccoons — what they are,
how they act, why they are as they are — and likewise a “theory” about most
of the other concepts we hold. The theories are less precise, less elaborate,
than a scientist’s theory, but they serve the same function. They provide a
crucial knowledge base that we rely on in thinking about an object, event,
or category; and they enable us to understand new facts we might encounter
about the object or category.
The Function of Explanatory Theories
We’ve already suggested that implicit “theories” influence how you categorize
things — that is, your decisions about whether a test case is or is not in a particular category. This was crucial in our discussion of the abused lemon, the
transformed raccoon, and the counterfeit bill. Your “theory” for a concept
was also crucial for our discussion of resemblance — guiding your decisions
about which features matter in judging resemblance and which ones do not.
344 •
C H A P T E R N I N E Concepts and Generic Knowledge
As a different example, imagine that you see someone at a party jump
fully clothed into a pool. Odds are good that you would decide this person
belongs in the category “drunk,” but why? Jumping into a pool in this way
surely isn’t part of the definition of being drunk, and it’s unlikely to be part
of the prototype (Medin & Ortony, 1989). But each of us has certain beliefs
about how drunks behave; we have, in essence, a “theory” of drunkenness.
This theory enables us to think through what being drunk will cause someone to do, and on this basis we would decide that, yes, someone who jumped
into the pool fully clothed probably was drunk.
You also draw on a “theory” when thinking about new possibilities for a
category. For example, could an airplane fly if it were made of wood? What if
it were ceramic? How about one made of whipped cream? You immediately
reject this last option, because you know that a plane’s function depends on
its aerodynamic properties, and those depend on the plane’s shape. Whipped
cream wouldn’t hold its shape, so it isn’t a candidate for airplane construction. This is an easy conclusion to draw — but only because your “airplane”
concept contains some ideas about why airplanes are as they are.
Your “theories” also affect how quickly you learn new concepts. Imagine
that you’re given a group of objects and must decide whether each belongs
in Category A or Category B. Category A, you’re told, includes objects that
are metal, have a regular surface, are of medium size, and are easy to grasp.
Category B, in contrast, includes objects that aren’t made of metal, have
irregular surfaces, and are small and hard to grasp. This sorting task would
be difficult — unless we give you another piece of information: namely, that
A WOODEN AIRPLANE?
Could an airplane be made of
wood? Made from ceramic?
Made from whipped cream? You
immediately reject the last possibility, because your “theory”
about airplanes tells you that
planes can fly only because of
their wings’ shape, and whipped
cream wouldn’t maintain this
shape. Planes can, however, be
made of wood — and this one
(the famous Spruce Goose) was.
Concepts as Theories
•
345
Category A includes objects that could serve as substitutes for a hammer.
With this clue, you immediately draw on your other knowledge about hammers, including your understanding of what a hammer is and how it’s used.
This understanding enables you to see why Category A’s features aren’t an
arbitrary hodgepodge; instead, the features form a coherent package. And
once you see this point, learning the experimenter’s task (distinguishing
Category A from Category B) is easy. (See Medin, 1989; Wattenmaker, Dewey,
Murphy, & Medin, 1986. For related findings, see Heit & Bott, 2000; Kaplan
& Murphy, 2000; Rehder & Ross, 2001.)
Inferences Based on Theories
If you meet my pet, Milo, and decide that he’s a dog, then you instantly know
a great deal about Milo — the sorts of things he’s likely to do (bark, beg for
treats, chase cats) and the sorts of things he’s unlikely to do (climb trees, play
chess, hibernate all winter). Likewise, if you learn some new fact about Milo,
you’ll be able to make broad use of that knowledge — applying it to other
creatures of his kind. If, for example, you learn that Milo is vulnerable to circovirus, you’ll probably conclude that other dogs are also vulnerable to this virus.
These examples remind us of one of the reasons categorization is so
important: Categorization enables you to apply your general knowledge (e.g.,
knowledge about dogs) to new cases you encounter (e.g., Milo). Conversely,
categorization enables you to draw broad conclusions from your experience
(so that things you learn about Milo can be applied to other dogs you meet).
All this is possible, though, only because you realize that Milo is a dog; without this simple realization, you wouldn’t be able to use your knowledge in
this way. But how exactly does this use-of-knowledge proceed?
Early research indicated that inferences about categories were guided by
typicality. In one study, participants, told a new fact about robins, were willing to infer that the new fact would also be true for ducks. If they were told
a new fact about ducks, however, they wouldn’t extrapolate to robins (Rips,
1975). Apparently, people were willing to make inferences from the typical
case to the whole category, but not from an atypical case to the category. (For
discussion of why people are more willing to draw conclusions from typical
cases, see Murphy & Ross, 2005.)
However, your inferences are also guided by your broader set of beliefs,
and so, once again, we find a role for the “theory” linked to each concept.
For example, if told that gazelle’s blood contains a certain enzyme, people are
willing to conclude that lion’s blood contains the same enzyme. However, if
told that lion’s blood contains the enzyme, people are less willing to conclude
that gazelle’s blood does too. What’s going on here? People find it easy to
believe the enzyme can be transmitted from gazelles to lions, because they
can easily imagine that lions sometimes eat gazelles; people have a harder
time imagining a mechanism that would transmit the enzyme in the reverse
direction. Likewise, if told that grass contains a certain chemical, people are
willing to believe that cows have the same chemical inside them. This makes
346 •
C H A P T E R N I N E Concepts and Generic Knowledge
WHY IS CATEGORIZATION SO IMPORTANT?
If you decide that Milo is a dog, then you instantly know a great deal
about him (e.g., that he’s likely to bark and chase cats, unlikely to climb
trees or play chess). In this way, categorization enables you to apply
your general knowledge to new cases. And if you learn something
new about Milo (e.g., that he’s at risk for a particular virus), you’re likely
to assume the same is true for other dogs. In this way, categorization
also enables you to draw broad conclusions from specific experiences.
perfect sense if people are thinking of the inference in terms of cause and
effect, relying on their beliefs about how these concepts are related to each
other (Medin, Coley, Storms, & Hayes, 2003; also see Heit, 2000; Heit &
Feeney, 2005; Rehder & Hastie, 2004).
Different Profiles for Different Concepts
This proposal about “theories” and background knowledge has another
implication: People may think about different concepts in different ways. For
example, most people believe that natural kinds (groups of objects that exist
naturally in the world, such as bushes or alligators or stones or mountains) are
as they are because of forces of nature that are consistent across the years. As a
result, the properties of these objects are relatively stable. Thus there are certain
properties that a bush must have in order to survive as a bush; certain properties that a stone must have because of its chemical composition. Things are
different, though, for artifacts (objects made by human beings). If we wished to
make a table with 15 legs rather than 4, or one made of gold, we could do this.
The design of tables is up to us; and the same is true for most artifacts.
This observation leads to the proposal that people will reason differently
about natural kinds and artifacts — because they have different beliefs about
why categories of either sort are as they are. We’ve already seen one result
consistent with this idea: the finding that children agree that toasters could
be turned into coffeepots but not that skunks could be turned into raccoons.
Plainly, the children had different ideas about artifacts (like toasters) than
they had about animate objects (like skunks). Other results confirm this pattern. In general, people tend to assume more stability and more homogeneity
Concepts as Theories
•
347
when reasoning about natural kinds than when reasoning about artifacts
(Atran, 1990; Coley, Medin, & Atran, 1997; Rehder & Hastie, 2004).
The diversity of concepts, as well as the role of beliefs, is also evident in
another context. Many concepts can be characterized in terms of their features (e.g., the features that most dogs have, the features that chairs usually
have, and so on; after Markman & Rein, 2013). Other concepts, though,
involve goal-derived categories, like “diet foods” or “exercise equipment”
(Barsalou, 1983, 1985). Your understanding of concepts like these depends
on your understanding of the goal (e.g., “losing weight”) and some causeand-effect beliefs about how a particular food might help you achieve that
goal. Similar points apply to relational categories (“rivalry,” “hunting”) and
event categories (“visits,” “dates,” “shopping trips”); here, too, you’re influenced by a web of beliefs about how various elements (the predator and the
prey; the shopper and the store) are related to each other.
Concepts and the Brain
The contrasts among different types of concepts are also reflected in neuroscience evidence. For example, fMRI scans show that different brain sites are
activated when people are thinking about living things than when thinking
about nonliving things (e.g., Chao, Weisberg, & Martin, 2002), and different
sites are activated when people are thinking about manufactured objects such
as tools rather than natural objects such as rocks (Gerlach, Law, & Paulson,
2002; Kellenbach et al., 2003).
These results suggest that different types of concepts are represented in different brain areas, and this point is confirmed by observations of people who
have suffered brain damage. In some cases, these people lose the ability to
name certain objects — a pattern termed anomia — or to answer simple questions about these objects (“Does a whale have legs?”). Often, the problem is
specific to certain categories, such that some patients lose the ability to name
living things but not nonliving things; other patients show the reverse pattern. (See Mahon & Caramazza, 2009; Mahon & Hickok, 2016. For broader
discussion, see Peru & Avesani, 2008; Phillips, Noppeney, Humphreys, &
Price, 2002; Rips et al., 2012; Warrington & Shallice, 1984.) Sometimes the
disruption caused by brain damage is even more specific, with some patients
losing the ability to answer questions about fruits and vegetables but still able
to answer questions about other objects, living or nonliving (see Figure 9.4).
Why does the brain separate things in this way? One proposal emphasizes
the idea that different types of information are essential for different concepts.
In this view, the recognition of living things may depend on perceptual properties (especially visual properties) that allow us to identify horses or trees or
other animate objects. In contrast, the recognition of nonliving things may
depend on their functional properties (Warrington & McCarthy, 1983, 1987).
As an interesting complication, though, brain scans also show that sensory
and motor areas in the brain are activated when people are thinking about
certain concepts (Mahon & Caramazza, 2009; Mahon & Hickok, 2016;
348 •
C H A P T E R N I N E Concepts and Generic Knowledge
FIGURE 9.4
DIFFERENT BRAIN SITES SUPPORT DIFFERENT CATEGORIES
Lesion data
Lesion results summarized
Persons
Animals
TP
Tools
A
IT
IT+
Persons: x = 59.8
Persons: x = 75.5
Persons: x = 91.7
Animals: x = 93.3
Animals: x = 80.1
Animals: x = 88.3
Tools: x = 96.0
Tools: x = 84.5
Tools: x = 78.5
B
Brain damage often causes anomia — an inability to name common objects. But the specific loss depends on
where exactly the brain damage has occurred. Panel A summarizes lesion data for patients who had difficulty
naming persons (top), animals (middle), or tools (bottom). The colors indicate the percentage of patients with
damage at each site: red, most patients; purple, few. Panel B offers a different summary of the data: Patients with
damage in the brain’s temporal pole (TP, shown in blue) had difficulty naming persons (only 59.8% correct) but
were easily able to name animals and tools. Patients with damage in the inferotemporal region (IT, shown in red)
had difficulty naming persons and animals but did somewhat better naming tools. Finally, patients with damage in
the lateral occipital region (IT+) had difficulty naming tools but did reasonably well naming animals and persons.
( after damasio , grabowski , tranel , hichwa , & damasio , 1996)
McRae & Jones, 2012). For example, when someone is thinking about the
concept “kick,” we can observe activation in brain areas that (in other circumstances) control the movement of the legs; when someone is thinking about
rainbows, we can detect activation in brain areas ordinarily involved in color
vision. Findings like these suggest that conceptual knowledge is intertwined
with knowledge about what particular objects look like (or sound like or feel
like) and also with knowledge about how one might interact with the object.
Concepts as Theories
•
349
TEST YOURSELF
9.Why is an (informal,
usually unstated)
“theory” needed in
judging the resemblance between
two objects?
10.What’s different
between your (informal, usually unstated)
theory of artifacts and
your theory of natural
kinds?
Some theorists go a step further and argue for a position referred to
as “embodied” or “grounded cognition.” The proposal is that the body’s
sensory and action systems play an essential role in all our cognitive
processes; it’s inevitable, then, that our concepts will include representations of perceptual properties and motor sequences associated with each
concept. (See Barsalou, 2008, 2016; Chrysikou, Csasanto, & ThompsonSchill, 2017; Pulvermüller, 2013. For a glimpse of the debate over this perspective, see Binder, 2016; Bottini, Bucur, & Crepaldi, 2016; Dove, 2016;
Goldinger, Papesh, Barnhart, Hansen, & Hout, 2016; Leshinskaya &
Caramazza, 2016; Reilly, Peele, Garcia, & Crutch, 2016. For a discussion of
how this approach might handle abstract concepts, see Borghi et al., 2017.)
Even if we hold the embodied cognition proposal to the side, the data
here fit well with a theme we’ve been developing throughout this chapter —
namely, that conceptual knowledge has many elements. These include a prototype, exemplars, a theory, and (we now add) representations of perceptual
properties and actions associated with the concept. Let’s also emphasize
that which of these elements you’ll focus on likely depends on your needs at
that moment. In other words, each of your concepts includes many types of
information, but when you’re using your conceptual knowledge, you call to
mind just the subset of information that’s needed for whatever task you’re
engaged in. (See Mahon & Hickok, 2016, especially pp. 949–950; also Yee
& Thompson-Schill, 2016.)
The Knowledge Network
Overall, our theorizing is going to need some complexities, but, within this
complexity, one idea has come up again and again: How you think about
your concepts, how you use your concepts, and what your concepts are, are
all shaped by a web of beliefs and background knowledge. But what does this
“web of beliefs” involve?
Traveling through the Network
to Retrieve Knowledge
In earlier chapters, we explored the idea that information in long-term
memory is represented by means of a network, with associative links connecting nodes to one another. Let’s now carry this proposal one step further.
The associative links don’t just tie together the various bits of knowledge;
they also help represent the knowledge. For example, you know that George
Washington was an American president. This simple idea can be represented
as an associative link between a node representing washington and a node
representing president. In other words, the link itself is a constituent of the
knowledge.
On this view, how do you retrieve knowledge from the network, so that
you can use what you know? Presumably, the retrieval relies on processes
we’ve described in other chapters — with activation spreading from one
350 •
C H A P T E R N I N E Concepts and Generic Knowledge
node to the next. This spread of activation is quick but does take time, and
the farther the activation must travel, the more time needed. This leads to a
prediction — that you’ll need less time to retrieve knowledge involving closely
related ideas, and more time to retrieve knowledge about more distant ideas.
Collins and Quillian (1969) tested this prediction many years ago, using the
sentence verification task described earlier in this chapter. Their participants
were shown sentences such as “A robin is a bird” or “Cats have claws” or
“Cats have hearts.” Mixed together with these obviously true sentences were
various false sentences (e.g., “A cat is a bird”), and in response to each sentence,
participants had to hit a “true” or “false” button as quickly as they could.
Participants presumably perform this task by “traveling” through the network, seeking a connection between nodes. When the participant finds the
connection from, say, the robin node to the birds node, this confirms that
there’s an associative path linking these nodes, which tells the participant
that the sentence about these two concepts is true. This travel should require
little time if the two nodes are directly linked by an association, as robin and
birds probably are (see Figure 9.5). In this case, we’d expect participants
to answer “true” rather quickly. The travel will require more time, however,
if the two nodes are connected only indirectly (e.g., robin and animals),
FIGURE 9.5
HYPOTHETICAL MEMORY STRUCTURE FOR KNOWLEDGE ABOUT ANIMALS
HAVE HEARTS
ANIMALS
BIRDS
CAN FLY
CATS
LAY EGGS
ROBIN
CANARY
EAT FOOD
BREATHE
HAVE SKIN
DOGS
HAVE CLAWS
CHESHIRE
ALLEY CAT
CHASE CATS
BARK
PURR
COLLIE
TERRIER
CAN SING
IS YELLOW
(etc.)
Collins and Quillian proposed that the memory system avoids redundant storage of connections between cats and
have hearts, and between dogs and have hearts, and so on for all the other animals. Instead, have hearts is stored
as a property of all animals. To confirm that cats have hearts, therefore, you must traverse two links: from cats to
animals, and from animals to have hearts.
( after collins & quillian , 1969)
The Knowledge Network
•
351
so that we’d expect slower responses to sentences that require a “two-step”
connection than to sentences that require a single connection.
Collins and Quillian also argued that there’s no point in storing in
memory the fact that cats have hearts and the fact that dogs have hearts
and the fact that squirrels have hearts. Instead, they proposed, it would
be more efficient just to store the fact that these various creatures are animals, and then the separate fact that animals have hearts. As a result, the
property “has a heart” would be associated with the animals node rather
than the nodes for each individual animal, and the same is true for all
the other properties of animals, as shown in the figure. According to this
logic, we should expect relatively slow responses to sentences like “Cats
have hearts,” since, to choose a response, a participant must locate the
linkage from cat to animals and then a second linkage from animals to
have hearts. We would expect a quicker response to “Cats have claws,”
because here there would be a direct connection between cat and the node
representing this property. (Why a direct connection? All cats have claws
but some other animals don’t, so this information couldn’t be entered at
the higher level.)
As Figure 9.6 shows, these predictions are borne out. Responses to
sentences like “A canary is a canary” take approximately 1 second (1,000 ms).
This is presumably the time it takes just to read the sentence and to move
your finger on the response button. Sentences like “A canary can sing”
require an additional step of traversing one link in memory and yield slower
responses. Sentences like “A canary can fly” require the traversing of two
links, from canary to birds and then from birds to can fly, so they are
correspondingly slower.
More recent data, however, add some complications. For example, we
saw earlier in the chapter that verifications are faster if a sentence involves
creatures close to the prototype — so that responses are faster to, say, “A
canary is a bird” than to “An ostrich is a bird.” This difference isn’t reflected
in Figure 9.6, nor is it explained by the layout in Figure 9.5. Clearly, then,
the Collins and Quillian view is incomplete.
In addition, the principle of “nonredundancy” proposed by Collins and
Quillian doesn’t always hold. For example, the property of “having feathers”
should, on their view, be associated with the birds node rather than
(redundantly) with the robin node, the pigeon node, and so on. This fits
with the fact that responses are relatively slow to sentences like “Sparrows
have feathers.” However, it turns out that participants respond rather quickly
to a sentence like “Peacocks have feathers.” This is because in observing peacocks, you often think about their prominent tail feathers (Conrad, 1972).
Therefore, even though it is informationally redundant, a strong association
between peacock and have feathers is likely to be established.
Even with these complications, we can often predict the speed of knowledge access by counting the number of nodes participants must traverse in
answering a question. This observation powerfully confirms the claim that
associative links play a pivotal role in knowledge representation.
352 •
C H A P T E R N I N E Concepts and Generic Knowledge
FIGURE 9.6
T IME NEEDED TO CONFIRM VARIOUS
SEMANTIC FACTS
A canary
has skin.
1,500
Mean response time (ms)
1,400
A canary
can sing.
1,300
A canary
is a bird.
1,200
1,100
A canary
can fly.
A canary is
an animal.
A canary is
a canary.
1,000
Property
900
Category
1
0
2
Levels to be traversed
In a sentence verification task, participants’ responses were fastest when the
test required them to traverse zero links in memory (“A canary is a canary”),
slower when the necessary ideas were separated by one link, and slower still
if the ideas were separated by two links. Responses were also slower if participants had to take the additional step of traversing the link from a category
label (“bird”) to the node representing a property of the category (can fly).
( after collins & quillian , 1969)
Propositional Networks
To represent the full fabric of your knowledge, however, we need more than
simple associations. After all, we need somehow to represent the contrast
between “Sam has a dog” and “Sam is a dog.” If all we had is an association
between sam and dog, we wouldn’t be able to tell these two ideas apart.
One widely endorsed proposal solves this problem with a focus on propositions, defined as the smallest units of knowledge that can be either true or
false (Anderson, 1976, 1980, 1993; Anderson & Bower, 1973). For example,
“Children love candy” is a proposition, but “Children” is not; “Susan likes
blue cars” is a proposition, but “blue cars” is not. Propositions are easily represented as sentences, but this is just a convenience. They can also be represented
in various nonlinguistic forms, including a structure of nodes and linkages, and
that’s exactly what Anderson’s model does.
The Knowledge Network
•
353
Figure 9.7 provides an example. Here, each ellipse identifies a single proposition. Associations connect an ellipse to ideas that are the proposition’s constituents, and the associations are labeled to specify the constituent’s role within that
proposition. This enables us to distinguish, say, the proposition “Dogs chase
cats” (shown in the figure) from the proposition “Cats chase dogs” (not shown).
This model shares many claims with the network theorizing we discussed
in earlier chapters. Nodes are connected by associative links. Some of these
links are stronger than others. The strength of a link depends on how frequently and recently it has been used. Once a node is activated, the process
of spreading activation causes nearby nodes to become activated as well. The
model is distinctive, however, in its attempt to represent knowledge in terms
of propositions, and the promise of this approach has attracted the support
of many researchers. (For recent discussion, see Salvucci, 2017; for some
alternative models, see Flusberg & McClelland, 2017; Kieras, 2017. For
more on how this network can store information, see Figure 9.8.)
Distributed Processing
In the model just described, individual ideas are represented with local representations. Each node represents one idea so that when that node is activated,
you’re thinking about that idea, and when you’re thinking about that idea,
that node is activated. Connectionist networks, in contrast, take a different
approach. They rely on distributed representations, in which each idea is represented, not by a certain set of nodes, but instead by a pattern of activation
across the network. To take a simple case, the concept “birthday” might be
represented by a pattern in which nodes b, f, h, n, p, and r are firing, whereas
the concept “computer” might be represented by a pattern in which nodes
c, g, h, m, o, and s are firing. Note that node h is part of both of these patterns and probably part of the pattern for other concepts as well. Therefore,
we can’t attach any meaning or interpretation to this node by itself; we can
only learn what’s being represented by looking at many nodes simultaneously
to find out what pattern of activation exists across the entire network. (For
more on local and distributed representations, see Chapter 4; also see the
related discussion of neural coding in Chapter 2.)
This reliance on distributed representation has important consequences
for how a connectionist network functions. Imagine being asked what sort
of computer you use. For you to respond, the idea “computer” needs to
trigger the idea “MacBook” (or “Toshiba” or whatever it is you have). In a
distributed network, this means that the many nodes representing the concept “computer” have to manage collectively to activate the many nodes
representing “MacBook.” To continue our simple illustration, node c has to
trigger node l at the same time that node g triggers node a, and so on, leading ultimately to the activation of the l-a-f-j-t-r combination that, let’s say,
represents “MacBook.” In short, a network using distributed representations
must use processes that are similarly distributed, so that one widespread
activation pattern can evoke a different (but equally widespread) pattern.
354 •
C H A P T E R N I N E Concepts and Generic Knowledge
FIGURE 9.7
N
ETWORK REPRESENTATIONS OF SOME
OF YOUR KNOWLEDGE ABOUT DOGS
CHEW
Relation
DOG
Agent
BONE
Agent
Agent
CHASE
Object
Subject
Relation PART
OF
Relation
Object Object
Relation
Object
CAT
EAT
MEAT
Your understanding of dogs — what they are, what they’re likely to do — is
represented by an interconnected network of propositions, with each proposition being indicated by an ellipse. Labels on the arrows indicate each node’s
role within the proposition.
( after anderson , 1980)
FIGURE 9.8
EPRESENTING EPISODES WITHIN
R
A PROPOSITIONAL NETWORK
LAST SPRING
Time
JACOB
Agent
Object
Relation
FEEDS
PIGEONS
IN
Relation
Location
TRAFALGAR
SQUARE
In order to represent episodes, the propositional network includes time and
location nodes. This fragment of a network represents two propositions: the
proposition that Jacob fed pigeons last spring, and the proposition that the
pigeons are in Trafalgar Square. Notice that no time node is associated with
the proposition about pigeons being in Trafalgar Square. Therefore, what’s
represented is that the feeding of the pigeons took place last spring but that
the pigeons are always in the square.
The Knowledge Network
•
355
COGNITION
outside the lab
Stereotypes
You have concepts of things (“chair,” “book,”
efficient means of organizing large quantities of
“kitchen”),
(“running,”
information (and so both serve the same function
“hiding,” “dancing”), and concepts for animals
as schemata, which we described in Chapter 8). But,
(“cat,” “cow,” “dragon”). But you also have con-
of course, a reliance on stereotypes can lead to a
cepts that apply to people. You understand what
list of toxic problems — racism, sexism, homophobia,
a “nurse” is, and a “toddler,” and a “nerd” or a
prejudice against anyone wearing a hijab, and more.
“jock.” You also have concepts for various reli-
Many factors fuel these ugly tendencies,
gious, racial, and ethnic groups (“Jew,” “Muslim,”
including the fact that people often act as if all
“African American,” “Italian”), various political
members of the stereotyped group are alike. They
groups (“radical,” “ultra-conservative”), and many
assume, for example, that a tall African American
others as well.
individual is probably a talented basketball player,
concepts
for
actions
In many ways, concepts representing your ideas
and that a Semitic-looking young man wearing a
about groups of people have the same profile as any
headscarf is probably a terrorist. These assump-
other concepts: It’s difficult to find a rigid definition
tions are, of course, indefensible because humans
for most of these groups, because we can usually
in any group differ from one another, and there’s
find individuals who are in the group even though
no justification for jumping to conclusions about
they don’t quite fit the definition. You also have a
someone just because you’ve decided he or she is
cluster of interwoven beliefs (a “theory”) about these
a member of a particular group.
groups — beliefs that link your ideas about the group
This kind of assumption, though, is widespread
to many other ideas. You also have a prototype in
enough so that social psychologists give it a name:
mind for the group, but here we typically use a dif-
the outgroup homogeneity effect. This term refers
ferent term: You have a stereotype for the group.
to the fact that most people are convinced that their
How are stereotypes different from prototypes?
“ingroup” (the group they belong to) is remarkably
Prototypes are a summary of your experience —
varied, while “outgroups” (groups they don’t belong
and so your prototype for “dog” can be thought
to) are quite homogeneous. In other words, no mat-
of as an average of all the dogs you’ve seen.
ter who you count as “they” and who you count as
Stereotypes, in contrast, are often acquired
“we,” you’re likely to agree that “they all think and
through social channels — with friends or family,
act alike; we, however, are wonderfully diverse.”
or perhaps public figures, shaping your ideas
In combating prejudice, then, it’s useful to real-
about what “lawyers” are like, or “Canadians,” or
ize that this assumption of homogeneity isn’t just
“Italians.” In addition, stereotypes often include an
wrong; it can also have ugly consequences. There
emotional or evaluative dimension, with the result
may be intellectual efficiency in thinking about wom-
that there are groups you’re afraid of, groups you
en, or the elderly, or politicians as if these groups
respect, groups you sneer at.
were uniform, but in doing so you fail to respect
Let’s acknowledge, though, that stereotypes can
the differences from one person to the next — and
serve the same cognitive function as prototypes.
may end up with beliefs, feelings, or actions that are
In both cases, these representations provide an
impossible to justify and often deeply harmful.
356 •
C H A P T E R N I N E Concepts and Generic Knowledge
In addition, the steps bringing this about must all occur simultaneously — in
parallel — with each other, so that one entire representation can smoothly
trigger the next. This is why connectionist models are said to involve parallel
distributed processing (PDP).
Many theorists argue that models of this sort make biological sense. We
know that the brain relies on parallel processing, with ongoing activity in
many regions simultaneously. We also know that the brain uses a “divide and
conquer” strategy, with complex tasks being broken down into small components, and with separate brain areas working on each component.
In addition, PDP models are remarkably powerful, and computers relying
on this sort of processing are often able to conquer problems that seemed
insoluble with other approaches. As a related point, PDP models have an
excellent capacity for detecting patterns in the input they receive, despite a
range of variations in how the pattern is implemented. The models can therefore recognize a variety of different sentences as all having the same structure,
and a variety of game positions as all inviting the same next move. As a result,
these models are impressively able to generalize what they have “learned” to
new, never-seen-before variations on the pattern. (For a broad view of what
connectionism can accomplish, see Flusberg & McClelland, 2017.)
Learning as the Setting of Connection Weights
How do PDP models manage to detect patterns? How do these models “learn”?
Recall that in any associative network, knowledge is represented by the associations themselves. To return to an earlier example, the knowledge that “George
Washington was president” is represented via a link between the nodes representing “Washington” and those representing “president.” When we first introduced
this example, we phrased it in terms of local representations, with individual
nodes having specific referents. The idea, however, is the same in a distributed
system. What it means to know this fact about Washington is to have a pattern
of connections among the many nodes that together represent “Washington”
and the many nodes that together represent “president.” Once these connections
are in place, activation of either pattern will lead to the activation of the other.
Notice, then, that knowledge refers to a potential rather than to a state. If
you know that Washington was a president, then the connections are in place
so that if the “Washington” pattern of activations occurs, this will lead to the
“president” pattern of activations. And this state of readiness will remain even
if you happen not to be thinking about Washington right now. In this way,
“knowing” something, in network terms, corresponds to how the activation
will flow if there is activation on the scene. This is different from “thinking
about” something, which corresponds to which nodes are active at a particular moment, with no comment about where that activation will spread next.
According to this view, “learning” involves adjustments of the connections
among nodes, so that after learning, activation will flow in a way that can represent the newly gained knowledge. Technically, we would say that learning
involves the adjustment of connection weights — the strength of the individual
The Knowledge Network
•
357
TEST YOURSELF
11.What does it mean
to say that knowledge can be represented via network
connections?
12.What is a propositional network?
13.Why do distributed
representations
require distributed
processing?
connections among nodes. Moreover, in this type of model, learning requires
the adjustment of many connection weights. We need to adjust the connections, for example, so that the thousands of nodes representing “Washington”
manage, together, to activate the thousands of nodes representing “president.”
In this way, learning, just like everything else in the connectionist scheme, is a
distributed process involving thousands of changes across the network.
Concepts: Putting the Pieces Together
We have now covered a lot of ground — discussing both individual concepts
and also how these concepts might be woven together, via the network, to
form larger patterns of knowledge. We’ve also talked about how the network
itself might be set up — with knowledge perhaps represented by propositions,
or perhaps via a connectionist network. But where does all of this leave us?
You might think there’s nothing glorious or complicated about knowing what
a dog is, or a lemon, or a fish. Your use of these concepts is effortless, and so is your
use of thousands of other concepts. No one over the age of 4 takes special pride
in knowing what an odd number is, nor do people find it challenging to make the
elementary sorts of judgments we’ve considered throughout this chapter.
As we’ve seen, though, human conceptual knowledge is impressively complex. At the very least, this knowledge contains several parts. We’ve suggested
that people have a prototype for most of their concepts as well as a set of
remembered exemplars, and use them for a range of judgments about the
relevant category. People also seem to have a set of beliefs about each concept they hold, and these beliefs reflect the person’s understanding of causeand-effect relationships — for example, why drunks act as they do, or how
enzymes found in gazelles might be transmitted to lions. These beliefs are
woven into the broader network that manages to store all the information in
your memory, and that network influences how you categorize items and also
how you reason about the objects in your world.
Apparently, then, even our simplest concepts require a multifaceted representation in our minds, and at least part of this representation (the “theory”)
seems reasonably sophisticated. It is all this richness, presumably, that makes
human conceptual knowledge extremely powerful and flexible — and so easy
to use in a remarkable range of circumstances.
COGNITIVE PSYCHOLOGY AND EDUCATION
learning new concepts
In your studies, you encounter many new terms. For example, in this book (and
in many others) you’ll find boldfaced terms introducing new concepts, and
often the book provides a helpful definition, perhaps in a glossary (as this book
358 •
C H A P T E R N I N E Concepts and Generic Knowledge
does). As the chapter argues, though, this mode of presentation doesn’t line up
all that well with the structure of human knowledge. The reason is that you
don’t have (or need) a definition for most of the concepts in your repertoire; in
fact, for many concepts, a definition may not even exist. And even when you
do know a definition, your use of the concept often relies on other information —
including a prototype for that term as well as a set of exemplars.
In addition, your use of conceptual information depends on a broader
fabric of knowledge, linking each concept to other things you know. This
broader knowledge encompasses what we’ve called your “theory” about
that concept — a theory that (among other things) explains why the concept’s
attributes are as they are. You use this theory in many ways; for example,
we’ve argued that whenever you rely on a prototype, you’re drawing conclusions based on the resemblance between the prototype and the new
case you’re thinking about, and that resemblance depends on your theory.
Specifically, it’s your theory that tells you which attributes to pay attention to in judging the resemblance, and which ones to ignore. (So if you’re
thinking about computers, for example, your “theory” about computers
tells you that the color of the machine’s case is irrelevant. In contrast, if
you’re identifying types of birds, your knowledge tells you that color is an
important attribute.)
What does all of this imply for the learning of new concepts? First, let’s
be clear that in some technical domains, concepts do have firm definitions.
(For example, in a statistics class, you learn the definition for the mean of a
set of numbers, and that term is precisely defined.) More broadly, though,
you should bear in mind that definitions tell you what’s generally true of a
concept, but rarely name attributes that are always in place. It’s also important not to be fooled into thinking that knowing a definition is the same as
understanding the concept. In fact, if you only know the definition, you may
end up using the concept foolishly. (And so you might misidentify a hairless
Chihuahua: “That couldn’t be a dog — it doesn’t have fur.”)
What other information do you need, in addition to the definition? At the
least, you should seek out examples of the new concept, because you’ll often
be able to draw analogies based on these examples. You also want to think
about what these examples have in common; that will help you develop a
prototype for the category.
In addition, many students (and many teachers) believe that when learning a new concept, it’s best to view example after example, so that you
really master the concept. Then, you can view example after example of the
next concept, so that you’ll learn that one too. But what if you’re trying to
learn about related concepts or categories? What if, for example, you’re an
art student trying to learn what distinguishes Picasso’s artwork from the
work of his contemporaries, or if you’re a medical student learning how
to distinguish the symptom patterns for various diseases? In these settings,
it’s best to hop back and forth with the examples — so that you examine
a couple of instances of this concept, then a couple of instances of that
one, then back to the first, and so on. This interweaving may slow
Cognitive Psychology and Education
•
359
LEARNING NEW
CONCEPTS
When learning to distinguish
two categories, it’s best to
hop back and forth between
the categories. To learn to
distinguish Monet’s art from
van Gogh’s, therefore, you
might view a painting by
Monet, then one by van Gogh,
then another by Monet, and
so on. This sequence will
lead to better learning than a
sequence of first viewing a
large block of Monet’s paintings and then a large block
of van Gogh’s. (The painting
on top is Monet’s The Artist’s
Garden in Argenteuil; the
painting on the bottom is
van Gogh’s Farmhouse in
Provence.)
360 •
down learning initially, but it will help you in the long run (leaving
you with a sharper and longer-lasting understanding) because you’ll learn
both the attributes that are shared within a category and also the attributes
that distinguish one category from another.
In viewing the examples, though, you also want to think about what
makes them count as examples — what is it about them that puts them into
the category? How are the examples different, and why are they all in the
same category despite these differences? Why are other candidates, apparently similar to these examples, not in the category? Are some of the qualities
of the examples predictable from other qualities? What caused these qualities to be as they are? These questions will help you to start building the
network of beliefs that provide your theory about this concept. These beliefs
will help you to understand and use the concept. But, as the chapter discusses,
these beliefs are also part of the concept — providing the knowledge base that
specifies, in your thoughts, what the concept is all about.
These various points put an extra burden on you and your teachers. It
would be easier if the teacher could simply provide a crisp definition for you
to memorize, and then you could go ahead and commit that definition to
memory. But that’s not what it means to learn a concept. Strict attention just
to a definition will leave you with a conceptual representation that’s not very
useful, and certainly far less rich than you want.
For more on this topic . . .
Brown, P. C., Roediger, H. L., & McDaniel, M. A. (2014). Make it stick: The science
of successful learning. New York, NY: Belknap Press.
Dorek, K., Brooks, L., Weaver, B., & Norman, G. (2012). Influence of familiar
features on diagnosis: Instantiated features in an applied setting. Journal of
Experimental Psychology: Applied, 18, 109–125.
Kim, N. S., & Ahn, W.-K. (2002). Clinical psychologists’ theory-based representations of mental disorders predict their diagnostic reasoning and memory.
Journal of Experimental Psychology: General, 131, 451–476.
C H A P T E R N I N E Concepts and Generic Knowledge
chapter review
SUMMARY
• People cannot provide definitions for most of
• Sometimes categorization doesn’t depend at
the concepts they use; this suggests that knowing a
concept and being able to use it competently do not
require knowing a definition. However, when trying to define a term, people mention properties that
are in fact closely associated with the concept. One
proposal, therefore, is that your knowledge specifies
what is typical for each concept, rather than naming properties that are truly definitive for the concept. Concepts based on typicality will have a family
resemblance structure, with different category members sharing features but with no features being
shared by the entire group.
all on whether the test case resembles a prototype
or a category exemplar. This is evident with some
abstract categories (“even number”) and some
weird cases (a mutilated lemon), but it’s also evident
with more mundane categories (“raccoon”). In
these examples, categorization seems to depend on
knowledge about a category’s essential properties.
• Concepts may be represented in the mind via
prototypes, with each prototype representing what
is most typical for that category. This implies that
categories will have graded membership, and many
research results are consistent with this prediction.
The results converge in identifying some category
members as “better” members of the category. This
is reflected in sentence verification tasks, production
tasks, explicit judgments of typicality, and so on.
• In addition, basic-level categories seem to be the
• Knowledge about essential properties is not just a
supplement to categorization via resemblance. Instead,
knowledge about essential properties may be a prerequisite for judgments of resemblance. With this knowledge, you’re able to assess resemblance with regard to
just those properties that truly matter for the category
and not be misled by irrelevant or accidental properties.
• The properties that are essential for a category
vary from one category to the next. The identification of these properties seems to depend on beliefs
held about the category, including causal beliefs that
specify why the category features are as they are.
These beliefs can be thought of as forming implicit
theories, and they describe the category not in isolation but in relation to various other concepts.
ones we learn earliest and use most often. Basiclevel categories (e.g., “chair”) are more homogeneous than their broader, superordinate categories
(“furniture”) and much broader than their subordinate categories (“armchair”). They are also usually
represented by a single word.
• Researchers have proposed that knowledge is
• Typicality results can also be explained with a
• To store all of knowledge, the network may need
model that relies on specific category exemplars,
and with category judgments being made by the
drawing of analogies to these remembered exemplars. The exemplar model can explain your ability
to view categories from a new perspective. Even so,
prototypes provide an efficient summary of what is
typical for the category. Perhaps it’s not surprising,
therefore, that your conceptual knowledge includes
exemplars and prototypes.
stored within the same memory network that we’ve
discussed in earlier chapters. Searching through this
network seems to resemble travel in the sense that
greater travel distances (more connections to be
traversed) require more time.
more than simple associations. One proposal is that the
network stores propositions, with different nodes each
playing the appropriate role within the proposition.
• A different proposal is that knowledge is contained in memory via distributed representations.
These representations require distributed processes,
including the processes that adjust connection
weights to allow the creation of new knowledge.
361
KEY TERMS
family resemblance (p. 329)
prototype theory (p. 329)
graded membership (p. 331)
sentence verification task (p. 331)
production task (p. 331)
rating task (p. 331)
basic-level categorization (p. 333)
exemplar-based reasoning (p. 335)
typicality (p. 337)
anomia (p. 348)
propositions (p. 353)
local representations (p. 354)
connectionist networks (p. 354)
distributed representations (p. 354)
parallel distributed processing (PDP) (p. 357)
connection weights (p. 357)
TEST YOURSELF AGAIN
1.Consider the word “chair.” Name some
attributes that might plausibly be included
in a definition for this word. But, then, can
you describe objects that you would count
as chairs even though they don’t have one or
more of these attributes?
2.What does it mean to say that there is a family
resemblance among the various animals that
we call “dogs”?
3.Why is graded membership a consequence
of representing the category in terms of a
prototype?
4.What tasks show us that concept judgments
often rely on prototypes and typicality?
5.Give an example of a basic-level category, and
then name some of the subcategories within this
basic-level grouping.
6.What is similar in the processes of categorizing
via a prototype and the processes of categorizing via an exemplar? What is different between
these two types of processes?
362
7.Give an example in which something is
definitely a category member even though it
has little resemblance to the prototype for the
category.
8.In judging similarity, why is it not enough
simply to count all of the properties that two
objects have in common?
9.Why is an (informal, usually unstated)
“theory” needed in judging the resemblance
between two objects?
10.What’s different between your (informal,
usually unstated) theory of artifacts and your
theory of natural kinds?
11.What does it mean to say that knowledge can
be represented via network connections?
12. What is a propositional network?
13.Why do distributed representations require
distributed processing?
THINK ABOUT IT
1.You easily understand the following sentence:
“At most colleges and universities, a large
number of students receive financial aid.”
But how do you manage to understand the
sentence? How is the concept of “financial
aid” represented in your mind? Do you have
a prototype (perhaps for “student on financial
aid”)? Do you have some number of exemplars? A theory? Can you specify what the
theory involves? What other concepts do you
need to understand in order to understand
“financial aid”?
E eBOOK DEMONSTRATIONS & ESSAYS
Go to http://digital.wwnorton.com/cognition7 for the online demonstrations and essays
relevant to this chapter. These can be found in the ebook.
Online Demonstrations
• Demonstration 9.1: The Search for Definitions
• Demonstration 9.2: Assessing Typicality
• Demonstration 9.3: Basic-Level Categories
COGNITION LABS
Go to http://digital.wwnorton.com/cognition7 to experience interactive online psychology
labs relevant to this chapter.
363
10
chapter
Language
what if…
On January 8, 2011, Congresswoman Gabby Giffords
was meeting with citizens outside a grocery store near
Tucson, Arizona. A man ran up to the crowd and began shooting. Six
people were killed; Giffords was among the others who were wounded.
A bullet had passed through her head, traveling the length of her brain’s
left side and causing extensive damage.
As a result of her brain injury, Giffords has suffered from many
profound difficulties, including a massive disruption of her language
capacity, and, among its many other implications, her case brought public attention to the disorder termed “aphasia” — a loss of the ability to
produce and understand ordinary language.
In the years since the shooting, though, Giffords has shown a wonderful degree of recovery. Just five months after the injury, an aide
announced that her ability to comprehend language had returned to a
level that was “close to normal, if not normal.” Her progress has been
slower for language production. Consider an interview she gave in early
2014. Giffords had, on the third anniversary of her shooting, decided to
celebrate life by skydiving. In a subsequent TV interview, she described
the experience: “Oh, wonderful sky. Gorgeous mountain. Blue skies. I like
a lot. A lot of fun. Peaceful, so peaceful.”
Giffords’s recovery is remarkable, but — sadly — not typical. The outcome for patients with aphasia is highly variable, and many recover
far less of their language ability than Giffords has. Her case is typical,
though, in other ways. Different brain areas control the comprehension
and the production of speech, so it’s common for one of these capacities
to be spared while the other is damaged, and the prospects for recovery
are generally better for language comprehension than for production.
And like Giffords, many patients with aphasia retain the ability to sing
even if they’ve lost the ability to speak — a clear indication that these
seemingly similar activities are controlled by different processes.
Giffords also shares with other patients the profound frustration of
aphasia. This condition is, after all, a disorder of language, not a disorder of thought. As a result, patients with aphasia can think normally but
complain (often with great difficulty) that they feel “trapped” in their
own heads, unable to express what they’re thinking. They are sometimes
forced to grunt and point in hopes of conveying their meaning; in other
cases, their speech is so slurred that others cannot understand them,
365
preview of chapter themes
•
anguage can be understood as having a hierarchical
L
structure — with units at each level being assembled to
form the larger units at the next level.
•
t each level in the hierarchy, we can combine and recomA
bine units, but the combinations seem to be governed by
various types of rules. The rules provide an explanation of
why some combinations of elements are rare and others
seem prohibited outright. Within the boundaries created by
these rules, though, language is generative, allowing any user
of the language to create a virtually unlimited number of new
forms (new sound combinations, new words, new phrases).
•
different set of principles describes how, moment by
A
moment, people interpret the sentences they encounter;
in this process, people are guided by many factors, including syntax, semantics, and contextual information.
•
In interpreting sentences, people seem to use a “compile
as you go” strategy, trying to figure out the role of each
word the moment it arrives. This approach is efficient but
can lead to error.
•
ur extraordinary skill in using language is made possible
O
in part by the fact that large portions of the brain are specialized for language use, making it clear that we are, in a
literal sense, a “linguistic species.”
•
inally, language surely influences our thoughts, but in an
F
indirect fashion: Language is one of many ways to draw
our attention to this or that aspect of the environment.
This shapes our experience, which in turn shapes our
cognition.
so they are caught in a situation of trying again and again to express
themselves — but often without success.
To understand the extent of this frustration, bear in mind that we use
language (whether it’s the spoken language most of us use or the sign
language of the deaf) to convey our ideas to one another, and our wishes
and our needs. Without language, cooperative endeavors would be a
thousand times more difficult — if possible at all. Without language, the
acquisition of knowledge would be enormously impaired. Plainly, then,
language capacity is crucial for us all, and in this chapter we’ll consider
the nature of this extraordinary and uniquely human skill.
The Organization of Language
Language use involves a special type of translation. I might, for example,
want to tell you about a happy event in my life, and so I need to translate my
ideas about the event into sounds that I can utter. You, in turn, detect those
sounds and need to convert them into some sort of comprehension. How
does this translation — from ideas to sounds, and then back to ideas — take
place?
The answer lies in the fact that language relies on well-defined patterns —
patterns in how individual words are used, patterns in how words are put
together into phrases. I follow those patterns when I express my ideas, and
the same patterns guide you in figuring out what I just said. In essence, then,
we’re both using the same “codebook,” with the result that (most of the time)
you can understand my messages, and I yours.
366 •
C H A P T E R T E N Language
But where does this “codebook” come from? And what’s in the codebook?
More concretely, what are the patterns of English (or whatever language
you speak) that — apparently — we all know and use? As a first step toward
tackling these issues, let’s note that language has a well-defined structure, as
depicted in Figure 10.1. At the highest level of the structure (not shown in
the figure) are the ideas intended by the speaker, or the ideas that the listener
derives from the input. These ideas are typically expressed in sentences —
coherent sequences of words that express the speaker’s intended meaning.
Sentences, in turn, are composed of phrases, which are composed of words.
Words are composed of morphemes, the smallest language units that carry
meaning. Some morphemes, like “umpire” or “talk,” are units that can stand
alone, and they usually refer to particular objects, ideas, or actions. Other morphemes get “bound” onto these “free” morphemes and add information crucial
for interpretation. Examples of bound morphemes in Figure 10.1 are the pasttense morpheme “ed” and the plural morpheme “s.” Then, finally, in spoken
language, morphemes are conveyed by sounds called phonemes, defined as the
smallest unit of sound that serves to distinguish words in a language.
Language is also organized in another way: Within each of these levels,
people can combine and recombine the units to produce novel utterances —
assembling phonemes into brand-new morphemes or assembling words into
THE HIERARCHY OF LINGUISTIC UNITS
The umpires talked to the players
PHRASE
The umpires
WORD
The
MORPHEME
The
PHONEME
i
talked to the players
umpires
umpire
v
mpayr
talked
to
the
the
play
er
s
e
SENTENCE
pley
r
z
s
talk
ed
to
z
t ck
t
tuw
players
e
FIGURE 10.1
It is useful to think of language as having a hierarchical structure. At the top of the hierarchy, there are sentences.
These are composed of phrases, which are themselves composed of words. The words are composed of
morphemes, and when the morphemes are pronounced, the units of sound are called “phonemes.” In describing phonemes, the symbols correspond to the actual sounds produced, independent of how these sounds are
expressed in ordinary writing.
The Organization of Language
•
367
TEST YOURSELF
1.What are morphemes?
What are phonemes?
brand-new phrases. Crucially, though, not all combinations are possible — so
that a new breakfast cereal, for example, might be called “Klof” but would
probably seem strange to English speakers if it were called “Ngof.” Likewise,
someone might utter the novel sentence “I admired the lurking octopi” but
almost certainly wouldn’t say, “Octopi admired the I lurking.” What lies behind
these points? Why are some sequences acceptable — even if strange — while
others seem awkward or even unacceptable? The answers to these questions
are crucial for any understanding of what language is.
Phonology
Let’s use the hierarchy in Figure 10.1 as a way to organize our examination
of language. We’ll start at the bottom of the hierarchy — with the sounds of
speech.
The Production of Speech
In ordinary breathing, air flows quietly out of the lungs and up through the
nose and mouth (see Figure 10.2). There will usually be some sort of sound,
though, if this airflow is interrupted or altered, and this fact is crucial for
vocal communication.
For example, within the larynx there are two flaps of muscular tissue
called the “vocal folds.” (These structures are also called the “vocal cords,”
although they’re not cords at all.) These folds can be rapidly opened and
closed, producing a buzzing sort of vibration known as voicing. You can feel
this vibration by putting your palm on your throat while you produce a [z]
sound. You’ll feel no vibration, though, if you hiss like a snake, producing a
sustained [s] sound. Try it! The [z] sound is voiced; the [s] is not.
You can also produce sound by narrowing the air passageway within the
mouth itself. For example, hiss like a snake again and pay attention to your
tongue’s position. To produce this sound, you placed your tongue’s tip near
the roof of your mouth, just behind your teeth; the [s] sound is the sound of
the air rushing through the narrow gap you created.
If the gap is somewhere else, a different sound results. For example, to
produce the [sh] sound (as in “shoot” or “shine”), the tongue is positioned
so that it creates a narrow gap a bit farther back in the mouth; air rushing
through this gap causes the desired sound. Alternatively, the narrow gap can
be more toward the front. Pronounce an [f] sound; in this case, the sound is
produced by air rushing between your bottom lip and your top teeth.
These various aspects of speech production provide a basis for categorizing speech sounds. We can distinguish sounds, first, according to how
the airflow is restricted; this is referred to as manner of production. Thus,
air is allowed to move through the nose for some speech sounds but not
others. Similarly, for some speech sounds, the flow of air is fully stopped for a
moment (e.g., [p], [b], and [t]). For other sounds, the air passage is restricted,
but air continues to flow (e.g., [f], [z], and [r]).
368 •
C H A P T E R T E N Language
FIGURE 10.2
THE HUMAN VOCAL TRACT
Dental
consonant
region
Palate
Soft palate
Nasal cavity
Oral cavity
Tongue
Lips
Vocal folds
(in the larynx)
Speech is produced by airflow from the lungs that passes through the larynx
and from there through the oral and nasal cavities. Different vowels are created by movements of the lips and tongue that change the size and shape
of the oral cavity. Consonants are produced by movements that temporarily
obstruct the airflow through the vocal tract.
Second, we can distinguish between sounds that are voiced — produced
with the vocal folds vibrating — and those that are not. The sounds of [v],
[z], and [n] (to name a few) are voiced; [f], [s], [t], and [k] are unvoiced. (You
can confirm this by running the hand-on-throat test while producing each of
these sounds.) Finally, sounds can be categorized according to where the airflow is restricted; this is referred to as place of articulation. For example, you
close your lips to produce “bilabial” sounds like [p] and [b]; you place your
top teeth close to your bottom lip to produce “labiodental” sounds like [f]
and [v]; and you place your tongue just behind your upper teeth to produce
“alveolar” sounds like [t] and [d].
This categorization scheme enables us to describe any speech sound in terms
of a few simple features. For example, what are the features of a [p] sound?
Phonology
•
369
First, we specify the manner of production: This sound is produced with air
moving through the mouth (not the nose) and with a full interruption to the
flow of air. Second, voicing: The [p] sound happens to be unvoiced. Third, place
of articulation: The [p] sound is bilabial. These features are all we need to identify the [p], and if any of these features changes, so does the sound’s identity.
In English, these features of sound production are combined and recombined to produce 40 or so different phonemes. Other languages use as few as
a dozen phonemes; still others use many more. (For example, there are 141
different phonemes in the language of Khoisan, spoken by the Bushmen of
Africa; Halle, 1990.) In all cases, though, the phonemes are created by simple
combinations of the features just described.
The Complexity of Speech Perception
This description of speech sounds invites a simple proposal about speech
perception. We’ve just said that each speech sound can be defined in terms of
a small number of features. Perhaps, then, all a perceiver needs to do is detect
these features, and with this done, the speech sounds are identified.
It turns out, though, that speech perception is more complicated. Consider
Figure 10.3, which shows the moment-by-moment sound amplitudes produced by a speaker uttering a brief greeting. It’s these amplitudes, in the form
of air-pressure changes, that reach the ear, and so, in an important sense, the
figure shows the pattern of input with which “real” speech perception begins.
Notice that within this stream of speech there are no markers to indicate
where one phoneme ends and the next begins. Likewise, there are, for the
most part, no gaps to indicate the boundaries between successive syllables
FIGURE 10.3
My
THE ACTUAL PATTERN OF SPEECH
name
is
Dan
Reis
−
berg
Shown here are the moment-by-moment sound amplitudes produced by the author uttering a greeting. Notice
that there is no gap between the sounds carrying the word “my” and the sounds carrying “name.” Nor is there
a gap between the sounds carrying “name” and the sounds carrying “is.” Therefore, the listener needs to figure
out where one sound stops and the next begins, a process known as “speech segmentation.”
370 •
C H A P T E R T E N Language
or successive words. Therefore, as your first step toward phoneme identification, you need to “slice” this stream into the appropriate segments — a step
known as speech segmentation.
For many people, this pattern comes as a surprise. Most of us are convinced that there are brief pauses between words in the speech that we hear,
and it’s these pauses, we assume, that mark the word boundaries. But this
perception turns out to be an illusion, and we are “hearing” pauses that
aren’t actually there. This is evident when we “hear” the pauses in the “wrong
places” and segment the speech stream in a way the speaker didn’t intend (see
Figure 10.4). The illusion is also revealed when we physically measure the
FIGURE 10.4
A MBIGUITY IN SEGMENTATION
“Boy, he must think we’re pretty stupid
to fall for that again.”
Almost every child has heard the story of Chicken Little. No one believed this
poor chicken when he announced, “The sky is falling!” It turns out, though,
that the acoustic signal — the actual sounds produced — would have been
the same if Chicken Little had exclaimed, “This guy is falling!” The difference
between these utterances (“The sky . . .” vs. “This guy . . .”) isn’t in the input.
Instead, the difference lies in how the listener segments the sounds.
Phonology
•
371
speech stream (as we did in order to create Figure 10.3) or when we listen
to speech we can’t understand — for example, speech in a foreign language.
In the latter circumstance, we lack the skill needed to segment the stream, so
we’re unable to “supply” the word boundaries. As a consequence, we hear
what is really there: a continuous, uninterrupted flow of sound. That is why
speech in a foreign language often sounds so fast.
Speech perception is further complicated by a phenomenon known as
coarticulation (Liberman, 1970; also Daniloff & Hammarberg, 1973). This
term refers to the fact that in producing speech, you don’t utter one phoneme
at a time. Instead, the phonemes overlap, so that while you’re producing
the [s] sound in “soup,” for example, your mouth is getting ready to say
the vowel. While uttering the vowel, you’re already starting to move your
tongue, lips, and teeth into position for producing the [p].
This overlap helps to make speech production faster and considerably
more fluent. But the overlap also has consequences for the sounds produced, so that the [s] you produce while getting ready for one upcoming
vowel is actually different from the [s] you produce while getting ready
for a different vowel. As a result, we can’t point to a specific acoustical
pattern and say, “This is the pattern of an [s] sound.” Instead, the acoustical pattern is different in different contexts. Speech perception therefore
has to “read past” these context differences in order to identify the phonemes produced.
Aids to Speech Perception
The need for segmentation in a continuous speech stream, the variations
caused by coarticulation, and also the variations from speaker to speaker
all make speech perception rather complex. Nonetheless, you manage to
perceive speech accurately and easily. How do you do it?
Part of the answer lies in the fact that the speech you encounter, day by
day, is surprisingly limited in its range. Each of us knows tens of thousands
of words, but most of these words are rarely used. In fact, we’ve known
for many years that the 50 most commonly used words in English make up
roughly half of the words you actually hear (Miller, 1951).
In addition, the perception of speech shares a crucial attribute with other
types of perception: a reliance on knowledge and expectations that supplement the input and guide your interpretation. In other words, speech perception (like perception in other domains) weaves together “bottom-up” and
“top-down” processes — processes that, on the one side, are driven by the
input itself, and, on the other side, are driven by the broader pattern of what
you know.
In perceiving speech, therefore, you don’t rely just on the stimuli you
receive (that’s the bottom-up part). Instead, you supplement this input with
other knowledge, guided by the context. This is evident, for example, in the
phonemic restoration effect. To demonstrate this effect, researchers start by
recording a bit of speech, and then they modify what they’ve recorded. For
example, they might remove the [s] sound in the middle of “legislatures” and
372 •
C H A P T E R T E N Language
replace the [s] with a brief burst of noise. This now-degraded stimulus can
then be presented to participants, embedded in a sentence such as
The state governors met with their respective legi*latures.
When asked about this stimulus, participants insist that they heard the
complete word, “legislatures,” accompanied by a burst of noise (Repp,
1992; Samuel, 1987, 1991). It seems, then, that they use the context to
figure out what the word must have been, but then they insist that they
actually heard the word. In fact, participants are often inaccurate if asked
when exactly they heard the noise burst. They can’t tell whether they heard
the noise during the second syllable of “legislatures” (so that it blotted
out the missing [s], forcing them to infer the missing sound) or at some
other point (so that they were able to hear the missing [s] with no
interference). Apparently, the top-down process literally changes what participants hear — leaving them with no way to distinguish what was heard
from what was inferred.
How much does the context in which we hear a word help us? In a
classic experiment, Pollack and Pickett (1964) recorded a number of
naturally occurring conversations. From these recordings, they spliced out
individual words and presented them in isolation to their research participants. With no context to guide them, participants were able to identify only half of the words. If restored to their original context, though,
the same stimuli were easy to identify. Apparently, the benefits of context
are considerable.
Categorical Perception
Speech perception also benefits from a pattern called categorical perception.
This term refers to the fact that people are much better at hearing the differences between categories of sounds than they are at hearing the variations
within a category of sounds. In other words, you’re very sensitive to the
differences between, say, a [g] sound and a [k], or the differences between a
[d] and a [t]. But you’re surprisingly insensitive to differences within each of
these categories, so you have a hard time distinguishing, say, one [p] sound
from another, somewhat different [p] sound. And, of course, this pattern is
precisely what you want, because it enables you to hear the differences that
matter without hearing (and being distracted by) inconsequential variations
within the category.
Demonstrations of categorical perception generally rely on a series of
stimuli, created by computer. The first stimulus in the series might be a [ba]
sound. Another stimulus might be a [ba] that has been distorted a tiny bit,
to make it a little bit closer to a [pa] sound. A third stimulus might be a [ba]
that has been distorted a bit more, so that it’s a notch closer to a [pa], and
so on. In this way we create a series of stimuli, each slightly different from
the one before, ranging from a clear [ba] sound at one extreme, through a
series of “compromise” sounds, until we reach at the other extreme a clear
[pa] sound.
CATEGORICAL
PERCEPTION IN
OTHER SPECIES
The pattern of categorical
perception isn’t limited to
language — or to humans. A
similar pattern, for example,
with much greater sensitivity to between-category
dif­ferences than to withincategory variations, has been
documented in the hearing
of the chinchilla.
Phonology
•
373
100
Percentage identifying
sounds as [pa]
Percentage
80
60
40
Percentage identifying
sounds as [ba]
20
0
CATEGORICAL
25
0
25
50
75
Voice-onset time (ms)
With computer speech, we can produce a
variety of compromises between a [pa] and
a [ba] sound, differing only in when the voicing begins (i.e., the voice-onset time, or VOT).
Panel A shows a plausible hypothesis about
how these sounds will be perceived: As the
sound becomes less and less like an ordinary
[ba], people should be less and less likely to
perceive it as a [ba]. Panel B, however, shows
the actual data: Research participants seem
indifferent to small variations in the [ba]
sound, and they categorize a sound with a
10 ms or 15 ms VOT in essentially the same
way that they categorize a sound with a 0
VOT. The categorizations also show an abrupt
categorical boundary between [pa] and [ba],
although there is no corresponding abrupt
change in the stimuli themselves.
(after lisker & abramson, 1970)
A Hypothetical identification data
100
Percentage identifying
sounds as [pa]
80
Percentage
FIGURE 10.5
PERCEPTION
60
40
20
Percentage identifying
sounds as [ba]
0
25
0
25
50
75
Voice-onset time (ms)
B Actual identification data
How do people perceive these various sounds? Figure 10.5A shows the pattern we might expect. After all, our stimuli are gradually shading from a clear
[ba] to a clear [pa]. Therefore, as we move through the series, we might expect
people to be less and less likely to identify each stimulus as a [ba], and correspondingly more and more likely to identify each as a [pa]. In the terms we
used in Chapter 9, this would be a “graded-membership” pattern: Test cases
close to the [ba] prototype should be reliably identified as [ba]; as we move
away from this prototype, cases should be harder and harder to categorize.
However, the actual data, shown in Figure 10.5B, don’t fit with this prediction. Even though the stimuli are gradually changing from one extreme to
another, participants “hear” an abrupt shift, so that roughly half the stimuli
374 •
C H A P T E R T E N Language
are reliably categorized as [ba] and half are reliably categorized as [pa].
Moreover, participants seem indifferent to the differences within each category.
Across the first dozen stimuli, the syllables are becoming less and less [ba]-like,
but this is not reflected in how the listeners identify the sounds. Likewise, across
the last dozen stimuli, the syllables are becoming more and more [pa]-like, but
again, this trend has little effect. What listeners seem to hear is either a [pa] or
a [ba], with no gradations inside of either category. (For early demonstrations,
see Liberman, Harris, Hoffman, & Griffith, 1957; Lisker & Abrahmson,
1970; for reviews, see Handel, 1989; Yeni-Komshian, 1993.)
It seems, then, that your perceptual apparatus is “tuned” to provide just
the information you need. After all, you want to know whether someone
advised you to “take a path” or “take a bath.” You certainly care whether
a friend said, “You’re the best” or “You’re the pest.” Plainly, the difference
between [b] and [p] matters to you, and this difference is clearly marked in
your perception. In contrast, you usually don’t care how exactly the speaker
pronounced “path” or “best” — that’s not information that matters for getting the meaning of these utterances. And here too, your perception serves
you well by largely ignoring these “subphonemic” variations. (For more on
the broad issue of speech perception, see Mattys, 2012.)
Combining Phonemes
English relies on just a few dozen phonemes, but these sounds can be combined
and recombined to produce thousands of different morphemes, which can
themselves be combined to create word after word after word. As we mentioned
earlier, though, there are rules governing these combinations, and users of the
language reliably respect these rules. So, in English, certain sounds (such as the
final sound in “going” or “flying”) can occur at the end of a word but not at
the beginning. Other combinations seem prohibited outright. For example, the
sequence “tlof” seems anomalous to English speakers; no words in English contain the “tl” combination within a single syllable. (The combination can, however, occur at the boundary between syllables — as in “motley” or “sweetly.”)
These limits, however, are simply facts about English; they are not at all a limit
on what human ears can hear or human tongues can produce, and other languages routinely use combinations that for English speakers seem unspeakable.
There are also rules governing the adjustments that occur when certain
phonemes are uttered one after another. For example, consider the “s” ending
that marks the English plural — as in “books,” “cats,” and “tapes.” In these
cases, the plural is pronounced as an [s]. In other contexts, though, the plural
ending is pronounced differently. Say these words out loud: “bags,” “duds,”
“pills.” If you listen carefully, you’ll realize that these words actually end with
a [z] sound, not an [s] sound.
English speakers all seem to know the rule that governs this distinction.
(The rule hinges on whether the base noun ends with a voiced or an unvoiced
sound; for classic statements of this rule, see Chomsky & Halle, 1968; Halle,
1990.) Moreover, they obey this rule even with novel, made-up cases. For
TEST YOURSELF
2.Define “voicing,”
“manner of production,” and “place of
articulation.”
3.What is speech segmentation, and why is
it an important step in
speech perception?
4.What is categorical
perception?
Phonology
•
375
COGNITION
outside the lab
“Read My Lips”
In 1988, presidential candidate George H. W. Bush
unmistakably hear the other sound. (Try it. There are
uttered the memorable instruction “Read my lips,”
many versions of this effect available on YouTube.)
and then he slowly enunciated “No . . . new . . .
It seems, then, that you have no choice about using
taxes.” Bush intended the initial phrase to mean
the lip cues, and when those cues are available to
something like, “Note what I’m saying. You can
you, they can change what you “hear.”
count on it.” Other speakers have picked up this
A different sort of evidence comes from settings
idiom, and today many people use the words
in which the input is easy to hear, but difficult to
“read my lips” to emphasize their message.
understand. Consider the case of someone who’s
Aside from this locution, though, what is lip-
had a year or two of training in a new language —
reading, and who uses it? People assume that
maybe someone who took two years of French in
lip-reading is a means of understanding speech
high school. This person now travels to Paris and is
based on visual cues, used when normal sound
able to communicate well enough in face-to-face
isn’t available. Of course, the set of cues avail-
conversation, but she’s hopelessly lost when trying
able to vision is limited, because many phonemes
to communicate in a phone call.
depend on movements or positions that are hid-
Can we document this pattern in the laboratory?
den inside of the mouth and throat. Even so, skilled
In one study, participants tried to understand some-
lip-readers (relying on a mix of visual cues, con-
one speaking in a language that the participants
text, and knowledge of the language) can glean
knew, but were not fluent in. (This is, of course, the
much of the content of the speech that they see.
situation of the French novice trying to get by in
However, we need to set aside the idea that
Paris.) In a second study, (English-speaking) par-
lip-reading is used only when the auditory signal
ticipants tried to understand someone speaking
is weak. Instead, lip-reading is an integral part of
English with a moderately strong foreign accent. In
ordinary speech perception. Of course, you often
a third study, the participants heard material that
don’t need lip-reading; if you did, you’d never be
was clearly spoken, with no unfamiliar accent, but
able to use the telephone or understand an inter-
was difficult to understand because the prose was
view on the radio. But even so, you rely on lip-
quite dense. (They were listening to a complex
reading in many settings — even if the acoustic
excerpt from the writings of the philosopher Im-
signal reaching your ears is perfectly clear.
manuel Kant.) In all cases, participants were able
Powerful evidence comes from the McGurk
to “hear” more if they could both see and hear the
effect, first described in a 1976 paper entitled “Hearing
speaker, in comparison to a condition in which there
Lips and Seeing Voices” (McGurk & MacDonald, 1976).
was no visual input.
In this effect, the audio track plainly conveys the
You shouldn’t be embarrassed, therefore, if you
sound of someone saying one sound (perhaps “ba”),
dread making a phone call in a language that’s not
but the carefully synchronized video shows some-
your native tongue. Whether you’re using your sec-
one uttering a different sound (“va”). If you listen to
ond language or your first, lip-reading is a normal part
the recording with eyes closed, you consistently hear
of speech perception, and at least part of what you
one sound; if you listen while watching the video, you
“hear” is actually coming to you through your eyes.
376 •
C H A P T E R T E N Language
example, I have one wug, and now I acquire another. Now, I have two . . .
what? Without hesitation, people pronounce “wugs” using the [z] ending — in
accord with the standard pattern. Even young children pronounce “wugs”
with a [z], and so, it seems, they too have internalized — and obey — the
relevant principles (Berko, 1958).
Morphemes and Words
A typical college graduate in the United States knows between 75,000 and
100,000 different words. These counts have been available for many years
(e.g., Oldfield, 1963; Zechmeister, Chronis, Cull, D’Anna, & Healy, 1995),
and there’s no reason to think they’re changing. For each word, the speaker
knows the word’s sound (the sequence of phonemes that make up the word)
and its orthography (the sequence of letters that spell the word). The speaker
also knows how to use the word within various phrases, governed by the
rules of syntax (see Figure 10.6). Finally, speakers know the meaning of
a word; they have a semantic representation for the word to go with the
phonological representation.
Building New Words
Estimates of vocabulary size, however, need to be interpreted with caution,
because the size of someone’s vocabulary is subject to change. One reason is
that new words are created all the time. For example, the world of computers
has demanded many new terms — with the result that someone who wants
to know something will often “google” it; many of us get information from
“blogs”; and most of us are no longer fooled by the “phishing” we sometimes
find in our “email.” The terms “software” and “hardware” have been around
for a while, but “spyware” and “malware” are relatively new.
FIGURE 10.6
(1)
(2)
(3)
(4)
KNOWING A WORD
* She can place the books on the table.
* She can place on the table.
* She can sleep the books on the table
* She can sleep on the table.
Part of what it means to “know a word” is knowing how to use a word. For
example, a verb like “place” requires an object — so that Sentence 1 (with an
object) sounds fine, but Sentence 2 is anomalous. Other words have other
requirements. “Sleep,” for example, does not take an object — so Sentence 3
is anomalous, but Sentence 4 is fine.
Morphemes and Words
•
377
TEST YOURSELF
5.Why is it difficult to
give an exact count
of the number of
words in someone’s
vocabulary?
Changes in social habits and in politics also lead to new vocabulary. It
can’t be surprising that slang terms come and go, but some additions to the
language seem to last. Changes in diet, for example, have put words like
“vegan,” “localvore/locavore,” and “paleo” into common use. The term
“metrosexual” has been around for a couple of decades, and likewise “buzzword.” It was only in 2012, though, that Time magazine listed “selfie” as one
of the year’s top ten buzzwords, and it was a 2016 vote in Great Britain that
had people talking about “Brexit.”
Often, these new words are created by combining or adjusting existing
words (and so “Brexit” combines “Britain” and “exit;” “paleo” is a shortened form of “Paleolithic”). In addition, once these new entries are in the
language, they can be combined with other elements — usually by adding
the appropriate morphemes. Imagine that you’ve just heard the word “hack”
for the first time. You know instantly that someone who does this activity
is a “hacker” and that the activity itself is “hacking,” and you understand
someone who says, “I’ve been hacked.”
Once again, therefore, note the generativity of language — that is, the
capacity to create an endless series of new combinations, all built from the
same set of fundamental units. Therefore, someone who “knows English” (or
someone who knows any language) hasn’t just memorized the vocabulary of
the language and some set of phrases. Instead, people who “know English”
know how to create new forms within the language: They know how to
combine morphemes to create new words, know how to “adjust” phonemes
when they’re put together into novel combinations, and so on. This knowledge isn’t conscious — and so most English speakers could not articulate the
principles governing the sequence of morphemes within a word, or why they
pronounce “wugs” with a [z] sound rather than an [s]. Nonetheless, speakers
honor these principles with remarkable consistency in their day-by-day use of
the language and in their day-to-day creation of novel words.
Syntax
The potential for producing new forms is even more remarkable when we
consider the upper levels in the language hierarchy — the levels of phrases
and sentences. This point becomes obvious when we ask: If you have 60,000
words in your vocabulary, or 80,000, how many sentences can you build
from those words?
Sentences range in length from the very brief (“Go!” or “I do”) to the
absurdly long. Most sentences, though, contain 20 words or fewer. With this
length limit, it has been estimated that there are 100,000,000,000,000,000,000
possible sentences in English (Pinker, 1994). If you could read these sentences
at the insane rate of 1,000 per second, you’d still need over 30,000 centuries to read through this list! (In fact, this estimate may be too low. Decades
before Pinker’s work, Miller, Galanter, & Pribram, 1960, estimated that the
number of possible sentences is actually 1030 — billions of times larger than
the estimate we’re using here.)
378 •
C H A P T E R T E N Language
Once again, though, there are limits on which combinations (i.e., which
sequences of words) are acceptable and which ones are not. For example, in
English you could say, “The boy hit the ball” but not “The boy hit ball the.”
Likewise, you could say, “The moose squashed the car” but not “The moose
squashed the” or just “Squashed the car.” Virtually any speaker of the language would agree that these errant sequences have something wrong in them,
but what exactly is the problem with these “bad” strings? The answer lies in
the rules of syntax — rules that govern the structure of a phrase or sentence.
One might think that the rules of syntax depend on meaning, so that meaningful sequences are accepted as “sentences” while meaningless sequences
are rejected as non-sentences. This suggestion, though, is wrong. As one
concern, many non-sentences do seem meaningful, and no one’s confused
when Sesame Street’s Cookie Monster insists “Me want cookie.” Likewise,
viewers understood the monster’s wistful comment in the 1935 movie Bride
of Frankenstein: “Alone, bad; friend, good.”
In addition, consider these two sentences:
’Twas brillig, and the slithy toves did gyre and gimble in the wabe.
Colorless green ideas sleep furiously.
(The first of these is from Lewis Carroll’s famous poem “Jabberwocky”; the
second was penned by the linguist Noam Chomsky.) These sentences are, of
course, without meaning: Colorless things aren’t green; ideas don’t sleep; toves
SYNTAX AND MORPHEMES IN “JABBERWOCKY”
In the poem “Jabberwocky,” Lewis Carroll relies on proper syntax
and appropriate use of morphemes to create gibberish that is
wonderfully English-like. “He left it dead, and with its head / He
went galumphing back.”
Syntax
•
379
aren’t slithy. Nonetheless, speakers of English, after a moment’s reflection,
regard these sequences as grammatically acceptable in a way that “Furiously
sleep ideas green colorless” is not. It seems, therefore, that we need principles
of syntax that are separate from considerations of semantics or sensibility.
Phrase Structure
YODA'S DIALECT
“Named must your fear be
before banish it you can.”
Yoda is, of course, a source
of great wisdom, and this
quotation is meaningful and
maybe even insightful. Even
so, the quotation is (at best)
syntactically odd. Apparently,
then, we need to distinguish
whether a word string is
meaningful from whether the
string is well formed according to the rules of syntax.
The rules of syntax take several forms, but they include rules that specify
which elements must appear in a phrase and (for some languages) that govern
the sequence of those elements. These phrase-structure rules also specify the
overall organization of the sentence — and therefore determine how the
various elements are linked to one another.
One way to depict phrase-structure rules is with a tree structure like the
one shown in Figure 10.7. You can read the structure from top to bottom,
and as you move from one level to the next, you can see that each element
(e.g., a noun phrase or a verb phrase) has been “expanded” in a way that’s
strictly governed by the phrase-structure rules.
Prescriptive Rules, Descriptive Rules
We need to be clear, though, about what sorts of rules we’re discussing. Let’s
begin with the fact that most of us were taught, at some stage of our education,
how to talk and write “properly.” We were taught never to say “ain’t.” Many
FIGURE 10.7
A PHRASE STRUCTURE TREE
S
NP
VP
V
det
A
N
The
ferocious
dragon
chased
NP
det
A
N
the
timid
mouse
The diagram shows that the overall sentence (S) consists of a noun phrase
(NP) plus a verb phrase (VP). The noun phrase is composed of a determiner
(det) followed by an adjective (A) and a noun (N). The verb phrase is
composed of a verb (V) followed by a noun phrase (NP).
380 •
C H A P T E R T E N Language
of us were scolded for writing in the passive voice or starting a sentence with
“And.” Warnings like these are the result of prescriptive rules — rules describing how something (in this case: language) is “supposed to be.” Language that
doesn’t follow these rules, it’s claimed, is “improper” or maybe even “bad.”
You should, however, be skeptical about these prescriptive rules. After all,
languages change with the passage of time, and what’s “proper” in one period
is often different from what seems right at other times. In the 1600s, for example, people used the pronouns “thou” and “ye,” but those words are gone from
modern usage. In more recent times, people just one generation back insisted
it was wrong to end a sentence with a preposition; modern speakers think this
prohibition is silly. Likewise, consider the split infinitive. Prominent writers of
the 18th and 19th centuries (e.g., Ben Franklin, William Wordsworth, Henry
James) commonly split their infinitives; grammarians of the early 20th century,
in contrast, energetically condemned this construction. Now, in the 21st
century, most English speakers seem entirely indifferent to whether their infinitives are split or not (and may not even know what a split infinitive is).
This pattern of change makes it difficult to justify prescriptive rules. Some
people, for example, still insist that split infinitives are improper and must be
avoided. This suggestion, however, seems to rest on the idea that the English
spoken in, say, 1926 was proper and correct, and that the English spoken a
few decades before or after this “Golden Age” is somehow inferior. It’s hard
to think of any basis for this claim, so it seems instead that this prescriptive
rule reflects only the habits and preferences of a particular group at a particular time — and there’s no reason why our usage should be governed by their
preferences. In addition, it’s not surprising that the groups that set these rules
are usually groups with high prestige or social standing (Labov, 2007). When
people strive to follow prescriptive rules, then, it’s often because they hope to
join (or, at least, be associated with) these elite groups.
Phrase-structure rules, in contrast, are not prescriptive; they are descriptive rules — that is, rules characterizing the language as it’s ordinarily used
by fluent speakers and listeners. There are, after all, strong regularities in the
way English is used, and the rules we’re discussing here describe these patterns. No value judgment is offered about whether these patterns constitute
“proper” or “good” English. These patterns simply describe how English is
structured — or perhaps we should say, what English is.
The Function of Phrase Structure
No one claims that language users are consciously aware of phrase-structure
rules. Instead, the idea is that people have somehow internalized these rules
and obey the rules in their use of, and judgments about, language.
For example, your intuitions about whether a sentence is well formed
or not respect phrase-structure rules — and so, if a sequence of words lacks
an element that should, according to the rules, be in place, you’ll probably
think there’s a mistake in the sequence. Likewise, you’ll balk at sequences of
words that include elements that (according to the rules) shouldn’t be there,
THE (SOMETIMES)
PECULIAR NATURE
OF PRESCRIPTIVE
RULES
According to an oftenrepeated story, an editor had
rearranged one of Winston
Churchill’s sentences to bring
it into alignment with “proper”
English. Specifically, the editor
rewrote the sentence to avoid
ending it in a preposition. In
response, the prime minister,
proud of his style, scribbled
this note: “This is the sort of
English up with which I will
not put.” (Often repeated or
not, we note that there’s debate about the historical roots
of this story!)
Syntax
•
381
or elements that should be in a different position within the string. These
points allow us to explain why you think sequences like these need some sort
of repair: “His argument emphasized in modern society” or “Susan appeared
cat in the door.”
Perhaps more important, phrase-structure rules help you understand the
sentences you hear or read, because syntax in general specifies the relationships among the words in each sentence. For example, the NP + VP sequence
typically divides a sentence into the “doer” (the NP) and some information
about that doer (the VP). Likewise, the V + NP sequence usually indicates the
action described by the sentence and then the recipient of that action. In this
way, the phrase structure of a sentence provides an initial “road map” that’s
useful in understanding the sentence. For a simple example, it’s syntax that
tells us who’s doing what when we hear “The boy chased the girl.” Without
syntax (e.g., if our sentences were merely lists of words, such as “boy, girl,
chased”), we’d have no way to know who was the chaser and who (if anyone)
was chaste. (Also see Figure 10.8.)
Sometimes, though, two different phrase structures can lead to the same
sequence of words, and if you encounter one of these sequences, you may
not know which structure was intended. How will this affect you? We’ve just
suggested that phrase structures guide interpretation, and so, with multiple
phrase structures available, there should be more than one way to interpret
the sentence. This turns out to be correct — often, with comical consequences
(see Figure 10.9).
TEST YOURSELF
6.What evidence tells us
that the rules of syntax can be separated
from considerations
of whether or not a
string of words has
meaning?
7.What are phrasestructure rules, and
what does it mean
that these rules are
“descriptive,” not
“prescriptive”?
Sentence Parsing
A sentence’s phrase structure, we’ve said, conveys crucial information about
who did what to whom. Once you know the phrase structure, therefore,
you’re well on your way to understanding the sentence. But how do you
figure out the phrase structure in the first place? This would be an easy question if sentences were uniform in their structure: “The boy hit the ball. The
girl drove the car. The elephant trampled the geraniums.” But, of course,
The large tomato
The
made
large tomato made
a satisfying splat
a satisfying
when
splat when it
it hit
hit the
the floor.
floor.
A
382 •
B
C H A P T E R T E N Language
FIGURE 10.8 PHRASE STRUCTURE
ORGANIZATION AIDS THE READER
Panel A shows a sentence written so that the breaks between
lines correspond to breaks between phrases; this makes
reading easier because the sentence has been visually “preorganized.” In Panel B, the sentence has been rewritten so
that the visual breaks don’t correspond to the boundaries
between phrases. Reading is now slower and more difficult.
FIGURE 10.9
PHRASE STRUCTURE AMBIGUITY
VP
V
NP
discuss
N
sex
VP
PP
with
NP
V
NP
JK
discuss
N
sex
PP
P
NP
with
JK
He wants to discuss sex with Jimmy Kimmel.
I saw the gorilla in my pajamas.
The shooting of the hunters was terrible.
They are roasting chickens.
Visiting relatives can be awful.
Two computers were reported stolen by the TV announcer.
Often, the words of a sentence are compatible with more than one phrase
structure; in such cases, the sentence will be ambiguous. Therefore, you can
understand the first sentence here either as describing a discussion with
Kimmel or as describing sex with Kimmel; both analyses of the verb phrase
are shown. Can you find both interpretations for the remaining sentences?
sentences are more variable than this, and this variation makes the identification of a sentence’s phrase structure much more difficult.
How, therefore, do you parse a sentence — that is, figure out each word’s
syntactic role? It seems plausible that you’d wait until the sentence’s end,
and only then go to work on figuring out the structure. With this strategy,
your comprehension might be slowed a little (because you’re waiting for the
sentence’s end), but you’d avoid errors, because your interpretation could be
guided by full information about the sentence’s content.
It turns out, though, that people don’t use this wait-for-all-the-information
strategy. Instead, they parse sentences as they hear them, trying to figure out
the role of each word the moment it arrives (e.g., Marcus, 2001; Savova,
Roy, Schmidt, & Tenenbaum, 2007; Tanenhaus & Trueswell, 2006). This
approach is efficient (since there’s no waiting) but, as we’ll see, can lead
to errors.
Sentence Parsing
•
383
NOAH’S ARC
Sometimes linguistic ambiguity involves the interpretation of a phrase’s organization. Sometimes, though, the ambiguity involves the interpretation of a single word.
Sometimes the ambiguity is evident in spoken language but not in written language.
Garden Paths
Even simple sentences can be ambiguous if you’re open-minded (or perverse)
enough:
Mary had a little lamb. (But I was quite hungry, so I had the lamb and
also a bowl of soup.)
Time flies like an arrow. (But fruit flies, in contrast, like a banana.)
Temporary ambiguity is also common inside a sentence. More precisely, the
early part of a sentence is often open to multiple interpretations, but then the
later part of the sentence clears things up. Consider this example:
The old man the ships.
In this sentence, most people read the initial three words as a noun phrase:
“the old man.” However, this interpretation leaves the sentence with no verb,
384 •
C H A P T E R T E N Language
so a different interpretation is needed, with the subject of the sentence being
“the old” and with “man” being the verb. (Who mans the ships? It is the old,
not the young. The old man the ships.) Likewise:
The secretary applauded for his efforts was soon promoted.
Here, people tend to read “applauded” as the sentence’s main verb, but it
isn’t. Instead, this sentence is just a shorthand way of answering the question,
“Which secretary was soon promoted?” (Answer: “The one who was
applauded for his efforts.”)
These examples are referred to as garden-path sentences: You’re
initially led to one interpretation (you are, as they say, “led down the
garden path”), but this interpretation then turns out to be wrong. So you
need to reject your first interpretation and find an alternative. Here are two
more examples:
Fat people eat accumulates.
Because he ran the second mile went quickly.
Garden-path sentences highlight the risk attached to the strategy of interpreting a sentence as it arrives: The information you need in order to understand these sentences arrives only late in the sequence, and so, to avoid an
interpretive dead end, you’d be better off remaining neutral about the sentence’s meaning until you’ve gathered enough information. That way, you’d
know that “the old man” couldn’t be the sentence’s subject, that “applauded”
couldn’t be the sentence’s main verb, and so on. But this isn’t what you do.
Instead, you commit yourself fairly early to one interpretation and then try
to “fit” subsequent words, as they arrive, into that interpretation. This stra­
tegy is often effective, but it does lead to the “double-take” reaction when
late-arriving information forces you to abandon your initial interpretation
(Grodner & Gibson, 2005).
Syntax as a Guide to Parsing
What is it that leads you down the garden path? Why do you initially choose
one interpretation of a sentence, one parsing, rather than another? Many
cues are relevant, because many types of information influence parsing. For
one, people usually seek the simplest phrase structure that will accommodate the words heard so far. This strategy is fine if the sentence structure
is indeed simple; the strategy produces problems, though, with more complex sentences. To see how this plays out, consider the earlier sentence, “The
secretary applauded for his efforts was soon promoted.” As you read
“The secretary applauded,” you had the option of interpreting this as a noun
phrase plus the beginning of a separate clause modifying “secretary.” This
is the correct interpretation, and it’s required by the way the sentence ends.
However, you ignored this possibility, at least initially, and went instead with
Sentence Parsing
•
385
a simpler interpretation — of a noun phrase plus verb, with no idea of a separate embedded clause.
People also tend to assume that they’ll be hearing (or reading) active-voice
sentences rather than passive-voice sentences, so they generally interpret a
sentence’s initial noun phrase as the “doer” of the action and not the recipient. As it happens, most of the sentences you encounter are active, not passive, so this assumption is usually correct (for early research, see Hornby,
1974; Slobin, 1966; Svartik, 1966). However, this assumption can slow you
down when you do encounter a passive sentence, and, of course, this assumption added to your difficulties with the “secretary” sentence: The embedded
clause there is in the passive voice (the secretary was applauded by someone else); your tendency to assume active voice, therefore, works against the
correct interpretation of this sentence.
Not surprisingly, parsing is also influenced by the function words that
appear in a sentence and by the various morphemes that signal syntactic role
(Bever, 1970). For example, people easily grasp the structure of “He gliply
rivitched the flidget.” That’s because the “-ly” morpheme indicates that “glip”
is an adverb; the “-ed” identifies “rivitch” as a verb; and “the” signals that
“flidget” is a noun — all excellent cues to the sentence structure. This factor, too, is relevant to the “secretary” sentence, which included none of the
helpful function words. Notice that we didn’t say, “The secretary who was
applauded . . .”; if we had said that, the chance of misunderstanding would
have been greatly reduced.
With all these factors stacked against you, it’s no wonder you were
(temporarily) confused about “the secretary.” Indeed, with all these factors
in place, garden-path sentences can sometimes be enormously difficult to
comprehend. For example, spend a moment puzzling over this (fully grammatical) sequence:
The horse raced past the barn fell.
(If you get stuck with this sentence, try adding the word “that” after “horse.”)
Background Knowledge as a Guide to Parsing
Parsing is also guided by background knowledge, and in general, people
try to parse sentences in a way that makes sense to them. So, for example, readers are unlikely to misread the headline Drunk Gets Six Months in
Violin Case (Gibson, 2006; Pinker, 1994; Sedivy, Tanenhaus, Chambers, &
Carlson, 1999). And this point, too, matters for the “secretary” sentence:
Your background knowledge tells you that women secretaries are more common than men, and this added to your confusion in figuring out who was
applauding and who was applauded.
How can we document these knowledge effects? Several studies have
tracked how people move their eyes while reading, and these movements
can tell us when the reading is going smoothly and when the reader
is confused. Let’s say, then, that we ask someone to read a garden-path
386 •
C H A P T E R T E N Language
FIGURE 10.10
INTERPRETING COMPLEX SENTENCES
A The detectives examined by the reporter
revealed the truth about the robbery.
B The evidence examined by the reporter
revealed the truth about the robbery.
Readers are momentarily confused when they reach the “by the reporter” phrase in Sentence A. That is because they initially interpreted “examined” as the sentence’s main verb. Readers aren’t confused by Sentence B,
though, because their background knowledge told them that “examined” couldn’t be the main verb (because
evidence is not capable of examining anything). Notice, though, that readers won’t be confused if the sentences
are presented as they are here — with a picture. In that case, the extralinguistic context guides interpretation and
helps readers avoid the garden path.
sentence. The moment the person realizes he has misinterpreted the words
so far, he’ll backtrack and reread the sentence’s start, and, with appropriate instruments, we can easily detect these backwards eye movements
(MacDonald, Pearlmutter, & Seidenberg, 1994; Trueswell, Tanenhaus, &
Garnsey, 1994).
Using this technique, investigators have examined the effects of plausibility on readers’ interpretations of the words they’re seeing. For example,
participants might be shown a sentence beginning “The detectives examined . . . ”; upon seeing this, the participants sensibly assume that “examined”
is the sentence’s main verb and are therefore puzzled when the sentence continues “by the reporter . . .” (see Figure 10.10A). We detect this puzzlement
in their eye movements: They pause and look back at “examined,” realizing
that their initial interpretation was wrong. Then, after this recalculation, they
press onward.
Things go differently, though, if the sentence begins “The evidence
examined . . . ” (see Figure 10.10B). Here, readers can draw on the fact that
“evidence” can’t examine anything, so “examined” can’t be the sentence’s
main verb. As a result, they’re not surprised when the sentence continues “by
the reporter . . .” Their understanding of the world had already told them that
Sentence Parsing
•
387
FIGURE 10.11
SEMANTIC AND SYNTACTIC PROCESSING
N400
250–300 ms
300–350 ms
350–400 ms
400–450 ms
450–500 ms
500–550 ms
550–600 ms
A Electrical activity in the brain after hearing a sentence that violates semantic expectations
LAN
B Electrical activity in the brain after hearing a sentence that violates
syntactic expectations
Semantic –5.5
–1.2
3.0
µV
Syntactic –3.1
–0.4
2.3
Many types of information influence parsing. The figures here show patterns of electrical activity on the scalp
(with different voltages represented by different colors). (Panel A) If the person hears a sentence that violates
semantic expectations (e.g., a sentence like, “He drinks his coffee with cream and dog”), this triggers a brain
wave termed the N400 (so-called because the wave involves a negative voltage roughly 400 ms after the trigger “dog” is encountered). (Panel B) If the person hears a sentence that violates syntactic expectations, though
(e.g., a sentence like, “He prefers to solve problems herself”), a different brain wave is observed — the so-called
left anterior negativity (LAN).
the first three words were the start of a passive sentence, not an active one.
(Also see Figures 10.11 and 10.12.)
The Extralinguistic Context
We’ve now mentioned several strategies that you use in parsing the sentences
you encounter. The role of these strategies is obvious when the strategies
mislead you, as they do with garden-path sentences. Bear in mind, though,
that the same strategies are used for all sentences and usually do lead to the
correct parsing.
It turns out, however, that our catalogue of strategies isn’t complete,
because you also make use of another factor: the context in which you
encounter sentences, including the conversational context. For example, the
garden-path problem is much less likely to occur in the following setting:
388 •
C H A P T E R T E N Language
FIGURE 10.12
Cz
N400 BRAIN WAVE
–6
N400
Amplitude (µV)
–4
–2
0
2
4
6
0
400
200
600
Time (ms)
Correct:
The Dutch trains are yellow and very crowded.
Semantic violation:
The Dutch trains are sour and very crowded.
World knowledge violation:
The Dutch trains are white and very crowded.
In parsing a sentence, you rely on your (nonlinguistic) knowledge about the world. This point is evident in a study
of electrical activity in the brain while people were hearing different types of sentences. Some of the sentences
were sensible and true (“The Dutch trains are yellow and very crowded”). Other sentences contained a semantic
anomaly (“The Dutch trains are sour and very crowded”), and this peculiarity produced the N400 brain wave.
The key, though, is that a virtually identical N400 was produced in a third condition in which sentences were
perfectly sensible but false: “The Dutch trains are white and very crowded.” (The falsity was immediately obvious
to the Dutch participants in this study.) Apparently, world knowledge (including knowledge about train color)
is a part of sentence processing from a very early stage.
(fig. 1 from hagoort et al., “integration of word meaning
and world knowledge in language comprehension,” science 304 [april 2004]: 438–441. © 2004 aaas. reprinted with permission
from aaas.)
Jack: Which horse fell?
Kate: The horse raced past the barn fell.
Just as important is the extralinguistic context — the physical and social
setting in which you encounter sentences. To see how this factor matters,
consider the following sentence:
Put the apple on the towel into the box.
Sentence Parsing
•
389
FIGURE 10.13
CONTEXT
THE EXTRALINGUISTIC
“Put the apple on the towel into the box.” Without the setting shown here, this sentence causes
momentary confusion: The listener will initially think
she’s supposed to put the apple onto the towel and
is then confused by the sentence’s last three words.
If, however, the sentence is spoken in a setting like
the one shown in this picture, there’s no confusion.
Now, the listener immediately sees the ambiguity
(which apple is being discussed?), counts on the
speaker to provide clarification for this point, and so
immediately understands “on the towel” as specification, not a destination.
TEST YOURSELF
8.What’s the evidence
that multiple factors
play a role in guiding
how you parse a
sentence?
9.What is a garden-path
sentence?
At its start, this sentence seems to be an instruction to put an apple onto a towel;
this interpretation must be abandoned, though, when the words “into the box”
arrive. Now, you realize that the box is the apple’s destination; “on the towel”
is simply a specification of which apple is to be moved. (Which apple should be
put into the box? The one that’s on the towel.) In short, this is another gardenpath sentence — initially inviting one analysis but eventually requiring another.
This confusion is avoided, however, if the sentence is spoken in the appropriate setting. Imagine that two apples are in view, as shown in Figure 10.13.
In this context, a listener hearing the sentence’s start (“Put the apple . . .”)
would immediately see the possibility for confusion (which apple is being
referred to?) and so would expect the speaker to specify which one is to be
moved. Therefore, when the phrase “on the towel” comes along, the listener
immediately understands it (correctly) as the needed specification. There is
no confusion and no garden path (Eberhard, Spivey-Knowlton, Sedivy, &
Tanenhaus, 1995; Tanenhaus & Spivey-Knowlton, 1996).
Prosody
One other cue is also useful in parsing: the rise and fall of speech intonation and the pattern of pauses. These pitch and rhythm cues, together called
prosody, can communicate a great deal of information. Prosody can, for example, reveal the mood of a speaker; it can also direct the listener’s attention by
specifying the focus or theme of a sentence (Jackendoff, 1972; also see Kraus
& Slater, 2016). Consider the simple sentence “Sol sipped the soda.” Now,
imagine how you’d pronounce this sentence in response to each of these questions: “Was it Sue who sipped the soda?”; “Did Sol gulp the soda?”; or “Did
390 •
C H A P T E R T E N Language
Sol sip the soup?” You’d probably say the same words (“Sol sipped the soda”)
in response to each of these queries, but you’d adjust the prosody in order to
highlight the information crucial for each question. (Try it. Imagine answering
each question, and pay attention to how you shift your pronunciation.)
Prosody can also render unambiguous a sentence that would otherwise be
entirely confusing (Beach, 1991). This is why printed versions of garden-path
sentences, and ambiguous sentences in general, are more likely to puzzle you,
because in print prosody provides no information. Imagine, therefore, that you
heard the sentence “The horse raced past the barn fell.” The speaker would probably pause momentarily between “horse” and “raced,” and again between “barn”
and “fell,” making it likely that you’d understand the sentence with no problem.
As a different example, consider two objects you might buy for your
home. One is a small box designed as a house for bluebirds. The other is a
small box that can be used by any type of bird, and the box happens to be
painted blue. In print, we’d call the first of these a “bluebird house,” and the
second a “blue birdhouse.” But now, pronounce these phrases out loud, and
you’ll notice how prosody serves to distinguish these two structures.
Some aspects of prosody depend on the language being spoken, and even
on someone’s dialect within a language. Other prosodic cues — especially
cues that signal the speaker’s emotions and attitudes — seem to be shared
across languages. This point was noted more than a century ago by Charles
Darwin (1871) and has been amply confirmed in the years since then (e.g.,
Bacharowski, 1999; Pittham & Scherer, 1993).
TEST YOURSELF
10. What is prosody?
11. Why are printed
versions of gardenpath sentences more
likely to puzzle you,
compared to spoken
versions of the same
sentences?
Pragmatics
What does it mean to “know a language” — to “know English,” for example?
It should be clear by now that the answer has many parts. Any competent
language user needs somehow to know (and obey) a rich set of rules about
how (and whether) elements can be combined. Language users rely on a further set of principles whenever they perceive and understand linguistic inputs.
Some of these principles are rooted in syntax; others depend on semantics
(e.g., knowing that detectives can “examine” but evidence can’t); still others
depend on prosody or on the extralinguistic context. All these factors then
seem to interact, so that your understanding of the sentences you hear (or see
in print) is guided by all these principles at the same time.
These points, however, still understate the complexity of language use and,
with that, the complexity of the knowledge someone must have in order to
use a language. This point becomes clear when we consider language use at
levels beyond the hierarchy shown in Figure 10.1 — for example, when we
consider language as it’s used in ordinary conversation. As an illustration, consider the following bit of dialogue (after Pinker, 1994; also see Gernsbacher
& Kaschak, 2013; Graesser & Forsyth, 2013; Zwaan, 2016):
Woman: I’m leaving you.
Man: Who is he?
Pragmatics
•
391
TEST YOURSELF
12. “What happened
to the roast beef?”
“The dog sure looks
happy.” Explain
what happened in
this conversational
exchange, and how
the exchange will be
understood.
You easily provide the soap-opera script that lies behind this exchange,
but you do so by drawing on a fabric of additional knowledge — in this case,
knowledge about the vicissitudes of romance. Likewise, in Chapter 1 we
talked about the importance of background knowledge in your understanding of a simple story. (It was the story that began, “Betsy wanted to bring
Jacob a present . . . .”). There, too, your understanding depended on you
providing a range of facts about gift-giving, piggy banks, and more. Without
those facts, the story would have been incomprehensible.
Your use of language also depends on your assumptions about how, in
general, people communicate with each other — assumptions that involve the
pragmatics of language use. For example, if someone asks, “Do you know
the time?” you understand this as a request that you report the time — even
though the question, understood literally, is a yes/no question about the
extent of your temporal knowledge.
What do the pragmatics of language — that is, your knowledge of how
language is ordinarily used — actually involve? Many years ago, philosopher
Paul Grice described the conversational “rules” in terms of a series of maxims that speakers follow and listeners count on (Grice, 1989). The “maxim
of relation,” for example, says that speakers should say things that are rele­
vant to the conversation. For example, imagine that someone asks, “What
happened to the roast beef?” and gets a reply, “The dog sure looks happy.”
Here, your assumption of relevance will most likely lead you to infer that
the dog must have stolen the meat. Likewise, the “maxim of quantity” specifies that a speaker shouldn’t be more informative than is necessary. On this
point, imagine that you ask someone, “What color are your eyes?” and he
responds, “My left eye is blue.” The extra detail here invites you to assume
that the speaker specified “left eye” for a reason — and so you’ll probably
infer that the person’s right eye is some other color. In these ways, listeners
count on speakers to be cooperative and collaborative, and speakers proceed
knowing that listeners make these assumptions. (For more on the collaborative nature of conversation and the assumptions that conversational partners
make, see Andor, 2011; Clark, 1996; Davis & Friedman, 2007; Graesser,
Millis, & Zwaan, 1997; Holtgraves, 2002; Noveck & Reboul, 2008; Noveck
& Sperber, 2005).
The Biological Roots of Language
Each of us uses language all the time — to learn, to gossip, to instruct, to persuade, to warn, to express affection. We use this tool as easily as we breathe;
we spend far more effort in choosing our clothes in the morning than we do
in choosing the words we will speak. But these observations must not hide
the facts that language is a remarkably complicated tool and that we are all
exquisitely skilled in its use.
How is all of this possible? How is it that ordinary human beings — even
ordinary two-and-a-half-year-olds — manage the extraordinary task of
392 •
C H A P T E R T E N Language
mastering and fluently using language? According to many authors, the
answer lies in the fact that humans are equipped with sophisticated neural
machinery specialized for learning, and then using, language. Let’s take a
quick look at this machinery.
Aphasias
As we described at the chapter’s start, damage to specific parts of the brain
can cause a disruption of language known as aphasia. Damage to the brain’s
left frontal lobe, especially a region known as Broca’s area (see Figure 10.14),
usually produces a pattern of symptoms known as nonfluent aphasia. People
with this disorder can understand language they hear but cannot write or
speak. In extreme cases, a patient with this disorder cannot utter any words
at all. In less severe cases, only a part of the patient’s vocabulary is lost, but
the patient’s speech becomes labored and fragmented, and articulating each
FIGURE 10.14
RAIN AREAS CRUCIAL FOR THE PERCEPTION AND
B
PRODUCTION OF LANGUAGE
Tongue Jaw
Throat
Lips
Motor projection areas
related to speech
Broca’s area
A
Wernicke’s area
Auditory projection area
B
Panel A shows some of the many brain regions that are crucial in supporting the comprehension and production
of language. For most individuals, most of these regions are in the left cerebral hemisphere (as shown here).
Broca’s area (named after the physician Paul Broca) is heavily involved in language production; Wernicke’s area
(named after the physician Karl Wernicke) plays a crucial role in language comprehension. Panel B shows a
photograph of the actual brain of Broca’s patient “Tan.” Because of his brain damage, this patient was no longer
able to say anything other than the syllable “Tan” — leading to the nickname that’s often used for him. This pattern (along with observations gained through Tan’s autopsy) led Broca to propose that a specific brain region
is crucial for speech.
The Biological Roots of Language
•
393
word requires special effort. One early study quoted a patient with aphasia
as saying, “Here . . . head . . . operation . . . here . . . speech . . . none . . .
talking . . . what . . . illness” (Luria, 1966, p. 406).
Different symptoms are associated with damage to a brain site known as
Wernicke’s area (again see Figure 10.14). Patients with this sort of damage
usually suffer from a pattern known as fluent aphasia. These patients can
talk freely, but they say very little. One patient, for example, uttered, “I was
over the other one, and then after they had been in the department, I was in
this one” (Geschwind, 1970, p. 904). Or another patient: “Oh, I’m taking
the word the wrong way to say, all of the barbers here whenever they stop
you it’s going around and around, if you know what I mean, that is tying
and tying for repucer, repuceration, well, we were trying the best that we
could while another time it was with the beds over there the same thing”
(Gardner, 1974, p. 68).
This distinction between fluent and nonfluent aphasia, however, captures
the data only in the broadest sense. One reason lies in the fact that — as we’ve
seen — language use involves the coordination of many different steps, many
different processes. These include processes needed to “look up” word meanings in your “mental dictionary,” processes needed to figure out the structural
relationships within a sentence, processes needed to integrate information
about a sentence’s structure with the meanings of the words within the sentence, and so on. Each of these processes relies on its own set of brain pathways, so damage to those pathways disrupts the process. As a result, the
language loss in aphasia can sometimes be quite specific, with impairment
just to a specific processing step (Cabeza & Nyberg, 2000; Demonet, Wise,
& Frackowiak, 1993; Martin, 2003).
Even with these complexities, the point here is that humans have a considerable amount of neural tissue that is specialized for language. Damage to
this tissue can disrupt language understanding, language production, or both.
In all cases, though, the data make it clear that our skill in using language
rests in part on the fact that we have a lot of neural apparatus devoted to
precisely this task.
The Biology of Language Learning
The biological roots of language also show up in another manner — in the
way that language is learned. This learning occurs remarkably rapidly, and
so, by the age of 3 or 4, almost every child is able to converse at a reasonable
level. Moreover, this learning can proceed in an astonishingly wide range of
environments. Children who talk a lot with adults learn language, and so
do children who talk very little with adults. In fact, children learn language
even if their communication with adults is strictly limited. Evidence on this
last point comes from children who are born deaf and have no opportunity
to learn sign language. (In some cases, this is because their caretakers don’t
know how to sign; in other cases, it’s because their caretakers choose not
to teach signing.) Even in these extreme cases, language emerges: Children
394 •
C H A P T E R T E N Language
SIGN LANGUAGE
Across the globe, humans speak many different languages —
English, Hindi, Mandarin, Quechua, to name just a few. Many
humans, though, communicate through sign language.
Actually, there are multiple sign languages, and so, for example, American Sign Language (ASL) is quite different from
South African Sign Language or Danish Sign Language. In
all cases, though, sign languages are truly languages, with
all of the richness and complexity of oral languages. Indeed,
sign languages show many of the fundamental properties
of oral languages, and so (for example) they have complex
grammars of their own.
in this situation invent their own gestural language (called “home sign”)
and teach the language to the people in their surroundings (Feldman,
Goldin-Meadow, & Gleitman, 1978; Goldin-Meadow, 2003, 2017; Senghas,
Román, & Mavillapalli, 2006).
How should we think about this? According to many psychologists, the
answer lies in highly sophisticated learning capacities that have specifically
evolved for language learning. Support for this claim comes from many
sources, including observations of specific-language impairment (SLI). Children with this disorder have normal intelligence and no problems with the
muscle movements needed to produce language. Nonetheless, they are slow
to learn language and, throughout their lives, have difficulty in understanding
and producing many sentences. They are also impaired on tasks designed to
test their linguistic knowledge. They have difficulty, for example, completing
passages like this one: “I like to blife. Today I blife. Tomorrow I will blife.
Yesterday I did the same thing. Yesterday I ______.” Most 4-year-olds know
that the answer is “Yesterday I blifed.” But adults with SLI cannot do this
task — apparently having failed to learn the simple rule of language involved
in forming the past tense of regular verbs (Bishop & Norbury, 2008;
Lai, Fisher, Hurst, Vargha-Khadem, & Monaco, 2001; van der Lely, 2005;
van der Lely & Pinker, 2014).
Claims about SLI remain controversial, but many authors point to this
disorder as evidence for brain mechanisms that are somehow specialized for
language learning. Disruption to these mechanisms throws language off track
but seems to leave other aspects of the brain’s functioning undisturbed.
The Processes of Language Learning
Even with these biological contributions, there’s no question that learning
plays an essential role in the acquisition of language. After all, children who
grow up in Paris learn to speak French; children who grow up in Beijing learn
to speak Chinese. In this rather obvious way, language learning depends on
the child’s picking up information from her environment.
The Biological Roots of Language
•
395
But what learning mechanisms are involved here? Part of the answer rests
on the fact that children are exquisitely sensitive to patterns and regularities
in what they hear, as though each child were an astute statistician, keeping track of the frequency-of-occurrence of this form or that. In one study,
8-month-old infants heard a 2-minute recording that sounded something like
“bidakupadotigolabubidaku.” These syllables were spoken in a monotonous
tone, with no difference in stress from one syllable to the next and no pauses
in between the syllables. But there was a pattern. The experimenters had
decided in advance to designate the sequence “bidaku” as a word. Therefore, they arranged the sequences so that if the infant heard “bida,” then
“ku” was sure to follow. For other syllables, there was no such pattern. For
instance, “daku” (the end of the nonsense word “bidaku”) would sometimes
be followed by “go,” sometimes by “pa,” and so on.
The babies reliably detected these patterns. In a subsequent test, babies
showed no surprise if they heard the string “bidakubidakubidaku.” From the
babies’ point of view, these were simply repetitions of a word they already
knew. However, the babies showed surprise if they were presented with the
string “dakupadakupadakupa.” This wasn’t a “word” they had heard before,
although they had heard each of its syllables many times. It seems, then,
that the babies had learned the “vocabulary” of this made-up language. They
had detected the statistical pattern of which syllables followed which, despite
their brief, passive exposure to these sounds and despite the absence of any
supporting cues such as pauses or shifts in intonation (Aslin, Saffran, &
Newport, 1998; Marcus, Vijayan, Rao, & Vishton, 1999; Saffran, 2003; Xu
& Garcia, 2008).
In addition, children don’t just detect patterns in the speech they hear.
Children also seem to derive broad principles from what they hear. Consider,
for example, how English-speaking children learn to form the past tense.
Initially, they proceed in a word-by-word fashion, so they memorize that the
past tense of “play” is “played,” the past tense of “climb” is “climbed,” and so
on. By age 3 or so, however, children seem to realize that they don’t have to
memorize each word’s past tense as a separate vocabulary item. Instead, they
realize they can produce the past tense by manipulating morphemes — that
is, by adding the “-ed” ending onto a word. This is, of course, an important
discovery for children, because this principle allows them to generate the past
tense for many new verbs, including verbs they’ve never encountered before.
However, children over-rely on this pattern, and their speech at this age
contains overregularization errors: They say things like “Yesterday we goed”
or “Yesterday I runned.” The same thing happens with other morphemes, so
that children of this age also overgeneralize their use of the plural ending —
they say things like, “I have two foots” or “I lost three tooths” (Marcus et al.,
1992). They also generalize the use of contractions; having heard “she isn’t”
and “you aren’t,” they say things like “I amn’t.”
It seems, then, that children (even young infants) are keenly sensitive to
patterns in the language that they’re learning, and they’re able to figure out the
(sometimes rather abstract) principles that govern these patterns. In addition,
396 •
C H A P T E R T E N Language
language learning relies on a theme that has been in view throughout this
chapter: Language has many elements (syntax, semantics, phonology, prosody, etc.), and these elements interact in ordinary language use (so that you
rely on a sentence’s syntactic form to figure out its meaning; you rely on
semantic cues in deciphering the syntax). In the same way, language learning
also relies on all these elements in an interacting fashion. For example, children rely on prosody (the rise and fall of pitch, the pattern of timing) as clues
to syntax, and adults speaking to children helpfully exaggerate these prosodic
signals, easing the children’s interpretive burden. Children also rely on their
vocabulary, listening for words they already know as clues helping them to
process more complex strings. Likewise, children rely on their knowledge of
semantic relationships as a basis for figuring out syntax — a process known
as semantic bootstrapping (Pinker, 1987). In this way, the very complexity of
language is both a burden for the child (because there’s so much to learn in
“learning a language”) and an aid (because the child can use each element as
a source of information in trying to figure out the other elements).
Animal Language
We suggested earlier that humans are biologically prepared for language
learning, and this claim has many implications. Among other points, can we
locate the genes that underlie this preparation? Many researchers claim that
we can, and they point to a gene called “FOXP2” as crucial; people who have
a mutated form of this gene are markedly impaired in their language learning
(e.g., Vargha-Khadem, Gadian, Copp, & Mishkin, 2005).
As a related point, if language learning is somehow tied to human genetics,
then we might expect not to find language capacity in other species. Of course,
many species do have sophisticated communication systems — including the
songs and clicks of dolphins and whales, the dances of honeybees, and the
various alarm calls of monkeys. These naturally occurring systems, however, are extremely limited — with small vocabularies and little (or perhaps
nothing) that corresponds to the rules of syntax that are evident in human
language. These systems will certainly not support the sort of generativity
that is a prominent feature of human language — and so these other species
don’t have anything approaching our capacity to produce or understand an
unending variety of new sentences.
Perhaps, though, these naturally occurring systems understate what
animals can do. Perhaps animals can do more if only we help them a bit.
To explore this issue, researchers have tried to train animals to use more
sophisticated forms of communication. Some researchers have tried to train
dolphins to communicate with humans; one project involved an African grey
parrot; other projects have focused on primates — asking what a chimpanzee,
gorilla, or bonobo might be capable of. The results from these studies are
impressive, but it’s notable that the greatest success involves animals that are
quite similar to humans genetically (e.g., Savage-Rumbaugh & Lewin, 1994;
Savage-Rumbaugh & Fields, 2000). For example, Kanzi, a male bonobo,
COMMUNICATION
AMONG VERVET
MONKEYS
Animals of many species
communicate with one another. For example, Vervet
monkeys give alarm calls
when they spot a nearby
predator. But they have distinct alarm calls for different
types of predator — so their
call when they see a leopard is different from their call
when they see an eagle or
a python. The fact remains,
though, that no naturally
occurring animal communication system comes close
to human language in richness or complexity.
The Biological Roots of Language
•
397
seems to understand icons on a keyboard as symbols that refer to other ideas,
and he also has some mastery of syntax — so he responds differently and
(usually) appropriately, using stuffed animals, to the instructions “Make the
doggie bite the snake” or “Make the snake bite the doggie.”
Kanzi’s abilities, though, after an enormous amount of careful training, are
way below those of the average 3- or 4-year-old human who has received no
explicit language training. (For example, as impressive as Kanzi is, he hasn’t
mastered the distinction between present, past, and future tense, although
every human child effortlessly learns this basic aspect of language.) Therefore, it seems that other species (especially those closely related to us) can
learn the rudiments of language, but nothing in their performance undercuts
the amazing differences between human language capacity and that in other
organisms.
“Wolf Children”
Before moving on, we should address one last point — one that concerns the
limits on our “biological preparation” for language. To put the matter simply,
our human biology gives us a fabulous start on language learning, but to turn
this “start” into “language capacity,” we also need a communicative partner.
In 1920, villagers in India discovered a wolf mother in her den together
with four cubs. Two were baby wolves, but the other two were human children, subsequently named Kamala and Amala. No one knows how they got
A MODERN WILD BOY
Ramu, a young boy discovered in India in 1976, appears to have been raised by wolves. He was deformed —
apparently from lying in cramped positions, as in a den. He couldn’t walk, and he drank by lapping with
his tongue. His favorite food was raw meat, which he seemed to be able to smell at a distance. After he
was found, he lived at the home for destitute children run by Mother Teresa in Lucknow, Uttar Pradesh. He
learned to bathe and dress himself but never learned to speak. He continued to prefer raw meat and would
often sneak out to prey upon fowl in the neighbor’s chicken coop. Ramu died at the age of about 10 in
February 1985.
398 •
C H A P T E R T E N Language
there or why the wolf adopted them. Roger Brown (1958) tells us what these
children were like:
Kamala was about eight years old and Amala was only one and onehalf. They were thoroughly wolfish in appearance and behavior: Hard
callus had developed on their knees and palms from going on all fours.
Their teeth were sharp edged. They moved their nostrils sniffing food.
Eating and drinking were accomplished by lowering their mouths to
the plate. They ate raw meat. . . . At night they prowled and sometimes
howled. They shunned other children but followed the dog and cat.
They slept rolled up together on the floor. . . . Amala died within a year
but Kamala lived to be eighteen. . . . In time, Kamala learned to walk
erect, to wear clothing, and even to speak a few words. (p. 100)
The outcome was similar for the 30 or so other wild children for whom
researchers have evidence. When found, they were all shockingly animal-like.
None could be rehabilitated to use language normally, although some (like
Kamala) did learn to speak a few words.
Of course, the data from these wild children are difficult to interpret,
partly because we don’t know why the children were abandoned in the first
place. (Is it possible that these children were abandoned because their human
parents detected some birth defect? If so, these children may have been
impaired in their functioning from the start.) Nonetheless, the consistency
of these findings underscores an important point: Language learning may
depend on both a human genome and a human environment.
TEST YOURSELF
13. W
hat is aphasia?
14. What are overregularization errors?
15. What do we learn
from the fact that socalled wolf-children
never gain full language proficiency?
Language and Thought
Virtually every human knows and uses a language. But it’s also important
that people speak different languages — for example, some of us speak
English, others German, and still others Abkhaz or Choctaw or Kanuri or
MYTHS ABOUT LANGUAGE AND THOUGHT
Many people believe that the native peoples of the
far north (including the Inuit) have an enormous
number of terms for various forms of snow and are
correspondingly skilled in discriminating types of
snow. It turns out, though, that the initial claim (the
number of terms for snow) is wrong; the Inuit have
roughly the same number of snow terms as do people living further south. In addition, if the Inuit people
are more skilled in discriminating snow types, is this
because of the language that they speak? Or is it
because their day-to-day lives require that they stay
alert to the differences among snow types?
(after
roberson, davies, & davidoff, 2000)
Language and Thought
•
399
Quanzhou. How do these differences matter? Is it possible that people who
speak different languages end up being different in their thought processes?
Linguistic Relativity
The notion that language shapes thought is generally attributed to the anthropologist Benjamin Whorf and is often referred to as the “Whorfian hypothesis.”
Whorf (e.g., 1956) argued that the language you speak forces you into certain modes of thought. He claimed, therefore, that people who speak different
languages inevitably think differently — a claim of linguistic relativity.
To test this claim, one line of work has examined how people perceive
colors, building on the fact that some languages have many terms for colors
(red, orange, mauve, puce, salmon, fawn, ocher, etc.) and others have few
(see Figure 10.15). Do these differences among languages affect perception?
Evidence suggests, in fact, that people who speak languages with a richer
color vocabulary may perceive colors differently — making finer and more
sharply defined distinctions (Özgen, 2004; Roberson, Davies, & Davidoff,
2000; Winawer et al., 2007).
Other studies have focused on other ways in which languages differ.
Some languages, for example, emphasize absolute directions (terms like the
English words “east” or “west” that are defined independently of which way
the speaker is facing at the moment). Other languages emphasize relative
directions (words like “right” or “left” that do depend on which way the
speaker is facing). Research suggests that these language differences can lead
to corresponding differences in how people remember — and perhaps how
they perceive — position (Majid, Bowerman, Kita, Haun, & Levinson, 2004;
Pederson et al., 1998).
Languages also differ in how they describe events. In English, we tend
to use active-voice sentences that name the agent of the action, even if the
action was accidental (“Sam made a mistake”). It sounds awkward or evasive to describe these events in other terms (“Mistakes were made”). In other
languages, including Japanese or Spanish, it’s common not to mention the
agent for an accidental event, and this in turn can shape memory: After viewing videos of accidental events, Japanese and Spanish speakers are less likely
than English speakers to remember the person who triggered the accident
(Boroditsky, 2011).
How should we think about all these results? One possibility — in line
with Whorf’s original hypothesis — is that language has a direct impact on
cognition, so that the categories recognized by your language become the
categories used in your thought. In this view, language has a unique effect
on cognition (because no other factor can shape cognition in this way), and
because language’s influence is unique, it is also irreversible: Once your language has led you to think in certain ways, you will forever think in those
ways. From this perspective, therefore, there are literally some ideas that, say,
a Japanese speaker can contemplate but that an English speaker cannot, and
vice versa — and likewise, say, for a Hopi or a French speaker.
400 •
C H A P T E R T E N Language
FIGURE 10.15
COLORS IN DIFFERENT LANGUAGES
English naming
5
Light
10
5Y 10Y
Pink
5
10
5G 10G
5
10
5R 10R
5
10
5P 10P
5
10
Yellow
5R
10
Pink
Orange
Blue
Green
Purple
Brown
Red
Red
Dark
Berinmo naming
5
10
Light
5Y 10Y
5
10
5G 10G
5
10
5R 10R
10
5P 10P
5
10
5R
10
Wap
Wap
Wor
Mehi
5
Mehi
Nol
Kel
Kel
Dark
The Berinmo people, living in Papua New Guinea, have only five words for describing colors, and so, for example,
they use a single word (“nol”) to describe colors that English speakers call “green” and colors we call “blue.”
(The letters and numbers in these panels refer to a system often used for classifying colors.) These differences,
from one language to the next, have an impact on how people perceive and remember colors. This effect is best
understood in terms of attention: Language can draw our attention to some aspect of the world and in this way
(after roberson, davies, & davidoff, 2000)
can shape our experience and, therefore, our cognition.
A different possibility is more modest — and also more plausible: The language you hear guides what you pay attention to, and what you pay attention
to shapes your thinking. In this view, language does have an influence, but
the influence is indirect: The influence works via the mechanisms of attention. Why is this distinction (direct effect vs. indirect effect) important? The
key is that other factors can also guide your attention, with the result that in
Language and Thought
•
401
many settings these factors will erase any impact that language might have.
Put differently, the idea here is that your language might bias your attention in one way, but other factors will bias your attention in the opposite
way — canceling out language’s impact. On this basis, the effects of language
on cognition might easily be reversible, and certainly not as fundamental as
Whorf proposed.
To see how this point plays out, let’s look at a concrete case. We’ve mentioned that when English speakers describe an event, our language usually
requires that we name (and so pay attention to) the actor who caused the
event; when a Spanish speaker describes the same event, her language doesn’t
have this requirement, and so it doesn’t force her to think about the actor. In
this way, the structure of each language influences what the person will pay
attention to, and the data tell us that this difference in focus has consequences
for thinking and for memory.
But we could, if we wished, simply give the Spanish speaker an instruction: “Pay attention to the actor.” Or we could make sure that the actor is
wearing a brightly colored coat, using a perceptual cue to guide attention.
These simple steps can (and often do) offset the bias created by language.
The logic is similar for the effect of language on color perception. If you’re
a speaker of Berinmo (a language spoken in New Guinea), your language
makes no distinction between “green” and “blue,” so it never leads you to
think about these as separate categories. If you’re an English speaker, your
language does make this distinction, and this can draw your attention to
what all green objects have in common and what all blue objects have in
common. If your attention is drawn to this point again and again, you’ll
gain familiarity with the distinction and eventually become better at making
the distinction. Once more, therefore, language does matter — but it matters
because of language’s impact on attention.
Again, let’s be clear on the argument here: If language directly and uniquely
shapes thought, then the effects of language on cognition will be systematic
and permanent. But the alternative is that it’s your experience that shapes
thought, and your experience depends on what you pay attention to, and
(finally) language is just one of the many factors guiding what you pay attention to. On this basis, the effects of language may sometimes be large, but can
be offset by a range of other influences. (For evidence, see Boroditsky, 2001;
and then Chen, 2007, or January & Kako, 2007. Also see Li, Abarbanell,
Gleitman, & Papafragou, 2011; Li & Gleitman, 2002.)
More than a half-century ago, Whorf argued for a strong claim — that the
language people speak plays a unique role in shaping their thought and has a
lifelong impact, determining what they can or cannot think, what ideas they can
or cannot consider. There is an element of truth here, because language can and
does shape cognition. But language’s impact is neither profound nor permanent, and there is no reason to accept Whorf’s ambitious proposal. (For more
on these issues, see Gleitman & Papafragou, 2012; Hanako & Smith, 2005;
Hermer-Vazquez, Spelke, & Katsnelson, 1999; Kay & Regier, 2007; Özgen &
Davies, 2002; Papafragou, Li, Choi, & Han, 2007; Stapel & Semin, 2007.)
402 •
C H A P T E R T E N Language
Bilingualism
There’s one more — and intriguing — way that language is said to influence cognition. It comes from cases in which someone knows more than one language.
Children raised in bilingual homes generally learn both languages quickly
and well (Kovelman, Shalinksy, Berens, & Petitto, 2008). Bilingual children
do tend to have smaller vocabularies, compared to monolingual children, but
this contrast is evident only at an early age, and bilingual children soon catch
up on this dimension (Bialystok, Craik, Green, & Gollan, 2009).
These findings surprise many people, on the expectation that bilingual
children would become confused — blurring together their languages and
getting mixed up about which words and which rules belong in each language. But this confusion seems not to occur. In fact, children who are raised
bilingually seem to develop skills that specifically help them avoid this sort
of confusion — so that they develop a skill of (say) turning off their Frenchbased habits in this setting so that they can speak uncompromised English,
and then turning off their English-based habits in that setting so that they can
speak fluent French. This skill obviously supports their language learning,
but it may also help them in other settings. (See Bialystok et al., 2009; Calvo
& Bialystok, 2013; Engel de Abreau, Cruz-Santos, Tourinho, Martion, &
Bialystok, 2012; Hernández, Costa, & Humphreys, 2012; Hilchey & Klein,
2011; Kroll, Bobb, & Hoshino, 2014; Pelham & Abrams, 2014; Zelazo,
2006.) In Chapter 5 we introduced the idea of executive control, and the suggestion here is that being raised bilingually may encourage better executive
control. As a result, bilinguals may be better at avoiding distraction, switching between competing tasks, or holding information in mind while working
on some other task.
There has, however, been considerable debate about these findings, and
not all experiments find a bilingual advantage in executive control. (See, e.g.,
Bialystok & Grundy, 2018; Costa, Hernández, Costa-Faidella, & SebastiánGalés, 2009; de Bruin, Treccani, & Della Salla, 2015; Goldsmith & Morton,
2018; Von Bastian, Souza & Gade, 2016; Zhou & Kross, 2016.) There is
some suggestion that this advantage only emerges with certain tasks or in
certain age groups (perhaps in children, but not adults). There is also some
indication that other forms of training can improve executive control — and
so bilingualism may be just one way to achieve this goal. Obviously, further
research is needed in this domain, especially since the alleged benefits of
bilingualism have important implications — for public policy, for education,
and for parenting. These implications become all the more intriguing when
we bear in mind that roughly a fifth of the population in the United States
speaks a language at home that is different from the English they use in other
settings; the proportion is even higher in some states, including California,
Texas, New Mexico, and Nevada (Shin & Kominski, 2010). These points
aside, though, research on bilingualism provides one more (and perhaps a
surprising) arena in which scholars continue to explore the ways in which
language use may shape cognition.
TEST YOURSELF
16. W
hat does it mean to
say that language’s
effects on cognition are indirect and
reversible?
Language and Thought
•
403
COGNITIVE PSYCHOLOGY AND EDUCATION
writing
Students are often required to do a lot of writing — for example, in an
essay exam or a term paper. Can cognitive psychology provide any help in this
activity — specifically, helping you to write more clearly and more persuasively?
Research tells us that people usually have an easier time understanding
active sentences than passive, and so (all things being equal) active sentences
are preferable. We also know that people approach a sentence with certain
parsing strategies, and that’s part of the reason why sentences are clearer if
the structure of the sentence is laid out early, with the details following, rather
than the other way around. Some guidelines refer to this as an advantage for
“right-branching sentences” rather than “left-branching sentences.” The idea
here is that the “branches” represent the syntactic and semantic complexity, and you want that complexity to arrive late, after the base structure is
established.
By the same logic, lists are easier to understand if they arrive late in the
sentence (“I went to the store with Juan, Fred, George, Sue, and Judy”), so
that they can be fitted into the structure, rather than arriving early (“Juan,
Fred, George, Sue, Judy, and I went to the store”) before the structure.
PEER EDITING
It is often useful to have a peer (a friend, perhaps) edit your prose (and you can then
do the same for the friend’s prose). These steps can lead to a large improvement in
how clearly you write!
404 •
C H A P T E R T E N Language
Readers are also helped by occasional words or phrases that signal the flow
of ideas in the material they’re reading. Sentences that begin “In contrast,”
or “Similarly,” or “However,” provide the reader with some advance warning
about what’s coming up and how it’s related to the ideas covered so far. This
warning, in turn, makes it easier for the reader to see how the new material fits into the framework established up to that point. The warning also
requires the writer to think about these relationships, and often that encourages the writer to do some fine-tuning of the sequence of sentences!
In addition, it’s important to remember that many people speak more
clearly than they write, and it’s interesting to ask why this is so. One reason
is prosody — the pattern of pitch changes and pauses that we use in speaking. These cannot be reproduced in writing — although prosodic cues can
sometimes be mimicked by the use of commas (to indicate pauses) or italics
(to indicate emphasis). These aspects of print can certainly be overused, but
they are in all cases important, and writers should probably pay more attention to them than they do — in order to gain in print some of the benefits that
(in spoken language) are provided by prosody.
But how should you use these cues correctly? One option is to rely on the
fact that as listeners and speakers we all know how to use prosodic cues, and
we can exploit that knowledge when we write by means of a simple trick:
reading your prose out loud. If you see a comma on the page but you’re not
inclined, as a speaker, to pause at that moment, then the comma is probably
unnecessary. Conversely, if you find yourself pausing as you read aloud but
there’s no comma, then you may need one.
Another advantage of spoken communication, as opposed to written, is
the prospect of immediate feedback. If you 
Download