The Symbolic Basis of Physical Intuition by Bruce Lawrence Sherin

advertisement
The Symbolic Basis of Physical Intuition
A Study of Two Symbol Systems in Physics Instruction
by
Bruce Lawrence Sherin
B.A. (Princeton University) 1985
M.A. (University of California, Berkeley) 1988
A dissertation submitted in partial satisfaction of the
requirements for the degree of
Doctor of Philosophy
in
Science and Mathematics Education
in the
GRADUATE DIVISION
of the
UNIVERSITY of CALIFORNIA, BERKELEY
Committee in charge:
Professor Andrea A. diSessa, Chair
Professor Rogers P. Hall
Professor Daniel S. Rokhsar
Professor Barbara Y. White
1996
1
The Symbolic Basis of Physical Intuition
A Study of Two Symbol Systems in Physics Instruction
© 1996
by
Bruce Lawrence Sherin
2
Abstract
The Symbolic Basis of Physical Intuition:
A Study of Two Symbol Systems in Physics Instruction
by
Bruce L. Sherin
Doctor of Philosophy in Science and Mathematics Education
University of California, Berkeley
Professor Andrea A. diSessa, Chair
This dissertation is a comparative study of the use of two symbol systems in physics
instruction. The first, algebraic notation, plays an important role as the language in which physicists
make precise and compact statements of physical laws and relations. Second, this research explores
the use of programming languages in physics instruction. Central to this endeavor is the notion that
programming languages can be elevated to the status of bona fide representational systems for
physics.
I undertook the cognitive project of characterizing the knowledge that would result from each
of these two instructional practices. To this end, I constructed a model of one aspect of the
knowledge associated with symbol use in physics. This model included two major types of
knowledge elements: (1) symbolic forms, which constitute a conceptual vocabulary in terms of
which physics expressions are understood, and (2) representational devices, which function as a set
of interpretive strategies. The model constitutes a partial theory of “meaningful symbol use” and
how it affects conceptual development.
The empirical basis of this work is a data corpus consisting of two parts, one which contains
videotapes of pairs of college students solving textbook physics problems using algebraic notation,
and one in which college students program computer simulations of various motions. The
videotapes in each half of the corpus were transcribed and analyzed in terms of the above model,
and the resulting analyses compared. This involved listing the specific symbolic forms and
representational devices employed by the students, as well as a measurement of the frequency of
use of the various knowledge elements.
A conclusion of this work is that algebra-physics can be characterized as a physics of balance and
equilibrium, and programming-physics a physics of processes and causation. More generally, this
3
work provides a theoretical and empirical basis for understanding how the use of particular symbol
systems affects students’ conceptualization.
4
Acknowledgements
This work, to a great extent, grew out of my interactions with my advisor, Andrea A. diSessa. He
shaped my views on thinking and learning generally, and also contributed directly to many of the
ideas in this thesis.
My views have also been strongly influenced by Rogers Hall and Barbara White. Rogers gave me
many new ways to think about representational artifacts—really, more than I could handle—and
Barbara kept me anchored to the world of physics instruction. Thanks also to Daniel Rokhsar, for
his physics-eye view.
Many other members of the EMST/SESAME community helped me along the way. Thanks
especially to the members of the Boxer Research Group, for listening to me talk about this work
for several years, and giving helpful feedback. Thanks also to the members of the Representational
Practices Group and the ThinkerTools Group for the added perspective provided by their own
research.
I could never have finished—or taken so long—without my friends to distract me. Thanks to the
Geebers for Geebing, the thing that they do best.
My parents have never been anything but supportive in my academic pursuits, a fact that I really
do appreciate. Thanks also to my grandparents, my sister Heidi, my parents-in-law, all my
brothers-in-law, sisters-in-law, nieces, nephews, and everybody else for generally being a mellow
bunch.
My wife Miriam, in a thousand ways, made this work possible. She did the last-minute tasks that I
couldn’t do, and kept me sane and happy. If she’s proud of me, then I’m satisfied.
5
Chapter 1. Introduction and Major Themes
4
26
319
64
9
6
33
fattened kids
fattened goat-heifers
pastured sheep
pastured he-goats
pastured ewes
pastured goats
weaned lambs…
Circa 2500 BCE, an inhabitant of Mesopotamia—probably a man—stands working on a clay
tablet, writing in cuneiform. He is recording information concerning livestock to be presented in
trade. The marks he makes are not a careful transcription of speech, but they are also not a
haphazard collection of information. The information is represented in a specific form, as a list.
This ancient man is making and using an external representation.
To a present day individual, the making of such a list is a mundane event. We make lists
before going shopping and to keep track of the tasks we need “to do.” But the timeless and
ubiquitous nature of written lists does not suggest that they are simple and unimportant
constructions; on the contrary, it implies that they have a power that transcends time and task.
Why are lists so useful that they are not only written on sheets of paper, they have even been
etched on clay tablets?
Some of the answers to this question are relatively straightforward. The list is useful to our
ancient Mesopotamian because he can use it later as an aid to memory. In general, the use of
written lists allows individuals to remember long lists of information with relative ease, as
compared with committing the same information to memory. In addition, lists, such as shopping
lists, can serve as resources during future action. While at the supermarket, we can procure the
items on our list one by one, checking them off as we go.
Furthermore, there is the striking fact that the list above is a real one: This particular list of
information was actually written by some inhabitant of Mesopotamia during the third
millennium, BCE (It is taken from Goody, 1977). Thus, the information has accomplished the
trick of traveling through time several millennia so that we can read it, a striking feat!
Some aspects of the power of lists and other external representations are somewhat less easy to
recognize and explain. In The Domestication of the Savage Mind, Jack Goody discusses the
important role played by lists during the early development of writing (Goody, 1977). According
6
to Goody, writing—and lists, in particular—did not simply constitute a means of performing the
same tasks, such as remembering existing information, more easily and efficiently. The changes
introduced by written lists were not only changes in degree or extent. Rather, he argues, the
development of lists changed people’s practices as well as the very nature of knowledge and
understanding.
My concern here is to show that these written forms were not simply by-products of the interaction
between writing and say, the economy, filling some hitherto hidden ‘need,’ but that they represented a
significant change not only in the nature of transactions, but also in the ‘modes of thought’ that
accompanied them, at least if we interpret ‘modes of thought’ in terms of the formal, cognitive and
linguistic operations which this new technology of the intellect opened up. (p. 81)
The point is this: As our list-writer works, he must bend his thought and action to the
framework provided by the list. For this reason, the thoughts of our list-writer, during the
moments of composing the list, are in part shaped by the structure of the list. Thus, the product
of this activity is also in part determined by the list structure. Goody even argues that the use of
lists led to the development of scientific taxonomies and eventually to the development of science
as we know it.
Step forward to the 20th century. Lists and other external representations are still quite
popular. External representations are an important part of everyday activity, as well as of work in
almost all disciplines. Here we turn to the scientific disciplines; in particular, we turn to the
domain of physics. Within the discipline of physics, equations and related symbolic expressions
are an extremely important external representational form. Physicists write equations and
manipulate symbols in order to perform extended and complex computations. Furthermore,
physicists use equations as a means of making precise and compact expressions of physical laws,
which then appear in academic papers and in textbooks.
As with Goody’s lists, the implications of equation-use in physics extend to the subtle and
profound. Equation-use is so pervasive in physics that, for the novice physics student, it may be
hard to escape the intuition that equations ARE physics knowledge. And it may seem like “doing
physics” means nothing more than using and manipulating symbols. This may be more than a
foolhardy misconception, however. At least one physicist believes that there is some truth behind
this appearance. In The Character of Physical Law, physicist Richard Feynman, expresses a similar
intuition (Feynman, 1965):
The burden of the lecture is just to emphasize the fact that it is impossible to explain honestly the
beauties of the laws of nature in a way that people can feel, without their having some deep
understanding in mathematics. I am sorry, but this seems to be the case. (p. 39)
7
Why not tell me in words instead of in symbols? Mathematics is just a language, and I want to be able
to translate the language. … But I do not think it is possible, because mathematics is not just another
language. Mathematics is a language plus reasoning; it is like a language plus logic. (p. 40)
The apparent enormous complexities of nature, with all its funny laws and rules, each of which has
been carefully explained to you, are really very closely interwoven. However, if you do not appreciate
the mathematics, you cannot see, among the great variety of facts, that logic permits you to go from one
to the other. (p. 41)
In these quotes, Feynman does not quite say that physics knowledge is nothing but equations,
but he is making a related claim. Here, he expresses the intuition that, in some manner, it is not
possible to fully understand physics without being an initiate into the physicist’s use of
mathematics and the associated symbols. Feynman tells his students that he cannot explain “in
words instead of symbols.” To understand physics, you must have the experience of living in the
tight net of physical relations expressed as equations, going from one to the other. You have to
write equations and solve them, derive one equation from another, and from this experience
develop a sense for the properties of the equations and the relations among them.
From Goody’s lists to Feynman’s equations, the power of external representations is manifest.
They allow us to easily remember long lists of facts for a duration that is essentially unlimited.
And they allow us to perform extended, context-free manipulations to compute results.
Furthermore, these authors point to the possibility of more subtle implications of external
representations for the nature of human knowledge. The first, an historical view, posits a role for
external representations in an historically developing body of knowledge. The second view boils
down to a physicist’s intuition that an appreciation of the beauties of physics is somehow tied up
with their expression in the language of mathematics.
Similar notions have appeared in many forms, and in a variety of literature. Jerome Bruner
spoke of external representations as a type of “cultural amplifier” (Bruner, 1966). Whorf
hypothesized that the vocabulary and syntax of a language shape the speaker’s view of reality
(Whorf, 1956). Lev Vygotsky was the father of a branch of psychology for which the
internalization of external signs plays a fundamental role in the development and character of
thought (Vygotsky, 1934/1986). And there is an extensive body of literature devoted to
determining the effect of literacy on thought (e.g., Goody & Watt, 1968; Scribner & Cole,
1981).
For each of the above researchers, knowledge and external representations are inextricably
related. But the diversity of views here will require some sorting, and the intuitions call for
expression in theoretically precise terms. Here, I will not adopt the claims or perspective of any of
8
these researchers, or add directly to any of their research programs. Unlike these anthropologists
and historians, I will not look across cultures or time for differences in thought or the nature of
knowledge associated with the use of various symbol systems. Instead, I will use the methods of
cognitive science to build a model of symbol use within a particular domain. Then, given this
model, I can begin to tentatively address some of the long-standing issues concerning external
representations and human knowledge.
A Tale of Two (Representational) Systems
The inquiry to be described here will be built around a comparative study of two symbol
systems within the domain of physics. The first of these is the traditional system of algebraic
notation. By “algebraic notation” I mean to include the full range of equations that physicists
write. Within the practice of physics, algebraic notation plays an important role as the language in
which physicists make precise and compact statements of physical laws and relations, and in which
they solve problems and derive new results. Here we will be particularly concerned with the use of
algebraic notation as it occurs in introductory physics instruction.
The use of the second representational system is novel: This research explores the use of
programming languages in physics instruction. Central to this endeavor is the notion that
programming languages can be elevated to the status of bona fide representational systems for
physics. This means treating programming as potentially having a similar status and performing a
similar function to algebraic notation in physics learning. It means that instead of representing
laws as equations, students will represent them as programs.
This proposal—to replace algebraic notation with a programming language—would certainly
constitute a significant change in the nature of physics and physics instruction. If programming
becomes a major part of the physics curriculum, then students will be engaged in very different
activities: They will be programming instead of solving traditional physics textbook problems.
And the knowledge that students take away from this new curriculum will be substantially
different, at least in some superficial respects. Students will learn programming and programming
algorithms, rather than some derivational and problem solving strategies that are particular to
textbook physics problems.
In addition, we must also keep in mind the more profound possibilities raised by the likes of
Goody and Feynman. Goody might tell us that we must expect to alter the nature of physics
knowledge in a more fundamental manner. And Feynman might worry that an approach to
9
physics that passes through programming may deprive students of an appreciation of the “beauty
of nature.”
These observations are more than interesting asides; they have serious implications for how the
comparative study at the heart of this inquiry must be understood. If external representations are
inextricably tied to the nature of physics knowledge, then it will not be reasonable to think of
programming as a new representation for doing the same old jobs. In Goody’s terms,
representations do not just fill some need, they also shape and determine the nature of the
knowledge. For this reason, it will not be precisely correct to ask whether programming or algebra
does certain jobs better. I cannot ask which of the systems better “represents” physics knowledge. I
cannot even really ask which of these representational systems will lead to better student learning
of physics. In my view, there is no physics that exists separate from these representational systems
and the associated tasks, that can be represented more accurately or taught better.
As an alternative, I thus propose to ask and answer the following questions: How does
“programming-physics” differ from “algebra-physics?” And: Is “programming-physics” a
respectable form of physics? These questions will provide the focus and anchor for this inquiry.
Two practices for introductory physics instruction
There is still another complication to be considered. I cannot assume that these symbol
systems, somehow defined prior to their application to physics, uniquely determine the associated
practices. In other words, for example, given any programming language, there are many possible
practices of programming-physics that one could build around that programming language. For
this study, I must therefore pick out a particular algebra-physics and a particular programmingphysics for study.
Specifying a practice of algebra-physics for study does not pose a significant difficulty. Since I
am interested in programming as an alternative for physics instruction, I will examine algebraphysics as it is practiced by students in traditional introductory physics courses. Because of the
great uniformity that exists across such traditional courses (which still make up most physics
instruction), this constitutes a relatively well-defined and uniform practice. Perhaps the most
prominent aspect of this practice is an emphasis on solving a certain class of problems. A problem
and solution, taken from the textbook by Tipler (1976), is shown in Figure Chapter 1. -1. This
problem is wholly typical of those that students solve during the first weeks of an introductory
10
course. Note that it employs a few equations (notably v = vo + at and ∆x = vot + 12 at 2 ) that
students are expected to have learned during this period.
A ball is thrown upward with an initial velocity of 30 m/s. If its acceleration is 10 m/s2
downward, how long does it take to reach its highest point, and what is the distance to the
highest point?
v = vo + at
0 = 30 m s + −10 m s 2 t
(
t=
)
30 m s
= 3.0s
10 m s 2
(
)
∆x = vot + 12 at 2 = (30 m s)(3.0s ) + 12 −10 m s 2 (3.0s )2
= +90m − 45m = 45m
Figure Chapter 1. -1. A typical problem and solution (from Tipler, 1976).
For the case of programming, there is less to build on, and there is certainly no uniformly
accepted method of employing programming languages in physics instruction. In this work, I will
draw on a practice of programming-physics that was developed by the Boxer Research Group at
U.C. Berkeley, and was implemented in 6th grade and high school classrooms (diSessa, 1989;
Sherin, diSessa, and Hammer, 1993). For the purposes of this study, this practice was transported
into a laboratory setting, and it was employed with university students who had already
completed two semesters of introductory physics. The resulting complications will be discussed
later.
Unlike the problem-based approach typical of traditional introductory physics courses, the
courses developed by the Boxer Research Group were all simulation-based; rather than solving
standard physics textbook problems, the students spent their time programming simulations of
motion. Figure Chapter 1. -2 shows an example program, which simulates the motion of a
dropped ball. This simulation is written in the Boxer programming environment (diSessa,
Abelson, and Ploger, 1991). At first glance, the external representations that appear in this figure
probably seem very different than the equations in Figure Chapter 1. -1—and they are!
Nonetheless, this research will uncover a surprising degree and kind of similarity. The Boxer
programming language and the practice of programming-physics employed in this study will be
described in some detail in Chapter 8.
11
Instructions
Make a realistic simulation of the motion of a ball
that is dropped from someone's hand.
Data
drop
setup
change
change
change
repeat
pos 0
vel 0
acc 1
23 tick
pos
vel
acc
253
23
1
Data
Data
Data
tick
Doit
change pos
pos + vel
Doit
change vel
vel + acc
Doit
fd vel
dot
Doit
Data
Data
Figure Chapter 1. -2. Simulation of a dropped ball.
Finally, I want to note that, just looking at Figure Chapter 1. -1 and Figure Chapter 1. -2, it is
evident that each of these practices involves more than a single external representation—they
involve more than just algebraic notation and programming statements. For example, both of the
above figures have textual elements, and the programming figure shows a representation of the
ball’s motion involving dots. Furthermore, even more external representations are involved in these
practices than appear in these figures. For instance, algebra-physics uses graphs and various types
of diagrams. Nonetheless, I have chosen to name these two practices after algebra and
programming because my attention will be focused tightly around these particular symbol
systems. In addition, I believe that these symbol systems play a critical role in constraining the
nature of their respective practices.
The advantageous properties of this study
Why is this particular comparative study the right place to explore the relation between
external representations and understanding? First, it is apparent that symbol use plays a substantial
and important role in the discipline of physics, both in instruction and in the work of practicing
experts. The process of learning physics traditionally involves a great deal of writing equations and
12
manipulating them to solve problems, and equations are the target form for physicists’ laws.
Because symbol use plays such an important role in physics, this study is pretty much guaranteed
to encounter some of the more interesting interactions between knowledge and external
representations. Second, although algebraic notation plays a central role in physics, it is not as
integrated in everyday activity as natural language; thus, its role should be somewhat easier to
fathom. The point is that describing the relationship between literacy and thought (Goody &
Watt, 1968), or thought and language (Vygotsky, 1934/1986) is probably a much harder
problem than describing the relationship between equations and physics understanding. We can
thus think of this project as a first and easier step toward these other, more ambitious
undertakings.
Third, the use of a comparative study has some benefits. I will be able to do more than
speculate about how physics built around an alternative symbol system might differ; I can actually
uncover some specific differences. It is also important that the two practices we are comparing are
not altogether different. If the two practices were too divergent, what type of comparisons might
we make? How, for example, would one compare physics and chemistry?
Finally, this study has interest beyond the more theoretical issues I have described. One of the
aims of this project is to explore the feasibility of a novel instructional approach for physics.
Preliminary research has found that the use of a programming language in physics instruction is a
promising alternative and that programming languages may even have certain advantages (as well
as certain disadvantages) over algebraic notation. The study described here will contribute to the
investigation of this novel instructional approach.
Overview of the Approach
Taken down to a single driving notion, this study can be seen as an exploration of the
possibility of replacing algebraic notation with a programming language in physics instruction.
But, when the question is viewed with a little care, such a study blows up to a much larger scale,
and it is evident that a number of subsidiary questions must be answered. First, we cannot possibly
understand the implications of replacing algebra with programming if we do not have some
notion of the role played by algebraic notation in traditional physics learning. So, already this
study has doubled in size; before even starting to look at the use of programming in physics, we
have to go back and understand certain aspects of traditional physics learning using algebraic
notation.
13
And we have seen that it is probably not going to be good enough to think simply in terms of
replacing algebraic notation with a programming language. Such a switch may have far-reaching
effects on the nature of physics knowledge and understanding. So, we really cannot do anything
like a simple comparison, there is some tricky business about the relationship between knowledge
and external representations. This is a problem to overcome, but it is also a positive feature of this
inquiry: One of the outcomes of this study will be some insight into this complex relationship.
In order to compare algebra-physics and programming-physics I proceeded as follows. I began
by collecting a data corpus consisting of two parts, one which contains observations of students
engaged in the practice of algebra-physics and another in which students were engaged in
programming-physics. The question then becomes: How can we go about comparing these two
sets of observations?
The key to producing a useful comparison of algebra-physics and programming-physics based
on these observations was the construction of a model of some aspects of symbol use in physics.
This model, which I will preview below in a moment, is a model in the spirit of cognitive science; I
attempted to account for a certain class of human behavior (symbol use in physics) by
hypothesizing that people possess and employ certain elements of knowledge.
This model is the central result of this research and the key to getting at the goals of this
inquiry. Viewed at a certain level, this model will be general enough to describe both varieties of
symbol use under consideration in this study. Thus, it can provide the categories for
understanding and comparing the observations of algebra-physics and programming-physics in
the data corpus. In addition, once we have the model, we can look at it with more general issues in
mind.
In the remainder of this section I will briefly describe the data corpus and the model of
symbol use on which my analysis is based.
The data corpus, in brief
All of the subjects in this study were UC Berkeley students enrolled in “Physics 7C.” Physics
7C is a third semester introductory course intended primarily for engineering majors of all
persuasions. The fact that these students were in Physics 7C implies that they had already
completed two semesters of instruction, Physics 7A and 7B. Many also had some instruction
during high school.
14
The subjects were divided into two distinct pools, an “algebra” pool and a “programming”
pool, with each pool consisting of 5 pairs of students. All experimental sessions were conducted in
a laboratory setting and these sessions were videotaped. Students in the algebra pool worked with
their partner at a blackboard to solve a pre-specified set of problems. Most of these problems were
fairly traditional textbook problems, but a few more unusual tasks were also included. The pairs
typically solved the problems in 5 1/2 hours spread over 5 sessions. The result was a total of 27
hours of videotape of these students using algebraic notation to solve physics problems. A subset
of this data, 10 hours, was selected for more focused analysis.
Programming pool subjects worked in pairs at computers. The students were first given about
four hours of training, usually requiring 2 two-hour sessions. Following the training, the students
were asked to program a set of simulations of physical motions. They typically required 4
additional two-hour sessions to complete these tasks. In sum, the result was a total of 53 hours of
videotape of students programming simulations. A subset of this data, 16 hours, was selected for
more focused analysis
All of the videotaped sessions selected for focused analysis were carefully transcribed and the
resulting transcripts were analyzed. The examples given in upcoming chapters are all taken from
these transcripts, and the model described is drawn from this analysis. However, I will not
describe the data corpus or my analysis techniques in more detail until Chapter 6. That chapter
will also include an account of the rationale for the above design.
Overview of the model: The Theory of Forms and Devices, in brief
As I mentioned above, the model will be a theory in the spirit of cognitive science, which
means it will posit knowledge elements to explain certain behaviors. The scope of the model will
be restricted in a number of senses. Most importantly, I will attempt to describe only a small
subset of the knowledge necessary to provide a complete account of symbol use in physics. This
subset is particularly important, I will argue, in the construction of novel symbolic expressions and
the interpretation of expressions, both in algebra-physics and programming-physics. Furthermore,
I will argue that this small subset of knowledge is precisely where our attention must be focused in
order to get at the relation between physics knowledge and external representations.
The model of symbol use will be built around two major theoretical constructs, what I call
“symbolic forms” and “representational devices.” I will now briefly discuss each of these
constructs. (A more complete discussion can be found in Chapters 3 and 4.)
15
On Symbolic Forms
An important contention of this work is that initiates in algebra-physics are capable of
recognizing a particular type of structure in physics expressions. More specifically, they are able to
see expressions in terms of a set of templates that I refer to as “symbol patterns,” each of which is
associated with a simple conceptual schema. The knowledge element involving this combination
of symbol pattern and schema I refer to as a “symbolic form” or just as a “form,” for short.
In a sense, symbolic forms constitute a conceptual vocabulary in terms of which physics
equations are written and understood. For example, one of the most common forms is what I call
“BALANCING .” In the conceptual schema associated with BALANCING , a situation is schematized as
involving two influences, such as two forces, in balance. Furthermore, the symbol pattern
associated with BALANCING involves two expressions separated by an equal sign:
=
BALANCING
For illustration, imagine an example in which a student is solving a problem involving a block
that has two opposing forces acting on it. Because the block is not moving, the student writes the
expression F1 = F2. Now, the question is, how precisely did the student know to write this
equation? The answer I will give to this question is that the BALANCING form has been engaged and,
when it is engaged, it pins down what the student needs to write at a certain level. In this case, it
specifies that the student needs to write two expressions separated by an equal sign, with each side
of the equation associated with one of the two influences in balance. Furthermore, this process can
also work in reverse; a student can look at an equation, like F1 = F2, and see it in terms of the
BALANCING
form.
Symbolic forms, in part, develop out of symbolic experience and then become a way of
experiencing the world of symbols. Because of the developed ability to be sensitive to the patterns
associated with forms, the symbolic world of experience is more structured and meaningful for
the physics initiate than one might first expect.
On Representational Devices
In general, a number of different symbolic forms may be recognized in any particular physics
expressions. The question then arises: What determines, at any particular time, the forms that a
student recognizes in an expression? The answer, I will attempt to show, is that initiates in physics
possess a repertoire of interpretive strategies that I call “representational devices.” Roughly,
16
representational devices set a stance within which symbolic forms are engaged. The term “device”
here is intended to call to mind “literary devices,” for reasons that I hope to make apparent in
Chapter 4.
I will argue that there are three major classes of representational devices. In one of these classes,
which I call Narrative devices, an equation is embedded in an imaginary process in which some
change occurs. For example, suppose that a student has derived the expression a=F/m, which gives
the acceleration of an object in terms of the mass and the force applied. The student can then
imagine a process, for example, in which the mass increases, and infer from the equation that the
acceleration must decrease. This would be an interpretation involving a Narrative device.
The other two classes of representational devices are what I call Static and Special Case
devices. In Static devices, an equation is seen as describing a moment in a motion. For example, an
equation may be seen as true only at the apex of a projectile’s trajectory. In Special Case devices, a
stance is set by considering some restricted set of cases to which the equation applies. For
example, a student may consider the behavior of an equation in an extreme or limiting case.
In sum, the Theory of Forms and Devices describes a collection of knowledge that allows
initiates in physics to see a certain type of meaningful structure in physics expressions. Two
specific types of knowledge are involved in this collection: symbolic forms, which correspond to a
set of meaningful structures that students can recognize in expressions, and representational
devices, which constitute a repertoire of interpretive stances. Primarily, it is by looking at the
forms and devices associated with algebra-physics and programming-physics that I will compare
these two practices.
In the remainder of this chapter, I will now work toward situating this work in relation to
broader issues and in relation to existing research. I will begin, in the next section, by attempting
to provide a definition of “external representation.” Then I will present a more in depth discussion
of a central issue that I have so far only touched upon: the relation between knowledge and
external representations. Finally, I will discuss how this work relates to existing research
concerning physics learning and instruction.
A Definition of ÒExternal RepresentationÓ
Much of what I say in this document will be specific to the external representational practices
described in the previous section. But I will also want to argue for some generality of my
17
observations, especially when these observations are framed in broader terms. However, in order to
make these claims of generality there is some additional work that must be done. If I want to say
that my observations apply more broadly, then I need to say something about the scope of
application. In particular, it would be helpful to have a definition of “external representation.”
So, what is an “external representation.” Let’s take a moment, first, to think about the type of
definition that is required here. My project is to build a model of some knowledge associated with
symbol use in physics. In essence, I am going to be providing a kind of description of a class of
human behaviors. Thus, the definition we need is a definition that characterizes this class of
behavior, that specifies it in contrast to other behaviors. Finding such a definition may very well be
difficult, since human activity is complicated and messy. In fact, there is no reason to believe that
there is a simple definition that lines up with our intuitions concerning what constitutes a
representation and still specifies a relatively circumscribable class of human activity.
An alternative approach, taken by a number of other researchers, is to define external
representation without explicit reference to behavior or cognition. They begin, instead, by
defining a “representational system.” For example, in a typical version of such a definition, a
representational system is defined in terms of correspondences between a “representing” and a
“represented world.” This type of definition of representation hooks up with our intuitive notion
of representation quite clearly—a representation is something that “stands for” something else.
Such correspondence definitions have their uses, and I will follow this line for a while in Chapter 7.
However, I believe that this is not the right type of definition for the present project. Since I am
building a model of symbol use, I need a definition that slices off a class of human activity.
Let’s take a stab at this type of definition. I will define an “external representational practice”
as a class of human activity that involves:
1.
A set of externally realizable marks. These marks will be organized, in some manner, either
in space or through time. They may be arrangements of objects, marks on paper, sounds, or
even smells.
2.
Conventions concerning the making of marks. There will be conventions—socially dictated
regularities in human behavior that are shared by many individuals—pertaining to how the
marks are organized in space and time, and how the making of marks is integrated with other
aspects of activity.
18
3.
Conventions concerning action associated with existing arrangements of marks. People who
have learned to participate in the practice will be sensitive to distinctions in the arrangement of
marks. This means that their action in the presence of the external representation—what they
say and do because of and with a particular collection of marks—will depend sensitively on
conventionally defined distinctions in the arrangement of the marks.
To see that this definition can actually do some work, let’s look at what it rules out:
•
A random pile of stones out in the middle of nowhere. This is not an instance of human
activity so it’s not even covered by the definition.
•
Some children make a haphazardly arranged pile of stones. This violates requirements (2) and
(3) of the definition. To the extent that there are conventions, they are not widely shared by
many individuals.
•
A table is set prior to a meal. In this case, requirement (2) is met since there are conventions
concerning the arrangement of the objects. However, requirement (3) is not met; fine
distinctions in the arrangement of objects are not important for future action.
•
The entrails of a goat are used to predict the future. Here, there is an arrangement of physical
marks that is interpreted, so requirement (3) is met. But the act of arranging the marks is not,
presumably, a normal component of the practice, so requirement (2) is not met and this is not
a representational practice.
Note that this is not really a technical definition and I am relying on the reader’s intuitions, to
a great extent, to apply it. Furthermore, the agreement of specific instances with the requirements
of the definition may be a matter of degree rather than clearly establishable fact. For example, a
sophisticated diner may know that the outermost utensils in a place setting are always supposed to
be used first. Thus, the arrangement of objects is significant for human action and dining becomes
a representational practice. My first response to this is that maybe this is not such a ridiculous
conclusion. To the extent that the arrangement of silverware communicates information from the
table setter to the diner, perhaps it is reasonable to think of this as a very simple representational
practice.
My second response is to again appeal to the type of definition that I am trying to create.
Since I am attempting to slice off a class of human behavior, there are certain to be some fuzzy
19
edges. The best I can really shoot for is a rough characterization of this class, something that gives
the reader a sense for the scope of the conclusions to be drawn here. The hope is that the
intuitively central cases (equations, graphs, writing) are clearly included by the definition, and that
intuitively excludable cases (the haphazard arrangement of stones) are excluded.
Thus, this definition does not specify properties of external representations that make them
absolutely unique among artifacts; most notably, I have not tried to argue that external
representations have the special property that they refer to or stand for ideas or other elements of
the environment. Instead, I have characterized a representational practice as involving the assigning
of significance to fine distinctions in the arrangements of elements of the environment.
Furthermore, the conventions for assigning this significance are shared among individuals. On
these dimensions, the difference between a representational practice and other activity is a matter
of degree; the most typical instances of representational practices involve a sensitivity to many and
numerous fine distinctions, with the associated conventions shared by many individuals.
Relations Between Knowledge and External Representations
I am working from the premise that replacing algebraic notation with programming must
change the nature of physics knowledge and understanding. But, to this point, I have only
proffered some vague notions and rough suggestions as to the manner in which knowledge and
external representations are related. I have Goody’s notion that lists shape knowledge and
Feynman’s intuition that an appreciation of physics is tied up with its mathematical expression.
But, more precisely, what can it mean to say that knowledge is inextricably representational? In
what ways might physics knowledge and external representations be related?
Here I will take some preliminary steps toward answering these questions. To do this, I will lay
out a very basic framework on which we can hang some of our intuitions concerning relations
between knowledge and external representations. However, I will not push too hard on these
intuitions and I certainly do not intend this discussion to be exhaustive in its stating of relations
between knowledge and external representation. My purpose is only to make a basic accounting of
some of these relations within a simple framework.
Direct versus indirect effects of external representations
In order to introduce the framework on which the rest of the discussion in this section
depends, I will begin with an important distinction between “direct” and “indirect” influences of
20
external representations on the knowledge of individuals. This distinction is illustrated nicely by
some competing viewpoints concerning the effects of literacy on thought. To begin, I turn to
Goody and Watt’s seminal article The Consequences of Literacy (Goody & Watt, 1968). In this
article, Goody and Watt’s agenda was to react against what they perceived to be the over-extreme
cultural relativism of the time. To respond to this relativism, they proposed that a number of
significant differences existed between “oral” and “literate” societies. Most importantly for our
concerns here, they presented a number of strong hypotheses concerning the effects of literacy on
human thought.
The arguments of Goody and Watt rest largely on the analysis of historical cases. For
example, they claim that the development of the specific type of alphabet used by the Ancient
Greeks was responsible for—or, at least, allowed—many of the important intellectual
achievements of the time. Included is the claim that the development of this type of literacy
allowed the development of a new sort of logic and even “scientific” thought.
Two researchers, Sylvia Scribner and Michael Cole, proposed to directly test these claims. In
The Psychology of Literacy, Scribner and Cole describe an anthropological research project
involving the Vai people of Northwestern Liberia (Scribner & Cole, 1981). For the purposes of
studying the consequences of literacy for individuals, the Vai provided an opportunity for a
potentially illuminating case study. The interesting fact about the Vai culture is that there are
three widely practiced forms of literacy, as well as widespread non-literacy. First, the Vai employ
an independently invented phonetic writing system—Vai script—that is not used elsewhere. No
formal institutional settings exist for the propagation of this script; rather, it is transmitted
informally among the Vai. In addition, there are Qur’anic schools where the Arabic script is
learned, and English writing is taught in the context of formal western-style schooling.
Scribner and Cole’s framing idea was that the Vai culture provided a natural laboratory in
which the effects of literacy could be studied separately from the effects of schooling. Their
method involved the administering of various psychological tests to individuals with differing
literacies and with differing schooling backgrounds. Through such a study, Scribner and Cole
reasoned, they could isolate the consequences of literacy for the mental abilities of individuals.
The results of this research, at least with regard to encompassing cognitive consequences, were
generally negative.
There is nothing in our findings that would lead us to speak of cognitive consequences of literacy with
the notion in mind that such consequences affect intellectual performance in all tasks to which the
human mind is put. Nothing in our data would support the statement quoted earlier that reading and
21
writing entail fundamental “cognitive restructurings” that control intellectual performance in all
domains (Scribner & Cole, 1988, p. 70).
Literacy, it seemed, did not lead Vai individuals to perform better on measures of general
cognitive ability, such as tests of logic or categorization tasks.
However, Scribner and Cole did find differences in more specific “skills” associated with each
of the individual forms of literacy. For example, it turns out that letter writing is one of the most
common uses of the Vai script. For this reason, Scribner and Cole hypothesized that Vai literacy
would require and foster the development of some specialized skills not required by oral
communication.
In writing, meaning is carried entirely by the text. An effective written communication requires
sensitivity to the informational needs of the reader and skill in use of elaborative linguistic techniques.
(Scribner & Cole, 1988, p. 66)
In fact, Scribner and Cole did find a correlation between Vai script literacy and communication
skills. Similarly, because the use of the Arabic script was associated with memorization tasks in
Qur’anic learning, they predicted that this Arabic literacy would be associated with increased
memory skills. This hypothesis was born out, though only in the types of memory tasks that were
similar to those found in the Qur’anic schools.
So Scribner and Cole’s results seem to indicate individual differences due to literacy at the
level of specific skills, but not at the level of more general cognitive capacities. Goody’s response to
these results was that, given the nature of the study performed, these results are not at all
surprising:
That is to say, we did not expect the ‘mastery of writing’ (of whatever form) to produce in itself an
immediate change in the intellectual operations of individuals … But if we are referring to an
operation like syllogistic reasoning, the expectation that ‘mastery of writing’ in itself would lead
directly to its adoption is patently absurd. The syllogism, as we know it, was a particular invention of a
particular place and time; there were forerunners of a sort in Mesopotamia just as there were
forerunners of Pythagoras’ Theorem; nor on a more general level is the process of deductive inference
unknown in oral societies. But we are talking about a particular kind of puzzle, ‘logic’, theorem, that
involves a graphic lay-out. In this sense the syllogism is consequent upon or implied in writing.
However its use as distinct from its invention does not demand a mastery of writing. Once invented, it
can be fed back into the activities of individual illiterates or even non-literates just as the same
individuals can be taught to operate the arithmetic table or, as Scribner and Cole point out, to decode
Rebus writing. (Goody, 1987, p. 221)
Goody (1987) believes that psychological tests that compare individuals in a society should not be
expected to uncover the more general effects predicted by Goody and Watt (1968). According to
Goody, we should expect the stronger consequences of literacy to be manifested only at the level
22
of societies, not at the level of individuals. Thus, we should not expect to see any strong
differences between Vai individuals due to differences in literacy.
To clarify this point, Goody makes a distinction between what he calls “mediated” and
“unmediated” effects of literacy (Goody, 1987). The idea here is that the consequences of literacy
for the thought processes of individuals need not arise through the direct use of writing by
individuals. Instead, the consequences of literacy may play themselves out at a society-wide level
and then be “fed back” into the thought of individuals. An individual need not even be literate to
be affected. Given this viewpoint, Goody and Watt’s original hypotheses need not be manifested
as differences between individuals in the Vai society.
There are many other extant criticisms of the hypotheses of Goody and Watt, such as the
claim that their arguments tend toward an inappropriate “technological determinism” (Street,
1984). But my purpose here is not to engage in a thorough analysis of particular hypotheses for the
consequences of literacy. Instead, my goal has been to use this debate as a context for illustrating
that hypotheses concerning the effects of external representations on individual knowledge may be
of two sorts: We can posit direct effects, which arise due to an individual’s own participation in a
representational practice, and indirect effects, which do not necessarily arise from actual use of the
representation. The main purpose of the simple framework shown in Figure Chapter 1. -4 is to
highlight this distinction.
Representational
Practice
Direct
Effects
Literate
Individual
External
Representation
"Products"
Indirect
Effects
Cultural
Mediation
Literate or
Non-Literate
Individual
Figure Chapter 1. -4. An individual’s knowledge can change through interaction with an external
representation. In addition, the products of this interactions may be “fed back” into the knowledge of literate
as well as non-literate individuals.
23
I will take a moment to walk the reader through Figure Chapter 1. -4. To begin, we have a
“literate” individual, defined as such by their ability to participate in a representational practice.
Following the structure provided by this practice, this individual interacts with an external
representation. Participation in this practice may have a direct (unmediated) affect on the
knowledge or thought processes of the literate individual. In addition, there may be some
outcome of this interaction, a “product” whose nature is shaped by the external representation and
the associated practice. This product may then be processed by society and ultimately fed back
into the knowledge of individual literates or non-literates. Also note that the products of
representational activity may feed back into a reshaping of the representational practice itself.
This simple framework will guide us in the discussion that follows. In the remainder of this
section I will go into slightly more detail concerning some particular notions regarding individual
knowledge and external representations. Where appropriate, I will relate these notions to the
framework presented in Figure Chapter 1. -4.
A Òmore powerful systemÓ
One stance that we can take is that the literate individual and external representation,
interacting as in Figure Chapter 1. -4, constitute a system whose properties we can study. The
point here, which has been made by many researchers, is that this system may in some ways be
“more powerful” than the individual acting alone. Before saying how this stance reflects on the
relationship between external representations and knowledge, I want to take a moment to explore
this view.
The “more powerful system” notion has many roots, one of which is in the work of Jerome
Bruner (Bruner, 1966). Bruner argues that cultures provide individuals with a number of “cultural
amplifiers,” artifacts and techniques that amplify the capacities of individuals, both physical and
mental. More recently Norman (1991), following Cole and Griffin (1980), argued that the
amplifier metaphor is somewhat misleading. Norman does allow that a person-artifact system can
indeed have enhanced capabilities when compared with an individual acting alone—in fact, this
observation is central for him—but he does not believe that this enhancement is accurately
described as a process of “amplification.” Instead, he argues that artifacts that enhance human
cognitive capabilities do so by changing the nature of the task that an individual must perform,
not by “amplifying” individual action.
24
One common way in which the system view is applied is to describe a person-artifact system
as having an increased memory capacity or increased knowledge base to draw on. Consider the
system consisting of a literate person together with a pencil and a sheet of paper. In a sense, this
system may be said to possess a “better memory” than the person acting alone.
In fact, at least one author, Merlin Donald, takes the system view, together with an emphasis
on memory-enhancing properties of external representations, as a central way to understand the
functioning of the modern human mind. In The Origins of the Modern Mind, Donald attempts to
lay out, in broad sweep, the evolution of the human mind (Donald, 1991). His story posits three
stages in this evolution. The first stage, associated with Homo Erectus, is the development of
“mimetic culture.” According to Donald, Homo Erectus was distinguished from apes in its ability
to “mime, or re-enact, events.” During the second stage, associated with the appearance of Homo
Sapiens, speech developed. In Donald’s scheme, both of these stages involved true biological
evolution in the sense that there were concomitant changes in human biology.
The third stage in Donald’s evolution of the mind is distinguished from these first two in that,
rather than involving any biological changes, it involved the development of what Donald calls
“external symbolic storage.” Donald’s external symbolic storage is precisely what we have been
calling external representations. For Donald, the “storage” feature gets the lion’s share of his
emphasis:
External symbolic storage must be regarded as a hardware change in human cognitive structure, albeit
a nonbiological hardware change. Its consequence for the cognitive architecture of humans was similar
to the consequence of providing the CPU of a computer with an external storage device, or more
accurately, with a link to a network (p. 17).
To be fair, Donald allows a somewhat more complex role for external representations than a
simple memory-aid story. Because his external symbolic storage supplements short-term as well as
long-term memory, its presence can alter the nature of short-term thinking processes. To Donald,
it is as if the very “architecture” of the mind has been augmented; he treats external
representations as comparable to an additional supply of memory that just happens to be
connected through the visual channel.
So what does the “more powerful system” view tell us about the relation between external
representations and the knowledge of individuals? The answer is that it does not tell us much. For
the simple reason that the focus of a system analysis is on the system and not on an individual, it
does not require many presumptions about how the thought processes of individuals must change.
However, the fact that the system view allows external representations to profoundly enhance
25
cognition, without requiring changes in the nature of thought processes, does have important
implications for this discussion. Because it can predict strong effects without positing any major,
or even permanent, changes in the thought processes of individuals, the system view frees us to
take the stance that there are no such important changes.
Incidentally, symbol use in physics is often understood through the lens of the “more powerful
system” view. The idea is that when equations are written on a sheet of paper, they can be
manipulated following a relatively simple set of rules. Thus computations and derivations, which
might be difficult or impossible otherwise, are made easy or, at least, possible. Again, the point is
that this does not necessitate any strong influence on the abilities of individuals in the absence of
external symbols. The person-representation system is capable of performing remarkable
computations, but these remarkable computations still only require the person to perform a
relatively simple task, the rule-based manipulation of symbols.
The Òresidual effectsÓ of symbol use
Restricting ourselves to the system view really begs the question of whether there are any
residual effects on individuals due to the use of external representations. So now let us consider
this question explicitly. Suppose a person is in interaction with an external representation as in
Figure Chapter 1. -4. What types of residual effects might there be due to the direct interaction
of person and representation? Here I consider a few possibilities.
Knowledge is adapted by and for symbol use
Recall that, as I explained above, Norman argued that a person-artifact system gets its
enhanced abilities by allowing the person to perform an alternative, easier task. The important
observation here is this: Although the artifact allows a person to perform a task that is somehow
easier, the individual may still need to learn or otherwise develop new capabilities in order to
perform this alternative task. Thus we see an opening for the possibility of residual effects. The
knowledge and capabilities of an individual may need to be changed and reorganized for the
specific tasks required by representational practices. In fact, the positive results found by Scribner
and Cole can be understood in this way. The representational practices associated with Vai script,
such as letter-writing, could require individuals to develop certain skills.
It is interesting to note that these required changes in what an individual must learn are not
always seen in a positive light. To illustrate, I begin with an analogy. Imagine that a person spends
a significant amount of time using some sort of physical support, such as a crutch. It is possible
26
that certain portions of this person’s physique will gradually be modified. Some muscles may
atrophy. Or, for example, if the person is using a wheelchair, the muscles in their arms might grow
stronger while the muscles in their legs grow weaker. In an analogous manner, Plato worried, in
the Phaedrus, that the use of writing would dampen a scholar’s ability to memorize information
(Plato, trans. 1871).
Similar observations can be made for the case of symbol use in physics. If students can learn to
remember and use certain equations, they may never need to learn the conceptual information
that is somehow embodied in these equations. They can know the equations and not their
meanings. In this way, the use of external representations could have a direct effect on the nature
of an individual’s understanding of physics, literally what an individual “knows.” A student may,
in a sense, just know enough to fill in the gaps between the equations and to use equations
effectively.
The point I am trying to make here is, I believe, relatively uncontroversial. If we spend time
engaged in a particular variety of symbol use, then our knowledge and capabilities will be adjusted
for the requirements of that variety of symbol use. We will learn what we need, and not learn what
we do not need. In this way, the knowledge of an individual may be adapted by and for symbol
use. Of course, in any particular case, we can argue about the type and degree of adaptations, but
the general point still holds.
A symbolic world of experience
Now I want to propose a second way to think about possible residual effects on individual
symbol users due to direct interaction. This viewpoint is not logically distinct from that presented
just above, but I believe that announcing it will help us to recognize and understand some
additional possible relations between knowledge and external representations. Here, I connect to
Feynman’s intuition that one must have a certain type of mathematical-symbolic experience in
order to appreciate the beauty of nature.
I summarize this viewpoint as follows. Start with the (admittedly vague) presumption that
individual knowledge derives from experience. For example, much of our physics-relevant
knowledge comes from our experiences living in the physical world. Before anybody takes their
first physics course, they know what a physical object is, just because they are a person living in the
world. They have held objects, kicked them, painted them, etc. Now here is the key move
associated with this viewpoint: We can simply take the experience of manipulating symbols to be a
27
new class of experience. Working with equations—reading them, writing them, manipulating
them, talking to other people about them—is a kind of experience that need not be any less
meaningful than our experience in the physical world. And this experience forms part of the basis
for our understanding of physics.
Instrumental Psychology and internalization
In speaking of a “symbolic world of experience,” I make it clear that the character of human
experience is shaped by humans. In part, our experience consists of a world of our own devising.
One theory of psychology, the “Instrumental Psychology” of Lev Vygotsky, puts this idea at its
center. For Vygotsky, it is essential that individuals are subject to stimuli not only from the
“natural environment,” they also actively modify their stimuli. Furthermore, the tools for
modifying and mastering the environment are inventions that have been perfected over the course
of human history, and which are somewhat specific to the culture in which the individual is
embedded.
Strikingly, for Vygotsky, not only is it the case that the knowledge of individuals is somehow
affected by stimuli of human devising, these stimuli happen to be the origins of the higher,
uniquely human forms of mental activity. External, social activity is internalized and gives human
thought its particular character. In addition, spoken language—a specific external
representation—is assigned a special role in this process. For Vygotsky, it is language that gives
order to the initially undifferentiated stream of infant thought.
In mastering external speech, the child starts from one word, then connects two or three words; a little
later, he advances from simple sentences to more complicated ones, and finally to coherent speech made
up of a series of such sentences; in other words, he proceeds from a part to the whole. In regard to
meaning, on the other hand, the first word of the child is a whole sentence. Semantically, the child
starts from the whole, from a meaningful complex, and only later begins to master the separate
semantic units, the meanings of words, and to divide his formerly undifferentiated thought into those
units. The external and the semantic aspects of speech develop in opposite directions—one from the
particular to the whole, from word to sentences, and the other from the whole to the particular, from
sentence to word. (Vygotsky, 1934/86, p. 219) [My emphasis.]
Vygotsky takes us close to what is perhaps the strongest form of “direct effect” one can
imagine, the view that thought is, in some sense, an internalized version of external symbolic
activity. For example, someone who adopts this most extreme of views might assert that human
thought is internalized speech. While Vygotsky does not quite propose this extreme view, he does
believe that the use of external signs is primarily responsible for the shaping of the highest levels of
individual thought.
28
Indirect effects of symbol use
In the preceding sub-section I discussed residual effects on the knowledge of individuals due
to direct interactions with external representations. Although they will not be of primary concern
in the rest of this work, I want to briefly consider some hypotheses concerning indirect effects. The
general form of these hypotheses is that the use of certain external representations shapes a
society’s knowledge products in a certain way, and these knowledge products are then internalized
by individuals. Again, these are indirect in the sense that the knowledge of an individual may be
affected even if they are not literate with the external representation in question.
One example of such an indirect effect I call “epistemic structuring.” To illustrate I turn once
again to Jack Goody. In The Domestication of the Savage Mind, Goody extends his discussion of
lists from purely administrative lists, such as the one with which I began this chapter, to lists
designed for a more academic purpose. In particular, he describes the “Onomasticon of
Amenope,” which was composed by an Egyptian around the end of the Second Millennium BCE.
Apparently, this document was the result of an attempt to make a truly exhaustive list of “all
things that exist.” It bears the title:
Beginning of the teaching for clearing the mind, for instruction of the ignorant and for learning all
things that exist: what Ptah created, what Thoth copied down, heaven with its affairs, earth and what is
in it, what the mountains belch forth, what is watered by the flood, all things upon which Re has shone
all that is grown on the back of the earth, excogitated by the scribe of the sacred books in the House of
Life, Amenope, son of Amenope. (p. 100)
For Goody, the fact that this document takes the form of a list is essential; the
representational form has the effect of shaping the ideas that are represented:
We can see here the dialectical effect of writing upon classification. On the one hand it sharpens the
outlines of the categories; one has to make a decision as to whether rain or dew is of the heavens or of
the earth. At the same time, it leads to questions about the nature of the classes through the very fact of
placing them together. How is the binary split into male and female related to the tripartite division
between man, stripling, and old man? (p. 102)
Collins and Ferguson (1993) would say that the list is functioning as an “epistemic form,” a
“generative framework” for knowledge. Goody’s point is that the mere fact of listing leads to
reflection and progress in elaborating the categories used for describing the world.
The idea here is that the “product” of the person interacting with the list is shaped by the
nature of the list structure. Then, once this “product” is produced, it can be fed back into the
knowledge of individuals, both literate and non-literate. Furthermore, after it is processed by
29
society, the product need not even be in the form of a list; distinctions can be taught to nonliterate as well as literate individuals, without restriction to the list structure.
It is important to note that these proposed indirect effects can vary greatly in how profoundly
they purport to alter the nature of individual cognition. For example, we could imagine taking
one of Amenope’s specific distinctions and teaching it to someone. This would not lead to a very
profound change in the nature of that individual’s thought. However, some of the indirect effects
proposed by researchers suggest more general changes in the nature of thinking. Above I
mentioned Goody and Watt’s hypothesis that the invention of the Greek alphabet allowed the
development of techniques for reasoning logically—techniques that could be learned by any
individual.
Similarly, other authors have argued that the appearance of literacy led to generally important
intellectual achievements. David Olson, for example, makes some of his own very strong
hypotheses concerning consequences associated with the development of writing (Olson, 1994).
According to Olson, an important feature of writing, in contrast to spoken language, is that it
must be interpreted without many of the contextual cues that are available during verbal
discourse. For this reason, it is much harder for writers to convey their intended “illocutionary
force.” Therefore, Olson argues, writing must contain additional clues that convey the author’s
stance to what is said. For example an author must tell the reader whether they mean their
statements to be taken literally or as metaphor; and an author must tell a reader whether they
intend to be presenting an hypothesis or a fact. Olson’s big leap is that, for this reason, the
development of writing had to be accompanied by developments in epistemology; the categories
available for describing knowledge had to be made explicit and refined. It is in this manner, he
argues, that categories like “fact” and “hypothesis” were born. With these new categories available,
the cognitive processes of individuals—even non-literate individuals—would be generally
enhanced.
Summary: Where I will stand
In this section I have briefly discussed some hypotheses concerning deep relations between
external representations and the knowledge of individuals. Centrally, I distinguished the direct
effects that arise from an individual’s use of an external representations with indirect effects, which
can even influence non-literates. In this project, it will be helpful to think of my focus as being on
30
direct, non-mediated effects of external representations. I will be interested in the knowledge and
abilities that must develop specifically for symbol use in physics.
Furthermore, I will lean heavily on the notion that physicists develop an intuition—they get a
feel for physics—through participation in what I have called “the symbolic world of experience.” I
will argue that, to know physics as physicists know it, one must be an initiate into the full world of
physics practice, including learning to write and manipulate symbols, and learning to say things
with and about equations in the manner of a physicist. Note that this view implicitly presumes
that experience with symbols is not any less meaningful or important than other types of
experience, such as daily experience interacting with objects in the physical world. In fact, a major
part of the work of this document will be to illustrate that, for physicists and even for mildly
advanced physics students, an equation is an element in a highly structured and meaningful world.
However, my own orientation stops somewhat short of strong versions of the internalization
view. I do not believe that symbol use is uniquely responsible for the development of human
thought. Nor do I believe that direct interaction with external representations tends to engender
broad changes in the character of the thought of individuals, such as a move toward more
“logical” or “scientific” thought. External representations are a meaningful and important element
of our experience, but activities with external representations are not unique in giving human
thought its character.
Finally, it should be noted that, throughout the discussion in this section, I have essentially
presumed that it is useful to speak of knowledge as something that is localized in the minds of
individuals. Though, to some, this may seem like an obvious and unproblematic move, some
researchers have questioned whether it is appropriate and scientifically productive. Researchers
such as Greeno and Moore (1993) and Hall (1996) take the stance that knowledge is most
profitably understood as a theoretical entity that spans person and elements of the environment.
Note that, since the environment may include external representations, this view provides a quite
difference sense in which knowledge and external representations may be related.
I will not adopt the stance that knowledge must be interactionally defined in this manner;
throughout this document, I will talk as if knowledge is something that can exist localized in
individuals. I will have somewhat more to say on this topic in Chapter 7, but I will make one
additional point here. The notion that knowledge must be interactionally defined is an element of
31
some versions of a more general viewpoint that is known under such names as “situated action.”1
Situated action, as compared to traditional cognitive science, tends to emphasize the influence of
social and historical factors on individual behavior, as well as the importance of the immediate
environment. In allowing that knowledge is, in some sense, dependent on specific practices,
including the use of cultural artifacts such as equations, I hope to be taking into account many of
the concerns that are central to situated action. Again, I will have more to say on this subject later.
Concerning Physics Learning and Instruction
A central goal of this inquiry is to contribute to research in physics learning and instruction.
We want to know what it means to understand physics, and why students do or do not succeed in
learning this topic. Of course, the emphasis here remains on symbol use, how the using of
equations contributes to and is part of physics understanding. In this section, I will describe how
the current project ties into work that has already been done.
Figure Chapter 1. -6. A typical misconception.
Intuitive knowledge and the learning of physics
In the early 1980’s, a number of studies appeared that seemed to indicate that physics
instruction was, stated simply, a disaster. Although students could solve many of the problems
that are typical of introductory instruction, such as the one shown in Figure Chapter 1. -1, they
apparently could not answer some very basic qualitative questions. For example, a typical question
asked students what happens when a ball is shot out of a spiral tube (Figure Chapter 1. -6). In
response, students frequently stated that the ball continues to move in a curved path after leaving
the tube (e.g., McCloskey, 1984). This answer is not correct; once the ball leaves the tube, it
1
See, for example, the special issue of the journal Cognitive Science that is devoted to this topic
(ÒSituated Action,Ó 1993).
32
moves in a straight line, continuing in the direction that it had at the moment it exited the tube.
These alternative answers were attributed to “misconceptions” or “alternative conceptions” by
researchers.2
The presence of misconceptions such as this one is shocking because it suggests that students
lack an understanding of some of the most fundamental aspects of Newtonian physics. For
example, Newton’s first law states that:
Every body perseveres in its state of rest, or of uniform motion in a right line, unless it is compelled to
change that state by forces impressed thereon. (Newton, 1687/1995)
What this means is that, if you go out into outer space and throw a baseball, the baseball continues
forever, moving with the same speed and in the same direction that you threw it. It never slows
down and it never turns, unless it happens to come near to a planet or asteroid. It does not matter
how the ball is launched; the ball doesn’t care and it doesn’t remember. Even if it is launched from
a spiral tube, once the ball leaves the tube, it just continues in a straight line. Thus the spiral tube
misconception is in contradiction with one of the first things that students learn in their
introductory physics classes, Newton’s first law.
And such errors appear to be commonplace. Halloun and Hestenes (1985b), in a study that
employed a multiple choice test given to hundreds of university physics students, found that 44%
of these students exhibited the belief that a force is required to maintain motion. (Though
individual students tended to answer questions inconsistently.) Nearly half, it seems, did not learn
Newton’s first law!
These results led to the obvious question: What is the problem with physics instruction?
Researchers hypothesized that part of the answer was that students spent the vast majority of their
time engaged in solving problems. And it was possible to get good at solving these problems
without really understanding the content of the equations that you were using. I will return to this
hypothesis below.
But the over-emphasis on problem solving could not be the whole story. The fact that objects
continue moving in the absence of forces seems like a straightforward notion, even if students were
not given much time to mull it over. This fact even seems like something that can be memorized,
whether or not you believe it. What is so intrinsically difficult about the notion that objects just
2
For a variety of accounts of student difficulties see, for example: R. Cohen, Eylon, and Ganiel, 1983;
Halloun and Hestenes, 1985a & 1985b; McDermott, 1984; Trowbridge and McDermott, 1980 & 1981;
Viennot, 1979. Also, see Smith, diSessa, and Roschelle (1993) for a critique of the misconceptions
33
keep moving in a straight line? In some manner, it seemed, students were actually resistant to
learning physics.
What researchers came to realize was that students’ existing knowledge of the physical world
has a strong impact on the learning of formal physics. In fact, students start physics instruction
with quite a lot of knowledge about the physical world. Furthermore, this knowledge is based in
quite a number of years of experience, so it might well be resistant to change.
So, one of the major outcomes of this “disaster” research was the realization that students have
prior knowledge of the physical world that is relevant to physics learning. The study of this prior
knowledge—often called “intuitive physics” or “naive physics” knowledge—has become an
endeavor of its own. When it takes its simplest form, naive physics research simply catalogs
student’s alternative conceptions (e.g., Halloun & Hestenes, 1985b). Further research attempted
to go beyond mere lists of misconceptions to develop generalizations about the nature of student
beliefs. These generalizations were directed at answering two distinct questions:
1. Can intuitive physics be formalized?
2. How is intuitive physics actually realized as knowledge in people?
The first question takes student’s answers to misconception-style questions and searches for a
specification of some underlying physics that these answers reflect. To understand the nature of
this approach, imagine that it was applied to the statements of expert physicists. In that case, the
assumption is that this methodology would produce Newtonian physics (Newton’s three laws,
literally) as the formalization of a physics underlying expert answers.
However, in searching for a similar formalization of intuitive physics, researchers must make
many questionable assumptions. In particular, they must interpret students’ answers to questions
as reflecting more general beliefs. In so doing, they beg the question of whether a formalization
exists.
Furthermore—and this is the point of the second question—even once you have found a
formalization, you still do not know how this physics is realized as knowledge in people. For
example, even if we know that expert physics formalizes as Newton’s laws, we still know very little
about what knowledge experts actually have. Most of the first formalizers, such as Halloun and
Hestenes (1985b) and Viennot (1979), gloss this issue with statements like: “[students show] the
perspective.
34
belief that, under no net force, an object slows down” (Halloun & Hestenes, 1985b). What is left
often by such assertions is how such a belief is realized as knowledge. Do students actually possess
propositional knowledge of this form? Or is there some more fundamental type of knowledge
that causes students to behave, in certain circumstances, as if they have this particular belief?
Throughout this manuscript, I intend particularly to build on the work of one researcher who
is self-consciously directed at answering the second question. In his Toward an epistemology of
physics, Andrea diSessa sets out to describe a portion of intuitive physics knowledge that he calls
the “sense-of-mechanism” (diSessa, 1993). The idea is that elements of the sense-of-mechanism
form the base level of our intuitive explanations of physical phenomena. As an example, diSessa
asks us to think about what happens when a hand is placed over the nozzle of a vacuum cleaner.
When this is done, the pitch of the vacuum cleaner increases. According to diSessa, the way people
explain this phenomena is they say that, because your hand is getting in the way of its work, the
vacuum cleaner has to start working much harder. The point is that this explanation relies on a
certain primitive notion: Things have to work harder in the presence of increased resistance if they
want to produce the same result.
diSessa’s program involves the identification of these primitive pieces of knowledge which he
calls “phenomenological primitives” or just “p-prims,” for short. (The p-prim that is employed to
explain the vacuum cleaner phenomenon is known as “OHM’S P -PRIM .”) The word
“phenomenological” appears in the name of this knowledge element in part because they are
presumed to be abstracted from our experience. OHM’S P -PRIM , for example, is abstracted from
the many experiences we have in the physical world in which things have to work harder in the
presence of increased resistance. diSessa lists a variety of p-prims, including p-prims pertaining to
force and motion (e.g., FORCE AS MOVER , DYING AWAY ) and p-prims pertaining to constraint
phenomena (e.g., GUIDING , ABSTRACT BALANCE). Furthermore, during physics instruction, the senseof-mechanism is refined and adapted, and comes to play an important role in expert physics
understanding. I will have more to say about diSessa’s theory in later chapters, particularly in
Chapter 5.
Problem solving in physics
Contemporaneous with this research into conceptual understanding and naive physics, a
second and essentially disjoint set of studies were examining how students—and, to a lesser
extent, expert physicists—solve textbook physics problems (Bhaskar & Simon, 1977; Chi,
35
Feltovich, & Glaser, 1981; Larkin, McDermott, Simon, & Simon, 1980). In its earliest
incarnations, these studies of physics problem solving were strongly influenced by prior research
into human problem solving, conceived more generally. For example, some early research in
artificial intelligence and cognitive science examined how people approached the solution of the
famous “Tower of Hanoi” puzzle. (Simon, 1975, is a typical example.) The Tower of Hanoi
puzzle involves a set of rings that can be moved among three pegs (Figure Chapter 1. -7). The
object of the puzzle is to move all of the rings from the left hand peg to the far right hand peg.
The catch is that at no time can a larger ring be on top of a smaller ring. Furthermore, only one
ring may be moved at a time.
Figure Chapter 1. -7. The Tower of Hanoi puzzle
An essential feature of this puzzle is that, at any time, there are no more than a few possible
moves. You can only move the topmost ring on any peg to one of the other pegs. And many such
moves will be prohibited because of the rule that larger rings cannot be on top of smaller rings.
Furthermore, the current state and desired end-state of the puzzle are clearly and simply defined.
In this early cognitive science literature, the solution of the Tower of Hanoi puzzle was
described as a search through the space of the possible states of the puzzle. From this viewpoint,
the most interesting question to be answered was exactly how a person goes about searching this
space; in other words, we would want to know how, at any time during the solution of the puzzle,
a person chooses among the few possible available moves. What Simon (1975) found was that
problem-solvers could be well described by a relatively simple strategy, what he called “meansends analysis.” In means-ends analysis, a problem solver chooses the move that appears to most
decrease the perceived difference between the current state and the goal state.
Initially, this research was extended to physics problem solving in a quite straightforward
manner (Bhaskar & Simon, 1977; Larkin et al., 1980). The notion was that moves in the space of
equations written on a sheet of paper could be treated much like moves in the solution of a puzzle.
More specifically, physics problem solving was seen to work like this: The student or expert reads
the problem and takes note of the quantities that are given in the problem and the quantities that
are needed. Then the problem solver writes, from memory, equations that contain these various
quantities as variables. Once the equations are written, the search can begin; the equations are
36
manipulated until a route from the given quantities to what is given is obtained. The problem
shown in Figure Chapter 1. -1 provides a simple example. In that problem, the student is given the
initial velocity, vo , the acceleration, a, and asked for the time. Once the equation v = vo + at is
written, it can simply be solved for the time, t.
Of course, in more complicated problems, some type of strategy is needed to find your way
through the space of equations from the givens to the solution, just as means-ends analysis was
needed in the Tower of Hanoi problem. Larkin and colleagues (1980) found the interesting result
that novices work by “backward chaining:” They start with the quantity that they want to know,
then write down an equation in which that quantity appears. Each of the unknowns in this new
equation then becomes a new quantity to be determined, and the process iterates. In contrast,
experts were found to “forward chain;” that is, work forward from the quantities given by writing
equations in which these quantities appear, and then using these equations to find new quantities.
ÒUnderstandingÓ and problem solving
I am now approaching the point where I will say how the current inquiry fits in with all of this
prior research. First, notice that there are some obvious limits in the above problem solving
research. As I have presented it, the physics problem solving research does not posit much of a role
for anything like “understanding” during problem solving and the use of equations. The story is
simply that you see the quantities that appear in the problem, you dredge equations out of
memory and write them down, then you manipulate the equations according to simple rules to
find an answer. So where is “understanding?” Larkin and colleagues (1980) themselves comment:
The major limitation we see in the current models is their use of an exceedingly primitive problem
representation. In fact, after the initial problem representation, these models work only with algebraic
quantities and principles described by algebraic equations. Thus, in a sense, they have no knowledge of
physics, but only the algebraic representation. This may actually not be too bad for capturing the
performance of novice solvers, who have little knowledge of physics, but much more of algebra.
However, it is certainly inadequate for capturing the work of more competent solvers.
But “understanding” may be tricky to define and to recognize. Perhaps, knowing equations
and knowing how to execute the above problem solving process is exactly what constitutes
understanding. Even if we adopt this improbable—and, I think, unpalatable—stance, it is still
possible to find some quite obvious and definite lacks in the above view of equation use. Most
obvious of these lacks is the fact that, if equations are only written from memory and are not
“understood” in any deeper sense, then how is it possible that people ever compose completely
new equations, corresponding to newly discovered principles and never-before-modeled systems?
37
There are some obvious ways out of this bind. One could maintain, quite reasonably, that the
above model of problem solving is only designed for a certain regime of problem solving involving
well-known principles and systems. But, if we accept this view, then we are simply left with the
new problem of describing the unaccounted for regimes of equation use and problem solving.
Furthermore, one might maintain that the composition of new equations is something that is only
done rarely and by scientific geniuses, like Einstein and Feynman. The rest of us can then make
relatively mechanical use of the equations that they develop for us. I will attempt to show, in
succeeding chapters, that this is simply not how things are.
The “naive physics” researchers may seem to have an advantage here, since they discuss
behavior that, at least intuitively, appears closer to “understanding.” But these researchers have yet
to tell us precisely how intuitive physics and conceptual knowledge is employed during the use of
equations. In what way do physicists make use of the sort of qualitative understanding that these
researchers are concerned with during problem solving? Is there any relation or are there simply
two sets of things to learn to become a physicist?
This is where I come in. In a sense, this inquiry will combine these two areas of research that,
to this point, have remained relatively disjoint. I will be examining phenomena akin to
misconceptions, naive theories, and p-prims, but in the context of problem solving and equation
use. Strikingly, we will see that there is actually a very strong connection between intuitive physics
knowledge and equation use; equations are understood and generated from knowledge that is
very closely related to intuitive physics knowledge. Describing this tight link between equations
and understanding is the central focus of this work.
Before proceeding, I want to note that the above account of the problem solving literature is a
little unfair. The same researchers have subsequently elaborated their work to account for some of
the difficulties that I have mentioned. A notable example is Larkin (1983), which I will discuss
extensively in Chapter 3. Also, the substantial success of the problem solving literature deserves
some emphasis. These researchers were very successful in building models that accurately
described subjects’ behavior in the regimes of problem solving that they studied. Thus, it is likely
that future research can profitably build on this work.
On the teaching of Òconceptual physicsÓ
I want to close this section with a note concerning how the results of the above research on
physics learning, particular the disastrous results reported by the misconceptions research, have
38
been interpreted as a prescription for improving physics instruction. The fact that intuitive physics
exists has led to the notion that it should be directly addressed in some way. Take the example of
the spiral tube misconception and students’ failure to understand the implications of Newton’s
First Law. Why not take the time to emphasize the (qualitative) principle that objects continue to
move in a straight line with a constant speed in the absence of intervening forces?
This prescription—emphasize the qualitative principles at the expense of problem solving—has
been taken up in a variety of courses that I will refer to as “Conceptual Physics” courses.3 These
courses operate under the principle that they should emphasize the “conceptual” content of
physics. In practice, this means that there is little problem solving, and few equations. Though, of
course, this varies greatly from course to course.
Given the viewpoint I am beginning to lay out here, it is clear that we should at least take the
time to reconsider this prescription for instruction. I have proposed replacing algebraic notation
with a programming language. Conceptual Physics courses make a similarly strong move, they
replace equations with plain language statements. If knowledge is inextricably representational,
then what are the consequences if the equations are removed altogether? What would Feynman
say?
I believe that the consequences are not necessarily bad; they may very well not be worse than
the consequences of my proposal to use programming languages. But I do want to make the point
that we need to understand what we’re doing if we choose to teach physics in the manner of
Conceptual Physics. I challenge the assumption that—in physics or any domain—we can separate
the “conceptual” from the symbolic elements of a practice for the purposes of instruction. In
removing equations from the mix, we change the constitution of understanding. This does not
imply that we cannot teach physics without equations. However, it does imply that equation-free
courses will result in an understanding of physics that is fundamentally different than physics-asunderstood-by-physicists.
Again, I must make clear that I do not mean to critique Conceptual Physics courses, rather I
hope to better understand the consequences of omitting equations. In fact, in the vein of these
more “conceptual” techniques, there is already a history of striking successes and exciting
proposals. For example, White and Horwitz designed and implemented a middle school
curriculum based around computer models. The students in this curriculum performed better on
3
See, for example, the textbook ÒConceptual PhysicsÓ (Hewitt, 1971). This textbook has been reprinted
in a number of more recent additions.
39
many tasks than students in a traditional high school course (White, 1993b). Clement, through his
technique of “bridging analogies,” has had success in training expert intuition (Clement, 1987).
And I suspect that even many working physicists could learn a great deal from Haertel’s
impressive reworking of the entire topic of electricity and magnetism in qualitative terms (Haertel,
1987). All of this published research is in addition to the many instructors at universities and high
schools that are enjoying success in designing and teaching Conceptual Physics courses. My work
here will help us to better understand the character of conceptual physics, so that we can
understand and better judge the results of such an approach, in clear contrast to alternatives.
Incidentally, the instructional move embodied in Conceptual Physics courses appears to be
common in much of the recent innovation in mathematics and science instruction. “Teaching for
understanding” seems to frequently involve a scaling back in the use of formalism. I do not intend
necessarily to question whether this is a productive instructional maneuver. However, if the
concerns expressed here are valid, then this sort of “teaching for understanding” may lead to new
sorts of understanding of mathematics and science.
Chapter Overview
Part 1. Algebra-physics and the Theory of Forms and Devices
The remainder of this dissertation is divided into two major parts. In Part 1 I will describe the
Theory of Forms and Devices, drawing on examples from the algebra-physics portion of the data
corpus. Thus, the first part of the dissertation will perform the dual roles of describing the theory
and applying it to describe algebra-physics.
Part 1 is divided into six chapters, Chapters 2 through 7. In Chapter 2 I take some steps
preparatory to the introduction of the major theoretical constructs in the following chapters. This
chapter will include some accounts of students using equations, accounts which are designed
simply to show that students can use equations in a meaningful manner. At the end of this chapter
I will also introduce the notion of “registrations.”
In Chapter 3 I will embark on a detailed discussion of symbolic forms. I will explain the
nature of symbolic forms in more detail, and I will individually discuss each of the specific
symbolic forms needed to account for the observations in the algebra data corpus. This will
include numerous examples from the data. In addition, I will argue against some competing
viewpoints, including what I call the “principle-based schemata” view.
40
Chapter 4 introduces and describes representational devices. As with symbolic forms, I will
discuss each of the individual devices needed to account for the observations in the data corpus
and I will draw examples from the data to illustrate my points. Along the way, I will have reason
to discuss what it means to “interpret” a physics equation. Importantly, we will see that the
“meaning” of a physics expression, as it is embodied in students’ interpretive utterances, is highly
dependent on the concerns of the moment. There is no single, universal manner in which meaning
is attached to equations.
In Chapter 5, I will speculate on the development of the knowledge system described in the
preceding chapters. My basic tactic here will be to draw on existing accounts of students’
knowledge as it is found prior to instruction, namely the research on intuitive physics knowledge
that I described above. Most prominently, I will draw on diSessa’s theory of the sense-ofmechanism. By comparing diSessa’s account to my own model, I will be able to comment on the
relationship between new and old knowledge, and I will construct hypothetical developmental
paths to describe how symbol use may contribute to the development of physics understanding.
One important result of this discussion will be the observation that expert “physical intuition” is
sensitively dependent on the details of the symbolic practices of physics.
In Chapter 6, I will present the results of applying the model systematically to describe the
whole of the algebra data corpus. In that chapter, I will describe the algebra corpus in some detail,
as well as the iterative process through which the data corpus was analyzed. In addition, the
chapter will contain an attempt to characterize aspects of the algebra corpus in a quantitative
manner. One reason for this quantitative characterization is to provide the reader with a means of
drawing conclusions concerning the prevalence of the phenomena that I describe, without having
to read the entirety of the data corpus. For example, through this quantitative characterization, we
will arrive at a measure of how often students construct novel equations and interpret equations.
Both of these phenomena turn out to be quite common. I will also present results concerning the
frequency of individual forms and representational devices.
The reader should be warned that, although the analysis process I will describe in Chapter 6 is
meticulous and detailed, it does not provide an exact and rigorous method of extracting my
model from the data. Instead, it is a highly heuristic process, and the recognition of specific
knowledge elements in the data should be understood as rooted in my intuition and theoretical
predispositions, as well as in the data itself. For this reason, much of the rhetorical burden of
arguing for my view must be carried by Chapters 2 through 5.
41
In Chapter 7—the last chapter of Part 1—I will reflect on the theory, draw out some
implications of the theory, and I will attempt to generalize the observations of the preceding
chapters outside the immediate concerns of symbol use in physics. Included in this chapter will be
a discussion of the implications of the preceding results for the relation between physics
knowledge and external representations. In addition, I will contrast my own theoretical position
with a variety of other theoretical accounts.
Part 2. Programming-physics, its analysis, and a comparison to algebra-physics
In Part 2, the discussion turns to programming-physics. This is the first time that
programming will appear following this introductory chapter. The primary concerns of this part of
the dissertation are to apply the Theory of Forms and Devices to describe programming-physics,
and to use the theory to compare programming-physics and algebra-physics.
Chapter 8 introduces the reader to the particular practice of programming-physics that was
the subject of this study. I will describe the programming language used and give examples to
illustrate the types of programs written by students.
In Chapter 9, I apply the Theory of Forms and Devices to programming-physics. This will
include a discussion of individual forms and devices with examples drawn from the programming
data corpus. Particular attention will be paid to the forms and devices that do not also appear in
algebra-physics.
Chapter 10 presents the systematic analysis of the programming data corpus. The content of
this chapter will be parallel to that of Chapter 6, in which I described the systematic analysis of the
algebra corpus.
In Chapter 11, I will use the analyses of algebra-physics and programming-physics in terms of
forms and devices to compare these two practices. To briefly preview the results, I found that the
distribution of forms in programming-physics was substantially similar to that in algebra-physics;
however, an important new class of forms appeared that was completely absent in the algebra
corpus. These new forms relate to schematizations of processes. With regard to representational
devices, there were some interesting differences in the frequency with which the various classes of
devices were employed.
Finally, in Chapter 12, I will summarize the thesis and trace the instructional implications of
this work. This will include a discussion of the implications of this work for traditional physics
42
instruction. I will argue that, given an improved understanding of how students use and interpret
equations, we may be able to recognize and remediate some types of student difficulties.
Finally, not least among the educational implications of this work is the fact that it considers
the possibility of a substantially altered practice of instruction, built around student programming.
In the last chapter, I will finally calibrate the strengths and weaknesses of this alternative approach,
to the extent that this study allows.
43
Chapter 2. Prelude to Theory
Rij − 12 gij R − Λgij = 8πGT ij
In the years around 1916, Albert Einstein presented the world with a new theory of gravitation.
And a truly new theory it was. In the existing Newtonian theory, gravity was “explained” by the
simple statement that there is an attractive force between every two masses. Pick any two masses,
m1 and m2, separated by a distance r. Then each of those masses feels a force, directed toward the
other mass, of size
F=
Gm 1 m 2
r2
In Einstein’s theory, the nature of gravity was wholly reconceptualized. Rather than explaining
gravity by an appeal to some unexplained “action at a distance,” Einstein reduced gravity to the
tendency of matter to distort space. An object, like the sun, reshapes the space around it. Then
when other objects—like the Earth—travel through this distorted space, their motion is directed
by the shape of space, just like a marble moving around the inside of a curved metal bowl. At the
heart of this new theory is Einstein’s “Field Equation,” written at the head of this chapter, which
relates the curvature of space to the distribution of matter. What we need to note about this
equation is that, somehow, a portion of Einstein’s reconceptualization of gravity is embodied in
this expression. Einstein began with the notion that matter distorts geometry, mixed in the old
Newtonian theory and some essential principles, and produced a new equation, one that did not
exist before.4
As we saw in Chapter 1, the existing literature concerning physics problem solving is not
capable of accounting for Einsteinian behavior of this sort. In that literature, problem solving is
dependent on a database of known equations, so it cannot explain the development of wholly new
4
Physicists might complain that textbooks do, in fact, contain ÒderivationsÓ of EinsteinÕs field
equation. However, a close examination of these ÒderivationsÓ reveals that they do not begin with
accepted physical laws and then perform logical manipulations to derive consequences of these laws.
Rather, the Field Equation is derived in part from some assumptions concerning what a geometric
theory of gravity must look like, such as what quantities the curvature of space may reasonably depend
on.
44
equations, or modifications based on conceptual considerations. However, it may not be fair or
appropriate to think of this as a deficiency in that literature or even to think of it as a task for
education research. The focus of the problem solving literature (and of this study) is on physics
learning and instruction. Our job here is to describe what students do and what we can reasonably
expect them to learn in an introductory course. And it’s possible that the construction of novel
equations is solely the province of physical geniuses, like Einstein and Feynman, or at least expert
physicists. Even if such invention does not require special mental prowess, it may nonetheless
depend on possessing some esoteric skills. Thus, the rest of us, including introductory physics
students, should perhaps only expect to use the equations that the experts write. If this is the case,
then accounting for this behavior may be less important for educationally directed research, both
because it is not really relevant to learning, and simply because it is relatively rare.
In the next sections of this chapter, I will take the first step toward showing that this view is
untenable. Equations have a perceived meaningful structure for all initiates in physics, even
introductory physics students. We will see that students can and do construct novel equations,
and they use equations in a meaningful manner, flexibly constructing their own interpretations. In
this chapter, I will begin by presenting some examples designed only to show that students can, at
least sometimes, construct and interpret equations. These examples will also provide a first taste of
what these behaviors look like. Then, in the latter part of this chapter, I will take some of the first
steps toward a theoretical account of this behavior. In subsequent chapters I will complete the
theoretical account and I will show through systematic quantitative analysis that these phenomena
are actually quite common.
Young Einsteins
Of all the tasks that I asked students to do, students found a task that I call “The Shoved
Block” problem the most exasperating. They found it exasperating, not because it is difficult, but
because it is a seemingly easy problem that produces counter-intuitive results. In the Shoved Block
problem, a block that rests on a table is given a hard shove. The block slides across the table and
eventually comes to rest because friction between the block and the table slows the block down.
The problem then asks this question: Suppose that the same experiment is done with a heavier
block and a lighter block. Assuming that both blocks are started with the same initial velocity,
which block travels farther?
45
v = vo
v = vo
Figure Chapter 2. -1. A heavier and a lighter block slide to a halt on a table. If they are shoved so as to have
the same initial speed, they travel exactly the same distance.
The counter-intuitive answer to this question is that the heavier and lighter block travel
precisely the same distance (refer to Figure Chapter 2. -1). In a later chapter I will describe how
students used equations to obtain this result. However, for the present purposes, I want to describe
how the dissatisfaction of a pair of students with this counter-intuitive result led them to invent
their “own brand of physics.”
Students actually express two competing intuitions concerning the Shoved Block. The most
common intuition is that the heavier block should travel less far. The reason given is that this is
true because the heavier block presses down harder on the table, and thus experiences a stronger
frictional force slowing it down. The alternative intuition is that the heavier block should travel
farther, since it has more momentum and is thus “harder to stop.” In fact, it is because these two
effects precisely cancel that the heavier and lighter block travel the same distance.
The pair of students I am going to discuss here adopted the latter intuition, that heavier blocks
should travel farther. And they clung to this intuition, even after they had used equations to derive
the “correct” result. This was especially true of one student in the pair, Karl5, who argued that it
was simply “common sense” that the heavier block should travel farther.
Karl
Yeah, that's true. But I still say that the heavier object will take the longer distance to stop
than a lighter object, just as a matter of common sense.
His partner, Mike, was somewhat less worried by the counter-intuitive result, but he was
willing to stop and discuss what might be changed in their solution. Eventually, Karl came to the
conclusion that “in real life” the coefficient of friction probably varies with weight in such a way
that heavier blocks travel farther.
Karl
I think that the only thing that it could be is that the coefficient of friction is not constant.
And the coefficient of friction actually varies with the weight.
The coefficient of friction is a parameter that determines the size of the frictional force. Students
are taught in introductory physics classes that this parameter can be treated as a constant that
46
depends only on the nature of the materials involved, such as whether the objects are made out of
wood or steel.
Given the counter-intuitive result, Mike and Karl were willing to reconsider the assumption
that the coefficient of friction is a constant. After some very minimal cajoling, (I said only “Do
you want to try putting in maybe some hypothetical weight dependence to that”), they set out to
compose their own expression for the coefficient of friction. Karl began by expressing their core
assumption:
Karl
I guess what we're saying is that the larger the weight, the less the coefficient of friction
would be.
To obtain their desired result—that heavier masses travel farther—they needed to have the
coefficient of friction decrease with increasing weight. In other words, friction must work less on a
heavier mass. Over approximately the next nine minutes, they laid out some additional properties
that they wanted their new expression to have. Here’s an excerpt (refer to Appendix A for a key to
the notations used in transcripts):
Karl
Well yeah maybe you could consider the frictional force as having two components. One
that goes to zero and the other one that's constant. So that one component would be
dependent on the weight. And the other component would be independent of the weight.
Mike
So, do you mean the sliding friction would be dependent on the weight?
Karl
Well I'm talking about the sliding friction would have two components. One component
would be fixed based on whatever it's made out of. The other component would be a
function of the normal force. The larger the normal force, the smaller that component. So
that it would approach a - it would approach a finite limit. It would approach a limit that
would never be zero, but the heavier the object, the less the coefficient of friction at the
same time. (…)
Mike
I don't remember reading that at all. [laughs]
Karl
See, I'm just inventing my own brand of physics here. But, if I had to come up with a way if I had to come up with a way that would get this equation to match with what I think is
experience, then I would have to - that's what I would have to say that the=
Mike
Actually, it wouldn't be hard to=
Karl
=the coefficient of friction has two components. One that's a constant and one that varies
inversely as the weight.
In this passage, Karl outlines what he wants from the new expression. He decided that the
coefficient of friction should have two components. One of these components is independent of
weight, the other decreases with increasing weight. After some fits and starts, Mark and Karl,
settled on this expression for the coefficient of friction:
µ
µ = µ 1 + C m2
5
All student names are fictitious.
47
Here µ1, µ2, and C are constants and ‘m’ is the mass. This expression does have some problems,
most notable of which is the fact that the coefficient of friction tends to infinity as the mass goes
to zero. But this expression does capture much of what Mike and Karl wanted to include, it
decreases with increasing mass, and it approaches a constant as the mass becomes infinitely large.
The most important thing to note about this expression is that you will not find it in any
textbook. Mike and Karl did not learn it anywhere, and they didn’t simply derive it by
manipulating equations that they already knew; instead, they built this expression from an
understanding of what they wanted it to do. As Karl says, he was “just inventing his own brand of
physics.” Their expression is not as fancy and subtle as Einstein’s Field equation, but it is clearly a
non-trivial, novel construction. It seems that the construction of novel equations is not only the
province of experts like Einstein.
The question for us to answer in the upcoming chapters is: How did they do this? How did
they know to write a ‘+’ instead of a ‘×’ between the two terms? How did they know to put the m
in the denominator? These are precisely the behaviors we need to be able to explain; we want to
know what knowledge gives students the ability to connect conceptual content and equations.
After a few more preliminaries, that is what we will do.
Sensible Folk
Episodes like the last one were a little uncommon in my data corpus, though not really rare. In
fact, like Mike and Karl, all of the pairs in my study demonstrated an ability to compose new
expressions at some point during their work, and without obvious difficulty. Nonetheless, these
episodes were uncommon because most of the physics problems that students are asked to solve
do not require them to invent novel equations. In this next example, I am going to describe an
episode that should appear almost painfully mundane to those who are familiar with introductory
physics. The important fact about this episode is that it includes an instance of a student
interpreting an equation. Such episodes are not uncommon because, although students to not need
to invent totally new expressions very often, they do frequently need to interpret equations that
they derive.
In this episode, Jack and Jim were asked to solve the following problem, in which a mass hangs
at the end of a spring:
A mass hangs from a spring attached to the ceiling. How does the equilibrium position of the mass
depend upon the spring constant, k, and the mass, m?
48
This problem was extremely easy for them and they spent only about 2 1/2 minutes working on
it. Their board work is reproduced in Figure Chapter 2. -2.
k
x
F=-kx
m
F=ma
kx=mg
x=
mg
k
Figure Chapter 2. -2. Jack and Jim's solution to the mass on a spring problem.6
Jack and Jim began by explaining that there is a force from gravity acting downward, a force
from the spring acting upward, and that these forces must be equal for the mass to hang
motionless:
Jim
Mmm. Well, there's the gravitational force acting down. [w. Arrow down; F=ma next to
arrow.] And then there is
Jack
a force due to the spring // holding it up.
// a force due to the spring which I believe is,, [w. F=-kx] Is equal
Jim
to that.
Jack and Jim know that the force of gravity on an object is equal to mg where m is the mass of the
object and g is the gravitational acceleration, a constant value. They also know that springs apply a
force proportional to their displacement. If a spring is stretched an amount x from its rest
length—the length it likes to be—then it applies a force kx, where k is a constant known as the
“spring constant.”
Jack and Jim went on to equate these two forces, writing kx=mg. Then they solved for the
displacement x to obtain their final expression:
x=
mg
k
Then Jim immediately explained that this is a completely sensible result:
Jim
Okay, and this makes sort of sense because you figure that, as you have a more massive
block hanging from the spring, [g. block] then you're position x is gonna increase, [g.~ x in
diagram] which is what this is showing. [g.~ m then k in x=mg/k] And that if you have a
6
All diagrams are my own reproductions of student work. In creating these reproductions I attempted to
retain the flavor of student drawing while rendering important elementsÑespecially symbolic
expressionsÑclearly.
49
stiffer spring, then you're position x is gonna decrease. [g. Uses fingers to indicate the
gap between the mass and ceiling.] That why it's in the denominator. So, the answer
makes sense.
Jim says that this expression makes sense for two reasons. First, he tells us that the equation says
that, if you increase the mass (with k held constant) then x increases. Second he considers the case
where the spring is made stiffer; i.e., k is increased. In this case, he says, the equation implies that x
will decrease. Presumably, these observations make sense to Jim because of knowledge that he has
concerning the behavior of springs.
I consider this statement by Jim to be an “interpretation.” Jim is not merely providing us with
a literal reading of the symbols on the page; he does not simply say “X equals M G over K.”
Rather, he tells us what the equation “says.” For example, one thing that the equation “says” is
that if you increase k then x decreases.
Even this very simply interpretation has a number of interesting characteristics that are worth
mentioning. In saying that as k increases x decreases, Jim is imagining a very particular process: k is
increasing rather than decreasing, and m and g are parameters that are implicitly held fixed.
Describing and accounting for these particulars is part of the job of this study. I would like to
know: What do interpretations look like, in general? Do they always involve holding some
parameters fixed and varying others? Furthermore, how does Jim know what the equation says?
How does he know, by looking at the equation, that if k increases then x will decrease?
Interpretations of this sort are a simple and important phenomenon that are not included in
existing accounts of equation use in physics. Interpretation has no place in a model in which
equations are just written from memory and then manipulated to produce a numerical result. This
might be acceptable if this behavior was rare or unimportant, but that is not the case. To those
familiar with introductory physics, Jim’s interpretation will appear so obvious and common that it
is almost unnoticeable. Later on, the systematic analysis will show just how frequently these
interpretations appear in my data corpus. In fact, almost every problem solving episode involves
some interpretations. And students interpret, in some manner, a relatively high percentage of the
equations that they write. So, in addition to any prima facie importance of these phenomena,
there is the simple fact that interpretations are extremely common in students’ problem solving,
and therefore should be taken into account in any model.
50
A Feel for the Numbers
The above examples begin to illustrate what it means for students to use equations
meaningfully. Mike and Karl composed an expression from an idea of what they wanted from the
expression. And Jim told us what an equations “says.” In both of these examples, the equations
were, in a sense, directly meaningful for the students; the meaning was somehow seen in the
structure of the equations itself.
The ability to check the sensibility of results is very important. “Is my work correct?” and
“Does this equation make sense?” are extremely useful questions to be able to ask and answer
when working with equations. The point I want to make with this next example is that equation
users have many methods available for determining if what they have written is correct; they do
not always need to see meaning directly in equations.
These alternative methods of checking the sensibility of expressions will not play as important
a role in this inquiry. I will not attempt to model the knowledge involved in these behaviors.
However, I want to do one brief example here in order to help clarify what I intend to exclude. In
this episode, Jon and Ella were working on a very standard problem:
Suppose a pitcher throws a baseball straight up at 100 mph. Ignoring air resistance, how high does it
go?
Jon and Ella’s path to the solution of this problem was very short. They began by writing, from
memory, an equation:
v 2f = vo2 + 2ax
This equation applies to cases in which an object moves with constant acceleration, either speeding
up or slowing down. It happens that gravity has (approximately) the effect of causing objects to
decelerate at a constant rate as they travel up into the air, so this equation is appropriate for the
task. Here, vf and vo are the final and initial velocity of the object, a is the acceleration, and x is the
distance traveled.
After the ball is thrown, it steadily decelerates at the fixed rate determined by gravity until, at
the peak, it stops momentarily. Thus, the ball’s maximum height can be found by substituting a
value of zero for vf and then solving for x. This is precisely what Jon and Ella did, and they wrote
51
vo2
=x
2a
When they substituted numerical values for vo and a, Jon and Ella obtained a result of 336ft for x.
Then they began to wonder if this is a reasonable result:
Ella
Yeah, okay. I guess that’s pretty reasonable.
Jon
Well that’s like a - a football field. [g. =336ft.]
Ella
Yeah. How far,,
Jon
A hundred miles an hour is,, (1.0) If you were in a car going a hundred miles an hour,
Ella
[laughs] straight up into the air.
Jon
straight up into the air, would you go up a football field? (0.5) I have no idea. Well it doesn’t
sound - it’s on the right order of magnitude maybe.
Ella
Yeah. Okay.
(5.0)
Jon
I mean, it’s certainly not gonna go (0.5) thirty six feet. That doesn’t seem far enough.
Cause a hundred miles an hour is pretty fast.
Ella
Hm-mm.
Jon
It’s not gonna go three thousand feet.
So, Jon and Ella concluded that 336 feet was a reasonable answer, or at least the right “order
of magnitude.” This method of checking the sensibility of a result is quite different than Jim’s
interpretation of an equation. Here, Jon and Ella did not look at the nature of the final symbolic
expression, rather they examined the numerical result that this expression produces. Once they
obtained the numerical result, they were able to draw on a wide variety of knowledge to get a feel
for whether this result was correct. The ball is thrown at roughly the speed that a car travels and it
goes about as high as a football field, does this make sense? Well, they argued, it is a little hard to
tell but “it’s on the right order of magnitude.”
Students have a selection of alternative routes available for checking the sensibility of their
work. As in the first two episodes I presented, they can see meaning “directly” in equations.
However, they also have some less direct methods available for checking whether their work
makes sense. For example, like Jon and Ella, they can consider the reasonableness of numerical
results. Or they can look at their expressions and see if they have the “right units.” As I
mentioned, these alternative routes will be less important in the work that follows and will not be a
central focus of my modeling efforts.
52
Steps Toward a Model
The above episodes, especially the first two, are intended to provide a first introduction to the
type of phenomena that I want to understand, as well as to show that they exist. In the first
example, Mike and Karl composed their own novel expression, starting from an understanding of
what they wanted from that expression. The observation that students have this capability has
implications beyond contexts where students have to invent their own equations. If students can
build equations as meaningful structures, why can’t they see existing equations, composed by
others, in terms of the same meaningful structures? The answer is that they can, and this is
precisely what was illustrated in the second episode. Jim told us what an equation said.
Furthermore, the second example provided a glimpse into what it means to “interpret” an
equation.
The task now is to build a model of symbol use in physics, in the spirit of cognitive science,
that incorporates these important phenomena. Before proceeding, it is worth a moment to
comment on what is involved in “the spirit of cognitive science.” One way that cognitive science
proceeds is by making models that are designed to make predictions and provide accounts of
human activity—what people say, do, and think. Furthermore, these models are of a particular
sort, they generally posit the existence of internal mental structures, including entities called
“knowledge.”
In spite of the fact that I will be working within this tradition, I will not develop a model that
is detailed enough that it can be “run” to generate predictions. I will posit the existence of
knowledge elements that are to be understood as elements of such a model. But to actually “run”
the model, in any sense, we would have to fill in many missing details with common sense and
intuition. This is required because the creation of anything approaching a complete model would
be very difficult given the present early stage of this research.
Note that this boils down to self-consciously applying a different tactic than that taken by the
existing physics problem solving literature. In that literature, certain phenomena are disregarded
and simplified situations are studied so that a runnable model is possible. Such an approach has
great merits, but simply will not do for my needs. We want to press on what “meaningful”
symbol use is, which precludes our ignoring phenomena like interpretation. This does not mean
that I am not throwing things out for the sake of simplicity. But preserving the phenomena that I
am interested in, even in a simplified form, is enough to make the creation of a runnable model
very difficult. As we will see, the interpretation of equations is a complex and idiosyncratic
53
business, sensitively dependent on details of context, and hence is not amenable to description by
any simple models.
Let me pin down a little further exactly what phenomena I propose to model and where my
attention will be focused. First, my model of symbol use in physics should primarily be
understood as a model that applies in the very moments of symbol use, when an individual is
actually working with an expression. The model will address questions such as: How does a
student know to write “+” instead of “×”? And: What kinds of things does a student say when
they have an equation in front of them? Interpersonal and developmental (learning-related)
phenomena will be treated as secondary, although I will do some tentative extending of the model
in each of these directions.
Second, my attention will be focused on only one particular slice of the knowledge associated
with symbol use in physics; prototypically, this is knowledge that is particularly evident when
novel expressions are constructed and expressions are interpreted. I will not worry about, for
example, the knowledge needed to correctly perform syntactic manipulations of expressions, nor
will I worry about how students know which equations to pull out of memory for a given task.7
This focus is justified because of the primary concern of this work, which is to understand the
relation between understanding and symbol use. The presumption is that, to understand that
relation, our effort is best focused at the places where symbols and understanding, in a sense,
come together.
First step: Registrations
Now we are finally ready to take some first steps in the direction of a model of symbol use in
physics. Let us begin by imagining a specific situation, a situation that is prototypical of our
concerns. A student is looking at a sheet of paper that has, written on it, a single equation. The
student has some particular interest in this equation, perhaps wondering if it’s written correctly,
perhaps wondering what its implications are. To take a first step toward constructing our model,
we can begin with a fundamental question: What does the student “see” when she looks at this
equation? On what features of the page of expressions does her behavior depend?
7
Chi et al. (1981) present evidence that students and experts can directly recognize problems as being of
a specific type and then equations are associated with particular problem types. This is one model of
how students draw on known equations from memory.
54
This may seem like a simple or silly question, but it is not altogether trivial. To illustrate, think
for a moment of the experience you have when looking at a word processing document on a
computer display. The text that you see is actually, at its lowest level, made up of a number of
small dots, pixels. Does this mean that you are seeing pixels when you look at the document on a
computer display? In a sense, yes. But, in an important sense, you also “see” letters. Notice that
there is no effort involved in going from the pixels to the letters, you just see the letters directly.
In fact, if you are reading a printed version of this document, the letters are also probably formed
from dots, but you are certainly not conscious of those dots as you read. And there are other
things that you “see” when you look at a text document. You can see words, groups of words,
paragraphs, and perhaps arbitrary groupings of text. You can even see the entire block of text on a
page as a unit.
Similar points have been made in a variety of literature with diverse concerns. It is a wellknown result that an account of perception which maintains that we simply perceive a neutral array
of colors is inadequate. Rather we see a world of objects and edges, a partly processed world of
things. I am not going to attempt at all a survey of the literature that has made related points.
However, I will borrow a term from Roschelle (1991). Roschelle uses the term “registration” to
describe the way people “carve up their sensory experience into parts, give labels to parts, and
assign those labeled parts significance.” This provides us with a way to rephrase the question we are
currently concerned with: How does our hypothetical student “carve up” her experience of the
equation written on the sheet of paper? Here is a preliminary list of some of the more important
registrations with which I will be concerned:
1. Individual symbols. If our hypothetical student is looking at a sheet of paper where the equation
v = vo + at is written, she may see the letter ‘v’ as an individual symbol. This is in contrast to
seeing some dots, or two straight lines that meet at a point. The student could also see the
letter ‘t’ or a ‘+’.
2. Numbers. Students can also see numbers. Jon and Ella saw the number “336” that they had
written on the blackboard. Note that this number can also be seen as comprised of three
individual symbols.
3. Whole equations. Our hypothetical student can see the equation v = vo + at as a single unit.
55
4. Groups of symbols. Arbitrary groups of symbols can also be seen as units. For instance, if I write
an expression for the total momentum of two objects as, P = m1v1 + m2v2, then m1v1 can be
seen as the momentum of one of the objects. Terms in expressions and symbols grouped in
parentheses are frequently seen as units, but arbitrary and complicated groupings are not
uncommon.
5. Symbol Patterns. Finally, part or all of an expression may be seen in terms of a number of
“symbol patterns”—patterns in the arrangement of symbols. These patterns will be discussed
extensively in the next chapter, along with many examples, so I will just give a brief
explanation here. Roughly, symbol patterns can be understood as being templates in terms of
which expressions are understood and written. In general, these are relatively simple templates.
For example, two symbol patterns that students see in expressions are:
=
+
+
…
The symbol pattern on the left is a template for an equation in which two expressions are set
equal. These two expressions can be of any sort—the
can be filled in with any expression.
Similarly, the symbol pattern on the right is a template for an expression in which a string of
terms are separated by plus signs. Note that symbol patterns differ from what I have called
groups of symbols because, in the case of groups, the arrangement of symbols within the group
is unimportant; the group is simply treated as a whole. For patterns, the arrangement of
symbols constitutes a recognizable configuration. Symbol patterns will play a key role in my
explanation of how students interpret expressions and compose novel expressions.
So, with this discussion of registrations, we have taken a first step into the physicist’s world of
symbols. What does a physicist or physics student see when looking at a sheet of equations? They
do not see an organic array of curving lines and marks; they see symbols, equations, terms, and
groups of various sorts. Furthermore, we will see that, for a physicist, a sheet of physics equations
is even more structured and meaningful than this first account implies. In the next chapter, we
dive hard into this world of structure and meaning.
56
Chapter 3. A Vocabulary of Forms
Rabbit or duck?
Art and Illusion, the well-known book by E. H. Gombrich concerning art and pictorial
representation, begins with a discussion of what Gombrich calls “the riddle of style” (Gombrich,
1960). For Gombrich, the riddle is this: “Why is it that different ages and different nations have
represented the visible world in such different ways? Will the paintings we accept as true to life
look as unconvincing to future generations as Egyptian paintings look to us?” Gombrich rejects the
view that the history of art can simply be understood as a development toward more accurate
representations of reality. However, Gombrich balances this rejection with an acceptance of a
certain kind of historical development. He believes that, over time, artists have developed an
increasingly varied and flexible “vocabulary of forms.” “But we have seen that in all styles the artist
has to rely on a vocabulary of forms and that it is the knowledge of this vocabulary rather than a
knowledge of things that distinguishes the skilled from the unskilled artist.” Every artist and every
person that looks at art learns the current vocabulary by looking at existing paintings. Then, new
paintings are created from this vocabulary and other paintings are seen in terms of the vocabulary.
The point I want to extract here is that the forms that we see when looking at a painting are
not somehow inherent in the painting, or even in the painting and the “raw” perceptual capabilities
of humans. People learn—and we, as individuals, learn—to see the forms, and then we bring this
capability with us when we look at new drawings and paintings.
This point is especially evident in some more marginal cases. Consider the simple drawing at
the top of this page, which I have taken from Gombrich. Is it a smiling duck? Is it a cute, but wary
rabbit? Whatever it is, this drawing consists of a very simple arrangement of lines. But when we
look at it we see more than those lines; to us, the image is much more than a simple sum of its
parts. If it was a simple sum of its parts, how could it possibly be both a rabbit and a duck?
Because we have learned to see certain forms, we see a meaningful image. In a sense, we bring the
rabbit and duck with us.
57
I can now add a basic point to the discussion of registrations I presented at the end of the last
chapter. The registrations available in any page of expressions is not only a function of the marks
on the paper, it also depends on the (learned) capabilities of the person looking at the display. To
see letters, for example, you have to know the alphabet. This will be my focus in this chapter: I will
describe the knowledge that people acquire that is associated with certain registrations in symbolic
expressions.
What is a Symbolic Form?
I am now ready to define the first major theoretical construct in my model of symbol use in
physics, what I will refer to as “symbolic forms” or just as “forms,” for short. A symbolic form is a
particular type of knowledge element that plays a central role in the composition and
interpretation of physics expressions. Each form has two components:
1. Symbol Pattern. Each form is associated with a specific symbol pattern. Recall that symbol
patterns are one of the types of registrations I described in the previous chapter. A symbol
pattern is akin to a template for an expression and symbolic expressions are composed out of
and seen through these templates. To be more precise, forms are knowledge elements that
include the ability to see a particular symbolic pattern.
2. Schema. Each symbolic form also includes a conceptual schema. The particular schemata
associated with forms turn out to be relatively simple structures, involving only a few entities
and a small number of simple relations among these entities. These schemata are similar to
diSessa’s (1993) “p-prims” and Johnson’s (1987) “image schemata,” which are also presumed
to have relatively simple structures. For example, as I discussed earlier, diSessa describes a
particular p-prim known as “Ohm’s p-prim” in which an agent works against some resistance
to produce a result. This p-prim can be understood, in part, as a simple schematization
involving just three entities, in terms of which physical situations are understood. Although I
will discuss both p-prims and image schemata in later chapters, the view that I will describe
shares many more of the commitments of diSessa’s theory.
One other point about forms is worthy of note at this very early stage. The schema component
of forms can allow specific inferences. I will understand this as working as follows: Generic
58
capacities (that I will not describe) act on the schemata to produce inferences. I will say more
about this in a moment.
To illustrate forms, I want to return to the very first example that I presented from my data
corpus. In Chapter 2, I described an episode in which Mike and Karl constructed their own
expression for the coefficient of friction:
µ
µ = µ 1 + C m2
When presenting this episode, I commented that this was an equation that you would not find in
any textbook; Mike and Karl had composed this equation on their own. Now we are ready to fill
in the first pieces of the puzzle in accounting for how they accomplished this feat.
Recall that the core of Mike and Karl’s specification for this equation was the notion that the
coefficient of friction should consist of two parts, one that is constant and one that varies inversely
with the weight:
Karl
=the coefficient of friction has two components. One that's a constant and one that varies
inversely as the weight.
I want to argue that two forms are evident in this specification, the first of which I call PARTS -OF -AWHOLE.
Like all forms, PARTS -OF -A-WHOLE consists of a symbol pattern and a schema. The symbol
pattern for PARTS -OF -A-WHOLE is two or more terms separated by plus (+) signs. And the schema
has entities of two types: There is a whole which is composed of two or more parts. This schema
can be seen behind Karl’s above statement; he says that the coefficient of friction consists of “two
components.” In other words, the coefficient of friction is a whole that has two parts.
Furthermore, the equation has the structure specified in the symbol pattern; it has two terms
separated by a plus sign.
To describe the various symbol patterns that we will encounter, I will employ a shorthand
notation. For PARTS -OF -A-WHOLE I write the symbol pattern as:
PARTS -OF -A-WHOLE
[
+
+…]
Here the ‘ ’ character refers to a term or group of symbols, typically a single symbol or a product
of two or more factors. The brackets around the whole pattern indicate that the entity
corresponding to the whole pattern is an element of the schema. The shorthand notation for
symbol patterns is briefly summarized in Appendix D.
59
The second form in Karl’s statement is what I call PROPORTIONALITY MINUS or just PROP-. The
schema for PROP- involves two entities: One quantity is seen as inversely proportional to a second
quantity. In Karl’s specification for this expression, it is the second component of the coefficient of
friction that is inversely proportional to the weight. As he says, there’s one component that’s
constant and one that “varies inversely as the weight.”
The symbol pattern for PROP- is relatively simple. When a student sees PROP- in an expression,
all they are attending to is the fact that some specific symbol appears in the denominator of an
expression. I write this as:
 … 
…x… 
PROP-
Here, x will usually correspond to a single symbol. In Mike and Karl’s expression, the m is the
symbol of interest and, in line with the PROP- form, it appears in the denominator of the second
term.
I want to digress for a moment here to make a couple of points about the forms we have seen
thus far. First, it may seem as though the list I am building will simply end up with a form
corresponding to each syntactic feature of algebraic notation. But this is not the case. We will see
that there are limits to what features people are sensitive to, and that identical patterns can
correspond to more than one form. Note, for example, that
PROP-
constitutes a very specific
hypothesis concerning what features are consequential for students during equation construction
and interpretation. All that is relevant for PROP- is that the symbol in question appears in the
denominator of the expression; it does not matter whether the quantity is raised to a power or
whether there are other factors surrounding it. With regard to this form, such details are not
consequential; both 1/m and 1/m2 can be seen as PROP-.
Second, I want to say a little about the inferences permitted by these two forms, just to give
the flavor of what I have in mind. Here is a sampling of inferences permitted by
PARTS -OF -A-
WHOLE:
•
If one part changes and the other parts are constant, the whole will change.
•
If one part increases and the other parts are constant, the whole will increase.
•
If one part decreases and the other parts are constant, the whole will decrease.
•
If the whole is necessarily constant and one part increases, then another part must decrease.
60
And here is a sampling of inferences permitted by
PROP-:
•
If the denominator varies, the whole varies.
•
If the denominator increases, the whole decreases.
•
If the denominator decreases, the whole increases.
Some of these inferences may be playing an implicit role in Mike and Karl’s construction of their
equation for the coefficient of friction. Recall that Mike and Karl began with the requirement that
µ must decrease with increasing weight:
Karl
I guess what we're saying is that the larger the weight, the less the coefficient of friction
would be.
Inferences of the above sort could confirm that this requirement is met. If the weight is increased,
PROP-
tells us that the second term in Mike and Karl’s expression for µ must decrease. Then PARTS -
OF -A-WHOLE
permits the inference that if the second part is decreased the whole also decreases,
since the first term is constant. Thus, the whole coefficient of friction decreases if the weight is
increased.
Two additional forms also played a role in Mike and Karl’s building of their novel equation
and I will just comment on them briefly here. The first of these forms has to do with the ‘C’ in
front of the second term. This factor was inserted by Mike and Karl almost as an afterthought, at
the very end of the episode. I take this to be a case in which the “COEFFICIENT” form is in play:
COEFFICIENT
[x
]
In COEFFICIENT, a single symbol or number, usually written on the left, multiplies a group of
factors to produce a result. A number of features are typical of the coefficient part of the schema,
and I will explain these features later in this chapter. For now, I just want to comment that the
coefficient is frequently seen as controlling the size of an effect. This, roughly, was Mike and
Karl’s reason for inserting the coefficient. As Karl wrote, he commented:
Karl
You find out that uh you know the effect is small, or you find out the effect is large or
whatever so you might have to screw around with it.
So, for Mike and Karl, the factor of C was a way for them to tune their expression.
Finally, there is one important part of Mike and Karl’s expression that I have yet to comment
on; I have not said anything about the equal sign and the µ that appears on the left side of the
61
equation. This constitutes an application of one of the simplest and most ubiquitous forms, the
IDENTITY
form.
x = […]
IDENTITY
In the IDENTITY form, a single symbol—usually written on the left—is separated from an arbitrary
expression by an equal sign. I will have more to say about IDENTITY later in this chapter; at this
point, however, I just want to note that identity transparently allows the inference that whatever is
true of the entity corresponding to the whole right hand side is also true of the entity on the left.
In the case of Mike and Karl’s equation for the coefficient of friction, since the whole expression
on the right decreases with increasing mass, µ can also be presumed to decrease with increasing
mass.
In the previous chapter I asked: How did these students know to write a ‘+’ instead of a ‘×’
between the two terms? And: How did they know to put the m in the denominator? Now I have
given the answers to these questions; or, at least, I have given a particular variety of explanation. I
have said that, to compose a new equation, these students had to essentially develop a specification
for that equation in terms of symbolic forms. Once this was done, the symbol patterns that are
part of these forms specified, at a certain level of detail, how the expression was to be written.
In addition to being knowledge for the construction of the expression, the forms are also the
way that the equation is seen and understood by Mike and Karl. Just as we can see a rabbit or a
duck in the drawing at the head of this chapter, Mike and Karl see
IDENTITY
PARTS -OF -A-WHOLE
and
when they look at their expression for µ. For Mike and Karl, their equation is not just a
collection of marks, or even a collection of unrelated symbols, it has a meaningful structure.
Forms and a More Traditional Problem
As I have mentioned, the episode discussed in the previous section is somewhat atypical. When
solving problems in an introductory physics course, it is rare that students must build elaborate
expressions from scratch. In this section, I want to turn to a more typical episode, involving more
familiar expressions. In addition, this episode is interesting because I believe it may provoke
skeptical readers to argue that knowledge other than forms is responsible for the students’ ability
to write the expressions that appear. Thus, I will need to do some work to argue that forms are,
nonetheless, a central part of the knowledge involved here.
62
In this example, I am again going to look at the work of Mike and Karl. The reason for
sticking with these students is that I want there to be no doubt that the students we see doing the
more typical work described here have the abilities demonstrated in the above section. In this
episode, Mike and Karl were working together at a blackboard to answer the following question:
For this problem, imagine that two objects are dropped from a great height. These two objects are
identical in size and shape, but one object has twice the mass of the other object. Because of air
resistance, both objects eventually reach terminal velocity.
(a) Compare the terminal velocities of the two objects. Are their terminal velocities the same? Is the
terminal velocity of one object twice as large as the terminal velocity of the other? (Hint: Keep in mind
that a steel ball falls more quickly than an identically shaped paper ball in the presence of air
resistance.)
This problem asks about a situation in which a ball has been dropped under the influences of
gravity and air resistance. Gravity is pulling the ball down and, as the ball rushes through the air,
the resistance of the air tends to oppose the downward fall. When the ball is first dropped, it
initially accelerates for a while—it goes faster and faster. But eventually, because of the opposition
of air resistance, the speed levels off at a constant value. This maximum speed that the ball reaches
is called the terminal velocity. The question that this problems asks is: How does the terminal
velocity differ for two objects, one of which is twice as large as the other?
(b)
(a)
(c)
Fair
Fair
Fgravity
(d)
Fgravity
Fgravity
Fair
Fgravity
Figure Chapter 3. -1. (a) An object is dropped. (b) & (c) As it speeds up, the opposing force of air resistance
increases. (d) At terminal velocity, the forces of air resistance and gravity are equal.
One way to understand what’s going on here is that there are two forces acting on the ball, a
force down from gravity and a force from air resistance acting upward. Initially the ball is not
moving and the force from air resistance is zero. But gradually the ball speeds up and the force
from air resistance grows until, at terminal velocity, it’s exactly equal to the force from gravity
(refer to Figure Chapter 3. -1).
Now let’s look at what Mike and Karl had to say about this problem. They began by
commenting that the hint given in the problem “basically answers the question,” since it gives
away the fact that the terminal velocity of a heavier object is higher. The discussion below
63
followed, in which, Mike and Karl agreed that what they called the “upward” and “downward
accelerations” of a dropped object must be equal when it reaches terminal velocity:
Karl
So, we have to figure out (0.5) what, how do you figure out what the // terminal velocity is.
Mike
// terminal velocity is.
Well, okay, so, you have a terminal velocity, You have terminal velocity, that means it
falls at a steady rate. // Right, which means the force opposing it
// Right, which means
Karl
Karl
It means it has an upward acceleration that's equal to the downward acceleration.
Mike
Ri:::ght.
Following this discussion, Mike drew the diagram shown in Figure Chapter 3. -2, talking as he
drew.
f(v)=
Fa
v m
Fg
Figure Chapter 3. -2. Mike and Karl’s air resistance diagram.
Mike
Air resistance. You have a ball going down with a certain velocity. A for:::ce,, You reach
terminal velocity when air resistance, the force from the air resistance is the same as the
force from gravity. [g. points to each of these arrows as he writes, with spread thumb and
forefinger indicating size] But, we also know that force of air resistance is a function of
velocity. [w. f(v)=] Right?
In this last comment, Mike reiterates the notion that the influences of air and gravity must be
equal at terminal velocity, but here he talks in terms of opposing forces, rather than opposing
accelerations. Notice that Mike’s drawing is in agreement with the account I presented in Figure
Chapter 3. -1.
Finally, in the next passage, Mike and Karl compose an expression for air resistance:
a(t ) = −g +
f (v)
m
Mike
So, at least we can agree on this and we can start our problem from this scenario. [g.
Indicates the diagram with a sweeping gesture] Right? Okay? So, at any time,, At any
time, the acceleration due to gravity is G, and the acceleration due to the resistance force
is F of V over M. [w. g + f(v)/m] This is mass M. [w. m next to the mass on the diagram]
Karl
Ah:::.
Mike
Right.
Karl
Yeah.
Mike
Okay, // now they're opposing so it's a minus.
Karl
// (So) as your mass goes up,, [g. m in equation]
64
Mike
So, this is negative G. [g. down, makes a negative sign in front of the g; w. arrow up
labeled "+" near the top of the diagram] Positive direction. You have negative G plus F of V
over M [gestures down and then up]. That's you're acceleration at any time. Right?
Karl
Well, wait. You want them to have,,
Mike
This is the acceleration at any time T. [w. a(t) next to the expression]
In this last passage, Mike and Karl’s conceptualization of the situation in terms of two competing
influences takes shape as an equation. The construction involves a few forms:
OPPOSITION,
COMPETING TERMS,
and IDENTITY.
COMPETING TERMS
±
OPPOSITION
–
x = […]
IDENTITY
We encountered the
IDENTITY
±…
form in the previous episode. COMPETING TERMS and OPPOSITION
are new, so I will explain them briefly:
COMPETING TERM s. Each of the competing influences corresponds to a term in the equation. As
I will discuss later, this is one of the most common ways that physicists see equations, as terms
competing to have their way.
OPPOSITIO n. The two competing influences are competing in a specific manner, they are
opposing. Opposition is frequently associated with the use of a negative sign. “Now they’re
opposing so it’s minus.”
A Fundamental Contrast: Forms versus Principle-Based Schemata
The basic presumption underlying my account of Mike and Karl’s above work is that they
conceptualized the situation as involving competing influences and these competing influences get
associated with terms in equations. But notice that there are aspects of the equation that this does
not explain. In particular, I have not said anything about how they knew to write f(v)/m for the
term corresponding to air resistance. Likely this has to do with the equation F=ma, an equation
that Mike and Karl certainly remember. They stated that the force of air resistance depends on
velocity, which led them to write this force as f(v). Then, since they were talking about opposing
accelerations, they used F=ma to write the acceleration as a=F/m.
So, at the least, forms cannot be the whole story in this construction episode—and they will
not be the whole story in most construction episodes. More formal considerations and
remembered equations must also play a role. In my rendition of the above episode, a remembered
65
equation was used to fill in one of the slots in a symbol pattern. However, the above episode is
open to some interpretations that are significantly different than the one I provided. Most
notably, one might argue that Mike and Karl have memorized the equation:
∑F=ma
Here, the ‘∑’ symbol means ‘summation.’ This equation says that the sum of all the forces on an
object is equal to the mass of the object multiplied by its acceleration. Thus, it may be possible to
see Mike and Karl’s work as solely a straightforward application of this equation; they write the
sum of the forces and then divide by the mass to get the acceleration.
This alternative viewpoint is an important enough (and strong enough) alternative that I want
to devote significant effort to discussing it and contrasting it with the view I am developing in this
thesis. In addition, working through this alternative will help to clarify what is particular about my
own viewpoint. But, I will not try to argue that this alternative view is incorrect, or that it is even
strictly competitive with the one I am developing in this thesis. In fact, I believe that a more
complete description of the knowledge involved in expert symbol use would include knowledge
of multiple sorts, including knowledge more closely tied to formally recognized principles, as well
as symbolic forms. However, I will try to argue that some student behaviors are better explained
by symbolic forms and that symbolic forms—and the related behavior—persist into expertise,
complementing more formal aspects of knowledge.
A version of this alternative view was worked out by Larkin (1983). Larkin begins by
distinguishing what she calls the “naive” and “physical” representations. (Here, when Larkin says
“representation” she is talking about internal mental representations, not external representations.)
In Larkin’s scheme, everyone—novice AND expert—constructs a naive representation as a first
step toward analyzing a physical situation. A naive representation involves only “familiar” entities
and objects; it essentially uses our everyday vocabulary for describing the world. In contrast, the
physical representation, which is only constructed by physicists, “contains fictitious entities such as
forces and energies.” Once the physical representation is constructed, the writing of an equation is
apparently relatively easy. Larkin tells us that: “Qualitative relations between these entities can be
‘seen’ directly in this physical representation, and these qualitative relations then guide the
application of quantitative principles.”
To this point, Larkin’s view may be said to share many broad similarities with my own. There
is a qualitative mental representation that guides the construction of an equation. But Larkin soon
66
takes a step that strongly differentiates her view. She says: “The physical representation of a
problem is closely tied to the instantiation of the quantitative physics principles describing the
problem.” In contrast, in my own view, forms are not closely tied to specific physical principles,
such as Newton’s Laws.
Let’s take a moment to see how Larkin’s model works. According to Larkin, experts possess a
number of schemata that are directly associated with physical principles. These schemata contain
what she calls “construction rules” and “extension rules.” The construction rules act on the naive
representation to produce the physical representation. And the extension rules act on an existing
physical representation to add to that representation.
Larkin provides two examples of these schemata, the “Forces Schema” and the “Work-Energy
Schema.” These schemata are quite closely related to physical principles, as these principles would
be presented in a physics textbook. For example, Larkin says that the Forces Schema “corresponds
to the physical principle that the total force on a system (along a particular direction) is equal to
the system’s mass times its acceleration (along that direction).” This is essentially Newton’s
Second Law, F=ma. The construction rules in this schema correspond to “force laws,” which are
the laws that allow a physicist to compute the forces on an object given the arrangement of objects
in a physical system. So the idea is this: A physicist looks at a physical system and sees the
arrangement of objects involved. The Forces Schema then provides force laws that allows the
physicist to determine the forces (a particular kind of fictitious entity) acting on the objects. Then
the physicist plugs these forces into the equation ∑F=ma and solves for whatever is desired. Note
that this equation, ∑F=ma, is a template of a sort, and it allows the construction of equations that
are not literally identical to those seen before. But, it is has particular limits on its range of
applicability: the entities in the sum must be forces and they must be set equal to ma.
Of course, as I have already suggested, it may be possible to apply Larkin’s account to describe
Mike and Karl’s work on the air resistance task. This would go as follows: Using the construction
rules provided by the activated Forces Schema, Mike and Karl write the force on each object; -mg
is the force of gravity, which is negative because it acts in the downward direction, and the force
of air resistance is simply taken to be f(v). These forces are then summed and set equal to ma:
ma = −mg + f (v )
(Presumable, Mike and Karl do this step mentally, since it doesn’t appear as one of the equations
they write.) Finally, this equation is divided by the mass, m, to produce an equation for a.
67
I must emphasize that one crucial difference between Larkin’s view and my own is the fact
that her schemata are closely tied to physical principles. (This is in addition to some more global
differences in orientation.) For her, it is crucial that the influences in competition are forces, since
Newton’s laws only talk about forces. And the above equation is only written because the schema
tells us to—it’s a simple matter of “plugging-in.” In contrast, in my view, the situation is
generically schematized as competing influences, a notion that is supposed to cut across physical
laws. Many types of entities, other than forces, may compete.
For this reason, I will refer to Larkin’s account as the “principle-based schemata” view. The
modifier “principle-based” is important here since, as we have seen, a type of schema also appears
in my own model. However, in this section, I will at times simply refer to Larkin’s view as “the
schema view.”
Now that I have presented the principle-based schemata view and contrasted it with my own
view (the “forms” view), I want to take steps toward arguing for my own account of some
episodes. In the remainder of this section I will discuss a few episodes that are specifically selected
to contrast these two views. In addition, in later sections and later chapters I will argue for the
usefulness of my viewpoint on other grounds.
Step 1: Students do, at least sometimes, see symbolic forms
The first step I want to take is to point out that I have already presented enough evidence to
show that the schema view cannot be the whole story. In particular, there is no way for Larkin’s
schema view to explain Mike and Karl’s ability to construct their own equation for the coefficient
of friction.
µ
µ = µ 1 + C m2
This is because there are simply no principles or existing equations to guide this construction.
Furthermore, once we have accepted that Mike and Karl have abilities that transcend what is
allowed by the principle-based schemata model, we have to live with this observation in all
accounts that we give of all types of problem solving. Both of the above two episodes involved the
same pair of students, Mike and Karl. And, just because they were working on a more familiar
task, there is no reason to believe that their form-related capabilities will simply go away or not be
employed. Suppose, for a moment, that something like Larkin’s Forces Schema is responsible for
68
Mike and Karl’s work on the air resistance task. Even if this is the case, because Mike and Karl have
the ability to see meaningful patterns—forms—they may, at any time, be seeing these patterns in
equations that they write.
I want to illustrate this point with an analogy. Suppose that we decided to teach some
mathematicians a little physics, so we teach them to write equations and solve problems.
Mathematicians already know a lot about equations. They know a linear equation when they see
one, and they can differentiate a linear equation from a quadratic equation. Thus, even if they
learn to write the expression v = vo + at from memory, they may still always see this equation as a
linear relation between the velocity and the time. The equation has a structure that a
mathematician can recognize.
I am trying to make an analogous point about forms. The coefficient of friction episode tells
us that Mike and Karl can sometimes compose expressions from symbolic forms. They may thus
see these forms in any equation that they write. If this is true, then the principle-based schemata
view cannot be “the whole story.”
Step 2: Deviations from the ÒschemaÓ model in the air resistance example
Mike and Karl’s work on the air resistance task warrants a closer look. Recall that they
bounced between describing the situation in terms of competing accelerations and in terms of
competing forces. This must, to some degree, be because Mike and Karl are aware that force and
acceleration are formally interchangeable via F=ma. However, I also take this as evidence that a
view of the situation in terms of opposing influences is the core driving element of these students’
work; thus, the specific nature of these influences is not of utmost importance to them at all stages
of the solution attempt. This observation really cuts to the heart of the difference between the
forms model and the principle-based schemata model. I am arguing that Mike and Karl’s work is
driven—at least, in part—by an understanding tied to a notion of opposing influences that cuts
across principles, not by a schema tied solely to forces and Newton’s laws.
For those who are less familiar with traditional physics practice, it is worth noting that Karl’s
use of the terms “upward” and “downward acceleration” would probably be considered
inappropriate by many physicists. To a physicist, the term “acceleration” is usually reserved for
describing actual changes in the speed of objects, not for virtual changes that may be canceled out
by other virtual changes. Thus, Mike and Karl’s work departs somewhat from a strict and careful
69
application of traditional methods that follow a universal schema associated with physical
principles.
Finally, notice that, as they work, Mike and Karl have interpretations readily available. For
example, Mike says that one acceleration is “opposing” the other. The fact that such interpretive
comments are ready-to-hand suggests that these notions are playing a role in these students’ work.
Step 3: Students do not rigorously deduce from principles
As the final step in my argument against the schema view, I want to present a battery of
evidence designed to further refute the notion that students (and experts) work by slavishly
following standard techniques and rigorously deducing from principles, even when these
techniques are available and appropriate. The purpose of this data is not only to argue against
certain aspects of the principle-based schemata view, but also to provide counter-evidence for a
broad range of opposing views and to give a feel for the character of student work. I will begin
with one more example concerning Mike and Karl; here, I look at their work on the “spring” task,
in which a mass hangs motionless at the end of a spring. In Chapter 2, I briefly examined Jack and
Jim’s work on this task.
The incident I want to recount actually took place during Mike and Karl’s final session, which
was in part used to wrap-up some issues and debrief the subjects. During this session, Mike and
Karl had returned to the spring task, and they had written the equation:
Fs = Fg
This equation says that, when the mass hangs motionless, the force upward due to the spring must
equal the force of gravity pulling down. Since this was a wrap-up session, I probed Mike and Karl
to see how, and in what way, they related this equation to known principles. I began by asking if
they could somehow include the equation F=ma in their solution. What I had in mind was that
they would say that, because they know that the sum of the forces is equal to ma, and the mass
isn’t accelerating (a = 0), the sum of the forces must be zero: Fs + Fg = 0 . Furthermore, if we are
willing to adjust some sign conventions, then this expression can be rewritten as Fs = Fg .
Immediately following my initial probe, Mike and Karl complained that it simply did not
make sense to employ F=ma in the solution of this problem:
Mike
It's not very natural.
Karl
We're looking at a static system.
70
So, I tried another approach. I pressed them as to how they knew that the forces of the spring and
gravity must be equal.
Mike
How do I know F S equals F G? Because it's in equilibrium.
Karl
It's defined that way. [They both laugh]
Mike
Because it's in equilibrium, right, so the forces they have to, um, add up to zero.
Bruce
Why do the forces have to add up to zero?
Mike
To be in equilibrium. [laughs] We're going in circles aren't we. If it's in equilibrium there's no
acceleration, is that what you're looking for?
In the above passage, Mike and Karl first seem to assert that it is essentially obvious that these
forces must be equal. But then, at the end of the passage, it seems like Mike has latched onto what
I had in mind, he says: “If it's in equilibrium there's no acceleration, is that what you're looking
for?” Strikingly, Mike’s further work made it clear that he was still not on my wavelength. He
went on to assert that the acceleration due to each force must be equal and wrote
as = ag
Apparently, he had in mind using F=ma simply to find an acceleration associated with each of the
forces, not to establish the dynamical truth of their equality.
Mike
There’s acceleration but they cancel, right? I mean there is - when there’s force on it and
there’s no - the result - there’s no resulting acceleration. There’s no net acceleration,
maybe that’s a better way to put it. (3.0) Cause one is accelerating one way [g. up] and the
other force is accelerating downward.
So, it seems that I had still not succeeded in getting Mike and Karl to state the argument that
I had in mind. Nonetheless, to this point in the episode, it is still possible that Mike and Karl were
well aware of how their equation followed form F=ma, but that they simply did not know what I
was looking for. Perhaps, for example, the explanation from F=ma is simply too simple to be
worth mentioning. To explore this possibility, I finally just told them what I had in mind, giving
the argument I presented above. Mike and Karl’s reaction to this made it clear that this was an
entirely novel idea to them. Karl was especially astounded and pleased by my little explanation of
how Fs = Fg followed from F=ma and a = 0. In fact, as the session continued and we moved on
to other issues, he occasionally returned to this issue to comment on how much he liked my
explanation:
Karl
I like - I like - this made sense to me because this took away the “well, this is true just
because.”
Bruce
Right.
Karl
And we get a lot of “just becauses.”
Mike
Uh-huh.
71
Karl
This explained why.
What I am hoping that this episode makes very clear is that fundamental principles like F=ma
are not always lurking secretly behind student’s work, simply unsaid. It is true that this episode
cannot quite confirm the kind of generality I have asserted for forms; for instance, it is possible
that Karl has something like a Forces Schema that is specialized for static situations. But this
episode can at least eliminate the most economical version of the principle-based schemata model,
in which all work flows from schemata associated with the most basic fundamental principles.
Of course, even if we accept that Mike and Karl’s work is not driven by principle-based
schemata, one could point out that these students are simply not experts and maintain that, in
expertise, behavior tends toward the description given by the principle-based-schemata view. It is
very likely that experts and more advanced students would be able to quickly produce the
argument that the equality of the two forces in the spring task follows from F=ma. Nonetheless,
we must keep in mind that these students are at least at an intermediate level of expertise. They
have completed two full semesters of physics and they can very reliably solve problems like the
spring task. The fact that this sort of behavior continues into a stage in which there is practiced
and reliable performance suggests that the associated knowledge may linger into expertise.
Adding to the plausibility of this statement is the observation that symbolic forms constitute a
useful type of knowledge. This knowledge is necessary, for example, in the construction of
completely novel expressions, such as Mike and Karl’s expression for µ. Because there is an
important niche for this knowledge in expertise, it is thus more plausible that it will continue to
play a role as students progress toward more expert behavior.
For the sake of completeness, I should mention that I would explain Mike and Karl’s
construction of the relation Fs = Fg by appeal to what I call the BALANCING form.
BALANCING
=
In BALANCING , two competing influences are seen as precisely equal and opposite. Note, once again,
the central point here: The BALANCING form is more general than any assertions about balanced
forces. Any types of influences can be balanced and, once they are seen as balanced, a relation in
line with the BALANCING symbol pattern can be written.
72
Some additional examples
Mike and Karl were quite typical of my subjects; as a whole, the students almost never worked
by a strict following of traditional techniques that follow from principles. Furthermore, when
students did appear to work from principles—or, at least, maintained that they were working
from principles—these applications of principles often seemed to play the role of an after-the-fact
rationalization for work that was already done. And even these rationalizations often involved
some hand-waving. Let’s look briefly at a few examples.
First, I return to Jack and Jim’s work on the spring task. Recall that they began by writing the
force due to the spring as F=-kx and the force due to gravity as F=ma. Then, in a passage that I
did not cite earlier, they proceeded to equate these two forces, writing kx=mg:
Jack
So, okay, so then these two have to be in equilibrium those are the only forces acting on
it. So then somehow I guess, um, (3.0) That negative sign is somewhat arbitrary [g. F=-kx]
depending on how you assign it. Of course, in this case, acceleration is gravity [g. a in
F=ma] Which would be negative so we really don't have to worry about that [g.~ F=-kx] So I
guess we end up with K X is M G. [w. kx=mg]
The delicate issue that Jack is trying to deal with is, if these forces are going to be equated, then
what happens to the negative sign in F=-kx? Jack knows that he should somehow end up with
kx=mg, but he makes a pretense of being careful; he says that the acceleration due to gravity is also
negative so the negative signs cancel. This is not quite up to the standards of a rigorous argument;
such an argument would require that Jack carefully associate the signs of terms and directions on
the diagram. Jack does not need to do this however, because he knows the answer before he starts.
This sort of hand-waving concerning signs is extremely common in my data corpus. If anything, it
is the norm.
Finally, I want to talk about some student work on a task I have yet to mention, what I call the
“Stranded Skater” problem.
Peggy Fleming (a one-time famous figure skater) is stuck on a patch of frictionless ice. Cleverly, she
takes off one of her ice skates and throws it as hard as she can.
(a) Roughly, how far does she travel?
(b) Roughly, how fast does she travel?
What happens in this situation is that, when Peggy throws the skate, she moves off in the direction
opposite to the direction in which the skate was thrown. I am going to talk about the work of two
pairs on this problem, Jack and Jim and Alan and Bob. Jack and Jim began by announcing that
“it's basically a conservation of momentum problem.” Then they immediately went on to write
the expression
73
m pf v1 = ms v2
where m pf is the mass of Peggy Fleming and ms is the mass of the skate. Here, Jack and Jim are
implicitly using the relation P=mv, where P stands for the momentum of an object. The question
is, how did Jack and Jim know to simply equate the momenta of the skater and the skate? I
believe that they see this problem in terms of BALANCING . After the skate is thrown, the movement
of the skater and skate must somehow balance out.
Figure Chapter 3. -3. Alan and Bob’s diagram for the Stranded Skater problem.
After they had written this relation, Jack moved to rationalize it in terms of the principle of
conservation of momentum. He wrote the following expression and said:
m pf v1 + ms v2 = 0
Jack
Actually, it should probably be … [w. MpfV+MsV=0] This case we'll assume that this
velocity would be a negative velocity [g. above v of skate] so that we can go to that. [g.
first equation] Cause one's gonna go one way, one's gonna go the other.
So, in this episode, Jack began by writing a relation that he believed to be true and then moved to,
after the fact, rationalize that relation in terms of known principles. The principle applied here is
the conservation of momentum, which says that the initial and final momenta of a system must be
equal in the absence of any externally applied forces. In this case the initial momentum is zero,
since everything in the system is at rest before the throwing of the skate. Thus, an expression for
the total final momentum can be correctly equated to zero. This is what Jack does in the above
equation.
I want to note that it was common for students to see the Stranded Skater problem in terms
of BALANCING as well as in terms of a related form, what I call CANCELING (0 =
-
Ê
). Alan and
Bob began their work on this problem by writing the expression:
0 = ms vs − m pv p
Alan
And we expect her to move to the left to compensate (her) mass of Peggy times the
velocity of Peggy. [w.mod 0=MsVs–MpVp]
74
Here, Alan argues that Peggy must move in the direction opposite that of the skate in order to
“compensate” for the motion of the skate. The important point here is that the use of a negative
sign between these terms is not derived from a principle or from mechanically following
techniques. Rather, it is introduced based on an understanding of what is going to happen and
what the expression should look like. In my view, Alan sees “competing influences” that cancel
and he makes sure that his expression is in line with this symbolic form.
Final comments on principle-based schemata
In this section I have contrasted my model with a model that is truly different, what I have
called “the principle-based schemata” view. This contrast is extremely important because it gets to
the heart of some of the essential features of my model. Allow me to elaborate briefly. There is a
conventionalized version of physics knowledge that is presented in physics textbooks. But it is an
error simply to presume that, when people learn physics, the knowledge they acquire somehow has
the same structure as the conventionalized knowledge presented in a textbook.8 Physics
knowledge, as realized in individuals, may mirror textbook physics, but it may not; the question
must be settled empirically.
I have attempted to argue that at least one slice of physics knowledge does not have the same
structure as textbook physics. In particular, the knowledge constituted by forms is not organized
like a textbook; it is not tightly related to basic physical principles. Instead, notions like BALANCING
are central. And these notions cut across physical principles—you can have balanced forces OR
balanced momenta.
However, principle-based knowledge must have been playing some role in the episodes I
presented above. For example, even if we accept that some students understand the Stranded
Skater problem in terms of BALANCING , we still have to explain how they knew to balance momenta,
in particular. I believe that this depends, in part, on students ability to recognize problem types
and to know what principles apply to those problem types.9
This observation implies that the knowledge employed in symbol use, in its full and finished
expert form, must certainly have components that are tied to physics principles. In fact, I believe
that there is likely to be knowledge that is somewhat like principle-based schemata. This returns
me again to the point that a full model of symbol use in physics will be quite complicated, more
8
9
Andrea A. diSessa has made this point in a number of articles. See, for example, diSessa, 1993.
See, for example, Chi et al., 1981.
75
complicated than I can deal with here. So, I do not mean to discard the principle-bases schemata
view, but I do hope that I have shown that there are cracks in this model—some pretty big cracks,
in fact—that make room for other important classes of knowledge, and that greatly broaden the
range of student activity that we can expect and explain.
To sum up this section, we have seen that, first, there are episodes that the principle-based
schemata model simply cannot handle. These episodes constitute evidence for the kind of
knowledge that I am calling “symbolic forms.” Furthermore, I have argued that, because this
knowledge exists, it is plausible to presume that it gets employed during more traditional types of
symbol use. In addition, I presented a number of examples to illustrate that students do not
always work by rigorous deduction from principles; instead, they wave their hands and rationalize
to get the answer that they already know they want. This constitutes evidence that forms play a
role in what, at first glance we might take to be the strict following of principle-based tradition.
Forms: A Primer
In this section I am going to discuss all the forms that are necessary to account for the behavior
of students in my data corpus. Because this knowledge system is not organized according to the
principles of physics—and because it apparently has no other simple structure—there are no
shortcuts for describing the breadth of forms, I have to describe each type of form and give
examples. However, for the reader that forges ahead, this section is not without its rewards. There
are many interesting details associated with particular forms, as well as important and subtle
points to be made along the way. In addition, the broad scope provided by this overview is
important. The hope is that it paints a picture of a physicist’s phenomenal world of symbols—how
physicists and physics students experience a page of expressions.
Keep in mind that this list is intended only to account for the particular data in my algebra
corpus. If I had given students a different set of tasks, looked at a different population, and
certainly if I was concerned with a different domain, the symbolic forms in this primer would be
different. Nonetheless, I believe that this list captures much of the breadth as well as the most
common and central forms of introductory algebra-physics. Adding tasks would likely discover
more, but my belief is that these additions would primarily allow us to fill in some less common
forms around the fringes of the form system. I hope to support this viewpoint when I present the
systematic analysis in Chapter 6.
76
In saying that this list of forms “accounts” for my data corpus, I do not mean to imply that
the relation between this list and my data is straightforward. In Chapter 6, I will describe the
procedure whereby I make contact with my data. However, I must warn the reader that, even
there, I will not be presenting an exact procedure for extracting this list of forms from my
transcripts. Instead, the procedure I will describe is highly heuristic, and the resulting categories
depend heavily on theoretical arguments, as well as the data corpus itself.
Although there is no very simple structure to the forms knowledge system, it is possible to
organize forms according to what I call “clusters.” 10 Within a given cluster, the various schemata
tend to have entities of the same or similar ontological type. For example, forms in the
Competing Terms Cluster are primarily concerned with influences. In addition, the forms in a
cluster tend to parse equations at the same level of detail. For example, some forms tend to deal
with equations at the level of terms, while others tend to involve individual symbols. A complete
list of the forms to be discussed here, organized by cluster, is provided in Figure Chapter 3. -4. In
addition, a list of these forms with brief descriptions is given in Appendix D.
10
In using this term I am following diSessa, 1993. I will make close contact with diSessa and his use of
this term in Chapter 5.
77
Competing Terms Cluster
Terms are Amounts Cluster
PARTS -OF -A-WHOLE
[
+
–
BASE ± CHANGE
[
± ∆]
BALANCING
=
WHOLE - PART
[
-
CANCELING
0=
COMPETING TERMS
±
OPPOSITION
± …
SAME AMOUNT
–
Dependence Cluster
DEPENDENCE […x…]
NO DEPENDENCE
SOLE DEPENDENCE
…]
]
=
Coefficient Cluster
COEFFICIENT
[x
]
SCALING
[n
]
[…]
[…x…]
Other
Multiplication Cluster
INTENSIVE •EXTENSIVE x × y
EXTENSIVE •EXTENSIVE
+
IDENTITY
DYING AWAY
x=…
[e− x…]
x×y
Proportionality Cluster
PROP+
…x… 
 … 
PROP-
 … 
…x… 
x
y
 
CANCELING(B) …x… 
…x… 
RATIO
Figure Chapter 3. -4. Symbolic forms by cluster.
The Competing Terms Cluster
To a physicist, a sheet of equations is, perhaps first and foremost, an arrangement of terms
that conflict and support, that oppose and balance. The Competing Terms Cluster contains the
forms related to seeing equations in this manner, as terms associated with influences in
competition. Frequently—though not always—these influences are forces, in the technical sense.
The forms in the Competing Terms Cluster are listed in Table Chapter 3. -1.
COMPETING TERMS
±
OPPOSITION
–
BALANCING
=
CANCELING
0=
± …
–
Table Chapter 3. -1. Forms in the Competing Terms Cluster.
78
We have already encountered a number of forms from this cluster. During Mike and Karl’s
work on the air resistance task, they composed the equation:
a(t ) = −g +
f (v)
m
I argued that their construction of this equation involved the COMPETING TERMS and OPPOSITION
forms. Mike and Karl schematized the situation as involving two opposing influences, sometimes
understood as forces and sometimes as accelerations. Recall that it was the
OPPOSITION
form that
was associated with Mike’s insertion of a minus sign: “now they're opposing so it's a minus.”
Many of the interesting details here, in fact, have to do with how students select signs (+ or –)
for the terms in their Competing Terms expressions. In cases in which these forms are in play,
students frequently draw diagrams and associate the signs of influences with directions in the
diagram. For example, a student may decide that influences acting to the right are positive, and
influences acting to the left are negative.
We also encountered some instances of BALANCING in previous episodes. I maintained that
BALANCING
was responsible for the equating of the forces in Mike and Karl’s work on the spring
task; the force up due to the spring and the force down due to gravity were seen as “in balance.” In
addition, I argued that Jack and Jim were BALANCING the momenta of the skater and the skate in
the Stranded Skater problem when they wrote:
m pf v1 = ms v2
The last form in this cluster, CANCELING, was encountered briefly in Alan and Bob’s work on
the skater problem:
0 = ms vs − m pv p
Alan
And we expect her to move to the left to compensate (her) mass of Peggy times the
velocity of Peggy. [w.mod 0=MsVs–MpVp]
The CANCELING and BALANCING forms are quite closely related. Both of these forms involve equal and
opposite influences. However, they do differ in some important respects. First, with respect to their
symbol patterns, CANCELING is usually written with terms separated by a minus sign and set equal to
zero, rather than as two terms separated by an equal sign. In addition, the associated schemata differ
in a subtle manner. The schema associated with the BALANCING form involves only two entities, the
two influences that are in balance. In contrast, CANCELING can involve three entities; two entities are
79
combined to produce a net result that happens to be zero. Furthermore, the relations between the
entities in CANCELING frequently include a weak type of ordering of the two competing influences.
Sometimes, for example, one of the influences is seen as a response to the other influence. This type
of relationship is evident in Alan’s statement just above. It is as if Peggy’s motion is a response to the
throwing of the skate; she throws the skate and then moves “to compensate.”
BALANCING is an extremely common and important form, and thus merits a little further
discussion here. BALANCING is important because it provides the central dynamical constraint of
many problem solving episodes. For example, in Mike and Karl’s work on the spring problem
described above, it was BALANCING that was really answering the “why” question behind all their
work. Why do we write Fs = Fg ? Because the forces must balance.
The next couple of examples involve student work on a task I call the “Buoyant Cube”
problem. In this problem, as shown in Figure Chapter 3. -5, an ice cube floats at the surface of a
glass of water, and the students need to find how much of the ice cube is below the surface.
An ice cube, with edges of length L, is placed in a large container of water.
(a) How far below the surface does the cube sink? (ρice = .92 g/cm3; ρwater = 1 g/cm3)
L
L
Figure Chapter 3. -5. The Buoyant Cube Problem.
In brief, this problem is typically solved as follows (refer to Figure Chapter 3. -6). Two forces
act on the floating cube, the force downward from gravity, Fg , and a buoyant force upward from
the water, Fb . In order for the cube to be motionless, these two forces must be equal.
Furthermore, the force of gravity is equal to the weight of the cube, which is the product of the
density of ice, the volume of the cube, and the acceleration due to gravity: ρi gVcube . And the
buoyant force is equal to the weight of the water that is displaced by the cube: ρw gV water . If the
amount that the cube sinks below the surface is given the label s, then the volume of water
displaced is equal to L2s. With these relations in place, it is possible to solve for the value of s as
shown in Figure Chapter 3. -6.
80
Fb = ρ wgV water
Fb = Fg
ρ wgV water = ρ i gVcube
( )
( )
ρ wg L2s = ρ i g L3
ρ
s= i L
ρw
s =.92L
S
Fg = ρ igV cube
Figure Chapter 3. -6. A solution to the Buoyant Cube problem.
For almost all students, this problem strongly suggests BALANCING . Frequently they even oblige
us by saying the word “balance:”
Jack
Um, so we know the mass going - the force down is M G and that has to be balanced by the
force of the wa=
Jim
=the displaced water.
What is truly striking, is that students seem to really possess the sense for
BALANCING
at quite a
fundamental level. In Mark and Roger’s work on this task, Roger carried most of the load, with
Mark expressing confusion along the way. After Roger produced a solution, Mark tried to
rationalize Roger’s work for himself. In particular, Mark was a little mystified by Roger’s first
expression, in which he had equated the weight of the ice cube and the weight of the displaced
water.
W w = Wi
Mark ultimately satisfied himself that this equation makes sense with this comment:
Mark
We know one is gonna float. And then just sort of, I guess, I don't know from like all our
physics things with proportionalities and everything, that we'll have to have something
equal, like the buoyancy force should be equal to the weight of the block of ice. [g. pulling
down block]
Here, Mark does not appeal to forces or anything like the notion that the net force must be zero.
Rather he simply states that, if the block is going to float, then something has got to balance: “we’ll
have to have something equal.”
Instead of being an appeal to principle, Mark’s argument is an appeal to experience.
Furthermore, the experiences he appeals to are of a special sort: “…from like all our physics things
with proportionalities and everything, that we'll have to have something equal… .” Mark, like all
other physics students, has a wealth of experience with equations, writing expressions, setting
81
things equal, and manipulating to find results. From this symbolic body of experience, Mark has
developed the sense that setting things equal is a sensible, intrinsically meaningful action, and that
it is typical physics. I will return to this point in later chapters.
The final point I want to make about BALANCING is that its prevalence may actually extend to
overuse; students sometimes employ BALANCING in a manner that could be called inappropriate. To
illustrate, I will return to the Shoved Block problem, which I discussed briefly in Chapter 2. Recall
that this problem involves two objects that are shoved in such a way that they both start moving
with the same initial speed. The question is, given that one block is twice as heavy as the other,
which block travels farther?
As I mentioned earlier, the correct answer is that both blocks travel the same distance. The
derivation of this fact can be quite short. (Refer to Figure Chapter 3. -7). After the shoved block is
released the only force acting on it is the force of friction, which is F f = µmg . Since this is the
only force, it alone is responsible for any acceleration that the block undergoes as it slides along
the table. For this reason, this force can simply be substituted for F in the equation F=ma to give
the equation µmg = ma . Then the m that appears on each side of the equation can be canceled to
produce the final expression µg=a.
Direction of motion
F f = µN = µmg
Friction
∑ F = ma
F f = ma
µmg = ma
µg = a
Figure Chapter 3. -7. A solution to the Shoved Block problem.
The important thing to note about the final result is that this expression for acceleration does not
involve the mass. This means that it simply does not matter what the mass is, the shoved block
always travels the same distance.
We will have reason to return to this problem a number of times, but, for now, I just want to
consider a single interesting episode. Alan and Bob had produced a solution very similar to the
one presented in Figure Chapter 3. -7. In brief, their work looked just right. However, after they
were done, Alan pointed to the equation µmg=ma and added this little addendum to their
explanation:
µmg=ma
Alan
What Bob didn't mention is that this is when the two forces are equal and the object is no
longer moving because the force frictional is equal to the force it received.
82
I believe that this interpretation is clearly inappropriate. When Alan looked at the equation
µmg=ma, he saw two influences in balance, the “force frictional” and the “force it received.” More
specifically, he sees this equation as specifying a condition for an end state, in which the block
comes to rest because these influences balance. However, as I have mentioned, once the block is
released, only one force acts on it, the force due to friction. In other words, only one side of this
equation corresponds to an influence! Thus, Alan is seeing BALANCING in an inappropriate context.
The Proportionality Cluster
When a physicist or physics student looks at an equation, the line that divides the top from the
bottom of a ratio—the numerator from the denominator—is a major landmark. Forms in the
Proportionality Cluster involve the seeing of individual symbols as either above or below this
important landmark. Earlier in this chapter I argued that one of these forms,
PROP-,
was involved
in Mike and Karl’s construction of their novel expression for the coefficient of friction:
µ
µ = µ1 + C 2
m
Karl
The coefficient of friction has two components. One that's a constant and one that varies
inversely as the weight.
The idea was that Karl’s statement specified that the second component of their expression should
be inversely proportional to the mass. Thus, the PROP- form was invoked and the ‘m’ symbol was
written in the denominator of the second term. Table Chapter 3. -2 contains a list of all the forms
in the Proportionality Cluster.
PROP+
PROP-
…x… 
 … 
 … 
…x… 
x
y
 
CANCELING(B) …x… 
…x… 
RATIO
Table Chapter 3. -2. Forms in the Proportionality Cluster
We also encountered some other clear examples of
PROP+
and PROP- in Chapter 2, though
they were not noted at the time. I have in mind the interpretation that Jim gave of the final result
of the spring problem:
83
x=
Jim
mg
k
Okay, and this makes sort of sense because you figure that, as you have a more massive
block hanging from the spring, [g. block] then your position x is gonna increase, [g.~ x in
diagram] which is what this is showing. [g.~ m then k] And that if you have a stiffer spring,
then your position x is gonna decrease. [g. uses fingers to indicate the gap between the
mass and ceiling] That why it's in the denominator. So, the answer makes sense.
In the first part of this interpretation, Jim sees the expression in terms of the PROP+ form, x is seen
as directly proportional to the mass: “as you have a more massive block hanging from the spring,
then you're position x is gonna increase.” And, in the second part of the interpretation, the
expression is seen through the lens of PROP-: “if you have a stiffer spring, then you're position x is
gonna decrease.”
There were many instances in which students interpreted final expressions of this sort by
announcing proportionality relations. For example, when solving the air resistance problem,
students typically derived an expression very similar to
v=
mg
k
for the terminal velocity. In this relation, k is just a constant. To cite an instance, after Alan and
Bob derived this equation, Alan commented:
Alan
So now, since the mass is in the numerator, we can just say velocity terminal is
proportional to mass [w. Vt∝m] and that would explain why the steel ball reaches a faster
terminal velocity than the paper ball.
And, concerning the same relation, Mark said simply:
Mark
So the terminal velocity of the heavier object would be greater.
Now I will move on to a form we have not before encountered, the RATIO form. In the RATIO
form, the symbol pattern involves an individual symbol in both the numerator and denominator
of an expression. The key to the RATIO form is that the ratio is seen as a comparison—it’s a
comparison of the quantities that appear on the top and the bottom of the ratio. In most cases, the
quantities involved are of the same type; for example, they may be both masses or both velocities.
To see an example, I return to Alan and Bob’s work on the Stranded Skater problem. Recall
that they began their work on this problem by writing this expression.
0 = ms vs − m pv p
84
With this expression in hand, they proceeded to solve for v p , the velocity of Peggy:
vp =
ms
vs
mp
Then Alan commented:
Alan
Well the mass of the skate [g. Ms] is approximately small - it’s smaller. It’s probably much
smaller than mass of Peggy. [w. Ms<< Mp] …So your velocity is gonna be really small. [w.
Vp<<Vs]
The cueing to activity of the RATIO form almost always leads to a certain question being asked:
Which is bigger, the numerator or the denominator? This question can be seen at the heart of
Alan’s comment. He notes that, not only is the mass of the skate smaller than Peggy’s mass, it is
probably much smaller. Alan’s concern with the relative size of numerator and denominator
suggests that he is seeing the RATIO form in this expression; in particular the ratio of the masses,
ms m p , is seen in terms of this form.
The RATIO form frequently appeared in a very similar context in the Buoyant Cube problem.
Student work on this problem usually produced an expression similar to the solution presented in
Figure Chapter 3. -6.
s=
ρi
L
ρw
This expression tended to encourage interpretations based on a comparison of the densities of ice
and water. The following passages are typical:
Jack
This does seem to make sense cause if ice had the same volume as water [g. ρi then ρw],
er, same density as water, then L would equal S which means the cube would be right at
the surface.
_______________
Mike
Does it seem sensible? Well, let's see. Um. If ice is lighter than water, [g. ρ(I) then ρ(w) in
last] then ice will float.
Karl
It makes sense because this is gonna be less // than one. [g. ρ(I)/ρ(w) in last]
// Less than one.
Mike
Karl
And we know that it's gonna float.
Mike
Right. So, yeah, that makes sense.
Karl
That makes sense.
The final form in the Proportionality Cluster is the CANCELING(B) form. This form bears the
name CANCELING(B) to distinguish it from the CANCELING form in the Competing Terms Cluster.
85
The reason for retaining the term “canceling” in both names is that students very commonly use
the word “cancel” in association with both of these forms.
Since, as we shall see in Chapter 6, I observed this form only rarely, I will present just a single
brief incident for illustration. This incident took place during an episode in which Mark and
Roger were discussing why all dropped objects fall at the same rate (neglecting air resistance). To
answer, Mark wrote the following expressions and commented:
a=
Mark
F
; F = mg
m
You know, however you have a force [g. F in a=F/m] that varies at the same proportion
directly - directly with the mass, don't you? The heavier - the heavier an object is, the more
force it will exert, right? Upon, like say I try to lift it. … So, if it's double the mass it's gonna
have double the force downward.…Two M G. Right? But the mass is also two M , so those
two M's cancel and you get - you're left with A equals G in all cases, aren't you.
The idea here is that, in a sense, Mark has mentally combined these two expressions to obtain
a=mg/m. This is Mark’s expression for the acceleration of a dropped object. With this expression
in mind, Mark imagines what would happen to the acceleration if the mass is doubled. The answer
is that nothing would happen, any change in the numerator is canceled by a corresponding change
in the denominator. This canceling of parallel changes in the numerator and denominator of a
ratio is the core of the CANCELING(B) form. Here, it constitutes the basis of Mark’s explanation
concerning why all objects fall at the same rate.
Forms and the limits of qualitative reasoning
This discussion of the Proportionality Cluster provides an appropriate context to consider an
important issue, what I will call the “qualitative limits” of forms. For some first examples, I want
to look to the air resistance problem. As I have discussed, two forces act on the dropped object in
the air resistance task, gravity and a force from air resistance. All students know an expression—a
force law—that specifies the force of gravity, Fg = mg. However most students do not know an
expression for the force due to air resistance. While some may have seen such an expression once
or twice, it is rare for students to have committed an equation to memory. For this reason, the
students have no option but to construct their own expression for the force of air resistance.
The “right” answer—the answer you will find in an introductory textbook—is
Fair = kv 2
86
where v is the current velocity of the object and k is a constant.11 In general, the students in my
study just stated that the force of air resistance should be proportional to velocity, and wrote
either the above expression or Fair = kv . Here are some typical examples:
Fair = kv
Bob
Okay, and it gets f-, it gets greater as the velocity increases because it's hitting more
atoms (0.5) of air.
_______________
R = µv
Mark
So this has to depend on velocity. [g. R] That's all I'm saying. Your resistance - the
resistor force depends on the velocity of the object. The higher the velocity the bigger the
resistance.
The important issue here is that the PROP+ form doesn’t specify whether the force of air resistance
should be kv or kv 2 . Some students ran headlong into this difficulty, realizing that understanding
that the force of air resistance must be proportional to the velocity was not sufficient to determine
what relation they should write. Jack went so far as to list the possibilities. In the following, the ‘∝‘
symbol means “is proportional to.”
Fu ∝ v
∝ v2
∝ v3
Jack
Somehow they’re related to the velocity but we’re not sure what that relationship is. All we
can say is that F U is either proportional to V [w. F∝v] or to V squared or to V cubed. [w.
∝v2; ∝v3]
This type of ambiguity seems to be a fundamental property of forms; there are limits to the
specificity with which forms describe equations. And this is true both when reading and writing
equations. When we see an equation in terms of PROP+ and say “it’s proportional,” it is simply not
consequential whether the proportionality relation is linear or quadratic.
The students in my study came up against these limits in a number of cases, and not only in
the context of the air resistance task. For example, when working on the spring task, Mike and
Karl could not remember whether the force due to a spring is F = kx or F = 1 2 kx 2 . Although
they knew that the expression needed to be consistent with PROP+, this was not sufficient to
distinguish between these two equations:
11
A more complete expression could include other parameters, such as the cross-sectional area of the
dropped object. However, for the present task it is sufficient to presume that these other parameters
are part of the constant, k, and that this constant is the same for both objects.
87
F = kx ; F = 1 2 kx 2
Karl
Okay, now, qualitatively, both K X and half K X squared do come out to be the same
answer because as (…) looking at it qualitatively, both half - both half K X squared and K
X, um, you know, increase as X increases.
Furthermore, this was no fleeting difficulty. Mike and Karl struggled with this question for about
25 minutes, but to no avail. Up to the level of their specification, these equations were
indistinguishable. As Karl says, both “increase as X increases.”
In Karl’s above statement, he expresses the intuition that two expressions are “qualitatively”
equivalent. Of course, what it means for two expressions to be qualitatively the same is far from
clear. Interestingly, however, there is a relevant body of literature that attempts to make this
intuition precise. This literature describes what the researchers involved call “qualitative reasoning
about physical systems” (deKleer & Brown, 1984; Forbus, 1984).
The qualitative reasoning research is concerned with describing a certain slice of physics
knowledge and reasoning. This work falls into the “formalization” category I described in Chapter
1; these researchers have attempted to create a formalization of a particular slice of physics
knowledge. The interesting observation, from the point-of-view of the current discussion, is that
the knowledge system formalized by the qualitative reasoning researchers apparently possesses
similar limits to those described here. For this reason, it merits a brief discussion.
Paradoxically, the focus of qualitative reasoning is on what physicists would call physical
quantities. The literature that discusses qualitative reasoning makes specific hypotheses about how
people reason about physical quantities to describe and predict behavior. One hypothesis concerns
the nature of the information about a quantity that people encode. The range of possible values
that a quantity can attain is seen as divided into a set of discrete regions, with the boundaries at
“physically interesting” places. Then, the hypothesis is that, for qualitative reasoning, only the
following information about a quantity is relevant: (1) the region the quantity is in, and (2) its
direction of change.
In addition, later literature in this vein made hypotheses about how people propagate
knowledge about changes in one quantity to changes in another. Of particular interest for us are
proportionality relations, which Forbus (1984) calls “qualitative proportionalities.” For example, if
we know that two quantities are qualitatively proportional, and we know the direction of change
of one quantity is such that it is increasing, then we know that the second quantity is also
increasing. Notice that, just as in the case of forms, it is not appropriate to ask whether the relation
is linear, second order, or otherwise. So the knowledge system formalized by the qualitative
88
reasoning literature appears to possess the same limits as forms. As with
PROP+,
a qualitative
proportionality relation between force and velocity says simply that force increases with increasing
velocity (or decreases with decreasing velocity), and that is the limit of the specification.
These “qualitative limits” constitute an essential part of the hypothesis associated with the
statement that people construct equations from and see equations in terms of symbolic forms.
Relations and symbolic expressions are only schematized at a certain level, and only some details
are relevant. Really, this should not be a surprising result. If every distinction was relevant—if we
didn’t carve up our experience and register it in some way—then our worlds would be an
overwhelming mass of details. Of course, with learning, we can learn to register more detail
productively; we can learn a larger vocabulary of forms. It is precisely such a developmental
process that Gombrich hypothesized to explain the evolution of art and individual artists. The
skilled physicist—as well as the skilled artist—learns an increasing vocabulary of forms in terms of
which to see the world, as well as for seeing the world as represented on paper.
The Terms are Amounts Cluster
Forms in the Proportionality Cluster, for the most part, address expressions at the level of
individual symbols, with special attention to whether these individual symbols are located above
or below the division line in a ratio. In this cluster, I am going to again be describing forms that,
like those in the Competing Terms Cluster, address expressions at the level of terms. However,
these forms differ in an important way from forms like competing terms and opposition. In
opposition and other forms in the Competing Terms Cluster, terms are seen as corresponding to
influences. In a very general sense, these influences are directed and they work to somehow move
an object toward their desired ends.
In contrast, in the Terms are Amounts Cluster, terms are not treated as influences, they are
treated more like a quantity of a generic substance. Rather than describing a battle between
competing terms, these expressions concern the collecting of this substance, putting some in and
maybe taking some away. Thus, while signs in the Competing Terms Cluster are commonly
associated with directions in physical space, signs in this cluster generally signal adding on or
taking away from the total amount of stuff. The forms in the Terms are Amounts Cluster are
listed in REF _Ref336930235 \* MERGEFORMAT Table Chapter 3. -3 .
89
parts-of-a-whole [ + + …]
base ± change [ ± _]
whole - part [ - ]
SAME AMOUNT
=
Table Chapter 3. -3. Forms in the “Terms are Amounts” Cluster.
The first form in this cluster, PARTS -OF -A-WHOLE, played a role in Mike and Karl’s construction
of their novel expression for the coefficient of friction. Their friction had two “components” that
together comprised the whole coefficient of friction. However, that episode does not really
provide the clearest and most illustrative example of that form; in particular, it is a little hard to
see the coefficient of friction as a “quantity of substance.” Indeed, this is probably something of a
marginal instance. For that reason, I want to present an episode in which a more typical instance of
the PARTS -OF -A-WHOLE form appears. This episode involves students engaged in solving a problem
we have not yet encountered, “Running in the Rain.”
Suppose that you need to cross the street during a steady downpour and you don’t have an umbrella. Is
it better to walk or run across the street? Make a simple computation, assuming that you’re shaped like
a tall rectangular crate. Also, you can assume that the rain is falling straight down. Would it affect your
result if the rain was falling at an angle?12
Students generally thought that this question was a little silly, but they were game to give it a
try. A diagram from Alan and Bob’s work on this task is shown in Figure Chapter 3. -8. Early on
in their work, Alan and Bob had a little argument concerning whether rain would hit only on the
top of the crate, or on the top and front side as you (the crate) moved across the street. Alan
started the discussion by taking the stance that rain would only strike the top of the crate:
Alan
It’s falling straight down. If the rain is falling straight down we're only concerned with how
much water is actually hitting this portion right here. [w. draws in top and darkens]
Figure Chapter 3. -8. Alan and Bob's diagram for the “Running in the Rain” problem.
12
This problem is taken from The Flying Circus of Physics (Walker, 1975).
90
But Bob disagreed, and the discussion below ensued:
Bob
That's if you're standing still!
Alan
If we're running, and this box is move=
Bob
=But see there's rain=
Alan
You're still only looking at the rain that hits the top of the box. [g. top of box]
Bob
Your picture - your picture's not accurate. Okay. There's rain everywhere. [w. draws rain
in front of the box]
Alan
Oh, so you're talking about the rain that=
Bob
=Okay, you run this way [w. arrow indicating direction of motion] - well, yeah, it's sort of
bizarre. Okay, it is bizarre.
Alan
So, you're saying it's gonna hit the raindrops on the side, so we need to account for this,
[w. darkens the front area]=
At the end of the last passage, Alan agrees that rain would hit the front of the crate. With
consensus on this fact, Alan and Bob went on to construct an expression for the total rain that
strikes the person which consisted of two parts, one part corresponding to what strikes the top of
the crate and one part corresponding to what strikes the front.
total rain =
Bob
# raindrops
+C
s
There's two - two sources of rain. Okay, rain that you walk into and rain that drops on you.
[g. rain hitting his head] Okay. Walk or run into. Okay, so, the number of rain per second
[g. #raindrops/s] is how much rain hits your head. … if you're a crate and you - you move,
move into some rain. Since you move from here to here, [g. two points on the board],
there's - since there's rain coming down at a constant rate [g. falling rain], there's always
rain in front of you. Okay, so there's raindrops in front of you. So if you walk forward you
hit it, hit that rain. Okay. And-and- I mean, you're gonna hit some rain when you walk
forward.
Alan and Bob’s expression follows from
PARTS -OF -A-WHOLE;
the total amount of rain is seen as
consisting of two parts. The first term corresponds to the rain that strikes the top of the crate, and
the second term corresponds to the rain that strikes the front of the crate. In this case, thinking of
the total rain as a total amount of “stuff” is relatively natural. You can almost imagine a container
filling up with all this rain, flowing in through the two possible surfaces.
One interesting aspect of the above expression deserves a brief comment. Notice that this
expression is not quite a syntactically correct equation. In the place of the first term there are just
some units, the number of raindrops per second. Furthermore, these aren’t even the correct units
for this term; ultimately, Alan and Bob built an expression in which the terms all had units of
volume. Given the “forms” view, such partially syntactic expressions are to be expected. Since
forms like
PARTS -OF -A-WHOLE
only describe equations at a certain level of description, it makes
sense that expressions might appear in a schematic form during early stages of their construction. I
91
believe that this is precisely what we are seeing in Alan and Bob’s expression. The units written in
the position of the first term can be taken as functioning as a placeholder for a more complete
expression to be written later. In fact, this is exactly how this expression functioned in Alan and
Bob’s work. Ultimately both the first and second term were elaborated in terms of some useful
parameters.
The next two forms in this cluster, BASE +CHANGE and BASE -CHANGE , are flip-sides of the same
coin. Both of these forms typically correspond to expressions involving two terms. The first term
is a “base” quantity, and the second term is an amount that is taken away or added to this base.
The BASE +CHANGE form may appear indistinguishable form
PARTS -OF -A-WHOLE,
but this is not the
case. In the case of PARTS -OF -A-WHOLE, the two “parts” are semantically indistinguishable; within
the schema, these two entities are of precisely the same sort. In contrast, in BASE +CHANGE there is a
semantic asymmetry between the entities corresponding to each of the terms; in particular, there
will be some sense in which one of these entities is a base to which something is added.
In fact, the key feature to look for in identifying BASE +CHANGE or BASE -CHANGE is the notion
that some quantity is made greater or made less. The typical case is one in which a student already
has an expression of some sort and they know, for example, that they want a new expression that
is larger than the first. So what do they do? They add something onto the first expression.
Let’s look at an example. Part (b) of the air resistance task asked students the following
question:
(b) Suppose that there was a wind blowing straight up when the objects were dropped, how would
your answer differ? What if the wind was blowing straight down?
Students approached this problem in a number of ways. Sometimes they performed a reference
frame analysis, and sometimes they simply treated the wind as an additional force. But I do not
want to get into these particular details at this point. I want, instead, to jump to the final result.
First, recall that the final expression for the first part of the air resistance task was:
vt =
mg
k
This equation expresses the terminal velocity as a function of the mass, the acceleration due to
gravity, and a constant, k. Most students eventually derived an equivalent expression, although
they used a variety of symbols to stand for the constant.
92
In Mark and Roger’s solution to Part (b), they treated the wind as an additional force. When
they were done, they had two new expressions for the terminal velocity.
vt =
mg − Fw
mg + Fw
; vt =
µ
µ
Mark and Roger knew that one of these expressions corresponded to the case where the wind is
blowing up and the other corresponded to the case where the wind is blowing down, but the
question was, which equation goes with which case? In the next passage Roger worked out the
answer to this question:
Roger
Okay, well for the up case - yeah, terminal velocity will be (0.5) // less [g. -Fw in Vt=(mgFw)/µ] by a factor F W over mu. [w. Fw/µ] Right?
// less
Mark
Mark
Yeah. // Yeah, by - By a constant. By a constant factor.
// And for the down case - okay, because we're assuming a constant wind right,
Roger
force?
Bruce
Uh-huh.
Mark
By a constant [g. Fw in (mg-Fw)/µ] - it'll be less by that much. [w.O Fw/µ, adds a minus sign]
And minus. And then you have to-
Roger
This would be greater by F W over mu. [w. +Fw/µ in lower area]
Roger decided that the expression with the minus sign goes with the case in which the wind is
blowing up. His reasoning was that, if the wind is blowing up, then the terminal velocity should be
less. And, following
BASE -CHANGE ,
in order to make a quantity less, you have to subtract
something.
There is a minor subtlety here. The force of wind, Fw , could potentially be treated as an
influence and we could see the expression mg − Fw in terms of COMPETING TERMS. In that case, the
sign of Fw would be chosen based on some association of signs with directions. But this is not
what Roger does in the above passage. He does not argue that there needs to be a negative sign in
front of the Fw because of the direction that the wind points. Rather, he argues that there needs
to be a negative sign because the terminal velocity must be less than it was in the base case—the
case in which there was not any wind. This is evidence for BASE -CHANGE .
I want to mention one other place that BASE +CHANGE appears, mainly because it involves an
equation that is learned very early and used frequently: v = vo + at . At a number of points during
the problem solving sessions, I pressed students about this equation, asking them such things as
93
how they knew it was true, and whether they knew it “by heart.” During the final session, I also
each asked pair whether they thought that this equation was “obvious.” Mike and Karl responded:
Mike
Well yeah it is obvious because, well velocity will equal to V naught if it's not being
disturbed. But if it's being acc- If there's a f- acceleration action on it, then uh- and that's
constant, you know then velocity will be decreasing as time goes on. Or increasing,
whatever it works, I mean whichever it does. So, it's like whatever it is and then plus a
correction on the acceleration. So yeah it makes sense. It's obvious, yes it is.
…
Karl
What's obvious to me is that you have the final velocity is obviously going to be equal to
the initial velocity plus however much, however faster it gets. That's what's obvious to me.
What's not necessarily positively obvious is that the amount, the amount that it gets
faster is the acceleration times the time.
Mike begins by answering that, yes, v = vo + at is obvious. And it is obvious for Mike because it is
consistent with the BASE +CHANGE form, there is a base, vo , plus some “correction.” Karl agrees that
the equation is obvious, but only in a limited sense. In particular, he’s not sure that it’s necessarily
obvious that the second term must be precisely “at”. He agrees that there needs to be an initial
velocity plus some change that depends on the acceleration, but he does not know exactly what
that change should look like.
Here, again, we have run into the limits of forms. Forms specify equations to be written, but
only up to a certain level of detail. The interpretation of v = vo + at given by Mike and Karl was
the one provided by almost all students; they typically saw this equation in terms of
BASE +CHANGE
and at no finer level of detail. In fact, it was not uncommon for students to write the second term
incorrectly. For example, rather than writing v = vo + at , Mark and Roger wrote:
v = vo + 12 at 2
Mark
Cause we have initial velocity [w. circles vo term] plus if you have an acceleration over a
certain time. [w. circles 1/2at2] Yeah, I think that's right.
Although Mark’s expression is consistent with BASE +CHANGE —it has an initial velocity plus a
change involving the acceleration—this expression does not correctly give the dependence of
velocity on time for cases of constant acceleration. In fact, the units of the second term are not
even correct.
Again, we have students seeing an equation in terms of a meaningful structure, but it is a
structure that does not recognize or determine all details of the equation. The point here is really
not any different than points I made just above. Nonetheless, some physicists and physics teachers
may be a little surprised by these episodes since, to the physics expert, the equation v = vo + at
94
seems very straightforward, at every level of detail. I will have more to say about this in Chapter
12.
The final two forms in the Terms are Amounts Cluster, WHOLE - PART and SAME AMOUNT ,
appear extremely rarely in my data corpus, so I will only mention them briefly here. In
PART,
WHOLE
-
rather than adding to or subtracting from some base quantity, a piece of a whole is
removed. Typically, this piece is an identifiable component of the original whole. SAME AMOUNT is
the counterpart of the BALANCING form. Rather than involving equal and opposite influences, it
states that two collections contain the same amount of stuff. A statement that energy is conserved,
for example, may be seen in terms of this form.
The ÒDependenceÓ Cluster
The forms in the Dependence Cluster are, in a sense, extremely simple. In fact, their symbol
patterns may hardly even merit being called “patterns.” Nonetheless, I will argue in Chapter 6 that
these forms are common and important. Table Chapter 3. -4 lists the forms in this cluster.
DEPENDENCE
NO DEPENDENCE
SOLE DEPENDENCE
[…x…]
[…]
[…x…]
Table Chapter 3. -4. Forms in the Dependence Cluster.
The forms in this cluster have to do with the simple fact of whether a specific individual
symbol appears in an expression or does not appear in an expression. Most basic of all these forms
is NO DEPENDENCE , whose symbol pattern involves the absence, rather than the presence of
symbols. When we see an expression in terms of NO DEPENDENCE , we are noting that a particular
symbol does not appear in the expression. For example, in the construction of their expression for
µ, Mike and Karl wanted to have one component that does not depend on the weight (mass). This
is an application of NO DEPENDENCe:
µ
µ = µ1 + C 2
m
Karl
Well yeah maybe you could consider the frictional force as having two components. One
that goes to zero and the other one that's constant. So that one component would be
dependent on the weight. And the other component would be independent of the weight.
A prototypical use of NO DEPENDENCE can be found in students’ work on the Shoved Block
problem. Recall that the “correct” answer to this question is that the heavier and lighter block
travel precisely the same distance, if started with the same initial speed. In order to arrive at this
95
conclusion, the students needed to derive an equation in which the mass doesn’t appear. Alan and
Bob derived such an expression and then commented:
a = gµ
Alan
Bob
Right, so, actually, they should both take the same.=
=Wait a minute. Oh, they both take the same! [Surprised tone]
…
Bob
So, no matter what the mass is, you're gonna get the same, the same acceleration.
Here, Alan and Bob are seeing their equation in terms of NO DEPENDENCE . They state that, because
no mass appears in this expression, the acceleration doesn’t depend on the mass, so both objects
travel the same distance. Note that this really does constitute a specific way of “seeing” this
expression. To see this expression in terms of NO DEPENDENCE , Alan and Bob have to do the trick
that is always associated with seeing equations in terms of forms; they have to blur certain features
into the background, and highlight other features for themselves. In this case, the absence of the
symbol ‘m’ is the key registered feature. Furthermore, notice the character of Alan and Bob’s
reaction in the above passage. NO DEPENDENCE is not something subtle; it is striking and obvious to
them, and elicits a strong reaction.
The flip-side of NO DEPENDENCE and the next step up in simplicity is DEPENDENCE. When
seeing an expression in terms of DEPENDENCE, we simply note the appearance of a particular
individual symbol. In the construction of expressions, this form is frequently used in tandem with
other forms. For example, in the early specifications of their expression for µ, Mike and Karl
stated that one component should be dependent on the weight. (Refer to the last Karl passage
above.) This was later elaborated to the statement that this component should be inversely
proportional to the weight. Similarly,
DEPENDENCE
usually played a role in students’ steps toward
the construction of an expression for the force of air resistance:
R = µv
Mark
So this has to depend on velocity. [g. R] That's all I'm saying. Your resistance - the
resistor force depends on the velocity of the object. The higher the velocity the bigger the
resistance.
DEPENDENCE and NO DEPENDENCE are the basic and most common forms in this cluster. The
one additional form is, in a sense, a slight variation on this theme. In SOLE DEPENDENCE , one
quantity is seen as depending only on a single other quantity. Consider this passage from Alan and
Bob’s work on Running in the Rain:
96
total rain(t) = #rain
t + ρrain hwx
s ( )
Alan
You're just gonna- (well) based upon that equation right there you're just gonna conclude
that you would feel more - you would have more raindrops hitting you if you walk because
it's gonna take you a longer time since the last term is just a constant. And you're just
saying it just depends on the time.
Bob
Right.
Alan
It's only dependent upon time.
In the above passage, Alan argues that the time needed to cross the street, t, is larger if you walk
than if you run. Thus the “total rain” is larger if you walk. (Based on an inference from PROP+.)
Furthermore, Alan goes on to emphasize that time is the only quantity that the total rain depends
on. Across the two cases of interest—walking and running—the time is the only varying quantity
that appears in the expression. This is an application of SOLE DEPENDENCE .
The Coefficient Cluster
In the COEFFICIENT form, a product of factors is seen as broken into two parts. One part is the
coefficient itself, which often involves only a single symbol and almost always is written on the
left. This is indicated in the symbol pattern shown in Table Chapter 3. -5.
COEFFICIENT
[x
]
SCALING
[n
]
Table Chapter 3. -5. Forms in the Coefficient Cluster
The COEFFICIENT form is distinguished from other forms because of some properties that are
typically possessed by the coefficient. Students will frequently comment, for example, that a
coefficient is “just a factor” or “just a number.” It is, in fact, the case that important dynamical
variables rarely appear in the position of the coefficient. For example, you will almost never see a
student treat a force as a coefficient. And you will not see velocities or positions that vary over the
course of a motion treated as coefficients. Instead, coefficients most frequently involve
parameters, such as the coefficient of friction, that, in a sense, define the circumstances under
which a motion is occurring.
The proportionality constant introduced by students in the air resistance problem is a typical
example of a coefficient.
F = kv 2
Jack
Right, all I did was introduce a constant of proportionality. [g. k] We have no idea what it
is.
97
Here, Jack sees this expression in terms of the
COEFFICIENT
form; the constant k is seen as
multiplying the rest of the expression. Furthermore, Jack’s comment that “We have no idea what
it is” is characteristic of the type of things that people say about symbols that are being treated as
coefficients. Jack makes this comment in an utterly flippant manner. The suggestion is that,
although they have no idea what this coefficient is, it doesn’t matter that they don’t know. For the
most part, coefficients only matter at some boring level of detail. They can tune the size of an
effect or influence, and in that way have a fine quantitative effect on a motion, but they are not
generally seen as influencing the overall character of a motion or the result in question.
These details constitute the stance that goes with the COEFFICIENT form. Notice that such a
stance constitutes a very specific orientation to an expression or portion of an expression. A
product of factors, such as abc, need not have any particular structure to it. We need not,
necessarily, see the ‘a’ as any different than the ‘b’ or ‘c.’ But, when a physicist does see
COEFFICIENT
here—when they do see the ‘a’ as a coefficient that multiplies ‘bc’— the implications
are significant: All of the above stance comes along in the bargain. This is part of what forms do.
They do not only blur features; in carving up an expression, they also highlight features of the
symbolic landscape and give meaning to those features. When a physicist sees the expression abc in
terms of the
COEFFICIENT
form, they are seeing a specific, meaningful structure.
The second and only other form in this cluster is scaling. Scaling is very similar to the
COEFFICIENT
form, with the single important difference being that, in scaling, the coefficient is
unitless. Suppose, for example, that we have an expression of the form
T2 = nT1
where n is a (unitless) number multiplied by a collection of factors T1 to produce the result T2 .
Because n is unitless, T2 is the same type of entity as T1, and we can think of the coefficient n as
an operator that acts to “scale up” or “scale down” T1. In addition, inferences associated with the
scaling form tended to be very salient in my data. Here are a few common examples:
•
If n > 1 then T2 > T1
•
If n < 1 then T2 < T1
•
If n = 1 then T2 = T1
•
If n >> 1 then T2 >> 1
98
The scaling form is nicely illustrated in some examples involving the Buoyant Cube problem.
Recall that the point of this problem is to find how much of an ice cube is below the surface of the
water when an ice cube floats at rest. Alan and Bob obtained this simple expression as a final result
and commented:
x =.92L
Bob
Um, okay. [w. x=.92L] Yeah. Point nine two L. (2.0) Does that make sense? Yeah.
Alan
Yeah, because like when you see an iceberg they say that most of the iceberg is below
the water, so that's why ships have to be really careful.
In this example, the result x is the same type of entity as L; both are lengths. The factor .92 scales
L to produce a length that is “most” of L, a result that Alan and Bob believe makes sense.
The Multiplication Cluster
The Multiplication Cluster is the final cluster to be discussed in this primer concerning forms.
To this point I have discussed five other clusters. Two of these clusters, Competing Terms and
Terms are Amounts, deal with expressions at the level of terms. Another two clusters,
Proportionality and Dependence, were concerned more with individual symbols. Forms in the
Dependence Cluster were sensitive to the simple presence or absence of these individual symbols.
And forms in the Proportionality Cluster were concerned with whether particular symbols
appeared above or below the division sign in a ratio.
The observation I am working toward here is that only the fifth cluster—the Coefficient
Cluster—was concerned with structure in the arrangement of a product of factors. Consider, once
again, an expression of the form abc. In terms of what forms can a student see this expression? The
Terms as Amounts forms and Competing Terms forms could allow the whole of the expression to
be seen as an influence or as some “stuff.” In addition, we could see this expression in terms of
DEPENDENCE
in three ways, as dependent on a, b, or c. Similarly we could see prop+ in three
separate ways, as focusing on any of the factors involved.
However, only
COEFFICIENT
breaks down this string of factors into a meaningful structure; it
splits off the first factor, a, and treats it differently. And this is a very specific structure that will
only be relevant in a limited number of cases. It is certainly not true that every time we multiply
two things, one can profitably be thought of as a coefficient. So we are left wondering: Isn’t there
some other way to see meaningful structure in a string of factors? Aren’t there some forms missing
here?
99
Of course, it may simply be the case that the vocabulary of forms just happens to be a little
sparse in this territory, at least for students working on the tasks in my study. In fact, to a certain
extent, I believe this is the case. However, there appear to be a couple of forms that cover some of
this territory. I will discuss these forms presently, but I do this with the qualification that these
forms were relatively rare in my data and their identification is somewhat more tentative then
others I have described.
INTENSIVE •EXTENSIVE
EXTENSIVE •EXTENSIVE
x×y
x×y
Table Chapter 3. -6. Forms in the Multiplication Cluster.
As shown in Table Chapter 3. -6, the two forms in the Multiplication Cluster are called
INTENSIVE •EXTENSIVE
and EXTENSIVE •EXTENSIVE . The use of the terms “intensive” and “extensive”
here is taken from some other researchers, who used them in similar contexts (Greeno, 1987; Hall,
Kibler, Wenger, & Truxaw, 1989). I will discuss this research at the end of this chapter.
Intensive and extensive quantities are distinguished in roughly the following away. An
intensive quantity specifies an amount of something per unit of something else. In contrast, an
extensive quantity is a number of units. So if, for example, we are solving an arithmetic word
problem that gives us a number of apples per basket, a number of baskets, and asks us how many
total apples there are, then the number of apples per basket is intensive, and the number of baskets
is extensive.
As the names imply, the INTENSIVE •EXTENSIVE form applies when an expression, such as xy, is
seen as a product of an intensive and an extensive quantity, and the EXTENSIVE •EXTENSIVE form
applies when such an expression is seen as a product of extensives. In my data, intensive quantities
were almost always densities or rates. So, the
INTENSIVE •EXTENSIVE
form usually involved the
product of a rate and a time or a density and a volume. The fact that this form applied in such a
narrow range was part of the reason that it was difficult to establish its general validity.
There are a few instances, however, in which the INTENSIVE •EXTENSIVE form was applied in a
more flexible manner, and seemed to slightly transcend this narrow range. A number of these
instances appeared while students worked on the Running in the Rain problem. For example,
while working on this problem, Jon and Ella defined a quantity that was essentially a flux rate for
the rain.
100
 # drops 
 A•s 
Jon
Do we have a rate for the raindrops? We need the number of drops per-…per area and per
second. [w. #drops/A•s]
So, Jon and Ella’s rate is the number of drops that flow through an area per unit time. After
defining this quantity, they went on to multiply it by the area of the top of the crate and the time
to cross the street, in order to get an expression for the total amount of rain that flows through the
top of the crate.
 # drops  A
(
)(t )
 A • s  head walk
Jon
And then we multiply that by the area of our head. [w.mod (#drops/A•s)(Ahead)] Oh, maybe
we should just say - Times the time we're in the street (for) walking. [w.mod (#drops/A•s)
(A head)(twalk)]
A number of the features of this little episode are typical of cases in which intensive•extensive is
applied. First, the use of the term “rate” is an important indicator. Second, explicit discussion of
the units involved seems to be common. Often the units of the factors are written out and used as
a guide in composing the expression. This is precisely what Jon and Ella did in this episode.
I will not give any additional examples of the intensive•extensive form here, but I do want to
foreshadow an observation that I will make in Chapter 12. Recall that the ‘at’ in the equation
v = vo + at , while somehow obvious for experts, was not obvious for the students in my study. I
believe that it is partly through the INTENSIVE •EXTENSIVE form that this term becomes obvious for
experts; in particular, they see it as a rate times a time. Furthermore, the fact that students do not
see intensive•extensive in this equation may be symptomatic of the overall low appearance of forms
in this cluster.
The
IDENTITY
form
To wrap up this primer, I want to discuss a pair of forms that do not really fit into any of the
above clusters. To begin, note that few of the forms discussed so far involve the use of an equal
sign. The only exceptions are BALANCING and SAME AMOUNT . Now I want to mention another form
involving an equal sign that covers a very large percentage of the expressions that are composed
and interpreted: the IDENTITY form. In IDENTITY, a single symbol, usually written on the left, is
separated from an arbitrary expression by an equal sign. I mentioned this form earlier in my
101
account of Mike and Karl’s construction of their novel expression for the coefficient of friction. It
was IDENTITY that allowed them to write the “µ=” on the left side of the expression.
IDENTITY
x = […]
IDENTITY is so common that it is nearly invisible and rarely commented upon directly. As I
mentioned,
IDENTITY
allows the results of inferences to transparently travel across the equal sign to
the symbol on the left. Whatever is true of the expression on the right is true of the quantity
identified with that expression. Furthermore,
IDENTITY
is not only involved in the construction of
expressions, expressions are frequently manipulated so as to be in alignment with IDENTITY before
they are interpreted. This was the case with the final expression from the spring task that Jim
interpreted:
x=
Jim
mg
k
Okay, and this makes sort of sense because you figure that, as you have a more massive
block hanging from the spring, [g. block] then you're position x is gonna increase, [g.~ x in
diagram] which is what this is showing. [g.~ m then k] And that if you have a stiffer spring,
then you're position x is gonna decrease. [g. Uses fingers to indicate the gap between the
mass and ceiling] That why it's in the denominator. So, the answer makes sense.
In this passage, Jim sees the right side of this expression in terms of PROP+ and PROP-. This means
that if x increases, the whole right side increases. IDENTITY tells Jim that what is true of the whole
right hand side, is true of x: If the whole right hand side increases, then x increases.
In truth, I have not been careful with my language in such situations. Throughout my
examples, whenever an equation was in line with
IDENTITY,
I spoke interchangeably of the whole
right hand side and the quantity identified with that expression. In what follows, I will continue to
be loose in this way. Only when precision is called for or when certain issues arise will I bother to
note the role being played by the IDENTITY form.
Meanings of the equal sign
While I am on the subject of the equal sign, I want to briefly mention some related literature
from mathematics education. This literature is concerned with how younger, mostly grade-school
students understand the “meaning” of an equal sign. Central to this literature is the observation
that novice algebra students tend to see an equal sign as a “do something signal” (Kieran, 1992;
Herscovics & Kieran, 1980). The equal sign is seen as simply separating the answer from the
computation to be performed, as in 5+4=9.
102
These researchers take the stance that there is nothing particularly wrong with this way of
understanding an equal sign. However they believe that it would be profitable if students’
understanding of the meaning of equal signs and equations were enlarged in a particular manner.
They have in mind an understanding in which expressions involving an equal sign are taken as
symmetric and transitive relations, as in expressions like 5+4=6+3. Such an understanding is seen
to be critical for more advanced mathematics in which equations are treated as objects in their
own right.
There is a certain temptation to attempt to match my forms,
IDENTITY
and BALANCING , with
the “do something” and “symmetric relation” versions of an equal sign. But note that such a
correspondence is not easily made. The equal sign in the
IDENTITY
form doesn’t necessarily specify
an action to be performed; rather, it states that one quantity can be identified with an expression.
In fact, the shift from writing the single symbol on the right to writing it on the left may be
significant. In the case of
IDENTITY,
the single symbol is not the computed result of a
computation; instead, it announces the nature of the expression to follow.
But I do not want to push this comparison too far. Keep in mind that my subjects are
advanced college students and the “equal sign” literature deals with novice algebra students;
therefore, we should expect significant differences. However, I believe that these mathematics
education researchers can draw some important morals from the presentation in this chapter. If
the “forms” view is correct then, as students progress to expertise, we should not expect them to
move toward possessing only one or two stances toward the meaning of an equal sign, identified
with distinctions like “process” versus “object.” Instead, we should expect something more like an
idiosyncratic vocabulary of stances, that depends on the domain and the tasks undertaken in that
domain. I will touch on related issues in the chapters that follow.
D YING A WAY
Lastly, as the final entry in this primer, I want to note one form that is interesting because it
suggests how the vocabulary of forms may be extended in expert directions. The form is called
DYING AWAY
and the symbol pattern involves exponentiation to a negative power.
DYING AWAY
[e− x…]
This form appeared just a few times in my data corpus. When working on Running in the
Rain, Jack and Jim hypothesized that the probability of getting struck by a raindrop would die
103
away exponentially with increasing velocity. Figure Chapter 3. -9 shows their graph and the
expression that they wrote. Their comments follow.
P
Ae−v
0
v
Figure Chapter 3. -9. Jack and Jim's graph and expression.
Jim
The faster you go, the less you're gonna get wet.
…
Jack
So, you could say the probability if you're traveling at infinite velocity is very extremely
small.
…
Jim
It’s probably- it’s probably if you fit a mathematical model to it, it's probably like an
exponential decaying curve, where you go up to infinity as you near zero velocity. And the
probability of getting hit by a raindrop the faster you go goes down and down but it never
reaches zero.
A number of features of this episode seem to characterize all of the few episodes in which
AWAY
DYING
appeared. First, talk about the limiting case, when the exponent goes to infinity, always
accompanied this form. Second, the use of this form was always accompanied by the drawing of a
graph, similar to Jack and Jim’s. The relationship between physics equations and graphs is an
interesting subject to which, unfortunately, I will not be able to devote significant time.
An interesting point to note about DYING AWAY is that, in a sense, it transcends the limits of
qualitative reasoning. Note that, according to the limits laid out by the qualitative reasoning
researchers, DYING AWAY and PROP- are equivalent—both say that one quantity decreases as another
quantity increases. But the DYING AWAY form is more specific than this brief description; it is
somehow tied to the particular shape of the curve associated with an exponential function, and the
notion of a gradual approach to a limiting value seems to be essential. Perhaps the use of graphs
and the vocabulary of limits serve as crutches that allow people to extend beyond the limits of the
qualitative reasoning system. For example, while certain distinctions may lie beyond the abilities
of qualitative reasoning, it may be easy for us to distinguish between the corresponding graphs. In
any case, I believe that we are seeing the beginnings of the extension of the vocabulary of forms to
include more expert features.
104
ÒPatternsÓ in Arithmetic Word Problems
Before wrapping up this discussion of forms, I want to mention a collection of research that
bears some clear similarities—both superficial and deep—to my symbolic forms-based model of
symbol use. This research comes out of mathematics education, and deals with observations of
how some much younger students solve arithmetic world problems such as the following: John has
five apples and Mary gives him three more, how many does he have?
The research in question is a collection of papers that identify what Greeno (1987) has called
“patterns” in arithmetic word problems. (See, for example, Carpenter & Moser, 1983; Riley,
Greeno, & Heller, 1983; Vergnaud, 1982). There is some variation in the patterns identified by
these researchers, but a single example will suffice here. Riley and colleagues list four categories of
arithmetic word problems, what they call “change,” “equalization,” “combine,” and “compare”
problems. In change and equalization problems, addition and subtraction are described as actions
that lead to increases or decreases in some quantity. More specifically, in change, an amount is
simply added onto another quantity, as in the case of John and Mary’s apples. In equalization, an
amount is added to a base quantity so that it becomes equal to some third quantity: How many
apples must Mary give to John so that he has the same number as Fred?
In contrast, combine and compare problems “involve static relations between quantities.” In
combine problems, there are two distinct quantities that remain unchanged: How many apples do
John and Mary have between them? And in compare problems the difference between two
unchanging quantities is determined.
In addition to additive patterns of this ilk, some researchers also identified a number of
“multiplicative patterns” for problems that involve multiplication and division. Here is one place in
the literature that the “intensive” and “extensive” quantities that I mentioned earlier appear. In
some researcher’s schemes, for example, there are multiplicative patterns that differ according to
whether they involve the products of intensives, extensives, or combinations of intensives and
extensives (Greeno, 1987).
One question we can ask of this research is: What kind of thing are these patterns? It seems
that, in the first iteration, researchers considered the patterns to be categories of problems. Riley et
al., for example, call them “problem types.” Given this orientation toward the status of patterns,
the Patterns research constitutes a particular kind of empirical research strategy. You take the
problems that students have to solve and divide them up into different categories. Then you see
how students perform on the different categories. If there are significant differences across
105
categories then you’ve found something interesting. And researchers such as Riley and colleagues
did find some strong differences in how students performed across category types. For example,
certain types of compare problems were found to be particularly difficult for younger students.
But, as the Patterns researchers were quick to point out, the fact that students perform
differently on the different problem types tells us something about how students are
understanding and solving these problems. This led to some hypotheses in which the patterns
played a role in cognitive models of problems solving. For example, in Riley et al.’s model,
students learn to see arithmetic problems in terms of the patterns, which are associated with
problem schematizations. Then problem solving operators are tied to these schemata.
Clearly, this whole story bears some strong similarities to my own. “Patterns” seem to live at a
similar level of abstraction to my “forms.” And even some specific forms seem similar to patterns.
For example,
PARTS -OF -A-WHOLE
and BASE +CHANGE are similar to Riley’s combine and change
categories.
There are some noteworthy differences, however, and I will begin with the obvious ones. First,
the concerns of the Patterns literature are, in a number of respects, quite different than my own.
For example, the Patterns research looks at much younger students solving arithmetic problems,
not at young adults working on physics problems. There are also some more subtle differences in
concerns. Most notably, the Patterns researchers are concerned with describing problem solving,
how students get from the statement of a problem to a solution. They want to build a model that
has the same problem solving capabilities and goes through the same steps as students. In contrast,
I am interested in describing how students understand expressions and in the role that symbol use
plays in understanding. Thus, I am interested in building a model that can describe, for example,
the interpretations of equations that students give. By way of illustration note that, in the Patterns
literature, patterns are things that are seen in problems, not necessarily in equations. This is because
of the problem solving orientation; in problem solving, you go from problem specification, to
equation, to result.
Now I will move on to the most important and, perhaps, most subtle difference. It is
possible—though not necessary—to read the Patterns literature as implying that all of the possible
word problems that students will ever encounter can be described by a relatively small number of
patterns. This is implied by the fact that these researchers use a relatively simple vocabulary
involving “sets” and a few kinds of actions and relations between sets to describe the patterns. For
example, Carpenter and Moser (1983) say that arithmetic word problems can be characterized
106
according to three dimensions: (1) whether there is “an active or static relationship between sets of
objects,” (2) whether the problem “involves a set-inclusion or set-subset relationship,” and (3) for
the case of problems that involve action, whether the action results “in an increase or decrease in
the initial given quantity.”
This sort of talk suggests a relatively easy to specify space of possibilities. In contrast, the
forms vocabulary that I have observed seems much more idiosyncratic and particular to physics.
Note, for example, that the “set” vocabulary doesn’t even apply to the influences that appear in
the Competing Terms forms. Furthermore, the semantics of influences that appear in these forms
may be somewhat particular to physics.
Taken as a critique of the Patterns literature, what I have just said is a little unfair. The
Patterns researchers are quite self conscious about looking at only a very limited class of problems,
as they are performed by a particular range of young students. It is certainly possible that the
patterns they have uncovered are sufficient to account for their data, and they do not explicitly
argue for the generality of their patterns or their vocabulary beyond this limited range. Overall, I
mean to portray a positive attitude toward this research. And, in fact, it has served as a useful
resource in my own endeavor: I used the patterns discovered by these researchers to aid in my
recognition of forms. But there is a possible trap to be recognized here if we do try to generalize
from the Patterns literature. We must be careful not to assume that the vocabulary and
distinctions that are useful for mathematicians and physicists, such as “sets,” will be the right
distinctions for describing student or even expert knowledge.
Conclusion: How Do I Know That Forms Are Real?
How do I know that forms are real? In this chapter, I have described symbolic forms and
listed the various kinds of forms that appeared in my data corpus. I have presented examples to
illustrate particular aspects of the hypotheses associated with positing the existence of forms, and I
have contrasted my view with alternatives and competitors. Still, I have not presented all of my
evidence that forms do exist and are important. Most importantly, I have not described the
details of how I make contact with my data corpus. This connection to the data will not be
described until Chapter 6, when I will explain how, through repeated iterations of systematic
analysis of the entire corpus, I arrived at the particular list of forms described here. In addition, I
will present specific results concerning the pervasiveness of the various types of events described
here.
107
However, I believe that the content of this chapter contains many elements that can contribute
to an argument for my viewpoint. First, I have argued against some important competitors; most
notably, what I referred to as the principle-based schemata model and other models that propose
that physics knowledge is closely tied to physical principles. Some of the arguments against these
competitors were relatively straightforward. As we saw, these competitors are not constructed with
an intent to deal with some of the phenomena that I am treating as central, namely the
interpretation of expressions and the construction of novel expressions.
In addition, I believe that the large number of examples presented in this section can add
significantly to the plausibility of the view that I am describing. We saw, of course, that students
can and do construct and interpret equations, and I tried to give a feel for what these behaviors
look like. Furthermore, many of the examples I presented were quite suggestive of the types of
phenomena I hypothesize to be associated with forms as knowledge elements. Here is a list of
some of these suggestive examples:
•
Mike and Karl, two students that demonstrated great prowess in their ability to construct
novel expressions, did not know that the equality of forces in the spring problem followed
form F=ma.
•
We saw that students do not always work by carefully and straightforwardly applying
principles. Instead, principles are often used as a way of rationalizing expressions after they are
already composed.
•
To explain why, in the Buoyant Cube problem, the weight of water displaced had to equal
the weight of the cube, Mark argued simply from the observation that, in his experience with
problems of this sort, you “have to have something equal.”
•
We saw that students are not always capable of choosing between certain kinds of related
expressions. For Mike and Karl, F = kx and F = 1 2 kx 2 are “qualitatively” equivalent.
•
Similarly, we saw that the equation v = vo + at was “obvious” for students, but only in a
certain way. How can students, in the same breath, say that v = vo + at is obvious, but that
they can’t be sure of the ‘at’ term? If we say that students are understanding this equation in
terms of the BASE +CHANGE form, then this phenomenon makes sense.
•
While working on Running in the Rain, Alan and Bob composed an expression with
“placeholders.”
108
The point is that these surprising phenomena are well accounted for by the forms hypothesis. It is
part of this hypothesis that students have knowledge for constructing equations that cuts across
physical principles. And it is an essential property of forms that certain features of equations
become consequential at some times, and not at others.
Most of all, I hope that the presentation in this chapter is plausible to some readers because it
maps a territory that they recognize, a world that they have lived in. The idea of this chapter has
been to do part of the job of describing the meaningful structure that a physicist and physics
student see when looking at a sheet of equations. They do not see a meaningless mass of details or
only a collection of individually recognizable symbols and equations. Like Gombrich’s painters,
they bring to a sheet of expressions a known “vocabulary of forms.”
109
Chapter 4. Representational Devices and the
Interpretation of Symbolic Expressions
x=
mg
k
“As you have a more massive block hanging from the spring, then you're position x is gonna
increase, which is what this is showing.” In this statement, Jim tells us what this equation “shows.”
But how did he conclude that this is the content of this equation? I have already argued that
symbolic forms are one piece of the answer to this question. In this statement, Jim announces a
proportionality relation between position and mass, which suggests that he is seeing the
PROP+
form. And the PROP+ form allows Jim to make various inferences, including the inference that, in
this case, increasing the mass will increase the position.
However, important elements of the story are still missing. The PROP+ relation between
position and mass is only one of the possible ways that a form can be seen in this equation. It
would be helpful if we could say something about what factors lead to Jim seeing this particular
form. Furthermore, Jim expresses this form in a very specific manner. He does not just say that x
“is proportional to” the mass, instead he talks through the specific case in which the block becomes
“more massive.” What does a “more massive” block have to do with this expression? Why does he
choose to talk about a more massive block rather than a less massive one?
Clearly a number of important aspects of Jim’s interpretation remain to be accounted for by
my model. One thing that might help here is a better understanding of what it means to
“interpret” an equation. If we understood, in general, what it means to interpret an equation, then
we might be in a position to explain the choices that Jim makes in his statement. The hope is that
we could find some general procedure for generating interpretations of equations, and then we
could view Jim’s interpretation as an application of this general procedure.
Let’s take a moment to work on what an interpretation of an equation could be. Intuitively, we
might say that to interpret something is to state its “meaning.” If this is right, then the
construction of a “theory of meaning” would be a significant first step toward understanding the
nature of interpretation. This might not seem like a very helpful observation. Why should it be
any easier to invent a theory of meaning than a theory of interpretation? However, it happens that
some potentially useful theories of meaning already exist. In particular, philosophers and linguists
110
have worked out theories of meaning for natural language sentences, so we could begin with these
existing theories and try to adapt them. Although I will ultimately adopt an alternative approach,
I want to take a moment to explore this possibility.
One type of theory of meaning involves what is called “model-theoretic semantics.” Modeltheoretic semantics associates the meaning of a sentence with its “truth conditions”—the situations
in the world, understood in terms of a certain class of models, in which the sentence is true. Of
course, employing such a definition requires that we have a way of determining, from a sentence,
what its truth conditions are. In model-theoretic semantics this works as follows: Individual words
and predicates get their meaning by referring to things in the world. Then the truth conditions of
a sentence are a function of the arrangement of words in the sentence and the referents of these
individual words.
To take this discussion a little closer to home, some mathematicians and philosophers have
tried to provide foundations for mathematics that employ a version of model-theoretic semantics
called set-theoretic semantics. They begin with a collection of axioms that are expressed in terms
of what are essentially meaningless symbols. Proofs are then performed by using the rules of logic
to deduce new statements from these axioms. All of these statements are still in terms of
meaningless symbols. The idea is that set-theoretic semantics specifies how the statements can be
given meaning, of a sort. The symbols in the axiomatic system are put in correspondence with a
model that consists of sets and various kinds of relations among sets. (See Lakoff, 1987, for a
relevant discussion and critique of set-theoretic and model-theoretic semantics.)
For either of these cases—for model-theoretic semantics or the special case of set-theoretic
semantics in mathematics—“interpretation” can be understood to be the announcing of the truth
conditions of a symbolic statement. So, to understand what an interpretation of an equation is, we
could embark on working through the details of this program for the case of physics equations.
This would mean that we would have to define a class of models—like the set-based models of
set-theoretic semantics—and then rules by which physics equations are put in correspondences
with circumstances in these models.
This could work as follows: We begin by defining a method for building models of physical
circumstances. These models could be something like Larkin’s “physical representation;” they will
include point-masses, forces, velocities and other entities from the physicist’s palette. Then
individual symbols, such as the letter ‘F’, can be put in correspondence with entities in these
models. In this case, ‘F’ might correspond to a force in a particular circumstance. Finally, the
111
arrangement of symbols in the equation somehow specifies a relation among these entities that
must hold. The set of circumstances in which this relation holds stated in the modeling
language—its truth conditions—constitutes the “meaning” of the equation.
Can such a project, in principle, be made to work? I believe that the project to invent a theory
of equation meaning is a useful and interesting project, and it is for that reason that I have
followed it this far. However, there is one important difficulty with this approach from the point
of view of the current work. Even if we can construct an internally consistent theory of equation
meaning, this theory may not be useful for describing students’ interpretive utterances or the
knowledge that generates those utterances.
Thus, although I will make some contact with theories of equation meaning at the end of this
chapter, this is not the approach I will adopt here. Rather, I propose to adopt an altogether
different approach, an empirical investigation of what, in practice, constitutes an interpretation of
an equation. In many ways this is an easier—if more tedious—project than the philosophical, a
priori approach. To execute this empirical project, I can begin by looking through my data corpus
for any utterance that sounds remotely as though it could be called an interpretation of an
expression. Along with Jim’s above interpretation, I have already presented many examples of
utterances that fall into this category:
µmg=ma
Alan
What Bob didn't mention is that this is when the two forces are equal and the object is no
longer moving because the force frictional is equal to the force it received.
________________
s=
Jack
ρi
L
ρw
This does seem to make sense cause if ice had the same volume as water [g. ρi then ρw],
er, same density as water, then L would equal S which means the cube would be right at
the surface.
________________
F = kx ; F = 1 2 kx 2
Karl
Okay, now, qualitatively, both K X and half K X squared do come out to be the same
answer because as (…) looking at it qualitatively, both half - both half K X squared and K
X, um, you know, increase as X increases.
________________
total rain(t) = #rain
t + ρrain hwx
s ( )
Alan
You're just gonna- (well) based upon that equation right there you're just gonna conclude
that you would feel more - you would have more raindrops hitting you if you walk because
112
it's gonna take you a longer time since the last term is just a constant. And you're just
saying it just depends on the time.
In a sense, such a list of examples constitutes a preliminary answer to my question; if you want
to know what an interpretation of an equation is, then look at these examples. Of course, an
answer of this sort would be far from satisfying, and I intend to do better. One obvious next step
would be to look for regularities that span these instances. For example, if every interpretive
statement seemed to involve varying some quantity and holding other quantities fixed, as in Jim’s
interpretation, that would be an important observation. Failing the existence of such universal
regularities, a further step would be to group interpretive utterances into categories that do seem
to share features. Then we could characterize the within-category regularities.
This last approach, the grouping of observed interpretive utterances into categories, is roughly
the tactic that I will employ. However, since this is a cognitive science-style inquiry, I will not talk
about these categories as simply constituting observed regularities in student discourse. Rather, in
the manner of cognitive science, I will explain these regularities in terms of mental structures,
specific abilities possessed by individuals. In particular, I will associate each of these categories of
utterances with an interpretive strategy that can be known by an individual. I call these interpretive
strategies “representational devices.” Here, the use of the term “device” is supposed to call to
mind “literary devices.” We will see why in a moment.
Representational devices are the second major theoretical construct in my model of symbol
use in physics, and a discussion of devices is the main concern of this chapter. Devices are to be
understood as a type of knowledge, and, as knowledge, devices constitute an important part of
what physics students must learn on their way to expertise. In the last chapter, we explored the
structure—a meaningful structure—that a physicist sees in a page of expressions. The point of the
present chapter is that, to enter into this meaningful world of symbol use, it is not sufficient to
know symbolic forms. In addition, initiates in physics must possess the complex repertoire of
interpretive strategies that I call “representational devices.” With forms, this repertoire of devices
defines the character of the meaningful symbolic world of physics.
Two very different examples
I will start by looking a little more closely at two examples that we have already encountered.
First, here is Jim’s interpretation of the spring problem result, reproduced once again in full.
113
x=
Jim
mg
k
Okay, and this makes sort of sense because you figure that, as you have a more massive
block hanging from the spring, [g. block] then you're position x is gonna increase, [g.~ x in
diagram] which is what this is showing. [g.~ m then k] And that if you have a stiffer spring,
then you're position x is gonna decrease. [g. Uses fingers to indicate the gap between the
mass and ceiling] That why it's in the denominator. So, the answer makes sense.
Jim’s statement involves two separate interpretations. In the first of these interpretations, Jim
describes what the equation says will happen to the position if the object becomes more massive.
In the second interpretation, he says what will happen if the spring is made stiffer; that is, k is
increased. It does turn out that a certain percentage of students’ interpretive utterances do involve
an imaginary process of this sort, in which some quantities are held fixed and one quantity is
varied. I call the device involved in such interpretations CHANGING PARAMETERS. Both of the
interpretations in Jim’s statement involve the use of this representational device.
Now we move on to a second instance that we have already encountered. Recall that at the
end of Alan and Bob’s work on the Shoved Block problem, Alan made what I took to be an
inappropriate statement concerning an equation. In particular, he incorrectly maintained that a key
equation stated a condition that must be satisfied for the motion to terminate:
µmg=ma
Alan
What Bob didn't mention is that this is when the two forces are equal and the object is no
longer moving because the force frictional is equal to the force it received.
At this point, it is not especially relevant for the discussion at hand that this interpretation is
inappropriate. What is important is that this interpretation is very different than Jim’s
interpretations of the spring equation. Note that it does not seem correct to describe Alan’s
interpretation in terms of quantities that are varied and quantities that are held fixed. Rather, Alan
projects this equation into a particular moment in the motion. I refer to the device that Alan
employs here as the SPECIFIC MOMENT device.
Jim’s and Alan’s interpretations are typical of two of the major classes of representational
devices, what I call “Narrative” and “Static” devices. I will not explain these major types further at
the moment; for now, I simply want to note that there are interpretations that seem to be of
significantly different types. Thus, we can already see that my project of putting interpretive
utterances into categories is not going to end up with all utterances in the same category.
114
On the form-device relation
I opened this chapter by wondering what factors determine which forms are seen in a given
equation, by a given individual at a particular time. I want to now discuss the role that
representational devices might play in determining the forms engaged. In this vein, I begin with a
hypothetical story concerning Jim’s interpretation of the spring equation, a story that I will
ultimately question:
1. Jim starts with the CHANGING PARAMETERS device and looks to see the effects of increasing the
mass.
2. This draws his attention to the ‘m’ variable, highlighting its location in the equation and
causing other features of the equation to be less salient. This leads to the
PROP+
form being
engaged.
3. Then PROP+ (with IDENTITY) allows the inference that, as mass increases, position increases.
The key feature of this story is that the choice of representational device comes first. The device is
chosen and then this choice can be said to be determining the symbolic form seen by Jim.
Similarly, in Alan’s interpretation, the SPECIFIC MOMENT device may lead Alan to see BALANCING
in the equation. Perhaps the choice to see the equation as describing a particular moment in the
motion causes Alan to “look for” BALANCING in the equation. However, there is another very
plausible alternative to this story: The form could come first and determine the device, rather than
the reverse. Perhaps, for Alan, the equation in question strongly cues a BALANCING interpretation.
This is especially plausible since, as I have said, students are predisposed to overuse BALANCING .
So which of these stories is correct? Do forms determine devices or devices determine forms?
My answer to this question is that neither story is always correct; I believe that there are multiple
ways that the process of interpretation is structured, including the two possibilities described
above. However, I will not undertake a description of these processes in this thesis. Instead, for the
sake of simplicity, I propose to adopt a more heuristic stance in which forms and devices are seen
as elements in a system of constraints that, working as a whole, adjusts itself until the system is
satisfied and an interpretation is produced. In adopting this model, I am intentionally glossing
some of the details of the interaction between forms and devices.
As shown in the simple schematic in Figure Chapter 4. -1, in addition to forms and devices,
the system includes the equation itself and what I have simply called the “rest of the context.” The
115
point of including the “rest of the context” is that the form and device engaged must depend on
considerations of the moment, such as what the question asks. This issue will be a major focus of
the final sections of this chapter.
symbolic expression ⇔ form ⇔ device ⇔ rest of the context
Figure Chapter 4. -1. The form-device relation.
Let me take a moment to clarify the roles that these two types of knowledge play in this story.
As shown in my schematic, forms are most closely connected with equations; they involve directly
seeing patterns in equations. Furthermore, everything to the right of forms in my schematic plays
a role in constraining what forms are seen in the equation. Now, I could have chosen to just wrap
everything to the right of forms in a black box and say that I am not going to explain it further in
this study. But I have attempted to do just a little bit better. My model includes an additional
layer between forms and the rest of the context, what I have called representational devices.
Representational devices are interpretive stances and strategies that more broadly characterize an
orientation to an equation, and thus influence the forms seen in a symbolic expression.
Representational devices and the construction of equations
I want to clarify one final point before diving in and systematically discussing each of the
types of representational devices. The schematic shown in Figure Chapter 4. -1 is designed to
apply to the construction of new equations, as well as to the generation of interpretive utterances
concerning equations that are already written. Notice that the arrow that connects forms to the
equation points in two directions. The idea is that, just like the other elements of the system, the
equation can change as the system adjusts itself.
All of this implies that the construction of equations involves representational devices. To
illustrate, on the way to composing the novel equation,
µ
µ = µ 1 + C m2
Karl commented that:
Karl
I guess what we're saying is that the larger the weight, the less the coefficient of friction
would be.
The thing to notice here is that a CHANGING PARAMETERS device is built into Karl’s specification for
this equation. In an important sense, equations are constructed with interpretations in mind.
116
The Variety of Devices
In this section, the project of categorizing interpretive utterances and listing representational
devices moves into full swing. As shown in Figure Chapter 4. -2, this discussion divides the
representational devices observed into three “classes:” Narrative, Static, and Special Case. I will now
discuss each of these three classes in turn.
Narrative
Static
CHANGING PARAMETERS
SPECIFIC MOMENT
PHYSICAL CHANGE
GENERIC MOMENT
CHANGING SITUATION
STEADY STATE
Special Case
STATIC FORCES
RESTRICTED VALUE
CONSERVATION
SPECIFIC VALUE
ACCOUNTING
LIMITING CASE
RELATIVE VALUES
Figure Chapter 4. -2. Representational devices by class.
The Narrative Class
Representational devices in the Narrative Class embed an equation in an imaginary process in
which some type of changes occur. As a whole, I call these devices “narrative” because they often
lead to interpretive utterances that are like a very short story or short argument. It is because of
these devices—and the fact that they constitute the most common class—that devices deserve
their name. The use of the term “device” is designed to suggest that representational devices are
like the literary devices for the very short stories in narrative interpretations. The devices in this
class are listed in Table Chapter 4. -1.
CHANGING PARAMETERS
PHYSICAL CHANGE
CHANGING SITUATION
Table Chapter 4. -1. Devices in the Narrative Class.
The analysis I will present in Chapter 6 will suggest that the CHANGING PARAMETERS device was
the most common of all representational devices in my data corpus. In CHANGING PARAMETERS,
some parameter is picked to vary and other parameters are held fixed. As I already mentioned,
Jim’s interpretation of x=mg/k involved the use of this device.
117
CHANGING PARAMETERS is frequently applied as a way to get at the properties of more elaborate
expressions. For example, during their work on the Running in the Rain problem, Jon and Ella
arrived at the following expression for the total rain that strikes a person as they run across a street.
In this expression, vR is the running velocity, how fast the person runs across the street.


h
w
 # drops  
+
=
wx
vd2 v R 
 A⋅s  
 v d 1+ 2

vR




Extracting meaning from a complicated expression of this sort can be quite difficult. By way of
interpretation, Ella tried to determine what this expression says about how the total rain will
change if the speed increases, that is, if the person runs instead of walking across the street.
Ella
Okay, V R is bigger. Then (10.0),, This is gonna get smaller, [w. arrow pointing to second
term] but this is gonna get bigger. [w. arrow pointing to first term] I think.
Jon
Yeah, I think - that gets smaller [g. second term] cause V walking is smaller than V
running.
Ella’s interpretation is based on the CHANGING PARAMETERS device; she imagines that v R increases
and that all of the other parameters are held fixed. The result, in this case, is somewhat
inconclusive. She infers that the second term will get bigger but that the first term will get smaller.
I will have more to say about these inconclusive outcomes later in the chapter.
The CHANGING PARAMETERS device is not only used as a way of extracting information from
final expressions; it is also used to check the sensibility of equations that constitute intermediate
steps in a solution. For an example, I return to Mike and Karl’s expression for the acceleration of a
dropped ball under the influence of gravity and air resistance:
a(t) = −g +
Karl
f (v)
m
Okay, this sign is negative [g. -g] this number [g. f(v)/m] as the mass- as you increase the
mass,, Well, one problem- from one ball to the other the mass is doubling. Right, so you
double this mass [g. m] so this number [g. f(v)/m] is getting smaller.
Here, Karl’s interpretation involves varying the mass and holding the other parameters fixed.
More specifically, Karl imagines that the mass is doubled. Such a choice is not very surprising since
the task under consideration—the Air Resistance problem—asks the students to compare the
118
terminal velocities of two masses, one of which is twice as heavy as the other. This occasional
dependence of device choice on the task will be a major focus of the final section of this chapter.
I have yet to comment on one very important feature of CHANGING PARAMETERS. In the
CHANGING PARAMETERS
device, the changes in question are not changes associated with a motion
under consideration. For example, in Karl’s interpretation above, the process he imagines involves
the mass doubling. This is not a process that takes place through the temporal sequence of the
motion, as the ball drops. The same is true of Ella’s interpretation. She is not telling the story of a
single individual running across the street and what happens along the way; rather, she is talking
about two separate instances in which someone runs across the street, but with different speeds.
There are, in fact, interpretations in which an equation is placed in the process of the motion
itself, and which concern changes in quantities that do occur through the temporal sequence of a
motion. I call the device associated with this separate category of interpretation the
CHANGE
PHYSICAL
device. Bob’s interpretation of this expression for the force of air resistance is an example:
Fair = kv
Bob
Okay, and it gets f-, it gets greater as the velocity increases because it's hitting more
atoms (0.5) of air.
The key thing to note about this little example is Bob’s statement that this happens “as the
velocity increases.” Just as in a CHANGING PARAMETERS interpretation, this interpretation involves a
quantity that is increasing, but in this case it is increasing through the time of the motion itself.
Later on in the same task, Alan provides us with another nice example involving PHYSICAL CHANGE :
∑F = mg - kv
Alan
And so I guess we were just assuming that this was [g. kv] de-accelerating - that we would
have gravity accelerating the ball and then as, as we got a certain velocity, [g. kv] this
would contribute to making accelerating closer - de-accelerate it and so we would just
have a function of G minus K V. And then we assumed that since we know that it reaches
terminal velocity, its acceleration's gonna be zero, we just assumed G minus K V would
equal zero.
Roughly, the story in Alan’s interpretation is that gravity starts the ball accelerating. Then, as the
ball speeds up, the kv term grows and grows until it eventually cancels mg and makes the
acceleration zero. Thus, the equation is placed in the story of the motion and the interpretation
tells us what happens to the equation as the ball falls.
So, interpretations based on PHYSICAL CHANGE involve changes in parameters that actually vary
during the motion itself. In contrast, in CHANGING PARAMETERS, an equation is placed in a story in
which one of the parameters that affects the motion, viewed as a whole, varies. These are
119
parameters that, like the mass, do not vary during the time of the motion itself. Note that the
changes induced by varying these parameters are typically changes in degree only, the motion is
still motion of the same sort and the same influences and entities are involved.
In the final device in this class, CHANGING SITUATIONS , an equation is given meaning by
embedding it in a comparison between the current situation and a situation that is fundamentally
different; that is, it differs by more than simply the values of the parameters that define the
motion. Recall, for example, that part (b) of the Air Resistance task asked students to consider
how the terminal velocity of a dropped object would change if there was a wind blowing straight
up or straight down. The addition of the wind is precisely the type of change I have in mind here
and, in fact, this problem did tend to engender the use of the changing situations device. In solving
this problem, students typically obtained an expression for the terminal velocity which was the
same as the expression obtained without wind, but had an additional term. Close to every pair
gave a CHANGING SITUATIONS interpretation of their version of this expression.
vt =
Karl
mg
− vw
k
Which is what I was saying [g. final expression] that the terminal velocity will be the velocthe terminal velocity in still air plus the velocity of the wind.
________________
vt =
Bob
m
g − vo
k
If the wind is blowing up, the velocity - terminal velocity is going to be less than it was in
this case. [g. vt=m/k g] I mean, we have the same mass. This is M K G. [w. vt=m/k g; draws
a box around it] This is no wind. [w. "no wind"] No wind. And this is with wind. [w. "w/ wind"
and boxes wind result] Okay, so if the wind is blowing up, the terminal velocity is gonna be
less.
________________
vt =
Roger
(mg − Fw )
µ
Okay, well for the up case - yeah, terminal velocity will be (0.5) less [g. -Fw] by a factor F W
over mu.
In each of these examples, the student’s interpretation of their expression is framed by a
comparison to the case in which there is no wind. The adding or taking away of the wind is a
change in the situation of the sort that is associated with the use of the CHANGING SITUATION device.
120
The Static Class
As we have just seen, Narrative devices embed an equation in a very short story that involves a
hypothetical process of change. In contrast, devices in the Static Class project the equation into a
static situation. Rather than playing a role in a story, the equation is treated more like a snapshot
of a moment. Furthermore, Static devices are often used in close concert with diagrams, which
can also be used to depict instants in a motion. The devices in the Static Class are listed in Table
Chapter 4. -2.
SPECIFIC MOMENT
GENERIC MOMENT
STEADY STATE
STATIC FORCES
CONSERVATION
ACCOUNTING
Table Chapter 4. -2. Devices in the Static Class.
Alan’s interpretation of the equation µmg=ma is an example of an interpretation involving the
SPECIFIC MOMENT
device.
µmg=ma
Alan
What Bob didn't mention is that this is when the two forces are equal and the object is no
longer moving because the force frictional is equal to the force it received.
In specific moment interpretations, an equation is treated as describing one particular moment in a
motion. As I have mentioned, in Alan’s interpretation above he associates this equation with the
particular instant in which the shoved block comes to rest.
SPECIFIC MOMENT is often employed (correctly) to describe a moment when two forces are in
balance. This was frequently the case in students’ work on the Air Resistance task, where this
device was used to treat an equation as describing the moment when the force of air resistance has
just become equal to the force of gravity. In the following excerpts, note the appearance of some
key suggestive phrases, such as “at some time” and “that’s when.”
R = mg
Roger
It means that the velocity won't increase anymore.
Mark
Okay.
Roger
That's after a certain point.
Mark
Oh, that's when the air resistance equals the=
Roger
=Yeah!
Mark
That's when the forces are equal then, right?
121
Roger
Yeah, because then you wouldn't have acceleration anymore. [g. g in mg]
Mark
Right. (…) When the force is zero the acceleration's zero.
Roger
Okay. [w. R=] I guess. Okay. After a certain time
Mark
R equals G.
Roger
At T, some T.
________________
Cv 2 = mg
Ella
The terminal velocity is just when the - I guess the kind of frictional force from the air
resistance equals the gravitational force?
In both of these excerpts, the equation is associated with the specific moment when the two forces
are equal and the terminal velocity has been reached.
I have stated a number of times that Alan’s interpretation of the equation µmg=ma is
inappropriate since it implies the existence of a second force that does not exist in the standard
account. Now I would like to deal with the question of what would constitute a more appropriate
interpretation of this expression. An alternative stance that can be taken toward this expression is
that it is not just true at the end of the motion, it is true at any time along the entirety of the
motion of the shoved block. More specifically, the equation can be thought of as providing a
means of computing the acceleration at any time along the motion. When we see an equation in
this manner, as true at any time, then we are employing the “GENERIC MOMENT ” device.
A very typical example of this device can be found in Mike and Karl’s work on the Air
Resistance task, when they composed an equation for the acceleration of a dropped ball. In the
following passage, note their use of the phrase “at any time.”
a(t ) = −g +
Mike
f (v)
m
So, at least we can agree on this and we can start our problem from this scenario. [g.
Indicates the diagram with a sweeping gesture] Right? Okay? So, at any time,, At any
time, the acceleration due to gravity is G, and the acceleration due to the resistance
force is F of V over M. [w. g + f(v)/m]
…
Mike
You have negative G plus F of V over M [gestures down and then up]. That's you're
acceleration at any time. Right?
Karl
Well, wait. You want them to have,,
Mike
This is the acceleration at any time T. [w. a(t)=]
In this episode, Mike and Karl composed an expression for the acceleration that is valid at any
time as the ball drops. It is true at the moment the ball is released, as the ball speeds up, and even
after it has reached terminal velocity.
122
At first blush, it may seem a little strange that it is possible to describe a “generic” instant in
such a varying sweep of events. Indeed, the ability to provide such descriptions—to find the level
of generality from among the particulars—is one of the great powers of physics, and the ability to
create and interpret such descriptions is one of the skills that physics students must learn.
In fact, there are a number of supports built into the representational practice of introductory
physics that aid students in constructing equations that are to be interpreted in the manner of the
GENERIC MOMENT
device. The ubiquitous “free-body diagram” can be seen as an example of a
support of this kind. In learning to use free-body diagrams, students are taught to isolate an object
and to draw all of the forces that act on that object. Mike and Karl’s diagram for the air resistance
task, which I presented in Chapter 3, is an example (refer to Figure Chapter 4. -3). In acting out
the set of procedures associated with the use of a free-body diagram, students are led to produce
equations that are true at any time during the motion, and which are amenable to GENERIC
MOMENT interpretations.
f(v)=
Fa
v m
Fg
Figure Chapter 4. -3. Mike and Karl’s air resistance diagram.
Readers familiar with the practice of introductory physics will be well aware that the use of
free-body diagrams to construct equations of this sort is extremely common. I will present just
one additional example here. When working on the spring task, Alan and Bob set the sum of the
forces equal to ma, saying that this described the spring at each point along its oscillation:
ma = -ky + mg
Alan
What's gonna happen is that, at most points, except for I believe this point right here,
[g. rest position of spring] you're gonna have a net force that's gonna cause the spring to
accelerate. [g. ma] Right? So, I'm gonna set this up as this equation [w.mod ma=-ky+mg]
The above diagram-oriented work is typical of most cases involving the
GENERIC MOMENT
device. There are some cases, however, in which a student simply emphasizes that a given relation
is true at any time along the motion and can be used to compute results that are true for any
particular time. For example, Mark and Roger manipulated the relation v = at to obtain an
expression for t and then commented:
123
t=
v
a
Mark
Oh, it gives us (st…) say we want to find the velocity, [g. v] we can if we know the time. [g.
t] Or, vice-versa, if we know the velocity we can find how far into the free fall it =
Roger
= I mean this is a relationship between p- T and V. [g. t, v]
As Mark and Roger emphasize, this relation is true at any time. You can pick a time that you want
to know about and get a velocity, or you can substitute a value for the velocity and compute the
associated time.
The next device in this cluster, STEADY STATE , is a companion to the above two. There is a
special case to be considered here, the case when there is no distinction between
GENERIC MOMENT
and SPECIFIC MOMENT . This will be true when no parameters of the system vary with time, as in the
case of the spring problem. Consider the following two examples:
kx = mg
Jack
So, okay, so then these two have to be in equilibrium those are the only forces acting
on it. … So I guess we end up with K X is M G. [w. kx=mg]
________________
1 kx 2
2
Mike
= mg
Now, when it’s hanging, you have the force from the spring pulling it up and the weight
from the mass pulling it down. And those two forces would have to equate [g. the force
diagram], and you’ve got that right here. [g. 1/2kx2=mg]
These examples can be seen to share aspects of both GENERIC MOMENT and SPECIFIC MOMENT . The
expressions are true at any time, but such a fact is trivially true since nothing about the system
varies with time. And, in fact, these examples can be seen to have features I have associated with
SPECIFIC MOMENT .
For example, there are balanced forces involved. And note Mike’s use of the
word “when” in the phrase “when it’s hanging.” Thus, I consider both of these examples to be
instances of a third device, the STEADY STATE device.
The STATIC FORCES device is designed to account for another kind of ambiguity. Above, I
argued that we could think of free-body diagrams as a representational support for the
construction of GENERIC MOMENT expressions. Students also learn to draw diagrams of this sort that
depict a particular moment in a motion, and that depict steady state situations. The problem is
that, in many cases, although it may be clear that students are projecting an equation into some
static situation involving forces, it may not be clear which of the three static situation the student
has in mind. Furthermore, there is no reason to believe that the student “intends” any of the three.
Consider, for illustration, the following example:
124
∑F = mg - kv
Bob
There's two forces. Okay. M G. And then there's a force upward which is dependent on
velocity, K V. [w. a little diagram showing up and down forces] Alright, so the sum of the
forces is equal to that. [g. ∑F equation]
I believe that there is a relatively well-developed nexus of equation construction activity here,
which takes on a kind of life of its own. Students spend a lot of time drawing diagrams with
forces, then building equations from these forces. In the STATIC FORCES device, it is as if an
equation is projected into one of these diagrams, rather than into a motion or physical situation.
Because these diagrams may describe the sort of situations associated with SPECIFIC MOMENT ,
GENERIC MOMENT , OR STEADY STATE, STATIC FORCES
interpretations cannot be treated as belonging
to one of these other categories.
Now on to the fifth device in this class, what I call the “CONSERVATION” device. The
CONSERVATION
device is named after its most common situation of use, problems that involve the
application of conservation principles. It is similar in some respects to SPECIFIC MOMENT , except
that each side of the equation is associated with a different moment in the motion. As an example,
I want to present some student work on a problem that I have not discussed much to this point,
the problem in which a ball is thrown straight up and students are asked to find the maximum
height reached. In Chapter 2, I briefly described Jon and Ella’s work on this task, with emphasis
on their discussion of whether a numerical result was reasonable. In their solution, Jon and Ella
employed the memorized equation vf2=vo 2+2ax.
In contrast, Alan and Bob used the principle of “conservation of energy” to solve this problem.
This principle tells Alan and Bob that the total energy of the ball is the same at any point along its
motion. They also know some other information that helps them here. They know that the total
energy at any moment is the sum of the “kinetic energy” and the “potential energy.” Furthermore
they know expressions for the kinetic and potential energy:
kinetic energy = 12 mv 2 ; potential energy = mgh
These expressions give the kinetic and potential energy in terms of the mass, m, the height, h, the
speed of the ball, v, and the acceleration due to gravity, g. With these relations and the
conservation of energy in mind, Alan and Bob assembled the following expression:
125
mgh = 12 mv 2
Alan
Because at the top - we take the initial energy which is only kinetic energy. Potential
energy at the height of zero on the planet we arbitrarily set that to zero. So we only have
kinetic energy which is one half M V squared. [g. 1/2mv2 ] At the top we're not - the
velocity's going to be zero. [g. top of graph] It's going to be standing still for a split
moment. So we only have potential energy that's only gonna be M G H.
In this expression, Alan equates the potential energy at the maximum height with kinetic energy at
the moment the ball is released. This makes sense because, as Alan argues, the total energy at the
top consists only of potential energy and the initial energy is purely kinetic. The key point to note
here is that the left side of the equation is associated with a particular moment in the toss, the
instant when the ball reaches its apex, and the right side is associated with the instant that the ball
is released. Each side is thus associated with a different moment in the motion (refer to Figure
=
0
Chapter 4. -4). This is the CONSERVATION device.
Etop = 12 mv2 + mgh = mgh
=
0
mgh = 12 mv 2
Einitial = 12 mv2 + mgh = 12 mv 2
Figure Chapter 4. -4. Solving the tossed ball problem using conservation of energy. The instant of release and
the instant when the apex is reached are each associated with a side of the equation.
The last device in the Static Class, ACCOUNTING, is of a somewhat different sort than the other
devices in this class. In applying this device, a student imagines that an equation provides an
accounting of all of some entity or substance. Although I did not find that this device was
especially common in my data corpus, it is associated with what might be the most stereotypical
stance that can be taken toward an equation: We imagine that the job of an equation is to tell us
“how much” or “how many.”
The ACCOUNTING device belongs in this class because interpretations based on ACCOUNTING can
be thought of as describing a static situation. In fact, like the other elements of this class it tends
to be closely tied to depictions in diagrams. To elaborate, when students use free-body diagrams
to construct GENERIC MOMENT or SPECIFIC MOMENT expressions, they walk around the diagram,
126
using it to systematically enumerate each of the forces that are acting. When expressions are being
constructed using
ACCOUNTING,
the diagram plays a similar role; it can help to systematically
account for all of the entity or substance that the equation is supposed to be computing.
The context provided by the Running in the Rain problem consistently engendered use of this
device. Alan and Bob’s work on this task, which we have already encountered, was typical. In this
example, note the systematic listing of “sources” and Bob’s use of the phrase “how much.”
total rain =
Bob
# raindrops
+C
s
There's two - two sources of rain. Okay, rain that you walk into and rain that drops on
you. [g. rain hitting his head] Okay. Walk or run into. Okay, so, the number of rain per
second [g. #raindrops/s] is how much rain hits your head. … if you're a crate and you you move, move into some rain.
Here, Bob’s stance is that this equation is doing the “accounting” job, it tells us how much rain
there is. Jon and Ella made a similar accounting of the rain that hits your head and face as you run
through the rain:
# drops   xwh w 2 x 

+
# drops total =

 A ⋅ s   vd
vr 

Ella
And uh this term [g. first term]. This is from the drops that hit you on your face. And this
term [g. second term] is from the drops that hit you on your head.
The Special Case Class
For interpretations based on all of the above devices, the statements made are presumed to be
true for any values of the quantities that appear in the expression. For example, when Jim says that
the expression x=mg/k implies that when m increases, x increases, this observation is assumed to be
true independent of the values of the quantities involved. In contrast, in Special Case devices,
conclusions are drawn for cases in which the values of quantities that appear in an expression are
somehow restricted. The devices in this class are listed in Table Chapter 4. -3.
RESTRICTED VALUE
SPECIFIC VALUE
LIMITING CASE
RELATIVE VALUES
Table Chapter 4. -3. Devices in the Special Case Class.
To illustrate, consider the following example from Alan and Bob’s work on the Mass on a
Spring problem. In this example, Alan checks the sensibility of the expression F = -ky that he has
127
written for the force applied by the spring. In his interpretation, Alan refers to a diagram he has
drawn, which I have reproduced in Figure Chapter 4. -5. The dotted line in the diagram marks
the location where y is equal to zero, the rest position of the spring in the absence of any applied
forces. Locations above this line are positive and locations below are negative.
y=0
m
Figure Chapter 4. -5. A diagram from Alan & Bob's work on the spring problem.
To check the sensibility of his equation for the force, Alan considers cases where the position
of the spring, y, is positive, negative and zero:
F = -ky
Alan
If you look at this point [g. point where dotted line intersects spring] there'd be no force
due to Hooke's law, and that's where it'd remain at rest. If you're looking above this axis
[g. dotted line then above dotted line] you're looking at a negative force so that's gonna
push it down. So, if you get closer to the ceiling, the spring force is gonna push you down.
If you're below this axis um you're going down but now you're gonna have a force because you're gonna have two negatives [g. -ky term] you're gonna have a force that's
sending you back up towards the ceiling. So it's just a question of signs.
In this interpretation, Alan concludes that his expression for the spring force makes sense because
his sign analysis yields forces that are in the correct directions. For example, for positions below
the y=0 line, y is negative. Substituting a negative value for y yields a positive value for the force,
which implies that the force is acting upward. This makes sense since a spring that is pulled away
from its rest position exerts a force back toward that rest position.
In the above example, Alan is applying what I refer to as the
RESTRICTED VALUE
device, he
checks whether the equation makes sense by determining its behavior when the range of some
quantities are restricted to a limited region. In this case, Alan considers the cases when the value of
y is restricted to being negative (less than zero), and then when it is restricted to being positive
(greater than zero).
In addition to considering the cases where y is positive and negative, Alan also considers the
special case when y is precisely zero, noting that the spring force is zero when y=0. This is an
instance of the SPECIFIC VALUE device. In SPECIFIC VALUE, rather than simply restricting a quantity
128
to a range of values, students consider the case when a quantity is restricted to a precise value.
Presuming a quantity to be zero is particularly common, but any value can potentially be used. In
this next example, Mark and Roger are trying to determine if the expression v = vo + at makes
sense, and they consider a few specific values. In addition to this expression, they have also written
a statement on the board specifying that the value of the acceleration is a=32 ft/s2.
v = vo + at ; a = 32 ft 2
s
Mark
Say it starts at zero, right? And it has an acceleration of this much. [g. 32ft/s2] How far
has it gone- what's the velocity after three seconds? Acceleration [g. 'a' in v=v o+at; g.
32ft/s2] times three, right? Cause every second it's increasing by thirty two feet per
second. This is you can say is equal to thirty-two feet per second, per second, right?
Roger
Uh-huh.
Mark
So per second it increases by this much [g. 32ft/s2] in velocity. Right? So, if it starts at
zero, after one second it's going this fast, after two it's going sixty-four, after three it's
going ninety six. See what I'm saying?
In his above statements, Mark presumes throughout that the value of the initial velocity, vo , is zero,
and then he considers several specific values for the time. Initially the velocity is zero, then after
one second, it’s 32, after two seconds, it’s 64, and after three seconds, it’s 96.
Notice that, in the above example, the values that Mark considers each correspond to a
particular moment in the motion. This is frequently the case when the SPECIFIC VALUE device is
applied; a student will elect to restrict a quantity to a specific value because that value corresponds
to some interesting time during the motion. Does this mean that these cases could also be
accounted for by the SPECIFIC MOMENT device? The answer is no, and to explain why I will refer
back to one of the examples I previously presented of SPECIFIC MOMENT . Here, Ella is looking at an
expression from the Air Resistance task, in which the forces of gravity and Air Resistance are
equated.
Cv 2 = mg
Ella
The terminal velocity is just when the - I guess the kind of frictional force from the air
resistance equals the gravitational force?
The point is that, in this interpretation, Ella does not restrict the expression to a specific moment
in the motion; rather, this equation is only true at a specific moment. When we apply SPECIFIC
MOMENT,
we are saying that an equation is only valid for describing a particular moment in a
motion. In contrast, when employing a Special Case device, we consider the behavior of an
expression within a narrower range than its presumed regime of validity.
129
The LIMITING CASE device is a version of SPECIFIC VALUE in which a quantity takes on an extreme
or limiting value. Recall that, for the Running in the Rain task, Jack and Jim composed an
exponential expression for the “probability” of being struck by a raindrop.
Ae −v
Jack
So, you could say the probability if you're traveling at infinite velocity is very extremely
small.
Here, Jack states that if the velocity, v, becomes infinite then the probability of being struck by a
raindrop is very small. This is an example of LIMITING CASE .
In Jon and Ella’s work on the Running in the Rain task we find an instance from the other
extreme. In this example, Jon and Ella are attempting to interpret another complicated expression
that they generated while working on this task. Recall that v R is the speed with which the person
crosses the street. Here they get some work out of considering the case where vR is zero:
vRh
# drops  
w
# drops running = 
wx 
+ 
 A ⋅ s   v v2 + v2 vR 

 d R d
Ella
Jon
Yeah. // And in an extreme case where you’re just standing still then V R [g. vr in bottom of
first term]
// You just put in-
Jon
goes to zero.
Ella
would equal zero.
Jon
And you get=
Ella
=Then this (…) go to zero then you’d be in trouble. [g. last term]
Jon
Yeah, that would be a problem.
Ella
Oh that’s because you’d be standing there forever. (3.0) So I guess that makes sense.
Jon
Oh, yeah. So it go to an infinite.
In this case, imagining that vR is zero led Jon and Ella to a striking conclusion: Although the first
term is simply zero, the second term in the expression is infinite, which means that the person
crossing the street becomes infinitely wet. This is very wet indeed! However, after a moment’s
worry, Jon and Ella decided that this result is precisely right. Traveling at zero speed means that
the street-crosser is “standing there forever,” in which case, it is not surprising that they would get
infinitely wet. Thus, the result of this interpretation is consistent with the conclusion that this
complicated expression is correct.
It is interesting note that focusing on one quantity in the expression and imagining that this
quantity is zero can highlight certain aspects of the structure of the equation and makes many
details irrelevant. For example, the radical in the first term is simply wiped out of consideration.
130
Of course, this interpretive strategy has the drawback that it does not allow Jon and Ella to draw
detailed conclusions concerning the validity of this portion of the expression.
When we examine the special case in which a quantity is zero or infinite, as in the above
example, this typically has the effect of highlighting broad aspects of the structure of an
expression. Consider what happens when we substitute zero for a quantity. If the quantity appears
in the numerator of a term, then that term is simply wiped out; the whole term becomes zero.
There is no need to know any of the other quantities that appear in the term, as long as the
denominator is non-zero. If the quantity in question happens to be in the denominator, then the
term blows up to infinity. Again, the values of other quantities in the term are irrelevant. Note that
this is just the kind of sensitivity to the structure of the equation that is associated with forms; a
key feature here is whether a symbol is in the numerator or denominator, and structure at the level
of terms is highlighted.
The final device in this class is the RELATIVE VALUES device. In all of the other devices in this
class, the value of a single quantity that appears in an expression is somehow restricted. In RELATIVE
VALUES,
two quantities are picked for attention, and the values of these two quantities are
restricted relative to each other. For example, we might specify that one quantity is larger than the
other, that one quantity is much larger than the other, or that the two quantities are precisely
equal.
A variety of examples can be found in the Buoyant Cube problem. As we have seen, each pair
of students derived a relation for the amount of the ice cube that lies below the surface of the
water, and this relation involved the ratio of the densities of ice and water multiplied by the length
of a side of the cube. To interpret this expression, students varied the relative values of the two
densities involved. Sometimes, as was stated in the problem, the density of ice was presumed to be
less than that of water. But, in some cases, students imagined that the densities were equal or even
that the density of ice was greater than the density of water.
D=
ρi
L
ρw
Mike
Does it seem sensible? Well, let's see. Um. If ice is lighter than water, [g. ρi then ρw in last]
then ice will float.
Karl
It makes sense because this is gonna be less // than one. [g. ρi/ρw in last]
// Less than one.
Mike
Karl
And we know that it's gonna float.
Mike
Right. So, yeah, that makes sense.
Karl
That makes sense.
131
________________
ρL
S= i
ρw
Jack
This does seem to make sense cause if ice had the same volume as water [g. ρi then ρw],
er, same density as water, then L would equal S which means the cube would be right at
the surface. [g. diagram]
________________
ρL
S= i
ρw
Jim
Well, if - if you - okay. Let's say rho of ice [g. ρi in last] wasn't point nine two, it was
something greater than water,
Bruce
Uh-huh.
Jim
So, it's something greater than one [g. ρi in last]. That means this would go up [g. L] and S
would go up [g. S in last] so which means it's sinking, which would make sense, cause if
the block was heavier then it would sink more.
The examples above provide clear instances of this last device. In the first passage, Mike and Karl
presume that, as stated in the problem, the density of ice is less than the density of water. In the
second example, Jack considers the case where the densities are equal, which implies that the
amount of the cube below the surface is equal to the length of a side of the cube. Finally, in the
last example, Jim considers the case where the density of ice is greater than the density of water,
which appears to suggest that the amount of the cube below the surface is greater than L. To Jim,
this implies that the cube is “sinking.” As we will see in a moment, there are some problems with
this last interpretation.
It is worth taking a moment to remind the reader of how I explained these examples in terms
of symbolic forms. Two forms are involved in the interpretations of these equations. First, the
ratio of the densities of ice and water is seen in terms of the RATIO form. In addition, this ratio,
treated as a unit, is seen as SCALING the length L.
With these forms in mind, we can now get a little of a feel for how devices and forms work
together in these interpretations. Recall that the RATIO form is associated with comparisons of the
quantities that appear in the numerator and denominator of the ratio. That’s what going on here,
the students are basing their interpretations on comparisons of the densities. Furthermore, the
forms are allowing inferences to be drawn. The RATIO form permits the conclusion, for example,
that if ρi is greater than ρw , then the ratio is greater than one. And SCALING allows inferences
132
concerning whether the amount of the cube below the surface is more or less than the length of an
entire side.
Inventive combinations of devices
To this point, I have been proceeding under the conceit that I can easily place interpretive
utterances into distinct categories. Although the majority do fall neatly into the categories I have
outlined, some interpretive utterances seem to involve aspects of more than one category. This is
not a problem as long as we think of representational devices as a repertoire of interpretive
strategies that can be combined in an inventive manner.
Note that this is one place that talking about representational devices as knowledge pays off.
Even though some interpretive utterances may not line up uniquely with the categories I defined,
we can still think of these utterances as generated by the set of devices that I listed. Thus, the
hypothesis is that representational devices constitute elements of knowledge that can be learned as
separate elements, and then employed, separately and together, to generate interpretive stances.
I want to present just one example of an utterance produced by an inventive combination of
devices. This example pertains to Mike and Karl’s work on the Air Resistance task. I have had a
number of occasions to discuss their equation for the acceleration:
a(t ) = −g +
f (v)
m
Like all of the subject pairs, Mike and Karl decided that the force of air resistance should be
proportional to velocity, and they wrote the relation f(v) = kv. Substituting this relation in the
above expression for the acceleration yielded
a(t ) = −g +
kv
m
Now I want to jump to Mike and Karl’s work on part (b) of the Air Resistance task, in which
the students were asked to consider the effects of a wind that blows straight up. To account for
this wind, Mike and Karl modified their expression so that the velocity of the wind, vw , was added
to the velocity of the ball:
a(t ) = −g +
k
(v + vw )
m
133
Finally, with this expression in hand, Karl proceeded to interpret it as follows:
Karl
So, if the wind is blowing up, it causes the terminal velocity, this acceleration [g.
k/m(v+vw)] to reach this acceleration [g. -g] faster.
Karl’s interpretation involves two imaginary processes through time, one process in which there is
wind, and one in which there is not. In both of these processes, the second term grows in time
until it cancels the first term. The point of Karl’s interpretation is that this happens more quickly
in the case in which there is wind.
I understand this interpretation as involving two representational devices: (1) There is a
comparison across two fundamentally different situations (wind and no wind) and (2) there is a
process through the time of the motion. Thus, this interpretation is based on an inventive
combination of two representational devices, CHANGING SITUATION and PHYSICAL CHANGE .
Inappropriate and Inconclusive Interpretations
In the previous section, we sifted through students’ repertoire of interpretive strategies. The
extent and flexibility of this repertoire is, in itself, striking. It seems that, when it comes to
interpreting equations and using them meaningfully, third semester physics students already
possess a great deal of expertise. However, although I tended to emphasize the positive in this
chapter, there are limits to students’ abilities. As we have already seen in some examples, students
do not always interpret equations flawlessly; they sometimes give inappropriate interpretations and
sometimes get stuck with inconclusive results. These inappropriate and inclusive interpretations are
the subject of this section.
To begin, as I have already mentioned, Alan’s BALANCING /SPECIFIC MOMENT interpretation of
the equation µmg=ma can be confidently described as “incorrect.” It presumes the existence of a
second influence and inappropriately associates the equation with a single moment in the motion,
the moment when the block comes to rest.
Students also produced a number of inappropriate interpretations of their final expression
from the Buoyant Cube problem. We encountered one of these interpretations during my recent
discussion of Special Case devices.
ρL
S= i
ρw
Jim
Well, if - if you - okay. Let's say rho of ice [g. ρi in last] wasn't point nine two [g. .92 on
block], it was something greater than water.
Bruce
Uh-huh.
134
Jim
So, it's something greater than one [g. ρi in last]. That means this would go up [g. L] and S
would go up [g. S in last] so which means it's sinking, which would make sense, cause if
the block was heavier then it would sink more.
In this passage, Jim imagines the case in which ρi is greater than ρw , which seems to imply that S
is greater than L. But there is a big problem with this interpretation. In order to derive the above
expression, Jack and Jim equated the upward and downward forces on the block and, in so doing,
they implicitly assumed that the block does float. However, if ρi were greater than ρw , then the
block would sink and the assumption that the forces are equal does not hold. This means that the
above equation cannot correctly be used to describe the case in which ρi is greater than ρw . Thus,
Jim’s interpretation is inappropriate.
It is interesting that such an innocuous seeming expression can be subject to serious errors in
interpretation. We might be tempted to think that, once this expression is derived, then all of the
hard work associated with solving the problem is complete. But the interpretation of an expression
can pose a serious challenge. In fact Alan and Bob made precisely the same error in their work on
the Buoyant Cube problem, though Bob immediately caught the mistake:
ρ L
x = ice
ρh2 o
Bob
So, okay, so if the density [g. ρice] is greater than water [g. ρh20]. Um, X is gonna be greater
than L. Um, okay, makes sense.
Alan
The density is // greater than (one).
Bob
// This - we assumed, we assumed it floated actually. [w. "floated"] We have
to assume it floated. If we didn't assume it floated, then this wouldn't give us the right
answer.
Now I want to consider another way in which limits in a student’s abilities can manifest
themselves during attempts to produce an interpretation. In some instances, students struggle to
produce any useful interpretation of an expression. The interesting thing about these instances is
not that students say things that are wrong, but that they have trouble saying anything cogent
about an equation or drawing a useful conclusion from an equation. An interesting set of related
cases involve what I call “inconclusive” interpretations. In these interpretations, a student sets out
to construct a certain type of interpretation of an expression, but, in some manner, the
interpretation fails to produce useful results.
We have already encountered some examples of inconclusive interpretations. In attempting to
interpret one of their complicated expression for the Running in the Rain task, Jon and Ella
explored a tactic that failed to produce useful results:
135


h
w
 # drops  
+
=
wx
vd2 v R 
 A⋅s  
1+
v
 d

v 2R




Ella
Okay, V R is bigger. Then (10.0),, This is gonna get smaller, [w. arrow pointing to second
term] but this is gonna get bigger. [w. arrow pointing to first term] I think.
Jon
Yeah, I think - that gets smaller [g. second term] cause V walking is smaller than V
running.
Jon and Ella have run into a problem here. They see that if vR gets bigger, then the second term in
their expression gets smaller. However, under these circumstances the first term gets larger. So,
given these conflicting observations what can Jon and Ella conclude about the overall expression?
The result is simply inconclusive; this method of interpreting does not tell them whether the total
rain becomes greater or less. Jon and Ella understood this difficulty clearly:
Jon
And, um, we’re trying to get it in a form that we can look - we can say, well V walking is less
than V running and then compare the equations. [g. expression] But right now since this
thing is getting bigger [g. second term] - I mean this thing is bigger when you are walking
and this thing [g. first term] is smaller when you’re walking,
Bruce
Hm-mm.
Jon
They sort of - we don’t know which one is bigger than the other.
Bruce
Hm-mm.
Jon
Or has more significance than the other.
So, sometimes students reach an impasse when trying to construct a useful interpretation of an
expression. In this case, I am not sure that an expert could make a CHANGING PARAMETERS
interpretation of this expression do much more work for them. Because Jon and Ella made some
errors during earlier steps in their work, they produced a somewhat less tractable and more
difficult to interpret expression than other students in the study.
Before wrapping up this section, I want to present one additional example of an inconclusive
interpretation. This example involves the Shoved Block problem, in which a heavier and lighter
block are shoved so that they have the same initial speed. The question that students must answer
is: Which block goes farther? As I have discussed, my subjects displayed two different intuitions
concerning this problem. The first intuition is that the heavier block should go less far because it
presses down harder on the table and thus is subject to a higher frictional force. The other
intuition produces the opposite conclusion; the heavier block should go farther because heavier
things have more momentum and are thus “harder to stop.” In their work on this problem, Alan
and Bob quickly mentioned both of these intuitions. Alan began with:
136
Alan
And then if the block was heavier, it'd just mean that our friction was greater, and it would
slow down quicker.
Then, shortly later, Alan stated the second intuition:
Alan
That seems kind of- Something seems strange about that because if you had like a block
coming at you, a big heavy block, and a small lighter block. Two different instances. And
you have this big heavier block coming toward you and you want it to slow down, it
certainly takes much more force for you to slow down the bigger heavier block than the the lighter one.
Bob
Oh wait, that's true!
Alan
Because, even though they're both working on the same frictional surfaces, just logically I
know the big heavier block takes-, I mean football is all about that, that's why you have
these big three hundred pound players.
So, which intuition is correct? Answering this question is a job for equations. Bob first attempted
to use only the equation F=ma to determine which intuition is right:
F = ma
Bob
I mean, F is obviously going up as you get um - The frictional force is going up as you get
heavier [w. an arrow going up under the F], and the mass is going up [w. an arrow under the
m]. So, I mean it depends if.
As indicated by the arrows that Bob drew, both the force and the mass increase as you go between
the light and heavy-block circumstances. Essentially, Bob is running a CHANGING PARAMETERS
device that involves simultaneously changing the force and the mass. His hope, we can presume, is
that he will find out how the acceleration changes as the force and mass are increased. However,
the result of this interpretation is inconclusive. The fact that the force is increasing implies that the
acceleration should also increase, but the increasing mass implies that the acceleration should
decrease. As we know, it turns out that these effects precisely cancel, but Bob’s above method is
not sufficient to determine this fact. Following this inconclusive interpretation, Alan and Bob went
on to do a more complete solution and they did determine that these two effects precisely cancel,
and therefore that both blocks travel the same distance.
The upshot of this section is that interpreting equations can be difficult. Sometimes students
produce incorrect interpretations, and sometimes they have trouble producing any useful
interpretation at all. These observations are important for several reasons.
First, the difficulties observed here point to issues that must be dealt with by physics
instruction. Our instruction should lead to students having the ability to extract correct and useful
interpretations from the expressions that they write. Such a goal is a worthy end in its own right,
137
but it is important to note that interpretations also play a role in the problem solving process. To
take a simple example, if a student decides that an expression does not “make sense,” then it is
likely that they will try to fix the solution that led to the expression. For such reasons, the ability
to correctly interpret expressions has implications for successful problem solving and therefore
must be addressed in instruction.
Furthermore, the observations in this section help to support the notion that it is important to
account for interpretive utterances in our models of symbol use. If interpretation were always
smooth and unproblematic, we could perhaps wrap it in a black box and take the ability to
generate interpretations for granted, just as I am taking for granted students’ ability to plug
numbers into an equation and their ability to perform syntactic manipulations. However, the
above examples make it clear that the act of interpreting an equation is not always unproblematic.
Finally, I believe that the examples in this section can generally help to support the fact that
interpretive phenomena are real—there really is a class of utterances worth calling “interpretations”
and there really is knowledge devoted to interpreting expressions. These examples help because,
like many other capabilities, the ability to interpret expressions reveals its existence and nature in
cases of breakdown. Alan’s interpretation of the equation µmg=ma sticks out like a sore thumb in
the protocol of the session in which it appears, and Jon and Ella’s frustration when they could not
generate a useful interpretation of their complicated expression was dramatic. Thus, these
inappropriate and inconclusive interpretations stand out as real phenomena to be accounted for.
And once we acknowledge that there are inappropriate interpretations, then we have already
implicitly acknowledged that there is such a thing as an interpretative utterance, of any sort.
Interpretation as ÒEmbeddingÓ
I have, at times, described the project in this chapter as taking interpretive utterances, putting
them in categories, and then characterizing the common features of utterances within each of the
categories. Now I want to take a moment to think about whether we can come up with any
characterizations that span the categories. Are there any overarching statements we can make
about all of the interpretive utterances that we have seen? Are there any general ways to
characterize interpretations of equations?
A possibility that I mentioned was that interpretation could be equivalent to the stating of
truth conditions. I will have more to say about this possibility and theories of equation meaning in
a moment. Another possibility is that interpreting an equation may boil down to making
138
correspondences between the equation and the motion that the equation is supposed to describe.
At least at first blush, this works well for some interpretations we have seen, such as SPECIFIC
MOMENT
interpretations, which project an equation into a single instant in a motion. However, it is
a little less clear how this correspondence characterization will work for some other varieties of
interpretations, such as those involving the CHANGING PARAMETERS device. CHANGING PARAMETERS
interpretations incorporate stories that span multiple versions of the same motion, and CHANGING
SITUATION
interpretations interpret an equation by considering motions other than the one that is
specifically the subject of study. Thus, it does not seem like the notion of correspondence between
equation and the physical world does a good job of characterizing the breadth of interpretations
that we have encountered, at least in its simplest form.
But my goal here is not to list a variety of possibilities and argue among the alternatives.
Thoroughly arguing for a choice among such alternatives would be an extensive undertaking.
Rather, I want to present what I think is a passable, rough characterization, what I call the
“interpretation as embedding” view. This characterization is not designed to serve as anything like
a definition or even as a rigorous description. Instead it is designed to function as a mnemonic
that captures some features of what I have found, particularly features that some may find
surprising and that distinguish my viewpoint from competitors. In this section and the next two, I
will work through some of the aspects of this “embedding” view of interpretation, and, in the
process, I will attempt to uncover some of the general conclusions concerning the nature of the
interpretation of equations that can be drawn from the observations in this chapter. Although I will
be building on the examples and observations presented earlier, my comments in these sections
will be, in general, much more speculative than those in the rest of this chapter, and I will be going
somewhat beyond the core of my theoretical position.
The first thing I will do is make clear what it means to interpret by embedding. To do this,
consider the equation that Jim interpreted:
x=
mg
k
Now I will present my own interpretation of this equation, which is designed specifically to help
illustrate the nature of interpretation by embedding:
Bruce
Suppose you’ve written this equation on a sheet of paper. Then you plug in some value for
the mass and get out a result. Then you can plug in another, larger value for the mass and
get out another result. If you do this, then the answer you get for the larger mass will be
bigger.
139
In this passage, I have interpreted this equation by placing it within a little story. We can, if we
like, construe this story as being about the equation itself, but we need not. The equation is only
one element in a story that includes other elements, playing other roles. The story includes, for
example, a person using the equation, writing it on a sheet of paper and plugging in values. And
there are narrative elements that are not necessitated by the equation, such as the choice to first
substitute one value, then substitute a larger value.
This is the essence of the “embedding” characterization of interpretations: The equation is set
within a context or framework that is larger than any possible listing of the entailments of the
equation. The framework is not fully dictated by the equation, and not every aspect of the
framework corresponds to some feature of the equation. Thus, in this view, it is simply not
appropriate to ask how features of the equation correspond to aspects of some world. If we are
looking to ask a question in this vein, we should instead ask how the equation articulates with or
fits with the other elements of the story; in other words, we should ask how the equation is
integrated into the larger framework of the interpretation.
Ambiguity and the Abstractness of Devices
There are some additional, somewhat more subtle points to be made about the nature of
embedding interpretations, as they appear in the student utterances in my data corpus. Consider
the following CHANGING PARAMETERS interpretation, in which Karl interprets an equation for the
acceleration from the Air Resistance task:
a(t) = −g +
Karl
f (v)
m
Okay, this sign is negative [g. -g] this number [g. f(v)/m] as the mass- as you increase the
mass,, Well, one problem- from one ball to the other the mass is doubling. Right, so you
double this mass [g. m] so this number [g. f(v)/m] is getting smaller.
Contrast Karl’s interpretation with the contrived example I presented in the previous section. In
the contrived example, certain aspects of the story I told were very clear. For example, it was quite
clear where the story took place and who the actors were; the story involved a person working with
an equation, writing things on a sheet of paper. However, in Karl’s above interpretation, it is
somewhat less clear where the story takes place. As in my contrived interpretation, this one
involves a changing mass. But we can imagine this change as occurring in either of two possible
settings. First, just as in my contrived example, we could treat Karl’s interpretation as taking place
140
in the world of a problem solver, in which someone is using the equation. In this setting, there is a
hypothetical person plugging-in different values for the mass and obtaining various results.
But we can also imagine Karl’s interpretation as taking place in an alternative setting, the world
of objects and motions. If we adopt this stance, then Karl’s story is about a sequence of physical
experiments in which someone takes objects of various mass to a great height and then drops
them.
So which of these settings does Karl have in mind in his above interpretation? What does Karl
mean when he says “as you increase the mass?” Does this increase take place in the physical world
or in the here-and-now of the problem solving world?
Of course, there is the possibility that Karl’s interpretation just happens to be more ambiguous
than the norm. To see that this is not the case, the reader is encouraged to flip back through this
chapter. Such a perusal will reveal that the majority of interpretations are ambiguous in the same
manner. However, I will present one additional very brief example here. This is a typical
interpretation of the final expression from part (b) of the Air Resistance task, in which the effect of
wind is considered.
vt =
Bob
m
g − vo
k
If the wind is blowing up, the velocity - terminal velocity is going to be less than it was in
this case. [g. vt=m/kg] I mean, we have the same mass. This is M K G. [w. vt=m/kg; draws
a box around it] This is no wind. [w. "no wind"] No wind. And this is with wind. [w. "w/ wind"
and boxes wind result] Okay, so if the wind is blowing up, the terminal velocity is gonna be
less.
As is typical of interpretations of this expression, Bob compares this result to the case in which
there is no wind. The question here is: what kind of things are these two “cases”? Are they two
real-world experiments? Or is each “case” a problem to be solved in a physics class?
I believe that this ambiguity is built into the nature of these interpretive utterances. Although
some interpretations may sound more or less like they take place in one of these two possible
“worlds,” in most cases it is hard to decide which world students have in mind. Furthermore, I do
not believe that this is just a matter of limitations in what we, as analysts can reliably say about
protocol data. Rather, I believe that there is a fundamental sense in which these utterances are
truly ambiguous. According to this view, it is incorrect to ask whether an interpretive utterance is
about the physical world or about the here-and-now world of symbol use. This means, for
example, that when a person says the word “velocity,” they mean neither the velocity symbol nor a
velocity in the world, they somehow mean both at once.
141
I want to propose a way of thinking about this observed ambiguity that I hope takes away a
little of the mystery and strangeness. The proposal is that we can think of the interpretation of
physics equations as involving embedding in worlds or settings that are abstractions of “real”
settings or real processes. The idea is that these specially adapted hypothetical settings involve less
information and are less thoroughly defined than real settings in which humans live and act. And
because they are more abstract, they can simultaneously be consistent with more than one of these
real settings for human activity.
To illustrate, I want to point out some of the other ways that student interpretations tend to
be ambiguous. In CHANGING PARAMETERS interpretations, we describe a process in which some
parameter increases or decreases. There are a couple of possibilities concerning how this change
might occur. First, the change could be a continuous change, in which the parameter is supposed
to pass through every possible value. Second, the parameter could change discontinuously, from
one value to another. Again, student utterances were generally ambiguous in this regard.
Furthermore, it is often not possible to determine whether the supposed change even happens
through time. Do you do one experiment and then the other? Or are the two experiments simply
two cases to be considered in parallel, with no assumed time-ordering?
Again, the idea is that the worlds and processes described in interpretations can be considered
to be abstract and there are thus limits to what we can say about these interpretations. We can say
that they embed the equation in a hypothetical process of change; i.e., they employ the
PARAMETERS
CHANGING
device. But we cannot say that this process takes place in one world or another, and
we cannot necessarily claim that the process is continuous or discrete. The point is that if we try to
understand Karl’s interpretation as taking place in a real setting—as describing some specific
experience that we could really have, such as dropping a ball or writing an equation—then we are
filling in details that are not part of or even implicit in Karl’s statement.
I believe that it is plausible that this very particular kind of abstractness is also characteristic of
the interpretations of physicists—not just the moderately expert students in this study. If this is
true, it points, once again, to the fact that there is a lot for students to learn here; there is much
that is special about interpreting physics equations. This is not really a surprising observation.
Physicists have a standard way of talking with and about equations, and it makes sense that there
will be some particularities to this discourse. In learning to interpret equations, students are
learning to make utterances about expressions that are like those that a physicist would make, and
it turns out that these utterances involve the sort of abstract constructions I have described here.
142
Postscript: A function for ambiguity
In this section, I have argued that interpretive utterances tend to be ambiguous in certain
specific respects. And my stance has been purely descriptive; I have simply argued that “this is how
physics discourse is.” The question arises of whether we can say why it is that interpretive
utterances are ambiguous in this way. Does it serve any obvious function?
I have one guess at an answer to this question. Above, I distinguished two possible settings in
which an equation could be embedded. The first is the world of problem solving, in which a
person is writing and manipulating equations. The second is the physical world that the equation is
presumed to describe. It is clear how an equation can be embedded in a story that takes place in
the first kind of world—equations are one of the things that appear in a problem solving setting.
But, what could it possibly mean to embed an equation in a story about the physical world?
Stories about Newton’s physical world include objects and motions, not equations.
In discarding correspondence as a fundamental means of connecting the symbolic and the
physical, we are left with a problem. Without correspondence, how can we make an equation have
anything at all to do with the physical world? This is where the above type of ambiguity can
actually help. The abstractness built into interpretations allows a blurring—a sort of welding
together—of the physical world and the here-and-now of the problem solving world.
Interpretations thus embed equations in a story that is simultaneously about the physical world
and the problem solving world. This permits a connection, of sorts, between equations and the
physical world.
On the ÒMeaningÓ of an Equation
I began this chapter by wondering if we could better understand interpretations of equations
by first constructing a “theory of meaning” for equations. The route that I ultimately took was a
different one, however. Rather than beginning with an a priori theory of meaning, I instead took
an empirical approach to discovering what an interpretation of an equation is, in practice. Now this
project can come full circle. Assuming that what we have seen in this chapter is broadly
characteristic of interpretation in physics, what can we say about the meaning of equations?
Let’s begin by looking closely at a brief example. In this example, Bob gives an interpretation
that involves the NO DEPENDENCE form and the CHANGING PARAMETERS DEVICE. The equation in
question is from the Shoved Block task.
143
a = gµ
Bob
So, no matter what the mass is, you're gonna get the same, the same acceleration.
According to Bob, this equation says that the acceleration of the block is independent of the mass.
That is Bob’s statement of a meaning for this equation. Now, notice that this interpretation has a
strange feature: The meaning that Bob gives for this equation involves a quantity that does not
even appear in the expression! For this reason, this interpretation is not consistent with simple
versions of a theory in which meaning is something that can be just “read out” of an equation; the
meaning cannot be intrinsic to the arrangement of symbols in the expression. In particular, there is
no way that the arrangement of symbols can explain why Bob chooses to mention that this
expression does not depend on the mass, rather than mentioning all of the other many parameters
that do not appear in this expression.
In this next example, the plot thickens further. Here, once again, is Karl’s interpretation of an
equation from the “wind” problem that combined the PHYSICAL CHANGE and CHANGING PARAMETERS
devices:
a(t ) = −g +
Karl
k
(v + vw )
m
So, if the wind is blowing up, it causes the terminal velocity, this acceleration [g. second
term] to reach this acceleration [g. first term] faster.
The noteworthy feature of this interpretation for the present discussion is that the meaning that
Karl assigns involves a comparison to a fundamentally different physical situation. In fact, as we
saw, many interpretations from the wind task involved the CHANGING SITUATION device and
comparisons of this sort. For example, students frequently interpreted the final result for the
terminal velocity by comparing it to the terminal velocity for the case in which there is no wind:
vt =
Bob
m
g − vo
k
If the wind is blowing up, the velocity - terminal velocity is going to be less than it was in
this case. [g. vt=m/k g] I mean, we have the same mass. This is M K G. [w. vt=m/k g; draws
a box around it] This is no wind. [w. "no wind"] No wind. And this is with wind. [w. "w/ wind"
and boxes wind result] Okay, so if the wind is blowing up, the terminal velocity is gonna be
less.
The point I want to make about these interpretations is that they are not solely about the physical
situation that the equation presumably describes. Instead, these interpretations are based around a
comparison to an altogether different physical situation. This observation forces us to the strange
144
conclusion that the meaning of the above equations, as embodied in students’ interpretations, has
something to do with a physical circumstance that, in a sense, is not “mentioned” in the equation.
This point is central to the embedding view of interpretation. The building of “meaning” for
an equation involves dragging in more than what is straightforwardly entailed by the symbols that
appear in the expression. The existence of CHANGING SITUATION interpretations, which involve
comparisons to altogether different physical circumstances, helps to make this point evident.
Because Bob’s interpretation involves a “less than” comparison, the meaning is intrinsically
relational. Furthermore, it is not clear that there is anything about the equation itself that instructs
the equation-user what physical circumstance to choose for comparison.
These observations lead to another set of questions. If anything can be used as part of the
meaning of an equation, what determines what extraneous material is dragged in? And, more
generally, if meaning is not a simple function of the arrangement of symbols in an equation, what
else determines the meaning of the expression?
symbolic expression ⇔ forms ⇔ devices ⇔ rest of the context
To investigate this question, we can return to my schematic of the form-device system. Recall
that in my heuristic model, the equation, forms, devices, and the rest of the context form a
system of constraints that adjusts itself to generate an interpretation. Thus, the forms seen by the
student and the choice of device are not only influenced by the equation, but also by the “rest of
the context.” The conclusion I draw from this is that the meaning of an equation as embodied in
interpretive utterances is context dependent; it depends upon the concerns of the moment. What
we choose to compare an equation to and what other entities and circumstances we drag in, will
depend on what’s going on in the here-and-now world of problem solving. With this conclusion in
mind, it is not surprising that students interpret the result of the wind task by comparing it to the
no-wind case since, only minutes before, they had solved the same problem for the case in which
there is no wind. Thus it makes sense that the meaning of the equation for students, in this context,
would involve a comparison between these two cases.
This driving of interpretation by concerns of the moment can take an even more trivial form.
In some cases, the choice of representational device can be seen to be driven directly by the
problem statement. Consider, for example, Bob’s above interpretation of the equation a=gµ in
which he notes that this expression does not depend on the mass. Bob has chosen to be concerned
with mass dependence for the simple reason that this is what the question asks; it asks whether a
145
heavier or a lighter block will travel farther. Similarly, interpretations of expressions written for the
Air Resistance task frequently involved
CHANGING PARAMETER
v=
Alan
devices in which the mass varied:
mg
k
So now, since the mass is in the numerator, we can just say velocity terminal is
proportional to mass [w. vt∝m] and that would explain why the steel ball reaches a faster
terminal velocity than the paper ball.
________________
a(t) = −g +
Karl
f (v)
m
Okay, this sign is negative [g. -g] this number [g. f(v)/m] as the mass- as you increase the
mass,, Well, one problem- from one ball to the other the mass is doubling. Right, so you
double this mass [g. m] so this number [g. f(v)/m] is getting smaller.
Again, it is not surprising that these students chose to interpret these equations in this manner
since the problem statement specifically asks how the terminal velocity would differ in the case
when the mass is twice as large.
The upshot of these last few paragraphs is that the meaning of an equation will be dependent
on the concerns of the moment. I want to emphasize, however, that I have tried to do more than
simply announce that equation meaning is “context dependent.” I believe that one of the main
contributions of this work is that I have provided a partial explanation of how concerns of the
moment interact with equation use. What I have argued is that people know a set of interpretive
stances and strategies that I call representational devices, and that these function as a layer between
forms and the “rest of the context.” The concerns of the moment thus feed through
representational devices to determine how equations are interpreted.
So, the bottom line is that equation meaning, as embodied in student utterances, is not a
simple function of the symbols in an equation. And there is no single, universal procedure for
reading out the meaning of an equation. Instead, meaning is built around an equation by the
symbol user via representational devices, and it depends in a complex manner on the concerns of
the moment.
In conclusion, we have once again taken a foray into the physicist’s world of symbols. In the
last chapter, we explored the structure—a meaningful structure—that a physicist sees in a page of
expressions. Now, in this chapter, I have argued that, to enter into this meaningful world of
symbol use, it is not sufficient to know symbolic forms. In addition, initiates in physics must
possess a complex repertoire of interpretive strategies that I have called “representational devices.”
146
With forms, this repertoire of devices defines the character of the meaningful symbolic world of
physics.
147
Chapter 5. Symbol Use and the Refinement of
Physical Intuition
The model described in the preceding chapters is designed to do a very limited job. To pin down
what this job is, it is helpful to think about time scale. So far in this project, I have only been
working to describe phenomena, such as the interpretation of an equation, with a time scale on the
order of a few seconds. But there are many other phenomena, with a range of associated time
scales, that could reasonably be taken to be part of this project. Here are some time scales that we
might want to consider:
1. The evolutionary time scale (~10 million years). As brains evolved so did cognition. This is, in
part, the sort of story I attributed to Merlin Donald in the first chapter.
1. The historical time scale (~10,000 years). Some authors have wondered whether the nature of
human thought has changed over the course of recorded history. This time scale is the purview
of researchers like David Olson and Jack Goody, both of whom have worried about the effects
of the development of writing on thought.
1. The sociocultural time scale (~100 years). We can imagine that there are ebbs and flows in the
thinking of particular cultures and of the practitioners within a domain.
1. The ontogenic time scale (~10 years). The ontogenic time scale is Piaget’s realm. Piaget was
interested in characterizing broad changes in the nature of thought, as individuals progress
from birth to adulthood. While Donald’s stages are phylogenic—they pertain to the
evolutionary development of the human species—Piaget’s stages describe the development of
individuals.
1. The learning time scale (~1 year). Even when no broad changes in the nature of thought are
implicated, minds can still change and we can call these changes “learning.” For example,
during their first one or two years in college, students can learn some physics.
1. The problem solving time scale (~10 minutes). This is the time scale of the problem solving
research I discussed in Chapter 1. Problem solving research is primarily concerned with
describing behavior of individuals that spans from a few minutes to a couple of hours, as they
work out the solution to problems.
148
1. The thought time scale (~1 second). The thought time scale is associated with phenomena like
the generation of an interpretation of an equation: A student looks at an equation, comes to
see it in a certain way, and then says something.
For the most part, I have thus far been dealing with phenomena in time scale (7), with a little
thrown in about (6). The “model of symbol use” I have presented is intended to describe short
time scale phenomena; the form-device system adjust itself in a few seconds. Of course, as I have
discussed, there are implications of my model for problem solving. Because students can see
equations as meaningful, they can construct novel equations and they are likely to take certain
subsequent actions. Nonetheless, my model does not really describe events on the order of
minutes. For example, I have not been directly concerned with the order and progression of
expressions that are written on the way to a solution.
In this chapter I want to extend my scope of interest, not to include phenomena in time scale
(6), but up yet another order of magnitude to the learning-related phenomena in time scale (5).
My goal here is to provide some account of the learning that takes place over the period of the
first few years of physics instruction. Since my data corpus does not include a longitudinal study,
this extension of my model will necessarily be somewhat more speculative than the core of the
model, which was presented in the previous two chapters.
There are, however, some reasons to believe that I can do pretty well with these speculations.
First, the students in my study are neither complete novices nor true experts, they are at an
intermediate level of expertise. For this reason, the “snapshot” that I have in my data corpus
should have useful properties from the point of view of speculating about the broad sweep of
development. Second, although the data corpus that is at the center of this work does not include
longitudinal observations, there is other research that I can draw on. In fact, the educational
research literature is replete with characterizations of the knowledge possessed by younger
students and of characterizations of “intuitive physics” knowledge, the knowledge that students
bring to physics instruction. Therefore, in essence, I have snapshots of some knowledge systems as
they exist prior to physics instruction, and some knowledge that exists during an intermediate
stage in physics instruction. Given these snapshots, I can speculate about the processes that connect
these knowledge systems.
149
Cognitive Resources: Prior Knowledge and Physics Learning
How do we interpolate between snapshots of knowledge systems? To deal with this question,
we need to develop a general means of thinking about how old resources come to be used in new
contexts. There is a set of related questions to be addressed here:
•
How must existing cognitive resources be adapted for use in new contexts?
•
Under what circumstances should we say that new resources have developed out of
existing resources, rather than old resources adapted?
•
When new resources develop, in what ways do these resources relate to the previously
existing resources?
All of these questions relate to the flexibility and generality of cognitive resources. Certainly, we
expect that cognitive resources will change and adapt. We must also expect that, as individuals
learn sufficiently new behaviors, new cognitive resources will develop out of the existing resources.
These issues arise in our study of the learning of symbol use in physics. We must ask: In what ways
does symbol use in physics make use of existing resources and in what ways does it require the
development of new cognitive resources?
In this paper, I will do little to answer these very general questions concerning the specificity of
resources and the manner in which new resources develop from existing resources. My main
contribution will be to point roughly to relevant prior resources and to indicate where I believe we
should say that new resources have been developed. To make this task simpler, I will introduce a
little terminology. I will use the term “cognitive resource” to describe relatively functionally
specific resources, and I will also use the term “cognitive resource cluster” to describe collections of
such resources, grouped according to use in narrowly defined activities.
Cognitive resources and the learning of physics
The above brief discussion of cognitive resources lays out the issues to be considered here very
generally. People have cognitive resources and, when they learn new things, this means that they
adapt these resources and develop new resources. Now I want to start to fill in the particulars of a
story of this sort for physics learning.
150
What are the relevant cognitive resources that exist prior to physics instruction? We can narrow
the scope here a little by saying that we are interested in resources that contribute to symbol use in
physics, as I have been describing it. Much prior knowledge is implicated in the development of
even this narrow slice of physics knowledge, so I will limit my discussion to describing a few of the
particularly interesting relations between existing and new resources.
First, students come to physics instruction with a cluster of resources that pertain to the
solving of mathematics word problems. I call these cognitive resources the “algebra problem
solving” cluster (refer to Figure Chapter 5. -1.) This knowledge is certainly relevant to learning to
solve the types of textbook problems that students encounter in their physics courses. The algebra
problem solving cluster is almost exclusively associated with schoolish activities and it is employed
in solving the traditional algebra problems that one would find in a school textbook. It includes
resources relevant to performing the small variety of tasks found there, such as solving given
equations, solving simple word problems, or deriving equations by manipulating given equations.
Of course, there is a wide variety of literature that attempts to describe the knowledge associated
with this activity. Of particular relevance for us is the “patterns” research, described in Chapter 3.
Recall that this research argued that students apprehend arithmetic word problems in terms of a
number of simple patterns, such as “part-part-whole” and “equalization.”
Quantity
e.g. qualitative
proportionalities
Algebra Problem
Solving
Intuitive Physics
e.g. p-prims
e.g. "patterns"
Algebra-Physics
e.g. forms, devices
Figure Chapter 5. -1. A resource cluster story for physics learning.
A second relevant cognitive resource cluster I refer to as the “intuitive physics” cluster. The
intuitive physics cluster is associated with the day-to-day activity that occurs throughout a
person’s life, interacting with the physical world. In particular, I have in mind the activities of
151
describing and explaining physical phenomena. Our capabilities in this regard are, of course, quite
extensive and varied. We can describe the geometrical arrangement of objects in space as well as
make predictions concerning the behavior of a physical system. In what follows, I will be
particularly interested in a particular resource within this cluster, diSessa’s “sense-of-mechanism,”
which I discussed in Chapter 1 (diSessa, 1993).
In large part, learning expert physics means learning to quantify the physical world. For this
reason I have included a cluster of resources relating to quantity in Figure Chapter 5. -1. As with
the other clusters of resources, I presume that students enter physics with many resources relating
to dealing with quantities in the world. Children learn to count objects and variously use numbers
across a variety of circumstances. In addition, people develop the ability to deal more generally
with amount, even when the entities involved are not simply denumerable, as when we are talking
about a quantity of liquid. Among the resources that we could consider, I have already discussed
the abilities that people possess in the area of “qualitative reasoning about physical systems.”
This third cluster of resources is less functionally specific than the above two; in fact, it
certainly plays some role in algebra problem solving and intuitive physics activities. In general, we
can expect there to be interesting overlaps and inter-relations among resource clusters. For the sake
of keeping this story simple, I am not going to worry about these inter-relations. My purpose here
is only to give a very simple account that highlights what I believe are some of the more important
resources involved.
The point I have been working toward here is that learning symbol use in physics involves the
formation of a new cluster of resources that I refer to as “algebra-physics.” This cluster involves its
own characteristic resources, including the form-device system I described in the previous two
chapters. Furthermore, I intend for the diagram in Figure Chapter 5. -1 to depict two different
kinds of relations between algebra-physics and the previously existing resource clusters. The first
of these relations is genetic: Algebra-physics resources develop, in part, out of the existing
resources. Second, once the new resources have developed, connections may remain between the
new resources and the previously existing resources.
My purpose here is not to elucidate the mechanisms involved in these two relations. For the
present, I do not intend this list of relations to be much more than an accounting of some of the (a
priori) possible relations between old and new resource clusters. The statement here is really a very
simple one: New resources can develop out of existing resources and, once developed, they may
maintain some types of relations to these prior resources.
152
Furthermore, I have no intention of even working through all of the possible relations shown in
Figure Chapter 5. -1. Instead, I will focus primarily on one particular relation between old and
new resources: the relation between diSessa’s sense-of-mechanism and the form-device system.
There are a few reasons for focusing on this particular relation. First, we shall see that the elements
in the sense-of-mechanism—p-prims—are closely related to symbolic forms (in the sense of both
types of relations that are supposed to be depicted in Figure Chapter 5. -1). Second, this focus will
help to show how symbol use can have effects that permeate right down to what, I will argue,
must be considered the roots of physics understanding. And finally, a number of important issues
relating to learning will be clarified by working through this particular relation.
The Developing Sense-of-Mechanism
In this section I begin my focus on the sense-of-mechanism and its role in the development of
the form-device system. I briefly discussed diSessa’s account of this cognitive resource in Chapter
1 and I will provide a somewhat extended discussion here. All of this presentation is derived from
diSessa’s Toward an Epistemology of Physics (1993) and the interested reader should look there for
the authoritative account.
The sense-of-mechanism is a portion of our intuitive physics knowledge—the knowledge that
we gain through our day-to-day experiences in the physical world. One function of this
knowledge is to contribute to our ability to interact in the physical world; it plays a role in the
pushing, pulling, throwing, and pouring that we do in order to live in the world. But the sense-ofmechanism also has some functions that apply more obviously and directly to physics learning: It
allows us to judge the plausibility of possible physical events, make predictions, and it plays a
crucial role in our construction of explanations of physical events.
The sense-of-mechanism consists of knowledge elements that diSessa calls “phenomenological
primitives” or just “p-prims” for short. They are called “primitives” because, as I have already
explained, p-prims constitute the primitive level of our explanations of physical phenomena. In
Chapter 1 I described an example situation in which someone places their hand over the nozzle of
a vacuum cleaner and the pitch increases. A typical explanation of this phenomenon, diSessa tells
us, is that the vacuum cleaner must work harder to overcome the resistance that is applied by the
hand. This boils down to an appeal to what diSessa calls OHM’S P -PRIM as the basis for the
explanation. In OHM’S P -PRIM , the situation is schematized as having an agent which works against
153
some resistance to produce a result. The idea is that OHM’S P -PRIM provides the primitive basis for
this typical explanation; the explanation goes precisely this deep and no deeper.
The “phenomenological” part of “p-prim” also merits some comment. P-prims are described
as phenomenological because they develop out of our experience in the physical world. We have
many experiences in the physical world, pushing and lifting objects, and p-prims are abstractions
of this experience. Furthermore, once they are developed, we come to see p-prims in the world. In
sum, p-prims are basic schematizations of the physical world that we learn to see through repeated
experience in the world.
Force and Agency
OHM’S P -PRIM
Constraint Phenomena
SPONTANEOUS RESISTANCE
SUPPORTING
FORCE AS MOVER
GUIDING
BLOCKING
DYING AWAY
Balance and Equilibrium
DYNAMIC BALANCE
ABSTRACT BALANCE
Figure Chapter 5. -2. A sampling of p-prims.
The variety of p-prims
In order to give a feel for the variety and scope of phenomena covered by the sense-ofmechanism, I want to discuss a selection of p-prims (refer to Figure Chapter 5. -2). I begin with
some p-prims from what diSessa calls the “Force and Agency” Cluster. OHM’S P -PRIM is an
example of a p-prim in this cluster. Recall that in OHM’S P -PRIM a situation is schematized as
involving some agent that works against a resistance to produce a result. A related p-prim in this
cluster is SPONTANEOUS RESISTANCE . The resistance in SPONTANEOUS RESISTANCE is different than that
in OHM’S P -PRIM because it is intrinsic to the patient of some imposed effort. For example, the
difficulty that we have in pushing a fairly heavy object can be attributed to SPONTANEOUS
RESISTANC e.
Compare this to the resistance that is imposed by a hand in the vacuum cleaner
situation.
I want to mention two other p-prims in the Force and Agency Cluster. The first,
MOVER,
FORCE AS
has clear relevance to physics learning. In FORCE AS MOVER , a push given to some object is
seen as causing a movement of the object in the same direction as the push. In some
154
circumstances, the predictions made by FORCE AS MOVER agree with Newtonian physics, but, in
other circumstances,
FORCE AS MOVER
contradicts the predictions of Newtonian physics. According
to Newton’s laws, an object only moves in the direction of an applied force if the object is initially
at rest, or if the push happens to be in the direction that the object is already moving. Otherwise,
the object will be deflected by the applied force.
The final p-prim I want to mention in the Force and Agency Cluster is DYING AWAY . Like
FORCE AS MOVER ,
this p-prim is associated with the drawing of non-Newtonian conclusions about
the world. The idea behind DYING AWAY is that all motion must, in due time, die away to nothing.
In contrast Newton’s laws predict that, in the absence of any applied forces, objects in motion
continue to move indefinitely.
A second cluster of p-prims pertains to constraint phenomena. These p-prims explain
phenomena by appeal to the constraints imposed by the geometric arrangement of physical
objects. For example, if a rolling ball runs into a wall, the ball stops. We might explain this by
saying that the wall blocked the ball’s motion, or the wall simply “got in the way.” This is an
application of the BLOCKING p-prim. Compare this explanation with one that states that the ball
stopped because the wall applied a force to it.
Another p-prim in this cluster, SUPPORTING , is a special case of BLOCKING in which the motion
opposed, or the motion that would have happened, is due to gravity. Why doesn’t a book placed
on a table fall? Because the table supports it.
The last p-prim from this cluster that I will mention is GUIDING . As an example, imagine a
metal ball rolling in a groove made in a wood surface. It is not surprising to us that the ball follows
the path of the groove. Note, again, that we explain this without appeal to forces; the groove
simply guides the ball because of its geometric nature.
Finally, I want to mention two p-prims from the “Balancing and Equilibrium” Cluster. The
first of these p-prims is DYNAMIC BALANCE . A situation involving two equal and opposite forces
would likely be explained by appeal to this p-prim. diSessa contrasts this p-prim with a second
that he calls ABSTRACT BALANCE. In ABSTRACT BALANCE, the balancing of the quantities involved is
required either by the definition of these quantities (as in one kilogram is 1000 grams), or because
of universal principles (such as the conservation of energy). As diSessa says: “abstractly balancing
things should or must balance; dynamic balancing is balancing by accident or conspiracy.” I will
discuss the relation between these p-prims and balance-related symbolic forms below.
155
A mechanism for p-prim activation
The above discussion is designed only to provide a feel for the scope of the sense-ofmechanism. diSessa’s list is somewhat longer and he suggests where many p-prims exist beyond
those that he names. My listing of p-prims will end here, however, and I will instead move on to
another piece of diSessa’s account of the sense of mechanism. To this point, I have not said very
much about the mechanism that determines which p-prims gets used at which time. So far, the
only mechanism that we have is “recognition,” p-prims are just recognized in circumstances.
But diSessa extends his account beyond the simple statement that p-prims are recognized. The
key question is when and how a p-prim is “cued to an active state.” diSessa argues that the
activation of a p-prim depends on other aspects of the current “mental context,” which includes
what we perceive in the world, what other p-prims are active, and any additional active
knowledge, including “conscious ideas.” Furthermore, diSessa defines two terms that are designed
to provide characterizations of how likely a given p-prim is to be activated. The first of these
terms, “cuing priority,” describes the likelihood that a given p-prim will be activated given some
perceived configuration of objects and events in the world. In effect, cuing priority tells us how
much activation must be contributed by other elements of the mental context in order to cue a
given p-prim in a particular physical context. In cases where the p-prim clearly applies—when it
has a high cuing priority—little additional activation is needed from the rest of the mental
context. Note that the cuing priority of a p-prim is context dependent; it depends on the nature
of the physical circumstance that is currently of concern. However, I will sometimes use the term
“cuing priority” more roughly as an overall measure of the frequency with which a given p-prim is
activated.
The second term concerning p-prim activation that diSessa defines is “reliability priority.”
Reliability priority provides a measure of how likely a p-prim is to stay activated once it is
activated. The point is that, once a p-prim is activated, this activation contributes to a subsequent
chain of mental events that may or may not involve the p-prim continuing to be activated. Taken
together, cuing priority and reliability priority constitute what diSessa calls “structured priorities.”
Although diSessa talks in the language of structured priorities throughout most of Toward an
Epistemology of Physics, he also provides a model of the sense-of-mechanism as a connectionist
network. P-prims are nodes in the network and there are weighted connections between these
nodes. Given this model, cuing and reliability priority can be reduced to behavior of the network
due to the values of these various weightings. Thus, in principle, we could discard the language of
156
structured priorities in favor of descriptions given solely in terms of this connectionist network.
diSessa argues, however, that is worth retaining “cuing priority” and “reliability priority” as
technical terms because these terms provide qualitative characterizations of the connectionist
network that are more easily put in correspondence with human behavior. For example, given the
activation of a high reliability p-prim, a person is more likely to stick to their characterization of a
situation in terms of this p-prim, and to assert a high level of confidence for this characterization.
The development of the sense-of-mechanism
With this account of the sense-of-mechanism in hand, we can begin to talk about how the
sense-of-mechanism develops during the learning of physics. I am going to begin here by
presenting what diSessa says about how this cognitive resource develops, and then I will add some
of my own details about the role that symbol use plays in this development. diSessa describes
three types of changes that the sense-of-mechanism undergoes during physics instructions:
1) Weightings change and the sense-of-mechanism is restructured. One type of development that
occurs is changes in the weights in the connectionist network, which can alternatively be thought
of as changes in the priorities of individual elements. We can imagine that these are incremental
adjustments to the weighting values that occur through repeated experiences in physics
instruction. However, although individual changes may be small, diSessa believes that these
incremental adjustments ultimately lead to a change in the overall character of the p-prim system.
Before any physics instruction the sense-of-mechanism is relatively flat, it has only very local
organization with individual p-prims having connections to only a few others. Certainly some pprims have higher priorities than others, but there are no central p-prims with extremely high
priority.
With the development of expertise, this situation changes. diSessa hypothesizes that the
priority of some elements is greatly increased and the priority of others greatly decreased. The
result is a system with central, high priority elements. Thus, there is a change in the character of
the sense-of-mechanism; it undergoes a transition from having little structure to having more
overall organization.
2) New p-prims develop. As students learn to attend to different pieces of the world and their
experience, new elements may be added to the sense-of-mechanism.
157
3) P-prims take on new functions. Finally, the new activities associated with classroom physics and
the new types of knowledge that are acquired provide opportunities for p-prims to perform new
functions. For example, p-prims may come to serve as “heuristic cues” for more formal
knowledge; when a p-prim is cued it can lead more or less directly to the invocation of some
formal knowledge or procedure. In addition, p-prims may play a role in “knowing a physical law.”
For example, we may in part understand Newton’s second law and F=ma through SPONTANEOUS
RESISTANCE,
the tendency of objects to continue moving in the direction that they are already
moving.
The result of all of these changes is a physical intuition that is refined and somewhat adapted
for use in expert physics. diSessa also provides a wealth of details concerning the specific
refinements that he expects to occur, of which I will only mention a few. Some parts of the senseof-mechanism can have little use in expert physics and thus should be substantially suppressed. For
example, in expert physics, constraint phenomena are no longer explained by a simple appeal to
the geometry of objects. Instead, these phenomena must be explained in terms of forces applied
by obstructing objects. Thus, the priority of p-prims, such as BLOCKING, that were previously
associated with constraint phenomena should be greatly decreased, and the cuing priorities of pprims in the Force and Agency Cluster should be increased for these circumstances.
The changes associated with constraint phenomena are typical of a more widespread trend
predicted by diSessa. He conjectures that, in general, the role of agency will be greatly expanded
in the sense-of-mechanism. As students learn to see more and more circumstances in terms of
forces, the range of application and priority of p-prims in the Force and Agency Cluster will be
increased.
Symbol use and changes in the sense-of-mechanism
The above sub-section described the types of changes that we can expect the sense-ofmechanism to undergo during physics instruction. But how does physics instruction cause these
changes? What, precisely, are the experiences that foster these developments? Since physics
students spend such a significant amount of time manipulating equations and solving problems, it
would be comforting to know that some tuning of the sense-of-mechanism occurs during the
symbol use that is typical of these activities. In this sub-section, I will show that this is plausible
and I will give a little of a feel for how I believe it works.
158
Let’s start by considering a hypothetical and clear type of circumstance. Imagine that a
student is working to understand a physical situation and two competing and contradictory pprims are cued to activation. From the point of view of the sense-of-mechanism, this cuing of
conflicting p-prims is a problem. Since, in its naive state, the sense-of-mechanism is rather flat and
only weakly organized, no p-prim has a much higher priority than any other p-prim. Thus, the pprim system is not very good at resolving this type of conflict.
Equations and symbol use can provide a way out for this stymied student. It is possible that,
by manipulating equations, the student can find a solution to the problem and thus resolve the
conflict. If this happens, then it is likely that the priority of the “winning” p-prim will be
incrementally increased, and the priority of the losing p-prim will be incrementally decreased.
Through such experiences the sense-of-mechanism can be nudged toward alignment with expert
intuition.
As we have seen, students frequently stated two conflicting intuitions when working on the
Shoved Block task. These two intuitions were: (1) The heavier block experiences a greater
frictional force, thus it slows downs more quickly and travels less far. (2) Heavier things are
“harder to stop” so the heavier block travels farther. In Chapter 4 we saw that Alan stated both of
these intuitions quite clearly:
Alan
And then if the block was heavier, it'd just mean that our friction was greater, and it would
slow down quicker.
________________
Alan
That seems kind of- Something seems strange about that because if you had like a block
coming at you, a big heavy block, and a small lighter block. Two different instances. And
you have this big heavier block coming toward you and you want it to slow down, it
certainly takes much more force for you to slow down the bigger heavier block than the the lighter one.
Bob
Oh wait, that's true!
Alan
Because, even though they're both working on the same frictional surfaces, just logically I
know the big heavier block takes-, I mean football is all about that, that's why you have
these big three hundred pound players.
After Alan stated these conflicting intuitions, Alan and Bob went on to use equations to find the
solution to this problem. As we have seen, they derived the following equation and noted that it
did not depend on the mass:
a = gµ
Alan
Right, so, actually, they should both take the same.=
Bob
=Wait a minute. Oh, they both take the same! [Surprised tone]
…
Bob
So, no matter what the mass is, you're gonna get the same, the same acceleration.
159
So, in this little episode, Alan and Bob have resolved this conflict between two intuitions in a way
that is a little surprising to them. It turns out that neither intuition is exactly right, both blocks
travel exactly the same distance.
Now I want to talk about this episode in terms of p-prims. The first intuition can be
considered to be a somewhat refined application of
FORCE AS MOVER ;
the presence of the frictional
force causes the block to slow down. The second intuition involves the SPONTANEOUS RESISTANCE pprim; the block has an intrinsic resistance to changing its speed by slowing down. In addition,
both intuitions probably involve the activation of OHM’S P -PRIM since there is an effort working
against a resistance to produce a result. But this last p-prim is applied somewhat differently in each
instance. In the case of the first intuition, the change in intrinsic resistance is not a salient aspect of
the difference between the heavy block and light block situations.
It is worth noting that both of these intuitions already involve somewhat refined applications
of the sense-of-mechanism. The use of FORCE AS MOVER to account for changes in speed rather than
as an explanation for motion in some direction constitutes a refined use of this p-prim. More
dramatically, it is likely that complete novices would explain this motion by an appeal to entirely
different p-prims such as DYING AWAY . If we apply DYING AWAY to explain the shoved block then we
state that the motion dies away simply because that is what motions do. In contrast, Alan and Bob
are capable of attributing the slowing down of the block to a particular agent, a force applied by
the table. This suggests significant progress in the direction of expertise.
So can we say anything about what effects this brief experience may have on Alan and Bob’s
senses-of-mechanism? First notice that this episode is not exactly the same as the hypothetical
circumstance described above, in which the conflict between two p-prims was resolved in favor of
one of them. In this case, symbol use does not choose between two competing p-prims, instead it
suggests that both of these intuitions have some validity.
One further observation from my data can help here. It turns out that the FORCE AS MOVER
intuition was somewhat more common and was always produced before the SPONTANEOUS
RESISTANCE
intuition. Furthermore, only one pair (Mike and Karl) expressed a preference for the
second intuition, and even that pair generated the FORCE AS MOVER intuition first. Of the other
pairs, only Alan and Bob generated the SPONTANEOUS RESISTANCE intuition without some explicit
prompting on my part, though all pairs were quick to acknowledge the plausibility of this second
intuition.
160
I therefore speculate that this episode may have the effect of incrementally increasing the
priority of the SPONTANEOUS RESISTANCE p-prim. In particular, I hypothesize that this experience
may increase the cuing priority of this p-prim for cases in which the mass of an object can be
treated as an intrinsic resistance. This is a rather satisfying outcome since the resistance of masses
to changes in motion is one of the fundamental aspects of Newton’s laws. In the next passage, Bob
sums up the results of this problem. As he does, note that he particularly emphasizes the validity
of the SPONTANEOUS RESISTANCE intuition:
Bob
So, well, when there isn't like equations down there sort of, I have a tendency just to sort
of say: oh yeah, it just=
Alan
=Right.=
Bob
=it would slow down quicker. But, so if - I mean, you do - That's what physics is, is you
look at the equations and figure it out. But, um, okay=
Bruce
=So, what do you think of that result that they slow - that they take the same amount of
time to stop?
Bob
Um=
Bruce
=Is it surprising? Is it not so surprising in retrospect?
Bob
In retrospect it isn't too surprising. But, um, I mean, I guess, I thought (2.0) Why did I
think that the heavier block would slow down quicker? I (st…) just wasn't, I mean. You're
pushing on it - pushing against with a greater force so (…) it seems like it would slow down
quicker. And that was my first thought. Um, but since it is bigger you obviously need a
greater force to stop it. So, I wasn't - I just wasn't thinking about that. I was just thinking:
Oh well, you got two blocks. One's experiencing a greater force and one, one - so this
one's gonna slow down quicker. But, I mean they're not equal. This one's bigger than this
one. Um, so, it isn't too surprising. I mean, it makes sense that, that you need to push, I
mean, you got this big boulder coming at you, you have to push harder than if a pebble's
rolling at you. Um,, To stop it.
In the above passage, Bob starts off by telling us what his “first thought” was, that the heavier
object should slow down faster. But then he goes on to tell us that there is a second effect that
must be considered and that, in retrospect, “isn’t too surprising.” He explains this second intuition
with an example: “it makes sense that, that you need to push, I mean, you got this big boulder
coming at you, you have to push harder than if a pebble's rolling at you.”
The point here is that Bob ends his summary by essentially emphasizing the importance of the
intrinsic resistance of masses. Although the SPONTANEOUS RESISTANCE p-prim did not “win” in this
problem—the outcome was a draw—from Bob’s point of view the moral is that the effects of
SPONTANEOUS RESISTANCE
must not be overlooked. Given his experience in this episode, it is
plausible that, in future episodes, Bob will be slightly more likely to see masses as having intrinsic
resistance. The priority of the SPONTANEOUS RESISTANCE p-prim has been incrementally increased.
Of course, it would be possible to give many competing analyses of this episode. The above pprim account is very speculative, and I have not tried to argue against competitors. This is
161
appropriate since my purpose here is only to illustrate, with an example, a possible mechanism
through which physical intuition could be altered through symbol use.
As a slight aside, I want to also mention that the use of equations can do more than adjudicate
between two p-prims; in addition, the use of equations will also tend to cue p-prims in places
where they might otherwise not be cued or shift how they are applied. This added cuing of certain
p-prims should lead to the increased priority of these p-prims. As an example, recall that during
their solution to the Shoved Block problem, Alan saw one of the written equations as expressing
BALANCING :
µmg=ma
Alan
What Bob didn't mention is that this is when the two forces are equal and the object is no
longer moving because the force frictional is equal to the force it received.
As I have pointed out, the use of BALANCING here is not really appropriate. Nonetheless, this does
illustrate how equations can lead to the activation of p-prims that might otherwise have not been
employed.
I believe that there are two important morals to be derived from the story of Alan and Bob’s
work on the Shoved Block problem. The first of these morals I have already suggested. Alan and
Bob’s experience here can lead to changes in the weighting of connections in the sense-ofmechanism; symbol use can have the effect of increasing the priority of certain p-prims and
suppressing others.
The second moral is a little more subtle. My telling of the above story essentially presumes
that Alan and Bob do not have a fully developed (expert) intuition at the time of the episode. In
particular, I suggested that their sense of SPONTANEOUS RESISTANCE may be a little weaker than that
of an expert physicist. However, even with this relatively unrefined physical intuition, Alan and
Bob were able to successfully complete this problem and produce the correct answer! So, from the
point of view of solving problems, it does not seem to matter very much that their physical
intuition was not fully developed.
This observation leads to the following question: How much work should the sense-ofmechanism and other aspects of physical intuition be able to do? It seems reasonable that
physicists will have an increased priority for SPONTANEOUS RESISTANCE . But is it necessary for the
sense-of-mechanism to become so finely tuned that it can not only activate both SPONTANEOUS
RESISTANCE
and FORCE AS MOVER , but that it can actually produce the result that these effects
precisely cancel?
162
Clearly there are limits to what we need from our physical intuition. Furthermore—and this is
the big point I have been working toward—these limits are in part determined by what symbol
use is capable of doing. Since, Alan and Bob can always rederive this result, it does not need to be
wired into their sense-of-mechanism.
This is akin to the external representation as “crutch” story that I told in the first chapter.
Symbol use can act as a support that allows certain aspects of the sense-of-mechanism to atrophy,
and that requires other aspects to be strengthened. More generally, the moral here is that the
character of physical intuition will depend on specific characteristics of the practice of symbol use
employed. This is true not only because weightings are adjusted through symbol use, but because
the physical intuition of physicists is adapted specifically for the purpose of complementing and
functioning in the activity of symbol use. Thus, the cluster of resources that corresponds to
physical intuition is not necessarily refined so that it approaches Newton’s laws as a limit. Instead,
there is the possibility that this refinement is idiosyncratic, with the idiosyncrasies dependent on
the specific capabilities and requirements of symbol use. I will return to this point in Chapter 7.
The Development of New Resources and the Relationship
Between Forms and P-prims
To this point, all of this discussion of the development of intuition has been in terms of
diSessa’s model of the sense-of-mechanism. This means that I have really only been talking about
changes inside the “intuitive physics” oval in Figure Chapter 5. -1. The story has been that, with
the development of expertise, the sense-of-mechanism gets refined and adapted for new roles. In
this section I want to talk about the new cluster of resources that develops for symbol use in
physics, what I have called “algebra-physics.”
I want to start by pointing out that the notion that the sense-of-mechanism can be refined
during symbol use is, from a certain point of view, slightly bizarre. Prior to physics instruction, the
sense-of-mechanism was built and tuned through experiences in the physical world. Thus, the
advent of physics instruction means not only that there is a whole new set of pressures acting on
the sense-of-mechanism, these pressures are of a new kind. Furthermore, it appears that p-prims
are going to be called upon to perform new types of functions.
These observations lead to some questions: How far can the sense-of-mechanism be stretched
before it should not even be called the sense-of-mechanism anymore? If p-prims are taking on
radically new functions and being cued in radically new circumstances, should they still be called
163
p-prims? Notice that, throughout the above discussion, I always talked as if the individual p-prim
elements themselves were unchanged. What changed were the weightings of connections between
these elements and the jobs these elements were to perform. But isn’t it also possible that the
character of individual elements could change?
In building models of the development that occurs during physics learning, there are thus
several possibilities that we could pursue with respect to modeling the development of the senseof-mechanism. Consider the following simple palette of modeling possibilities:
1. The sense-of-mechanism is only tuned and adapted. In this model, individual elements remain
essentially unchanged. Weightings change and the system is adapted for new uses.
1. The sense-of-mechanism mutates into something new. It is possible that the sense-of-mechanism
could be transformed into something that is different enough that it is not even appropriate to
call it the sense-of-mechanism anymore. Even though the new system may develop out of this
old one, the individual elements may be substantially different than p-prims. This model
presumes that the sense-of-mechanism, as we knew it prior to physics instruction, no longer
exists.
1. The sense-of-mechanism continues to exist, but new related resources develop. In this model, a
new cognitive resource develops from prior resources and it maintains links to these prior
resources. Since the prior resources continue to exist, they may be tuned and adapted in the
manner of model (1).
I believe that I have already made it clear that I plan to adopt model (3): Algebra-physics
requires the development of new cognitive resources. More particularly, I plan to argue that forms,
in part, develop from p-prims and they retain close ties to p-prims. My reasons for choosing this
particular model (and the reasons that I believe it is the most accurate model) are rooted in the
argument I began in Chapter 3 and which I will complete in the next two chapters. In particular, I
have been arguing that physics learning involves the development of knowledge elements that I
call symbolic forms and that these knowledge elements have certain specific properties. Some of
these properties, such as the inclusion in forms of the ability to see symbol patterns, distinguish
forms from p-prims. Furthermore, as we will see in a moment, although some forms bear a close
relation to specific p-prims, the majority of individual forms do not line up with individual p-
164
prims. Thus, the view I have been arguing for asserts the existence of a new cognitive resource that
bears some relation to the sense-of-mechanism, but nonetheless is substantially different.
Intuitive Physics
(p-prims)
Algebra-Physics
(forms)
Figure Chapter 5. -3. Forms develop in part out of p-prims but the sense-of-mechanism continues to exist.
In addition, there are many reasons to believe that the sense-of-mechanism continues to exist
in something close to its original form. Even expert physicists must continue to participate in the
same old activities in the physical and social worlds that require the sense-of-mechanism. They
still have to push and pull objects. And they still have to participate in the usual informal discourse
about the behavior of the physical world. For this reason, I believe that it makes sense to talk as if
the sense-of-mechanism continues to exist in a form very close to that described by diSessa.
Forms are abstracted, in part, from the symbolic world of experience
So my view is that a new cognitive resource develops that includes symbolic forms as elements.
Forms are descended from and maintain fairly direct connections to p-prims, as well as to other
previously existing resources. Furthermore, I want to emphasize that I think of forms as being very
similar knowledge elements to p-prims. I take them to be nearly primitive knowledge structures
that involve simple schematizations of situations.
But now I want to point to some of the differences between forms and p-prims, differences
that should clarify why I choose to treat symbolic forms as new knowledge. Forms differ from pprims because they are not abstracted only from experiences in the physical world, they are
abstracted from the worlds of physical AND symbolic experience, and can be cued in both.
What does it mean that forms are “abstracted out of symbolic experience?” I will not provide
much detail concerning this process beyond the statement that I intend it to be much like the
abstracting of p-prims out of physical experience. For illustration, first consider a p-prim example:
People have repeated experiences in the physical world pushing on objects that are at rest to get
165
them moving. And when you push on something at rest, it moves in the direction that you push it.
These repeated experiences can lead to a simple schematization of these similar events of the sort
embodied in the FORCE AS MOVER p-prim, which says that pushed objects move in the direction of a
push. Similarly, students have experience in the symbolic world adding numbers onto other
numbers. And when they add a number onto another number, they get a result that is bigger than
the original number. This can lead to the abstraction of the BASE +CHANGE form.
Primarily, I intend for my discussion in Chapter 3 to provide support for this view, but I will
remind the reader of a brief example here. Recall that when working on the Buoyant Cube
problem, Mark asserted that the weight of the ice and the weight of the displaced water (the
buoyant force) had to be equal for the following reason:
W w = Wi
Mark
We know one is gonna float. And then just sort of, I guess, I don't know from like all our
physics things with proportionalities and everything, that we'll have to have something
equal, like the buoyancy force should be equal to the weight of the block of ice. [g. pulling
down block]
As I mentioned earlier, Mark is arguing simply from an appeal to his experiences of symbol use, he
says “we have to have something equal.” When students enter physics instruction, DYNAMIC
BALANCING
is a primitive element of the way they understand and explain physical phenomena. For
situations in which there is no motion because opposing forces are in balance, no further
explanation is necessary. In this example, however, Mark appeals to a new sort of intuition,
intuition that is rooted in symbolic experience.
Similar stories can be told about many symbolic forms. Let’s think about the
form for a moment. Certainly
DEPENDENCE
DEPENDENCE
has some roots in physical experience. We have
experiences in which the outcomes of events depend on aspects of the physical world. But we also
have relevant experiences in the symbolic world. For example, we can plug numbers into
expressions and get out numerical results. And these results depend on the values you plug in; if
you plug in a different number, you get a different result.
Hypotheses concerning relations between individual forms and p-prims
I want to take a moment to hypothesize about the relations between specific forms and pprims. To begin, it should be clear to the reader that I intend to suggest a close relationship
between forms and p-prims that have the same or similar names. For example, the
form is supposed to be closely related to the
DYING AWAY
166
DYING AWAY
p-prim. Similarly, the BALANCING form
bears a close relation to the DYNAMIC BALANCING p-prim. The hypothesis is that, in such cases, the
individual form is a direct descendant of the p-prim and that activating the form will have a
strong likelihood of activating the p-prim.
I also hypothesize that there are connections of varying strengths between forms and p-prims
in cases where there is not a simple correspondence between individual elements. In addition,
there are connections between individual forms and other previously existing cognitive resources.
Here, proceeding by cluster of forms, I will speculate about some of the stronger connections.
The Competing Terms Cluster. Across the board, forms in the Competing Terms Cluster will be
closely tied to Force and Agency p-prims. A key observation here is that these forms involve what
diSessa calls the same “attributes” or “base vocabulary” as Force and Agency p-prims. What this
means is that these particular forms and p-prims tend to break down a physical situation into
similar components—both involve agents, usually forces, working for results. Once a form or pprim with a given base vocabulary is activated, the likelihood of elements with the same base
vocabulary being activated is increased. In addition, there will be close connections between some
forms in this cluster and Balance and Equilibrium p-prims, such as between the BALANCING form
and the DYNAMIC BALANCING p-prim.
The Proportionality Cluster. The genetic roots and connections of Proportionality forms are
complex and multifarious. First, these forms will be tied to Force and Agency p-prims. In
particular, I hypothesize that PROP+ and PROP- are strongly connected to physical notions of effort
and resistance. To illustrate, consider the equation F=ma rewritten as a=F/m. The right side of this
equation can be seen in terms of the PROP- form—the acceleration is inversely proportional to the
mass. The hypothesis is that the use of PROP- here will tend to activate the SPONTANEOUS
RESISTANCE
p-prim with the mass seen as an intrinsic resistance that resists acceleration. In addition,
the right hand side may be seen in terms of
PROP-
PROP+.
Taken together, the activation of PROP+ and
may tend to cue OHM’S P -PRIM with, in this case, force taken as the effort, mass as the
resistance, and acceleration as the result.
In fact, the very same story may be told about the equation associated with Ohm’s
Law—V=IR—the law for which OHM’S P -PRIM is named. If we rewrite this equation as I=V/R it
says that the current, I, is proportional to the voltage, V, and inversely proportional to the
electrical resistance, R. In this case, the voltage can be thought of as the “effort” and, of course, the
167
electrical resistance as the “resistance.” Again, the point is that the
PROP+
and PROP- forms will be
connected to p-prims involving effort and resistance.
It is important to note that Proportionality forms also have their roots in cognitive resources
other than the sense-of-mechanism. For example, I noted that some researchers have made
“qualitative proportionality” relations a central element in their accounts of how people reason
about physical quantities. Proportionality forms will remain connected to these quantity-related
resources.
The Dependence Cluster. The connections of Dependence forms are also complex and diverse. As
I will argue in Chapter 11, we can understand DEPENDENCE as subsuming a wide range of
asymmetric relations that can exist between entities of a variety of types. For example, one event
can cause another event, or one quantity can determine the size of another quantity. And such
relations can be found in many p-prims. In OHM’S P -PRIM , the size of the result depends on the
both the amount of effort and the amount of resistance; in
FORCE AS MOVER
the amount of motion
depends on the size of the force; and in GUIDING the direction of motion depends on the shape of
the constraining geometry. To a varying extent, Dependence forms should be tied to all of these
p-prims.
This wide applicability of the DEPENDENCE form can be understood if we think of
DEPENDENCE
as similar to what diSessa calls a meta-p-prim—a p-prim that embodies a higher level abstraction
that is common to many p-prims. Furthermore, it is interesting to note that, because dependence
lives at a higher level of abstraction, much of the nuance associated with the diversity of p-prims is
lost. A similar loss of nuance can be seen in the PROP+ and PROP- forms—proportionality is a more
generic relation than that which holds, for example, between an effort and result.
I believe that this washing out of nuance is a fundamental feature of the move from intuitive
physics to algebra-physics. One of the hallmarks of expert physics practice is its ability to quantify
the entirety of the physical world; everything is described in terms of numbers and relations
between numbers, and all equations look the same whether the quantities that appear are forces or
velocities. I call this tension between the homogenizing influence of algebra and the nuance
inherent in intuitive physics the “fundamental tension” of physics instruction. The set of forms, as
I have listed them, constitute the end product of this tension; much of the nuance inherent in
intuitive physics is lost, but more distinctions are preserved than those inherent in the syntax of
symbolic expressions.
168
The Terms are Amounts, Multiplication, and Coefficient clusters. Forms in the remaining three
clusters primarily have their origins in cognitive resources other than the sense-of-mechanism. The
PARTS -OF -A-WHOLE
form is an example. As I have mentioned, similar knowledge structures were
posited to account for arithmetic problem solving behavior by the “patterns” researchers. Thus,
this form will have primarily developed from and have connections to cognitive resources in what I
have called the “algebra problem solving” and “quantity” resource clusters. There are some
exceptions among these forms, however, and a noteworthy example is the
SAME AMOUNT
form. In
addition to being strongly connected to algebra problem solving resources, I believe that SAME
AMOUNT
is tied to the ABSTRACT BALANCING p-prim. An example situation that might activate this
form and p-prim is when the principle of Conservation of Energy is applied. If we write
Efinal=Einitial, then this relation between the initial and final energies can be seen in terms of SAME
AMOUNT
and understood as required by ABSTRACT BALANCING.
This concludes my discussion concerning the connections between individual symbolic forms
and previously existing resources. Before moving on to the final section of this chapter, however, I
want to observe that p-prims in the Constraint Cluster were not mentioned at all in this discussion
of connections between individual forms and p-prims. This is for the simple reason that, as
hypothesized by diSessa, Constraint p-prims play little role in expert physics. As I have
mentioned, with expertise the physical circumstances associated with these p-prims must come to
be understood in terms of the action of forces.
Finale: Symbol Use and the Development of Expert Intuition
Let’s summarize where we have been in this chapter. First, we began with “snapshots” of
cognitive resources and I focused on the relation between two of these resources. One, the senseof-mechanism, exists prior to any physics instruction. The second resource—the form-device
system—is a new resource that develops during physics instruction. My goal was somehow to
interpolate between these two resources to describe some features of the development of physics
knowledge. In this regard, I made a couple of points. First, I described the sense-of-mechanism
and how it might develop because of symbol use. I stated that the experiences of symbol use
would lead to certain elements being reinforced and I argued that the sense-of-mechanism would
have to develop so as to be adapted for the specific roles required of it by symbol use. Second, I
proposed some specific relations between the sense-of-mechanism and the new resources. I argued
169
that the form-device system develops in part from the sense-of-mechanism. In addition, I
maintained that the sense-of-mechanism continues to exist in a refined form and is connected to
these new resources. In particular, I argued for a strong relation between p-prims and forms. Now,
with this broad picture in hand, I can clarify some points I made earlier and I can make some final
observations.
In the early part of this chapter I talked about some of the details concerning how I expect the
sense-of-mechanism to be refined during symbol use. I argued that symbol use can have the effect
of reinforcing some p-prims and suppressing others. Now I am in a position to elaborate just a
little on how this works. First, we know that forms can be cued directly by equations, they are
essentially “seen” in expressions. Furthermore, I have also asserted that the activation of forms can
lead to the activation of related p-prims. This equation–form–p-prim chain is thus a
particular—and rather direct—mechanism by which p-prims can be activated during symbol use.
As an example, take the episode in which Alan sees the equation mgµ=ma in terms of the
BALANCING
form. Once this form is cued, this can lead directly to the activation of the DYNAMIC
BALANCING
p-prim. The upshot of such an episode is that both the form and p-prim are reinforced.
The broad picture can also help me to clarify my stance on another issue: What is expert
“physical intuition?” It will be my stance that both the refined sense-of-mechanism and the newly
developed form-device system constitute components of an expert’s physical intuition. I realize
that it may seem a little irregular to think of forms as part of physical intuition since they are
derived in part from experience with symbol use. Nonetheless I believe that it is appropriate to
think of forms as part of the very roots of expert intuition. I will argue for this stance in Chapter 7.
Finally, with the model laid out in this chapter, I am now in a position to state one of the
fundamental arguments that I will make in this document. The idea is that the forms available in a
symbolic practice and the frequency of use of those forms have a strong influence on the nature of
expert intuition. If a form gets used frequently, then that form and related p-prims both get
reinforced. This is one way to think about how the details of a symbolic practice, such as algebraphysics, can determine the nature of expert physical intuition. Algebra-physics involves a
“vocabulary of forms,” and the nature of these forms and their frequency of use is indicative of the
character of expert intuition that is associated with algebra-physics.
Furthermore, this argument will provide one of my fundamental points of leverage in
comparing algebra-physics and programming-physics. What I propose is to actually count the
frequency of individual forms and devices as they appear in my programming-physics and
170
algebra-physics data. When this is done, I will essentially have characterizations of these two
practices that I can contrast. In the next chapter, I will show how I went about obtaining these
counts for algebra-physics, and I will present the results.
171
Chapter 6. Systematic Analysis of the Algebra
Data Corpus
To this point, I have devoted myself to describing my model of symbol use and arguing more
broadly for my theoretical orientation. I have presented my two main theoretical
constructs—symbolic forms and representational devices—and I have talked about how these new
resources develop out of existing prior resources during physics learning. Although the main goal
of the preceding chapters was to lay out the theoretical view, my presentation was not totally bare
of support for that view. I argued against competing viewpoints where appropriate. And I
employed examples from my data corpus to illustrate my points and to add to the plausibility of
my position.
In this chapter, I reveal more of the details of the methodology used to make contact with my
data corpus. This will include a more comprehensive portrayal of the participating subjects as well
as a description of the methods used for collecting and analyzing the data corpus. In brief, my
methods involved first videotaping students solving a range of physics problems and transcribing
the resulting videotapes. A sizable subset of the resulting transcripts were then systematically
coded. This required going through the corpus equation-by-equation and utterance-by-utterance
looking for forms and devices.
There are two major reasons that the effort necessary to produce this systematic analysis is
warranted. The first reason is that it is through the systematic analysis that the lists of forms and
devices, as I presented them in Chapters 3 and 4, were developed. Through an iterative process of
coding and recoding, these lists were refined until they were just sufficient to account for all
required aspects of the data. Thus, in recounting for the reader how the systematic analysis was
performed, I am essentially explaining how I arrived at the lists of forms and devices.
Second, the results of the systematic analysis paint a particular kind of picture of the data
corpus. These results can provide the reader with a sense for broad features of the data, such as
how regularly students constructed and interpreted equations, and how frequently each type of
symbolic form and representational device appeared. In addition—though more
speculatively—these frequencies can be taken to provide a picture of algebra-physics knowledge.
One of the goals of this work is to understand the character of expert intuition and, as I began to
argue in the previous chapter, the form-device system is one important component of expert
intuition. The underlying presumption here is that the frequency with which specific forms and
172
devices are used reflects the priority of these elements—in diSessa’s technical sense—within expert
intuition.
There are, however, some limits to the analysis I will present here. First, the analysis process I
describe in this chapter does not provide an exact means of extracting categories from a data
corpus, and I did not develop a specification so that my results could be corroborated by another
coder. Although I believe that my lists of forms and devices are supported by my data and this
analysis, the production of these categories was still strongly driven by theoretical concerns and
intuition; as I will describe here, it was a highly heuristic process. Even so, I hope that the reader
will understand this analysis as a useful first step toward future, more exact and replicable
methods of employing analyses of this sort to identify knowledge elements.
There are also limits in this analysis, not only due to the techniques I employed, but also due
to the nature of the data corpus itself. Most notably, the relatively small number of subjects and
limited task selection limit the extent to which we can expect to generalize from this corpus to
make claims about the nature of algebra physics. Even with these limits, I believe that the results of
this analysis provide a useful window into algebra-physics and, when similar data is obtained for
programming-physics, the results will allow a revealing comparison of these two practices.
In the following sections of this chapter I will describe how and what data was collected, how
that data was analyzed, and the results of the analysis.
About the Algebra Data Corpus
Subjects
All of the participants in the study were students at U.C. Berkeley and were currently enrolled
in “Physics 7C.” Physics 7C is a third semester introductory course in a series entitled “Physics for
Scientists and Engineers.” The title tells us at whom this course is targeted; it is for students who
intend to major in a “hard” science or engineering discipline. In contrast, students that intend to
go to medical school or who are headed toward a “softer” form of engineering (e.g., architecture)
generally enroll in the “Physics 8” series.
Two preceding courses in the Physics 7 series, Physics 7A and 7B, are prerequisites for
participation in 7C. Physics 7A is primarily an introduction to classical mechanics and 7B covers
topics in electricity and magnetism, as well as some thermodynamics. The third semester, Physics
173
7C, covers optics and electromagnetic waves, in addition to some modern physics topics,
including special relativity and quantum physics.
It is thus the case that the subjects in this study were fairly advanced physics students at a
prestigious university. Nonetheless, these students were certainly not experts in physics; rather,
they were at an intermediate level of expertise, somewhere between true novices and practicing
physicists. For several reasons, this intermediate level of expertise was necessary for this study.
First, since I wanted to make generalizations about the nature of algebra-physics practice, I needed
students that were already somewhat on their way to learning this practice. However, I also
wanted to be able to speculate about the development of algebra-physics knowledge. For this
reason, I did not want to look at true experts; I needed students that were still at an early enough
stage of their studies that I could expect to observe the types of difficulties encountered by
students in introductory courses.
Finally, there was an additional pressure to not have subjects closer to the novice end of the
spectrum. It was of great practical benefit to have students whose behavior on the given tasks was
relatively stable over the period of the semester. Since the tasks were primarily drawn from
mechanics topics, the new material learned in Physics 7C should not have contributed too
significantly to the ability of students to accomplish the tasks. This meant that each pair could
spread their sessions over the semester in a manner convenient with their schedules.
To recruit participants, I went to a Physics 7C lecture early in the semester and asked for
volunteers. Students were told that they would be paid $5 an hour for their work, and interested
students filled out sheets with their names and telephone numbers. They were permitted to
volunteer in pairs or singly. Later, I shuffled the stack of sheets and telephoned students at
random. I scheduled 7 pairs hoping that at least 5 would complete the study. In fact, precisely 5
of the 7 pairs completed all of the tasks.
In total, 21 students volunteered to participate, 17 men and 4 women. (This high percentage
of male volunteers is not surprising given that the class was predominantly male.) Of the 14
students that began the study, 11 were men and 3 women, with 4 pairings of two men, and 3
pairings of a man and a woman. The specific pairings chosen were almost entirely dictated by the
students’ scheduling constraints. Unfortunately, the two pairs that dropped out of the study were
both male/female pairings. One of these pairs dropped because the male had to leave school for
personal reasons. The other pair asked to drop out, pleading that they were simply too busy. In
fact, both members of this latter pair seemed very uncomfortable during the two sessions that
174
they attended, and I was contemplating asking them if they wanted to cease participation in the
study.
It may seem as though five pairs is a relatively small number for this study, but even this
number pushed the boundaries of what was practical. As will become clear shortly, the methods
employed required the transcription and analysis of a large number of hours of videotape, for both
the algebra and programming portions of the study. Furthermore, the small number of subjects
does not pose a significant difficulty for most of the claims that I will make. At no point will I try
to make claims concerning differences and similarities between individual students or pairs of
students. Instead, all of the (algebra-physics) data is amalgamated into a single set of numbers
before drawing conclusions.
The Tasks
The participants in the study were asked to solve a range of problems and to answer a variety
of questions. These tasks, as presented to the students, are reproduced in full in Appendix B. As
shown in Table Chapter 6. -1, the tasks were divided into four types, administered in the order
shown.
Open-Ended
Main Body
“Quick” Tasks
Cross Training
4 questions
11 problems
7 questions
4 problems
1/2 - 1 hour
3 - 3 1/2 hours
1/2 hour
1/2 - 1 hour
Table Chapter 6. -1. The four main types of tasks in order of execution.
Students began their work in the first session with four “open-ended” questions. These
questions did not ask students to solve specific problems or derive particular equations; instead,
the purpose of these tasks was to get students talking with only very limited prompts. The
questions were introduced with the following written instructions:
These first few tasks are slightly different than the others I’m going to ask you to do later.
For these, you should just explain as much as possible about what happens in the
situations described.
The “Shoved Block” task, which was described in earlier chapters, was amongst the situations
asked about in these open-ended questions. Another of the questions asked students simply to talk
about what happens when a ball is tossed into the air:
175
(1) A ball is tossed straight up in the air. What happens? Make sure to say something
about forces.
Following these open-ended questions, students moved on to the main body of tasks, which
consisted of 11 problems to be solved. The solving of these tasks occupied the vast majority of
student time in the study and work on these tasks was always spread over several sessions. There
were a range of problems here. Some of these problems, such as the Mass on a Spring problem,
were completely typical of problems that students solve during the first semester of introductory
physics. Others were somewhat more unusual. For example, one of these problems asked students
to answer questions about an alternative universe—“The Geeb Universe”—in which the laws of
physics differ from our own.
After the main body of problems students were given two additional types of tasks. The first,
which I simply call the “Quick” tasks, was a series of seven questions that students were specifically
told they should expect to answer quickly. In general, this portion of the sessions provided an
opportunity for me to ask students some specific questions and probe for particular information
without disrupting the normal problem solving activity I wanted to observe during the main body
of problems. For example, one of these questions asked students whether the equation v = vo + at
is “obvious.”
(3) Some students say that the equation v=vo+at is “obvious.” What do you think?
The final type of tasks constituted what I refer to as “Cross Training.” During this portion of
the study, I taught students a modified practice of algebra-physics that was specifically designed
to incorporate some of the features of programming representations. For example, in this
modified practice, an equal sign was sometimes replaced with an arrow that points from one side
of the equation to the other. The students were first given some written instructions that explained
how to solve problems using this modified representational system, then they were asked to solve
four problems using the modified system. The results of the Cross Training will not be discussed
in this document due to limits of space. Interested readers can examine the instructions and tasks
given to students, which are provided along with the other tasks in Appendix B.
For the purposes of the systematic analysis a subset of the above tasks was selected. This was
necessary because it would have been prohibitively time-consuming to analyze the entirety of the
data corpus in detail. In narrowing the tasks for analysis, I included tasks that, for the most part,
are typical of the algebra-physics practice in introductory courses. This is sensible because one of
the purposes of the systematic analysis is to paint a picture of this practice. Use of this criterion
176
eliminated the “Quick” and “Cross Training” tasks from the systematic analysis, as well as some
of the more unusual problems, such as the “Geeb Universe” task. The most unusual problem
included in the systematic analysis was “Running in the Rain.”
A second, minor criterion was also employed to narrow the scope. I wanted to have coverage
from the point of view of the “physical principles” involved in the problems, without too much
redundancy. For example, the main body of tasks included two problems that are traditionally
solved using the principle of Conservation of Momentum. Only one of these two problems was
included in the systematic analysis.
1. Shoved Block
Now the person gives the block a shove so that it slides across the table and then
comes to rest. Talk about the forces and what’s happening. How does the situation
differ if the block is heavier?
2. Vertical Pitch
(a) Suppose a pitcher throws a baseball straight up at 100 mph. Ignoring air
resistance, how high does it go? (b) How long does it take to reach that height?
3. Air Resistance
For this problem, imagine that two objects are dropped from a great height. These
two objects are identical in size and shape, but one object has twice the mass of the
other object. Because of air resistance, both objects eventually reach terminal
velocity.
(a) Compare the terminal velocities of the two objects. Are their terminal velocities the
same? Is the terminal velocity of one object twice as large as the terminal velocity of
the other? (Hint: Keep in mind that a steel ball falls more quickly than an identically
shaped paper ball in the presence of air resistance.)
4. Air Resistance
with Wind
(b) Suppose that there was a wind blowing straight up when the objects were
dropped, how would your answer differ? What if the wind was blowing straight down?
5. Mass on a
Spring
A mass hangs from a spring attached to the ceiling. How does the equilibrium
position of the mass depend upon the spring constant, k, and the mass, m?
6. Stranded Skater
Peggy Fleming (a one-time famous figure skater) is stuck on a patch of frictionless
ice. Cleverly, she takes off one of her ice skates and throws it as hard as she can. (a)
Roughly, how far does she travel? (b) Roughly, how fast does she travel?
7. Buoyant Cube
An ice cube, with edges of length L, is placed in a large container of water. How far
below the surface does the cube sink? (ρice = .92 g/cm3; ρwater = 1 g/cm3)
8. Running in the
Rain
Suppose that you need to cross the street during a steady downpour and you don’t
have an umbrella. Is it better to walk or run across the street? Make a simple
computation, assuming that you’re shaped like a tall rectangular crate. Also, you can
assume that the rain is falling straight down. Would it affect your result if the rain was
falling at an angle?
Table Chapter 6. -2. The eight tasks selected for systematic analysis.
Table Chapter 6. -2 lists the seven tasks selected for systematic analysis, along with the
complete text of each of these tasks. However, notice that part (b) of the Air Resistance problem
has been listed as a separate entry in this table. In what follows, I will consider part (b) of the Air
177
Resistance problem to be a separate task, since students frequently spent a significant amount of
time on this portion of the problem and their work on the two portions was easily separable. Thus,
throughout the rest of this discussion, I will speak in terms of the eight tasks selected for
systematic analysis.
Finally, I want to note that, since the forms and devices we observe are likely to depend
critically on the specific tasks selected, it is important to have a selection that is appropriately
representative. Although I did, as described above, employ some criteria, the ultimate selection of
tasks was still somewhat restricted and arbitrary. This is one of the most important limits on the
conclusions that can be drawn from this analysis and I will discuss this issue further in later
sections of this chapter, and in more detail in Chapter 10.
Details of the experimental sessions
All of the experimental sessions were performed in a laboratory setting. Each pair worked
standing at a whiteboard, and a single video camera was positioned so as to record the movements
of the two students as well as everything they wrote on the board. I operated the video camera
which, at most times, remained fixed in a single location and orientation. Occasionally I
repositioned the camera to account for student movements.
At all times, I did my best to maintain an open and relaxed atmosphere. Students were
encouraged to comment on whatever came to mind, and to air any concerns they might have
about their solutions. In fact, part of the purpose of the open-ended tasks (the first tasks given to
the students) was to encourage a slightly more open attitude to the tasks that followed. In
addition, I believe that having a partner present helped to make students more comfortable. The
hope was that this “open and relaxed” atmosphere would be somewhat characteristic of the group
and pair problem solving that students do outside of their courses—more characteristic than the
“performance” atmosphere that laboratory sessions might tend to engender.
Furthermore, having students work in pairs provided a natural way to encourage them to
explain what they were doing and why. When, in education research, we attempt to study
students while they work independently to solve problems, the students must frequently be
prompted to reflect aloud on their actions. In contrast, when a student works with a partner, the
partner must be consulted on each step of the solution process. Furthermore, using pairs is
certainly a more ecologically valid means of getting students to explain their thought processes.
Presumably, students do not typically reflect aloud as they solve physics problems alone in their
178
dormitory rooms. But discussing work with a partner while solving a problem in tandem is a
natural part of algebra-physics discourse.
Each task proceeded as follows. First, I handed each student a sheet of paper that had the
problem or question written on it. The students could ask for clarification after reading the
question or at any time while they worked on the problem. Once the students had read the
problem and asked any immediate questions, they proceeded to solve the problem and, when
they were done, I asked them a few questions about their solution.
Ideally, I always hoped that students would be able to proceed from start to end of the
solution without interruption. However, I did interrupt students occasionally for various reasons. I
sometimes prompted students for clarification. In addition, when students said that they were
“stuck” or they had spent a significant amount of time exploring what I knew to be an unviable
avenue, I interrupted them and gave them some hints to start them on a more useful track.
It is important to understand that these interruptions and hints pose only a very minor
difficulty for this study. It is true that these hints, by their nature, disrupt the natural flow of
problem solving. However, this study is not centrally concerned with the “flow of problem
solving.” In this systematic analysis, I will not attempt to describe the process by which a solution
is reached. At no time, for example, will I attempt to characterize the order in which equations are
deployed. Instead, I am interested in some specific kinds of events, namely the construction and
interpretation of equations.
What is potentially more problematic is the fact that my prompts to explain solutions, even
though primarily at the end of episodes and generally minimal, could tend to alter the frequency
of these events. This is true, first, because such prompts can directly provoke interpretation events.
In addition—and perhaps more profoundly—these prompts could tend to suggest to students
that I am particularly interested in behavior of this sort.
The reason for including these prompts is that, no matter what I did, the experimental sessions
were not going to be precisely the same as a session with a few undergraduates working together
on a problem set. Furthermore, the bigger danger—I worried—was that students would tend to
be particularly hesitant to voice their thoughts in a laboratory setting, in front of a video camera.
For this reason, as I said, I felt it was appropriate to cultivate a relaxed atmosphere, one open to
reflection and comment. Thus, I may have, in some sense, over-stimulated the production of
interpretations, and this must be kept in mind. However, importantly, this worry does not pose
difficulties for the assertion that students possess the knowledge necessary to interpret expressions.
179
Sessions were typically between 1 and 1.5 hours in length, with pairs participating in between 4
and 6 sessions (refer to Table Chapter 6. -3). The total time spent by all students working on tasks
was approximately 27 hours. When these 27 hours are narrowed to include only student work on
the tasks selected for systematic analysis, the result is 11 hours of videotape.
# of Sessions
Total Hours
Analyzed Hours
Mike&Ken
Jack&Jim
Mark&Roger
Jon&Ella
Alan&Bob
4
5.0
2.0
5
4.5
1.5
5
6.0
2.0
6
6.0
3.0
6
5.5
2.5
Totals
26
27.0
11.5
Table Chapter 6. -3. The time that students spent in the experimental sessions.
Analysis Techniques
The above sections described the nature of the algebra data corpus: the participants involved,
the tasks they performed, and the experimental setting. Now I want to describe the method used
to analyze the resulting videotapes and I will begin with a brief overview of the technique
employed. When I began the analysis, I had not yet narrowed my focus to only 8 tasks, but I did
quickly decide to exclude the Quick Tasks, Cross Training, and one task from the main body
which pertained to electrical circuits. I then made a first pass at transcribing student work on all of
the remaining tasks—a total of about 20 hours of videotape.
Next, one-by-one, I selected individual episodes for coding. Here an “episode” is one pair’s
work on a single task. After an episode was selected, I viewed the corresponding videotape again
and polished the transcript. Then the episode was coded and the coding technique refined so that
the episode was comfortably accommodated. In part, refining the coding technique meant
adjusting the lists of forms and devices so as to appropriately cover the data. Then another episode
was selected and the process iterated. The idea of this initial coding of a few episodes was to refine
the coding techniques to a passable level.
After coding 7 individual episodes in this manner, I felt that the coding scheme was stable
enough to apply to a large portion of the data corpus. At this point I narrowed my focus to 8
tasks, as described above, corresponding to 40 problem solving episodes. All of the corresponding
tapes were watched again, and the transcripts polished. Then these 40 episodes were coded, the
coding techniques refined once more, and then all 40 episodes were coded one final time.
The following sub-sections fill in the details of this analysis technique.
180
Transcription
The first step in analyzing any episode was to transcribe that episode into a written form. For
several reasons, transcription is extremely helpful in performing almost any type of fine-grained
analyses of videotape-based data. First, executing this type of analysis requires reviewing episodes
repeatedly and it is typically faster to read a transcript than to watch a tape, particularly once the
analyst has some familiarity with the videotape. In addition, analyses of the sort I will describe
here require the comparison of key moments within and across episodes. This sort of comparison
is very difficult to do without transcripts. In fact, even if I wanted to review the original videotapes
at every stage of the analysis, I would likely need transcripts to serve as an index to the tapes.
As anyone who has transcribed spoken utterances is aware, verbal communication is somewhat
different than written communication. For this reason, it is necessary to develop some special
techniques to record what people say. To a certain extent, this means inventing an extended set of
punctuation for use in the transcripts. The special notations that I used are described in Appendix
A. In addition, it was also necessary for me to keep track of certain types of non-linguistic
acts—what people do other than what they say. As described in Appendix A, these non-linguistic
acts were also recorded in the transcripts, set off in brackets. I will give an example in a moment.
For the purposes of this study it was extremely important to have an accurate record of
everything that students wrote on the whiteboard. Creating this record posed some difficulties.
One option would have been to simply describe what students wrote and drew on the board
within the written transcript. However, it is often very difficult and time-consuming to describe
diagrams and equations in textual form. Thus I chose to supplement the written transcripts with
drawn images of the board: Any time that an equation was written, it was added to the drawn
image and a more modest notation was made in the written transcript.
The images of students’ board work were created using standard computer-based illustration
tools, such as Adobe Illustrator. More recent versions of these tools have one feature that is
particularly useful in this regard: They allow images to be built up in layers. This was extremely
useful because by building up board images in layers I was able to keep track of successive
additions and modifications to what was written.
I will illustrate how this worked with an example. In some cases, students proceeded in a very
neat and orderly manner, writing one equation under another. In such cases, it was sufficient for
me to simply reproduce the final state of the whiteboard, as it appeared at the end of the problem.
Mike and Karl’s work on the Air Resistance task, shown in Figure Chapter 6. -1, proceeded in
181
such an orderly manner. They worked from the top to the bottom of what is shown, without
modifying or erasing along the way. Of course, this reproduction of their board work does not
contain all information concerning the order in which things were drawn and written, but
combined with the written transcripts it is possible to reconstruct this ordering information.
+
(kv)
f(v)= Fa
m
v
Fg
a(t)= g +
f(v)
m
kv
=0
m
−Fg + kv
= −g +
a(v)=
v=
m
m
g
k
Figure Chapter 6. -1. Whiteboard image from Mike and Karl's work on the air resistance task.
+
+
(kv)
f(v)= Fa
+
(kv)
f(v)= Fa
(kv)
f(v)= Fa
m
Fw
m
m
Fw
Fw
v
Fg
a(t)= g +
v
v
Fg
f(v)
m
mph
kv +50
= −g +
=0
m
−Fg + kv
a(v)=
m
m
v=
g
k
Fg
f(v)
m
h
k +50mp
kv
(v
+
v
)
== −g
+
=0
w
−g + m = 0
m
−Fg + kv
a(v)=
m − V
m
v=
g
w
k
a(t)= g +
f(v)
m
h
k +50mp
kv
(v
+
v
)
== −g
+
=0
w
−g + m = 0
m
−Fg + kv
a(v)=
m −+ V
m
v=
g
w
k
a(t)= g +
Figure Chapter 6. -2. Three successive board images showing Mike and Karl's modifications of their Air
Resistance work to account for wind. Changes made in each panel are marked with gray arrows.
When students erased and modified their work on the board, the situation was somewhat
more complicated. As students made modifications, these were added as layers over their existing
182
work. For example, after completing the Air Resistance task, Mike and Karl modified the solution
shown in Figure Chapter 6. -1 to account for the presence of a wind. Figure Chapter 6. -2 shows
three successive sets of modifications that they made to their work. The changes made in each
panel are indicated with gray arrows.
Figure Chapter 6. -3 contains the segment of transcript that corresponds to the modifications
made in the first panel of Figure Chapter 6. -2. Note that the first and last utterances in this
segment of transcript contain notations concerning changes to what is written on the board. These
written changes are set off in brackets and are preceded by a “w.” written just inside the bracket.
These notations tell us that, during the first utterance, Karl drew a circle around the kv/m term, as
shown in Figure Chapter 6. -2. Similarly, during the last utterance, Karl wrote “+50mph” above
this term. The point of this example is to illustrate how the combination of written transcript and
board image allow the progress of student work to be reconstructed. Note, for example, that it is
not necessary to say in the transcript precisely where the “+50mph” was written on the whiteboard.
Karl
Let me go back to, let me go back to this term. This term right here. Okay. [w. circles
kv/m term] This term right here is the force opposing
Mike
The acceleration
Karl
The acceleration due to gravity. And it's proportional to the velocity of the ball.
Mike
Right.
…
Karl
Well, what I'm saying is that the, this force [g. circled term], this force is relative- This
force is proportional to the velocity. But it's proportional to the velocity of the relative air
and ball system. It doesn't care about the really absolute, the absolute, uh, speed of the
ball relative to some outside observer.
Mike
It's just the relative between the two?
Karl
(This) only cares about the speed relative to between the ball and the air. That's what cThat's what counts this factor. [g. circled term] So, if you have a speed, if you have a
wind that's blowing up initially, then this speed [g. circled term] is the speed relative to
the ground plus say fifty K P H. [w. +50mph)]
Figure Chapter 6. -3. Portion of the transcript that corresponds to the modifications shown in the first panel of
Figure Chapter 6. -2.
Notice that the transcript given in Figure Chapter 6. -3 contains some additional nonlinguistic acts that are preceded by “g.” rather than “w.” These non-linguistic acts are gestures;
they indicate that the student pointed to something written on the board. Such notations are
essential to understanding many student utterances involving written expressions. For example,
when Karl says in this segment:
Karl
Well, what I'm saying is that the, this force [g. circled term], this force is relative- This
force is proportional to the velocity. …
183
it would be difficult to understand what he means by “this force” if we did not know what he was
pointing to as he made this statement.
There is an important subtlety relating to the transcription of gestures. When a person points
to a given location on the whiteboard, how do we know what they are pointing to? For example, if
Karl points right at the vertex of the letter ‘v’ in the kv/m term, he could be intending to point at
this letter, the kv/m term, at the entire equation, or even at a region of the board. Given this
inherent ambiguity of gestures, one might decide that it is more appropriate to transcribe gestures
in a more neutral manner. For example, we could divide the whiteboard into a fine grid and then
just note where in the grid the student pointed. However, I do not believe that it is necessary to
use this extremely neutral method. In most cases, I found that—as in the example above—I could
use other aspects of the context to comfortably determine what students intended to indicate
with a pointing gesture.
Using context to disambiguate the nature of an utterance for transcription is not
unprecedented; we must do precisely the same thing in transcribing speech. For example, if a
student says: “I didn’t hear what you said,” we know to write the word “hear” and not “here,”
because that is what makes sense given the context of the sentence. Imagine the extreme case in
which we restrict ourselves to a purely phonetic rendering of vocalizations. Although this would
likely be useful for some purposes, it would be very difficult and time consuming, and researchers
do not consider it to be necessary for all types of analysis. I take the same stance with regard to
transcribing gestures; I use context—where I can—to determine what a student intended to point
to.
The base-level coding scheme
Once an episode is transcribed in the manner described above, the coding of that episode can
commence. First, note that the goal of the systematic analysis is to code every episode in terms of
symbolic forms and representational devices. But where should we look for forms and devices in a
transcript? Transcripts certainly do not come equipped with annotations that say “Look for
symbolic forms here” and “Look for representational devices here.”
For this reason, there needs to be an intervening step between the production of a transcript of
an episode and the coding of that episode in terms of forms and devices. In this intervening step
in the analysis, which I call the “base-level” coding, the events that must be explained in terms of
184
forms and devices are identified. In addition, the base-level coding provides some basic
information concerning the character of the data corpus.
The base-level analysis is designed to identify the two types of events that must be explained
in terms of forms and devices:
(1) A symbolic expression is constructed by a student.
(1) A student interprets an expression.
Identifying events of type (1) requires an expression-centered analysis of the episode. In order to
find these events, I have to look at every expression that is written and decide if it was
constructed. Events of type (2) require an utterance-centered analysis. To identify these events, I
have to look at every utterance made by each student and decide if it is an interpretation of an
expression. In what follows I describe, in turn, the expression-centered and utterance-centered
components of the base-level analysis.
-mg
F
m
µ
mg
F=µN
f
Figure Chapter 6. -4. Board image from Jack & Jim's work on the Shoved Block task.
Expression-centered coding: Identifying constructed expressions
In the expression-centered coding, I look systematically at every expression that is written in
the episode under analysis. With a few exceptions, every one of these expressions is coded as
remembered, derived, constructed, or as a combination of two or more of these categories. Of
course, we need to have a way of deciding into which of these categories a given expressionwriting event falls, and I will tackle that issue in a moment. But there is an even more basic issue
that must be considered first; namely, I have to specify what counts as “an expression.” To begin,
isolated symbols written on the whiteboard do not count as expressions. For example, if a student
writes the letter ‘m’ on an object in a diagram to indicate that this object has a mass equal to m, I
do not treat this as an expression. However, any sequence of more than one symbol does count as
185
an expression. Thus, there are 3 expressions in the board image shown in Figure Chapter 6. -4; mg,
-mg and Ff=µN are all expressions since they involve more than one symbol, but the isolated
symbols F, m, and µ do not count as expressions. Notice that expressions need not have equal
signs.
There are some more difficult cases to consider in deciding what constitutes an expression to
be coded. As in the example shown in Figure Chapter 6. -2, students sometimes modified
expressions that were already written. In such cases, it is necessary to keep track of both the
original expression and the expression produced by the modification, since forms and devices
could play a role in both events. When actually coding my data, I kept track of modifications
separately. Ultimately, however, I pooled all the data so that expressions produced by
modifications were treated the same as other expressions.
After the expressions in an episode were recognized, each one was coded as remembered,
derived, constructed, or as a combination of two or more of these categories.
Remembered expressions
A remembered expression is one that a student appears to have simply written out “from
memory.” For example, students will often begin work on a task by writing F=ma on the
whiteboard, usually with little comment. When this happens, we can be relatively certain that the
student is not rederiving the expression or reconstructing the expression.
A variety of types of evidence can be brought to bear in determining that an equation is
remembered. If a student begins a problem by immediately writing:
x = xo + vot + 12 at 2
I would code this as remembered. There are several reasons that this is justified: (1) This expression
is relatively complicated and it was written without explanation or much evident prior work. (2) I
am aware that this is an expression that is taught in introductory physics courses. (3) When this
expression is taught it is often written in exactly this manner, using these symbols (x for position,
xo for initial position, etc.) and with the terms in this order.
For further illustration, consider the following two equations and the comments that students
made as they wrote them:
186
v 2f = vi2 + 2ad
Jim
So, I think I recall the equation for this being V F squared equals V I squared plus 2 A D?
[w. Vf2=Vi2+2ad]
________________
x = xo + vot + 12 at 2
Ella
So how does that equation go? It goes like X equals [w. x=]
Jon
X // naught.
// X naught [w. xo]
Ella
Jon
But we don’t need to know that.
Ella
plus V
Jon
T
Ella
naught T plus one half A T squared. [w. vot+1/2at2]
In each of the above cases, the students wrote an equation that is usually taught early on in an
introductory physics course, and they wrote it in a standard form. Furthermore, the equations are
apparently written with little effort. In fact, in the first excerpt, Jim actually tells us that he
“recalls” the equation that he is writing. And Ella starts out by saying “how does that equation
go?”, a clear sign that she is trying to remember an equation that she has used before. Each of
these expression-writing events can confidently be coded as remembered.
Derived expressions
A derived expression is one that is written by applying algebraic rules-of-manipulation to one
or more expressions that are already written on the whiteboard. In the prototypical case, a single
equation is manipulated to produce a new equation, as kx=mg can be manipulated to produce
x=mg/k. Of course new expressions may also be derived by combining multiple lines.
For the subjects in my study, the use of symbolic manipulation to produce new expressions
was a quick and fairly effortless process, and almost universally error-free. Thus, students often
said little during such derivations. For example, when solving the Vertical Toss problem, Mike
and Karl began by writing this expression for the distance
s = vot − 12 gt 2
After doing some work they found an expression for the time that it takes for the ball to reach its
maximum height:
187
v
t= o
g
Then Mike said “plug it back in” and he proceeded to substitute this relation for the time into the
above expression for the distance. He did this with little comment, only mumbling a “V naught”
here and there. When he was done he had written three new lines:
v 
v 
s = vo  o  − 12 g o 
 g
 g
2
v2
v2
= o − 12 o
g
g
2
v
s = 12 o
g
Each of these lines would be coded as derived. The first line was produced by substituting the
relation for time into the expression for distance. Then the second line was obtained by
manipulating the first line, and the third by manipulating the second.
Constructed expressions
Except for some rare and marginal cases that I will discuss in a moment, all remaining
expressions fall into the category of constructed expressions. If a student did not remember an
expression and they did not derive it, then where did the expression come from? The answer is
that they must have just invented it. These invented expressions were coded as constructed.
In the preceding chapters, I described a number of episodes in which students constructed
expressions. Mike and Karl’s construction of an expression for the coefficient of friction that
depends on mass—described in Chapter 2—is a clear example.
µ
µ = µ 1 + C m2
The evidence that this expression was constructed is the complement of the evidence I would use
to argue that an expression was remembered. First, it was preceded by significant work and
accompanied by an explanation. Second, it is not recognizable as a relation that is typically taught;
you will not find this expression in any physics textbook.
Combinations of remembered, derived, and constructed
188
Frequently, expressions do not fall neatly into one of the above three categories. This is not a
significant difficulty; in such cases, multiple codings are assigned. For example, suppose that when
solving the Mass on a Spring problem a student wrote
Fs = Fg
to indicate that the force of the spring must be equal to the force from gravity. Further suppose
that this is the very first thing written—no other expressions are yet on the whiteboard. Then if
the student writes
kx = mg
then this second expression would be coded as both derived and remembered. To understand this
coding, imagine that the students had written the following two expressions for the forces of
springs and gravity:
Fs = kx ; Fg = mg
Then they could have derived kx = mg by substituting these two relations into Fs = Fg . However,
since at the time that kx = mg was written these two expressions were not written on the board, I
treat the “kx” and “mg” as drawn from memory. The general rule being applied here is this:
Whenever an expression is produced using expressions that are currently written on the
whiteboard it receives a derived coding. And whenever an expression draws on known expressions
that are not currently written anywhere on the whiteboard it receives a remembered coding. In this
case, the expression kx = mg draws on both remembered expressions and expressions written on
the whiteboard.
Marginal cases
Even allowing for combinations of the above categories, some expression-writing events still do
not fall into any of these categories and must be treated separately. I start by noting that the
“remembered” category is intended to be quite narrowly defined. When a student writes
x = xo + vot + 12 at 2
189
this is coded as remembered because the student is producing a literally remembered string of
symbols. You will find this expression written in textbooks using the same letters and written in
precisely this order. In coding expressions, I do allow some deviations from this strict rule. For
example, occasionally students will write the above expression with the terms on the right hand
side in a different order. But, for an expression to be coded as remembered, it should be evident
that the student is working from a literally remembered string.
Constructed expressions, on the other hand, are intended to lie at the other extreme, they are
supposed to constitute true inventions by the student. However, there are some cases that lie
between these two extremes and which I therefore treat separately. There are three cases to be
considered here, and I consider each in turn.
Remembered templates. Students seem to employ a set of templates that define a narrow class of
expressions. Unlike true “remembered” expressions, these templates do not literally determine the
symbols that are to be written—they do not specify that particular symbols to be written in a
specific order. Nonetheless, the expressions produced by these templates seem to fall short of the
other extreme, constructed expressions; although they have the template-like character associated
with symbol patterns, they are more specific to content than forms. An example of a remembered
template is the following expression template for the volume of a rectangular parallelepiped:
volume = length × width × height
When students employ this template, they do not always write literally the same expression; for
example, they might write V=lwh or V=xyz. Although these expressions are thus not literally
remembered strings, it would certainly be inappropriate to claim that a student invented such
expressions. Furthermore, like remembered expressions, expressions associated with remembered
templates are usually written quickly and with little comment. This is further evidence that they
should be treated more like remembered expressions than novel constructions. Table Chapter 6. 4 contains a list of all of the remembered templates that appeared in the algebra data corpus.
Remembered Templates
190
volume = length × width × height
area = length × width
total = rate × time
total = density × volume
average =
sum of items
number of items
change = final − initial
Table Chapter 6. -4. The remembered templates used in the data corpus.
Value assignment statements. This class of expressions are always of the form variable=number. For
example, if a student writes vo =10 this would be coded as a value assignment statement. Such
expressions are not something that one could copy out of a textbook, they are specific to the
circumstance. Nonetheless, they are hardly worth calling “inventions.” Furthermore, these very
simple expressions are actually quite common—common enough that they would make a
significant contribution to any category that they were placed in. For this reason, I believe it is
worthwhile to separate out these expressions and treat them as a special case to be coded
independently.
Principle-based expressions. This last category is the most subtle and I believe the potentially most
controversial of these marginal cases. “Principle-based” expressions constitute another class of
expressions that are not literally remembered but also may not be worthy of being called
“inventions.” To explain this category, I have to return to the discussion of principle-based
schemata that appeared in Chapter 3. In that chapter I argued against the alternative view that
students always derive novel expressions by beginning with physical principles and then deriving
equations from those principles. In this alternative view, students possess schemata that are
associated with physical principles and these schemata provide templates for the construction of
expressions. For example, there is a template of the form
initial momentum = final momentum
that is associated with the principle of Conservation of Momentum.
I argued in Chapter 3 that, even when students claimed to be working from principles of this
sort, symbolic forms could be seen to be playing a role in their constructions. Furthermore, they
seemed to often use principles more for after-the-fact rationalization than as strict guides to
191
construction. I stand by these claims. However, in a very small number of cases, students made a
show of working from principles and there was no evidence to the contrary. In these cases the
resulting expression was coded as principle-based.
In addition, throughout this work I have taken the stance that multiple types of knowledge
must be involved in any construction episode. Therefore, even in construction episodes where
forms and devices are clearly implicated, knowledge tied to principles may be playing some more
or less auxiliary role. To keep track of such occasions, where appropriate I coded expressions as
both constructed and principle-based.
k
x
F=-kx
m
F=ma
Figure Chapter 6. -5. Jack & Jim's diagram for the Mass on a Spring problem.
An example is provided by Jack and Jim’s work on the Mass on a Spring problem. They began
by drawing the diagram shown in Figure Chapter 6. -5 and commenting that there were two
forces, a force upward due to the spring and a force downward from gravity. Then Jack wrote an
expression in which these two forces were equated:
kx = mg
Jack
So, okay, so then these two have to be in equilibrium those are the only forces acting on
it. So then somehow I guess, um, (3.0) That negative sign is somewhat arbitrary [g. F=-kx]
depending on how you assign it. Of course, in this case, acceleration is gravity [g. a in
F=ma] Which would be negative so we really don't have to worry about that [g.~ F=-kx] So I
guess we end up with k x is m g. [w. kx=mg]
This expression is actually assigned three codings: derived, constructed, and principle-based. It is
derived because it makes use of expressions that are already written on the whiteboard, namely
“kx” from F=-kx and the “ma” from F=ma. Furthermore, I argued in Chapter 3 that the
construction of this expression could not be completely explained by the statement that Jack is
applying known principles; the noteworthy feature in this regard was Jack’s “hand-waving”
treatment of the signs involved. For this reason, the expression is coded as constructed.
Finally, I believe that knowledge that is more closely tied to principles plays some role in
constructions of this sort. Although I have argued that it is not justified to claim that the work
192
here is being driven solely by a problem solving schema that strictly guides what is done, my guess
is that there is nonetheless knowledge active that is closely related to formal principles and
explicitly sanctioned procedures. For example, in this case it is likely relevant that Jack and Jim
have learned to draw force diagrams and to write equations involving the forces identified. For this
reason, I also code this expression-writing event as principle-based.
As an argument for including this additional information in the coding scheme, note that one
point of the systematic analysis is to paint a picture of the data corpus for readers. By including
information concerning how many expressions may plausibly have been constructed from
principles, I am allowing readers—especially skeptical readers—to draw some of their own
conclusions concerning the nature of algebra-physics.
Summary of expression-centered coding
In summary, the first part of the base-level analysis is the expression-centered coding. In this
coding, every expression in an episode is identified, then each of these expressions is coded as
remembered, derived, constructed, as one of three “marginal” cases, or as a combination of these
categories. (Refer to Table Chapter 6. -5.) The main function of this component of the base-level
analysis is to identify constructed expressions so that these can later be coded in terms of forms
and devices.
Expression-Centered
Coding Scheme
Remembered
Derived
Constructed
Marginal Cases:
Remembered Template
Value Assignment
Principle-Based
Combinations
Table Chapter 6. -5. Summary of categories for coding expressions in the base-level analysis.
Utterance-centered coding: Identifying interpretations of equations
The second component of the base-level analysis is the utterance-centered coding. The point
here is to identify places in an episode where a student “interprets” an equation. This is an
193
inherently tricky process since part of the purpose of analyzing the data corpus is to determine
what constitutes an interpretation of an equation.
My way out of this bind is to cast a wide net. In order to identify interpretive utterances, I
began by operationally defining an interpretation to be any statement that refers to or incorporates
part or all of an expression. For example, a student might point to an expression as they say
something, or a student might mention a portion of an expression in an utterance. Consider this
interpretation given by Jon & Ella, which I discussed in Chapter 4.


h
w

#
drops


+
=
wx
vd2 v R 
 A⋅s  
 v d 1+ 2

vR




Ella
Okay, V R is bigger. Then (10.0),, This is gonna get smaller, [w. arrow pointing to second
term] but this is gonna get bigger. [w. arrow pointing to first term] I think.
Jon
Yeah, I think - that gets smaller [g. second term] cause V walking is smaller than V
running.
Here, as Ella talks, she indicates portions of an expression by drawing arrows on the whiteboard.
Then, in Jon’s response, he also indicates a part of the expression, this time by simply pointing at
it. These statements would be coded as interpretations because an equation is being referred to by
the students.
This operational definition may appear to be very broad; however, I needed to narrow it only
slightly to have a useful rule for identifying interpretations. In particular, four types of utterances
that would otherwise fall under the broad rubric of this definition, were exempted:
(1) Literal reading of expressions. I excluded utterances in which a student simply reads the symbols
that appear in an expression. Bob’s statement below is an example:
2
v
h = 12
g
Bob
Um, so, H equals one half V squared over G. Right?
(2) Descriptions of problem solving actions. Students frequently recounted the steps that they took
to arrive at a solution. For example, a student might have stated that they had started with a
certain equation and then solved for a given variable in that equation. Utterances of this sort were
not considered to be interpretations. This next passage, in which Jim recounts his work on the
Vertical Toss problem, is typical:
194
Jim
But, what I did first of all … unit conversions which I've erased. And then I remember the
formula which I've also erased, V F squared equals V I squared plus two A D [w. rewrites
equation] So, using this I plugged in the converted values [g. 0=…] and I found the
distance, which for me was three hundred and thirty six feet. Using this I plugged it back
into the formula which is over here [g. x=…] and you do get a quadratic, depending on T.
Which I plugged into the ever reliable quadratic formula up here and it spat the value of
four point six eight seconds.
(3) Comments made during constructions. As students struggled to construct novel expressions, they
made many comments about the expression they were working on. These utterances are not
treated as interpretations. Instead, as described below, they were taken into account when coding
the construction effort itself—the expression-writing event—in terms of forms and devices. In
general, only utterances following the writing of a novel expression were considered to be
candidates for interpretive utterances; interpretive utterances were always interpretations of
existing, fully written expressions.
(4) Association of symbols with a physical entity. Sometimes a student will indicate a symbol or
group of symbols and associate it with a physical entity. Jack’s statement below is an example:
x=
Jack
mg
k
(An:::d) x equals m g over k. [w. x=mg/k] Where x is the distance from the ceiling.
Here, Jack tells us what ‘x’ is, it is the distance of the block from the ceiling. Similarly, a student
might point to ‘mv’ written in an expression and associate it with the momentum of some object
in the problem situation. Such statements were not treated as interpretations.
Since this last exemption may seem problematic I will elaborate on my rationale. In brief, the
reason for excluding these utterances is that, although it may be reasonable to describe them as
“interpretations,” these statements should not be explainable in terms of symbolic forms and
representational devices. Recall that, in my discussion of registrations in Chapter 2, I listed four
types of important registrations: (1) individual symbols, (2) numbers, (3) whole equations, (4)
groups of symbols such as individual terms, and (5) symbol patterns. Forms are associated only
with the last type of registration, symbol patterns. In contrast, utterances that associate a symbol
or portion of an expression with a physical entity involve registrations of type (1) or (4). These
types of registrations do not involve seeing the type of expression-spanning structure that is
associated with symbol patterns. For example, if I point to ‘mv’ in an expression and associate it
with the momentum of some object, then I am not recognizing structure in the expression, only a
component of the expression, whose substructure is not relevant. For this reason, utterances of this
195
sort should not be explainable in terms of forms and devices, and, since the point of this portion
of the analysis is to identify events that should be explainable in these terms, these utterances
should be excluded.
So, the above operational definition plus the four exemptions were used to identify
interpretive utterances in episodes. Only one minor potential complication remains to be
mentioned. Before deciding whether an utterance is an interpretive utterance, the transcript must
essentially be segmented into utterances. Another way to say this is that we need to have a way of
knowing where one interpretive utterance ends and the next begins. The bottom line is that this
turns out to not be much of a difficulty since interpretive utterances tended to be fairly isolated
events. For example, a pair of students might solve an entire problem without interpreting any
expressions and then provide an interpretation of the final result. Or a student might interpret an
expression and then go on to begin deriving some other expression. Furthermore, since the
majority of interpretive utterances are isolated in this manner, it was possible to bootstrap out of
my developing understanding of the nature of interpretations to deal with the few difficult cases.
In other words, once I had developed the categories that define what an interpretation of an
equation is, I could use these categories, in difficult cases, to break a stream into separate
interpretive utterances.
A sample base-level coding
Before going on to describe the next stage of the analysis, I want to present a base-level coding
of an entire problem solving episode. The episode is Jack and Jim’s work on the Mass on a Spring
problem, and it is one of the shortest episodes included in the systematic analysis—it was only 2
1/2 minutes in duration. The complete board image from their work is shown in Figure Chapter
6. -6.
196
k
x
F=-kx
m
F=ma
kx=mg
x=
mg
k
Figure Chapter 6. -6. Reproduction of Jack & Jim's board work for the Mass on a Spring problem.
The entire transcript for this episode is given in Figure Chapter 6. -7 and is annotated with the
appropriate base-level codings. Perusing this transcript and the board image tells us that Jack and
Jim began by writing two equations from memory: F=-kx and F=ma. Then they wrote the
expression kx=mg, which is coded as a combination of derived, constructed, and principle-based for
reasons discussed above. The last expression that they wrote, x=mg/k, was derived by manipulating
kx=mg. Finally, the episode concludes with two interpretive utterances, both given during Jim’s
final turn in the conversation. Note that this is an instance in which a knowledge of standard
interpretation types is used to segment a series of contiguous statements into two separate
interpretive utterances. In this case, Jim gives two CHANGING PARAMETERS interpretations of the
equation x=mg/k. In the first, he imagines that the mass varies while all other parameters are held
fixed, and in the second the spring constant varies.
197
Jack
Jim
Jack
Jim
Jack
Jim
Jack
Jim
Jack
Jim
Jack
Jim
Bruce
So, let's see, so we've got the ceiling. [w. draws ceiling] With our
trusty chunk o' spring. [w. draws spring] And a mass. [w. mass w/
label.]
Spring constant K. [w. k]
Okay, um, and we're assuming that it's at equilibrium so it's at a
distance we'll call this X. [w. marks off height, x] From the ceiling
and that's pretty much constant. Um. (Do you) put in force
vectors, the various forces acting on it?
Mmm. Well, there's the gravitational force acting down. [w. arrow
down; w. F=ma] And then there is
a force due to the spring // holding it up.
// a force due to the spring which I
believe is,, [w. F=-kx] Is equal to that.
Right, cause the::: The spring potential is one half K X squared.
So, okay, so then these two have to be in equilibrium those are
the only forces acting on it. So then somehow I guess, um, (3.0)
That negative sign is somewhat arbitrary [g. F=-kx] depending on
how you assign it of course in this case acceleration is gravity [g. a
in F=ma] Which would be negative so we really don't have to worry
about that [g.~ F=-kx] So I guess we end up with K X is M G. [w.
kx=mg]
Hm-mm.
An:::d how,, How does the equilibrium position of the mass
depend on the spring constant. [Read from sheet.] So I guess
we're solving // for X. [w. underlines x in mg=kx]
// for X equals M G over K.
(An:::d) X equals M G over K. [w. x=mg/k] Where X is the distance
from the ceiling.
Okay, and this makes sort of sense because you figure that, as
you have a more massive block hanging from the spring, [g. block] then
you're position x is gonna increase, [g.~ x in diagram] which is what
this is showing. [g.~ m then k in x=mg/k.] And that if you have a
stiffer spring, then you're position x is gonna decrease. [g. Uses fingers
to indicate the gap between the mass and ceiling.] That why it's in the
denominator. So, the answer makes sense. [Looking at Bruce as
he says this last.]
Hmm-mm, hmm-mm. (6.0) Okay. Is that all it asks on that one?
F=ma.
Remembered expression.
F=-kx.
Remembered expression.
kx=mg.
Constructed+derived+
principle-based expression.
x=mg/k.
Derived expression
“as you have a more massive
block hanging from the
spring…”
Interpretive utterance.
“if you have a stiffer spring …”
Interpretive utterance.
Figure Chapter 6. -7. Complete transcript of Jack and Jim’s work on the Mass on a Spring task. Base-level
codings are given in the right hand column.
Coding symbolic forms
Following the base-level analysis of an episode, the next stage of the analysis is to code every
interpretive utterance and construction event in terms of forms and devices. In this section I will
discuss the coding of forms. The first thing to note about the coding of the algebra data corpus in
terms of forms is that this process requires doing two things simultaneously: (1) The list of forms
must be defined and (2) each individual event must be coded in terms of the forms that are on
198
this list. In what follows, I will separately discuss each of these two components of the coding of
forms.
Defining the set of forms
As I described above, iterative coding and recoding of excerpts was an essential element of
defining the set of forms. I began with a preliminary list derived from a pilot study. Then I coded
7 selected episodes, refining the set of forms along the way. Finally, I coded the work of all pairs
on the 8 tasks selected for systematic analysis, refined the categories, and then coded all of these
episodes again.
I did not develop hard and fast rules for looking at a data corpus and specifying the symbolic
forms involved. Instead, following diSessa (1993), I employed a list of heuristic principles. diSessa
employed a similar list of principles to identify knowledge elements in the sense-of-mechanism,
and some of my principles are similar or identical to his.
1. Coverage of the data corpus. The set of symbolic forms must be able to account for every
constructed expression and interpretive utterance identified according to the procedure
described above. If it cannot, then new forms must be added or the set adjusted in some other
manner.
1. Parsimony. In tension with the requirements of the coverage principle is the need for parsimony
in theoretical work. It would not be scientifically useful, for example, to account for every
event in terms of a unique symbolic form. Instead, we want the minimum set that provides
adequate coverage.
1. Inclusion of a symbol pattern. Every symbolic form, by definition, involves a symbol pattern.
Thus, when positing the existence of a form it must be possible to specify a particular symbol
pattern that is associated with the form. This heuristic can help greatly in narrowing the
possible candidates for forms. For example, we might notice that students use intuitions like
“blocking” or “guiding” during problem solving. However, if we cannot associate these
intuitions with specific symbol patterns, then they cannot be considered legitimate candidates
for forms. Furthermore, by recognizing syntactic structures that exist in algebraic expressions,
we can find additional candidates for symbolic forms.
199
1. Symbol patterns only preserve distinctions consequential for interpretive behavior. Of course, not
all of the candidate symbol patterns that we find by examining syntactic structure will be
associated with any forms. The heuristic here is that symbol patterns should only involve a
sensitivity to distinctions that are consequential for the behavior associated with construction
and interpretation events. For example, we saw that when the students in this study
interpreted expressions, it was not generally consequential whether symbols appeared to the
first, second, or a higher power. Thus, these syntactic distinctions are not preserved in symbol
patterns.
1. Impenetrability. Symbolic forms are the base-level “vocabulary” in terms of which people
specify and understand equations. This vocabulary should be “impenetrable” in the sense that
elements cannot be decomposed in terms of other knowledge.
1. Continuity with prior knowledge. In defining the set of forms, we must be sensitive to the fact
that, as discussed in Chapter 5, this knowledge develops out of existing cognitive resources. In
addition to constituting a very general constraint on the nature of any forms we might
propose, this principle also provides us with a means to locate candidate symbolic forms. For
example, we can look to diSessa’s list of p-prims for candidates and we can look at the
“patterns in arithmetic word problems” that, as discussed in Chapter 3, have been uncovered
by some researchers in mathematics education. In fact, both of these sources were extremely
helpful for uncovering candidate forms.
1. Approximation to expertise. Finally, we must keep in mind that we know something about the
behavior that physics instruction is ultimately trying to produce; students are supposedly
learning to use equations “in an expert manner.” Even if we do not know the precise nature of
expert physics knowledge, we do know, in many cases, what equations and solution methods
experts use. Thus, if we presume that the set of symbolic forms will ultimately contribute to
expert physics problem solving, we can make some guesses concerning the reasonableness of
specific forms. For example, since we are looking at somewhat advanced students, we should
see forms that are useful for constructing the types of expressions that experts write. The many
Competing Terms expressions involving forces are a good example here. And forms should be
associated with intuitions that we expect to play a useful role in expertise.
200
The items on the above list differ considerably in how specific they are to the project at hand.
The first two principles are quite general and could potentially be applied to a wide range of
cognitive modeling projects. In contrast, principles (3), (4), and (5) are the most specific to this
project and they do the most work in narrowing the survey guided by these heuristics to symbolic
forms. Principle (6) is another very general heuristic. In attempting to identify knowledge we
should always be sensitive to the constructivist principle that this knowledge must develop from
existing knowledge. (See diSessa, 1993, Smith et al., 1993, and Roschelle, 1991 for more
discussion of this point.) However, in saying that we should look to p-prims and “patterns” for
continuity with forms, I have specialized this principle to the current project and I have made it
much easier to apply. Finally, though the last principle is not quite universal, it applies to a broad
range of circumstances that might be termed “instructional.” When people are learning to
participate in an existing practice, we can use a recognition of the established norms of that
practice to guide our search for the knowledge necessary to participate in the practice.
Recognizing forms in specific events
Once the list of forms is defined, individual events within episodes can then be coded in terms
of this list. Of course, as I have made clear, these tasks were not actually done sequentially; the list
of forms was defined through an iterative process of coding and recoding individual events.
Nevertheless, in order to simplify this discussion, it is productive to think of these two steps as
happening one after the other. First the list of forms is defined, then we write rules for recognizing
individual forms in a given event. Appendix D has a list of all the forms coded, a description of
each form, and a summary of the clues I used to recognize the form. Here, I will just briefly
discuss the types of evidence that were used to recognize forms in individual events.
1. Symbol patterns. We can begin by looking at the expression that is being written or
interpreted. What possible symbol patterns are there in this expression? This source of
evidence does not greatly restrict the forms that can be coded since a given expression may be
consistent with many symbol patterns and the associated forms. For example, almost any
expression is consistent with PROP+. However, this evidence can help in some cases. For
example, not every expression is consistent with PARTS -OF -A-WHOLE; the expression must have
more than one term.
201
1. Verbal clues. Of course, what students say is the main source of evidence. In some cases, the use
of specific words is considered evidence for certain forms. For example, if a student uses the
phrases “depends on” or “is a function of,” these are clues that the dependence form should be
coded for an event (refer to Appendix D). In addition, it is also very useful to look at what
sort of physical entities are mentioned in a statement and what level of the expression these
entities correspond to. Consider the following two examples, both of which were coded as
involving the BALANCING form:
kx=mg
Jack
So, okay, so then these two have to be in equilibrium those are the only forces acting on
it.
________________
µmg=ma
Alan
What Bob didn't mention is that this is when the two forces are equal and the object is no
longer moving because the force frictional is equal to the force it received.
There are several clues in these statements that point to the BALANCING form. In both cases, the
student is talking about influences—forces, in particular. This is a clue that forms in the
Competing Terms cluster are involved. Furthermore, these influences are associated with sides
of the equation; one force goes with the left side of the equation and one force goes with the
right side. This is evidence that the student is seeing the expression in terms of the symbol
pattern associated with BALANCING . Contrast this with a statement in which a student mentions
entities corresponding to individual symbols in these expressions, such as the mass or the
spring constant.
3. Inferences. This is really a special case of verbal evidence. In Chapter 3 I stated that individual
symbolic forms permit specific inferences. For example, if b/x is seen in terms of the PROPform, and x increases, then we know that b/x decreases. If a student states an inference
associated with a form then this constitutes partial evidence for the form.
3. Graphical concomitants. Some forms and form clusters have “graphical concomitants”—items
that are a commonly drawn on the whiteboard in association with the form. Graphs are one
type of graphical concomitant. For example, as I have mentioned, the
DYING AWAY
form is
frequently seen in combination with a graph of an exponentially decaying function. In
addition, free-body diagrams and diagrams of other sorts are used in a variety of ways that are
characteristic of particular forms. Forms in the Competing Terms cluster are frequently
202
accompanied by the use of a diagram in which the competing influences are labeled with
arrows and the signs of influences—positive or negative—are associated with directions in the
diagram.
3. Symbolic conventions. I have already suggested that there are some quite narrow conventions
associated with symbol patterns. For example, when an expression is written in accordance
with the COEFFICIENT form, it is almost always written with the coefficient on the left.
Similarly, expressions in-line with the IDENTITY symbol pattern almost always have the single
symbol on the left side of the equal sign: x=[…] . When expressions are in agreement with
these more narrow conventions, rather than just agreeing with the symbol pattern, more
broadly construed, this constitutes additional evidence for a form. Furthermore, conventions
can also extend beyond structural regularities. For example, when a symbol is being treated as
a coefficient, it is common to use a ‘C’ or a ‘K’ for that symbol. Again, agreement with such
conventions constitutes evidence that a form is operating.
3. Global as well as local evidence. Finally, the coding of an individual event need not be based
solely on evidence local to the event. For example, sometimes a student will repeat an
interpretive utterance later in an episode, in slightly different terms. Such a later interpretation
may guide us in coding the earlier episode. More generally, if a pair of students has been
speaking in terms of influences throughout an episode, this may skew the coding of forms
toward forms in the Competing Terms cluster. Furthermore, I believe this acceptance of
global evidence may be further extended to cross episode and even student boundaries. For
example, suppose that a student gives a very suggestive interpretation of an equation that,
nonetheless, poses problems when we attempt to code it in terms of forms. If other students
have, in a similar situation, given a clearer—but somehow similar—interpretation of the same
expression, we can use the clearer interpretation to support a particular coding of the
problematic instance.
On the Òsubstitution machineryÓ
In this section, I have thus far explained how I can take every constructed expression and
interpretive utterance and code it in terms of symbolic forms. However, just as in the base-level
analysis, there are some exceptions and marginal cases. Stated bluntly, it turns out that not quite
every relevant event can be coded in terms of forms. In some cases, even when a device is clearly
203
involved, students appear to be drawing on resources other than forms to do the work that we
would normally expect from forms.
Here I will only attempt to briefly characterize these additional cognitive resources. When
students enter college physics instruction at UC Berkeley they have a number of well-developed
skills pertaining to the use of symbolic expression. Among these skills are the ability to manipulate
expressions to derive new expressions, and the ability to substitute numbers into expressions. If the
students in my study were given an expression and some numbers to substitute into that
expression, they could perform this substitution and obtain a result with extreme reliability. The
point here is that some interpretations given by students seem to draw on elements of this
“substitution machinery” rather than on symbolic forms. This is particularly true of many
interpretations involving Special Case devices.
As an example, consider Alan’s interpretation of Hooke’s law, which I presented earlier:
F = -ky
Alan
If you look at this point [g. point where dotted line intersects spring] there'd be no force
due to Hooke's law, and that's where it'd remain at rest. If you're looking above this axis
[g. dotted line then above dotted line] you're looking at a negative force so that's gonna
push it down. So, if you get closer to the ceiling, the spring force is gonna push you down.
If you're below this axis um you're going down but now you're gonna have a force because you're gonna have two negatives [g. -ky term] you're gonna have a force that's
sending you back up towards the ceiling. So it's just a question of signs.
In this statement, Alan draws on aspects of his substitution machinery. He knows that substituting
a value of zero for y will give a result of zero for the force. And he knows that substituting a
negative value will yield a positive value for the force. Interpretations of this sort are not given any
coding in terms of forms (unless some forms are also operating). Instead, they are simply assigned
to their own separate category.
This is not to say that these events are completely ignored in the analysis. These events do
involve representational devices, so they must be included if a full accounting of devices is to be
made. Furthermore, these interpretive utterances appear to perform very similar functions to the
interpretations that are central to this study. In the above interpretation, for example, Alan is
explaining why an expression “makes sense.” Checking and explaining the sensibility of
expressions is one of the main functions of interpretive acts. Thus, these events are counted as
interpretive acts and coded in terms of devices; they are simply not assigned form codings.
204
Coding representational devices
When coding representational devices, their is a fundamental difficulty equivalent to the one
associated with coding forms. In order to code the representational devices in the algebra data
corpus, I needed to simultaneously define the set of devices and code individual construction and
interpretation events. I now discuss, in turn, each of these aspects of the coding of devices.
Defining the set of devices
As with forms, the key to finding the set of devices was iterative analysis of the data corpus. I
began with a preliminary set of devices derived from my pilot study and then refined this set
through an iterative analysis based on a set of heuristic principles:
1. Coverage of the data corpus. This is the same as the heuristic principle employed to identify
forms.
1. Parsimony. Same as for forms.
1. Close relation to the structure of interpretive utterances. Because they are more closely related to
visible aspects of behavior, devices are somewhat easier to recognize than forms. When
attempting to account for student behavior, forms function at a relatively deep explanatory
level. As analysts, we must therefore fill in several steps to get from forms to any visible aspects
of behavior. In contrast, devices relate much more directly to behavior; in particular, they are
closely related to the structure of interpretive utterances. As I said in Chapter 4, to find the set
of devices it is reasonable to think of taking all of the interpretive utterances in the data corpus
and putting them in categories. In fact, an important technique I employed for identifying
candidate devices was to look for turns-of-phrase that re-occur in interpretive utterances.
Examples of such phrases are: “At any time…”, “This is when…” and “As this goes up… .”
1. Restriction to reliably codable distinctions and allowance for ambiguity. When inventing categories
for devices there is a temptation to want to fill in “logical gaps” in these categories or to want
to break categories into multiple parts based on a priori possible distinctions. For example, it is
possible that the CHANGING PARAMETERS category should be broken into two distinct subcategories, one in which the change happens continuously, and one in which the change
happens as discrete hops between values. In practice, however, it is not possible to reliably place
individual interpretive utterances into one of these hypothetical categories or the other;
205
student utterances seem to usually be ambiguous in this regard. Consider, for example, this
interpretation by Jim which we have examined many times:
x=
Jim
mg
k
Okay, and this makes sort of sense because you figure that, as you have a more massive
block hanging from the spring, [g. block] then you're position x is gonna increase, [g.~ x in
diagram] which is what this is showing. [g.~ m then k].
The question here is: Is Jim talking about a continuous increase in the mass or is he talking
about a discrete jump in the mass from one value to another? I believe that this utterances is
simply ambiguous in this regard. In defining the categories for devices, I only retained
distinctions that could be reliably made when coding events.
5. Continuity with prior knowledge. This principle is equivalent to the principle of the same name
for forms. However, for the case of devices, there is less to work with in the existing literature.
Although researchers have studied the discourse of mathematics classrooms, this literature has
not adopted a focus close enough to the concerns of this work to be directly useful for
identifying individual representational devices. An exception is the work of Nemirovsky (in
press) which I will discuss in Chapter 7.
5. Approximation to expertise. Again, the idea here is essentially the same as for forms; we can
presume that most representational devices will be useful for doing the types of tasks that
physicists must perform.
Recognizing devices in specific events
Because of the somewhat more behavioral character of devices, recognizing them in particular
events is somewhat easier than recognizing forms. A list of all representational devices coded and
clues for recognizing them are given in Appendix E. Here, I again list only the types of evidence
used.
1. Verbal clues. The most important clues for coding an event in terms of devices are found in
what the students say. In fact, the phrases used can be very strong clues for a specific device.
Examples of such phrases were mentioned just above—“At any time…”, “This is when…,”
“As this goes up… ”—and more can be found in Appendix E. In addition, we can attend to
the “plot line” of an interpretive utterance. For example, does the interpretive utterance focus
206
on a particular quantity that appears in the expression? If so, is that quantity staying fixed or
varying?
1. Graphical concomitants. As with forms, other external representations drawn on the whiteboard
can provide clues concerning devices. For example, Static devices are generally seen in tandem
with a diagram of the physical situation. Because Static devices typically involve associating an
equation with a “moment” in a motion, it is relatively straightforward to draw a diagram to go
with the device. Although diagrams for Narrative devices are somewhat trickier, they are still
possible; a series of pictures may be drawn to depict the changes in a narrative. Figure Chapter
6. -8 shows a portion of the board image for Alan and Bob’s work on the Air Resistance task.
Note that they have drawn separate diagrams for the heavier and lighter mass. Such diagrams
can function as a graphical concomitant for the CHANGING PARAMETERS device.
m
∑F =
2m
mg − kv
Figure Chapter 6. -8. Some of Alan and Bob's work from the Air Resistance Task. A separate diagram is
drawn for the heavier and lighter mass.
3. Global as well as local evidence. As with forms, it is necessary to use global as well as local
evidence to code individual events in terms of devices. We can look at other events involving
the same students, or at other pairs in similar circumstances.
Coding of the sample episode
Before presenting the results of the systematic analysis, I want to briefly show how Jack and
Jim’s work on the Mass on a Spring task was coded in terms of forms and devices. This is the
episode for which I presented the base-level analysis in the previous section. Recall that Jack and
Jim began by drawing a diagram and then writing the equations F=ma and F=-kx. Since these two
expressions were coded as remembered in the base-level analysis, it is not incumbent upon me to
account for them in terms of forms and devices. Next, Jack wrote the expression kx=mg, which I
coded as constructed+derived+principle-based:
207
Jack
Right, cause the::: The spring potential is one half K X squared. So,
okay, so then these two have to be in equilibrium, those are the only
forces acting on it. So then somehow I guess, um, (3.0) That negative
sign is somewhat arbitrary [g. F=-kx] depending on how you assign it of kx=mg.
course in this case acceleration is gravity [g. a in F=ma] Which would be Form: BALANCING
negative so we really don't have to worry about that [g.~ F=-kx] So I
Device: STEADY STATE
guess we end up with K X is M G. [w. kx=mg]
Because this expression is coded as, in part, constructed, I need to analyze the forms and
devices involved. First, the balancing form is implicated here. One important piece of evidence for
this coding is Jack’s use of the term “equilibrium” (refer to Appendix D). In addition, note that
Jack is talking about two influences that correspond to sides of the expression. He says “these two
have to be in equilibrium, those are the only forces acting on it.” This is strong evidence for
BALANCING .
The device involved here is the STEADY STATE device. When Jack says “these two have
to be in equilibrium,” he is describing a static, unchanging situation that continues through time.
Finally, after a single manipulation to produce the equation x=mg/k, Jim made two
interpretive utterances:
Jim
Okay, and this makes sort of sense because you figure that, as
you have a more massive block hanging from the spring, [g. block] then
you're position X is gonna increase, [g.~ x in diagram] which is what
this is showing. [g.~ m then k in x=mg/k.] And that if you have a
stiffer spring, then you're position X is gonna decrease. [g. Uses fingers
to indicate the gap between the mass and ceiling.] That why it's in the
denominator. So, the answer makes sense. [Looking at Bruce as
he says this last.]
“as you have a more massive
block …”
Form: PROP+
Device: CHANGING
PARAMETERS
“if you have a stiffer spring …”
Form: PROPDevice: CHANGING
PARAMETERS
Each of these interpretive utterances must be coded in terms of forms and devices. In both cases,
Jim picked out a single quantity and imagined that it varied; in the first, the mass varied, in the
second, the spring constant varied. Thus, both of these interpretations involve the changing
parameters device. There are also a number of clues here concerning the forms involved. The fact
that Jim attends to the equation at the level of individual symbols is an important clue. In
addition, a clear inference is evident in each of these utterances: Jim infers that the position will
increase if m increases, and the position will decrease if k increases. Thus, the first interpretive
utterance involves the PROP+ form; the position is seen as proportional to the mass. And the
second interpretive utterance involves the PROP- form, with the position seen as inversely
proportional to the spring constant.
208
Analysis Summary
For future reference, the steps for analyzing a given episode (one pair’s work on a single task)
are summarized in Figure Chapter 6. -9. In the base-level phase of the analysis, symbolic
expressions and interpretive utterances are identified. Next, the identified expressions are further
coded as remembered, derived, constructed, or as belonging to one of the marginal cases. Finally,
in the last phase of the analysis of an episode, each constructed expression and interpretive
utterance is coded in terms of forms and devices. In this last phase, some events may not receive
any codings in terms of forms and instead are coded as involving the substitution machinery.
Episode Transcript
Symbolic Expressions
Interptive Utterances
Base-Level
Analysis
Remembered
Derived
Constructed
Forms
Marginal
Cases
Forms
Devices
Devices
Coding of
Forms and
Devices
Figure Chapter 6. -9. Overview of the analysis of an episode.
Summary of Results
Base-level results
I will begin the summary of results by presenting the outcomes of the base-level analysis.
Recall that the primary purpose of the base-level analysis was to identify events in the transcript
that must be coded in terms of forms and devices. Nonetheless, the base-level analysis does, in its
own right, reveal some important features of the data corpus.
The expression-centered component of the base-level analysis found a total of 547 expressions
in the episodes selected for systematic analysis. Of these expressions, 470 were written “from
scratch” and 77 were modifications of existing expressions (refer to Table Chapter 6. -6).
Expressions “from scratch”
Modifications
Total expressions
470
77
547
Table Chapter 6. -6. Expressions found in the base-level analysis.
209
Table Chapter 6. -7 shows the result of coding these 547 expressions as remembered, derived,
constructed, or as belonging to one of the three marginal cases. The sum of all codings is somewhat
higher than 547 since individual events may be given multiple codings. The far right hand column
in Table Chapter 6. -7, labeled “With Other Codings,” is intended to provide some additional
information in this regard. The numbers in this column tell how many instances of a given coding
were made in combination with some other coding. For example, 33 of the expressions coded as
remembered were also assigned another coding.
Remembered
Derived
Constructed
Marginal Cases
Total
Codings
With Other
Codings
157
287
75
110
33
46
42
51
Table Chapter 6. -7. Base-level coding of expressions.
The highlighted row in Table Chapter 6. -7 is worthy of special attention since the primary
goal of the expression-centered coding is to identify constructed expressions. As shown in the table,
75 constructed expressions were identified, 42 of which were assigned some additional coding.
Table Chapter 6. -8 provides a breakdown by type of the 110 marginal cases. It is worth
noting that 41 of these expressions were the extremely simple type of expression that I called
“value assignments.” The reader should also note that 32 of the 36 expressions coded as principlebased also received some other coding. Thus, a very small number of expressions were coded as
purely principle-based.
Total
Codings
With Other
Codings
33
41
36
19
0
32
Remembered Template
Value Assignment
Principle-based
Table Chapter 6. -8. Breakdown of marginal cases by type.
The results of the second component of the base-level analysis—the utterance-centered
analysis—are fairly simple to report since this analysis only identifies interpretive utterances; it
does not further code them in any way. As stated in Table Chapter 6. -9, a total of 144
interpretive utterances were identified in the episodes included in the systematic analysis. One
210
other piece of information I will provide here is that these 144 interpretive utterances were
associated with 93 expressions. The number of associated expressions is less than 144 since some
expressions were interpreted more than once by students.
144
93
Interpretive Utterances
Interpreted Expressions
Table Chapter 6. -9. Results of the utterance-centered analysis.
These 144 interpretive utterances, together with the 75 constructed expressions identified,
constitute the events that must be coded during the next stage of the analysis. In total there are
thus 219 individual events that must be coded in terms of forms and devices (refer to Table
Chapter 6. -10). It is these 219 events that I iteratively coded to arrive at the final sets of forms
and devices.
144
75
219
Interpretive Utterances
Constructed Expressions
Total Events
Table Chapter 6. -10. Events that must be coded in terms of forms and devices.
Before commenting further on these results, I want to present one additional table of values
from the base-level analysis. Table Chapter 6. -11 gives some of the results of this analysis by
subject pair. Note first that there was fairly substantial variation in the total number of expressions
written by the pairs. Mike and Karl wrote only 72 expressions, the smallest number in the study.
In contrast, Alan and Bob wrote 136, the most of any pair.
However, this variation across students in the total expressions written does not seem to drive
differences in the number of construction and interpretation events. Interestingly, there do not
appear to be strong correlations of the total number of expressions with interpretive utterances
(r=-.07), constructed expressions (r=.11), or the total interpretation and construction events
(r=.06).
Expressions
Interpretive Utterances
Constructed Expressions
Interpretations+Constructions
M&K
A&B
J&E
M&R
J&J
Totals
72
31
13
44
136
33
13
46
122
26
15
41
106
33
12
45
111
21
22
43
547
144
79
219
Table Chapter 6. -11. Results of the base-level analysis by subject pair.
211
I want to speculate briefly on this lack of correlation to total expressions. For illustration, I will
compare Alan and Bob with Mike and Karl. It turns out that Alan and Bob wrote 18 value
assignment statements while Mike and Karl only wrote 3. These expressions are generally not a
very important aspect of student work, when students write a value assignment statement they are
usually just transcribing aspects of the problem statement onto the whiteboard. Similarly, there
was a big difference between these two pairs in the number of expressions coded as derived. For
Mike and Karl there were 30; for Alan and Bob there were 75. In part, this difference reflects a
particular stylistic variation between the two pairs: Alan and Bob tended to write out many steps
when manipulating expressions. Thus, my hypothesis is that the large differences in total number
of expressions are traceable to certain stylistic differences that are generally not closely related to
construction and interpretation events.
This is not to say that I do not expect there to be differences among students in how
frequently they interpret and construct expressions. In fact, Table Chapter 6. -11 already suggests
possible differences among my subjects. But, interestingly, even where there are differences in the
number of interpretations or constructions, there is some balancing out of these differences in the
last row of the table. In fact, the data does indicate a negative correlation between interpretive
utterances and constructed expressions (r=-.94). A possible reason for this is that statements made
during the construction of an expression might, to some extent, obviate the need for later
interpretation of the statement. (Recall that statements made prior to or during the writing of an
expression were used only in coding the expression; they were not treated as interpretive
utterances.) But, given the small number of subjects this observation is somewhat shaky and I will
not draw on it in further discussions. In any case, this particular constancy is actually quite
desirable for this study since it means that pairs contributed roughly equally to the forms and
devices coded in the next phase of the analysis.
Now I will sum up the results of the base-level analysis. I believe that the most important
conclusion to be drawn from the base-level analysis is that the construction and interpretation of
expressions are not rare events. Recall that, from the very first chapter of this dissertation, I have
argued that this work is important precisely because these phenomena have not been accounted for
by existing research. However, if it had turned out that these phenomena were simply rare—if, for
example, students almost never constructed expressions—then we might have decided that it was
reasonable to ignore these phenomena. At the least we would have had to accord them less
importance than they have been given by this research.
212
Since this point is extremely important, it is worth taking a moment to argue that the results
presented in this section can be interpreted as suggesting that construction and interpretation are
“not rare.” First, I believe that the sheer number of events is somewhat convincing. As we saw,
there were a total of 219 construction and interpretation events spread over 11.5 hours of student
problem solving work.
And these results can profitably be viewed from some other angles. First, let’s compare the
frequency of interpretations and constructions with the total number of expressions written. Of
the 547 expressions that students wrote, 75—14%—received a coding of constructed. This is not
at all a vanishingly small percentage. Furthermore, there were a total of 144 interpretive utterances
applied to 93 separate expressions. This means that 17% of all the expressions written were
interpreted in some manner. Again, this is a non-trivial fraction of all expressions.
Finally, I want to consider the prevalence of these events from one additional angle, by
episode. It turns out that of the 40 total episodes included in the systematic analysis, 34 included
construction events, 37 included interpretation events, and 39 included either a construction or
interpretation. Furthermore, in the one odd episode that involved no constructions or
interpretations, the students solved the task without writing any expressions. Thus, in every
episode that could potentially have involved interpretations or constructions, these events
appeared. Taken altogether, I believe it is reasonable to describe these results as indicating that the
construction and interpretation of expressions are not “rare” events.
Results from the coding of symbolic forms
The results of coding all of the 219 events identified by the base-level analysis in terms of
symbolic forms are given in Table Chapter 6. -12. Results are listed both by cluster and by
individual form. (Refer to Appendix D for a brief description of each form.) The values in the
column labeled “count” represent the raw number of events coded with the corresponding form or
cluster. The numbers in the “percent” column were obtained by dividing this raw count by the
total number of codings; that is, by dividing the associated count by the total counts of all forms.
These percentages thus provide a measure of the relative frequency with which the various forms
were coded.
213
Count
Percent
BALANCING
75
30
12
28
27%
11%
4%
10%
CANCELING
5
Proportionality
PROP+
PROP-
Competing Terms
COMPETING TERMS
OPPOSITION
RATIO
CANCELING(B)
Count
Percent
SOLE DEPENDENCE
63
40
16
7
23%
15%
6%
3%
2%
Coefficient
14
5%
77
52
28%
19%
COEFFICIENT
SCALING
5
9
2%
3%
14
9
2
5%
3%
1%
Multiplication
INTENSIVE •EXTENSIVE
EXTENSIVE •EXTENSIV
5
5
0
2%
2%
0%
2
1
1%
0%
Dependence
DEPENDENCE
NO DEPENDENCE
E
Terms are Amounts
PARTS -OF -A-WHOLE
BASE ± CHANGE
38
10
23
14%
4%
8%
WHOLE - PART
1
4
0%
1%
SAME AMOUNT
Other
DYING AWAY
NO EVIDENCE
Table Chapter 6. -12. Results for the systematic coding of symbolic forms.
As I discussed above, some events could not be coded in terms of forms. Instead, these events
seemed to employ aspects of what I referred to as the “substitution machinery.” Of the 219 total
events, 12 received no form codings and were coded as involving only the substitution machinery.
In Figure Chapter 6. -10, the breakdown of codings by cluster is displayed in a pie chart.
Looking at this figure, it is evident that no one cluster of forms is especially dominant. Rather,
three clusters stand out in this chart. The Competing Terms and Proportionality clusters had the
most codings, with the Dependence cluster running a close third. Together, these three clusters
account for 77% of all codings. Of the remaining three clusters, Terms are Amounts, was most
common, occupying a middle range in the data. Finally, the Coefficient and Multiplication
clusters were coded rarely, together accounting for only 7% of the total number of codings.
214
Multiplication
2% Other
Coefficient
1%
5%
Competing Terms
27%
Terms are Amounts
14%
Dependence
23%
Proportionality
28%
Figure Chapter 6. -10. Percentage of codings by cluster.
Referring to Table Chapter 6. -12 we can see that the two most common individual forms
were PROP+ and DEPENDENCE. These two forms alone account for about a third of all codings. It is
interesting to note that both PROP+ and DEPENDENCE involve a sensitivity to very basic aspects of
the structure of expressions. Recall that, when a student sees
DEPENDENCE,
they are only
recognizing that a particular symbol appears somewhere in an expression. PROP+ also does not
involve a sensitivity to very complex aspects of structure. Seeing
PROP+
means simply attending to
the fact that a particular symbol appears in the numerator of an expression. (In some cases, as in
expressions like F=-kx, there may not even be a denominator!) Thus, a very large percentage of the
work of seeing meaning in symbolic expressions can be traced to very basic aspects of the structure
of expressions.
The next two most common forms, COMPETING TERMS and BALANCING , together made up
about 20% of all codings. Thus, over 50% of all of the codings are accounted for by only 4 of the
20 forms in my list. This suggests the interesting conclusion that a relatively small number of
forms are accounting for much of the meaning associated with expressions.
The prevalence of the COMPETING TERMS and BALANCING forms can be better understood if we
examine the type of contexts in which these forms appeared. Recall that both of these forms are
from the Competing Terms cluster and that forms in this cluster involve treating an equation at
the level of terms, with terms corresponding to competing influences. The point I want to make is
that, in many instances, the influences involved were forces, in the technical sense of this term. In
fact, of the 28 cases in which BALANCING was coded, 24 involved balanced forces. The COMPETING
TERMS
form was deployed in a slightly more varied manner, with 11 of the 30 instances involving
forces. I believe that these observations help to put the prevalence of these two forms in
215
perspective, and can provide the reader with a sense for the types of circumstances in which these
forms were coded. Since forces can be used in a productive manner to solve many of the tasks that
I gave to students, it is not surprising that these two forms appeared frequently in the data corpus.
Some other aspects of Table Chapter 6. -12 require further explanation. The reader may have
noticed that several of the forms listed received a very small number of codings. For example,
DYING AWAY
was coded twice and WHOLE - PART was only coded once. If these were the only events
in which I had seen these forms, I would have been hesitant to conclude that they exist. However,
in all of these cases I saw additional instances of the forms in episodes not included in the
systematic analysis.
The very small number of codings in the Multiplication cluster, including zero codings for the
EXTENSIVE •EXTENSIVE
form, also requires some comment. In part this result is indicative of some
real sparseness in this territory—these forms simply do not appear very often. But such a result
must be at least somewhat troubling. Does this mean that students are nearly incapable of
constructing expressions involving the product of factors? Are they simply not very sensitive to this
type of structure in expressions? Or is it that the particular tasks I asked students to perform
simply did not require them to use forms relating to the product of factors?
Remembered Templates
count
volume = length × width × height
area = length × width
total = rate × time
total = density × volume
10
3
5
8
2
sum of items
average = number of items
change = final − initial
3
Table Chapter 6. -13. Remembered Templates with counts. The top four templates relate to multiplication
forms.
The answer seems to be, in part, that my analysis in terms of symbolic forms actually misses
some of the action relating to the product of factors. Recall that when expressions were coded in
the base-level analysis a number of expressions were placed in three “Marginal Cases.” One of
these Marginal Cases was what I called “Remembered Templates.” The full list of Remembered
Templates is reproduced in Table Chapter 6. -13 along with the number of times that each of
these templates was employed. The important fact to notice is that four of the six templates
shown in Table Chapter 6. -13 could be constructed or interpreted in terms of the two forms in
216
the Multiplication cluster. In particular, the rate × time and density × volume templates could be
usefully interpreted in terms of the
INTENSIVE •EXTENSIVE
form. And the templates for computing
volume and area could be usefully interpreted in terms of the
EXTENSIVE •EXTENSIVE
form. This, in
part, is where work relating to the product of factors is hiding in my analysis.
So why not code the associated expression-writing events in terms of the INTENSIVE •EXTENSIVE
and the EXTENSIVE •EXTENSIVE forms? The reason has to do with what these events actually look
like in the data corpus. In particular, students generally wrote the expressions associated with these
templates quickly and with little comment. For example, they did not try to explain the fact that
the area of a rectangle was the product of its length and width; instead, this seemed more like a
remembered relation that they were just using. Thus, I believe that to code these events as
constructed would be misleading. However, it is still important to note that these templates, while
more limited in their potential range of use than forms, are still somewhat flexible. This is
especially true of the rate × time and density × volume templates. The “rate” in rate × time, for
example, can be any type of rate, such as a velocity or an acceleration.
Now I will comment more generally on the meaning of the numerical results presented in this
sub-section. The results in Table Chapter 6. -12 state the frequency with which particular symbolic
forms appeared in the data corpus. Of course, there are many limits to how far we can trust these
numbers, and I will comment on these limits shortly. But, accepting these numbers for the
moment, what can we conclude about the nature of student knowledge and the character of
“expert intuition?” As I mentioned briefly in the introduction to this chapter, the move I want to
make is to presume that the frequency with which forms are employed by students reflects the
priority of the associated knowledge elements possessed by students. Here I am using the term
“priority” in the sense I ascribed to diSessa: It is a measure of the likelihood of a knowledge
element being activated. In diSessa’s picture of the sense-of-mechanism, intuitive physics
knowledge gets its character from the priorities of individual p-prims. Similarly, I am now
asserting that expert intuition, in part, takes its character from the relative priorities of individual
forms.
There are two reasons that I believe that coding frequencies can be presumed to reflect the
priority of elements. The first reason is that, by definition, higher priority elements are more likely
to be used. So, if elements appear more often, this is a sign that they have a higher priority. The
second reason—which is closely related to the first—is that the problem solving activity that is
contained in the data corpus is precisely the kind of activity that is supposed to determine the
217
priority of individual forms. If a particular form is called on repeatedly during problem solving
then the priority of that form should be increased. Thus, this is another reason that we can
presume that the forms that appear frequently are the high priority forms.
There are some problems with these arguments for interpreting frequencies as priorities,
however. In particular, recall that the cuing priority of a knowledge element reflects its likelihood
of being activated in a particular circumstance. According to the strict usage of this term, cuing
priority is context dependence; there is no overall cuing priority that spans contexts. Thus, the
“priorities” that I measured here must be understood to be a coarser measure of the likelihood of
elements being used than would be given by a true measurement of context-specific priorities. In
essence, I have a measure of the priority of elements averaged over the circumstances in the study.
Nonetheless, I believe that this “coarser” measure is still useful for getting a feel for the rough
character of expert intuition.
The frequencies in Table Chapter 6. -12 also provide some indirect evidence concerning other
aspects of expert intuition. In the last chapter I hypothesized that forms are tightly connected to
some specific types of previously existing cognitive resources, especially p-prims. When a
particular form is activated, it is more likely that related p-prims will be activated. And when pprims are activated, they are reinforced and their priorities are increased. Thus the frequency data
not only provides information about the priorities of forms, it also provides some indirect
evidence concerning the priorities of p-prims in expert intuition, at least for the medium-level
experts that participated in this study.
Given the viewpoint that the coding frequencies tell us about the character of intuition, I want
to reflect just a little more about the above results. In the discussion of the development of expert
intuition in the previous chapter, I explained that diSessa hypothesizes that agency plays an
expanded role in expert intuition; and, in particular, much of this agency is localized in the
concept of force. For example, when talking about a book resting on a table, a novice might say
that the table simply “supports” the book. In contrast, a physicists sees the table as an agent—the
table applies a force upward on the book that opposes the gravitational force.
The high frequency of forms in the Competing Terms cluster may be indicative of this
expansion in the role of agency. In these forms, terms are associated with influences that oppose
and support, and frequently they are associated with forces. Thus, the high use of forms in the
Competing Terms cluster could be viewed as part of the developmental trend toward seeing all
objects as potential agents that apply forces.
218
The relatively high frequency of one form in the Competing Terms
cluster—BALANCING —might be considered surprising. From the point of view of standardized
physics, situations in which influences are in balance are special cases—the influences just happen
to be equal. Compare the COMPETING TERMS form. COMPETING TERMS is, in a sense, more general; it
is used in situations where influences are seen as competing, but not necessarily in balance or
canceling. Referring to Table Chapter 6. -12 we see that BALANCING was coded roughly the same
number of times as COMPETING TERMS. Does this mean that expert intuition is skewed toward
seeing situations in terms of BALANCING ? It is possible that this result is simply an artifact of the
particular selection of tasks that I gave students. However, I believe that we are seeing something
real and important here. Recall, for example, Alan’s inappropriate BALANCING interpretation of the
expression mgµ=ma. As I have mentioned, I believe that this is indicative of a tendency to overuse
BALANCING .
The high frequency of Proportionality forms may also, in part, reflect the developmental
trend toward an increasing role for agency. As I argued in Chapter 5, the PROP+ and PROP- forms
may be related to the notions of effort and resistance. I also believe that the prevalence of these
forms is due to the fact that, in a fundamental sense, learning algebra-physics means learning to
quantify the physical world. As I have mentioned, proportionality relations may play an important
role in our abilities for reasoning about quantities; therefore, it is not surprising that these
associated symbolic forms appear frequently in algebra-physics.
M&K
A&B
J&E
M&R
J&J
Totals
Competing Terms
Proportionalit
Dependence
Terms are Amounts
Coefficient
Multiplication
16
14
9
11
2
0
19
14
13
15
3
0
15
9
15
2
2
2
16
18
18
5
2
1
9
22
8
5
5
2
75
77
63
38
14
5
Totals
52
64
45
60
51
272
Table Chapter 6. -14. Form cluster codings broken down by subject pair.
All of the data concerning symbolic forms that I have presented so far has been aggregated to
produce the total instances found in all episodes. Before concluding this section, I want to
decompose this data along two dimensions. Table Chapter 6. -14 breaks down the coding of
symbolic form clusters by subject pair. Looking at this table, we see that the distribution is not the
same for all subjects. This suggests several reasons for caution in interpreting the results presented
219
in this section. First, if the differences in Table Chapter 6. -14 are indicative of real differences
that exist among individuals, then it could mean that my results depend sensitively on the
particular subjects that participated in the study. In addition, these observations limit the extent to
which we should even attempt to look for anything like a universal distribution.
Of course, I have always assumed that there would be individual differences in students’ formdevice knowledge. The behavior of the form-device system is determined by many cuing and
reliability priorities, and there is no reason to believe that these priorities would be precisely the
same across subjects. I have even explicitly argued that these priorities would change during
learning; thus, since my subjects may be more or less far along in their learning of physics, there is
particular reason to expect that priorities should differ. Nonetheless, I was hoping that there
would be some generalities that could be extracted by an analysis of this sort.
There are a few across-pair uniformities in Table Chapter 6. -14. For example, for all subject
pairs, the Multiplication Cluster received less than 5% of all codings for that pair, and the
Proportionality Cluster always received at least 20% of all codings. Also, the top three clusters in
Table Chapter 6. -14— Competing Terms, Proportionality, and Dependence—always accounted
for at least 70% of the codings for a pair. I do not mean to play down the importance of
differences here, only note that there are some similarities across subject pairs.
When, the results are broken down by task, as in Table Chapter 6. -15, we see that the results
for each task were quite different. To begin, note the extreme variation in total codings across
tasks. This is not too surprising since students spent quite a different amount of time on each task.
But there is also great variation in the distribution of forms within tasks. For example, the Air
Resistance task had the most total codings, 73, but had no codings of forms in the Terms are
Amounts cluster. Similarly, the Running in the Rain task had no codings of Competing Terms
forms. Again, such results are not surprising since different tasks require different methods.
However, even though this non-uniformity across problems may not be surprising, it may still
have negative implications with regard to the universality of the frequencies derived from my data
corpus. In particular, as with the variation across pairs, it implies that when averaging over tasks to
get a measure of an “overall” priority, the result is sensitively dependent on the selection of tasks
that I gave to students. I will comment on this issue in the last section of this chapter, and I will
discuss it in slightly more detail in Chapter 10.
220
Shove
Competing Terms
Proportionalit
Dependence
Terms are Amounts
Coefficient
Multiplication
Totals
4
15
15
4
Toss
Skater
Buo anc
8
3
8
6
3
1
6
1
11
Spring
Air
Wind
9
8
5
3
25
22
22
21
4
3
14
Rain
Totals
19
17
5
1
5
75
77
63
38
14
5
47
272
4
38
12
14
21
25
73
42
Table Chapter 6. -15. Form cluster codings broken down by problem.
Results from the coding of representational devices
The result of coding all 219 events identified by the base-level analysis in terms of
representational devices is shown in Table Chapter 6. -16. In addition, the results collected by
class are displayed in a pie chart in Figure Chapter 6. -11. Referring to this pie chart, it is clear that
Narrative devices dominate. In fact, the three devices in this class are three of the four most
common individual devices, and the
CHANGING PARAMETERS
device far outstrips all other devices,
accounting for a third of all codings.
Narrative
CHANGING PARAMETERS
CHANGING SITUATIONS
PHYSICAL CHANGE
Special Case
SPECIFIC VALUE
LIMITING CASE
RESTRICTED VALUE
RELATIVE VALUES
Count
Percent
128
79
21
28
54%
33%
9%
12%
26
9
6
11%
4%
3%
3
8
1%
3%
Count
Percent
83
17
5
15
35%
7%
2%
6%
17
6
23
7%
3%
10%
2
2
1%
Static
SPECIFIC MOMENT
GENERIC MOMENT
STEADY STATE
STATIC FORCES
CONSERVATION
ACCOUNTING
Other
NO EVIDENCE
Table Chapter 6. -16. Results for the systematic coding of representational devices.
The drop off is precipitous as we move to the next two classes of devices. Though devices in
the Static class received a respectable 35% of the total codings, this is somewhat less than the 54%
of codings accounted for by Narrative devices. Finally, Special Case devices constitute a mere
11% of the total.
221
Special Case
11%
Other
1%
Narrative
53%
Static
35%
Figure Chapter 6. -11. Representational devices by class.
These results point to Narrative devices as a clear winner. Not only are these the most
prevalent devices by class, every one of the individual devices in this class seems to be extremely
common. And one device in particular,
CHANGING PARAMETERS,
far outstrips all others. This leads
to the question: Is this what we would expect?
From at least one perspective this result is somewhat surprising. To illustrate, I want to take a
moment to reflect on what students are taught in introductory physics classes. Representational
devices, per se, are not an explicit component of introductory physics courses. You certainly will
not find sections on the SPECIFIC MOMENT and CHANGING PARAMETERS devices in a physics textbook.
However, some representational devices are loosely associated with more or less explicit elements
of instruction. For example, instructors have been known to exhort students to consider “limiting”
or “special cases.” Although these instructors likely do not think of these heuristics as “strategies
for interpreting equations,” they may occasionally think of them as a way to check whether an
equation “makes sense.” Thus, some Special Case devices may actually be closely related to
relatively explicit—though peripheral—aspects of instruction.
Static devices are also related, though somewhat less directly, to some explicit and more
central aspects of instruction. I argued in Chapter 4 that we can think of free-body diagrams as a
means of constructing equations to be interpreted with some devices in this class. Students learn
free-body diagrams as an explicit technique for solving problems—you isolate an object, then
draw all of the forces acting on the object. And many introductory textbooks explicitly mention
free-body diagrams. Thus, although instructors are not teaching interpretive strategies, per se,
these particular interpretive strategies do appear to be frequently embedded in well-worn and
clearly recognized elements of problem solving practice.
In contrast, Narrative devices are neither expressly taught nor tied to any explicitly recognized
aspects of practice. Rather, Narrative devices seem to constitute an altogether tacit aspect of
physics practice that cuts across the gamut of problem solving activities. It is for this reason that
222
the results presented in this section can be considered surprising. It is slightly ironic that the one
class of interpretive heuristics that is most explicitly taught—the Special Case class—turns out to
be comparatively rare. Instead, the majority of interpretations of expressions involve Narrative
devices, which are neither recognized nor expressly taught.
As with forms, it is helpful to break down the results of the device analysis by subject pair and
task. Table Chapter 6. -17 gives the devices in each class as they were coded for each subject pair.
Again there are notable differences. Jack and Jim received more than three times as many
Narrative codings as Static codings, while, in contrast, Alan and Bob actually received slightly less
Narrative codings than Static codings. These differences lead to the same caveats as those
mentioned for the form-related results.
But, again, there are some weak across-pair uniformities. The Special Case Class was always
the least common, accounting for less than 20% of the codings for each pair, while the Narrative
Class always accounted for at least 40%. Furthermore, the Static Class fell in between for all pairs
except Alan and Bob.
M&K
Narrative
Static
Special Case
Totals
25
14
9
48
A&B
J&E
21
26
6
53
21
16
5
42
M&R
30
18
1
49
J&J
Totals
31
9
5
46
128
83
26
237
Table Chapter 6. -17. Device class codings broken down by subject pair.
Table Chapter 6. -18 breaks down the devices in each class by task. As we saw with the forms
data, there is great variation across task. The total codings for individual tasks extend over a great
range, from 9 to 60, and the distribution within problems appears somewhat different. There are
some weak suggestions of uniformity here, however. For 5 of the 8 tasks, the ranking of the classes
by frequency is the same as for the overall distribution. Furthermore, the remaining 3 tasks happen
to be the 3 tasks with the lowest total number of codings. Still, the variation here is reason for
caution in interpreting the aggregated results.
223
Shove
Narrative
Static
Special Case
Totals
22
4
3
29
Toss
2
11
1
14
Skater
6
3
9
Buoyancy
2
9
7
18
Spring
11
10
5
26
Air
Wind
40
19
1
60
25
14
4
43
Rain
Totals
26
10
2
38
128
83
26
237
Table Chapter 6. -18. Device class codings broken down by task.
Conclusions to be Drawn From the Systematic Analysis
What was the purpose of this analysis? I want to take a moment to reflect on the nature of the
conclusions that we can draw from the presentation in this chapter. To be sure, the results
presented in this chapter are not the output from any typical experimental design. For example,
there were not a “control” group and an “experimental” group that were compared. Although in a
later chapter I will compare algebra-physics and programming-physics data, the central results of
this chapter do not involve comparisons of any sort. So what do all of these numbers tell us?
The analysis described in this chapter, rather than producing a comparison, can be understood
as producing a description of algebra-physics knowledge. This description has several elements.
First, the lists of forms and devices themselves—even without any frequencies—constitute part of
the description of algebra-physics. Although I first presented these lists in Chapter 3 and 4, the
current chapter adds an account of how I made contact between these lists and my data corpus.
Providing this account is perhaps the most important purpose of this chapter. Along with the
examples presented in chapters 3 and 4, this chapter does the important work of clarifying the
relationship between my theoretical categories and the data I had available.
In filling in frequencies and providing other numerical results, the analysis presented in this
chapter also paints a more detailed picture than is provided by a simple listing of forms and
devices. There are two different senses in which these results provide a “picture.” First, the
numerical results obtained here perform the function of summarizing the data corpus for the
reader. By looking at these numbers, a reader can get a broad feel for the features of the corpus,
such as the frequency of interpretation and construction events, without having to watch every
videotape or read every transcript. This summarizing function is one of the usual purposes of
characterizing a data corpus with statistics.
In addition to providing an overview of the data corpus, these results can also be understood as
providing a picture of algebra-physics knowledge. Earlier in this chapter I argued that the frequency
224
results provide information concerning physical intuition, because they can be interpreted as an
indication of the priorities of knowledge elements. But any such interpretation of the results is
necessarily more speculative.
However there are reasons for caution in interpreting all of these results, particular for any
conclusions relating to the nature of individual algebra-physics knowledge. First, as I have
commented, the lists of forms and devices were not derived from the data using exact techniques.
Instead, I employed a heuristic process rooted in my intuitions and theoretical predispositions, as
well as the data. Furthermore, the content of these lists was not confirmed to any extent by a
second coder. Because of the heuristic nature of the process, training another coder would have
been an extensive activity in itself. Thus, I hope that the reader will understand this analysis as a
first step toward future, more exact and replicable methods of using analyses of this sort to
identify knowledge elements.
Second, even accepting the lists of forms and devices as I presented them, the coding of
individual episodes was not an entirely straightforward process; thus, the numbers that I associated
with the various categories could be disputed. The coding of episodes in terms of given categories
(forms and devices) would actually be more amenable to confirmation by additional coders than
the production of the categories themselves, and I hope to do this in the near future. Nonetheless,
there is still a significant effort required to train coders in this procedure and, due to constraints of
time, such a confirmation was not included here.
Third, even if the numbers presented in Table Chapter 6. -12 and Table Chapter 6. -16 are
taken as, in some sense, accurately describing my particular data corpus, there are still reasons for
caution. I commented, in my discussion of the experimental sessions, that the sessions themselves
may not be entirely typical of any type of naturally occuring activity, such as a group problem
solving session. Most notably, in occasionally prompting students to explain their work, I may
have tended to increase the frequency of interpretation events.
In addition, we saw that the frequency results varied across subjects pairs. This is a problem
because it suggests that my results could be sensitive to the particular subjects that participated,
and because it casts doubt on the possibility of generalizing across subjects. These worries were
slightly moderated by the observation of some weak uniformities across subjects.
The results also showed a strong variation across tasks. This result is far from surprising, but it
nonetheless poses difficulties for drawing conclusions. Of course, I did attempt to engage students
in a spread of tasks, so there is some hope that the variation averages to a useful measure of the
225
overall priority of elements. But I did not attempt to rigorously establish that the selection of tasks
is, in any sense, representative. In fact, it is not even clear what dimensions should be used to
determine the “representativeness” of tasks. Thus, we must accept the likelihood that changing the
selection of tasks would have altered the results obtained, and that my selection may not be
especially representative.
Nonetheless, I believe we can still draw some general—though limited—conclusions. To end
this chapter, I want to present these conclusions, and attempt to calibrate the confidence we can
have in each of them. First, I hope to have shown that construction and interpretation were not
rare events, and that it was not unusual for students at this level to have the ability to construct
and interpret equations. All pairs constructed and interpreted expressions approximately the same
number of times per task, and there was at least a single construction or interpretation event in
every problem solving episode except one. Given the variation observed across problems, the
precise frequency with which these events occur is somewhat more questionable. Furthermore, my
prompts may have tended to skew my results toward seeing more interpretations. But the fact
that all of the subject pairs demonstrated the ability to construct and interpret on numerous
occasions—whether prompted or not—means that we can be relatively certain that most students
have the requisite capabilities. And the universal existence of these capabilities strongly suggests
that this knowledge and behavior are important.
Assessing confidence in the lists of forms and devices is somewhat more difficult. The
production of these lists was the component of my analysis most dependent on heuristic
procedures and my own intuition. Thus, our confidence in these lists must depend significantly on
the plausibility that follows from the arguments in the preceding chapters.
But, if we accept the categories, what can we conclude from the particular frequency counts I
presented in Table Chapter 6. -12 and Table Chapter 6. -16? Even with the worries about crosssubject and cross-task variation, I believe it is appropriate to take a few tentative conclusions into
succeeding chapters. First, beginning with the device-related results, the aggregated counts
indicated quite large differences between classes: Static devices were coded almost three times as
much as Special Case devices, and Narrative almost five times as much as Special Case. Of course,
this distribution was not uniformly represented across subjects and tasks, but there were some
weak reflections of this distribution across these dimensions. Thus, for the chapters that follow, I
will tentatively presume that Narrative devices are generally most common in algebra-physics,
Special Case rarely used, and Static devices in the middle.
226
We are in a similar position with regard to the frequency of forms. Again, the analysis found
some substantial differences across clusters. For example, there were 77 Proportionality codings
and only 5 codings in the Multiplication cluster, and the three most common clusters accounted
for at least 70% of the codings for every subject pair. In addition, there were substantial
differences in how many times the various individual forms were coded. The
DEPENDENCE
PROP+
and
forms were coded 52 and 50 times respectively, while some forms received only a
few codings. More generally, as I have commented, the four most common forms—PROP+,
DEPENDENCE, COMPETING TERMS,
and BALANCING —accounted for 50% of all codings.
Furthermore, I believe that these results are also somewhat plausible. The proportionality and
dependence forms can appropriately be used to interpret expressions in a wide range of
circumstances. And, because of its observed relation to forces, the high frequency of Competing
Terms forms is not very surprising.
Thus, even with the observed variations across tasks and subjects, I believe that it is appropriate
to take these results as indicating something about the character of algebra-physics. In what
follows, I will tentatively presume that the Competing Terms, Proportionality, and Dependence
clusters are somewhat more common than the other three clusters, and that the common forms
within these clusters— PROP+, DEPENDENCE, COMPETING TERMS, and BALANCING —are particularly
important in algebra-physics.
So, the result of this analysis is an account of how I arrived at the lists of forms and devices, as
well as a picture of the data corpus that can also be taken as a rough image of algebra-physics
knowledge. In later chapters, I will want to compare this image to a similarly constructed image of
programming-physics knowledge. At that time, we will see that the results of the corresponding
analyses are sufficient to uncover a number of revealing contrasts.
Finally, I want to conclude with one additional important conclusion that can be drawn from
the analysis presented in this chapter. I believe that the results of this analysis can be taken as
evidence for my theoretical viewpoint, understood more broadly. Included in this view is the notion
that students use expressions meaningfully; they can build symbolic expressions from an
understanding of what they want the expression to convey, and they can interpret expressions.
Also included in my viewpoint is the notion that the cognitive resources that are central to
meaningful symbol use have symbolic forms and representational devices as elements.
In a sense, the systematic analysis has put the totality of this viewpoint to test by attempting to
use it to describe the data corpus. In the process, we have seen that there are a few marginal cases
227
which could not be accounted for in terms of forms and devices. For example, some of the action
involving the product of factors had to be accounted for by “Remembered Templates,” rather
than forms and devices. And some interpretive utterances that could be coded in terms of devices
had to be described as employing the “substitution machinery,” rather than forms. However, I do
not interpret the existence of these marginal cases as signs of the breakdown of the theory.
Instead, I take them as symptomatic of the complexity that the theoretical view presumes to exist.
The form-device system is just one cognitive resource and we should not expect this resource to
do everything. What is impressive, I believe, is that this analysis suggests that forms and devices
do quite a lot of work in students’ problem solving. We saw many instances of students
constructing and interpreting expression and the vast majority of these could, in part, be explained
by an appeal symbolic forms and representational devices.
228
Chapter 7. Reflections on the Theory
For the physicist and physics student, a sheet of physics expressions is structured and meaningful.
This is a statement that I have made a number of times. But what, precisely, does this statement
mean? In what sense is a sheet of expressions more or less “meaningful?” Let’s back up for a
moment and work up to a more precise and comprehensible version of this statement.
We begin with a thought experiment. Imagine that we take a sheet of paper and fill it with
physics expressions, perhaps with the solution to a standard physics problem. Then we take this
sheet of paper and drop it on an island where the inhabitants have never seen any sort of writing
before. What will they make of this sheet of paper? If they have never seen paper before, they may
very well be fascinated by the sheet of paper as an object in itself. But what of the markings on the
paper? Whether the marks on the paper are interesting to the islanders or not, the details in the
arrangement of marks will not be consequential for them. Since they are not privy to the
conventions that we attach to these markings, the markings, in their details, are no more
significant for the islanders than the arrangement of grooves in the bark of a tree. Just as with these
grooves, fine differences in the markings are not important.
Now imagine that we send in a team to retrieve the sheet of paper from the island of nonliterate peoples. Once we have the sheet of paper, we then take it and show it to some people who
are literate in English, but who know no physics and little mathematics. What will these people see
when they look at the sheet of physics expressions? They will certainly recognize many individual
letters and perhaps some mathematical symbols like the equal sign and a plus sign. It is also likely
that they will recognize that these letters are arranged into short horizontal strings.
Next, we show the sheet of expressions to some students who are in the very early stages of
physics instruction. These students will likely recognize that the sheet is filled with equations.
Maybe they have learned the equation v=vo +at. If this equation is written on the sheet of paper
they might recognize it. They might then know, for example, that the v in the expression stands
for “velocity,” and they might know something about the conditions under which this equation is
used.
In the final step of this thought experiment we show the sheet of expressions to some people
who have been significantly initiated into the practice of physics. These initiates will recognize
many of the expressions that appear on the sheet of paper. If F=ma is written they will recognize it
and may associate it with Newton’s second law. If F=-kx is written they may know that this is
229
Hooke’s law for the force applied by a spring. Most importantly, I have argued that not only will
they recognize individual symbols and whole expressions, they will recognize structures in these
expressions that I have called “symbol patterns,” and these symbol patterns are associated with a
particular kind of knowledge element that I have called “symbolic forms.”
The purpose of this thought experiment was to display a continuum of cases in which the page
of expressions is increasingly “meaningful” for the people involved. As we progressed through the
cases, the expressions were more and more meaningful in a particular sense: The individuals
involved were sensitive to more varieties of structure in the markings, with the perceived structures
associated with knowledge possessed by the individuals. This sensitivity to details of structure
combined with rich associations with knowledge is what it means for the page of expressions to be
“meaningful.”
For the physics initiate, a page of expressions is richly meaningful in this sense. To this point, I
have been engaged in showing how I believe this works. I have worked on describing what the
physics initiate sees when they look at an expression—what features they are sensitive to and how
what they recognize depends on concerns of the moment. Really, I have focused on just one
aspect of this story, the aspect associated with the seeing of symbol patterns in expressions. The
result was a map of this one aspect of the physicist’s meaningful world of symbolic experience.
Characterizing the knowledge associated with the seeing of this meaningful structure has been
the job of the preceding five chapters. The resulting model of this cognitive resource—the formdevice system—is the core theoretical result of this dissertation. In this final chapter of Part 1, I
want to reflect a little on the nature of this model and the larger theoretical view in which it is
embedded. This job will be divided into two parts. First, I will summarize the model and engage
in some quite speculative extrapolation, drawing out some of the model’s implications, and
attempting to generalize the model to other representational practices. A central goal of this first
portion of the chapter will be to extract implications from the model concerning the relationship
between external representations and physics knowledge. Then, in the second portion of this
chapter I will contrast the viewpoint I have adopted with four other research perspectives.
How Physics Knowledge is ÒIntrinsically RepresentationalÓ
In Chapter 2, I began my foray into the physicist’s world of symbol use with some examples
designed to show that students do sometimes construct and interpret equations. These examples
were important because they provided a first glimpse into the phenomena that are the focus of this
230
project. In fact, a significant product of this research is the identification and description of these
phenomena.
Then, at the end of Chapter 2 and in subsequent chapters, I proceeded to describe how this
construction and interpretation works. I described my model of the cognitive resources involved
and how they develop:
Registrations. First, following Roschelle (1991), I introduced the term “registration,” to refer to
how people “carve up their sensory experience into parts.” I listed several kinds of registrations that
are relevant to symbol use in physics: individual symbols, numbers, whole equations, groups of
symbols, and symbol patterns.
A vocabulary of symbolic forms. Next I introduced the first of the two major theoretical constructs
in this work, what I have called “symbolic forms.” A symbolic form consists of two components:
(1) a symbol pattern and (2) a conceptual schema. Taken together the set of forms constitutes a
conceptual vocabulary in terms of which expressions are composed and understood.
Representational devices and the interpretation of symbolic expressions. In addition to symbolic
forms, I argued that physics initiates possess a complex repertoire of interpretive strategies that I
called “representational devices.” The project to identify representational devices involved, in part,
the collecting and categorizing of interpretive utterances.
The refinement of physical intuition. After describing the form-device system, I speculated on how
this knowledge develops during physics learning. I argued that the form-device resource develops
out of existing resources, including intuitive physics knowledge and knowledge relating to the
solving of algebra problems, and that these previously existing resources continue to exist and to
be refined.
The above summary, together with the specific lists of forms and devices that I provided in
Chapters 3 and 4, constitutes the core of the Theory of Forms and Devices. Now, in the
remainder of this section, I want to reflect on this model and extract some of its implications. As I
mentioned, this reflection will be speculative; I will be going somewhat beyond what can be
claimed with any confidence given the data and arguments I have thus far presented.
One key issue to which I want to pay particular attention is what I have called the “relation
between knowledge and external representations.” In Chapter 1 I discussed some researchers for
231
whom this relation is variously deep and fundamental. Exemplars that I referred to often were
Jack Goody, who posited a central role for external representations in the historically developing
knowledge of society, and Richard Feynman, who argued that to truly appreciate the beauty of
nature it is necessary to have the experience of manipulating physics expressions.
The possibility of these deep relations led me to urge caution in a few places. I pointed out
that, given these possibilities, algebra-physics and programming-physics practice may involve
fundamentally different understandings of physics. I was also led to question some of the
assumptions underlying the teaching of “Conceptual Physics” courses, which teach physics largely
without the use of equations. In particular, in this thesis, I challenge the assumption, usually
implicit, that an expert’s understanding can be separated into two pieces, one conceptual, the
other symbolic-mathematical, so that the conceptual part can be taught separately. To anticipate
some of what follows, this assumption is called into question by the observation that the formdevice system does not fall neatly into either of these two categories; it is both conceptual and
symbolic-mathematical. To be clear, I have not meant to dispute whether Conceptual Physics is a
useful instructional method; rather, my intent has been to push for a better understanding of the
results of such instruction.
In the present section, I will explore what the model described in the preceding chapters can
tell us about the relation between external representations and physics knowledge. My approach
here will be to argue for the stance that physics knowledge is “intrinsically representational,” and
to attempt to describe the precise sense in which this is true. In making this claim, I mean to be
asserting that the physics knowledge of individuals is in some way inseparable from the external
representations that we use.
Thus, in what follows, I will use the model to argue that physics knowledge is intrinsically
representational, and to state some specifics of the deep relation between physics knowledge and
external representations. However, I will not be able to discuss this deep relation in full generality.
This is the case because I have only studied a small portion of the knowledge involved in physics
understanding. As I will argue below, a more complete discussion of the relation between
knowledge and external representations would require a more thorough analysis of the knowledge
involved.
Before jumping in, I want to remind the reader of a few details of my discussion of the
relation between knowledge and external representations in Chapter 1. In that chapter, I began by
distinguishing between “direct” and “indirect” effects of external representations on individual
232
knowledge. Direct effects arise out of an individual’s actual interaction with an external
representation. In contrast, indirect effects arise because a person working with an external
representation may produce various products that can then be “fed back” into the knowledge of
individuals. This distinction was built into the simple framework that is reproduced in Figure
Chapter 7. -1.
Representational
Practice
Direct
Effects
Literate
Individual
External
Representation
"Products"
Indirect
Effects
Cultural
Mediation
Literate or
Non-Literate
Individual
Figure Chapter 7. -1. Framework separating direct and indirect effects of external representation on the
knowledge of individuals.
The model I described in the preceding chapters will primarily allow us to draw conclusions
concerning direct effects. In what follows, my discussion is organized around two views of direct
effects described in Chapter 1: (1) the point that knowledge must be adapted by and for symbol use
and (2) the notion that literates derive intuitions from a symbolic world of experience. I will now
reflect on my model from each of these two points of view.
Knowledge is adapted by and for symbol use
In Chapter 1 I mentioned that it is sometimes illuminating to treat a person and external
representation as a system whose cognitive abilities are more powerful than a person acting alone.
For example, the system consisting of a literate individual together with a pencil and a sheet of
paper may be said to have a “better memory” than the individual alone. An important feature of
this stance is it allows us to state that external representations significantly enhance the cognitive
abilities of individuals, without positing any substantial changes in the character of individual
thought processes in the absence of the external representation.
233
However, I went on to argue that we should still expect certain types of “residual effects” on
the knowledge of individuals. Because external representations essentially alter the tasks that
individuals must perform, their knowledge and thought processes must be adapted in order for
individuals to participate successfully in a representational practice. In short, I argued that
although there may not be changes in the overall character of thought, the knowledge of
individuals must still be adapted by and for symbol use.
In my discussion in the previous chapters, I described some of the specifics of how the
knowledge of initiates is refined by symbol use in physics. Recall the hypothetical scenario I
presented in Chapter 5. I described a situation in which two competing p-prims were activated
and I explained how symbol use could resolve the conflict between these two p-prims. Then I
argued that the result of such an experience would be that the priority of the “winning” p-prim
was incrementally increased. In such a manner, incremental changes could be made to the senseof-mechanism and the sense-of-mechanism gradually refined. This is a “residual effect” on
individual knowledge due to direct interaction with an external representation. Similarly, I argued
that incremental changes could follow from p-prims being cued during symbol use.
In addition to being adapted by symbol use, it is also useful to think of physics knowledge as
being adapted for symbol use. In order to construct a novel expression, a student’s notions must be
cast in symbolic forms. This requires that existing resources, like the sense-of-mechanism, must be
refined so that this is possible.
An important point—which I also made in Chapter 5—is that the precise nature of the
refinements made in this way depends intimately on idiosyncrasies of the symbolic practice. To
see how important this effect is, consider an alternate viewpoint. We might hypothesize that, with
the development of expertise, physical intuition develops toward an ideal form that is determined
by the conceptual structure of Newtonian physics and, ultimately, by some sort of fundamental
properties of physical reality. In this view, there are still limits to what physical intuition can do;
we still will not expect that physical intuition, without the aid of symbol use, can make detailed
predictions in all circumstances. But, within the limits of its nature, physical intuition gets as close
as it can to this ideal form.
The view I am presenting here differs sharply with this alternative hypothesis. I have argued
that an expert’s physical intuition is specifically adapted for the purpose of complementing and
functioning in the activity of symbol use. The point here is that it makes sense to say that physics
knowledge is intrinsically representational, not because symbol use can directly cause changes in
234
individual knowledge, but because the precise way that knowledge develops depends on the
details of the symbolic practice. Not only is there a “residual effect” of symbol use, this residual
effect bears the unique fingerprint of the symbolic practice that left it behind.
For example, consider the prevalence of the BALANCING form in my data corpus and recall that
this suggests that the BALANCING p-prim has a relatively high priority in expert intuition. As I
argued in Chapter 6, there is nothing particularly central about situations involving balanced
influences in the conceptual structure of Newtonian physics. Situations that involve such balanced
influences are special cases; the influences just happen to be equal. Thus, the prevalence of
BALANCING
intuitions in expert physics may have more to do with the nature of algebra-physics
practice than with the conceptual structure of Newtonian physics. In retrospect, this is not
surprising. The business of algebra-physics is the writing of equations, constraints that express
equalities. And if we want to write an equation, we have to find things in the world that we can set
equal. Thus, the intuition of the expert physicist must, to a certain extent, be geared particularly
toward searching out BALANCING in the world.
So what does all of this have to do with my model? Couldn’t we have made most of these
arguments without all of the work necessary to develop my model of the form-device system? It is
true that it is very nearly a trivial statement to say that the knowledge of physics initiates must, to
some extent, be adapted specially for symbol use. For example, physics students must learn
certain specific equations and they must learn specialized techniques for solving problems and
deriving new equations.
However, what my model adds—and what is really interesting—is that symbol use in physics
draws directly and significantly on knowledge at the level of physical intuition. We might have
found that people write and understand symbolic expressions only at the level of whole equations
associated with formal principles. Instead, I argued that people see meaningful structure at the
level described by symbolic forms which, in turn, are strongly tied to elements of the sense-ofmechanism. Thus, it is because of the fact that physicists see this type of meaningful structure that
the details of symbol use can have direct effects that percolate down to the very roots of physics
understanding.
A symbolic world of experience
Now we move on to the second way of thinking about how direct interaction with external
representations can lead to residual changes in individual knowledge. The idea is that, through
235
participation in a representational practice, a new variety of intuition can be born, rooted in this
experience. This is essentially the notion that Feynman expresses in the quotes presented in
Chapter 1. Feynman apparently believes that the experience of manipulating mathematical
expressions to solve problems and derive results contributes in some significant way to a person’s
appreciation of “the beauties of nature.” One of the central goals of this dissertation has been to
put some theoretical meat on this idea.
The theoretical meat can be found in my discussion of symbolic forms and representational
devices. Let us take a minute to reflect on the nature of the form-device knowledge system.
Importantly, this resource has much the character of the intuitive knowledge systems that we have
discussed, such as the sense-of-mechanism. It has relatively small and simple elements that are
multiply connected to the world and other knowledge. And the form-device system is only
weakly structured; there are no central organizing elements that tie the whole system together and
give it a straightforward organization.
For these reasons, I believe it makes sense to consider the form-device system to be a
component of expert physical intuition. Thus, the suggestion here is to think of expert physical
intuition as consisting of refined versions of pre-existing resources, such as the sense-ofmechanism, together with the form-device knowledge system. If we accept this proposal, then
there is a component of expert physical intuition that has its roots solidly in symbolic experience.
This is what I mean when I say that there is a “symbolic basis of physical intuition.”
Throughout the preceding chapters, we have seen examples of forms and devices functioning
like other intuitive resources. For example, just like p-prims, symbolic forms constitute a primitive
level in terms of which physical phenomena are explained and understood. I want to present an
example to emphasize this point. In this episode, which I also discussed in Chapter 3, Mark and
Roger were attempting to explain why all objects fall at the same rate in the absence of air
resistance. Eventually, Mark wrote two expressions and explained as follows:
a=
Mark
F
; F = mg
m
You know, however you have a force [g. F in a=F/m] that varies at the same proportion
directly - directly with the mass, don't you? The heavier - the heavier an object is, the
more force it will exert, right? Upon, like say I try to lift it. … So, if it's double the mass it's
gonna have double the force downward.…Two M G. Right? But the mass is also two M , so
those two M's cancel and you get - you're left with A equals G in all cases, aren't you.
Above, Mark has “mentally” combined these two expressions to obtain a=mg/m, which is an
expression for the acceleration of the dropped ball. He then goes on to explain that any changes in
236
the mass will have no effect on the acceleration, since any change in the numerator is canceled by a
corresponding change in the denominator.
So, how does Mark understand the fact that all objects fall at the same rate? What is his
explanation? I would argue that Mark’s explanation is based on the CANCELING(B) form. At the
heart of his explanation is the idea that there are two effects—one associated with the force and
one with the mass—that precisely cancel. Furthermore, there is no reason to presume that Mark’s
understanding of “canceling” is entirely separate from his experience with its mathematical
manifestation. This is precisely what it means to say that Mark’s explanation is based on the
CANCELING(B)
form. His explanation and understanding of the fact that all objects fall at the same
rate is rooted, at least in part, in his symbolic experiences of canceling. This is a very strong sense
in which physics knowledge is “intrinsically representational.”
The same point can be made from another angle. Suppose we were going to teach a
Conceptual Physics course and we believed that the knowledge of physicists could be separated
into “symbolic-mathematical” and “conceptual-physical” components. I have argued that
symbolic forms are part of the knowledge that expert physicists possess. This leads to the
question: Are forms “symbolic-mathematical” or “conceptual-physical?”
The answer is that symbolic forms do not seem to fit neatly into either of these categories;
they have their origins in both physical and symbolic experience. For example, consider the
dependence form. This is rooted in our physical experiences of causation and physical
dependence, but it also derives from our experience of symbolic dependence using expressions
and plugging various numbers into an expression to get a result. Thus, the physical and symbolic
are welded together in this single knowledge element.
In conclusion, this section has argued for a strong sense in which physics knowledge is
intrinsically representational. I maintained that, because it has the character of intuitive
knowledge, the form-device knowledge system should be considered a component of expert
physical intuition. And the knowledge elements of this system—basic elements of an experts
physical intuition—are themselves rooted in symbol use.
Summary: How would Conceptual Physics knowledge differ?
What do the comments in this section imply about the consequences of a move to Conceptual
Physics instruction? First, my arguments imply that we can expect a differently tuned sense-ofmechanism. In algebra-physics, the sense-of-mechanism and other pre-existing elements of
237
physical intuition are specially tuned for the role that they play in symbol use. In contrast, a
practice of Conceptual Physics may tend to tune the resources more for producing a certain type
of qualitative explanations of phenomena.
Overall, we may guess that intuition will be tuned more for providing correct predictions and
explanations in the absence of equations. This, in fact, is part of the rationale behind Conceptual
Physics instruction. In Chapter 1, I discussed some of the striking failures of traditional
instruction. This included the observation that students do not seem to learn some of the basics of
Newton’s laws, such as the fact that objects continue to move in a straight line in the absence of
applied forces. Given the fact that Conceptual Physics focuses on issues of this sort, it is very
plausible that students will do better in this regard. And, in fact, as I discussed in Chapter 1, there
is already much evidence of success.
But do students come away from Conceptual Physics instruction with an “appreciation of the
beauties of nature?” Of course, this depends what we decide counts as such an appreciation. What
we can say is that students will at least come away with a somewhat different appreciation of
physics, a divergent understanding. I have argued that, for initiates in algebra-physics, there is a
component of physical intuition that is rooted in symbolic experience, and this component will
necessarily be less developed in Conceptual Physics instruction. Again, I do not mean to imply
that Conceptual Physics is any less a “real” or respectable version of physics, and I certainly do not
mean that Conceptual Physics cannot provide useful preparation for algebra-physics. I am simply
trying to say that, at an important and deep level, the understanding associated with conceptual
physics is different.
The Theory Generalized
In the above sections I reviewed some of the specifics of my model of symbol use in physics,
and I drew out some of the implication of this model for the relation between external
representations and physics understanding. Strictly, all of the claims made above are specific to
the particular representational practice that I studied, algebra-physics. Now I want to consider the
extent to which my theory and the associated claims can be generalized outside of this narrow
realm. Of course, a major task of this thesis—which I will undertake in Part 2—is the extension of
the theoretical framework to programming-physics. But, in this chapter, I want to first think more
broadly about the generality of the framework and its implications. I will begin with a brief
discussion of where and how the form-device model of symbol use can be generalized and
238
extended. Then, as far as is possible, I will attempt to make a few broad claims about the relation
between knowledge and external representations.
How general is the form-device model?
Clearly, many aspects of my discussion of registrations, symbolic forms, and representational
devices are particular to the use of algebraic notation in physics. For example, since each individual
form is associated with a specific algebraic symbol pattern, the individual forms that I listed are
inescapably tied to algebraic notation. And since some devices have to do with projecting an
equation into a physical motion, these devices, at least narrowly understood, must be specific to
physics.
Nonetheless it is possible that my main theoretical constructs, understood more generally,
may have wider applicability. Take registrations, for example. Although specific registrations will
certainly not generalize across representations and practices, it seems like a reasonable project to
characterize, for any representational practice, the features and structures to which individuals are
sensitive. Thus my hypothesis is that this notion—which, recall, I borrowed from Roschelle
(1991)—has wide applicability.
However, the two novel theoretical constructs introduced in this research, symbolic forms and
representational devices, will likely be less generalizable to other representational practices. Note
that the hypothesis associated with the positing of these constructs is really rather strong.
According to the model that I presented, the seeing of meaning in expressions involves a very
particular sort of two-layer system: Representational devices set a stance, and expressions are seen
in terms of a number of specific symbolic forms, each of which is associated with a well-defined
symbol pattern. It seems unlikely to me that this very particular model will extend to all other
representational practices. Consider natural language, for example. It is quite implausible that we
would could capture much of the range of people’s abilities to construct meaningful statements in
terms of a few simple conceptual schemata and symbolic structures. The extensive efforts exerted
by linguists to construct grammars for natural language and to understand the nature of
utterances are sufficient testimony to this fact.
Registrations, forms, and devices in Ògraphing-physicsÓ
In some cases of “near transfer,” however, I believe that the form-device framework may turn
out to be quite helpful. Of course, one of the major presumptions of this work is that this model
can be extended to describe programming-physics, and I will undertake that project in the next
239
part of this dissertation. In this section, I want to do a small amount of work to try out the model
on another closely related practice: graphing-physics. My purpose here is just to give a feel for how
the framework presented in the preceding chapters might be applied to other practices.
For the case of graphing-physics, the framework seems to work quite well at the level of
theoretical constructs (registrations, forms, and devices). In what follows, I briefly discuss how
each of the major theoretical constructs might be manifested in graphing-physics.
Registrations. As in algebra-physics, there are many possible types of structure to which people
might be sensitive. Some examples are:
•
The overall shape of a curve.
•
The average slope of a segment.
•
The location (relative to the two axes) of a particular location on a curve.
Figure Chapter 7. -2. A possible graphing-physics graph.
Symbolic forms. Associated with these registrations, there are likely a variety of symbolic forms in
graphing-physics. For example, students could learn to associate “DYING AWAY ” with the overall
shape of the graph shown in Figure Chapter 7. -2. Similarly, there could be a “ STEEP ” form that
includes the ability to register a high-valued slope in a segment and associated with a conceptual
notion that is something like rapid increase or rapid change (refer to the boxed region in Figure
Chapter 7. -2).
Representational devices. In the case of representational devices, I believe that the framework can be
extended quite directly from my analysis of algebra-physics. Not only do I believe that we can
find devices in graphing-physics, I believe that some of these devices will be very similar to devices
encountered in algebra. For example, in the manner of Narrative devices, a student could trace
240
along a graph, telling the story of a motion as they go. Similarly, a particular location on a curve
might be associated with a specific moment in a motion. Such an interpretation would be similar
to the Static interpretations I described in Chapter 4.
It turns out that at least one researcher, Ricardo Nemirovsky, has argued for a theoretical
framework very close to the one presented in this section (Nemirovsky, in press). Nemirovsky
describes what he calls “mathematical narratives,” which he defines to be narratives that are
“articulated with mathematical symbols.” Strikingly, he gives examples that are very reminiscent
of the examples I have given of Narrative interpretations. Consider, for example, the following
graph and narrative, which Nemirovsky provides for illustration:
Figure Chapter 7. -3. Example graph from Nemirovsky (in press).
1.
2.
3.
First it rained more and more and it started to become steady (pointing to piece (a) of graph).
Then it rained steadily (marking piece (b) of graph).
Then it rained more and more (pointing to piece (c) of Graph).
Furthermore, in presenting this example, Nemirovsky argues that a major feature of such
narratives is the splitting of the continuous graph into phases or episodes. Such an observation is
closely related to many of the points I made in discussing symbolic forms. What Nemirovsky is
saying is that people recognize pieces of graphs, like the segments in the graph in Figure Chapter
7. -3. In fact, Nemirovsky goes on to hypothesize that people learn a “grammar of graphical
shapes” for graphs, a view that I take to be equivalent to saying that they learn symbolic forms
(Nemirovsky, 1992). As an example, he explains that the two “iconic shapes” shown in Figure
Chapter 7. -4 can express the distinction between “growing steadily” and “growing but slowing
down.” Thus, because of the great similarity of the view expressed by Nemirovsky, we can look to
these papers for arguments and evidence that the framework presented here can be applied to
graphing.
241
Figure Chapter 7. -4. Possible elements of a “grammar of graphical shapes” (Nemirovsky, 1992).
Representations and understanding, in general
Finally, I want to conclude this section with one last piece of reflection and speculation before
turning to some contrasts. In Chapter 1, I framed this project, to a certain extent, as a special case.
I raised the issue generally of the relation between individual knowledge and external
representations, and I argued that equation use in physics was a good place to begin to study these
issues. Now that I have completed the first stage of this project, I would like to pause and ask
what more general morals we can draw. What can the work here tell us about the relation between
knowledge and external representations, especially for the more pervasive symbol systems in our
activity, such as natural language and writing?
First, notice that I did not even do the whole job for equation use in physics. I only looked at
one new cognitive resource that develops—the form-device system—and I talked a little about
how a few other existing resources, especially the sense-of-mechanism, would change through
symbol use. To tell a complete story, I would really need to do more analysis of this sort. And, in
principle, there could be as many effects as there are relevant knowledge systems.
Unfortunately, the same is true for any external representations. Really, the project here must
be repeated for each external representational practice that we are interested in. There’s no
shortcut from doing the knowledge analysis; we need to understand the relevant knowledge
systems and how they change, and then we can delineate the specific relations that hold. There is
no reason to believe that there will be simple generalizations, or that we can make assertions
without understanding the nuances of the knowledge systems involved. Furthermore, for
representations that, like natural language and writing, appear across much of our activity, the
knowledge systems involved will be extremely numerous.
This conclusion may seem disappointing. Rather than answering any questions about, for
example, the relationship between thought and literacy, I have simply proposed a quite extensive
program of research. Nonetheless, before we set out on this program in future work, there are a
few helpful morals we can draw from the analysis presented in Part 1 of this document. First, a
242
point that I made just above is important: For pervasive external representations, there will be
effects on many types of knowledge. The interesting thing about this observation is that it seems
to be in agreement with the notion that the existence of pervasive representations, like natural
language, has extensive effects on the nature of human thought. Furthermore, if we accept the
program I propose here, we can have these extensive effects without positing any absolutely unique
influences for specific representations like language. There may be many and multifarious effects,
and these effects may permeate through and relate to much of our knowledge, but they can do so
without producing any simple, global effects on the character of thought.
This does not mean that the project to study the relationship between thought and natural
language is impossible. In fact, I believe that there are likely some typical types of interactions that
an analysis can uncover. But these specific types of interactions are what we should be looking for,
not for single, all-encompassing effects.
A second moral we can draw from Part 1 is that new knowledge that develops through symbol
use, knowledge that is closely tied to the use of an external representation, may be more
interesting than one might first think. I have tried to argue that representation-related knowledge
cannot be relegated to the realm of non-meaningful algorithms and procedures. This moral is
especially important to keep in mind for technical representations, such as those used in science
and mathematics.
The final moral really subsumes the last two. Throughout this work, I have pushed for a view
of symbol use as meaningful activity, much like other activity. As humans, we remake our world,
populating it with representations, and these representations become part of our experience. This
viewpoint smoothes things out a little; it suggests a stance in which external representations do not
play an absolutely unique role. Nonetheless, because it treats representations as a meaningful and
pervasive part of our experience, it still leaves the door open for deep effects on the nature of
individual knowledge.
Comparison 1: The Symbol-Referent View of Representation
In the above section, I reflected on the properties of the theory presented in the preceding
chapters. Now I will go still further to bring out some features of the theory by contrasting it with
a number of alternative viewpoints. Actually, I have already presented a few such contrasts. Most
notably, I took some time to contrast symbolic forms with what I called the “principle-based
schemata” view. In many ways, that viewpoint was very closely related to the one presented in this
243
thesis. Like forms and devices, principle-based schemata are hypothesized, in part, to explain how
certain types of physics equations are written. Now, in this and upcoming sections, I will travel a
little farther afield and look at some viewpoints relating to the use of external representations that
have been developed for a more wide-ranging collection of purposes.
The first alternative view I will present is what I will call the “symbol-referent” view of
representations. Although this view is not always made explicit, I will argue that it is a common
view in educational literature. For this reason, this first contrast is perhaps the most important of
the ones I will present, and I will therefore devote the most time to it.
In what follows, I will first simply lay out the symbol-referent viewpoint and how it is applied,
then I will undertake to contrast it with the viewpoint adopted in this manuscript. In contrasting
the symbol-referent view with my own, the point will not be to critique this view. Instead, my
purpose will be to see what the symbol-referent view can and cannot do for us, and to understand
what sort of theory it is.
The basic symbol-referent view
To explain the symbol-referent view of representation, I will begin with an exposition by
Stephen Palmer (1977). In turning to Palmer, I am moving somewhat far from the concerns of
this work. Palmer is actually concerned with representations of a very different sort; his central
interest is in internal “mental” representations. Nonetheless, it is useful to begin with Palmer
because he lays out his position quite clearly. Furthermore, as we shall see, he has been cited by
researchers with concerns closer to those of this work.
Palmer begins by stating that a representation is “first and foremost, something that stands for
something else.” He then goes on to say that this description implies the existence of two
“worlds,” a “represented world” and a “representing world,” with correspondences between these
worlds. Thus, he argues, to fully specify a representational system, one must lay out the following
five aspects:
4. What the represented world is.
(5) What the representing world is.
(6) What aspects of the represented world are being modeled.
(7) What aspects of the representing world are doing the modeling.
(8) What are the correspondences between the two worlds.
244
This, in a nutshell, is the symbol-referent view of representation. There are two separate
worlds—one relating to the symbols and the other to the referents of those symbols—and there
are correspondences between specific aspects of these worlds.
Figure Chapter 7. -5. A represented world and some representations.
(From Palmer, 1977. Adapted with permission.)
Palmer provides a very clear example to illustrate what he has in mind. In Figure Chapter 7. 5, the represented world—a collection of rectangles of various widths and heights—is shown in
(A). Then, in (B) through (G), Palmer gives various alternative representations of this represented
world. For example, in (B), the height of the rectangles is represented by the length of some lines.
Similarly, in (C) and (D), the width and area of the rectangles are represented by line-length.
Thus, in each of these three representations, the same representing dimension is doing the
representing, but different aspects of the represented world have been selected. In other words, for
each of these, the answers to question (4) are the same, but the answers to question (3) differ.
Like (B), representations (E), (F), (G), and (H) all select the height of the rectangles for
attention; however, they represent this dimension in different ways. Thus, for these
representations, we would give the same answers to question (3), but different answers to question
(4).
This example suggests that the symbol-referent view can be applied to distinguish among
representations of various types. In fact, it turns out that sorting representations is one of the main
uses of the symbol-referent view. Palmer does a little of this. For example, he uses the term
“informationally equivalent” to describe representations that represent the same aspects of the
represented world.
245
Views like Palmer’s have appeared in multiple guises, and designed for many purposes. One
very notable version is presented by Nelson Goodman in his book Languages of Art (Goodman,
1976). Goodman’s purpose in this work is quite different than Palmer’s. Rather than being
interested in mental representations, he is interested in fleshing out some intuitive distinctions that
he believes exist among art forms. For example, a guiding question for Goodman is: “Why is
there only one original of a painting while a musical score can legitimately be played over and
over?”
Goodman’s view is elaborate and well worked out, but I will only give a taste of it here. Most
important to this project is Goodman’s discussion, in Chapter 4 of his book, of what he calls “The
Theory of Notation.” His purpose in this chapter is to work up to defining a class of
representational systems that he calls notations. Toward this end, he begins by defining a symbol
scheme as consisting of “characters, usually with modes of combining them to form others,” and
he goes on to explain that “characters are certain utterances or inscriptions or marks” (p. 131).
Then, later, he defines a symbol system as consisting of a “symbol scheme correlated with a field of
reference” (p. 143). Ultimately, Goodman goes on to describe requirements for the symbol
scheme and its relation to the field of reference that define when a representational system is
“notational.”
I am leaving out many interesting details here. Goodman actually develops his view quite
generally since he wants it to apply to a wide range of representational systems, including
paintings as well as musical and mathematical notations. For this reason, what he means by
“character” is quite flexible, and he is careful to include a great range of types of correspondence
between the symbol scheme and field of reference. Nonetheless, it is clear that there are some
basic similarities between the views of Goodman and Palmer. Both talk in terms of two worlds,
with correspondences between those worlds.
Now I will take a first step toward illustrating how the symbol-referent view is applied for
various purposes. One possible application of this view is to guide the design of representations, a
use to which Donald Norman (1991) has put the symbol-referent view. Among other things,
Norman is interested in judging the appropriateness of particular representations for specific tasks.
Toward this end, he discusses what he calls the “naturalness” of a representation. At issue, is how
difficult it is to describe the mapping between the represented and representing worlds:
The "naturalness" of a mapping is related to the directness of the mapping, where directness can be
measured by the complexity of the relationship between representation and value, measured by the
length of the description of that mapping.
246
The basic idea here is to look at the match between the represented and representing worlds, but
Norman allows for some complexity in this story. He is careful to point out that it matters what
“terms” are used in the mapping description, and he emphasizes that the proper terms are
“psychological, perceptual primitives.” In addition, he also maintains that, after working with a
representation for a while, experts can derive new mapping terms, thus increasing the feel of
naturalness.
Some useful examples appear in Norman (1991), as well as in Zhang and Norman (1994). In
the latter paper, following Stevens (1946), the authors describe four types of “psychological scales”
that one might want to represent within a certain class of representations: nominal, ordinal, interval,
and ratio. In a nominal scale, there is no ordering among the instances, the instances are simply of
different types. Next up this ladder is an ordinal scale, in which instances can be ordered. In an
interval scale, the separation between ordered instances has meaning. Finally, instances in a ratio
scale are true magnitudes; they are values relative to a zero measurement point.
Figure Chapter 7. -6. Two representations of nominal data. In (a) the representing scale is not matched. In
(b), the representing scale is matched.
(From Zhang & Norman, 1994. Adapted with permission.)
Although Zhang and Norman are not completely strict in this regard, they generally argue for
a matching between the scale types of the representation and the represented information: “Thus,
in order for a representation to be efficient and accurate, the scale types of the represented and
representing dimensions should match.” This matching of scale types is illustrated in Figure
Chapter 7. -6, which is adapted from Zhang and Norman (1994). In this example, the data to be
represented is the type of each file in a computer directory. Thus, the represented scale here is
nominal; there is no sense in which the instances are ordered. Figure Chapter 7. -6(a) uses a
representing scale that is not matched to this data, a ratio scale. Figure Chapter 7. -6(b) uses a
scale that is matched; each file type is represented by a different shape.
247
The symbol-referent view and mathematics education
Now I will discuss a second type of application of the symbol-referent view that is much closer
to the concerns of this project. The symbol-referent view has been used to help explain certain
types of difficulties that students have in using external representations and, following these
diagnoses, to aid in making prescriptions for instruction. This has especially been the case in
mathematics instruction. James Kaput’s Toward a theory of symbol use in mathematics is a notable
example, both because it is relevant, and because Kaput’s account is rich and extensive. I will now
discuss this work in some detail in order to illustrate how the symbol-referent view has been
applied in theorizing about mathematics learning and instruction (Kaput, 1987).
Kaput presents some definitions that are entirely along the lines of the symbol-referent view, as
we have already encountered it. He begins by explicitly recounting Palmer’s definition, just as I
cited it above. Then he lays out some further definitions that are very reminiscent of Goodman.
He defines a symbol scheme as “a concretely realizable collection of characters together with more
or less explicit rules for identifying and combining them,” and a symbol system as “a symbol
scheme S together with a field of reference F, and a systematic rule of correspondence c between
them … .” Clearly, Kaput is operating within the territory of the symbol-referent view.
Kaput goes on to specialize this view to mathematics in some interesting ways. He argues that
mathematical symbol systems are a special kind of representational system in which the
represented world is itself a symbol system. All of mathematics is then built up through a chain of
reference, which begins from “primary referents,” and then builds layers of symbol system on top
of symbol system, each one becoming the represented world for the succeeding layer.
This repeated consolidation of mathematics in layers is very interesting. In fact, there may be
some interesting connections to my own view here, since Kaput allows the consolidation of
previous experience with mathematical manipulations into meaningful structures. However, I will
not explore this line in what follows.
Instead, I want to work toward seeing how Kaput uses the symbol-referent view to make some
specific claims about mathematics learning. There are actually a number of interesting comments
that he makes, but here I will just focus on one important piece of his discussion. Kaput lists two
types of “cognitive activity” associated with symbol systems: (1) Reading information from a
symbol system or encoding information into a symbol system, and (2) the elaboration that occurs
once information has been encoded in a symbol system. Kaput further divides the latter category
into two types of activity, syntactic elaboration, in which the individual directly manipulates
248
symbols (either in their head or on a sheet of paper), and semantic elaboration, in which the
individual reasons with the referents of the symbols. As an example, a person might add 23 and 9
by mentally visualizing symbols and performing two-column arithmetic (syntactic elaboration), or
by directly invoking knowledge of whole-number arithmetic (semantic elaboration).
Now, with these last distinctions laid out, Kaput is in a position to understand—or at least
provide a certain type of account of—some of the central difficulties experienced by students.
One of these difficulties is the observation that students seem to manipulate symbols without
understanding what they mean. For example, students may learn arithmetic algorithms, but not
know why they work, or how to modify and apply them. In Kaput’s terms, we can understand
students as stuck within the representing world of the symbol scheme. They work with symbols,
performing syntactic elaborations, but they do not read information out of this scheme to the
reference field, or perform the corresponding semantic elaborations.
Thus, for Kaput, symbol systems are a two-edged sword. One of the great benefits of symbol
systems is that they allow us to perform manipulations (syntactic elaborations) free of many of the
details of the reference field. But with this power comes a responsibility:
In each case we have alluded to, one sees an enormous increase in power resulting from the
development of a symbol system whose scheme’s syntax and correspondence faithfully preserve certain
relationships in its reference field. ... For individuals, however, the power and freedom of the
formalism brings an added responsibility—the ability to interpret the formal procedures back in the
reference field(s). Otherwise, all one has are meaningless symbols. (p. 180)
This quote exemplifies one of the most prominent ways that the symbol-referent view is used to
understand student difficulties in mathematics. The point is that working with mathematical
symbols is powerful, and it’s a power that we certainly want to give to our students. But, if we give
students this power, then they must—in Kaput’s words—live up to their responsibilities. They
need to know how to translate from the formalism to the reference field. If they do not, then all
that they have are “meaningless symbols.”
Though not always quite as explicitly, similar notions show up in many places in the
mathematics education literature, and we will see more examples in a moment. For now, I just
want to briefly mention one other related manifestation of a symbol-referent-like argument.
Mathematics educators are often concerned with the relations between symbols and
“embodiments,” which are more or less specific types of physical manifestations of mathematical
abstractions. For example, embodiments are frequently discussed in relation to instruction on
fractions, using such examples as a pie cut into slices or a collection of beads with several removed
249
(e.g.,Post, Wachsmuth, Lesh, & Behr, 1985; Chaffe-Stengel & Noddings, 1982). In these cases,
the relation between the embodiment and the symbolic representation is often understood in
terms much like the symbol-referent view.
In addition, similar discussions appear in relation to the use of “manipulatives” in instruction,
which can be treated as a sort of embodiment. For example, Resnick and Omanson (1987)
undertook to give students a better understanding of arithmetic algorithms by carefully teaching
them correspondences between these algorithms and a particular manipulative, Dienes Blocks.
Interestingly, a careful empirical study of this instructional technique revealed largely negative
results.
How the symbol-referent view is converted to a cognitive theory
Now that I have described the symbol-referent view, I will contrast it with my own account of
symbol-use in physics. My purpose here will not be to argue that my view is correct and the
symbol-referent view is incorrect, though there will be places that I argue for caution in applying
the symbol-referent view. Instead, my position will be that these are theories of a different kind,
with different strengths and regimes of application.
To begin, recall that the Theory of Forms and Devices is a cognitive theory; in order to
describe and explain certain types of behavior, I posit the existence of particular mental entities. In
contrast, reflect for a moment on the nature of the symbol and referent worlds. Are they
knowledge systems? Are they cognitive entities of any sort? Although we could invent a theory in
which these worlds are cognitive entities (as we shall see in a moment), a pure version of the
symbol-referent view need not take these worlds to be cognitive in nature. Instead, in adopting the
symbol-referent view, we choose to study symbol systems, rather than cognitive systems, and these
symbol systems involve correspondences between a symbol scheme and a field of reference.
This observation does not invalidate the symbol-referent view. We have already seen that this
view may have some important uses, for example in distinguishing among types of
representations. The point is just to notice that symbol-referent theories are of a different sort than
the form-device theory. Nemirovsky (1994) makes a similar point in distinguish “symbol
systems” from “symbol use:”
With “symbol system” I refer to the analysis of mathematical representations in terms of rules. … On
the other hand with “symbol-use” I refer to the actual concrete use of mathematical symbols by
someone, for a purpose, and as part of a chain of meaningful events. Symbol-use may enact some rules
of the symbol system that is being used, but the central point is that it cannot be reduced to those rules.
250
However, it turns out that there are cases where the symbol-referent view essentially gets
converted into a cognitive model. Sometimes this is implicit and piecemeal, but sometimes it is
quite explicit. For example, in Syntax and Semantics in Learning to Subtract, Lauren Resnick paints
a picture of a particular debate in mathematics education (Resnick, 1982):
Until quite recently, discussions of meaningfulness in arithmetic learning have been characterized by a
confrontation between those who advocate learning algorithms and those who argue for learning basic
concepts.
Her thesis in this particular paper is that it is not enough to learn “algorithms” or “basic concepts,”
they need to be hooked up correctly:
Proponents of conceptually oriented instruction have frequently argued that a major cause of difficulty
in learning arithmetic is a failure to relate rules of computational procedure to their underlying
mathematical concepts. ... The data presented here show that intuitions concerning the importance of
conceptual understanding are correct in a crucial respect: Difficulties in learning are often a result of
failure to understand the concepts on which procedures are based. But the data also show that even
when the basic concepts are quite well understood, they may remain unrelated to computational
procedure.
She goes on to display an algorithm for subtraction and says:
Because performing this algorithm requires no understanding of the base system, I claim that it
includes no semantics .... The evidence I present here suggests quite strongly that many children may
learn the syntactic constraints of written subtraction without connecting them to the semantic
information that underlies the algorithm.
The details of Resnick’s views – how, exactly, she believes that this “hooking up” occurs – are
actually quite relevant to this dissertation. However, for the present, I just want to look at how
Resnick paints the issues. The image she presents is of two separate categories of cognitive
resources, one associated with algorithms, the other with understanding of the “underlying
mathematical concepts.” This view appears very clearly in Resnick and Omanson’s (1987)
discussion of mapping instruction, which I mentioned above.
The idea was that the child came to mapping instruction with a mental representation of subtraction in
blocks that was attached to a rich knowledge base, including the principles of subtraction discussed in
this chapter. The child also entered instruction with a mental representation of written subtraction
that was purely syntactic in nature. According to the hypothesis, the detailed, step-by-step alternation
between the two routines would allow the child to attach to the written symbols the principles he or
she already knew with respect to the blocks. This would happen, according to our hypothesis, because
the child would come to think of the written symbols as representing the blocks—thus the emphasis in
our instruction on using the writing to record the results of blocks action.
Again, the idea is that there are two categories of mental resources, one associated with “written
subtraction” and “syntactic” manipulations of symbols, the other associated with a “rich
knowledge base.” Due to the negative outcome of their study, Resnick and Omanson came to
251
somewhat recant the view expressed above. In fact, in the above passage they are describing a view
that they no longer held at the end of the study! Nonetheless, this passage and the ones above are
interesting for how they make use of the symbol-referent view. In these passages, I argue, the
symbol-referent view is embodied as a cognitive theory. In addition, this cognitive theory is of a
particular sort: The symbol world and referent worlds have been associated with distinct
categories of cognitive resources.
Furthermore, in associating the symbol and referent worlds with distinct categories of
cognitive systems, there seems to be an implicit presumption that each of the two categories of
cognitive resources has a somewhat different character. The cognitive resources associated with the
symbol world are described as “syntactic” and “procedural,” while those associated with the
referent world are “semantic” and “conceptual.” These types of descriptions tend to accord the
symbol-related resources a second-class status. The description of the symbol world constraints as
syntactic implies a certain kind of simplicity. We expect that, for some notations, the constraints
will be easy to specify. If the conventions can be thoroughly spelled-out, then work within the
symbol world becomes simple, almost trivial, and with this simplicity comes the idea that the
symbol world is impoverished. If symbol-related knowledge is only about the meaning-free
manipulation of symbols according to a simple set of rules, then any “understanding” must be
localized to the world of semantics, in referent-related knowledge.
This image of two categories of knowledge is a very powerful one, and it is strongly tied to the
notion, discussed earlier, that student problems are traceable to a failure to “translate” back to the
reference world. However, given the viewpoint I have presented in this work, there is some reason
to be skeptical about cognitive models of this form. Notice that, in the model that I presented,
there is no clear breakdown into categories of this sort. In particular, the form-device resource
does not belong clearly in either the symbolic of non-symbolic category. It is knowledge that is
clearly tied to symbols, but I have argued that it should be considered a component of our
deepest conceptual understanding. Again, this is not necessarily a problem with the symbolreferent view, only a problem with this particular type of manifestation of it as a cognitive theory,
though this discussion does suggest reason for caution.
The problem with this view is not the description of student behavior—at its worst—as
meaningless formal manipulation. It is possible that this particular family of cognitive models may
be useful for providing rough descriptions of earlier, less successful, mathematical behavior. The
problem is the assumption that symbols and conceptual knowledge remain separate even with the
252
development of expertise. Notice, in some of the Resnick quotes above, the implication that
“concepts” exist to be learned separate from symbolic procedures: “…even when the basic
concepts are quite will understood, they may remain unrelated to computational procedure.”
In my view, as people learn to use computational arithmetic in its symbolic form, conceptual
understanding develops, with symbols as legitimate participants in meaning. Educationally, this
means that we cannot simply direct our instruction at forging links between an existing body of
conceptual knowledge and formal procedures. The integration and development that must occur
around forms is a non-trivial affair that depends, among other things, on the nature of the
computational manipulations employed.
To conclude this section, I want to emphasize that I do not mean to be overly critical of the
research I used to exemplify the symbol-referent view. I chose these researchers because they make
their assumptions clear and explicit, and because their work is rich and useful. Instead, my real
hope in trying to draw out the symbol-referent view has been to ultimately draw attention to these
assumptions where they have been left implicit. I believe that much recent educational innovation
in mathematics and science has been implicitly driven by the symbol-referent view and the
presumption that formal symbolic methods are somehow conceptually impoverished. These
assumptions, I believe, deserve to be considered more carefully.
Is the Theory of Forms and Devices a symbol-referent theory?
I want to continue my contrasting of the symbol-referent view and the Theory of Forms and
Devices. In this sub-section, however, I want to begin by turning the discussion around a little.
The question I want to ask here is: Can we see the Theory of Forms and Devices in terms of the
symbol-referent view? For example, we might decide that the collection of forms counts as a
reference field or represented world, with the external symbolic display playing the role of the
representing world.
Of course, whether this counts as a version of the symbol-referent view depends on how
loosely that view is defined. However, I believe that if the symbol-referent view is defined so as to
include this relation between forms and symbolic displays, then we have a version of the symbolreferent view that is so loose as to not be useful. Viewed at a very high level, form-device theory
involves mental structures that come into play depending, in part, on the particular symbols that a
person sees. But this structure is far from specific to the form-device theory; a structure in which
the activation of mental elements depends, in some way, on the elements of the environment
253
present, is extremely generic. Therefore, if we decide that any type of association between mental
elements and symbolic displays counts as a symbol-referent relation, then the symbol-referent
view will not help particularly in distinguishing among cognitive models.
It is also helpful to look more closely at the particular type of association that holds between
forms and written symbolic expressions. This association is really somewhat looser than the sort of
alignments that are characteristic of stronger versions of the symbol-referent view. Going back to
my discussion of Palmer, recall his presumption that the representing and represented worlds are
divided into “aspects,” and then there are fairly strict correspondences between these aspects. In
contrast, associations between symbolic forms and equations do not have this structure. First,
there is no “world” that is neatly sliced up by a collection of forms. A student’s repertoire of forms
are a loose collection of related schemata, and this repertoire, taken as a set, does not decompose
any motion or class of physical situations. In addition, the alignment between forms and symbolic
expressions is not strict. Any expression can be seen in terms of forms in multiple and overlapping
ways. Thus, unlike Palmer’s example representations in Figure Chapter 7. -5, there is not any sense
in which there are clear aspects of the representing world that do the representing, and which
correspond to aspects of the represented world.
This is not to say that the relation between symbolic forms and equations is totally
unconstrained and not characterizable. I have described quite specific relations that exist between
forms and equations. The point is that these relations are not of the same clean sort as those
typically associated with the symbol-referent view.
Actually, these last comments suggest some reasons to worry that the symbol-referent view,
even in its non-cognitive form, will not be appropriate for some analytic tasks. In applying the
symbol-referent view to help understand instructional difficulties, the presumption is generally
made that this view will be useful, at some level, for describing what people do with symbols.
However, the observations and arguments that I have presented in this document may indicate
that this presumption does not hold in all cases. If the form-device theory is an accurate account
of symbol use, with its multiplicity and context dependence, then my observations may cast
doubt on our ability to give a useful symbol-referent account. We may not, in general, be able to
find clean refers-to relations that are either explicit or implicit in symbolic activity.
This observation has fairly wide-ranging implications. We are very accustomed to thinking of
representations as things that, as Palmer says, “stand for something else,” according to rules that
we could find if we worked hard enough. But we should realize this is an assumption that might
254
not hold, or, at the least, may hold only in limited cases, such as the toy representation in Figure
Chapter 7. -5.
I believe that a more neutral stance is required, at least initially, with regard to how
representations function in human activity. We populate our experience with all kinds of artifacts
and, for most of these, we would not presume that we could find correspondences between
aspects of these artifacts and other entities. My proposal is to take a similar stance in studying the
use of external representations: We should attempt to characterize symbol use as we find it in the
world, perhaps positing knowledge along the way, without the assumption that we will find that
representations “refer.”
Comparison 2: Similar Views From Mathematics Education
I will now counterbalance the discussion in the preceding section with a comparison to some
mathematics education research that is somewhat closer to the viewpoint expressed in this work.
Actually, most of this research was already discussed in earlier chapters, where appropriate, so my
task here will be to just draw together those earlier discussions and summarize the comparison to
the present work.
Recall that, in Chapter 3, I discussed what have been called “patterns” in arithmetic word
problems (Greeno, 1987; Carpenter & Moser, 1983; Riley et al., 1983; Vergnaud, 1982). This
research asserts that students learn to recognize a relatively small number of schemata that guide
the solution of arithmetic word problems. These schemata are given names like “change”,
“equalization,” and “compare.” As I discussed, there are some relatively clear similarities between
this view and my own. Most notably, these schemata seem to live at a very similar level of
abstraction as forms, and there are even some specific forms that are closely related, such as
BASE +CHANGE .
However, there are also some clear differences. For example, the patterns research
is more interested in the role that the schemata play in driving a solution process than the role they
play as ways of understanding expressions.
Another set of closely related research, which I also discussed in Chapter 3, appeared under
the topic “meanings of an equal sign.” Researchers in mathematics education have contrasted
different stances that students adopt toward equations and the equal sign (Kieran, 1992;
Herscovics and Kieran, 1980). In some cases, students see an equal sign as a “do something”
signal. The contrasting stance is when an equation is treated as an object in its own right, with the
equal sign specifying a symmetric, transitive relation. Kieran (1992) argues that learning to adopt
255
the object stance is a developmental achievement, with the process stance being more natural early
on in instruction, but both stances are seen as playing a useful role in expertise.
There is an important sense in which both of these collections of research may differ from the
view expressed in this project. For these other researchers, a relatively small number of stances or
schemata suffice to characterize the way students understand and write equations. Kieran
describes two main categories of stances toward an equation. And the patterns research
characterizes all of its schemata in terms of operations on and relations between sets. In fact, one
gets the feeling that we could write down all of the possible schemata, just by imagining the
possible operations that are possible on sets.
But this procedure would fail to generate many of the symbolic forms that I have identified.
Notice, for example, that the COMPETING TERMS form cannot be understood in terms of sets; it is
fundamentally about influences. The point here is that, in contrast to this mathematics education
research, I generally believe that stances and meanings will be multiple and varied.
These differences are partly explainable by the fact that there are some important differences
in the regime studied, as compared to the present project, not least of which is that most of this
mathematics-related research was done with much younger students, working on a limited class of
problems. Furthermore, these researchers do not suggest that their categories necessarily extend
beyond their regime of study. Nonetheless, I believe it is appropriate to sound a note of caution
here. We cannot presume that it will be easy to guess a repertoire of symbolic forms given any
individual viewpoint, such as that provided by operations on sets. Furthermore, as students learn
to use equations in new domains, we should expect new forms and devices to develop—neither
the vocabulary of sets, nor any other vocabulary, will suffice for all domains.
Note that this last comment relates to one of the central themes of this work. In saying that
new forms may develop for each domain, I am saying that there is important symbol-related
knowledge that is domain dependent. This is contrary to a view in which mathematics is a tool
that transcends domains of application.
Finally, I want to close this section with a discussion of the well-known “Students-Professors”
problem (Clement, 1982; Kieran, 1992). This is a seemingly very easy problem that surprised
researchers by being quite difficult for students:
Write an equation using the variables S and P to represent the following statement: “There are six
times as many students as professors at this university.” Use S for the number of students and P for the
number of professors.
256
The correct answer to this question is the expression S=6P. Interestingly, Clement (1982) found
that only 43% of a population of university engineering students solved this problem correctly. In
addition, of those that got it wrong, 68% reversed the expression, writing 6S=P. Furthermore,
Clement found that this latter result was quite robust; students did not rapidly change their
responses, and the mistake appeared even when some hints were included in the question.
Clement explains these observations by saying that students apply one of three cognitive
schemes in solving the students-professors problems. In the first, which he calls “word order
match,” the equation is essentially written to follow the ordering of key words in the problem
statement. The second scheme, “static comparison,” involves a direct comparison between two
related groups, in this case, the 6 students and 1 professor. Evidence for this is that, when writing
their expressions, students made statements of the form: For each 6 students there’s a professor.
Clement describes these first two cognitive schemes as involving a “passive” association, and both
produce the reversed expression 6S=P.
Finally, in the last scheme, “operational equality,” the student imagines operating on the
number of professors to get the number of students. Clement characterizes this last scheme as
“active” rather than passive, and it is the only one of the three schemes that produces the correct
answer. Clement comments: “the key to understanding correct translations lies in the ability to
invent an operation … and to realize that it is precisely this action that is symbolized on the right
side of the equation… .”
Clement’s distinction between “passive” and “active” schemes can be discussed in terms of
Kieran’s (1992) distinction between structural and procedural stances toward equations, with the
interesting conclusion that a procedural stance (which is presumed by Kieran to be
developmentally less advanced) seems to be more successful here. Kieran (1992), in fact, makes a
similar comment.
But we can also understand Clement’s observations in terms of forms and devices. In the static
comparison scheme, there is likely some type of Static device in play, along with a BALANCING -like
form. In contrast, in the operational equality scheme, the student employs a device that narrates an
operation on the number of professors (i.e., employs a Narrative device), perhaps by imaging the
substitution of a value for P into the expression. Furthermore, this latter scheme would be
associated with some sort of forms involving an asymmetric relation between quantities, such as
DEPENDENCE
or the SCALING form.
257
A little more care would be required to accurately account for Clement’s results in terms of
forms and devices. A more appropriate explanation might very well involve some forms and
devices other than the ones that appeared in my data corpus. For example, I believe that there are
some differences in nuance in the BALANCING -like form that appears in Clement’s static comparison
scheme. And the plugging-in-numbers device that appears in the active scheme does not precisely
match with any of the devices that I have previous mentioned, though it is clearly in the ballpark
of the Narrative devices I discussed.
Nonetheless, I believe that the form-device theory can nicely account for Clement’s
observations, just as it subsumes the other research discussed in this section. What we are seeing
here is that the form-device view can capture many of the insights associated with this crosssection of mathematics education research. We do not need entirely new theoretical work to
account for Clement’s schemes, the patterns research, or the contrast between structural and
procedural stances toward equations.
Comparison 3: Situated Cognition and the Ecological Perspective
Throughout this work, I have taken an explicitly cognitive approach, positing mental structures
as my key claims. But one collection of orientations that has appeared under such names as
“situated cognition” and “situativity theory,” questions the validity of this entire approach
(“Situated Action,” 1993). In this section, I will discuss this alternative view and how it relates to
the present work.
Let us take a moment to reflect on the program of cognitive science. As I have said, cognitive
science works by positing knowledge that is presumed to be possessed by individuals. Using such a
framework, we can think about, for example, how to describe the results of school instruction. In
school, a cognitive scientist might say, people learn certain kinds of knowledge and then carry that
knowledge out into the world, using it where appropriate. This essentially means positing invariant
structures in the minds of individuals—structures that are the same across multiple situations, such
as school and home.
Furthermore, cognitive science makes some presumptions about the nature of these invariant
structures. In particular, they are presumed to live at something like an “informational level.”
Cognitivists are not only interested in pointing to neurochemical regularities in the brain, they
want to be able to, perhaps, state that there are mental structures associated with such things as
“concepts” and “propositions.”
258
For illustration, take a toy cognitive theory that purports to describe the knowledge that
people learn in mathematics instruction. We might hypothesize, for example, that students learn
arithmetic algorithms in their mathematics classes, and then, once these are learned, they carry
these algorithms around in their heads, using them in the world where appropriate. According to
this toy theory, this is what it means to “know math.”
But there is much evidence that this toy theory is far from good enough. Often cited in this
regard are the results of the “Adult Math Project,” a project conducted by Jean Lave and others
(Lave, Murtaugh, & de la Rocha, 1984). Among other things, this project set out to study the
mathematics that adults use while shopping in the supermarket. To this end, the researchers
followed people into a supermarket, watched what they did, and attempted to characterize their
mathematical activity. Looking at the observations and analyses produced by this project, it is
clear that it is far from sufficient to describe shoppers as simply taking algorithms learned in
school and applying them in supermarkets. Instead, in solving best-buy problems, for example, the
behavior of shoppers was wide-ranging and idiosyncratic, depending on the concerns of the
moment and what was available in the environment.
Results of this sort have been interpreted by some as a reason to reconsider some of the
fundamental assumptions of cognitive science. The program of describing knowledge seems
plausible if we believe that people learn procedures, like arithmetic algorithms, and then apply
those procedures in the world. But what if this is not the case, and the behavior of individuals is
highly idiosyncratic and context dependent? In that case, the whole project of looking for
knowledge, at least as cognitive science has traditionally done it, may not make sense, since it may
not be possible to identify structural invariants of an interesting sort.
This is where situated cognition comes in. The view of situated cognition is that cognitive
science has over-emphasized the determining effects of cognitive structure at the expense of the
influence of the material, social, and historical environments. For the case of supermarket
arithmetic, this means that you cannot expect to say much about this variety of math without
knowing about the details of the material and cultural context. This is frequently taken to suggest
that the very nature of what we mean by “knowledge” must be changed: Situated cognition
theorists often maintain that we should no longer think of knowledge as something localized in
the heads of individuals; instead, they say, we should think of “knowing” as existing only in the
interaction between person and environment.
259
I believe that observations like those of the Adult Math Project are entirely sensible, and I take
the observed sensitive dependence on the situation as absolutely fundamental. In fact, my focus on
external representations is partly for this reason. However, I do not believe that these observations
require that we change the very nature of what we mean by knowledge. Instead, I believe that
these observations simply tell us something about the type of cognitive theories that we need to
build: We need to push down in grain-size, and the problem of describing knowledge is probably
going to be much harder than one might initially have thought, but I do not believe that this is a
reason to give up on knowledge, understood in the traditional sense.
Even with this major point of difference between my own view and that of situated cognition
researchers, it is nonetheless the case that their concerns are often close to my own. As I
mentioned, one of the concerns of situated cognition is to draw attention to the material
situatedness of cognition. Because of this emphasis, these researchers, like myself, tend toward a
particular concern with the role of external representations. For this reason, a more detailed
comparison is merited.
There is no uniform theory of situated cognition, only a collection of loosely related views,
and there is no way I can do justice to this variety. Instead, I will just look closely at two
researchers from a particular branch of situated cognition—the “ecological perspective”— that
happens to have concerns close to those of this work. Greeno and Engle (1995), who ascribe to this
perspective, describe their approach as follows:
we consider cognitive processes such as reasoning, understanding, and representing to be accomplished
by systems that include humans interacting with each other and with available material resources,
rather than as processes that only occur inside individual minds.
The idea is to look at a person working with symbols in some situation as a whole system of
interacting elements that act together, perhaps toward some end. The analysis then looks for
regularities in the interactions among elements, and tries to describe how the various constraints
within the system lead to some goals being reached. Thus, in contrast to cognitive science, the
structural invariants identified by the ecological perspective will, in general, span person and
environment.
In their particular study, Greeno and Engle looked at people using a physical device they call a
“winch machine.” This device consists of blocks that slide along numbered tracks, with the blocks
connected by strings to spools of various diameters. When the spools are turned, the blocks slide
across the track by an amount that depends on the number of turns and the diameter of the
spools. Greeno and Engle gave seventh and eighth grade students the task of making a table of a
260
block’s position as a function of the number of turns of the spool, for spools of various sizes. Thus,
in this situation, the elements of the “ecology” include features of the winch machine, the written
tables, and the students.
Greeno and Engle performed several types of analyses on their transcripts of this activity. One
analysis was based around what they called “activity nests.” They decomposed transcripts of
activity into nested “functions,” saying which student activities contributed toward which
functions. In addition, they performed an analysis of “semiotic networks,” attempting to identify
networks of refers-to relations among elements of the ecology. Altogether, their image is of a
system, which includes students and external representations, interacting (in ways that they
describe) to achieve certain functions.
Learning, here, is a gradual attunement to constraints and affordances in the situation. This
means that a person develops a sensitivity to certain features of the environment, and can react to
these features in certain ways. This is how Greeno and Engel account for the types of change in
behavior associated with learning: If students are attuned differently, then they will interact
differently with the rest of the ecology, and the behavior of the system changes.
Though it may sound quite different, there are actually some similarities between Greeno’s
ecological view and the view presented here. Most notably, my argument that students become
sensitive to certain registrations, especially what I called symbol patterns, might be understood as
an attunement to an affordance. But there are important differences, even beyond the different
orientation to the nature of knowledge. In my discussion of the symbol-referent view, I already
questioned the possibility of finding refers-to relations in symbol-related activity. In addition,
Greeno and Engel’s function-based analysis departs from the analyses described in this document.
One of the main objectives of an ecological analysis is to explain how the interactions and
constraints in a system conspire to allow the system to reach some goal or perform some function.
In my analysis—as I have noted on a number of occasions—I did not attempt to explain how the
use of particular forms and devices contribute to the reaching of any particular goals, such as the
solving of a problem.
Before wrapping up this section, I want to look at one other piece of research, also done from
an ecological perspective, but which is closer to my own program in population and domain focus.
Rogers Hall looked at undergraduate computer science majors solving motion and work problems
involving linear rates (Hall et al., 1989; Hall, 1990). For example, one problem was as follows:
Two trains leave the station at the same time. They travel in opposite directions. One train travels 60
km/h and the other 100 km/h. In how many hours will they be 880 km apart?
261
This problem can be solved using algebraic expressions, manipulated in the usual ways, and it is
likely that Hall’s college students have some experience with these techniques. However, Hall
found that even competent problem solvers engaged in techniques outside of these formally
recognized activities. Of particular interest was what Hall called “model-based reasoning,” in
which “a student ‘executes’ a model of the problem situation along the dimension defined by an
unknown quantity such as time, distance, or work.” For example, a student might step through
time, figuring out the location of each train at each moment of time. In addition to model-based
reasoning, Hall characterized a large number of tactics used by competent solvers to reach
solutions.
In summary, what Hall found was that these competent problem solvers do not just employ
the few formally recognized—and supposedly very general—techniques they learned in their
mathematics classrooms. Instead, they employ a wide variety of tactics and seem to construct
techniques as needed. Given these observations, Hall argued for a “constructive” rather than a
“recalled” view of problem solving. Notice that this view is very reminiscent of some of my
comments concerning equation use and problem solving in physics. I explicitly contrasted my own
account with what I called the “principle-based schemata” view of physics problem solving. In that
view, people possess schemata associated with physical principles, and these schemata dictate what
equations are written and how they are used to solve a problem. In contrast, the Theory of Forms
and Devices allows for the construction of novel expressions, and for interaction with equations at
a level other than that specified by formal principles.
Thus, there are some important similarities with Hall’s work. The differences come when, like
Greeno, Hall situates his observations within an ecological perspective (Hall, 1990). Like Greeno
and Engel, Hall is concerned with a discussion of how the ecology achieves certain goals. For
example, he discusses how the use of the various tactics he observed supports different types of
inferences. Nonetheless, the concerns of both Greeno and Hall are very close to those of this work.
They, like me, attend to the more informal and flexible aspects of activity with external
representations, thus escaping from the relatively rigid view of symbol use that sometimes prevails
in cognitive science.
Comparison 4: A Bodily Basis for Meaning?
Finally, I want to conclude this chapter with a comparison to a viewpoint that, strictly
speaking, is not particularly concerned with external representations. However, this is a worthwhile
262
comparison to include here because it can serve to highlight certain important features of my
stance. In this section, I will address two related works: Women, Fire, and Dangerous Things, by
George Lakoff and Mark Johnson’s The Body in the Mind: The Bodily Basis of Meaning,
Imagination and Reason (Johnson, 1987; Lakoff, 1987).
Since Lakoff’s discussion, to a certain extent, subsumes Johnson’s, I will begin by following
Lakoff. Lakoff’s goals are quite broad and ambitious. In Women, Fire, and Dangerous Things he
sets out to define a new version of philosophical realism he calls “experiential realism” or
“experientialism.” Like many rejections of traditional realism, Lakoff begins by asserting that
there is not an objectively knowable world; our concepts do not exist separate from us to be
discovered. Then, what distinguishes experientialism is the notion that the concepts that we use
for understanding the world depend, in particular, on the bodily nature of our experience.
More specifically, the idea, for both Lakoff and Johnson, is that we have set of “preconceptual
experiences” that are “directly” meaningful. Then, our concepts get their meaning only indirectly,
by being tied to this preconceptual experience in certain specific ways. Lakoff explains:
“Experientialism claims that conceptual structure is meaningful because it is embodied, that is, it
arises from, and is tied to, our preconceptual bodily experiences” (p. 267).
Furthermore, the program seeks to identify types of “structure” that exist in our preconceptual
experience. Lakoff states that there are at least two kinds of structure. The first “basic-level
structure,” is closely related to what psychologists have called basic-level categories. The second
type of structure is called “kinesthetic image-schematic structure.” These “image schemata” are
primitive schemata that derive from the basic structure of our experience in the world. It is here
that Lakoff makes principle contact with Johnson’s book, which provides an extensive discussion
of these image schemata. One of Johnson’s principle examples is what he calls the “containment”
schema:
Our encounter with containment and boundedness is one of the most pervasive features of our bodily
experience. We are intimately aware of our bodies as three-dimensional containers into which we put
certain things (food, water, air) and out of which other things emerge (food and water wastes, air,
blood, etc.). From the beginning, we experience constant physical containment in our surroundings
(those things that envelop us). We move in and out of rooms, clothes, vehicles, and numerous kinds of
bounded spaces. We manipulate objects, placing them in containers (cups, boxes, cans, bag, etc.). In
each of these cases there are repeatable spatial and temporal organizations. In other words, there are
typical schemata for physical containment. (p. 21)
Other examples of image schemata given by Lakoff and Johnson are “balance,” the “path”
schema, “part-whole,” “up-down,” as well as some schemata relating to forces, including
“blockage,” and “compulsion.”
263
Lakoff and Johnson’s program is not only concerned with identifying structure in
preconceptual experience. In addition, they set out to describe how “abstract conceptual structure”
arises from the preconceptual structure. Lakoff lists two ways that this occurs:
1. By metaphorical projection from the domain of the physical to abstract domains.
2. By the projection from basic-level categories to superordinate and subordinate categories.
The idea is that abstract conceptual structures get meaning by being tied to preconceptual
structure. In some cases, for example, this happens through metaphors to directly understood
domains.
One of the examples that Lakoff provides is the “more is up; less is down” metaphor. He gives
many example of how this metaphor is manifested in our language: “The crime rate keeps rising.”
“That stock has fallen again.” The point is that this is a case in which a particular abstract target
domain, “quantity,” is understood indirectly by being tied to the directly understood source
domain of “verticality.”
Interesting, like “patterns” in arithmetic word problems, image schemata seem to live at a level
of abstraction that is similar to that of forms. Varieties of the part-whole schema have appeared in
all three cases, and Johnson even has his own version of a balance schema. This is important, if only
because it means that Johnson’s list of image schemata can be used as a source of potential forms.
It is particularly illuminating to look at one relevant example given by Johnson. Johnson asserts
that “mathematical equality” is understood by metaphorical project from a version of the balance
schema:
Here the mundane TWIN -PAN BALANCE schema is projected onto the highly abstract realm of
mathematical computation and equation solution. The basis for this projection is a point-by-point
mapping of entities from the physical realm into the abstract mathematical realm … . (p. 90)
This sounds very similar to some of my own assertions, but there are important differences, even
beyond the observation that Johnson glosses what I have taken to be rich territory with a single
schema. The big difference is that the whole Lakoff-Johnson view is based on the presumption
that certain types of experience occupy a privileged status. There is the more concrete,
preconceptual variety of experience, and more abstract categories of experience. Furthermore, the
more abstract experience, like that associated with mathematics, is only understood indirectly, by
metaphorical project from structures in more basic experience.
264
In contrast, in my view, there is not a stable background of preconceptual structures associated
with a special category of experience. Instead, there is a constant re-churning of our
understanding. One of the main points here has been to argue that through experiences with
symbols—and all other sorts of experience—the most primitive levels of our understanding
continue to evolve.
In fact, it is from the comparison to Lakoff and Johnson that this manuscript gets its title.
Recall that the sub-title of Johnson’s book is The Bodily Basis of Meaning, Imagination, and Reason,
a phrase that attempts to capture Johnson’s view that all meaning is ultimately traceable to bodily
experience. In contrast, the thesis of this work has been that experience with symbols is just as
basic as any other kind of experience, and that it has effects that percolate down to the most basic
levels of our understanding and intuition. Thus, my argument is that there is not only a bodily
basis of physical intuition, there is also a symbolic basis.
ReflectionÕs End
In this chapter, I wrapped up Part 1 of this thesis by reflecting on the nature of the theory
outlined in the preceding chapters. This actually was done in two parts. In the first half of this
chapter, I reflected on some properties of the theory itself. Then, in the second half, I worked by
contrasting my view to other research perspectives.
In reflecting on the nature of the Theory of Forms and Device, I discussed the sense in which
the theory implies that physics understanding is “intrinsically representational.” There were
essentially two answers. First, I argued that physics knowledge is refined specifically by and for
symbol use. Second, I argued that, through symbolic experience, a whole new resource develops,
the form-device system, and that this resource should be considered a component of physical
intuition.
In the first half of this chapter I also briefly discussed how the theory and its conclusions
might be generalized. I described how the Theory of Forms and Devices could be extended to
another representational practice, graphing-physics. In addition, I speculated concerning more
general implications for the relation between knowledge and external representations. My
conclusions in this regard were not definitive. I argued that there is simply no shortcut to avoid a
thorough analysis of the knowledge systems involved in any practice of symbol use, and that there
may potentially be as many relations between knowledge and external representations as there are
knowledge systems.
265
Then, in the second half of this chapter, I contrasted my view to four other research
perspectives. Among the important conclusions was the simple observation that the theory
presented here is a cognitive theory, which distinguishes it from pure symbol-referent theories or
an ecological perspective.
In addition, I tried generally to contrast my view with perspectives that tend to treat symbolic
knowledge and experience as semantically impoverished. I do not believe that there are two
categories of cognitive resources—one syntactic and non-meaningful, the other non-syntactic and
meaningful—as sometimes implied by applications of the symbol-referent view. And I argued
that unmediated bodily experience is not the only kind of meaningful experience. These last
points, perhaps, are the central moral of Part 1 of this work.
266
Chapter 8. An Introduction to Programming-Physics
From Ehrlich, 1973. Reprinted with permission.
A new technological innovation—the computer—is steadily becoming as much of a fixture in
schools as chalkboards and pencils. While some tout the computer as a potential catalyst for a
revolution in education, others believe that this technological marvel will turn out to be just
another educational flash in the pan. Certainly, the boundless enthusiasm that sometimes seems to
prevail may need some tempering as we proceed in our exploration of the potential of computers
in instruction, and it would doubtless be profitable for more researchers to adopt stances of
skepticism. As has been pointed out by Larry Cuban (1986, 1993) and others, (e.g., D. K. Cohen,
1988) broad change in our educational institutions does not come easily.
Although computers may not be the key ingredient in an incipient educational revolution,
there are some simple and compelling reasons to believe that some sort of computation or
information technology will ultimately become a permanent fixture in classrooms. Central among
these reasons is the simple fact that computers are useful for a wide range of tasks. The point is
that we do not need to argue that computers possess some very general properties, like
“interactivity” or “dynamism,” that explain why computers should be especially useful for
educational purposes. We simply need to note that, like all good tools, computers make some
tasks easier and new tasks possible. Thus, for the very same reasons that computers have
established a beachhead in homes and offices, we can expect to continue to find them in schools.
267
Take the particular example of word processing software. Students should find this software
useful for many of the same reasons that business-people do.
I do not mean to imply that there are not unique instructional activities based around
computers—activities that would not be possible without computers. The existence of this new
tool with its specific capabilities does allow for some novel instructional applications. These
applications might even, in the long run, profoundly change how certain subjects are taught. The
second part of this dissertation, which begins with this chapter, is essentially an exploration of one
such novel application within the domain of physics. Here, we once again take up the discussion
begun in Chapter 1 and undertake to explore the possibility of improving physics instruction by
teaching students to program computers and to create certain kinds of physics-relevant programs.
A brief history of programming-physics
The proposal I will explore here, to teach physics through programming, is not new. To see its
roots, we must go back to a time prior to the use of computers in instruction. Even before the
wide appearance of computers in classrooms, some educators had realized the value of a related
technique, the use of numerical methods in physics instruction. For example, in his Lectures on
Physics, Richard Feynman used numerical methods as a key piece of his exposition on forces and
F=ma (Feynman, Leighton, and Sands, 1963).
As we progress, I hope to illustrate that the use of numerical methods is closely related to
programming-physics. In fact, many of the techniques and motivations underlying Feynman’s
numerical solutions also lie at the heart of programming-physics, with the important difference
that the availability of computers helps to make numerical methods more practical. For that
reason, it is worthwhile to begin this brief history with a little detail on Feynman’s exposition.
Feynman begins his discussion by asking this question: Suppose that we know the position and
velocity of an object at a time t, how can we find the position at a slightly later time t+ε? (The
symbol ε here stands for a small interval of time.) In answer, Feynman presents the following
expression:
x (t + ε ) = x (t ) + εv(t )
This equation says that, to find the position at the time t+ε, you add the quantity εv(t) to the
current position, where v(t) is the velocity at time t. This makes sense since, by definition, the
268
velocity of an object is the amount that its position changes per unit time, so εv(t) is the amount
that the position changes in the time ε. In a similar manner, we can write an expression for the
velocity at the time t+ε:
v(t + ε ) = v(t ) + εa(t )
Here a(t) is the acceleration at the time t. Note that both of these relations are only
approximations because they neglect the fact that the velocity and acceleration may change over
the time interval ε.
Furthermore, if we know the force on an object as a function of its position and
velocity—F(x, v)—then we are in a position to employ a very powerful and general procedure for
predicting the motion of an object. The procedure works like this:
(1) Start with the position and velocity at some time t: x(t), v(t).
(1) Using the function F(x, v) and the equation F=ma, find the acceleration at the time t: a(t).
(1) Substitute x(t) and v(t) into the above expression for the position to obtain x(t+ε).
(1) Substitute v(t) and a(t) into the above expression for the velocity to obtain v(t+ε).
(1) Go back to step (1) using x(t+ε) and v(t+ε) and iterate.
Using this method we can trace the motion of an object forward into the indefinite future. We
start with the position and velocity at any time and then we step forward, one moment at a time,
along the motion. Of course, if each iteration only takes us a very small step forward in time, then
the calculations involved may be laborious.
The beauty of this method is its extreme generality and simplicity. As long as you know how
to find the force on an object, you can use a=F/m to find the acceleration, and then plug in and
iterate to find the motion of the object. The computations may be tedious, but the method will
work for every motion.
The alternative approach—what students are usually taught to do—is to essentially write and
solve differential equations to obtain analytic, closed-form solutions. The problem is that this
traditional approach quickly runs into mathematical difficulties. Not only is the mathematics
involved difficult, the majority of differential equations cannot be solved analytically at all. For
that reason, various tricks, approximations, and alternative techniques must be developed, each of
269
which tends to be specific to only a small class of problems. Among this variety of techniques, the
underlying simplicity may be lost.
In his Lectures, Feynman went on to apply the numerical method in gory detail for two
motions, the oscillation of a mass on a spring, and the motion of a planet around the sun. Then,
following his last computation Feynman commented:
So, as we said, we began this chapter not knowing how to calculate even the motion of a mass on a
spring. Now, armed with the tremendous power of Newton’s laws, we can not only calculate such
simple motions but also, given only a machine to handle the arithmetic, even the tremendously
complex motions of the planets to as high a degree of precision as we wish! (p. 9-9 Volume 1)
As this quote suggests, the reason that Feynman spends time on these laborious computations is
that he is trying to convey to his students the fundamental power and generality of Newton’s
theory. One of the great virtues of numerical methods for instruction is that it makes this
generality and power fairly clear. As we shall see, this is one of the arguments that researchers
make in espousing the merits of programming-physics for instruction.
Nevertheless, in Feynman’s three-volume text, numerical methods are not to be seen outside
of this one exposition. There are certainly many reasons for this, not least of which is that
Feynman probably has some commitment to covering the material in its traditional form.
However, a major factor must also be the lack of a “machine to handle the arithmetic.” Although
computers were in use by researchers at the time of Feynman’s lectures, they were not available in
classrooms.
But computers were to appear on the scene shortly and, with them, the possibility of realizing
the benefits of numerical methods. In fact, it wasn’t long after the publication of Feynman’s
lectures that books appeared, some directed at physics teachers, that explained how to use the
teaching of programming in physics instruction (e.g., Bork, 1967; Ehrlich, 1973).The books were
written with the computers of the day in mind. They expected that students would be
programming in FORTRAN using punch cards, and they assumed that students would be seeing
output on a line-by-line printing device. This means that the user never looked at the sort of
computer display to which we have become accustomed; instead, all output was printed on a sheet
of paper that scrolled out of the top of a printer.
In fact, much cleverness and ingenuity was directed toward inventing methods for producing
interesting displays with these printing devices. The example shown in Figure Chapter 8. -1, which
is taken from Ehrlich (1973), shows the electric potential around two point charges. Notice that
the image is produced using only ‘+’ and ‘-’ symbols. (See also the image at the head of this
270
chapter.) This image is quite effective, given the limits of the medium. But, no matter how clever
these early pioneers were, it was not possible for them to produce dynamic displays on a line
printer; they were restricted to static images.
Figure Chapter 8. -1. The electric potential around two point charges printed using only '+' and '-' symbols.
(From Ehrlich, 1973. Reprinted with permission.)
The publishing of Mindstorms by Seymour Papert is an important landmark in this brief
history (Papert, 1980). In Mindstorms Papert described his vision for a child-centered classroom.
At the heart of this vision was a new programming language, Logo, which was designed specially
for use by students. Though Mindstorms does contain some physics-related examples, Papert was
not especially concerned with programming-physics. Nonetheless this is an important landmark
for this history because of Papert’s well-known espousal of student programming, generally, as a
useful ingredient in instruction.
It is also worth noting that Logo was riding the crest of what was then the newest computer
technology. Notably, Logo was designed to be used with video displays rather than printers for
output. In this regard, a central innovation of the Logo language was the ability to program “turtle
graphics.” In Logo, students can give programming instructions to a small triangular object,
known as the “turtle,” which can be made to move around the video display. Thus, in contrast to
the early days of printed output, students could produce dynamic displays in which objects
actually appeared to move on a computer screen.
In subsequent years, as computers have become more powerful and user-friendly, the move in
physics education has been toward uses other than student programming. Rather, applications
271
such as pre-made simulation environments, computer tutorials, and real-time data-collection tools
have come to dominate. (For a broad picture, see Redish and Risley, 1988.)
In some respects, this proliferation of alternative uses can be understood as a consequence of
the increased power and capabilities of computers. It is not surprising that the development of new
capabilities, such as the ability to produce dynamic output, would lead instructors to invent new
applications for computers in physics instruction. But I believe that this shift in application also has
to do with some changes in the prevailing attitude toward computers. As applications have
become increasingly powerful and elaborate, programming has become more and more the
province of expert programmers. In this age of large, powerful, and flashy applications, the notion
that students can easily program their own simulations may sound less plausible to the average
teacher.
But student programming in physics has not been abandoned through the 1980’s and 1990’s.
Materials are still being produced for teachers, though they are frequently targeted for auxiliary
courses in numerical techniques (e.g., Gould & Tobochnik, 1988). And similar approaches are
showing up in alternative forms, such as in applications of spreadsheets to physics instruction (e.g.,
Misner & Cooney, 1991).
A notable example of a currently active program is the M.U.P.P.E.T. project at the University
of Maryland, which is an ambitious attempt to integrate Pascal programming into the university
physics curriculum (Redish & Wilson, 1993; McDonald, Redish, & Wilson, 1988). Rather than
simply supplementing the standard curriculum with a few programming activities, Redish and
colleagues have set out to “rethink the curriculum entirely from the ground up.” One of their
main points is that limits in students’ mathematical abilities, as well as the inherent difficulty of
much of the mathematics involved in physics, has placed many constraints on the physics
curriculum: Topics must be ordered in a way that parallels the development of student
mathematical knowledge, and some topics and phenomena cannot be studied at all because the
mathematics involved is extremely difficult, even for expert physicists. By introducing numerical
methods implemented on a computer, many of these constraints are loosened. As Feynman
realized, the same, very general, numerical methods can be applied to study a very wide range of
physical phenomena. This allows the curriculum to be ordered in a more “natural” manner, and
the range of phenomena that can be studied is greatly broadened.
272
The current project
I am now ready to discuss where the present research fits in this brief history. The work I am
going to describe in the second part of this dissertation is unique because of its tight focus on the
representational language itself. Earlier research has been very concerned with what student
programming allows. As Redish and colleagues have argued, if your students can program, then
you can re-order the topics in the curriculum, and you can study topics and phenomena that were
formerly inaccessible to students. From this point of view, the primary importance of the
programming language is that it provides a means of bringing the power of the computer into
play; if you want to use the computer to do some laborious computations for you, then you have
no choice but to give the computer instructions using a programming language. In this view,
algebra retains its status as the primary representational system of physics, and programming plays
an auxiliary role as the language we use to communicate with computers.
In contrast, I am interested in studying programming languages as representational systems in
their own right. I want to elevate programming languages to the status of bona fide
representational systems for expressing physical laws and relations, and I want to study the
properties of these systems. This means adopting the stance that a programming language can
function much like algebraic notation in physics learning. I believe that this move constitutes the
next logical step in this history and, when we take this step, we will be moving one notch further
toward a true programming-physics.
It is at this point that we can make contact with the broader concerns of this project. One of
the main goals of this dissertation is to argue for a tight relationship between knowledge and
external representations. In Chapter 1 I made the point that, if this tight relation exists, then
replacing algebraic notation with a programming language may have a fundamental effect on how
students understand physics. With that argument, the stage was set for the overall task of this
research project: a comparison of algebra-physics and programming-physics knowledge.
The first segment of this project, a description of a certain component of algebra-physics
knowledge, was completed in Part 1 of this dissertation. Now, in Part 2, I will move on to the
complementary segment, the production of a description of programming-physics knowledge
that can be compared to the results for algebra-physics. The remainder of this chapter lays out
some of the preliminaries for this second segment of the project. I will begin with an introduction
to the programming language used in my study, and then I will provide a description of the
particular practice of programming-physics that I taught to my subjects.
273
An Introduction to Boxer Programming
All student programming in this study was done in a new and rather unique programming
environment known as “Boxer,” which employs its own programming language. The Boxer
programming environment and language are currently under development at U. C. Berkeley
(diSessa & Abelson, 1986 ; diSessa et al., 1991). In many ways, Boxer can be understood to be a
direct descendent of Logo, and it has been designed with the specific aim of correcting some of
the shortcomings of Logo while maintaining many of the same goals and strengths (diSessa,
1986). For example, as we shall see in a moment, Boxer involves a form of graphics programming
that is very similar to the turtle graphics found in Logo. But perhaps the most important feature
that Boxer shares with Logo is its overall philosophy: Unlike most programming languages, Boxer
is designed especially with education in mind and it is intended to be used by students of a wide
range of ages and abilities, from elementary school students through adults.
The primary reason that I chose Boxer for this study is that it is easy to use and learn. In my
pilot work I found that it was only necessary to give the students in this study a one-hour
introduction to Boxer programming before they could begin working on some physics-related
tasks. This is especially striking given the fact that many of the students had little prior
programming experience. This property is not only extremely useful for research studies. Easy
learnability is also essential for applications of programming to subject matter instruction, since we
do not want students to get bogged down in the details of programming when trying to learn
about, for example, physics and mathematics.
The Boxer programming environment gets its named because, in Boxer, everything on the
computer display appears inside a box. Programs appear in boxes called “do-it” boxes, graphics
appears in “graphics” boxes, and variables appear in “data” boxes. Figure Chapter 8. -2 shows a
simple Boxer graphics program, with a graphics box on the left and a do-it box on the right. The
triangular object inside the graphics box is called a “sprite.” Boxer’s sprites, like Logo’s Turtles, are
the vehicles of all graphics programming. When given programming commands, sprites move
around the computer display within the confines of a graphics box. The do-it box to the right of
the graphics box contains a short program that tells the sprite to go forward 50 turtle steps, then
turn right 45 degrees, and finally go forward another 35 turtle steps.
274
a-program
fd 50
rt 45
fd 35
Doit
Data
Figure Chapter 8. -2. A Boxer graphics program.
The program shown in Figure Chapter 8. -3 illustrates the use of variables. Variables take on a
particularly simple form in Boxer, they are just boxes whose contents can be changed by
programming instructions. The key programming command that is used for working with
variables is the “change” command. The program shown in Figure Chapter 8. -3 uses the change
command to change the contents of the variable named bob to 3, then to change the contents of
the variable hal to 5, and finally to change jane to the sum of bob and hal. Notice that, in the final
statement, a do-it box has been used to group the two variables to be added. Do-it boxes can be
used to group terms in a Boxer statement much as one would use parentheses in an algebraic
expression.
jane
bob
hal
8
3
5
Data
Data
Data
a-program
change bob 3
change hal 5
change jane
bob + hal
Doit
Doit
Figure Chapter 8. -3. Programming with variables in Boxer.
It is worth taking a moment to note the common syntax shared by the programming
commands used in these examples. To use most Boxer commands, we write a command name
followed by a list of arguments:
command-name <argument-1> <argument-2> …
275
For example, both the fd and rt commands each take a single argument, a number that specifies
the number of steps to go forward or the number of degrees to turn. In contrast, the change
command takes two arguments, one that specifies the variable to be changed and another that tells
Boxer what new value to insert in this variable.
draw-square
repeat 4
fd 50
rt 90
Doit
Doit
Data
Figure Chapter 8. -4. A program to draw a square.
We will make quite frequent use of one additional command, repeat, which provides a means
of repeating a set of programming instructions. The two arguments to the repeat command
specify the number of repeat cycles and the set of instructions to be repeated. Figure Chapter 8. -4
shows a program that uses the repeat command to draw a square by repeating, four times, the
commands fd 50 and rt 90.
Finally, my study made use of the fact that students can define and call their own procedures.
To create a new procedure in Boxer, we write a set of instructions in a do-it box and then give that
box a name. This is illustrated by the program in Figure Chapter 8. -5. The procedure named
draw-pattern uses the named procedure draw-square to create a pattern of four tiled squares.
draw-pattern
draw-square
repeat 4
repeat 4
draw-square
rt 90
Doit
Doit
fd 50
rt 90
Doit
Doit
Data
Figure Chapter 8. -5. A program that uses a named procedure to draw a pattern.
These four programs illustrate most of what the reader will need to know to understand the
examples that follow. We will see that the combination of simple sprite-graphics, variable
programming with the change command, and the use of repeat were sufficient for the students in
276
my study to do quite a lot of interesting physics programming. I will fill in additional details as
we proceed.
A Practice of Programming-Physics
I am now ready to introduce the reader to the practice of programming-physics that was
employed in this study. As I mentioned in Chapter 1, the particular practice that I used was
adapted from work done by the Boxer Research Group at UC Berkeley. In that research,
programming was used as part of physics courses that the Boxer Research Group designed for 6th
graders and high school students (diSessa, 1989; Sherin et al., 1993).
To illustrate this practice, I will present a few examples from the brief curriculum given to the
college students in this study. The programming-physics training was intermingled with the more
general instruction in Boxer programming. To start, the students were taught only how to create
the sort of sprite graphics programs shown in Figure Chapter 8. -2. Following this
instruction—and prior to any instruction in the use of variables—the students were asked to
perform their first physics-related tasks. In one of these first tasks, they were simply told to “make
a realistic simulation of the motion of a ball that is dropped from someone's hand.” In response,
they created programs like that shown in Figure Chapter 8. -6, which was made by two students
in this study, Anne and Clem.
drop
setup
fd 1 dot
fd 3 dot
fd 5 dot
fd 7 dot
fd 9 dot
fd 11 dot
fd 13 dot
fd 15 dot
fd 17 dot
fd 19 dot
fd 21 dot
fd 23 dot
fd 25 dot
fd 27 dot
fd 29 dot
fd 31 dot
fd 33 dot
fd 35 dot
fd 37 dot
fd 39 dot
fd 41 dot
Doit
Data
Figure Chapter 8. -6. Anne and Clem's simulation of a dropped ball without variables.
277
Figure Chapter 8. -6 shows two boxes, a graphics box containing a sprite—the sprite, in this
case, is shaped like a ball—and a do-it box named drop, which contains the program itself. Anne
and Clem’s program begins by executing a setup command, which positions the sprite at the top
of the graphics box and orients it so it is prepared to move downward. Each of the remaining lines
in the program contains a fd command followed by the dot command. The dot command causes
the sprite to draw a dot at its current location. Thus we see that this program causes the sprite to
move forward in larger and larger steps, leaving a trail of dots in its wake. The sprite moves
forward one step and then leaves a dot, then three steps, then five steps, etc.
It is important to note how this sort of program was understood by students. All of the pairs
tended to assume that each of the movements of the sprite takes the same amount of time, an
assumption that was reinforced by later aspects of the curriculum. Given this assumption, Anne
and Clem’s program suggests that the ball is speeding up. Furthermore, note that the amount that
the sprite steps forward increases by two each time that it moves. Thus, not only is the speed
increasing, it is increasing at a constant rate for each time interval. Anne and Clem chose this
constant increase to reflect the fact that, neglecting air resistance, a dropped ball undergoes a
constant acceleration.
a
b
c
Data
counter
drive
0.000
setup
change counter 20
repeat 23
fd counter
dot
Data
Doit
repeat 40
change counter
counter - .5
Doit
fd counter
dot
Doit
Doit
Figure Chapter 8. -7. Using variables to simulate constant deceleration.
Of course, creating simulations in this manner, using only a sequence of forward commands,
would be quite tedious for students and not very illuminating. So, after making only a few
programs along the lines of Figure Chapter 8. -6, the students were taught to use variables and to
apply them to create simulations. An early application of variables to simulation programming is
278
shown in Figure Chapter 8. -7. In this task, the students’ job was to write a program that causes
the car-shaped sprite to move from point (a) to point (b) at a constant speed, and then slow down
at a constant rate as it moves from point (b) to point (c).
In the student simulation shown in Figure Chapter 8. -7 this motion is achieved using a
variable named counter. The program, which is named drive, begins with a setup command,
which places the car at point (a) in preparation for the motion. Then, in the next line, a change
command is used to initialize the value of counter to 20. The remainder of the program consists
of two repeat commands, each of which causes a short program to repeat a fixed number of times.
The first repeat tells the sprite to, 23 times, move forward the amount in the counter variable and
make a dot. Since the value in counter doesn’t change during this repeat, the sprite moves forward
the same amount on each cycle of the repeat, leaving a trail of equally spaced dots. Thus, this first
repeat takes the car from point (a) to point (b) at a constant speed, as required by the task.
The second repeat statement is similar but with one important difference: In addition to a fd
and dot command, it also contains the line
change counter
counter - .5
Doit
which causes the value of counter to be decreased by .5 on each cycle of the repeat. Thus the car
steps forward in smaller and smaller increments as it moves from (b) to (c). In this way, the
students have simulated a constant deceleration, but without resort to a long string of forward
commands, as in Figure Chapter 8. -6.
The simple program in Figure Chapter 8. -7 embodies most of the key structures and
techniques needed to create a wide range of simulations. There is a variable, in this case named
counter, that corresponds to the velocity of the moving object. This variable is initialized at the
start of the program then, within a repeat statement, values are added or subtracted from this
variable to simulate an accelerated motion.
279
a
b
c
Data
pos
vel
acc
850.000
0.000
-0.500
Data
Data
Data
tick
change vel
change pos
drive
setup
change
change
change
repeat
pos 0
acc 0
vel 20
23 tick
vel + acc
Doit
change acc -.5
repeat 40 tick
pos + vel
Doit
Doit
fd vel
dot
Doit
Data
Figure Chapter 8. -8. A simulation using the Tick Model.
After the students were given a few opportunities to write their own simulations using
variables, they were taught a standardized version of this structure called the “Tick Model.” The
simulation in Figure Chapter 8. -7, rewritten to employ the Tick Model, is shown in Figure
Chapter 8. -8. This new version of the simulation uses three variables, pos, vel, and acc, which
correspond to the position, velocity and acceleration of the object. Furthermore, each cycle of a
repeat is handled by a separate procedure named tick. Each time that this procedure is executed,
the value of acc is added onto vel, vel is added onto pos, and the sprite is told to move forward an
amount equal to vel.
Comparing Figure Chapter 8. -8 with Figure Chapter 8. -7 we see that, when the Tick Model
is used, the contents of drive are somewhat simplified. This new version of drive begins by
initializing all of the variables, then repeats the tick procedure 23 times. Since the value of acc is
zero during this portion of the program, the sprite just steps forward at a constant speed. Then,
following these first 23 ticks, the acceleration is changed to -.5 and tick repeated 40 more times.
During each cycle of this second repeat statement the value in vel is thus decreased by .5, and the
sprite slows to a halt.
Once the Tick Model was introduced, students were directed to use it in all of their
simulations. To facilitate this, their programming environment was modified so that, when a
particular function key was pressed, the three variables and the tick procedure appeared on the
display as shown in Figure Chapter 8. -9.
280
pos
vel
acc
850.000
0.000
-0.500
Data
Data
Data
tick
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
Data
Figure Chapter 8. -9. The Tick Model.
The above examples should serve to capture the essence of the practice of programmingphysics used in this study. Now I want to take a moment to reflect on some of the noteworthy
features of this practice. To begin, notice that the goal of each task is the production of a
simulation of some particular motion. For example, one task in my study gave students only the
following instructions:
In this task, you must imagine that the block shown is attached to a spring so that it oscillates back and
forth across the surface of a frictionless table. Make a simulation of this motion.
As in this example, instructions were typically very brief; they usually just described a motion in
simple terms and then asked students to create a simulation of that motion. Note that it would be
possible to create a practice of programming-physics based on tasks with very different types of
goals. For example, it is possible to write programs to obtain answers to textbook-like questions.
In addition, the Tick Model forces a very specific structure onto students’ simulations: It
essentially uses the numerical method described by Feynman. Alternatively, for example, it is
possible to create simulations of motion with constant acceleration by using this equation, which
gives the position as a function of time:
S = So + vot + 12 at 2
A simulation based on this equation might work by employing a time variable that is constantly
incremented. On each cycle, the value of this time variable would be substituted into the above
position equation and then the sprite would simply be placed at the corresponding location. Such
an approach has the benefit of producing a simulation that is more accurate than that produced by
simulations based around the Tick Model. However, it relies on the existence of an analytic
281
solution in which to substitute values for the time. In contrast, like Feynman’s numerical approach,
the Tick Model is highly general and can easily be applied in cases in which it is not easy to obtain
an analytic solution. A more extensive comparison of the Tick Model to alternatives can be found
in Sherin (1992).
Sample Programming Episodes
In the previous section, I presented a number of examples of finished programs, without any
discussion of how those programs came to be written. Now I want to move into a discussion of
the construction process, and I will present some brief descriptions of programming episodes, taken
from the work of the university students in my study. In addition to giving the reader a feel for
what the construction process was like, these accounts will also provide an opportunity for me to
describe more of the tasks given to students, as well as further examples of the programs they
created.
As preamble, I will begin by attempting to evoke the setting in which these programming
episodes occurred. At any time, a single pair of students was in the laboratory, and the students
sat together at a single computer, usually with one student typing at the keyboard and the other
operating the mouse. As compared with the algebra-physics sessions, there was thus somewhat less
of a “performance” aspect to the sessions, since the students sat facing a computer monitor, rather
than standing in front of a whiteboard. Two cameras were positioned quite close to the students,
one behind and one to the side. During the sessions, I sat off to the side and operated the cameras
or simply observed.
Tim and Steve: Simulating the drop
I will open with an account of Tim and Steve’s work on what proved to be one of the least
difficult simulation tasks. Recall that, before they were even introduced to the Tick Model,
students were asked to program a simulation of a ball that is released from rest and simply drops
to the ground. In response, Tim and Steve, like all of the students in the study, wrote a program
that consisted only of a string of fd commands (refer to Figure Chapter 8. -6).
282
POS
VEL
ACC
-539.000
-98.000
-9.800
Data
Data
Data
TICK
change vel
DROP2
setup
change
change
change
repeat
pos 0
vel 0
acc -9.8
10 tick
Doit
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
Data
Data
Figure Chapter 8. -10. Tim and Steve's simulation of a dropped ball using the Tick Model. In this initial
version, the ball goes up instead of down.
Later in the study, all of the students were asked to redo this simulation using the Tick
Model. This turned out to be rather straightforward for all the pairs, including Tim and Steve,
who wrote the program shown Figure Chapter 8. -10 in only about 3 minutes. Here’s a
description of what they did in those 3 minutes: First, Tim placed the tick model box in their
working area on the display. (Recall that, to do this, all he had to do was press a single function
key.) Then, with virtually no comment or discussion, he typed the program shown in the box
named drop2. This program first executes setup and initializes the variable—pos and vel to zero
and acc to -9.8—then it repeats the tick procedure 10 times.
This program would have produced an acceptable simulation except for one hitch: It turns out
that when this simulation was run, the ball moved upward instead of downward. The reason is
that, in this case, it happens that the setup command positions the ball at the top of the screen
with the sprite facing downward.13 Presumably, Tim and Steve had assumed that the sprite was
facing upward and, since it was shaped like a ball in this task, there was no easy way for them to
determine the sprite’s orientation. However, once the simulation was run this problem was easy to
recognize and diagnose. Tim observed simply that “It went the wrong way,” and then proceeded
13
The behavior of the setup command varied across tasks.
283
to modify the program so that the acceleration was initialized to a positive value, rather than a
negative. With this change, the program performed appropriately.
POS
VEL
ACC
539.000
98.000
9.800
Data
Data
Data
TICK
change vel
DROP2
setup
change
change
change
repeat
pos 0
vel 0
acc 9.8
10 tick
Doit
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
Data
Data
Figure Chapter 8. -11. Tim and Steve's corrected simulation.
There were not many interesting twists and turns in the story of this episode; the task was
completed in short order, and with only a minor hitch. Nonetheless, there are a number of
important points to be made about Tim and Steve’s work on this simulation. First, it is
worthwhile to comment on the fact that this task seemed to be almost trivial, a fact that is largely
due to the power of the programming structures that had been taught to the students. Notice how
simple drop2 is; all it has to do is initialize the variables and then repeat the tick procedure the
desired number of times, a technique that Tim and Steve had seen in a number of example
programs. As we will see, this initialize-and-then-repeat structure appeared in almost every
simulation program.
A second point illustrated by this episode is the importance of feedback from the graphical
display. The first time that Tim and Steve ran their program, the ball moved in the wrong
direction, a clear indication that their program needed to be changed. In this case, the error
indicated was not of a very deep nature; it was due to an idiosyncrasy in how the setup procedure
had been defined for the task. In general, however, feedback from the graphical display played an
important and fundamental role in the development of simulations. In some cases this feedback
indicated that radical changes were required, while, in others, this feedback simply led students to
284
adjust some of the quantities in their programs in order to make their simulations more
aesthetically pleasing. For example, in writing their drop simulations, students usually adjusted
the value of the acceleration and the number of repeats so that the ball stopped near the bottom of
the graphics box.
Tim and Steve: Adding air resistance to the drop simulation
In this next example, we trace Tim and Steve’s work into a more advanced and somewhat
more difficult task. In this task, the students were asked to include air resistance in their
simulation of a dropped ball, the very simulation that we just discussed above. The instructions
given were brief, but they included a couple of important hints and suggestions:
(a) Make a new simulation of the DROP that includes air resistance. (Hint: Air resistance is bigger if
you're moving faster.)
(b) Try running your simulation for a very long time. (The ball might go off the bottom and come in
the top; that's okay.) If you've done your simulation correctly, the speed of the ball should eventually
reach a constant value. (The "terminal velocity.")
Recall that the algebra-physics subjects were also given a task that involved air resistance. In
that task, the students were essentially asked to find how the terminal velocity of an object
depends on its mass. This task was moderately difficult for students since, in general,
introductory physics instruction does not take up the topic of air resistance very often; students
usually solve one or two problems pertaining to air resistance in their early physics careers. For this
reason, the students in my algebra-physics study needed to invent their own expressions to
account for the resistive effect of the air. Most commonly, they invented a relation for the force
due to air resistance, such as:
Fair = kv
The programming task placed students in a similar predicament. On a sheet of paper, Tim
and Steve drew an object with an arrow acting down from the bottom and an arrow acting up
from the top (refer to Figure Chapter 8. -12). They knew that the force acting down was equal to
mg—the force of gravity—but, like the algebra-physics students, they did not know immediately
how to find the force of air resistance. In the following discussion, this issue is worked out:
285
W=mg
Figure Chapter 8. -12. Tim and Steve's paper diagram for the air resistance task.
Tim
…
Tim
Steve
Tim
Steve
Tim
Steve
Tim
Steve
Anyway. Well, when you first start out there’s no air resistance. I mean there’s no - but as
you go faster the little v- little vector here starts growing. [g. arrow at top of diagram].
So we have to have the size, is bigger if you’re moving faster. (4.0) (…)
I don’t think we really have to worry about that. It just decreases.
The faster you’re moving the bigger the - so what is this. So this is - this is something
times velocity.
Right.
Right?
Yeah, okay. So it’s a function of velocity, right?
Yeah.
Okay. So what should we do? Just make it like one times - point one times V or
something?
The above discussion is very similar to discussions in which the algebra-physics students
constructed equations for the force of air resistance. Like their predecessors in this study, Tim and
Steve came to the conclusion that the force of air resistance should be a constant times the velocity
of the falling object. With this decided, they created a variable named res and modified their
program to compute a value for this variable:
drop2
setup
change
change
change
repeat
Doit
pos 0
vel 0
acc 9.8
10 tick
POS
VEL
ACC
0
0
9.8
Data
Data
Data
RES
Data
TICK
change Res
vel * .1
Doit
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
Data
At this point I need to pause and note that Tim and Steve have just done something that we
have not yet seen: They have modified the content of the tick procedure. In contrast, in all of the
previous examples, the students did all of their programming in a separate procedure, which then
286
called tick. This approach will not work here because lines added to the drop procedure above the
repeat command only affect the initial conditions of the simulation; they cannot dictate changes
that must occur during the time of the motion. By putting their new statement inside of the tick
procedure, Tim and Steve have constructed their simulation so that the value of res will be
recomputed at each moment of the motion.
Since this issue will play a fairly important role in future discussions, I want to say a little more
about how this simulation differs from the examples we have already examined. In the previous
examples that used the Tick Model, the value of the acceleration was constant throughout the
motion. However, in the present task and some others that we will see, the acceleration varies over
the time of the motion. Tim and Steve were quite clearly aware of this issue. They knew that the
effect of air resistance would grow, causing the acceleration of the object to decrease through time
until it was zero.
Steve
We have to alter the acceleration, huh. Like decrease it in- with time?
…
Tim
Yeah, basically it’s um, as this gets bigger [g. arrow at top of diagram], the acce - the
value of eight- nine point eight gets smaller. Until it reaches zero and there’s no
acceleration at all.
In its present state, Tim and Steve’ program computes a value for the variable res, but this
variable does not affect any other aspects of the simulation. Thus, the students still needed to
modify the tick procedure so that the size of the air resistance was somehow taken into account in
the acceleration. Tim and Steve were aware of this, and they knew that they wanted a statement
that began with “change acc … ” inside the tick procedure. However, as they set out to write this
line, they realized that it wasn’t clear to them exactly where to place it among the other lines in
tick:
Tim
No we have to (then account) in the acceleration. So we have to change the acceleration.
Do we change the acceleration before we change the velocity? [g. change vel line in tick]
Or do we change acceleration after we change velocity? Which also changes the
resistance. [g. change Res line] Oh, I’m getting confused. What’s the order in which you
change things?
(3.0)
Tim
You see what I’m sayin’?
Steve
Uh:::.
Tim
They all like change each other. (Around).
[They laugh.]
Tim
So it matters which order you do it in, right?
Steve
I think the velocity’s first, right?
Tim
I don’t know.
Steve
Cause the first velocity is-
287
Tim
no air resistance.
Steve
Yeah, and then it’s just this acceleration.
Tim
Is that true?
Steve
Well, not really, huh. (1.0) I don’t know, I’m confused too.
(3.0)
Tim
Well you have to change resistance because you change the velocity, because (0.5) the
velocity (1.0) - the new velocity is dependent upon the acceleration which is dependent
upon the resistance. So we should change resistance, then acceleration, then velocity.
Right? Does that make sense, or am I just talking nonsense.
In the above discussion, Tim and Steve decided on a specific ordering for the various lines that
must appear in the tick procedure: res would be changed first, then acc, then vel, then pos. They
next modified tick to reflect this decision. This required only the addition of a single new line to
compute the value of the acceleration.
TICK
change Res
vel * .1
Doit
change acc
acc - res
Doit
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
There are a couple of important points to be made concerning the line that Tim and Steve
have added to compute the acceleration. First, it appears that they are treating res like an
acceleration, rather than a force. This may be intentional, but it is more likely that they have
simply not considered the issue carefully. In fact, students in the algebra-physics portion of my
study were often ambivalent as to whether expressions they had written for “air resistance” should
be treated as a force or an acceleration. I will say more about this in a moment.
Second, there is a major problem with the statement that Tim and Steve have just written. A
correct expression for the acceleration here would subtract some type of air resistance term from a
term associated with gravity:
change acc
grav - res
Doit
Furthermore, the term associated with gravity should remain constant throughout the simulation
since gravity, too a good approximation, applies a constant force to a falling object. It is possible
288
that this is what Tim and Steve had in mind; but rather then having a constant gravity term, they
have used the variable named acc, and this variable changes as the simulation runs.
It turns out that this substitution causes some striking behavior. When Tim and Steve ran their
program, rather than just moving downward, the ball oscillated up and down. The students found
this quite amusing.
Steve
Oh my god. Oh no:::.
[They both laugh.]
Tim
[Runs the program again.]
Steve
But when you drop balls it’s not gonna do that though dude!
It took quite a while for Tim and Steve to correct this problem and ultimately get a working
simulation. In the process they considered many alternatives, including the possibility that it
would be necessary for the program to explicitly check for a condition which signals that terminal
velocity has been reached. Eventually, with a little help from me, they modified their acceleration
line to be:
change acc
mass * 9.8 - res
/ mass
Doit
Doit
If rendered in algebraic notation, this statement might be written as:
a=
mg − Fair
m
When this modification was made, the program appeared as shown in Figure Chapter 8. -13.
Tim and Steve ran this program, and were quite pleased and excited by the results. They could
tell that their program was working properly by watching how the values of certain variables
changed during the simulation. In particular, they noticed that the acceleration died away to zero
and that the velocity leveled off at a constant value, thus indicating that a constant velocity was
reached.
289
drop2
setup
change
change
change
repeat
pos 0
vel 0
acc 9.8
1090 tick
Doit
POS
VEL
ACC
RES
MASS
450.800
19.600
0.0000
98.000
10
Data
Data
Data
Data
Data
TICK
change Res
vel * 5
Doit
change acc
mass * 9.8 - res
/ mass
Doit
Doit
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
redisplay
dot
Doit
Data
Data
Figure Chapter 8. -13. Tim and Steve's final air resistance program.
Finally, when Tim and Steve told me that they were satisfied with their program, I asked
them to explain why their simulation reaches a terminal velocity. Tim explained:
change Res
vel * 5
Doit
change acc
mass * 9.8 - res
/ mass
Doit
Doit
Tim
Well it reaches terminal velocity when the resistance is equal to the weight. And, the
weight’s constant. (6.0) We - we had it multiplies by the acceleration [g. mass * 9.8] which
is the change in velocity which is a different acceleration than the one you use. (…)
(6.0)
Tim
Well, basically, as this - this [g. change res line] eventually gets close to this number [g.
mass * 9.8] and the difference [g. [mass * 9.8 - res]] becomes (next to nothing.) And once
that happens there’s no more change in velocity. Velocity becomes constant.
…
Tim
Well, whatever force vector’s going down, that um, cause it to, um, (3.0) accelerate. And,
as you - when you accelerate, the resistance gets bigger. So, it kinda like - so, every time
you accelerate a little more, the resistance gets a little bigger.
Bruce
Hm-mm.
Tim
So eventually, um, this grows and this grows and eventually it’ll reach a point where
they’re equal.
Tim’s explanation in the above passage is based around only two lines in the program, the two
lines they have added to the tick procedure to compute the resistance and acceleration. What he
tells us is that, as the ball falls, the value in the resistance variable grows. Furthermore, as it grows,
the value computed for the acceleration becomes less and less until, eventually, the resistance and
290
gravity terms cancel and the acceleration is zero. “So eventually, um, this grows and this grows
and eventually it’ll reach a point where they’re equal.”
With Tim’s above explanation, this episode wound down to a conclusion. Before wrapping up
this chapter, I want to highlight a few of the more important points to be made concerning Tim
and Steve’s work in this episode. This will help to prepare the ground for the more thorough and
principled analyses in upcoming chapters.
The power of the Tick structure
To begin, let’s look at the structure of Tim and Steve’s simulation program. First, as in the
Drop simulation, the initialize-and-then-repeat structure is evident in this program. Furthermore,
in this episode we also learned that any quantities that change through the time of the simulation
must be recomputed inside the tick procedure. In particular, in cases in which the acceleration is
not constant, a new value of the acceleration must be computed on each cycle of the repeat.
program
setup
change pos #
change vel #
repeat ## tick
Doit
POS
VEL
ACC
450.800
19.600
0.0000
Data
Data
Data
TICK
change force ??
change acc
force / mass
Doit
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
Data
Figure Chapter 8. -14. A programming structure for handling cases of non-constant acceleration.
Now that we know how to deal with cases of non-constant acceleration, the power of this
technique becomes evident. Using the basic structure embodied in Tim and Steve’s simulation
and generalized in Figure Chapter 8. -14, we can create a wide range of simulations. The template
shown in this figure essentially uses the initialize-and-then-repeat structure: The position and
velocity are initialized, and then tick is repeated some number of times. In addition, two key lines
have been added at the top of the tick procedure; the force is computed and then substituted to
find the acceleration. Thus, if we know the force on an object and the initial values of position and
velocity, we can simulate its motion.
291
This is precisely the power and generality that appealed to Feynman and Redish. Compare the
program produced by Tim and Steve with what is possible in algebra-physics. While some tasks
are comparatively easy for students in algebra-physics, such as computing the terminal velocity,
others are considerably more difficult. For example, it is harder to find how much time the ball
takes to reach terminal velocity, or how far the ball falls in a given time. To answer these questions
one must solve a differential equation, a task that ranges from difficult to extremely difficult,
depending on the precise function used for the force of air resistance, and the mathematical
expertise of the student.
In contrast, writing a simulation that includes air resistance is not that much harder than
simulating a drop without air resistance. Tim and Steve’s simulation only contains two additional
lines of programming. Furthermore, this simulation can be used to answer even those questions
that are so difficult in algebra-physics. For example, the distance that the ball falls in a given time
can simply be read out of the pos variable, as long as one knows how to deal with unit
considerations. Though the result is only approximate, it can easily be refined to any desired
accuracy, with only the time available as a limit.
Some similarities to algebra-physics
Another noteworthy feature of this episode is the fact that it contains construction and
interpretation incidents that are strongly reminiscent of those we encountered in algebra-physics.
The most obvious of these is Tim and Steve’s construction of an expression for air resistance.
Given the similarity of this construction incident to incidents in algebra-physics, it seems likely
that we will be able to explain this event much as we explained algebra-physics constructions. In
addition, notice that Tim’s final interpretation of the program is also reminiscent of PHYSICAL
CHANGE
interpretations from algebra-physics.
However, there were also cases in which similarities to algebra were actually the source of
problems for students. In fact, I will argue later that the difficulty that Tim and Steve had in
correcting their expression for acceleration was, in part, due to confusion traceable to overlap with
algebra-physics. As I will attempt to show, the relationship between programming statements and
algebraic expressions may sometimes be a little subtle.
Programming is tyrannical in enforcing ordering
Finally, recall that Tim and Steve essentially discussed what order the various quantities—res,
vel, and pos—should be changed. In programming, this seems like a very sensible question, and
292
answering this question is essential to the completion of the task. However, as I will argue later,
not only is it not necessary to answer this question in algebra-physics, the question hardly even
makes sense. The general point here is that programming is tyrannical in forcing an ordering on
how we see the physical world. One of the most important tasks of the succeeding chapters will be
to explore the consequences of this enforced ordering.
A Preview of the Comparison
In the preceding sections, we took our first steps into programming-physics. The episodes
presented were designed to prepare the ground for the discussions and examples that follow, and
to start us toward working through some of the noteworthy features of programming-physics.
Ultimately, my goal is to complete the project, laid out in Chapter 1, of describing how the
understanding associated with programming-physics would differ from physics-as-understood-by
physicists.
This task—which constitutes the major objective of this research—will be taken up in the next
three chapters. I will begin in Chapter 9 by describing how the Theory of Forms and Devices can
be extended to describe programming-physics. This will be followed in Chapter 10 by the results
of a systematic analysis of the programming-physics data corpus in terms of forms and devices.
Finally, the analyses in Chapters 9 and 10 , the work of Part 1, and a few additional ingredients
will be drawn together to produce a comparison of programming and algebra-physics in Chapter
11.
To conclude the present chapter, I will present a brief overview of this comparison, and a little
of what I hope to capture.
The comparison based on the theory of forms and devices
The Theory of Forms and Devices will provide the major basis of my comparison and most of
the work of the succeeding chapters be will aimed in that direction. The first step in this portion
of the comparison will be to extend the theory to programming-physics. This will involve working
through whether and in what manner the known forms and devices can be applied, and also
determining where new forms and devices must be posited. As I hinted in the last episode, there
will be some fundamental similarities to build upon, and many of the forms and devices seen in
algebra-physics will have similar counterparts in programming-physics. Nonetheless, we will see
that programming-physics requires the use of some distinct forms and devices, and it does not
293
draw on some of those that appeared in algebra-physics. Finding these differences will constitute
the core of my comparison of the two practices. In addition, as for algebra-physics, I will attempt
to measure the frequency with which these various knowledge elements appeared in
programming-physics.
I do not want to give away too much more of the results at this point, but I do want to at least
bring the reader up to my own jumping off point for this endeavor. The Theory of Forms and
Devices was, in part, designed to capture two informal conjectures concerning important
differences between programming-physics and algebra-physics. Ultimately, the theory allowed
these conjectures to be stated more formally, but here I will just briefly present these conjectures
in their original, informal terms.
Programming-physics privileges a different intuitive vocabulary. The first of these informal
conjectures is based in the notion that different symbol systems allow different “intuitive
conceptions” to be more or less explicitly represented, and that those that are directly supported
become privileged in the associated practice of physics. Because these are the intuitive conceptions
that can be represented within the symbolic practice, we must funnel our work through these
privileged conceptions in order to write new expressions. And when we look at existing
expressions, these intuitive conceptions are what we will see. The upshot of this is a more
prominent role for the privileged intuitions within physics understanding. My analysis in terms of
symbolic forms is, in part, an attempt to capture this “informal conjecture.”
This first informal conjecture also included some specific guesses concerning what sort of
intuitive conceptions would be supported by each of the symbolic practices. Some of these guesses
were based on the observation that, in contrast to algebraic statements, programming statements
are directed and rigidly ordered. In describing programming statements as “directed,” I am trying
to capture the fact that many programming statements force an asymmetry on the programmer.
To illustrate, compare the following algebra and programming statements:
F 1 = F2
change force1 force2
The algebra statement specifies a symmetric relation; it says that F2 is equal to F1, just as much as
it says that F1 is equal to F2. In contrast, the programming statement is an asymmetric, directed
relation; it says to put the value of force2 into force1.
294
Furthermore, every line in a program must be ordered. There is substantial flexibility in how
algebraic relations can be arranged on a sheet of paper but, as Tim and Steve discovered,
programming enforces a strict ordering on the statements that appear.
Because of the directed and ordered properties of programming, I speculated that
programming would support some intuitions that are not well supported by algebraic notation.
For example, since causal intuitions require an ordering of the world’s influences and effects, I
hypothesized that these intuitions might be more strongly drawn upon in programming-physics.
Furthermore, I conjectured that intuitions based in symmetric relations, such as those relating to
balance and equilibrium, would appear less in programming physics. In the next chapters, these
informal conjectures will be made more precise and we will see the degree to which they are born
out.
Programs are easier to interpret than equations. The second informal conjecture that my analysis
attempted to capture was the notion that programs might be easier to interpret than equations.
This conjecture derives from the observation that programs can be “mentally run” to generate
motion phenomena (Sherin et al., 1993). The idea is that a programmer can do what a computer
does and step, line-by-line, through a program, a procedure that appears relatively
straightforward. In contrast, when I began this work, I was not aware of any obvious candidates
for a universal technique that allows the generation of an associated motion phenomenon from an
equation.
In the chapters that follow, we will investigate how easy or difficult it really was for students to
interpret the sort of programs that appeared in this study. As we will see, the analysis of the
devices involved in programming-physics can go toward answering this question. I will argue that,
while programming does involve some new devices that can be roughly associated with the mental
tracing of a program, the interpretation of a program is still a complex business, and it draws on
many of the same representational devices encountered in the analysis of algebra-physics.
Other fundamental differences
The form/device analysis, described just above, is suited to drawing out a certain class of
differences between programming-physics and algebra-physics. This class of differences relates to
how people interact with and see meaning in the symbolic expressions that are employed in each
practice. Of course, there are many profound differences that will not be captured by this analysis,
and many alternative ways of talking about the differences that I will uncover. For example, in
295
working through the form/device analysis, there will be no place in which it is clearly appropriate
to comment on the simple fact that what a student does in each practice is very different. In
algebra-physics, a student manipulates equations to solve problems and derive new equations. In
contrast, in the practice of programming-physics employed in this study, a student writes
simulation programs. Furthermore, these differences in what one actually does in each practice are
tied to some fundamental differences between programming languages and algebraic notation.
Here I will mention just two very important differences:
Algebra has rules of manipulation. Algebraic statements can be manipulated using rules that act on
the symbols in the expressions to produce new, correct statements. For example, given an
equation, we can manipulate the equation to solve for an unknown. In contrast, programming
languages do not come equipped with rules for manipulating the symbols in a program to produce
new programs.
Programs can be run on computers. Programming statements can be typed into a computer and
then run to produce various types of output, including dynamic displays of motion that are
generated by the program.
I will, to a minor extent, comment on these sorts of differences when, in Chapter 11, I
compare programming-physics and algebra-physics. Many of the related issues will be collected
under the rubric of what I call “channels of feedback.” That discussion will be treated as largely
separate from my analysis in terms of forms and devices, but, in the process, I will show how both
sets of issues can be fit within a single, broader view.
296
Chapter 9. The Construction and Interpretation of
Simulation Programs
In Part 1 of this document, we explored the structured and meaningful world of equation use in
physics. I argued that, for a physicist and physics student, a sheet of equations is not just a
collection of individual symbols, manipulated according to a set of mechanical rules. Rather, I
attempted to show that initiates in physics can see certain types of meaningful structure in these
symbolic displays, and that they possess a repertoire of interpretive strategies. My account of these
capabilities took the form of the positing of two specific types of knowledge, what I called
symbolic forms and representational devices.
The purpose of this chapter is to perform a similar analysis for programming-physics. As for a
page of equations, we want to know what sorts of structure an initiate in programming-physics
sees in a program that appears on a computer display or that is written on a sheet of paper. This
means adapting and applying the Theory of Forms and Devices to programming-physics. We will
see that, with no change to its basic outline, the theory can be applied directly to programming; in
fact, it turns out that there are even counterparts for many of the specific forms and devices that
we encountered in algebra-physics. However, there are some notable absences and some
programming-specific forms and devices for which there are no similar counterparts in algebraphysics. Our job here will be to see where and how similar forms and devices appear in
programming-physics, and where new and somewhat different elements must be posited.
Of course, in moving from algebra to programming we are setting out on less sturdy ground.
Recall that the ability to see meaning in algebraic expressions is learned. Algebraic expressions do
not radiate meaning in a way that is obvious for any person that looks at them, and physics
students are certainly not born with the ability to recognize the type of structure that physicists see
in equations. Rather, equations were meaningful for the students in my study because they had
spent at least two semesters working with physics equations, listening to physics instructors talk
about these equations, and talking to other students with and about equations.
The situation in this examination of programming-physics is quite different. We do not have a
population of students that have spent several semesters in initiation into a practice. Instead, the
programming-physics students had only a very brief introduction to the programming of
simulations—between one and two hours. This observation has important implications for what
we can expect from this analysis and how we must understand the results. First, because their
297
experience is limited, the programming-physics students will have less of an ability to see structure
in a program. Second, more than in the examination of algebra-physics, I will be limited in how
much I can claim to be describing a stable and uniform practice. Because their shared background
is limited, at least with regard to simulation programming, there may be more variation across
students. Furthermore, the behavior of individual students may be less stable through the time of
the study since I am observing them during the early stages of their learning of programmingphysics.
Nonetheless, for several reasons, the situation is not as problematic as one might think. As we
shall see, it is possible for students to borrow heavily from their experience in algebra-physics when
understanding simulation programs. This adds greatly to the possible structure that students can
see in programs. Furthermore, it is not the case that students received no instruction. All of the
students were taught to use a set of simple but very powerful programming structures, and ways
of understanding these structures were implied in the instruction. The result is that the students in
this study were actually able to see quite a lot of structure in the programs that they wrote.
On Line-Spanning Structure
We are now ready to begin the project of applying the Theory of Forms and Devices to
programming. The first step will be to identify the places where programs are constructed and
interpreted much like algebraic expressions, and to identify where construction and interpretation
are very different in programming. When this is done, we can focus our attention on the places of
divergence. In fact, I will argue that there is both substantial overlap and interesting divergence.
To begin this discussion, I need to first take a moment to reflect on the types of meaningful
structure that we observed in algebra-physics. Although this point was not emphasized, my
description of algebra-physics always focused on individual equations or portions of individual
equations. I talked about how a given (single) equation was constructed or how a given equation
was interpreted. This implicitly presumed that, for my subjects, there were not symbolic forms
that span equations—a position that I arrived at after some analysis of my data corpus.
Now, in turning to programming, we are faced with a similar issue of what level to focus our
attention for analysis. One way to approach this issue is to ask: What gets interpreted? Is an entire
program always interpreted? Are individual lines of programming interpreted? Or are there more
arbitrary portions of programs that are interpreted?
298
In programming, I will indeed claim that there are forms that span lines of programming.
Then, what I hope to show is that it is in within-line structure that a great overlap between
programming and algebra appears. And the locus of substantial divergence will be the linespanning structure in programming.
The claim that there are no line-spanning symbolic forms in algebra-physics (as it appeared in
my data corpus) is a very significant claim, and one that I believe readers may find controversial.
The primary argument I will make in favor of this position is that, although students certainly do
associate meaning with the arrangement of algebraic expressions—and thus with line-spanning
structure—this is not the same sort of meaningful structure that is defined to be involved with
symbolic forms. Thus, my argument comes down to a clarification of what counts as a symbolic
form. I will explain just a little further here.
When working on an algebra-physics task, students often juxtapose multiple algebraic
expressions, and these statements are usually arranged in some manner. For example, it is
common to arrange statements so that they reflect the sequence in which the statements were
derived. Furthermore, it is certainly the case that this ordering is registered by students, in the
technical sense I introduced in Chapter 2. This is a plausible assertion because the ordering of
statements plays a role in the solution process. In addition, because it reflects the solution process,
the arrangement of statements is meaningful, in a certain sense, for students.
However, this is not necessarily the sort of meaningful structure that is associated with
symbolic forms. In forms, there are conceptual schemata that include physics-related entities (e.g.,
physical quantities) and relations among these entities. These schemata are then connected to
symbol patterns, which parse an arrangement of symbols in some manner. In this way, we can
conceptualize a physical circumstance in terms of the conceptual schemata, and then reflect this
conceptualization in an arrangement of symbols.
Though the arrangement of algebraic equations did carry meaning for the students in my
study, it was meaning of a different sort. For example, in order to explain how they solved a
problem, it was common for students to follow the lines they had written, explaining what they
had done in each step. Thus, the arrangement of lines can reflect the relationship between
components of a solution process—or even the relationships among components of an argument
or explanation. Although these are meaningful relationships, they are not relationships among
physics relevant entities.
299
It is important to note that it need not have been the case that this was the only type of
meaningful line-spanning structure used by students. In fact, some of the specific line-spanning
symbolic forms that I will identify in programming could have had counterparts in my algebraphysics corpus. When I describe these programming forms, I will comment on where and whether
they could have had algebraic counterparts. However, when I do describe a hypothetical algebraic
counterpart, I think it will be clear to readers familiar with algebra-physics that the associated
behavior is somewhat unusual, and not typical of algebra-physics. Thus, I hope that looking at the
line-spanning forms in programming, and contrasting them with what is done in algebra, will help
to clarify the claim that there are no line-spanning forms in algebra-physics, as it is traditionally
practiced.
One important clue that there is really something different going on in the programmingphysics corpus will be some episodes in which students have long and explicit discussions
concerning how the lines in a program should be ordered. Furthermore, in these episodes,
students do not make their decisions based solely on the behavior of the simulation; they choose
the ordering to reflect a conceptualization of how things are in the world. In contrast, I did not see
extended discussions about how algebraic statements should be ordered. Instead, to the extent
that there was an ordering, this usually just fell out as part of the solution process.
It is not hard to make guesses about the origins of this difference. As I discussed at the end of
Chapter 8, one important difference between programming and algebraic notation is that
programming is rigid in enforcing an ordering on the lines that are written. Although people do
tend to arrange algebraic expressions in certain traditional and useful ways, programming is more
strict and absolute in requiring that statements be ordered, and that ordering is meaningful in one
particular way. Furthermore, there are even ways of arranging statements into groups and giving
those groups names.
Again, I believe that this is a controversial point. But I do hope that skeptical readers will
withhold final judgment until they see the type of meaningful structure that students associated
with programs. In discussing the forms and devices of programming in the upcoming chapters, I
will add to this argument as appropriate.
Where Programs Can Be Interpreted Like Equations
The point of the above discussion was to warn the reader that, in programming, line-spanning
structure is going to be important. While the entirety of a program or procedure is rarely
300
interpreted at once, portions of a program containing multiple lines are often interpreted, and
forms can span these lines. As I will try to show, there are a variety of ways that portions of a
program are selected for interpretation, including groups of lines as well as individual statements.
However, this said, it does turn out that one of the most common sub-units of a program that
students interpret is an individual line of programming. Furthermore, it is here that the
interpretation of programs often resembles the interpretation of algebraic expressions. In this
section, I want to spend some time on the construction and interpretation of individual lines of
programming; in particular, I want to look at a range of cases in which single programming
statements are treated in a manner very similar to algebraic equations.
We already encountered some relevant examples when, in Chapter 8, we examined Tim and
Steve’s construction of a simulation of a dropped ball with air resistance. Recall that, using their
original simulation of a dropped ball as a starting point, Tim and Steve began by adding a single
line to compute the value of a resistance variable:
change Res
vel * .1
Doit
When writing this line they commented:
Tim
The faster you’re moving the bigger the - so what is this. So this is - this is something
times velocity.
Steve
Right.
Tim
Right?
Steve
Yeah, okay. So it’s a function of velocity, right?
Tim
Yeah.
Steve
Okay. So what should we do? Just make it like one times - point one times V or
something?
The point here is that this single line of programming was constructed and interpreted in a
manner that strongly resembles the construction and interpretation of similar expressions in
algebra-physics. We encountered the following examples in Chapter 3:
R = µv
Mark
So this has to depend on velocity. [g. R] That's all I'm saying. Your resistance - the
resistor force depends on the velocity of the object. The higher the velocity the bigger the
resistance.
_______________
Fair = kv
Bob
Okay, and it gets f-, it gets greater as the velocity increases because it's hitting more
atoms (0.5) of air.
301
The account of Tim and Steve’s programming statement in terms of forms and devices closely
parallels my earlier accounts of these algebra-physics constructions. Let’s begin with the role
played by the DEPENDENCE form in these construction efforts. Recall that the dependence form is
associated with a symbol pattern that specifies only that a given symbol—in this case, the
velocity—appears somewhere in an expression.
DEPENDENCE
[…x…]
In his above statement Mark says “this has to depend on the velocity,” evidence that the
DEPENDENCE
form is engaged here. There is also evidence for DEPENDENCE in one of Steve’s
statements above; he says that the resistance is “a function of” velocity. As noted in Appendix D,
terminology that employs the phrase “function of” is strong evidence for the
DEPENDENCE
form.
The PROP+ form is also a key element in the construction of the algebraic expression for the
force of air resistance.
PROP+
…x… 
 … 
Evidence for PROP+ can be found in both Mark and Bob’s statements. Mark says “The higher the
velocity the bigger the resistance,” and Bob says “it gets greater as the velocity increases.”
Similarly, in constructing the programming version of this expression, Tim said “the faster you’re
moving the bigger the [resistance].”
Two additional forms are also involved in each of these constructions. First, the IDENTITY form
is implicated since these expressions are all commonly understood as specifying how to compute
the value of a single unknown quantity.
IDENTITY
x = […]
And COEFFICIENT is implicated in the writing of the factor that multiples the velocity in each of
these expressions.
COEFFICIENT
[x
]
There is actually a minor difference here between the programming and algebra cases that
bears mentioning. Notice that, in the programming case, Tim and Steve wrote a specific
numerical value for the coefficient, rather than a symbol like ‘C’ or ‘K.’ This turns out to be
indicative of a rather consistent difference between programming and algebra-physics. As I will
302
discuss later, it was somewhat more common for the programming students to work with specific
values for quantities, rather than with those quantities represented by letters.
Finally, the representational device involved in Tim and Steve’s programming construction is a
straightforward counterpart of the device in the algebra-physics constructions. Note that each of
the above statements describes a change through the time of the motion. For example, Bob says
“it gets greater as the velocity increases.” Thus we can say that the PHYSICAL CHANGE device is
involved.
The existence of this particular overlap between programming-physics and algebra-physics is
not too surprising. There are lines of programming that are straightforwardly related to algebraic
expressions, and it is very plausible that these lines would be constructed and interpreted in a
manner very similar to that of algebra-physics expressions. These observations are not surprising
particularly given the fact that my student subjects had significant experience in algebra-physics
upon entry in the study.
DROP2
setup
change
change
change
repeat
Doit
pos 0
vel 0
acc 9.8
1090 tick
POS
VEL
ACC
RES
MASS
450.800
19.600
0.0000
98.000
10
Data
Data
Data
Data
Data
TICK
change Res
vel * 5
Doit
change acc
mass * 9.8 - res
/ mass
Doit
Doit
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
Data
Data
Figure Chapter 9. -1. Tim and Steve's simulation of a dropped ball with air resistance.
This sort of adaptation of algebra-physics knowledge can account for a surprising percentage
of the interesting work in programming-physics. For illustration, consider the final version of Tim
and Steve’s air resistance simulation, which is reproduced in Figure Chapter 9. -1. As I mentioned
in Chapter 8, this program is nearly identical to their simulation of a dropped ball without air
resistance, with the primary difference being the two added lines at the top of the tick procedure.
The first of these lines is the statement that computes the value of the resistance force, which I
303
discussed just above. The second new line uses this resistance to compute the acceleration at each
moment of the drop.
change acc
mass * 9.8 - res
/ mass
Doit
Doit
This second line is also analogous to some algebraic expressions that I discussed in Part 1. Recall,
for example, the extended discussion in Chapter 3 of Mike and Karl’s
MOMENT
COMPETING TERMS/GENERIC
expression for the acceleration of a dropped ball under the influence of gravity and air
resistance:
a(t ) = −g +
f (v)
m
Clearly, there is a great similarity between this equation and Tim and Steve’s programming
statement.
Thus, it seems that some of the core elements of Tim and Steve’s work on this fairly difficult
simulation task can be accounted for with forms and devices that are straightforward counterparts
of those encountered in algebra-physics. Later, I will argue that there are some important aspects
of this and similar episodes that cannot be accounted for in this way. For instance, the placement
of these two lines inside the tick procedure, as well as their ordering within tick, needs some
explanation. Nonetheless, the importance of knowledge that carries over from algebra-physics is
very clear.
I want to take a moment to narrow down where, within programming, algebra-like
constructions and interpretations are localized. In the above examples, the interpreted statements
all used the change command. However, it turns out that algebra-like expressions are not limited
to change statements. More generally, any command that takes a numerical argument can have
one of these algebra-like expressions in the place of that argument. For example, the fd command
takes a single number as an argument, and that argument may take the form of an algebra-like
expression. Similarly, the first argument to repeat specifies the number of times to repeat a series
of commands. However, the vast majority of algebra-like expressions did appear as part of a
change command, as in the above examples.
304
It is important to note that the first argument in a change statement must be a single variable,
it cannot be an expression. For this reason, the algebraic counterparts of statements involving the
change command are limited to expressions of the form
x=[…],
the class of expressions that can be interpreted in terms of the IDENTITY form. Notably absent
from this class of expressions are many that would often be interpreted in terms of
BALANCING ,
such as those that state the equality of forces. An example is the equating of forces in the Mass on
a Spring problem solved by the algebra students: kx = mg. This is not to say that we could not
devise a means of working expressions of this sort into simulation programs. For example, if
students were taught to use conditionals (e.g., “if” statements) then the expressions which specify
conditions might very well be of this form. I will say more about this in Chapter 11.
More examples of algebra-like interpretations: The Shoved Block simulation
We have seen that a student’s ability to see meaning in an algebraic expression can be
borrowed and applied to the understanding of a class of single programming lines, particularly
lines involving the change command. At this stage, I could proceed methodically through each of
the forms and devices that we encountered in algebra-physics, providing an example episode in
which these forms and devices are applied to single programming statements. But these episodes
would be so similar to their algebraic counterparts that such an exercise would do little more than
replicate the examples and arguments given in Chapters 3 and 4. For this reason, I will not provide
programming examples of every form and device that appeared in algebra-physics and which also
appears in programming-physics. However, I do want to present just a few more examples in
order to illustrate how these algebra-like interpretation and construction episodes fit within the
context of student work on a programming task.
In each of the examples I will present here students were working on a programming
counterpart of the “Shoved Block” problem. Recall that in the algebra-physics version of this task,
the subjects were asked to consider a situation in which a heavier and lighter block, both resting on
a table with friction, are each given a hard shove so that the blocks start off quickly but eventually
come to rest. The problem then was to figure out which block travels a greater distance, given that
they start off with the same initial velocity.
305
In the corresponding programming task, the students were asked to program a simulation of a
the shoved block motion. Although these written instructions do not specify that the block is
always given the same initial velocity, this was clarified verbally for the students:
Imagine that a block is resting on a table with friction. Program a simulation for the case in which the
block is given a short, hard shove.
How does the motion differ if the block is heavier? Can you show this in your simulation? Modify
your simulation if necessary.
As a first step in this task, the students usually began by creating a simulation in which the
block simply slows down at a constant rate. Greg and Fred’s simulation, shown in Figure Chapter
9. -2, is a typical example. In the procedure named shove, their variables are initialized so that the
velocity is positive and the acceleration is negative, and then the tick procedure is repeated 150
times. Notice that this simulation program is nearly identical to the simulations of a drop we
encountered in Chapter 8, with the single exception that the acceleration is negative instead of
positive. In short, this is a straightforward application of the initialize-and-then-repeat structure
that the students had learned to implement.
Data
shove
setup
change
change
change
repeat
Doit
pos
vel
acc
150
0
12
-.08
tick
POS
VEL
ACC
894.000
-0.0000
-0.080
Data
Data
Data
TICK
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
Data
Figure Chapter 9. -2. Greg and Fred's simulation of a shoved block that slides to a halt.
But Greg and Fred’s simulation only addresses the first part of the task. In the second part,
the students were asked to modify their simulation to show how the motion would differ if the
block were heavier. As I discussed in Part 1 of this manuscript, it turns out that both blocks travel
the same distance. But, this answer was not obvious to the students. As in algebra-physics, most of
the students began with the notion that a heavier mass should slow down more quickly, and hence
306
travel a shorter distance, because of the increased frictional force. Here, I will focus on programs
that implement this initial, not entirely correct, intuition.
Greg and Fred were no exception in this regard. They began with the algebraic expression for
the force due to friction F = µN, writing this equation on a sheet of paper. Then, by looking at this
equation, they concluded that the frictional force would be greater if the block was heavier.
Fred
So, increasing the mass is just gonna increase the frictional force.
Greg
[nods]
Fred
The resistance.
Greg
Uh. By- What? Linearly?
Fred
Sure.
Greg
Yeah. Linearly.
With this information in hand, they went on to modify their program to take into account the
value of the mass. After making a variable named mass, they altered the line that initializes the
value of the acceleration. In this new version, rather than simply placing a constant value into acc,
they computed a value based on the mass:
shove
setup
change
change
change
repeat
pos
vel
acc
150
0
12
-.08 - mass
tick
Doit
Greg
So we can say mass data box is equal, uh, to acceleration plus - or acceleration minus a
certain (1.0) number that we give it. The mass. Say, the mass or something.
Fred
You're just saying that we just, um, decrease the acceleration or reduce the magnitude of
the acceleration by a fixed amount.
Greg
Yeah. No, I'm just makin' it - tryin' to make it organized, so we know that it's because of
the mass, you know.
…
Fred
So basically you're saying this is an additional acceleration factor.
Greg
Hm-mm.
Fred
Provided by the increased mass.
Greg
Hm-mm. (1.0) Does that make sense? Kind of? So, if you make this bigger [g. mass
variable], our mass, it's gonna decrease it even more. That makes sense, right?
This is precisely the type of form-device construction act that I focused on in my examination
of algebra-physics, and two forms play a prominent role in the construction of this line. First, the
BASE +CHANGE
form is prominent. Recall that -.08 was the value of the acceleration that Greg and
Fred used in their original simulation. The increasing of the mass is seen as leading to a change
from this base value. Fred says: “… we just, um, decrease the acceleration or reduce the
magnitude of the acceleration by a fixed amount.” In addition, the
307
PROP+
form is implicated
here: “So, if you make this bigger [g. mass variable], our mass, it's gonna decrease it even more.”
Also notice that the CHANGING PARAMETERS device is clearly evident in this final statement.
There are a couple of points to be made about this construction episode. First notice that the
variable that Greg and Fred have named “mass” is really playing the role of the change in mass,
since a value of zero simply produces the base case. Greg and Fred realized this shortly after the
above episode:
Greg
Yeah. I'm tryin' to make it, like, as simple as he did it here, see [g. pos, vel, acc variables]
So if you just input a mass, then it'll just do it for you.
Fred
But then remember that this point oh eight was- it had a mass to begin with. So you're
doing, like, we're adding more weight [g. mass variable] on top of the first mass.
(1.0)
Greg
Uh:::, yeah. Well, that's what he wanted. A bigger mass, right? [g. instructions] [reads] If
the block is heavier.
Fred
Okay, so, that's what we'll pretend. That first we have a weight, and we're adding more
weight on top of it.
Following this realization, they altered this key line of programming to be in accordance with a
more standard interpretation of the mass variable:
change acc
- mass
Doit
A second interesting point concerning this episode relates to the role played by the equation
F=µN. Notice that, once Greg and Fred had written this equation, they could have essentially just
transcribed it into their program. For example, using the fact that the normal force is N=mg, they
could have written:
change force
- mu * mass * g
Doit
change acc
force / mass
Doit
Or simply:
change acc
- mu * g
Doit
Instead, rather than simply transcribing an equation into their program, they took a detour
through the realm of symbolic forms. First they interpreted the equation F=µN in terms of forms,
then they wrote a programming statement that was in accordance with those forms. This
308
observation is interesting because it hints very strongly that forms are a central element of the
cognitive currency in both programming-physics and algebra-physics.
I will return to Greg and Fred’s work on this task shortly, but I want to digress for a moment
to point out that most of the other subject pairs took a very similar route in writing their shoved
block simulations. For example, Ned and David began by commenting that a heavier block
would slow down more quickly:
David
[reads] How does the motion differ if the block is heavier? Can you show this in your
simulation?
Ned
It's just more - it's (just) - the deceleration would be greater.
Then they wrote the following line to compute the acceleration from the mass:
change acc
mass / -100
Doit
David
Uh just the (st…) the acceleration is dependent on the mass. And the greater the mass,
the greater, the greater the acceleration is.
Bruce
Uh-huh.=
David
=To the left.
Bruce
Uh-huh. I see. The greater the acceleration is to the left. Right. (4.0) And where did the
divided by the slash minus a hundred come from?
Ned
That - we just chose that arbitrarily.
David
Actually, it's cause we, we used a hundred kilograms as a fir- as, as the first mass that we
used.
Bruce
Uh-huh.
David
And we, and we had that as being minus one acceleration in the first equation, so we just
used it.
Bruce
Hm-mm.
David
It didn't really mean anything.
Bruce
// [Points to screen.]
David
// I guess, uh, - (…)
Bruce
Go ahead.
David
I guess that could be like the coefficient of friction or something.
As in Greg and Fred’s work, the PROP+ form is implicated in Ned and David’s construction:
“And the greater the mass, the greater the acceleration is.” The DEPENDENCE form also appears
clearly in David’s first statement “the acceleration is dependent on the mass.” Interestingly, the
factor of -100 in the denominator seems to be understood in two different ways by Ned and
David. In some instances, it is treated as a COEFFICIENT that scales the mass in computing the
acceleration. In other instances, it appears to be understood as corresponding to the original value
they used for the mass. In that case, the
RATIO
form is implicated, and the acceleration is
computed from the ratio of the current mass and a reference value for the mass.
309
In writing their statement to compute the acceleration, Tim and Steve also employed such
reference values.
change acc
st_acc * enter_weight / st_wgt
Doit
In this statement, st_wgt and st_acc were the “standard weight” and “standard acceleration.” The
variable named enter_weight was the current value of the weight, which they presumed would be
entered each time by a user of their simulation. So, as in Ned and David’s statement, the ratio of
the current mass to a reference is important.
In all of the above episodes, the students were composing an expression that specified how the
acceleration of the shoved block depended on the mass. So, what we are seeing here is that there is
at least one specific niche in this programming-physics task where students engaged in some very
algebra-like construction and interpretation. Interestingly, it turns out that there was also a second
niche in this task where algebra-like expressions were written. This second niche arises because, if
the acceleration is altered and the initial velocity remains unchanged, then a different number of
ticks will be required to slow precisely to a halt. To account for this, Greg and Fred’s final shove
procedure appeared as follows:
shove
setup
change
change
change
change
pos 0
vel 12
mass .08
acc
- mass
Doit
repeat
150 - mass * 2
tick
Doit
Doit
In this new version of the program, the repeat statement has been modified so that the number of
cycles of the repeat will vary depending on the value of the mass. In particular, if the mass is
larger, then the number of repeats will be less. This is in agreement with the assumption that a
larger mass should experience a greater deceleration and thus stop more quickly. Greg and Fred
explained:
Fred
So, we have this idea that increasing the mass will, um, decrease the acceleration and
also decrease the number of steps it'll go. Right? Greg?
…
Greg
That was- that was our idea that - that, um, [g. a range on the path] a mass, if you have a
lighter mass it'll travel this distance. But if you have a heavier mass, [g. smaller range] it'll
travel, like, you know, less than that distance.
310
Bruce
I see.
Greg
With the same push. So that's why we thought that this repeat [g. 150 - mass * 2] is kind of
like a distance. So we tried to make it less than our initial, um, one fifty. [g. 150] So we
subtracted the mass [g. - mass * 2] a factor of the mass.
This is yet another example of a very algebra-like construction of a single programming statement.
Again, the main device active here is CHANGING PARAMETERS and the BASE -CHANGE , PROP+, and
COEFFICIENT
forms are implicated. Notice that the base value of 150 for the number of repeats
corresponds to the value used in their original simulation when the acceleration was -.08.
Once more, the big point here is that students constructed and interpreted certain
programming statements in a manner that was very similar to the construction and interpretation
of equations. This observation has many important implications for this research project. It means
that students can borrow from algebra-physics in their understanding of programming
expressions, and thus that we can have higher expectations for the ability of students to see
structure in programs. In addition, the observation that there is this area of overlap between
programming and algebra-physics can be seen as a first step in establishing that a programming
language can do some of the work of algebraic notation.
The Symbolic Forms of Programming-Physics
Now I will begin to discuss how an analysis of programming-physics in terms of symbolic
forms departs from what we observed in the analysis of algebra-physics. Thus far we have seen that
many of the forms in programming-physics knowledge are straightforward counterparts of the
forms of algebra-physics. However, even though many of the forms may be the same, the
frequency with which these forms appear may vary greatly between the two practices. This is one
type of difference that the form-based analysis can uncover, and such a frequency analysis will be
the job of the next chapter, when I report on the results of the systematic analysis of the
programming data corpus.
Instead, my task now is to describe the programming-physics forms for which there are no
obvious counterparts in algebra-physics. For the most part, these new forms involve a sensitivity to
line-spanning structuring in programming. First I will describe a new cluster of forms, the Process
Cluster. Then I will describe one additional form that did not appear in algebra-physics, SERIAL
DEPENDENCE,
which is an important line-spanning addition to the Dependence Cluster.
311
A new cluster: The ÒProcess ClusterÓ
When programs are being constructed, programming requires that the statements be arranged
in a sequence. This means that we will need to posit knowledge that accounts for the capability of
students to successfully order programming statements when constructing new programs. In my
analysis of algebra-physics, I asked questions such as: How did the students know to write ‘+’
instead of ‘×’? Now, in programming-physics, we need to be able to answer the question: How
did the students know to write that particular statement at that location in the program?
SEQUENTIAL PROCESSES
SETUP -LOOP
VARIATION
CONSTANCY
[
[
[
[
[
[
[
[
[
…
…
…
]
]
]
]
change a …
change b …
…
repeat …
]
…
change x …
change x …
…
]
]
]
]
Table Chapter 9. -1. Forms in the Process Cluster.
Just as in algebra-physics, the answer to this sort of question is frequently mundane.
Sometimes when a student writes an equation, it is reasonable to hypothesize that they are simply
drawing on a database of remembered equations. The same is true in programming: In some
cases, students know how to order the lines in a program because they are following, more or less
exactly, the arrangement they have seen in other programs. Of course, as in algebra-physics, there
are also cases that require truly novel constructions. For the most part, I will account for the ability
of students to produce novel arrangements of statements by positing a new cluster of forms that I
call the “Process Cluster.” The forms in this cluster are listed in Table Chapter 9. -1.
To get this discussion of Process forms started, I want to begin with an example of the type of
construction act that needs to be explained. Recall that, in Chapter 8, I described how Tim and
Steve produced a simulation of a dropped ball. Following their work on that simulation, they
312
reworked their program to simulate a motion in a which a ball, rather than simply being dropped,
is tossed straight up and then falls back down. The first working version of this new simulation
appeared as follows:
drop
toss
setup
change
change
change
repeat
drop
change acc -9.8
repeat 10 tick
pos 0
vel 98
acc -9.8
10 tick
Doit
Doit
Execution of this program starts in the box named toss. Toss begins by executing setup to position
the ball at its starting point and orient it facing upward. Then each of the three variables are
initialized and tick is repeated 10 times, taking the ball up to the apex of its motion. Following
the repeat statement, there is a single line that calls a procedure named drop, which takes the ball
from its apex back down to the bottom of its motion.
One thing that needs explaining here is how Tim and Steve knew to order the lines in toss. To
begin, notice that the first five lines of this procedure, from setup to the repeat statement, are in
line with the initialize-and-then-repeat structure, which we have already encountered a number of
times. Just as this structure is familiar to readers, it was also familiar to Tim and Steve; they had
already seen several examples and created a few programs following this structure. Thus, I believe
that it is reasonable to hypothesize that the sequence of statements
setup
initialize variables
repeat ## tick
is a structure that Tim and Steve essentially remember; it is part of a known repertoire of program
structures. (I will have more to say about this structure in a moment.)
But in appending a call to drop at the bottom of the toss procedure, Tim and Steve are
departing from the initialize-and-then-repeat structure. How did Tim and Steve know to write
this particular line in this location? To an experienced programmer, this may seem like an almost
trivial move, hardly worthy of an explanation. But, I hope to show that working to describe the
knowledge involved here actually extracts some of the most important differences between
algebra-physics and programming-physics.
The key to the explanation here is to note that, in writing the program in this manner, Tim
have Steve have broken the motion into two segments, the upward motion to the apex and the
313
downward motion. With this in mind, the toss procedure can be understood as consisting of two
sequenced parts that correspond to these two portions of the motion. This is illustrated in Figure
Chapter 9. -3.
toss
setup
change
change
change
repeat
pos 0
vel 98
acc -9.8
10 tick
drop
Ball goes up.
Ball comes down.
Doit
Figure Chapter 9. -3. Tim and Steve's Toss procedure broken into two parts.
When a procedure is seen in this manner, as consisting of two or more sequential pieces that
correspond to segments of a motion, I say that the SEQUENTIAL PROCESSES form is active. The
entities in the conceptual schema associated with SEQUENTIAL PROCESSES are two or more processes,
and the relationship that holds among these processes is simply that they are ordered.
Furthermore, the associated symbol pattern is:
SEQUENTIAL PROCESSES
[
[
[
…
…
…
]
]
]
In this notation, “[…]” stands for a single programming statement or a group of statements.
Thus, like other forms, when the SEQUENTIAL PROCESSES form is engaged it constrains what
students write when constructing new expressions. Furthermore, unlike the way that line-spanning
symbol arrangements are used in algebra, this arrangement follows a conceptualization of the
physical. In the present case, once Tim and Steve have understood the motion as consisting of two
parts—the upward motion and the drop—then the SEQUENTIAL PROCESSES form specifies that the
statements corresponding to these parts of the motion will be written one above the other, with an
order corresponding to the ordering of the processes. This is how I account for the fact that Tim
and Steve know to write “drop” at the bottom of the toss procedure.
As I hope to show in the next chapter, the SEQUENTIAL PROCESSES form was comparatively
common and accounted for many of the episodes in which students constructed truly novel
arrangements of statements. I will present one more example here to illustrate further. This
example relates to a simulation task that we have not yet encountered, the simulation of a mass
that oscillates horizontally on a spring. Early in their work on this task, when they were operating
314
in a somewhat experimental mode, Adam and Jerry produced a simulation that did not make use
of the Tick Model structure:
repeat 50
change pos 50
change pos 0
change pos -50
Doit
Adam
We could have one program that makes it this way. [g. right] One - and then one that
makes it go that way [g. left] then this way then that way. [g. back and forth]
Their idea, in writing this short program, was to simulate an oscillation by sequentially placing the
block at three positions: +50, 0, and then -50. If this was done repeatedly, then the block would
appear to oscillate, albeit in a somewhat jumpy and discontinuous manner.14 I explain Adam and
Jerry’s sequencing of these three statements by appeal to the SEQUENTIAL PROCESSES form; the
motion is broken down into sequential sub-processes that correspond to ordered statements in the
program.
Notice that, except for ordering, SEQUENTIAL PROCESSES treats all of the processes involved as
equivalent. As far as the schematization associated with this form is concerned, none of the
processes plays a special role. In contrast, we can imagine that it would be possible to develop
standard ways of breaking a motion into sub-processes in which the sub-processes are nonequivalent.
One such breakdown, associated with the initialize-repeat structure, is what I call the SETUP LOOP
form. In SETUP -LOOP , the program is seen as decomposed into a portion that gets everything
ready for the motion—for example, positioning the sprite and initializing variables—followed by
a repeat statement that repeatedly executes a procedure that specifies what transpires at each
moment of the motion. To take a simple example, Tim and Steve’s drop simulation, which we
examined in Chapter 8, can be seen in terms of SETUP -LOOP (refer to Figure Chapter 9. -4). Note
that, since SETUP -LOOP breaks down a motion into sub-processes associated with program parts, it
may be thought of as a special case of SEQUENTIAL PROCESSES.
14
When implemented, this program actually produces no motion whatsoever. The sprite does not move
in the graphics box, only the value of the variable is changed, since the fd statement must be used to
cause the sprite to move. Adam and Jerry made this correction in later versions of their program.
315
drop2
setup
change pos 0
change vel 0
change acc 9.8
Setup
repeat 10 tick
Loop
Doit
Figure Chapter 9. -4. The SETUP -LOOP structure in Tim and Steve's drop simulation.
The
SETUP -LOOP
form can be seen in the final interpretation of the drop simulation given by
almost every pair of students, though it is sometimes a little subtle and difficult to recognize.
Here are a few examples:
drop
setup
change
change
change
repeat
vel 0
acc 2
pos 0
22 tick
Doit
Clem
Change velocity, change [g.~ setup lines] - we initialize everything. [g.~ pos, vel, acc
variables] And then, um, and then we repeated the tick twenty two times. [g.m along
repeat line] [This last said with strong suggestion of "that's it."]
_______________
drop2
setup
change acc 2
change vel 0
repeat 23 tick
Doit
Ned
It's just pretty much the tick model. [g. the tick model.] We just started the acceleration [g.~ 2 in change acc 2] or, the acceleration's constant at two.
Bruce
Hm-mm.
Ned
That's just gravity. Velocity's [g.~ 0 in change vel 0] zero because you're letting it go [g.
release from top of graphics box], so the initial is zero.
Bruce
Hm-mm.
Ned
And we just repeated the tick. [g. repeated circling motion around tick box]
Bruce
Okay.
Ned
Twenty three just so it stops at the bottom. [g. along motion; g. bottom of graphics box.]
_______________
drop2
setup
change
change
change
repeat
pos 0
vel 0
acc 9.8
10 tick
Doit
Tim
Um, we just did the regular setup.
Bruce
Hm-mm.
316
Tim
To take us up there [g.m top of path] and then set position and velocity to zero. [g.m
change pos 0; change vel 0] And then we just have acceleration, nine point eight. [g.m
change acc 9.8] Which is so it’s gravity.
Bruce
Hm-mm.
Tim
And then we just repeat it ten times it just (0.5) falls. [g.m along path in g-box]
At first blush, it may seem like the above passages are little more than a simple reading of the lines
that appear in the various drop procedures. But, if we look closely, there are some clues that the
students are seeing the program in terms of SETUP -LOOP. First note the use of the words “initial”
and “initialize” in describing the lines above the repeat statement. This suggests an interpretation
of these lines that is in accord with SETUP -LOOP.
In addition, a more subtle—and perhaps more speculative—observation can be made
concerning the phrasing these students used in describing the repeat statements in these programs.
Tim said: “And then we just repeat it ten times it just falls.” And Ned said something quite
similar: “And we just repeated the tick.” Both of these statements suggest something obvious or
automatic about the repeating of tick at the end of these programs, a stance that I believe is closely
associated with the
SETUP -LOOP
form. We can think of the setup portion of the program as getting
things all ready to go—like winding up a spring or cocking the hammer on a pistol. Then, after
everything is ready, you just pull the trigger and the motion happens according to the parameters
you have defined. According to this stance, repeating tick is just like turning on the machine that
generates the motion from the initial conditions. This is the SETUP -LOOP stance: Get things ready
and then just let it run.
I want to make a side comment about
SEQUENTIAL PROCESSES
SETUP -LOOP
before proceeding. In my discussion of
above, I spoke of the initialize-and-then-repeat structure as if it was
something that students just remembered, like a remembered equation. In contrast, in this
discussion of
SETUP -LOOP,
I have been talking about this structure as the sort of meaningful pattern
associated with symbolic forms. Although this may seem confusing in the present context, it is
actually completely consistent with all of my earlier discussion of forms. The point is that, even
though the initialize-repeat structure may be “memorized” by students, this does not necessarily
mean that they are incapable of seeing it as meaningful. We encountered precisely the same
situation in algebra-physics. Any algebraic expression that a student writes from memory, such as
F=-kx, can also be interpreted by students and seen as meaningful in various ways.
The
SETUP -LOOP
form sets up a distinction between what is inside the loop structure and what
is outside. It turns out that this boundary between the inside and outside of the loop is a very
317
significant landmark for students, just as the line that separates the numerator from denominator
was a critical landmark in algebra-physics. This is an important enough point that I want to take a
moment just to present a pair of examples in which students’ sensitivity to this landmark is
evident. The first of these examples involves Greg and Fred’s work on the air resistance
simulation. The following program is taken from an intermediate stage of their work:
drop
setup
change
change
change
change
change
repeat
vel 0
air 0
mass 20
force mass * acc - vel
acc force / mass
sqrt
tick
1200 / acc
Doit
Doit
The feature of this program that is important for our current concerns is that the lines that
compute the force and acceleration are outside the loop—they are in the setup area—rather than
inside the repeat loop. For this reason, the force and acceleration will not be recomputed at each
point along the motion, as they should be to produce a correct simulation. Ultimately, Fred
caught on to this:
Fred
Now notice that this equation [g.m change force line] won’t do anything because it’s not
inside a loop. First of all, our velocity is zero. [g.m change vel 0]
…
Fred
None of this - none of this stuff is inside a loop. [g. down from top of drop box to repeat
line] Force times mass times acceleration minus vel. [g. along change force line]…And,
the whole repeating thing [g.O repeat line] doesn’t even touch what’s up here. [g. above
repeat]
Similarly, during their work on the spring simulation, Greg and Fred commented frequently
on whether various quantities should be computed inside or outside the repeat loop:
Fred
Yeah. So, whatever it is, if we were to give initial displacement it would be before a loop.
Greg
Uh-huh.
Fred
And inside a loop, position would change and - and acceleration would change. Hey that’s
pretty easy.
The point here is simply that comments of this sort were quite ubiquitous. The loop boundary was
an extremely significant landmark and it frequently played a role in student discussions.
Like
SETUP -LOOP,
the last two forms in this cluster also depend strongly on the landmark
associated with the loop boundary. In particular, these two forms involve registering the presence
of an individual programming statement as either inside or outside the loop. The key idea here is
318
that, if you want a quantity to vary through the time of a motion, then that quantity must be
computed within the repeat loop. And, conversely, if you want the quantity to be constant, it
must be computed outside the loop. I call the corresponding forms VARIATION and CONSTANCY.
Greg and Fred’s statements above provide some first examples of episodes involving these
forms. In those episodes, note that Greg and Fred were sensitive to whether a given statement was
inside or outside the loop. For example, Fred said: “Now notice that this equation won’t do
anything because it’s not inside a loop.”
To further illustrate, let’s look at how Greg and Fred’s work on the Air Resistance task
proceeded. As I noted, the version of the Air Resistance simulation presented above has the
problem that the change force and change acc lines are outside the loop. Fred was aware of this
difficulty, and spent quite a while explaining to Greg why the lines must be moved inside the
loop:
Fred
No, we don’t change it there. Air is - the air term is gonna be something that changes
during the calculation or during the fall. Remember that you said it’s dependent on
velocity, right?
…
Fred
So it’s actually gonna change the force each point, each tick.
Greg
Hm-mm.
Fred
So we’re gonna have to put change something.
…
Fred
It- it won’t calculate velocity but I’m just saying that, well, it’s gonna repeat this thing a
couple of times, right? [g. repeat line] And it’s changing velocity in tick. [g. change vel line
in tick] But we have to say, some time inside the repeat [g. repeat line], that our frictional
force is changing. And that would change acceleration.
Fred’s argument here is that since the “air term” changes through the motion, it must be
recomputed at each point. Ultimately, Greg was convinced and they went on to modify their
simulation so it appeared as follows.
drop
setup
change
change
change
repeat
vel 0
air 0
mass 20
200
change force mass * acc - vel
change acc force / mass
tick
Doit
Doit
In this simulation program, the lines which compute the force and acceleration have now been
moved inside the loop—along with tick, they are part of the short procedure that is executed on
319
each cycle of the repeat. This is an example of a construction effort that involves the VARIATION
form; because they want these quantities to vary, they know to compute them inside the loop.
Another observation that can be inserted here is that the above episode—and many of the
other passages I have thus far presented in this sub-section—are excerpts of extended and explicit
discussions concerning the location of lines in a program. The existence of these discussions
supports my point that, unlike algebra, it is important to posit symbolic forms to explain how
students construct the line ordering in their programs.
I want to digress to point out an important property of VARIATION and CONSTANCY: These
forms constrain where statements are written, but they do not, in most cases, uniquely specify
exactly where a statement should be placed in a program. This observation is parallel to points I
made during my discussion of algebra-physics. In that discussion we saw that not all features of an
algebraic expression are specified by the forms engaged in a construction effort. For example,
when we construct an expression from PROP+, or recognize an expression in terms of PROP+, it is
not significant whether the quantity in question is raised to the first, second, or third power.
Similarly, when programs are seen in terms of
VARIATION
and CONSTANCY, some aspects of the
ordering of statements are not relevant, all that matters is whether the lines are inside or outside
the loop.
Thus, what we are seeing is that, although programming enforces an ordering of all
statements, every detail of this ordering will not always be treated as meaningful. This is a
favorable situation, since the ordering of statements in a program is frequently not consequential,
even for the operation of the program. For example, it often does not matter how the initialization
statements in a simulation are ordered. Programming does force us to select an ordering for these
statements, but it is usually not appropriate to ascribe any particular meaning to the sequence
selected.
So, VARIATION and CONSTANCY, together with SEQUENTIAL PROCESSES and SETUP -LOOP,
constitute the four forms in the Process Cluster. Now I want to conclude this discussion of the
Process Cluster with a comment concerning the novel nature of these forms. Notice that, for the
sort of structure associated with this cluster, there is not much that can be borrowed from algebraphysics. Thus, much of the ability that these students have for recognizing this sort of structure
must have been learned during the few hours of the experimental sessions. For this reason, we are
most likely seeing only a fraction of the possible meaningful patterns that students could learn to
recognize at this level. If we developed a true practice of programming-physics, with many
320
participants engaged over a significant duration of time, I believe that the participants would learn
to recognize a wider range of patterns of this sort.
More specifically, just as
PROCESSES
SETUP -LOOP
does away with the uniformity inherent in SEQUENTIAL
by breaking a process into non-equivalent sub-processes, it is possible that a true practice
would include a rich vocabulary of forms, each of which breaks a process into sub-processes in a
unique manner. I believe that this is one of the major ways that the vocabulary of forms would be
extended in a more complete programming-physics practice, at least one built roughly along the
lines we are exploring here.
I did see some hints of possible directions in which the sub-process vocabulary might be
extended. My guesses in this regard are quite speculative, and there was certainly not enough
uniformity across subjects for me to add any forms to the list I have already given, but it is worth
mentioning a single example just to illustrate what I have in mind. This example is taken from
Greg and Fred’s work on their simulation of a mass oscillating on a spring. During their work on
this task, these students spent a fair amount of time discussing an issue that was prominent for
many of the subjects: How does the mass get moving? After some discussion, Greg and Fred
decided that there were two different methods they could use to initiate the motion: Either
displace the block away from the origin and release it, or just give the block a slap.
Fred
… I think there’s two ways you can think of this.
Greg
Yeah.
Fred
That, somehow there’s gonna be something to start the motion okay.
Greg
Yeah.
Fred
And it’s either gonna to be a displacement of the - of the::: // whatever it is.
// Hm-mm.
Greg
Fred
The block.
Greg
Hm-mm.
Fred
Or it’s gonna be a slap, an applied force.
321
go2
❝ Initial
Shove ❞
setup
change
change
change
change
change
change
vel 0
pos 0
k 5
mass 10
force 7
acc
- force / mass
Doit
tick
repeat 300
change acc
- k * pos / mass
Doit
change force - pos * k
tick
Doit
Doit
Figure Chapter 9. -5. Greg and Fred's Spring simulation that uses an "initial slap."
During the time they spent on this simulation, Greg and Fred created simulations
corresponding to each of these two possibilities. Figure Chapter 9. -5 shows a version of their
simulation that uses the latter method, what they came to call an “initial shove,” to initiate the
motion of the block. The lines associated with this initial shove are indicated in the figure. In the
first of these lines, a variable named force, which corresponds to the force of the applied slap, is set
to a particular value. Then, in the next statement, an acceleration is computed from this force.
Finally, there is a call to tick that essentially actuates this acceleration; one cycle of time goes by
with the acceleration set to the value determined in the previous line.
The structure employed in this program can be thought of as an extension to SETUP -LOOP.
Between the initialization lines and the repeat command, there is an intervening sub-process that
plays the role of initiating the motion—it gives the kick that gets the motion going. So, instead of
just SETUP -LOOP, we have a tripartite structure that might be called SETUP -INITIATE -LOOP. The
point here is that this SETUP -INITIATE -LOOP structure is just one of the many possible ways that the
vocabulary of sub-processes might be extended. I believe that if we were studying a true practice
of programming-physics, rather than just a mock-up in which students had only a couple hours of
experience, then the Process Cluster would be enriched to contain additional forms of this sort.
Before concluding this section, I want to briefly connect to some closely related work. Other
researchers have argued that programming knowledge includes what have been called “templates”
(Linn, Sloane, & Clancy, 1987; Mann, 1991) or “plans” (Ehrlich & Soloway, 1984; Soloway,
1986). These templates are schematic structures—like the ones I have been describing—that
constrain, at certain levels, the programs that experts and students write. For example, Soloway
322
(1986) describes what he calls the “sentinel-controlled running-total loop plan.” This template
reads in one value after another, adding these values to a total, and terminating when a specific
“sentinel” value is encountered:
initialize
ask user for a value
if input is not the sentinel value then
add new value into running total
loop back to input
This template describes the structure of a program at a certain level and degree of specificity.
In the template literature, it is usually emphasized that templates can exist at many levels; they can
exist at the level of individual programming statements, or closer to the level of entire programs.
Furthermore larger structures can be filled in to varying degrees of specificity. Notice that, in
allowing templates to exist in such diversity, these researchers are defining “template” in a very
general way. This is not a difficulty, but it suggests that we should be able to identify subcategories of templates that do different sorts of jobs, and which are tied to differently to other
knowledge. In fact, Ehrlich and Soloway (1984) do describe two “main kinds” of plans, “Control
Flow Plans,” and “Variable Plans.”
My position here is that, at least to a first approximation, it is reasonable to view the Process
forms—and perhaps programming forms, in general—as a particular sub-category of template.
The templates in this sub-category specify programs at a certain level, and have a certain degree of
specificity. Note that the structures in the Process forms have only a small number of components,
each of which is quite generic. For example, the SETUP -LOOP form has just two parts, and the
SEQUENTIAL PROCESSES
form is even looser and simpler; it is just a sequence of equivalent ordered
components, with the content of these components unspecified. This is in contrast to the
comparative complexity and specificity of Soloway’s sentinel-controlled running-total loop plan.
There is another way to understand the special characteristics of this sub-category of template.
In the templates literature, it is often emphasized that templates are associated with goals or
functions. If you have some goal that you want to accomplish, then you use the template that
accomplishes that goal. In my account of forms, the story of use is somewhat different. We do
not pull out a form to serve a function; instead, when constructing a program, a form is activated
because of the meaning of the construct—the associated conceptual schema. Thus, Process forms
are a particular sub-category of template that correspond to how we conceptualize a process, and
thus allow the generative construction of simulations out of meaningful components.
323
A new Dependence form:
SERIAL DEPENDENCE
Next, I will discuss another form that appeared in programming, but which I did not observe
in algebra-physics. Like the forms in the Process Cluster, this form is responsible for determining
how the lines in a program should be arranged; nonetheless, it is of a somewhat different sort. The
conceptual schema associated with this new form does not have processes as the fundamental
entities in its description. Rather it is more like the forms we encountered in algebra-physics in
being concerned with physical quantities and their relations.
I consider this new form, which I call SERIAL DEPENDENCE , to be an addition to the
Dependence Cluster. In my discussion of algebra-physics, I listed three forms in this cluster: NO
DEPENDENCE, DEPENDENCE,
and SOLE DEPENDENCE . Stated simply, the forms in this cluster are
concerned with whether one quantity or an expression depends on some other quantity. With
regard to symbol pattern, this means simply registering whether or not a given symbol appears, in
any manner, in an expression. As an example, I cited Mark’s statement that the force of air
resistance must depend on the velocity of the falling object:
R = µv
Mark
So this has to depend on velocity. [g. R] That's all I'm saying. Your resistance - the
resistor force depends on the velocity of the object. The higher the velocity the bigger the
resistance.
In programming-physics, the notion of dependence gets extended in an interesting way.
When writing simulations, students were often concerned with chains of dependent quantities,
such as d depends on c, which depends on b, which depends on a. Within a program, these chains
show up as the following symbol pattern:
change b
change c
SERIAL DEPENDENCE
change d
... a ...
... b ...
... c ...
...
In this programming symbol pattern, b is computed from a, c is computed from b, d is computed
from c, etc. This sequencing of linked dependence relations, which students wrote into and read
out of programs, is the meaningful pattern I call SERIAL DEPENDENCE
We already encountered an episode involving SERIAL DEPENDENCE in Chapter 8, when I gave an
extended description of Tim and Steve’s work on the Air Resistance simulation. Recall that, in
324
that episode, Tim raised the issue of how the lines in their modified tick procedure should be
ordered. In the end, they arrived at the following result:
TICK
change Res
vel * .1
Doit
change acc
acc - res
Doit
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
Tim
No we have to (then account) in the acceleration. So we have to change the acceleration.
Do we change the acceleration before we change the velocity? [g. change vel line in tick]
Or do we change acceleration after we change velocity? Which also changes the
resistance. [g. change Res line] Oh, I’m getting confused. What’s the order in which you
change things?
…
Tim
Well you have to change resistance because you change the velocity, because (0.5) the
velocity (1.0) - the new velocity is dependent upon the acceleration which is dependent
upon the resistance. So we should change resistance, then acceleration, then velocity.
Right? Does that make sense, or am I just talking nonsense.
The question that Tim and Steve were wrestling with was, given that a number of quantities must
be computed in the tick procedure, “What’s the order in which you change things?” What they
decided was to compute res (a quantity associated with the air resistance) from the current value
of the velocity. Then they got an acceleration from this value of res, and finally a new velocity and
position from the acceleration.
Tim’s assertion is that this ordering somehow actually reflects the structure of dependence in
this situation, “the new velocity is dependent upon the acceleration which is dependent upon the
resistance.” Of course, he does not provide much of an argument for this view in the above
passage. Nevertheless, this ordering does turn out to be in agreement with one important way of
conceptualizing the structure of Newtonian physics: Find the force using a “force law” and then
use F=ma to find the acceleration. I will have more to say about this structure in Chapter 11.
SERIAL DEPENDENCE issues arose frequently during student work on the Air Resistance
simulation, as in Tim and Steve’s discussion above, and also in the context of the Spring
simulation. Ned and David’s brief discussion below is the tail end of a typical passage.
325
TICK
change vel
vel + acc
Doit
change pos
pos + vel
Doit
change force
-1 * pos
* k
Doit
Doit
change acc
force / mass
Doit
fd vel
dot
Doit
Ned
Does the force change because the acceleration changes?
David
The force changes ’cause X changes. And so that, and so=
Ned
=That causes acceleration to change.
In this example, Ned and David decided on a particular dependence structure; the position
determines the force which determines the acceleration. Again note that, like the Process forms,
SERIAL DEPENDENCE
constrains the arrangement of lines in a simulation program.
Similarly, the question of which quantities should determine which others played a central role
in Adam and Jerry’s work on the Spring simulation. Early on, Adam strongly argued that the
position was at the root of the dependence structure:
Adam
Well basically, the position is ba- um, the acceleration, everything,
Bruce
Hm-mm.
Adam
is based on where this is, the position, [g. pos in change acc line]
In accordance with this viewpoint, Adam and Jerry ultimately settled on the version of the tick
procedure shown below. (In this program, the variable k was elsewhere set to a negative value.)
TICK
change force
k * pos
Doit
change acc
force / mass
Doit
change vel
vel + acc
Doit
change pos
fd vel
dot
pos + vel
Doit
Doit
Finally, after they had produced a complete and working simulation, Adam and Jerry had a
number of discussions which involved recognition of SERIAL DEPENDENCE in this procedure. Here is
326
an extended excerpt where Adam once again argues for the primacy of position in determining
the other quantities:
Adam
And the position is the most important thing here. If we made it the other way around, we
don’t care about the acceleration, we care about where it is.
Jerry
Hm-mm.
Adam
See acceleration [g. change acc line] derives from the position. [g. change force line]
Jerry
We don’t know how acceleration changes like- the only th-
Adam
We wouldn’t know how position changes (…) acceleration I guess.
Jerry
Hm-mm.
Adam
That’s the best way. It’s easier just to think about it as position affecting acceleration.
Jerry
Hm-mm.
Adam
Than acceleration affecting position.
Jerry
Hm-mm.
Bruce
It’s easier to think of it that way?
Adam
Well it - because - I mean, what we’re // thinking of is that force depends on how far we pull
it. If we do it your way, we have to know how far - how the force depends on how fast you
pull it back. Or something. Something of that term.
// That’s what happens.
Jerry
Jerry
Hm-mm.
Adam
Yeah.
Jerry
This is like - this is like saying, okay, given a position, this is the acceleration. Whereas
the other way would be saying here is our acceleration, now where is it along the-
In this and the preceding passages, the students involved were engaged with a quite unusual set
of issues. The question under consideration—which quantities determine which others—is not a
question that arises very often in algebra-physics. In fact, as I mentioned in Chapter 8, such a
question may not even make sense within algebra-physics. But I will forestall a full discussion of
these issues until Chapter 11. My goal for this section has simply been to establish that the SERIAL
DEPENDENCE
form corresponds to a new type of meaningful pattern that my students learned to
see in their simulation programs and which, because it spans multiple statements, constrains the
arrangement of statements in novel constructions.
Representational Devices and the Interpretation of Simulation Programs
One of the original intuitions with which I began this work was that programs should be easier
to interpret than algebraic expressions. As I said in Chapter 8, this intuition is rooted in the notion
that programs can be “mentally run” to generate motion phenomena. If this is correct—if
interpreting a program always means mentally running it—then the interpretation of programs
may be, overall, a very different kind of thing than the interpretation of equations. Furthermore, if
327
there is a single, preferred mode for interpreting programs, this could mean that interpreting
programs is much easier than the interpretation of algebraic expressions.
However, we have already seen some evidence that this is not the case. Earlier in this chapter I
presented a number of examples in which individual programming statements were interpreted
much like the expressions of algebra-physics. For example, David’s interpretation of this
statement is very similar to the CHANGING PARAMETERS interpretations we encountered in algebraphysics.
change acc
mass / -100
Doit
David
Uh just the (st…) the acceleration is dependent on the mass. And the greater the mass,
the greater, the greater the acceleration is.
Furthermore, it turns out that the use of algebra-like devices even extends beyond the
interpretation of individual programming statements. In some cases, the devices of algebra-physics
were used to interpret groups of statements selected from a program. For illustration, consider
Ned and David’s Air Resistance simulation:
drop2
setup
change
change
change
change
change
TICK
change vel
pos 0
drag 0
mass 5.0
grav 2.0
force
mass * grav
Doit
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
change acc
change vel 0
repeat 700 tick
force - drag
/ mass
Doit
Doit
Doit
change drag
vel / 1.001
Doit
dot
Doit
The above program is the final version of the simulation that Ned and David produced, and it
correctly simulates the approach of the dropped object to terminal velocity. Like all of my
subjects, Ned and David were somewhat impressed when their simulation actually worked, even
though they were fairly clear on the physical principles involved. Thus it was natural for them to
try to talk through why this simulation has the property that it approaches a terminal velocity. The
final explanation, given by David, was based solely around two key lines from the program:
328
change acc
force - drag
/ mass
Doit
Doit
change drag
vel / 1.001
Doit
David
Um, yeah, because at, at the terminal velocity, this value for vel [g. vel in change drag
line], the drag will equal the force. [g. drag, force, drag in change acc line] So then you
have no - no more acceleration. And so as - as uh - as velocity approaches this value [g.
1.001] times the force, [g. force],
Bruce
Hm-mm.
David
um, the acceleration decreases. And so, as it gets close, the acceleration decreases
slower and slower and slower until finally it reaches zero.
What David is saying here is that, as the velocity increases, the value of the drag variable will
increase until it equals the value of force. At that moment the first line returns zero and terminal
velocity has been reached. Furthermore, by looking at the second line, he is able to tell that this
cancellation will occur when the velocity is equal to 1.001 times the force.
There are two important points to be made about this passage. The first point is that, although
it involves multiple lines, it is really quite similar to the PHYSICAL CHANGE interpretations we
encountered in algebra-physics. As in those interpretations, David looks at a symbolic display and
talks through how the variables in that display vary through the time of the motion. The following
interpretation, which I originally presented in Chapter 4, is clearly of the same ilk:
∑F = mg - kv
Alan
And so I guess we were just assuming that this was [g. kv] de-accelerating - that we would
have gravity accelerating the ball and then as, as we got a certain velocity, [g. kv] this
would contribute to making accelerating closer - de-accelerate it and so we would just
have a function of G minus K V. And then we assumed that since we know that it reaches
terminal velocity, its acceleration's gonna be zero, we just assumed G minus K V would
equal zero.
A second point to be made about David’s interpretation is that, if his interpretation is at all
representative, then understanding a program may mean focusing on only a small piece. Keep in
mind that it really does matter what the rest of the program is; these two lines alone are not
sufficient to generate the correct behavior. But David’s explanation does not mention the other
lines in the program; he instead makes implicit presumptions about the behavior of these other
lines. In David’s explanation, the rest of the program fades into the background.
Really, I could have made this same point at the beginning of this chapter when I observed
that students frequently give interpretations of single lines in a program. Just as in David’s
explanation, those interpretations often required implicit assumptions about the rest of the
329
program. But David’s interpretation makes it strikingly clear that students may try to understand
global properties of a program by looking only at a small piece.
Clearly, the interpretation of a program is not a wholly different and much easier task than the
interpretation of an equation. Individual and groups of statements are interpreted through devices
that are straightforward counterparts of the devices of algebra-physics. If anything, this might be
taken as suggesting that interpreting programs is a little trickier than interpreting equations, since
some cleverness must go into selecting the lines for focus. But there are a number of factors to
weigh, and I will defer the question of which, on balance, is “easier” until Chapter 11.
There are some more profound ways in which the interpreting of programs departs from the
interpreting of equations, and talking about these points of strong departure is the task of the
remainder of this section. What I will argue is that there are two critical and quite common
programming-physics devices for which there no algebraic counterparts, what I call TRACING and
RECURRENCE.
A new Narrative device:
TRACING
I have taken pains to argue that not all algebra-physics interpretations involve mentally
stepping through a program—what has sometimes been called “mental simulation” (e.g.,
Letovsky, Pinto, Lampert, Soloway, 1987). However, there is one common representational device
that can reasonably be considered to be a variety of mental running. Interpretations based on this
new Narrative device, which I call TRACING , take their narrative form from the line-by-line
structure of a program.
For a first example of a TRACING interpretation, I turn to Adam and Jerry’s work on the Spring
simulation. During my discussion of
SEQUENTIAL PROCESSES,
I described how Adam and Jerry
created a preliminary version of this simulation which worked by successively placing the block at
three distinct locations. After modifying this program to leave a trail of dots, Adam checked that
he had something sensible by tracing through what they had written:
repeat 50
change pos 50
dot
change pos 0
dot
change pos -50
dot
Doit
Adam
Make a dot right there. [g. g-box] Change to fifty. Make a dot. [g. a point to the right of
center in the g-box] Go back to here make a dot. [g. center] Go here, make a dot. [g. left]
And repeat.
330
In this interpretation, Adam steps through the program one line at a time. As he does, he points
into the graphics box on the display, moving his hand as the block would move for each
statement. This is a TRACING interpretation, a narrative interpretation that follows the structure of
the written program.
To further clarify the nature of TRACING , it helps to distinguish it from PHYSICAL CHANGE , since
both of these devices involve a description of a phenomenon that follows through the time of the
motion. The difference enters in how these descriptions are related to the written expression. In
the case of PHYSICAL CHANGE , one or two lines in the program are selected, and then the
interpretation describes how quantities in the expression vary through time. In contrast, rather
than focusing on how the quantities in a single line vary through time, a TRACING interpretation
proceeds, line-by-line, through a program. Compare David’s interpretation of the Air Resistance
program described just above. The action in that interpretation concerns how the value of the drag
variable changes through time, with attention focused on two statements. But the action in
Adam’s
TRACING
interpretation flows from line to line in the procedure.
Interestingly, stepping through the lines of a program will not, in all cases, correspond to
stepping through the time of the motion. For illustration, consider Ned and David’s
interpretation of this tick procedure, which has been modified for the Spring simulation:
TICK
change vel
vel + acc
Doit
change pos
pos + vel
Doit
change force
pos * k
Doit
change acc
force / mass
Doit
fd vel
dot
Doit
David
See, cause if we start off with zero acceleration, velocity would be velocity. And it'll go
once. [g. moves finger short distance to right] Position'll be position plus velocity. Force
would- (0.5) position times K. And then acceleration'd be force divided by mass. And then,
it'll use that acc, right, the next time around?
In this passage, David proceeds through the lines in tick under the assumption that the
acceleration is zero. This is clearly an instance of the TRACING device; in fact, as noted in Appendix
E, TRACING interpretations frequently take the form of a series of statements separated by the
phrase “and then,” as in David’s statement above. The point I want to make about this
interpretation is that, although it follows the lines in the program, it is not precisely correct to
331
think about the sequence of operations as happening through time. In the physical system that this
simulation purports to model, all of these parameters are constantly changing; they do not change
one at a time as in this program. In essence, the act of programming forces the students to
explode each instant of the motion into a series of actions that happen through what I will call
“pseudo-time.” Similar issues arose in my discussion of SERIAL DEPENDENCE and I will discuss this
issue further in Chapter 11.
TRACING interpretations do not always follow at the level of individual lines. In some cases, a
program is parsed in chunks—perhaps as dictated by a Process form—and then the interpretation
steps through these chunks. Compare the following two examples which I originally presented in
my discussion of
SETUP -LOOP.
Tim mentions each line in his version of the Drop simulation, but
Clem, in a very simple TRACING interpretation, just says “we initialize everything” when stepping
past the “setup” portion of the program.
drop2
setup
change
change
change
repeat
pos 0
vel 0
acc 9.8
10 tick
Doit
Tim
Um, we just did the regular setup.
Bruce
Hm-mm.
Tim
To take us up there [g.m top of path] and then set position and velocity to zero. [g.m
change pos 0; change vel 0] And then we just have acceleration, nine point eight. [g.m
change acc 9.8] Which is so it’s gravity.
Bruce
Hm-mm.
Tim
And then we just repeat it ten times it just (0.5) falls. [g.m along path in g-box]
_______________
drop
setup
change
change
change
repeat
vel 0
acc 2
pos 0
22 tick
Doit
Clem
Change velocity, change [g.~ setup lines] - we initialize everything. [g.~ pos, vel, acc
variables] And then, um, and then we repeated the tick twenty two times. [g.m along
repeat line] [This last said with strong suggestion of "that's it."]
The TRACING device is our first point of strong departure from the devices of algebra-physics.
It is closely related to a commonsense notion of mental running, a phenomenon that is likely to be
unique to programming. Furthermore, as I hope to show in the next chapter, TRACING is not an
inconsequential addition; it turns out to be one of the most common devices in programmingphysics.
332
A new Static device: RECURRENCE
The second new device in programming-physics, which I call RECURRENCE, constitutes an
addition to the Static Class and is a new type of stance toward individual statements. A brief
review here of Static devices is warranted. Devices in the Static class include
SPECIFIC MOMENT ,
in
which the stance is taken that an expression is true only at some particular instant in a motion, and
GENERIC MOMENT ,
in which the stance is adopted that an expression is true at any moment. Mike
and Karl’s interpretation of the following expression from their work on the Air Resistance task is
an example of GENERIC MOMENT . In Chapter 4, I emphasized Mike’s use of the phrase “at any
time.”
a(t ) = −g +
Mike
f (v)
m
So, at least we can agree on this and we can start our problem from this scenario. [g.
Indicates the diagram with a sweeping gesture.] Right? Okay? So, at any time,, At
any time, the acceleration due to gravity is G, and the acceleration due to the
resistance force is F of V over M. [w. g + f(v)/m.]
Now, in contrast, consider Fred’s interpretation of an Air Resistance simulation. Notice that,
where Mike would have said “at any time,” Fred says “each tick.”
drop
setup
change
change
change
repeat
vel 0
air 0
mass 20
200
change force mass * acc - vel
change acc force / mass
tick
Doit
Doit
Fred
No, we don’t change it there. Air is - the air term is gonna be something that changes
during the calculation or during the fall. Remember that you said it’s dependent on
velocity, right?
…
Fred
So it’s actually gonna change the force each point, each tick.
Greg
Hm-mm.
Fred
So we’re gonna have to put change something.
The point of this example is that, rather than adopting the stance that an expression states what is
true at any time, Fred adopts the stance that the change force line states what transpires at each
moment along the motion. Given only this example, this may seem like a subtle and perhaps
inconsequential distinction. However, there are cases where this difference in stance actually has
very strong implications. Allow me to explain.
333
Earlier I tried to show that it is common for programming statements involving the change
command to be constructed and interpreted in a manner very similar to certain classes of
equations. For example, the statement
change acc
mass * 9.8 - res
/ mass
Doit
Doit
can be understood very much like the corresponding equation:
a( t ) =
mg − Fair
m
However, this simple transcription between programming and algebraic notation does not always
work. Consider the following statement from the standard tick procedure along with Tim’s
interpretation:
change vel
vel + acc
Doit
Tim
So, every time - every tick it gains nine point eight. The acceleration increases the
velocity by nine point eight.
If we were to translate this line into algebraic notation in a straightforward manner, then we might
write v=v+a. Clearly there is something wrong with this translated expression, since we can cancel
the two velocity terms to obtain 0 = a.
The translation fails because each of the two vel’s that appear in the programming statement
does not refer to the same quantity. The vel in the second argument stands for the old value of the
velocity, the one produced the last time through the tick model, and the vel in the first argument
is a new value of the velocity that is to be computed from that old value. The result, as Tim tells us
in his RECURRENCE interpretation, is that the value in the velocity variable changes by 9.8. Thus, a
more correct translation into algebraic notation might look like: vnew = vold +a.
So, in this case, it is somewhat inappropriate to think of this programming expression as a
constraint among various quantities that holds at any instant. Instead it is best read as describing
what transpires on each tick, that the value of acc is added onto the velocity. This is an instance in
which the difference between RECURRENCE and GENERIC MOMENT is critical.
It was very common for RECURRENCE interpretations to be given for lines of the form
change x x+y, as in the change vel and change pos lines in tick. But, as we have already seen in
334
Greg and Fred’s interpretation above—my first
RECURRENCE
example—there are some exceptions.
For further illustration, I will present an episode in which Ned and David invented a quite unique
solution for the Air Resistance task. Rather than computing the acceleration at each point by
finding the total force and dividing by the mass, they created a program in which the value of the
acceleration was scaled down on each iteration of the tick procedure. In the following passage,
note David’s use of the phrase “every time you repeat.”15
TICK
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
change acc
acc / drag
Doit
dot
Doit
David
What if it's like acc divided by two? Acc divided by two, acc divided by two, acc divided by
two, acc divided by two, acc divided by two.
Ned
What?
David
So, like you - so, it's not like - so that every time you repeat the acceleration's cut down by
a factor. Not like subtracted but it's cut down by a factor. (And) if it keeps going infinitely
many times, eventually it's gonna just keep on,, You know what I mean?
This is a construction involving RECURRENCE because the statement is understood as specifying
how, on each iteration, the new value of acc is computed from the previous value.
This close and subtle relation between GENERIC MOMENT and RECURRENCE was actually a cause
of some difficulty for students; in general, students were perhaps not as disposed to see
RECURRENCE
as they could profitably have been. For an example, we return to Tim and Steve’s
work on the Air Resistance simulation, which we have had reason to examine on numerous
occasions. Recall that, in my discussion of SERIAL DEPENDENCE , I presented an early version of Tim
and Steve’s modified tick procedure:
15
Interestingly, this simulation is in agreement with the assumption that the force of air resistance is a
linear function of velocity: F=-kv.
335
TICK
change Res
vel * .1
Doit
change acc
acc - res
Doit
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
As I discussed in Chapter 8, there is a problem with the change acc statement in this procedure,
and the resulting simulation actually produces oscillations—the ball does not just fall, it moves up
and down. Tim and Steve were surprised and confused by this behavior, and they were never able
to produce an adequate explanation. After rewriting the change acc statement as
change acc
mass * acc - res
/ mass
Doit
Doit
they repeatedly gave
GENERIC MOMENT /COMPETING TERMS
interpretations of the following sort:
Tim
We just said, this is - the resistance is a vector going up.
Bruce
Uh-huh.
Tim
And then this is the - mass times acceleration is a vector going down. [g. mass * acc]
Bruce
I see.
Tim
And we subtracted it to get a force. The total for - the sum of the forces. Divided that by
mass to get acceleration.
Here, the mass * acc term is seen as corresponding to the force of gravity acting down, rather than
being associated with the previous value of the acceleration.
Of course, Tim and Steve’s error can be easily explained as a minor slip-up; it is almost certain
that they intend the acc in mass * acc to stand for the acceleration due to gravity. In fact, after
some pretty strong coaxing, they rewrote this line as:
change acc
mass * 9.8 - res
/ mass
Doit
Doit
and they seemed to think that this was in agreement with their original intent. Nonetheless their
lack of inclination to adopt a RECURRENCE stance had important implications here. If Tim and
336
Steve could have productively adopted
RECURRENCE,
they would have been in a better position to
understand the strange behavior produced by their original simulation, and perhaps corrected the
difficulty.
This story has a very interesting postscript. After completing their work on the Air Resistance
task, Tim and Steve—like all students—were asked to create the simulation of a mass oscillating
on a spring. After reading the instructions for this new task, Tim and Steve realized that they had
a leg-up on this one, they already knew how to make a simulation that oscillates! They simply
copied their original oscillating air resistance simulation, modified only so that the oscillations
were horizontal rather than vertical.
tick
change res
vel * k
Doit
change acc
mass * acc - res
/ mass
Doit
Doit
change vel
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
This simulation worked quite well and these students were satisfied. Unfortunately, however, Tim
and Steve’s understanding of this simulation was questionable from the point of view of
Newtonian physics. In the following passage, Tim essentially gives a COMPETING TERMS
interpretation of the above statement for the acceleration:
Tim
Just like try - well you can just - basically this- this thing [g.m.O (mass*acc - res)/mass] is
just basically dealing with like (10.0) I guess, forces?
Bruce
Hm-mm.
Tim
You have two opposing, two different force vectors [g. points in opposite directions with
each hand], and they’re- and how they’re - how they’re (effect- changing) each other?
Bruce
Hm-mm. And what’re those - what are those two different forces?
(16.0)
Tim
One’s kinda the - (2.0) one has to do with I guess the mass of the ball or whatever this is.
Block. And like, (1.0) - when you like pull it out, [g. pulling block out to the right] it has um the spring wants to pull it back that way [g. left] so it starts moving and it gets - there’s this
force vector I guess (4.0),, It kinda has to do with like the, I don’t know, inertia? Of the
block? Once it starts moving? That way. I don’t know.
(4.0)
Bruce
Hm-mm.
Tim
And the other - other force is the force that the spring exerts on the block.
337
Bruce
Hm-mm.
Tim
So there’s like the force of the block and the force of the spring.
In this passage it is clear that Tim understands the change acc line as computing the total force
and dividing by the mass. One of these forces is associated with the force exerted by the spring
and the other is associated with the “inertia” of the block. This is not a correct description of the
situation from the point of view of Newtonian mechanics. A physicist would say that there is only
one force on the block, the force exerted by the spring.
Although Tim’s interpretation is not really appropriate, it turns out that this simulation is
correct. Not only does it produce oscillations, it is also in agreement with the usual assumption
that the force exerted by a spring is linear in the position, though it is not a trivial exercise to see
that this is the case. But a first step toward a correct understanding of this unusual simulation
would be to adopt a RECURRENCE stance toward the change acc statement.
One possible conclusion we can draw from these episodes is that Tim and Steve are not very
strongly disposed to employ
RECURRENCE.
Conceivably, if they were more experienced in
programming-physics, Tim and Steve would be more aware that some statements require this
type of interpretation, and they might be able to come up with useful interpretations of even these
tricky expressions. Part of the problem here is almost certainly interference from all of their
experience with GENERIC MOMENT interpretation in algebra-physics. Expertise in programmingphysics might very well lead to an expansion in how frequently
RECURRENCE
is employed, and to
more flexible use.
Conclusion
We have now taken our first steps into the structured and meaningful world of programs. As
for algebra-physics, this step took the form of laying out the knowledge—forms and
devices—that is associated with recognizing a certain type of meaningful structure in programs.
Along the way, I frequently emphasized the caveat that these students have much less experience
with programming-physics than did their opposite numbers in algebra-physics. Nonetheless, it
turns out that they quickly learned to see a great deal of meaningful, physics-relevant structure in
programs. Part of the reason for this, I argued, is that much can be borrowed from algebraphysics, particularly for interpreting certain types of individual programming statements.
Even though much can be borrowed from algebra, most of my time in this chapter was spent
describing the forms and devices in programming-physics for which there were no obvious
338
algebraic counterparts. In the forms arena, I argued that the big story was the addition of forms
involving patterns that span lines of programming. Most of these line-spanning forms were in a
new cluster—the Process Cluster. Also, we encountered a new type of Dependence form,
DEPENDENCE,
SERIAL
which involves chained dependence relations of multiple physical quantities.
The story for devices was a little shorter and less dramatic—there were only two new devices.
One of these, RECURRENCE, is a new type of stance toward single programming statements. The
other new device, TRACING , is an absolutely critical and extremely common new device that is
closely related to what we might call “mental running.”
With this chapter, I have laid out the basic elements in the form-device system of
programming-physics. We have seen where and how forms like those in algebra-physics can be
applied, and I have described the new forms that appear. Now, we are ready to further pin down
some of the details of this practice. In the next chapter, I will attempt to establish through a
systematic analysis that the Theory of Forms and Devices can adequately account for the
necessary phenomena in programming-physics—the subset of phenomena that they should
account for—much as I did in the systematic analysis of the algebra-physics corpus described in
Chapter 6. I will attempt to show, in this manner, that the theory works, that the lists of forms
and devices presented here are sufficient, and I will obtain specific counts indicating the relative
frequency of the various forms and devices.
339
Chapter 10. Systematic Analysis of the Programming
Data Corpus
The broad strokes of my analysis of programming-physics are now in place. Because I gave a
number of examples—and simply because I showed how it could be done—I believe that the
discussion of the previous chapter makes it plausible that the Theory of Forms and Devices can be
applied to programming-physics. Now, as I did for algebra-physics, I will fill in the
methodological basis for the assertions made in that previous chapter. In particular, I will describe
how my account of the forms and devices of programming-physics was arrived at through a
process of systematically coding and recoding a large portion of my programming data corpus,
and I will present the detailed results of this systematic analysis.
To clarify, there are essentially two types of results of this systematic analysis. The first result
is simply the lists of forms and devices that are required to describe the data corpus. This “result”
was already presented in the previous chapter and now, in this chapter, I will say how I arrived at
these lists. The second type of result is the set of tallies of the specific forms and devices. These
tallies provide a summary of my data and a more detailed image of the practice of programmingphysics. Just as for algebra-physics—and with the same caveats—the tallying of specific forms and
devices paints a particular kind of picture of programming-physics knowledge. The key
presumption is that these tallies reflect the priority weightings of knowledge elements within the
form-device knowledge system.
Of course, one of the main reasons for producing the characterization of programming-physics
is to compare the set of numbers obtained to those obtained in the systematic analysis of algebraphysics. And, to a certain extent, I will do this comparison in the present chapter. As I present
various numerical results, I will try to provide some added perspective by weighing them against
the analogous values from algebra-physics. However, I will not go too far in drawing conclusions
and interpreting the differences observed. That job will be left primarily to the next chapter,
where I will pull together all of my various observations concerning programming-physics and I
will sum up the differences between programming-physics and algebra-physics. In contrast, the
main work of this chapter will be focused on methodological groundwork. I will first spend
significant time on describing what data I collected, how it was collected, and on my analysis
techniques. Then, when this is done, I will present the numerical results with some preliminary
comparisons to algebra-physics.
340
As for the analysis in Chapter 6, there are some limits to the analysis I will present here and
most of the same caveats apply. The analysis process I will describe is highly heuristic, and was not
corroborated by additional coders. In addition, there are limits to what we can conclude due to
the small number of subjects involved and the limited task selection. Nevertheless, I believe that
that results allow for a revealing comparison of the two practices.
About the Data Corpus
Subjects
The participants in the programming portion of this study were all currently enrolled in
“Physics 7C” at U.C. Berkeley, a third semester introductory physics course for scientists and
engineers. This is the same course from which I drew students for the algebra pool and a brief
description of this and preceding courses can be found in Chapter 6. However, since the two
portions of the study were conducted during successive semesters—algebra in the fall and
programming in the spring—the programming pool subjects were actually drawn from a different
class of students taught by a different professor.
Recruitment in this phase of the study worked precisely as it did in the first phase. I went to a
Physics 7C lecture early in the semester and made a brief announcement asking for volunteers.
The students were told they would be paid $6 per hour instead of the $5 per hour paid to the
algebra subjects. I felt that this slight increase was necessary since, as I will discuss below, the
programming study required a somewhat greater time commitment from students. Thus, I
wanted additional inducement for students to stay in the study for its duration.
Once I had a list of volunteers, I selected names from the list at random and telephoned
students. As for the algebra-physics study, I scheduled 7 pairs hoping that at least 5 would finish
the study. It happened that all 7 pairs completed all of the programming tasks.
Of the 52 students that volunteered, 41 were men and 11 women. This high percentage of
male volunteers is not surprising given the fact that the class is predominantly male. The fourteen
students that participated in the study consisted of 12 men and 2 women, with 5 pairings of two
men and 2 pairings of a man and a women. Even though I had 52 total volunteers, the selection
and pairing of subjects was largely dictated by scheduling constraints. The students in Physics 7C
have very busy—and often very similar—schedules, and students needed to agree to a weekly 2hour block, not already occupied, in order to participate in the study.
341
It is important to emphasize that I did not screen for programming background in any way;
enrollment in Physics 7C was the only requirement and I did not mention programming in my
recruitment announcement. It turned out that all of the participating students except one reported
that they had received some prior programming instruction, but the range of experience was great.
For example, Tim was a computer science major and stated that he knew how to program in
Basic, Pascal, Scheme, C, and C++. In contrast, Steve, who was Tim’s partner and had gone to the
same high school as Tim, reported that his only programming experience was some brief
instruction he had received in junior high school. Although some students reported experience
between these extremes—for instance, some had taken a single college programming
course—Steve’s experience was closer to the norm.
As I stated above, unlike the algebra subjects, all 7 programming pairs completed the entirety
of the tasks. Thus, I was faced with selecting 5 of these pairs for analysis. I could have chosen these
5 at random but I opted, instead, to make my work easier. It turned out that the speech of 2 of
the 7 pairs was much more difficult to understand, which posed problems during early attempts
at transcription. For that reason, I chose not to include these two pairs in the systematic analysis.
This is certainly not a “principled” maneuver, and language issues are actually important and
interesting for this work. However, given the exploratory nature of this research and the already
high burden of transcription and analysis, I believe that this choice was reasonable.
Tasks and Training
Students’ work progressed through a series of programming tasks and exercises, all presented
within the Boxer programming environment. As shown in Table Chapter 10. -1, the tasks were
divided into 4 types. First, the students were asked to complete a number of training exercises
that mixed instruction in Boxer programming with more specific training in the creation of
motion simulations. This instruction began with an introduction to sprite graphics in boxer—just
making sprites move around the display using the fd and rt commands. Early on, students were
also taught to use the repeat command and to use do-it boxes to make separate procedures.
Training Tasks and
Exercises
Main Body
“Quick” Tasks
Cross Training
3.5 hours
5.0 hours
0.5 hours
1.5 hours
Table Chapter 10. -1. The four main types of tasks in order of execution.
342
In the next phase of the training, students were given their first simulation tasks. This included
simulating the motion of a car that accelerates as it moves in a straight line. In addition, students
were asked to simulate the motion of a ball that is dropped and a ball that is tossed straight up.
Since they had not yet been instructed in the use of variables in Boxer, the students were limited
to using a string of forward commands to simulate these motions. An example is Amy and
Clem’s first simulation of a dropped ball, which I presented in Figure 8-6.
Following these first simulation tasks, the students received some additional Boxer training,
mainly having to do with the use of variables. Then they were given some tasks which required
them to use variables to simulate motions, once again involving the motion of an accelerating car.
Finally, there was a brief tutorial on the Tick Model which included examples of its use. All
together, the training took about 3 or 4 hours for students to complete, which generally accounted
for the first 2 full sessions.
The main body of tasks, which immediately followed these training tasks and exercises, took
about 5 hours for students to complete and constituted roughly half of the total time they spent
in the study. The tasks in the main body were, at least in their statement, quite simple. Each of
these tasks just briefly described a motion and asked students to create a simulation of that
motion. For example, as I described earlier, one task asked students to simulate the motion of a
mass oscillating back and forth at the end of a spring:
In this task, you must imagine that the block shown is attached to a spring so that it oscillates back and
forth across the surface of a frictionless table.
(a) Make a simulation of this motion.
(b) If the force, mass, and spring constant don't appear in your simulation, then add them.
This task, though simple in specification, typically took students about 45 minutes to complete.
The study concluded with two more brief segments, the “Quick” tasks and “Cross Training.”
Recall that there were counterparts for each of these segments in the algebra study, but those tasks
were not analyzed and I did not comment on them very frequently. In the programming Quick
tasks, like their algebra counterparts, students were asked a number of short questions that were
designed to be answered briefly. For example, one question asked:
Some students say that the tick model is "obvious." What do you think?
The purpose of these questions was to allow a forum for an open discussion of some issues,
without disrupting the flow of the simulation tasks.
343
Finally, the purpose of the Cross Training segment was to engage students in a slightly more
algebra-like practice of programming-physics. This meant adapting some of the programs they
had already written to obtain real numerical results. For example, one question asked students to
use their Toss simulation to find out how high a ball would go if it was thrown straight up at a
speed of 100 miles per hour. This type of task proved to be quite hard for students, in part
because of the difficulty of associating units with the numbers that appear in a simulation
program. Like the algebra Cross Training, the programming Cross Training tasks were not
analyzed and will not be discussed further in this document. The instructions for all of the tasks in
the Main Body, the Quick Tasks, and the Cross Training tasks are reproduced in Appendix C.
1. Shoved Block
Imagine that a block is resting on a table with friction. Program a simulation for the
case in which the block is given a short, hard shove.
How does the motion differ if the block is heavier? Can you show this in your
simulation? Modify your simulation if necessary.
2. Dropped Ball
The idea here is to make a realistic simulation of the motion of a ball that is dropped
from someone's hand.
Using the tick model, redo your simulation of a dropped ball.
3. Tossed Ball
Now we want to imagine a situation in which a ball is tossed exactly straight up and
then falls back down to where it started.
Using the tick model, redo your simulation of a ball tossed straight up.
4. Air Resistance
(a) Make a new simulation of the DROP that includes air resistance. (Hint: Air
resistance is bigger if you're moving faster.)
(b) Try running your simulation for a very long time. (The ball might go off the bottom
and come back in the top, that's okay.) If you've done you're simulation correctly, the
speed of the ball should eventually reach a constant value. (The "terminal velocity.")
5. More Air
Resistance
Your simulation of a ball dropped with air resistance is included in this box.
(a) How does the terminal velocity of a dropped ball depend upon the mass of the
ball? For example, what happens to the terminal velocity if the mass doubles? Can
you show this with your simulation? Modify your simulation if necessary.
(b) Suppose there was a wind blowing straight up. How would your simulation and the
terminal velocity be different?
(c) What if the wind is blowing straight down?
6. Mass on a
Spring
In this task, you must imagine that the block shown is attached to a spring so that it
oscillates back and forth across the surface of a frictionless table.
(a) Make a simulation of this motion.
(b) If the force, mass, and spring constant don't appear in your simulation, then add
them.
Table Chapter 10. -2. The 6 tasks selected for systematic analysis.
344
Since it would have been prohibitively time-consuming to analyze the videotape data
associated with all of the tasks, only a subset was included in the systematic analysis. In narrowing
the tasks for analysis, my goal was to have a selection that could be seen as representing a relatively
consistent practice of programming-physics. This meant eliminating the Quick and Cross
Training tasks, since they deviated considerably from what I took to be the core of the practice.
Furthermore, it made sense to eliminate the training tasks, since the students were just being
introduced to programming-physics during this phase of their work. This left the Main Body as a
possible source of tasks and, indeed, the Main Body contains a collection of tasks that are of
roughly the same sort. Table Chapter 10. -2 shows the 6 tasks that were selected for inclusion in
the systematic analysis.
It is worth a moment to explore the correspondence between the programming and algebra
tasks included in the systematic analysis (refer to Table Chapter 10. -3). Students in both
conditions were given a version of the Shoved Block task and were asked to deal with the issue of
how the motion differs if the mass is increased. There were also tasks included in both analyses
pertaining to dropped and tossed balls. In the algebra condition, students were asked how high a
ball travels when it is tossed up with a given velocity. The programming students were asked to
simulate two related motions: a ball that is tossed straight up and a ball that is just dropped. In
addition, both conditions included two tasks pertaining to motion with air resistance, as well as a
task involving a mass on a spring.
Algebra
Programming
Shoved Block
Shoved Block
Vertical Pitch
Dropped Ball
Tossed Ball
Air Resistance
Air Resistance
Air Resistance with Wind
More Air Resistance
Mass on a Spring
Mass on a Spring
Stranded Skater
Buoyant Cube
Running in the Rain
Table Chapter 10. -3. The correspondence between the algebra and programming tasks included in the
systematic analysis.
However, there were no programming counterparts for the last three algebra-physics tasks
listed in Table Chapter 10. -3. This is partly because it is difficult to construct programming
analogs for these tasks. To take a case in point, it is not easy to use programming to deal with
345
situations, such as the one in the Stranded Skater problem, that are usually handled using
conservation principles. In addition, it was also necessary to include fewer programming tasks
simply because these tasks took longer for the students to complete.
One final comment about Table Chapter 10. -3: The correspondences indicated in this table
are, in some ways, really quite weak. Although the corresponding tasks pertain to physical
phenomena that are similar in some respects, there are many ways in which the tasks and
phenomena under study are extremely different. For example, in the algebra version of the Mass
on a Spring task, the mass hangs motionless at the end of the spring. In contrast, in the
programming version of this task, the mass oscillates horizontally at the end of a spring. I will say
more about these differences in Chapter 11.
Thus, although the apparent correspondences in Table Chapter 10. -3 are indicative of a real
effort to maintain parallels, we must not forget that this simple table hides strong differences
between the algebra and programming tasks. Furthermore, there is really no way that I could have
forced the tasks in each practice to be identical. Programming-physics and algebra-physics tasks
are, by their nature, different and any attempt to alleviate these differences would alter the very
practices I am trying to study. I will discuss this issue further when I discuss the “matching of
conditions” at the end of this chapter.
Details of the experimental sessions
All of the programming sessions were conducted in a laboratory setting. Each pair sat at a
single computer and, typically, one student operated the mouse while the other typed at the
keyboard. Two cameras were used to capture student activity. One camera, positioned behind the
students, was raised high so that it could see over the students’ heads and record the computer
display. The second camera was positioned off to the side so that the students were visible in
profile. The resulting two images were mixed together into a single videotape image for later
viewing, with the profile image in a small inset in one corner. In this study and my pilot studies, I
found that this was an adequate technique for recording pairs of students while they worked at a
computer.
During the experimental sessions, students essentially progressed through a series of Boxer
boxes, each with its own written instructions. The students would open the box, read the
instructions, program the simulation, and then explain their work to me when they were done. I
tried not to interrupt the students very often while they were in the middle of programming a
346
simulation; for the most part, I saved my questions until the simulation was completed. My
strategy in this regard was similar to that employed for the algebra study, with a single exception:
I was quick to help students over certain kinds of programming-related difficulties. This included
helping them deal with Boxer bugs and system crashes, and also intervening when they were
stymied by some of the nitty-gritty of programming syntax.
A typical session lasted between 1.5 and 2 hours and, as shown in Table Chapter 10. -5, all of
the pairs except for Greg and Fred came to the laboratory on 6 separate occasions. Although Greg
and Fred required 8 sessions to complete the tasks, I do not believe that this is indicative of any
systematic difference in their working style. Actually, across all of the pairs, there was fairly wide
variation in the time spent on individual tasks. For example, the time spent on the Air Resistance
task ranged between 25 minutes and 1.5 hours, with Greg and Fred taking about 50 minutes. This
variation just happened to balance out for the other pairs, so that the same number of sessions was
required for each.
# of Sessions
Total Hours
Analyzed Hours
Adam&Jerry
Ned&David
Amy&Clem
Tim&Steve
Greg&Fred
6
10.5
3.0
6
9.0
3.0
6
9.5
2.0
6
11.0
3.5
8
13.0
4.5
Totals
32
53.0
16.0
Table Chapter 10. -5. The time students spent in the experimental sessions.
The total time spent by students in the experimental sessions was about 53 hours, with roughly
18 hours spent on the training exercises, 25 hours on the Main Body, and the remaining 10 on the
Quick tasks and Cross Training. When these 53 hours are narrowed to include only work on the
tasks included in the systematic analysis, the result is about 16 hours of videotape. This is
somewhat more than the 11.5 hours that were analyzed in the algebra portion of the study.
Analysis Techniques
The next step here is to describe how, precisely, I went about analyzing the collection of
videotapes described above. As was the case in the algebra study, the analysis involved an iterative
coding and recoding of transcripts. In overview, I began by first viewing and transcribing all of the
videotapes corresponding to the Main Body of tasks—about 25 hours of videotape. Then, one at a
time, I selected individual episodes for analysis. (Recall that an “episode” is one pair’s work on one
task.) For each of these episodes, I polished the transcript, coded the episode, and refined my
347
coding scheme as necessary. After 5 episodes were coded in this manner, I felt that the coding
scheme was stable enough to apply to a large portion of the data. At this point, I proceeded to
code the 30 episodes corresponding to student work on the six tasks listed in Table Chapter 10. 2. I began by viewing each of these episodes again and polishing all of the transcripts. Then three
full iterations of coding these 30 episodes and adjusting the coding scheme were required before a
stable coding was reached.
Now, in the remainder of this section, I will fill in the details of how the analysis was
conducted.
Transcription
The analysis of every episode was based on a transcription of that episode. For the
programming data corpus, creating transcripts that contained all of the needed detail required the
development of some specialized techniques. Not only did I need to record everything that the
students said, I needed to have information concerning what appeared on the computer display at
any time, and I needed to record what students were pointing to as they spoke. This was
accomplished, as it was in the algebra study, by building a transcript that consisted of two
components. In the first component, I recorded all student verbalizations and made brief notes
concerning changes made to the program. In the second, supplementary component, I kept track
of the successive states through which the program passed. To facilitate this tracking of the
program, I created a set of tools in the Boxer environment so that, as I read through the written
component of a transcript, I could essentially “play back” the series of changes that were made to
a program.
348
POS
VEL
ACC
539.000
98.000
9.800
Data
Data
Data
TICK
change vel
DROP2
setup
change
change
change
repeat
pos 0
vel 0
acc 9.8
10 tick
Doit
vel + acc
Doit
change pos
pos + vel
Doit
fd vel
dot
Doit
Data
Data
Figure Chapter 10. -6. Tim and Steve's simulation of a dropped ball.
My approach to the recording of gestures was analogous to that used for recording gestures in
the algebra study; where possible, I treated students gestures as indicating a particular item in the
display, such as a programming statement or variable. For various practical reasons, this was
slightly more difficult than in the algebra study. The most important difference was the simple
fact that the characters on the computer display were much smaller than the symbols written on
the whiteboard by the algebra subjects. In fact, a student’s fingertip could often span several
characters on the display! However, this difficulty was somewhat moderated by the fact that the
programming students did not always gesture with their hands; they frequently used the mouse
pointer or the typing cursor to indicate a specific location on the screen (refer to the notations
described in Appendix A). For example, in the following passage, Tim explains the simulation of
the dropped ball shown in Figure Chapter 10. -6, using the mouse pointer to indicate various
items on the display. Here, the “g.m” notation indicates a gesture employing the mouse pointer.
Tim
Um, we just did the regular setup.
Bruce
Hm-mm.
Tim
To take us up there [g.m top of dot trail in graphics box] and then set position
and velocity to zero. [g.m change pos 0; change vel 0] And then we just have
acceleration, nine point eight. [g.m change acc 9.8] Which is so it’s gravity.
Bruce
Hm-mm.
Tim
And then we just repeat it ten times it just (0.5) falls. [g.m along dot trail in
graphics box]
In this passage, Tim points back and forth between the trail of dots in the graphics box and
various lines in the program.
349
Actually, the students were quite expressive in their use of the mouse pointer. They did not
only “point” with it, they employed a range of different gestures. For example, a circling motion
was often used to more clearly include and exclude portions of the display from the target of the
gesture. And sometimes students acted out motions by moving the mouse pointer over a graphics
box. Of course, the range of possible gestures was more limited than what students could do with
their hands, but this was still a quite powerful means of gesturing and one that is well-tuned to a
computer display.
The base-level coding scheme
Once an episode was transcribed, the goal was ultimately to code the resulting transcript in
terms of forms and devices. However, as in the algebra analysis, an intervening step was required
before the jump to forms and devices could be made. The problem is that the raw transcripts do
not tell us what events within the transcripts need to be coded. Thus, before coding the forms and
devices, there needed to be a prior step in which the relevant events were identified. In addition,
this preliminary coding provides some information concerning the character of the data corpus.
In general, the base-level analysis was geared toward identifying two types of events which
would then be coded in terms of forms and devices:
(1) A programming expression is constructed by a student.
(1) A student interprets an expression.
As for algebra, the identification of these two types of events requires two different kinds of
analyses. First, events of type (1) require an expression-centered analysis that focuses on coding
pieces of a program. And events of type (2) require an utterance-centered analysis. This latter
analysis takes utterances from the transcript and decides if they qualify as interpretations. I will
now, in turn, describe each of these two kinds of analyses.
Expression-centered coding: Identifying constructed expressions
The first part of the base-level coding is the expression-centered analysis. To begin, I want to
remind the reader of how this analysis worked in the algebra study. The idea there was to look at
every expression that a student wrote and decide if it was remembered, derived, or constructed.
However, before coding the expressions in this manner I had to decide what counted as “an
expression.” In the end, I decided to treat any strings of more than one symbol as an expression.
350
This included equations as we usually understand them, as well as strings of symbols that did not
include an equal sign.
Thus, the very first step in adapting my analysis scheme to programming-physics is to decide
what counts as “an expression” in programming. It turns out that, although many adjustments will
be necessary to apply my analysis techniques to the programming data, this first issue is perhaps
the trickiest. In the case of algebra, I treated a page of equations as a set of individual expressions
to be coded separately. However, in Chapter 9 I argued that, unlike in algebra, programmingphysics involves symbolic forms that span statements. Furthermore, even more than in algebra, it
is clearly the case that a program is more than just a list of independent statements.
Nonetheless, it turned out that an analysis based primarily around individual programming
statements worked quite well for most circumstances. I looked at each individual programming
statement, such as change vel 25 or change acc force / mass and coded it as remembered, derived,
or constructed, just as I did for algebraic expressions. The only exception to this rule was when
statements were moved or copied as blocks. These cases were treated as single events to be coded.
So, the basic move here is to treat a program very much like I treated a blackboard full of
equations, coding each individual statement. However, as we will see, there is one very important
difference that is not captured by this gloss. In coding a statement as remembered, derived, or
constructed, the location of the statement within the program must be taken into account. Thus,
unlike algebra, I treated a program as a separate but ordered collection of statements.
In a sense, this approach is only an approximate technique for identifying expression-writing
events in programming. A more precise approach would be to identify clues in student activity
that bound expression-writing events. Nonetheless, I believe that the statement-oriented approach
I employed is a reasonable one. We have already seen that there is a great deal of interesting action
at the level of individual lines. Furthermore, we should realize that any sort of practical breakdown
is likely to be somewhat arbitrary. All I am really looking for here is a means of pinning codes to
the data in order to force some type of systematicity on the analysis, and to obtain some rough
measures of the frequency of various forms and devices. This statement-based approach does this
job in a way that is feasible to implement.
There are a few other fine points to be mentioned here concerning what should be counted as
an expression. First, it is sometimes the case that programming statements serve as arguments to
other programming statements. The primary case that appeared in my data corpus involved the
351
repeat command. For example, consider this line taken from Greg and Fred’s Air Resistance
simulation:
repeat 200
change force mass * acc - vel
change acc force / mass
tick
Doit
In such cases, the repeat statement, treated as a whole, was considered to be an expression. In
addition, each individual statement inside the repeated procedure was also treated as an
expression to be coded.
Some other fine points have to do with modifications. Just as in algebra, when a statement was
modified, the resulting statement was treated as a new expression to be coded. Note that this
includes the moving of a statement to a new location within a program.
In addition, there are a few types of modifications and additions that I chose not to count as
expression-writing events:
•
The correction of “typos” or other simple errors in programming syntax.
•
Adjustments to numerical values that appear within a program. In many cases,
however, there were interpretive utterances associated with these adjustments (which
were coded).
•
The placing of the Tick Model. It turned out that the Tick Model was created
automatically as a first step in every student simulation.
With these last few exceptions, I have laid out the basic guidelines I used in identifying
expressions. Once the expressions were identified each one was coded as remembered, derived, or
constructed. Again, the analysis strategies here are very similar to those of the algebra study, but a
few comments on each of these categories are merited.
Remembered
A remembered expression was one that a student appeared to have just written “from
memory.” The evidence for coding an expression as remembered was essentially the same as that
used in the algebra study. If a complicated expression was produced without substantial effort or
comment, this constituted evidence that an expression was remembered. Furthermore, if I
recognized the expression as one that students were familiar with, this added weight to the
coding.
352
There are just a few programming-specific comments to be made concerning this category.
First, in some cases, students used the “copy and paste” feature of the programming environment
to duplicate statements. The resulting expressions were coded as remembered.
Second, lines that are essentially transcribed versions of algebraic expressions were treated as
remembered, with a single exception that I will discuss below. For example, if a student wrote
change force
- k * pos
Doit
this was, in most cases, understood to be a transcribed version of the remembered equation F=-kx.
In addition, in some instances, students performed extended derivations on a sheet of paper and
then transcribed the result into their programs. These transcribed statements were also coded as
remembered.
The rationale for this is that, first, these transcribed statements do not require active
construction. The transcribing of algebraic expressions into programming is, in many cases,
relatively straightforward. In addition, I wanted to reserve the derived coding for expressions that
were written based on manipulations of existing programming statements (see below). Thus, in
coding these transcribed statements as remembered, I am using this as something of a catchall
category for some kinds of expressions that are not derived and do not involve active construction.
(Keep in mind that the primary goal of the base-level analysis is to identify constructed
expressions.)
Finally, some programming statements are no more than a one word call to a procedure such
as tick. These statements were usually coded as remembered since they required no active
construction on the part of the student. However, in some cases the placement of these simple
statements constituted a novel construction. I will say more about this in a moment.
Derived
In the algebra study, an expression was coded as derived when it was written by applying rulesof-manipulation to existing expressions. Since there are no formal rules-of-manipulation associated
with programming languages, it may seem that this coding should not apply here. However,
although it is somewhat rare, there are some programming events which I believe are aptly treated
as derived.
The most clear instance of derivation in programming is when algebraic manipulations are
applied to algebra-like expressions that appear within a programming statement. To illustrate,
353
consider the following line which appeared in an early version of Ned and David’s Shoved Block
simulation:
repeat
30 /
tick
-1 * acc
Doit
Doit
Later in their work, the expression for the number of repeats was simplified, an event that was
coded as derived:
repeat
-30 / acc
tick
Doit
Furthermore, there were a number of very simple, almost trivial, changes that I treated as
derivations. One example of such an event was when the name of a variable was changed to
something that a student deemed more appropriate. In addition, a derived coding was given when
a numerical value that appeared in a program was replaced by a variable, as when
change force
-.3 * pos
Doit
was changed to
change force
- k * pos
Doit
Finally, and most interestingly, there were a few more dramatic instances of programmingspecific modifications in which programs were replaced by equivalent programs. For example,
Tim and Steve’s work on the Toss simulation eventually reached the following stage:
toss
drop
setup
change
change
change
repeat
drop
change acc -9.8
repeat 9 tick
pos 0
vel 98
acc -9.8
10 tick
Doit
Doit
I will tell this story in more detail later. For now, I just want to mention that Tim and Steve
eventually realized that the above simulation could be replaced with the following equivalent
program:
354
toss
setup
change
change
change
repeat
pos 0
vel 98
acc -9.8
19 tick
Doit
Instances of this sort, in which programming statements were replaced by simplified but
equivalent alternatives, were coded as derived.
Constructed
Except for a few other cases that I will discuss in a moment, all of the remaining expressionwriting events fall into this last category. Constructed expressions were expressions that were
invented on the spot by students. I have argued that, in some cases, these constructions are much
like the expressions constructed by the algebra subjects. To illustrate this, in Chapter 9 I described
Tim and Steve’s construction of an expression for the force of air resistance
change Res
vel * .1
Doit
and Greg and Fred’s construction of an expression for the acceleration in the Shoved Block task:
change acc -.08 - mass
However, there is one very important, programming-specific issue to be dealt with here. Even
when a programming statement is a precise duplicate of a statement from an earlier task, the
choice of its location within the program may constitute an act of construction of the sort under
investigation here. It was essential to keep track of these events since, like all these construction
events, they must be accounted for in terms of forms and devices.
With regard to the ordering of statements, students had less experience to draw on, since their
experience from algebra-physics was not as directly relevant. Really, the arrangement of
statements in a program could only be coded as remembered if it followed the ordering of
statements in one of the programs that the students wrote earlier, or that they were shown during
their training. Nonetheless, it turned out that, in the vast majority of cases, the location of
statements could be treated as remembered. This is because almost all of the students’ programs
followed the same structure—the initialize-and-then-repeat structure that I have mentioned on
355
several occasions. It is primarily when students departed from or added to this basic structure that
I treated the location as constructed.
I have one last comment on the constructed category. In general, when work was done on a
sheet of paper and then an expression was transcribed into a program, I coded the expression
writing-event as remembered. There was an important exception to this rule however. If a single
expression was constructed on a sheet of paper and then directly transcribed, without any prior
paper-based manipulations, I coded the expression as constructed. The rationale for this is that the
forms that could be seen in the programming expression are basically the same as those in the
algebra version of the expression. And, in fact, it was sometimes hard to tell whether what the
students wrote on the sheet of paper was supposed to be a programming statement or an algebraic
expression.
Marginal cases
As in algebra, there were a few marginal cases that were treated separately. In general, these are
expressions that, while not literally remembered, were also not worthy of being called “novel
constructions.” The two marginal cases here are completely analogous to their algebra-physics
counterparts and the reader is referred to Chapter 6 for a more complete discussion. First, as in
algebra, some expressions were coded as principle-based. This coding was much more rare in
programming-physics, however.
As we will see, codings of the second marginal case—value-assignments—were extremely
common in programming. Recall that in algebra-physics, value assignment statements were
expressions in which a single quantity is equated to a numerical value, such as vo =0. Similarly, value
assignment statements in programming set a variable to a specific value, as in change acc -9.8.
Although these marginal cases were kept track of separately in the base-level analysis, they
were not treated further in later stages since these events do not involve the construction of an
expression.
Summary of expression-centered coding
Table Chapter 10. -7 provides an overview of the expression-centered component of the baselevel analysis. In summary, after transcribing an episode, I identified the expressions that were
written in the episode. Each of the identified expressions was then coded as remembered, derived,
constructed or as one of the marginal cases. The main function of this component of the base-level
analysis was to identify constructed expressions, so that these could later be coded in terms of
356
forms and devices. Any individual programming statement may be considered constructed because
of its content or its location within the program.
Expression-Centered
Coding Scheme
Remembered
Derived
Constructed
Marginal Cases:
Value Assignment
Principle-Based
Combinations
Table Chapter 10. -7. Summary of categories for coding programming expressions in the base-level analysis.
Utterance-centered coding: Identifying interpretations of equations
The second component of the base-level analysis was the utterance-centered coding. The
purpose of the utterance centered coding was to identify all of the utterances in an episode in
which a student “interpreted” all or part of a program. In some respects, this component of the
base-level analysis was easier to adapt for use in programming than was the expression-centered
coding. The expression-centered coding of the programming corpus was complicated by the fact
that it was difficult to know how to segment a program into “expressions.” In contrast, in the
utterance-centered coding, the goal is to pick out “utterances.” While the segmenting of a
programming transcript into utterances can certainly be a difficult task, it is largely the same task
in programming-physics that it was in algebra-physics.
In my analysis of the algebra corpus, my tactic for capturing all of the interpretive utterances
in an episode was to cast a wide net. Here I will use the same approach. Operationally, this means
taking every statement that referred to or incorporated any part of an expression. Each of the
following examples includes an interpretive utterance. Note how, in each sample passage, the
students point to some part of the program as they speak.
357
shove
setup
change
change
change
change
pos 0
vel 12
mass .08
acc
- mass
Doit
repeat
150 - mass * 2
tick
Doit
Doit
Greg
… So that's why we thought that this repeat [g. 150 - mass * 2] is kind of like a
distance. So we tried to make it less than our initial, um, one fifty. [g. 150] So we
subtracted the mass [g. - mass * 2] a factor of the mass.
_______________
drop2
setup
change
change
change
repeat
pos 0
vel 0
acc 9.8
10 tick
Doit
Tim
Um, we just did the regular setup.
Bruce
Hm-mm.
Tim
To take us up there [g.m top of path] and then set position and velocity to zero. [g.m
change pos 0; change vel 0] And then we just have acceleration, nine point eight.
[g.m change acc 9.8] Which is so it’s gravity.
Bruce
Hm-mm.
Tim
And then we just repeat it ten times it just (0.5) falls. [g.m along path in g-box]
_______________
change acc
force - drag
/ mass
Doit
Doit
change drag
vel / 1.001
Doit
David
Um, yeah, because at, at the terminal velocity, this value for vel [g. vel in change
drag line], the drag will equal the force. [g. drag, force, drag in change acc
line] So then you have no - no more acceleration. And so as - as uh - as velocity
approaches this value [g. 1.001] times the force, [g. force],
Bruce
Hm-mm.
David
um, the acceleration decreases. And so, as it gets close, the acceleration decreases
slower and slower and slower until finally it reaches zero.
A sample base-level coding
Before moving on to the next phase of the analysis, I want to go through the base-level analysis
of one very brief episode. This episode is Tim and Steve’s work on the Drop simulation, and it
was one of the shortest episodes in my corpus, lasting only about 6 minutes. Tim and Steve’s
358
finished simulation for this task was shown earlier in Figure Chapter 10. -6. This episode was also
discussed in Chapter 8.
The entire transcript for this episode is provided in Table Chapter 10. -9, annotated with the
appropriate base-level codings. Reading through this transcript, we see that the very first thing that
these students did was to press the key that creates the Tick Model box. Students nearly always
began work on a task by placing a Tick Model box in their working environment and, as I
specified above, this act was not coded in any way.
After placing the Tick Model box, Tim and Steve proceeded quickly to write all of the lines in
the drop2 procedure. The first statement in this procedure, setup, is a simple call to a procedure
that takes no arguments. Such one word calls are coded as remembered. Three initialization lines
follow this first statement, each of which is coded as a value assignment. Finally, the repeat
command at the end of the procedure is coded as remembered.
So, none of the expression-writing events in this episode were coded as constructed. The fact
that all of these lines were written quickly and with little discussion is good justification for these
codings. Furthermore, this program follows the initialize-and-then-repeat structure, so there is no
need to code any of the statements as constructed due to their location within the program.
Up to this point in the transcript, the episode is not very interesting, at least from the point of
view of forms and devices. In fact, Tim and Steve’s statements are little more than some partial
narration of what they were typing into the program. When Tim and Steve tried their program,
however, things got a little bit more exciting. As I described in Chapter 8, the sprite initially went
in the wrong direction—it moved upward instead of falling downward. To correct this problem,
the students had to initialize their acceleration variable to a negative value, rather than a positive
value. According to the specification layed out above, this modification does not get coded in any
way because Tim and Steve have only altered a numerical value that appears in a program.
However, as is common, there was an interpretive utterance associated with this modification.
After this modification, Tim and Steve were satisfied with their program and they explained
their simulation to me. The episode then concluded with two final interpretive utterances.
359
[They make a tick model box and then look at their old work
silently.]
Tim
Tim
Steve
Tim
Steve
Tim
Steve
Tim
Tim
We need a drop two.
So. We should do setup first, right? [w. setup] Then we wanna-,,
(10.0) Set the position, [g. pos:()] Set the velocity. [g. vel:()]
Acceleration we can set to [g.m acc:()]
nine point eight?
To negative nine point eight I guess, huh.
Yeah.
That’s it, isn’t it? Then the tick? (Right?)
Yeah.
[w. change pos 0; change vel 0; change acc -9.8. ]
Repeat. [w. repeat |] How many times? (1.0) … Let’s say ten.
[w.mod repeat 10 tick] Tick.
tick model box
Uncoded.
setup
Remembered expression.
change pos 0
change vel 0
change acc -9.8
Value assignment.
repeat 10 tick
Remembered expression.
[Tim runs the simulation and the ball goes up.]
Tim
It went the wrong way. [laughs]
[T runs it a few more times.]
Steve Yeah, cause - oh, this is subtracting it. [g. change vel
[vel + acc]]
Tim
So this should be a positive in here. [w.mod change acc 9.8]
That’s kinda weird.
[They run the program a few times and it looks okay. They call me over to
see.]
Bruce You wanna just explain what it does? How it works?
Tim
Um, we just did the regular setup.
Bruce Hm-mm.
Tim
To take us up there [g.m top of path] and then set
position and velocity to zero. [g.m change pos 0;
change vel 0] And then we just have acceleration,
nine point eight. [g.m change acc 9.8] Which is so it’s
gravity.
Bruce Hm-mm.
Tim
And then we just repeat it ten times it just (0.5) falls.
[g.m along path in g-box]
Bruce Hm-mm.
Tim
So, every time - every tick it gains nine point eight.
The acceleration increases the velocity by nine point
eight.
Bruce Hm-mm. Okay. Sounds good.
“cause…is subtracting
it.”
Interpretive utterance.
change acc 9.8
Uncoded.
“Um, we just did the
regular setup.…repeat
it ten times it just falls.”
Interpretive utterance.
“…every tick it gains
nine point eight …”
Interpretive utterance.
Table Chapter 10. -9. Complete transcript of Tim and Steve’s work on the Drop simulation. Base-level
codings are given in the right-hand column.
360
Coding forms and devices
The base-level analysis identified two types of important events in the data, events in which an
expression was constructed and events in which students produced an interpretive utterance. After
the base-level analysis was complete, the next task was to code each of these important events in
terms of forms and devices. Just as in the algebra-physics analysis, the coding of events in terms of
forms and the coding in terms of devices each required me to perform two tasks simultaneously:
(1) define the lists of forms and devices and (2) code each individual event in terms of the forms
and devices that are on these lists.
Principles for defining the set of forms
Principles for defining the set of devices
Coverage of the data corpus
Coverage of the data corpus
Overall parsimony of the set
Overall parsimony of the set
Inclusion of a symbol pattern
Close relation to the structure of interpretive
utterances
Symbol patterns only preserve distinctions
consequential for interpretive behavior
Restriction to reliably codable distinctions and
allowance for ambiguity
Impenetrability
Continuity with prior knowledge
Continuity with prior knowledge
Approximation to expertise
Approximation to expertise
Table Chapter 10. -11. Summary of the heuristic principles used to guide in defining the lists of forms and
devices.
Evidence for recognizing forms in specific
events
Evidence for recognizing devices in specific
events
Corresponding symbol pattern
Specific verbal clues
Specific verbal clues
Corresponding inferences
Related graphical concomitants
Related graphical concomitants
Corresponding symbolic conventions
Global as well as local evidence
Global as well as local evidence
Table Chapter 10. -12. Summary of the types of evidence used to recognize forms and devices in particular
events.
Although the resulting lists were different, the procedure employed here was essentially the
same as that used for the analysis of the algebra corpus. I iteratively coded and recoded the data
until the lists of forms and devices and individual codings were stable. This process began with
lists of forms and devices derived from a pilot study. Then I coded 5 focus episodes, refining my
categories as I went. Finally, I took three full passes at all of the episodes included in the
361
systematic analysis. As for the algebra analysis, this whole process relied on a set of heuristic
principles and some specific rules for recognizing individual forms and devices. The principles
used in defining the lists of forms and devices are summarized in Table Chapter 10. -11. In
addition, Table Chapter 10. -12 has a summary of the types of evidence used for recognizing
forms and devices in specific events. The material summarized in these tables is described in more
detail in Chapter 6, and the specific rules for recognizing individual forms and devices are given in
Appendices D and E.
On the Òsubstitution and execution machineryÓ
Before proceeding, I need to add one small addendum concerning the coding of forms. Recall
that in the analysis of the algebra corpus, some of the events identified by the base-level analysis
could not be coded in terms of forms. In some cases, I argued, another class of resources did the
work that I might otherwise attribute to forms. The idea was that students come to their learning
of physics with a number of well-developed skills relating to the use of symbolic expressions.
Among these skills are the ability to manipulate expressions to derive new expressions, and the
ability to substitute numbers into expressions. I argued that some interpretations given by students
seemed to draw on elements of this “substitution machinery” rather than on symbolic forms.
This same substitution machinery can play a role in producing interpretations of programming
expressions, especially in the case of the “algebra-like” programming statements discussed in
Chapter 9. Furthermore, in the case of programming, this category needs to be extended to
include some general “execution machinery.” The point is that the students in my study quickly
developed the ability to “mentally” execute certain programming instructions. For example,
students looking at the simple sprite graphics program:
fd 50
rt 90
fd 30
…
could mentally step through this program, showing with their hands how the sprite would move as
each instruction was executed, first moving forward, then turning right, and then moving forward
again. Such behavior does not require inferences based on symbolic forms. Instead, it draws on
specific knowledge concerning each statement that appears in the program. Thus, events of this
sort are not assigned any coding in terms of forms.
362
Coding the sample episode
Before presenting the results of the analysis, I want to briefly show how the sample episode
discussed above was coded in terms of forms and devices. Since there were no constructed
expressions in the sample episode, only the three interpretive utterances need to be coded. In the
first of these interpretive utterances, Steve points to the line change vel vel + acc and says: “this is
subtracting it.” I coded this utterance as involving the BASE ±CHANGE form, with the idea that the
acc in vel+acc is seen as applying a change to the base value in vel. The device associated with this
utterance is RECURRENCE since the “subtracting” described by Steve transpires on each tick of the
program:
Steve
Yeah, cause - oh, this is subtracting it. [g. change vel
[vel + acc]]
Tim
So this should be a positive in here. [w.mod change acc 9.8]
That’s kinda weird.
“cause…this is
subtracting it.”
Form: BASE±CHANGE
Device: RECURRENCE
The evidence for these codings here are actually quite weak; in fact, this is a case of especially
weak local evidence. In truth, there are many other ways of interpreting Steve’s statement. This is
where the “Global as well as local evidence” heuristic comes in. Even though there is not much
evidence in this specific event, I saw many other examples where more extensive, but similar,
interpretations were given of this same programming statement. Thus, I was able to use those
other examples to help in coding this event. In fact, as we shall see in a moment, the last
interpretation in this very episode can help in this manner.
Both the second and third interpretive utterances in this excerpt showed up as examples in
Chapter 9, so we have a head start on analyzing them. The first of these was a TRACING
interpretation:
363
Tim
Bruce
Tim
To take us up there [g.m top of path] and then set
position and velocity to zero. [g.m change pos 0;
change vel 0] And then we just have acceleration,
nine point eight. [g.m change acc 9.8] Which is so it’s
gravity.
Hm-mm.
And then we just repeat it ten times it just (0.5) falls.
[g.m along path in g-box]
“Um, we just did the
regular setup.…repeat
it ten times it just falls.”
Form: SETUP-LOOP
Device: TRACING
To see that the device here is TRACING , note that Tim’s utterance follows the structure of the
program, with statements separated by the phrase “and then.” And, in Chapter 9, I argued that
the
SETUP -LOOP
form is behind this interpretation, since Tim is associating portions of the
program with two sub-processes, one that prepares for the motion, and one in which the object
“just falls.”
The episode concluded with this third and final interpretive utterance:
So, every time - every tick it gains nine point eight.
The acceleration increases the velocity by nine point
eight.
Tim
“…every tick it gains
nine point eight …”
Form: BASE+CHANGE
Device: RECURRENCE
As I argued in Chapter 9, the device here is RECURRENCE, since Tim is announcing what transpires
on each Tick of the program. The form associated with this event is BASE ±CHANGE . As in the first
interpretive utterance in this episode, the original velocity in the line change vel vel + acc is seen as
a base onto which the acceleration is added. Tim’s use of the word “gains” here is evidence for this
coding.
Summary of Results
Base-level results
I am now ready to present the results of the systematic analysis of the programming corpus.
As I said at the start of this chapter, I will provide some perspective by comparing these results to
those obtained for algebra-physics. However, I will not go much beyond a simple comparison of
the raw numerical values. Instead, my careful thinking about the implications of any differences
will be reserved for the next chapter.
First, I will summarize the outcome of the base-level analysis. As shown in Table Chapter 10. 13, the expression-centered component of the base-level analysis identified a total of 246
364
expression-writing events. Of these 246 events, 70 involved the modification of existing
expressions, and 13 were events in which an existing statement was moved to a different location
in the program.
Expressions “from scratch”
Modifications
Moves
Total expression-writing events
163
70
13
246
Table Chapter 10. -13. Expressions found in the base-level analysis.
Table Chapter 10. -15 contains the results of coding these 246 events as remembered, derived,
constructed, or as belonging to one of the marginal cases. Note that the codings do not total to
246 since, as in the algebra analysis, events may receive multiple codings.
Remembered
Derived
Constructed
Marginal Cases
Total
Codings
With Other
Codings
93
12
61
88
11
2
12
5
Table Chapter 10. -15. Base-level coding of expressions.
Since the purpose of this component of the base-level analysis is to identify constructed
expressions, the highlighted row in Table Chapter 10. -15 merits special attention. It is important
to keep in mind that statements coded as constructed may have received this coding because of the
location of the statement, because of how the statement itself was constructed, or for both reasons.
Of the 61 events coded as constructed, 24 were coded this way because of location, 25 because of
the statement itself, and 2 because of both.
Table Chapter 10. -17 has a breakdown by type of the 88 marginal cases. Notice that the great
majority of these were “value assignment” statements. In fact, these 83 statements constitute a
significant fraction—right around one-third—of the 246 total expression-writing events.
Total
Codings
With Other
Codings
83
5
0
5
Value Assignment
Principle-based
Table Chapter 10. -17. Breakdown of the marginal cases.
365
The results of the second component of the base-level analysis, the utterance-centered
component, are fairly simple to report since this component only identifies interpretive utterances,
it does not code them further in any way. As shown in Table Chapter 10. -19, a total of 195
interpretive utterances were identified in the programming corpus. When this number is
combined with the 61 expression-construction episodes, the result is 256 total events that must be
coded in terms of forms and devices.
Interpretive Utterances
Constructed Expressions
Total Events
195
61
256
Table Chapter 10. -19. Programming corpus events coded in terms of forms and devices.
Now I want to take a moment to compare some of these values to those obtained in the
algebra analysis. First, notice that the 246 total expression-writing events is less than half the total
from the algebra corpus, which was 547. Of course, such a result is not necessarily very meaningful
in itself, since the students in each condition spent a different amount of time engaged in a
different number of tasks. In fact, the programming students did do fewer total tasks—6 instead
of 8—but the total time spent by the programming students on the coded tasks was 16.0 hours,
which is more than the 11.3 hours spent by the algebra students. Although a comparison based in
expressions/hour or expressions/task would produce different results, in either case, the “rate” for
programming is somewhat lower.
Interpretive Utterances
Constructed Expressions
Total Events
144
72
219
Table Chapter 10. -21. Algebra corpus events coded in terms of forms and devices.
The comparison gets a little more interesting if we look at the number of expressionconstruction and interpretation events. Comparing Table Chapter 10. -19 and Table Chapter 10. 21, we see that, although the number of expression-writing events in programming was much
lower, the total number of events to be coded was actually a little higher in programming, 256 as
compared to 219. Again, it is not completely clear how to compare these values appropriately. The
number of programming construction and interpretation events is much higher in comparison to
the total number of expressions written, but the number of events per hour is slightly lower
(16/hour versus 19/hour).
366
Whether we choose to normalize by expression, by task, or by hour, the above results paint an
image of a particular class of differences between the programming and algebra-physics corpuses:
The programming students wrote fewer expressions, but interacted meaningfully with
(constructed or interpreted) a larger fraction of these expressions than the algebra subjects.
Furthermore, these two effects balanced out to a certain extent, so that the number of
construction and interpretation events per hour were in the same ballpark.
Assuming, for the moment, that this result reflects a real difference between programmingphysics and algebra-physics, what can we make of this observation? There could be a number of
reasons for the differences observed here. Some of these reasons probably have to do with the fact
that the algebra subjects were quite familiar with the practice involved while the programming
subjects were learning a new practice. Given this difference in familiarity, one would expect the
programming subjects to proceed more slowly, and thus to generate fewer equations per hour.
Furthermore, since they have less of a body of experience to draw on, it is not surprising that they
needed to invent more of the expressions that they wrote, and that they were more likely to
discuss and interpret expressions.
But I also think that these results point to some real differences—differences that would
remain even if the students were equally familiar with the two practices. Recall that in the algebraphysics corpus, many of the total expressions that students wrote were produced by manipulating
expressions that were already written on the whiteboard. In fact, of the 547 expressions in the
algebra-physics corpus, 241 were produced only by manipulating existing expressions (i.e., they
were coded as derived and received no other codings). This is not surprising. In caricature, the
practice of algebra-physics involves first composing expressions, then pushing those expressions
through intermediate states, and finally doing something with the final expression obtained. The
“intermediate” expressions do not need to be constructed and are often not commented upon. In
contrast, programming-physics does not require the generation of these intermediate expressions.
In programming-physics, you compose the simulation and run it, then the computer does the rest.
There’s no generating of additional expressions, just the running of the original ones. Thus, you
might expect more expressions in algebra-physics, and you might expect students to interact
meaningfully with a higher percentage of programming expressions.
However, even if the numbers presented in the above tables are presumed to appropriately
describe my data corpus, there are reasons to doubt that these results somehow describe
programming-physics as it would exist outside my study. The results of the base-level analysis by
367
subject pair are shown in Table Chapter 10. -22. Notice that, as for algebra-physics, there is
substantial variation across pairs. This observation casts some doubt on whether the rough
generalizations above concerning differences between algebra and programming are, in any way,
universal.
N&D
Expressions
Interpretive Utterances
Constructed Expressions
Interpretations+Constructions
47
38
12
50
G&F
50
52
11
63
A&C
36
18
5
23
T&S
44
35
9
44
A&J
69
52
24
76
Totals
246
195
61
256
Table Chapter 10. -22. Results of the base-level analysis by subject pair.
There are a few other interesting comments to be made concerning Table Chapter 10. -22.
Recall that, in the algebra analysis, the number of interpretation and construction events were not
very strongly correlated with the total number of expressions. Here, more of the variation across
subjects in interpretations (r=.83) and constructions (r=.99) appears to be explainable by
differences in the total number of expressions written (as compared to r=.07 and r=.11 for
algebra). Interestingly, these observations, taken as a whole, can be seen to agree with the rough
differences between algebra and programming described above. If a large component of algebraphysics is the writing of expressions, mostly derived, that are somewhat independent of
interpretation and construction activity, then it makes sense that the number of interpretations
and constructions would not track closely with the total expressions. Similarly, because
programming does not include this component, this could explain the observation that there is
more tracking of construction and interpretation with the total number of expressions.
It is also worth commenting that the strong negative correlation between interpretive
utterances and construction events, observed in algebra, does not show up in Table Chapter 10. 22. Instead, there is a weaker, positive correlation (r=.75). I believe that more data is needed
before we can determine whether there are any interesting generalizations to be made concerning
differences across individuals, particularly due to the fact that any such differences would be
partly obscured by the fact that all of my data is for pairs of students.
I will conclude this discussion of the base-level results with the same point I made in
summarizing the base-level results of the algebra analysis: I believe that the results of this section
can be taken as suggesting that the construction and interpretation of expressions are not rare in
simulation programming. As I commented above, the number of construction and interpretation
368
events per hour was just a little less here than in the algebra corpus. In the algebra corpus there
were 219 events spread over 11.3 hours and the programming corpus had 256 events spread over
16.0 hours. Furthermore, there were many more of these important events in programming as
compared to the total number of expressions written.
Results from the coding of symbolic forms
The results of coding the 256 events identified in the base-level analysis in terms of symbolic
forms are given in Table Chapter 10. -24, with results listed both by cluster and individual form.
The values in the “Count” column are the raw number of times that each form was coded. The
numbers in the “percent” column were obtained by dividing these raw counts by the total counts
of all forms. The reader can refer to Appendix D for a brief descriptions of each form.
As I discussed in earlier sections of this chapter, it was not appropriate to code some of the
identified events in terms of forms. I argued that other resources were implicated in these events,
particularly what I called the “substitution and execution machinery.” Of the 256 total events, 27
fell into this auxiliary category.
369
Count
Percent
CANCELING
62
35
17
0
10
18%
10%
5%
0%
3%
Proportionality
PROP+
PROP-
63
45
13
18%
13%
4%
COEFFICIENT
RATIO
CANCELING(B)
1
2
Terms are Amount
PARTS -OF -A-WHOLE
Competing Terms
COMPETING TERMS
OPPOSITION
BALANCING
BASE ± CHANGE
WHOLE - PART
SAME AMOUNT
Count
Percent
81
43
5
6
27
24%
13%
2%
1%
8%
SCALING
38
29
9
11%
9%
3%
1%
0%
Multiplication
EXTENSIVE / INTENSIVE
4
4
1%
1%
36
11%
EXTENSIVE •EXTENSIVE
0
0%
0
37
0
0
0%
11%
0%
0%
Process
VARIATION
56
29
9
13
16%
9%
3%
4%
CONSTANCY
5
1%
0
1
0%
0%
Dependence
DEPENDENCE
NO DEPENDENCE
SOLE DEPENDENCE
SERIAL DEPENDENCE
Coefficient
SEQUENT. PROCESSES
SETUP -LOOP
Other
DYING AWAY
NO EVIDENCE
Table Chapter 10. -24. Results for the systematic coding of the programming corpus in terms of symbolic forms.
To begin discussion of these results, I want to examine the relative frequency of forms by
cluster. In Figure Chapter 10. -7, the breakdown of codings by cluster is displayed in a pie chart.
Glancing at this chart, none of the clusters stand out as a clear winner. The Dependence
Cluster—including
SERIAL DEPENDENCE —is
the most common with 24% of all codings. At the
next level are the Proportionality, Competing Terms, and Process clusters, with 18%, 18%, and
16% respectively. In the third echelon are the Coefficient and Terms are Amounts clusters, both
with 11%. Finally, the Multiplication Cluster received the fewest number of codings, only about
1%.
370
Terms are Amounts
11%
Multiplication
1%
Dependence
24%
Coefficient
11%
Process
16%
Propotionality
18%
Competing Terms
18%
Figure Chapter 10. -7. Percentage of form codings by cluster.
In many respects, the distribution revealed in Figure Chapter 10. -7 is quite similar to the
distribution observed in algebra-physics. Table Chapter 10. -26 provides a head-to-head
comparison of the relative frequency of the clusters in algebra-physics and programming-physics.
Notice that, in both cases, there are essentially three tiers of clusters. First, in both algebra and
programming, the Proportionality, Competing Terms, and Dependence clusters are the most
common. Then there is a middle tier with Terms are Amounts and Coefficient. And, finally, the
Multiplication Cluster is at the bottom in both practices. Programming also has the Process
Cluster, which falls between the top two tiers.
Algebra
Proportionality
Competing Terms
Dependence
Terms are Amounts
Coefficient
Multiplication
Process
28%
27%
23%
14%
5%
2%
-
Programming
18%
18%
24%
11%
11%
1%
16%
Table Chapter 10. -26. Comparison of form frequencies by cluster.
I believe that this similarity in distribution probably has a lot to do with a fact I have discussed
often, that individual programming statements can be interpreted much like algebraic expressions.
In fact, the similarity revealed in Table Chapter 10. -26 is accentuated if we throw out all codings
of the programming-specific forms—the Process forms and SERIAL DEPENDENCE . When any
371
codings of these forms are simply neglected, then the results are as shown in Table Chapter 10. 28, which reveals an even greater similarity between the programming and algebra distribution.
Algebra
Proportionality
Competing Terms
Dependence
Terms are Amounts
Coefficient
Multiplication
28%
27%
23%
14%
5%
2%
Programming
24%
24%
21%
14%
15%
2%
Table Chapter 10. -28. Comparison of form frequencies by cluster with codings of programming-specific forms
neglected.
Let’s take a moment to think about what this observation means. Note that, in casting out
SERIAL DEPENDENCE
and the Process forms, I have eliminated all of the forms that have to do with
line-spanning structure. Thus, Table Chapter 10. -28 suggests that the type of meaningful
structure that students see within programming statements is very similar to what they see within
algebraic expressions. This observation will have a number of implications for the discussion in
Chapter 11.
The most noticeable difference revealed in Table Chapter 10. -28 is the relative frequency of
the Coefficient forms. This, in some degree, likely has to do with idiosyncrasies of the specific
tasks I selected to include in the systematic analysis. However, I believe that this difference also
reflects an important feature of programming-physics. While working on the programmingphysics tasks, the students often spent a large amount of time “tuning” their programs; that is,
they would go through many cycles of running their program, and adjusting the various numerical
parameters that appeared in the program. Frequently, these numerical parameters were
coefficients, such as the one that appears in an expression for the force due to a spring: F= -kx.
Thus, because of the prevalence of this tuning activity, extra attention was focused on coefficients.
The comparison by cluster is summed up by the two bar graphs shown in Figure Chapter 10. 9 and Figure Chapter 10. -11, both of which display the number of counts by cluster. Notice that
the display in Figure Chapter 10. -11 uses light and dark bars to distinguish counts associated
with the programming-specific forms. These two figures primarily serve to highlight the
differences I have already mentioned. First, without the programming-specific forms, the
distributions have a rough overall similarity, with the Coefficient Cluster as the major exception.
In addition, looking at these figures highlights the fact that the Process forms were not
372
particularly rare in programming, there was a significant number of codings of these forms as
compared to forms in the other clusters. Finally, Figure Chapter 10. -11 makes clear the
importance of SERIAL DEPENDENCE . SERIAL DEPENDENCE adds a non-negligible amount to the
Dependence Cluster, making that cluster the most common in the programming-physics corpus.
80
70
Algebra
Forms
60
Counts
50
40
30
20
10
00
Competing
terms
Proportionality
Dependence
Terms are
Amounts
Coefficient
Multiplication
Other
Figure Chapter 10. -9. Algebra forms by cluster.
80
Serial
Dependenc
70
Programming
Forms
60
Counts
50
40
30
20
10
0
Competing
Terms
Proportionality
Dependence
Terms are
Amounts
Coefficient
Multiplication
Process
Figure Chapter 10. -11. Programming forms by cluster. Light portions of bars indicate counts of programmingspecific forms.
I will now turn my attention from clusters to a discussion of individual forms. As in algebraphysics, a few forms account for many of the codings, though the distribution is a little more
uniform in programming. Table Chapter 10. -30 and Table Chapter 10. -32 show the most
common forms in programming and algebra. In programming, the PROP+ and DEPENDENCE forms
373
leads the way, followed by
COMPETING TERMS
and BASE ±CHANGE . The story was precisely the same
in algebra, with the exception that BALANCING appears in the number four slot, between COMPETING
TERMS
and BASE ±CHANGE . Of course, this is a very significant exception—it is a very important
observation that one of the most common forms in algebra-physics did not appear at all in
programming-physics. This observation will play an important role in the discussion in Chapter
11.
Count
45
43
35
37
29
29
27
PROP+
DEPENDENCE
COMPETING TERMS
BASE ±CHANGE
SEQUENTIAL PROCESSES
COEFFICIENT
SERIAL DEPENDENCE
Percent
13%
13%
11%
11%
9%
9%
8%
Table Chapter 10. -30. The most common forms in programming-physics. All other forms received below 5%
of codings.
Count
52
40
30
28
23
PROP+
DEPENDENCE
COMPETING TERMS
BALANCING
BASE ±CHANGE
Percent
19%
15%
11%
10%
8%
Table Chapter 10. -32. The most common forms in algebra-physics. All other forms received below 6% of
codings.
It is also important to notice that Table Chapter 10. -30 contains two of the programmingspecific forms, SEQUENTIAL PROCESSES and SERIAL DEPENDENCE . This suggests that these new forms
are not inconsequential additions. The analysis of the results by cluster pointed to a similar
conclusion.
In summary, the distribution of forms in the programming-physics corpus seems to share a
common core with the algebra-physics corpus. PROP+, DEPENDENCE, and COMPETING TERMS were
again very common individual forms, and the associated clusters were very common in the coding
of both corpuses. The most interesting differences were associated with forms that appeared only
in one practice or the other. The Process forms and SERIAL DEPENDENCE appeared only in
374
programming, and with a reasonable regularity. Furthermore, the absence from programmingphysics of one of the most common forms—BALANCING —constitutes a significant observation.
N&D
G&F
A&C
Competing Terms
Proportionalit
Dependence
Terms are Amounts
Coefficient
Multiplication
Process
13
19
11
6
11
2
6
15
15
24
13
13
5
5
2
4
16
Totals
68
96
T&S
A&J
Totals
7
23
11
8
6
7
1
6
6
13
36
8
7
1
21
62
63
81
37
38
4
56
23
62
92
341
Table Chapter 10. -34. Form cluster codings broken down by subject pair.
The above discussion, to a certain extent, presumed not only that the numerical frequencies
described my data corpus, but that the overall average of my corpus was somehow representative
of programming-physics practice. As usual, I will break down my frequencies in order to help
estimate the extent to which this latter assumption is merited. Table Chapter 10. -34, which gives
the results for form cluster codings broken down by subject pair, shows substantial variation across
pairs. This suggests the usual caveats: My results may depend sensitively on the particular subjects
that participated in my study, and, because of individual differences, it may not make sense to
look for a universal distribution.
As in the algebra corpus, however, it is possible to find some weak suggestions of across-pair
uniformities. For example, the Multiplication Cluster received less than 3% of the codings for
every pair. In addition, the three least common clusters overall—Terms are Amounts, Coefficient,
and Multiplication—always received less than 28% of the total codings. In contrast, the three
most common clusters—Competing Terms, Proportionality, and Dependence—always received
more than 50% for all pairs. These observations do not eliminate the importance of the observed
differences, but they do hint at a degree of similarity.
Finally, the results by task are given in Table Chapter 10. -35. There is substantial variation
across tasks, which again suggests that my overall results may depend sensitively on the particular
tasks included in the study. I will comment on this issue further later in the chapter.
375
Shove
Drop
Toss
Air
Competing Terms
Proportionalit
Dependence
Terms are Amounts
Coefficient
Multiplication
Process
24
15
17
10
4
2
10
9
15
Totals
72
13
9
137
More Air
42
25
30
4
21
3
Spring
8
9
20
10
7
54
Totals
12
5
16
3
20
62
63
81
37
38
4
56
56
341
Table Chapter 10. -35. Form cluster codings broken down by task.
Results from the coding of representational devices
The results for the systematic coding of representational devices are summarized in Table
Chapter 10. -37. (Refer to Appendix E for a brief description of each device.) In addition, the
results collected by class are shown in a pie chart in Figure Chapter 10. -13. Looking at this chart,
it is clear that Narrative devices dominate. Furthermore, not only is the Narrative class the most
common overall, the three most common individual devices are members of this class: CHANGING
PARAMETERS , TRACING ,
and PHYSICAL CHANGE .
Narrative
CHANGING
PARAMETERS
CHANGING SITUATIONS
PHYSICAL CHANGE
TRACING
Special Case
SPECIFIC VALUE
LIMITING CASE
RESTRICTED VALUE
RELATIVE VALUES
Count
Percent
195
73
63%
24%
18
44
60
6%
14%
19%
GENERIC MOMENT
44
19
6
14%
6%
2%
CONSERVATION
14
5
5%
2%
Other
Static
SPECIFIC MOMENT
STEADY STATE
STATIC FORCES
ACCOUNTING
RECURRENCE
NO EVIDENCE
Count
Percent
67
2
22%
1%
8
0
13
3%
0%
4%
0
4
41
0%
1%
13%
1
1
0%
Table Chapter 10. -37. Results for the systematic coding of the programming-corpus in terms of
representational devices.
376
Special Case
14%
Static
22%
Narrative
63%
Figure Chapter 10. -13. Percentage of device codings by class.
If we compare the results summarized in Figure Chapter 10. -13 to the results by class for
algebra-physics, we find the same basic distribution (refer to Table Chapter 10. -39). In both
cases, Narrative devices were the most common, followed by Static and then Special Case devices.
However, in programming, the prevalence of the Narrative Class appears to be somewhat
accentuated, mostly at the expense of the Static Class. Although, as I will discuss in a moment,
this difference may be an artifact of my task selection or the variation across individuals, I will
argue in Chapter 11 that this diminishing in the use of Static devices is precisely what one might
expect given the differing nature of algebra-physics and programming-physics.
Algebra
Narrative
Static
Special Case
54%
35%
11%
Programming
63%
22%
14%
Table Chapter 10. -39. Comparison of device frequencies by class.
The comparison by class is further illustrated by the bar graphs in Figure Chapter 10. -15 and
Figure Chapter 10. -17, both of which display the raw number of counts by class. An examination
of these two figures clearly illustrates the significant role played by the two new devices,
TRACING
and RECURRENCE. In fact, the prevalence of the Narrative Class is largely due to TRACING ; it
accounts for about 30% of all Narrative codings. And the majority of Static codings—about
60%—were associated with
RECURRENCE.
377
140
Algebra
Devices
120
Counts
100
80
60
40
20
0
Narrative
Static
Special Case
Figure Chapter 10. -15. Algebra devices by class.
200
Programming
180
Devices
Tracing
160
140
Counts
120
100
80
60
Recurrence
40
20
0
Narrative
Static
Special Case
Figure Chapter 10. -17. Programming devices by class. Light portions of bars indicate counts of programmingspecific forms.
As in my analysis of forms, it is interesting to look at what happens if we throw out the devices
that are associated with line-spanning structure. This only means eliminating TRACING , since that is
the only device that is uniquely associated with line-spanning structure. When counts associated
with the TRACING device are thrown out, then we obtain the results shown in Table Chapter 10. 41.
378
Algebra
Narrative
Static
Special Case
54%
35%
11%
Programming
54%
27%
18%
Table Chapter 10. -41. Comparison with tracing eliminated.
A first observation to make about this table is that the distribution has gotten closer to the
distribution observed in algebra-physics. But this table allows another important observation that is
somewhat more surprising. Although Table Chapter 10. -41 shows approximately equal
percentages for Narrative devices, it seems to indicate a shift from Static to Special Case devices in
programming. This is partly a reflection of the diminishing of Static interpretations for the reasons
mentioned above. However, I believe it is also suggestive of another important feature of
programming-physics. As I will discuss in the next chapter, programming forces students to
embed expressions in the particular—in order to run a program, you need to make selections
concerning the numerical values that will be in the expressions that you have written, and you need
to make certain other choices concerning the programming context in which the expressions are to
be embedded. These features of programming may tend to make Special Case interpretations
more prevalent.
In summary, the basic ranking of device classes in the programming-physics corpus was
substantially similar to that obtained for algebra-physics: The Narrative Class was most common,
followed by the Static Class and the Special Case Class. However, the results for programming
were somewhat skewed away from the Static Class to the two other classes, especially to Narrative
devices. In addition, the two new programming-specific devices were both quite common.
To conclude this section, I must again note that the above frequency differences between
programming and algebra might be an artifact of the task selection or the variation across
individuals. In order to help calibrate this effect, it is helpful to break down these results by subject
pair and task. The results for device class codings by subject pair are given in Table Chapter 10. 43. This table indicates some fairly large differences across pairs, both in the total number of
codings and their distribution. For example, Adam and Jerry had more than five times as many
Narrative as Static codings, while the other pairs had closer to 2 times as many Narrative codings
as either other class. But, again, there are some uniformities. For example, the Narrative Class was
the most common for all subjects and it always received over 49% of the total codings. In
contrast, the Special Case Class never received more than 27% of the codings for a pair.
379
N&D
Narrative
Static
Special Case
Totals
G&F
33
16
18
67
A&C
48
20
7
75
13
5
5
23
T&S
A&J
33
15
8
56
Totals
68
12
6
86
195
68
44
307
Table Chapter 10. -43. Device class codings broken down by subject pair.
Table Chapter 10. -45 has the results broken down by simulation task. Again we see the usual
strong variation across task, both in number and distribution. However, there is some indication
that the prevalence of the Narrative Class may be reflected across tasks.
Shove
Drop
Narrative
Static
Special Case
41
7
10
3
3
Totals
58
6
Toss
Air
18
More Air
Spring
Totals
2
55
36
15
38
3
8
40
19
9
195
68
44
20
106
49
68
307
Table Chapter 10. -45. Device class codings broken down by task.
Methodological Issues
I will now close this chapter by reflecting on the types of conclusions that we can draw from all
the numbers presented in this chapter and in Chapter 6. There are a number of issues to be
addressed and caveats to have in mind, and I want to have a discussion of these issues behind us
before presenting my more interpretive comparisons in the next chapter.
Just as the results of Chapter 6 can be taken to provide a “description” of algebra-physics,
understood in multiple senses, the results presented in this chapter furnish a description of
programming-physics. There are several components of this description:
1. The very idea that elements like symbolic forms and representational devices exist.
1. The lists of specific forms and devices needed to account for my data.
1. The frequency of individual forms and devices.
Of course, the first two “results” were essentially presented in previous chapters. However, in
describing the systematic analysis techniques in this chapter, I have supplied some of the
methodological basis underlying the assertions of those preceding chapters.
Much went into the building of the lists of forms and devices, and attaching forms and
devices to events in the data corpus. This was a highly heuristic process—I did not describe an
380
exact and rigorous technique for extracting the coding categories from the data. Furthermore, the
coding of individual episodes may be questioned since I did not develop a scheme so that my
codings could be corroborated by a second coder. For these reasons, the analysis performed here
must be understood to be somewhat preliminary.
However, these are not the issues that I want to consider in this section. For the moment, I
want to assume that my results in some way describe the data corpus that I collected. Even if we
accept this assumption, a host of difficult issues arise. In particular, I want to think about issues
that cluster roughly around this question: In what sense can I be said to be describing practices of
algebra-physics or programming-physics that have some existence outside of the limits of this
study? Most of these issues have been commented on elsewhere. Here, I draw them together and
summarize my position.
Issues relating to subjects
First, there are a few issues relating to the subjects that participated in my study. The subjects
in both pools were drawn from a population of students at a specific stage in their learning of
physics. They were not expert physicists, but they were also certainly not complete novices.
Instead, they were at an intermediate stage, at the tail end of the introductory sequence. As I
discussed in Chapter 6, the decision to look at students in this intermediate range was a principled
choice; I wanted to be able to say something about the nature of expertise, as well as to see some
of the phenomena associated with learning difficulties.
But, because I only have information from this one time-slice, any projections up and down
the learning timeline must be understood to be speculative. Most importantly, throughout this
work I have implied that the form-device knowledge system continues to exist into expertise,
albeit with many refinements. But this speculation must ultimately be verified with new data
involving expert physicists. It may be that we cannot understand my results as describing expert
physics practice, not even in its broad strokes.
Of course, I do believe that the form-device system continues to play an important role in
expertise. I should note, however, that some refinement of the methods employed in this study
would likely be required to see an expert’s form-device system in action. In particular, the tasks
used in this study would probably be very easy for experts; they could probably solve them in a
rote and mechanical manner. To really see an expert’s form-device knowledge at work, we would
need to have tasks that challenge the experts, at least a little.
381
It is also worth commenting on the fact that, for both pools, the number of subjects was
relatively small. Nonetheless, this is not a “small N” study in the usual sense. Notice that, I did
not try to characterize individual students or pairs of students in any way. For example, an
alternative study might have given the algebra and programming subjects identical pre-tests and
post-tests. Then I could have made claims that sounded like: “8 of the 10 programming subjects
did X on the post-test, while only 4 of the 10 algebra subjects did X.” Claims of this sort rely on
producing characterizations of individual students. But I did not make any claims of this sort.
Instead, my data was always aggregated across the subjects in each pool. Furthermore, for each
pool, over 200 individual events were examined and coded.
Nonetheless, even though I aggregated my data before drawing conclusions, the observed
variation across subject pairs—and small number of subjects—could be reason for concern. First,
because there appear to be sizable variations across pairs, it is less likely that my overall sample of
construction and interpretation events reflects the overall distribution in the practice.
Second—and more profoundly—the existence of these differences casts doubt on the whole
program of looking for a universal set of weightings. I have always presumed that individual
differences would exist; but, in performing this analysis, I was hoping that there were rough
universals to be captured. The variation across subjects is reason to worry that this might not be
the case. Although these worries are somewhat moderated by the observation of some weak
similarities, these problems cannot be completely avoided and, where necessary, I have tried to
temper my claims appropriately.
There are also some issues relating to the subject population that are specific to programmingphysics. As I have commented at numerous times, it is of great relevance to my examination of
programming-physics that the students came to the study with experience in algebra-physics.
Thus, the results described here are almost certainly not the same as what we would find if I had
examined students that were introduced to physics solely through a programming-based practice.
The reasons for choosing to study experienced physics students in the programming portion
of the study were largely pragmatic. For one thing, it would simply have been too time
consuming to create, from scratch, a population of true programming-physics initiates.
Furthermore, there are some reasons that this approach is not too problematic. First, my analysis
was always tightly focused around the external representations, I only looked at what people said
while actually looking at and pointing to particular programming expressions. This means that my
analysis is only revealing where and how understanding gets connected up to programming
382
expressions; it is not giving a full picture of my students’ physical intuition or other physics-related
knowledge. This is what we want. Of course, because these students have significant experience in
seeing structure in algebraic expressions, there is likely still some skewing of my programmingphysics results in the direction of algebra-physics. But this just means that any differences
uncovered here are likely, if anything, to be exaggerated in a study of true programming-physics
initiates.
Issues relating to tasks
There are also a number of sensitive issues relating to my selection of tasks. In this chapter and
Chapter 6, we saw that there was a rather larger variation in codings across the tasks included in
the study. This implies that if I had included different tasks, I would have obtained different
frequencies for specific forms and devices. Clearly what we really want is a selection of tasks that
are somehow representative of the practice we are trying to study. But, although this sounds
sensible, it is a very hard goal to reach in practice. What are the boundaries of the practice we
should study? What are the dimensions along which we should categorize tasks in order to judge
representativeness?
Suppose for example, that we chose to categorize tasks and judge representativeness according
to the physical concepts that the tasks traditionally involve. In that case, we would want a selection
of tasks having to do with the concept of force, a selection having to do with momentum, etc. But,
such a principled selection could easily be very skewed in the forms that are engaged. Even if we
choose some tasks having to do with forces, we could easily choose “too many” or “too few” tasks
that involve balanced forces.
Probably the best way to select tasks would be based on a post-hoc categorization. For
example, now that I have an idea of how forms and devices tend to line up with tasks, I could go
through a physics textbook and do an analysis of the form and device distribution as realized in
the problems in a textbook. Then I could select tasks that mirror this distribution and iterate my
study. If I got the same results as for the textbook analysis, then I could know that I was getting
somewhere. Of course, all of this relies on my knowing what the forms and devices are and how to
associate them with tasks, things that I did not know when I began this study.
This issue of what tasks are representative is even more sticky for programming-physics, since
there is no independently existing practice that I can try to mirror. My selection of tasks can only
383
be based on guesses concerning what a practice based on programming would look like, and a set
of specific orientations concerning how best to use programming in instruction.
This problem of task selection is one of the biggest difficulties in drawing conclusions from
the frequency results. However, there are some moderating factors. Most importantly, when we
ultimately set out to draw conclusions, we will have more than just the quantitative results to lean
on; we will also have all of the specific examples I presented of forms and devices, as well as our
understanding of what these elements are and how they are tied to the representational forms.
These other resources can help us substantially in determining the plausibility of potential
conclusions.
The problem of ÒcausesÓ
Throughout this work, I have usually been careful to frame my study as a comparison of two
practices, rather than as a comparison of two symbol systems. Nonetheless, this work is intended,
in some sense, to be about symbol systems, and this is reflected in my choice of names for the two
practices, programming-physics and algebra-physics. However, there remains the possibility that
the symbol systems themselves are somehow irrelevant to all of the results I have obtained here. In
other words, even though I have found interesting differences in the forms and devices in
programming-physics and algebra-physics, how can I know that these differences are due to
differing “properties” of the two symbol systems? In asking this question we implicitly take the
view that symbol-use is one aspect of a larger practice; thus, any differences that we see between
two practices may potentially be better described as traceable to other elements of the practices.
And there is one very important example to be considered here. Programming-physics, as I
have constructed it, involves teaching a particular “difference” model, the “tick model.” Thus, one
could argue that any differences that I see could be due to the fact that programming-physics
teaches a difference model, and algebra-physics teaches something like a constraint-based model.
My primary response to this critique is that this issue, as posed, contains assumption that I do
not believe hold. Namely, in posing this question we assume that these models exist separately
from the symbol system involved; or, at least, that a variety of models can be taught in
conjunction with a given symbol system. Although there is certainly much looseness in the physics
practices that a symbol system can be embedded in, there are also severe limitations on how a
given system can be used naturally.
384
In fact, I tried for several months to construct a difference model practice of physics that
employed algebra, and did not come near to success. It is possible to employ a difference model
using something like algebraic notation, and what Feynman does in his introductory textbook is
an example. But, amongst the reasons that Feynman quickly abandons this approach, is that the
method he plays out is tedious and would be largely impractical to employ in many tasks over a
semester. Furthermore, note that to use algebraic notation to state difference relations, we have to
modify or at least add to traditional algebraic notation.
The point here is that the external representations we use, especially central ones like algebraic
notation, are inextricably tied up with the other elements of a practice that is built around the
representation. Another way to say this is that when “external representation” is suitably defined,
then it has properties that constrain the surrounding physics practice. I believe that those
constraints are critical in defining the practice, and that’s why I focus my attention there.
Furthermore, I have tried to structure my research so that it gets at these points of constraint. I
did not look at all statements that students made, only statements made while pointing to or
writing expressions. Thus, even though these constraints certainly leave room for much “play” in
the practices that we build, I hope to have gotten at the nature of the guiding influence imposed
by the external representation.
Final comments on the comparison: The matching of conditions
The comparison I am attempting to perform in this study is a difficult one. In most studies
that involve comparisons, such as “treatment” studies, one gives two different treatments, and
then compares subjects’ performance on the same tasks. Here, I am comparing subjects engaged in
different tasks, of a substantially different nature. Thus, some rethinking of what constitutes a
useful comparison is required. How much should we force conditions to be the same, when these
differences are just what we are interested in?
Broadly, I have uncovered two classes of assertions that I want to make that may be
problematic. The first has to do with representativeness. I would like to be able to argue that my
descriptions of algebra-physics and programming-physics are somehow representative of these
practices as they exist (or could exist) outside of my study. This does not require that I match the
two conditions; on the contrary, it requires that I somehow ensure that I am looking at a
representative sample of the practices in their true or natural forms.
385
The second class of problems has to do with assertions about causes. I would like to be able to
trace differences that I observe in the practices to factors closely related to symbol-use. This
requirement does suggest a need to “match” the two groups in certain ways. For example, I would
like to be able to exclude the possibility that the observed differences are due to differences in the
two populations. To the extent that the populations are closely matched, this possibility can be
more plausibly denied.
Conclusions
With all of the above results, and keeping all of the above caveats in mind, what can we
conclude? First, in a sense, the systematic analysis has put the totality of my viewpoint to test by
employing it to describe the data corpus. In the process, we did encounter a few marginal cases,
such as those involving the “Substitution and Execution Machinery,” which could not be
accounted for in terms of forms. However, as I argued in Chapter 6, the existence of these
marginal cases does not necessarily signal the breakdown of the theory, since my viewpoint
already presumes a certain degree of complexity. The form-device system is just one amongst
many resources and we should not expect this resource to account for all phenomena, even all of
the phenomena in a limited territory. On the contrary, I believe that it is impressive that the formdevice theory can do as much work as it does. And now we have seen evidence that this is true not
only of algebra-physics, but also of programming-physics.
Related to this first point is the observation that the interpretation of expressions and the
construction of novel expressions are not rare events, both in programming-physics and in algebraphysics. Since this observation should survive all of our worries about tasks and subjects—with the
sole caveat that, as discussed in Chapter 6, my prompts may have tended to over-encourage
interpretation—we can be relatively sure that I am not studying a collection of marginal
phenomena. This observation is extremely important with regard to drawing conclusions from
these results. Even if the systematic analyses of the two corpuses found very strong and reliable
differences, we could not draw very interesting conclusions from these differences if the
phenomena involved were relatively rare. We want to be able to presume that these phenomena are
common enough that an analysis of them tells us something important about the two practices.
Now let us consider some of the particular differences uncovered by the analysis of
programming-physics. We saw that, not only were there differences in the frequency of individual
forms, there were differences in the very forms that appeared. There were some new forms in
386
programming-physics—the Process forms and
SERIAL DEPENDENCE.
Since these new forms were
common across a variety of programming-tasks, and since they did not appear at all in my
algebra-physics corpus, there is probably a result here that we can count on.
Similarly, there were also some symbolic forms that appeared only in algebra-physics.
Probably the most important instance here is the BALANCING form, the fourth most common form
in algebra-physics. I believe this is another result we can count on and draw conclusions from.
In addition, there are some results pertaining to devices that I believe are relatively sound. In
particular, the analysis of the programming corpus revealed two new devices in programmingphysics: TRACING and RECURRENCE. Again, because these were common in the programming
corpus and did not appear at all in the algebra corpus, I believe it is appropriate to use this result in
what follows.
The above results, which are primarily based on forms and devices that appeared only in one
corpus or the other, are the main results I will build on. I believe that results of this sort—results
that depend only on the lists of forms and devices—are particularly plausible because it is often
easy to see how they follow from differences in the representations involved. However, in the next
chapter I will also want to argue for some differences between programming-physics and algebraphysics that are more speculative because they depend on some detailed aspects of the frequencies
obtained by the two analyses. For example, I will base part of my comparison on the observation
that, when line-spanning structure was ignored, the distribution of programming forms and
devices was close to that of algebra. Although this observation is on shaky ground because of
across-task and across-pair variations, the close similarities in these distributions is quite plausible.
This is especially true given my observation, in Chapter 9, that the construction and interpretation
of individual programming statements can be very similar to that of equations.
In addition, I will also make use of the observation that there seemed to be a general shift
from Static devices to the other two classes in programming. Again, this observation is based on
more shaky aspects of my frequency results. However, because this result is generally plausible
and—as I will argue—it is consistent with other results, I will take this shift in frequency as
suggestive of real differences between programming and algebra.
Those are the basic outlines of the comparison revealed by the systematic analysis of the two
corpuses. Now, in the next chapter, I will take these results as a given and move on to a more
interpretive and speculative comparison of programming-physics and algebra-physics.
387
Chapter 11. A Comparison of Algebra-Physics and
Programming-Physics
It is now time to, once again, step back from the forest of details and get some perspective. The
big picture I have been arguing for looks like this: Students come to their learning of physics with
quite a lot of relevant cognitive resources. As I discussed in Chapter 5, some of these resources
have to do with getting along in the physical world. In addition, there are some relevant cognitive
resources that have more schoolish origins, such as resources that develop during mathematics
instruction.
So, students have these previously existing cognitive resources. Then, when they start learning
physics, the existing resources get refined and new resources develop. In particular, this research
has been concerned with discussing a set of resources that are associated with seeing a certain type
of meaning in physics expressions. This set of resources is the form-device knowledge system.
This new knowledge system develops to do some very particular jobs. In a sense, the formdevice system acts like a way-station between the previously existing cognitive resources and
symbolic expressions. If you want to express an idea in terms of an external symbolic expression,
then that idea must feed through forms and devices. In addition, the form-device system allows
us to see a certain type of meaning in existing expressions.
Because of its special role as intermediary and connection to symbolic expressions, an
examination of the form-device system tells us a great deal about how the use of a particular
practice of symbol use influences a student’s developing understanding of physics. First, because it
is the link to pre-existing cognitive resources, an examination of the form-device system tells us
something about how these resources are refined and adapted during physics learning involving a
particular symbol system. In addition, the specific composition of the form-device system is itself
of interest. I have argued that, once developed, the form-device system should be considered to
be part of “physical intuition.”
An examination of the form-device system is also of importance from the point-of-view of
improving physics instruction. As physics educators, one of the main difficulties that we
face—and one of our most important goals—is to help students to use symbolic expressions with
understanding. We want connections to existing resources to be made during symbol use and we
want a student’s physical intuition to be refined. So the properties of the two symbolic practices
388
revealed by this analysis are of special importance. We would like to know: How easy is it to use
the expressions in a practice “with understanding?”
The task of the present chapter is to draw together all of the preceding analyses to produce my
final comparison of programming-physics and algebra-physics. Contributing to this comparison,
we have in hand an account of the forms and devices in each of these two practices, as well as some
more or less reliable numerical values for the frequencies of these elements in each practice. In
what follows, I will begin by drawing conclusions based on my analyses of symbolic forms, and
then I will turn to representational devices.
In addition, following the discussions of forms and devices, I will do a little additional work.
Since the analysis of forms and devices cannot capture all of the interesting differences between
algebra-physics and programming-physics, I will build some additional theory to capture some of
these differences. This will be embodied in a discussion of what I call “channels of feedback,” and
we will come to see the form-device system as just one of these multiple “channels.”
Before setting out, I want to say something about the tactic I will employ here for drawing
conclusions. Recall that, in my presentation of both systematic analyses—Chapters 6 and 10—I
pointed out that there are a number of reasons to question my particular quantitative results.
Among the most important of these reasons were the observations of variation across individuals
and tasks. Though, for these same reasons, some caution is required in this chapter, my intent here
will be to try and draw as many speculative conclusions as possible. To do this, my approach will
be, in general, to begin with differences between algebra-physics and programming-physics, as
suggested by the systematic analyses. Then, I will attempt to back up these results, where
appropriate, with arguments for their plausibility. These arguments will be based in the qualitative
observations presented in earlier chapters, as well as some more straightforward observations
concerning each of the representational systems and how they are typically used.
A Comparison Based In Symbolic Forms
Recall that, as I discussed at the end of Chapter 8, the analysis in terms of symbolic forms was
designed partly to capture a number of informal hypotheses. The basic notion was that
programming tends to privilege a somewhat different “intuitive vocabulary.” More specifically, I
hypothesized that some important differences would follow from the fact that, in contrast to
algebraic expressions, programming statements are ordered and directed. For example, since causal
intuitions require an ordering of the world’s influences and effects, I hypothesized that these
389
intuitions might be more strongly drawn upon in programming-physics. Furthermore, I
conjectured that intuitions based in symmetric relations, such as those relating to balance and
equilibrium, would appear less in programming-physics.
The Theory of Forms and Devices has provided us with a means of making these informal
hypotheses more precise. Given that theory, these informal hypotheses can be recast as questions
about the symbolic forms involved in each practice, as well as questions about how existing
intuitive resources might be refined through symbol use. Furthermore, with these more precise
specifications and the preceding results in hand, we are in a position to discuss whether some of
the specific hypotheses are born out. Is algebra-physics more about balance and equilibrium? Can
we identify any sense in which programming is more causal?
Let us turn to the results and see what conclusions we can begin to draw. The basic outlines of
the differences revealed in the preceding chapters are not too difficult to state succinctly. I argued
that programming-physics involves several new forms—a new dependence form called SERIAL
DEPENDENCE,
and a new cluster of forms that I called the Process Cluster. The existence of these
forms constitutes a substantial departure from algebra-physics. However, we saw that when these
forms were excluded from the analysis, a substrate of similarity was revealed. In particular, the
frequencies of forms, collected by cluster, were very similar (refer to Table Chapter 10. -28). I
argued in Chapter 10 that this similarity was quite plausible, given my qualitative observations.
Algebra
Proportionality
Competing Terms
Dependence
Terms are Amounts
Coefficient
Multiplication
28%
27%
23%
14%
5%
2%
Programming
24%
24%
21%
14%
15%
2%
Table Chapter 11. -47. Comparison of form frequencies by cluster with codings of programming-specific forms
neglected.
The major exception to the similarity revealed by the restricted analysis was a boost in the
frequency of programming’s Coefficient Cluster. Although this boost may just be an artifact of
individual variation or the limited task selection, I believe that it is indicative of a real difference
between programming-physics and algebra-physics. In the previous chapter, I maintained that this
difference has to do with the fact that the “tuning” of numerical values was a major activity in
390
programming-physics. In fact, I will argue later in this chapter that one of the most important
features of programming-physics is that it forces students to deal with specific numerical values.
Neglecting for a moment this boost in the Coefficient Cluster, we are left with the observation
of a fundamental similarity in the forms of programming and algebra, at least at the level of
clusters. There is a quite simple and important way to interpret this observation. As I have
discussed, each of the programming-specific forms has to do with line-spanning structure. Thus,
in excluding these forms from the analysis, we are essentially examining the types of meaningful
structure that can be supported by individual programming statements. This seems to indicate
that individual programming statements and individual algebraic statements can actually support
very similar symbolic forms and that they seem to engender these forms with roughly similar
frequency. Given this observation, it may seem that we are moving toward a negative result on my
hypotheses concerning differences between algebra and programming.
Unbalanced BALANCING
But there are a number of reasons that this result is misleading. One reason is because, as I will
argue shortly, some of the evidence for my informal hypotheses is hiding in line-spanning
structure, which is excluded from Table Chapter 10. -28. More subtly, in collecting the results by
cluster, we miss many of the interesting differences that are associated with the individual forms
within clusters.
One fact that is totally hidden when the results are collected by cluster is that some of the
forms that appeared in algebra-physics, even some that were quite common, did not appear at all
in programming-physics. For example, the
PARTS -OF -A-WHOLE
form received no codings in the
programming analysis. In this case, I believe that the absence of this form has much to do with the
selection of programming tasks. It turns out that the majority of the algebra codings of this form
were associated with a single task—the Running in the Rain task—and there was no analog of this
task in the programming corpus. Furthermore, it is not too hard to imagine a case in which
programming expressions could be interpreted in terms of
PARTS -OF -A-WHOLE.
For example, we
can imagine a program in which there is an expression for the total mass of a compound object:
change Mtotal M1 + M2.
However, I believe that the absence of some forms in programming is indicative of real and
important differences between programming and algebra. Most notable amongst these absences is
the lack of any codings for the BALANCING form. One reason to think that this constitutes a real
391
difference between these two practices is that BALANCING was quite prominent in algebra-physics: It
was the fourth most common form overall, alone accounting for 10% of all codings, and it
appeared regularly across a wide variety of tasks.
Another reason to believe that this result points to a real difference between programming and
algebra is that it is a plausible result, given some simple observations concerning the nature of these
two practices. In particular, situations involving BALANCING turn out to be somewhat easier to deal
with in algebra-physics than situations that do not involve BALANCING . For example, when we have
balanced forces or balanced momenta, then we can just set the quantities equal and solve, in the
process avoiding differential equations. Because of the simplification that results, the practice of
algebra-physics tends to seek out situations that can be understood in terms of
BALANCING .
In
contrast, BALANCING situations are not especially easy or natural in programming-physics.
To illustrate this point, I want to think about the spring-related tasks that appeared in each
portion of this study. Recall that I asked the students in the algebra pool to solve a problem in
which a mass hangs motionless at the end of a spring. In this case, there are two forces in balance,
the force due to gravity acting downward and the force of the spring acting up. Thus we can write
Fs = Fg
and, with this expression written, the path to a solution is relatively straightforward. Of course, it
is possible to use algebraic notation to solve problems involving a block that oscillates at the end of
a spring, but to fully derive the time dependence of the block’s position requires solving a
differential equation, or that we remember the solution to that equation. For this reason, it is
common in algebra-physics instruction to seek out the easier case in which the block is motionless
and the forces are in balance.
This is not to say that situations involving balance and equilibrium are not fundamental and
important phenomena in their own right. The physics of balance is not a poor man’s physics and I
do not mean to imply that the use of algebra constrains physics to the study of unimportant
phenomena. In fact, there is reason for us to have special interest in static phenomena—cases in
which all objects in a system are at rest relative to each other. For example, these cases happen to
be very important to engineers and architects, who need to design buildings and bridges that do
not move (i.e., that do not fall over).
392
The situation in programming is quite different. To begin, notice that there is no way to write
symmetric BALANCING relations like Fs = Fg , at least within the practice of programming-physics
that I employed in my study. In particular, the programming statement
change Fs Fg
is not equivalent to Fs = Fg . Rather than stating a symmetric constraint, this programming
statement specifies a directed action: Assign the value in Fg to the variable Fs.
Furthermore, it is not even clear what it would mean to use programming to solve a problem
involving the static case. We could write a program that has, at its heart, a single line that
computes how far the spring is stretched:
change pos
mass * g / k
Doit
Then our program could use the result of this computation to simply position a graphical object at
the corresponding location on the display. But this would not be a very interesting program and it
presumes that we have already used algebraic methods to derive the relation embodied in the
above statement.
Furthermore, the oscillating motion is very easy in programming; you just type the expression
for the force in the Tick Model and then let the simulation run. So, unlike the algebra case, there
is no incentive to seek out a situation involving BALANCING ; if anything, all of the motivating factors
suggest that we do otherwise. Since dynamic simulations are easy and interesting in programming,
we might as well do them.
So there is reason to believe that algebra-physics really does tend to more frequently draw on
notions of BALANCING . We could attempt to modify our practices so that such differences in
tendencies were moderated. For example, we could expand our practice of programming-physics
to include “conditional” statements, such as:
if
force1 = force2
do-something
Doit
Doit
This statement says to execute the procedure named do-something if the variable force1 happens
to equal the variable force2. If we allowed these conditional statements in our practice of
programming, there would be a niche for symmetric equality relations that might lend themselves
to BALANCING interpretations.
393
However, I believe that even in modified practices of this sort,
BALANCING
will still play a less
prominent role than it does in algebra-physics. This is true, if for no other reason, because we can
deal with time-varying situations so easily. There is simply no reason to lean as heavily on cases in
which influences happen to balance.
Ordering and processes
So, BALANCING is a symbolic form that is very common in algebra-physics, but which did not
appear at all in programming-physics. Now I want to turn this discussion to the forms that
appeared only in programming. Recall that, in my original “informal hypotheses,” I conjectured
that there would be important differences between programming and algebra traceable to the fact
that the lines in programs are strongly ordered. In part, the effects of this strong ordering are
reflected in the existence of the Process Cluster, a whole new cluster of forms that are unique to
programming-physics. The existence of this new cluster is extremely important; as I will argue, it
suggests that a whole new wing of intuitive knowledge is picked up and strongly supported by
programming.
The points here are so important that I want to support them with a careful discussion of an
example episode. This episode involves Tim and Steve’s work on the Toss simulation, which we
briefly encountered in Chapter 9. Here I want to go through this episode in some detail, telling
the whole story so that I can extract the relevant morals.
Like all of the students, Tim and Steve proceeded through a standard progression of tasks,
creating their Drop simulation just prior to embarking on the simulation of a tossed ball. Upon
starting work on the Toss, they commented that the second portion of the motion, during which
the ball drops from its apex to the ground, would be the same as the motion produced by their
Drop simulation.
Tim
It’s a toss. [g. tossing motion]
Steve
All the way up and back down.
Tim
Yeah. So the um - well the drop’s gonna be the same. Right?
Steve
Yeah.
Tim
As it was before.
This observation, that the second half of the Toss motion would be identical to the motion in the
Drop simulation, was common among my subjects.
394
Following this initial discussion, Tim and Steve copied their work from the Drop simulation
and pasted it into their current working environment. At that point, they had the following
procedures:
toss
drop
setup
setup
change
change
change
repeat
Doit
pos 0
vel 0
acc 9.8
10 tick
Doit
Next they worked to fill out the toss procedure and adapt drop to be appropriate for its current
use. The result was as follows:
toss
drop
setup
change
change
change
repeat
drop
change acc 9.8
repeat 10 tick
pos 0
vel 98
acc -9.8
10 tick
Doit
Doit
Recall that, in Chapter 9, I argued that the SEQUENTIAL PROCESSES form was involved in Tim
and Steve’s construction of this toss procedure. The story I told there was that these students saw
the simulation as broken into two parts, a part that takes the ball up and then a call to drop, which
is supposed to take the ball down (refer to Figure Chapter 11. -1).
toss
setup
change
change
change
repeat
pos 0
vel 98
acc -9.8
10 tick
drop
Ball goes up.
Ball comes down.
Doit
Figure Chapter 11. -1. Tim and Steve's toss procedure seen in terms of SEQUENTIAL PROCESSES.
When Tim and Steve executed this simulation, the sprite began by moving upward in the
graphics box, slowing down along the way. This part was as expected. But after the sprite stopped
at its apex, it began to move upwards again, accelerating off the top of the graphics box. Looking
at the graphical display, it was obvious to the students that something was amiss:
Steve
Oh, shoot.
Tim
What happened? Setup. [Runs]
Steve
It kept going the same way.
395
The reason for this incorrect behavior is the change in the sign of acc at the apex of the
motion. To explain: The ball/sprite starts off at the bottom of the graphics box, pointing upwards.
This means that the downward direction is essentially defined to be negative. Thus, when the
program starts, the ball’s initially positive velocity causes it to move upward, and the initially
negative acceleration decreases this velocity, making the velocity smaller and smaller until it is zero
at the apex. At this point, for the ball to come down, we would need the velocity to become
negative so that the ball moves in the negative direction. This requires that the acceleration
continue to be negative. If the acceleration is changed to have a positive value at this point, as in
Tim and Steve’s simulation, this makes the velocity become positive again, and the ball continues
upward.
Without making any comments whatsoever, Tim and Steve immediately modified their drop
procedure so that it made the acceleration -9.8, instead of 9.8. Thus, their program looked as
follows:
drop
toss
setup
change
change
change
repeat
drop
pos 0
vel 98
acc -9.8
10 tick
change acc -9.8
repeat 10 tick
Doit
Doit
This program, when run, produced behavior that was essentially correct and Tim and Steve were
largely satisfied. The only feature of their program that bothered them was that the motion
appeared to end slightly below its starting point. For that reason, they modified their drop
procedure so that tick was only repeated 9 times, rather than 10.
There are some interesting points to made about Tim and Steve’s simulation as it stands at
this point in time. First, because the change statement in drop now changes the acceleration
variable to -9.8—the value it already contains—this statement has no effect and can be safely
deleted. Furthermore, when this statement is eliminated, the above program essentially contains
two sequential repeat statements, both of which repeatedly execute the same tick procedure.
These two statements can thus be combined into a single longer repeat statement. Tim and Steve
recognized these points soon after completing their successful simulation:
Steve
So we don’t even need to have that change, huh? [g. change acc line in drop]
Tim
That can’t be right.
Steve
Cause it’s still the same.
Tim
Then why does it go back down?
396
Steve
I don’t’ know.
Tim
Oh, cause we have an initial velocity right. It starts out positive [g. moves finger up along
path], it just keeps getting more and more negative.
Steve
Oh. So you could just repeat that nineteen times. [g. repeat line in toss]
Tim
Yeah. [w.mod repeat 19 tick in toss]
After this exchange, Tim and Steve acted on these observations to produce this final version of the
simulation, which generates identical behavior to their previous simulation.
toss
setup
change
change
change
repeat
pos 0
vel 98
acc -9.8
19 tick
Doit
After confirming that this new version of the simulation worked, Tim commented:
Tim
We don’t need this. [g. drop procedure] That’s right, cause gravity’s constant.
Now I want to take some time to reflect on what happened in the above episode. Initially,
Tim and Steve had a program that broke the motion into two parts, a part that takes the ball up to
the apex and a part that takes the ball back down. Then this program was transformed into a new
version that treated the motion as a single whole. These two views of a toss are really quite
dramatically different, but is there a sense in which either version is preferable? Have Tim and
Steve learned anything important here?
Fg = mg
F g = mg F = ma
F g = ma
mg
a=
m
a=g
Figure Chapter 11. -2. A free-body diagram for a tossed ball which shows all of the forces acting on the ball.
This diagram is valid at every point along the motion.
To begin to answer these questions, let us first look at these issues from the point of view of a
physicist. Figure Chapter 11. -2 shows a free-body diagram for the tossed ball in flight. The only
force acting on the ball is the force of gravity, Fg =mg. This force can be used to compute the
acceleration of the ball using the relation F=ma to yield, a=g, as shown in the figure.
The important thing to note about this figure is that it applies to every instant along the
motion. It can just as correctly be taken to describe a moment when the ball is on the way up, or
when it is on the way down. This means that the same local physical mechanism generates the
397
entirety of the motion. There is always a constant acceleration downward, which causes the ball to
slow as it moves upward, then to speed up in the downward direction. Thus, there is a deep sense
in which the motion is the same at every point, and it is not “physically natural” to break the
motion into two parts.
This does not mean that it is incorrect to talk about the motion as having two parts; there may
very well be occasions in which this point of view is helpful and illuminating. However, if a student
does learn to see the toss motion as a single whole, then they may be on the way to learning
something profound. The power and subtlety of the formal physics approach here is that it reveals
a common, underlying mechanism that generates the entire motion.
Now let’s think about where Tim and Steve stand with respect to these issues. Initially, it is
not obvious to them that the toss motion could be treated as a single uniform process. This may
not seem too surprising, given the fact that it is subtle to recognize the underlying uniformity. On
the other hand, Tim and Steve have almost certainly received direct and fairly extensive
instruction relating to this issue. The fact that the acceleration is constant and downward
throughout the motion of any projectile is something that is repeated frequently in introductory
physics courses. That Tim and Steve have received such instruction is somewhat confirmed by
Tim’s final statement: “That’s right, cause gravity’s constant.” Nonetheless, in their first pass at
this simulation, Tim and Steve thought that they needed to alter the sign of the acceleration at the
apex. Did they just forget or was there still some lingering sense in which they didn’t really “get
it?” If so, why wasn’t the algebra-based instruction more successful?
The major point here is that programming forces you to engage with these issues in a way that
algebra does not. The question of how and whether to slice up the motion has clear and direct
implications for the program; you have to decide whether to have one loop or two. In algebraphysics, as it is usually practiced, the relevant differences are not so clearly manifested as
differences in the external representation. Given this observation, it is not too surprising that
algebra-physics instruction leaves some undiscovered intuitive cobwebs in this territory, and that
the act of programming tends to expose these cobwebs.
In fact, programming even does a little better than helping students to engage with these
issues, it actually helps them in working them through and getting them straight. Notice that, in
the above episode, Tim and Steve almost seemed to rediscover the fact that a toss can be treated
as a single uniform process. When their first program did not work, they were forced to make the
acceleration a constant throughout. Then, once this change was made, a brief inspection of their
398
program revealed that the two parts of their program could be combined. Thus, the act of
programming a simulation directly supported these students in coming to see a toss as a single,
uniform process.
Let’s take a moment to think about an analogous situation we encountered in algebra-physics.
Recall that when working on the Shoved Block task, students frequently expressed two competing
intuitions concerning which block would travel farther. One intuition was that the heavier block
should travel less far because it experiences a greater frictional force; the second intuition was that
the heavier block should actually travel farther, since it has a greater momentum and is thus
“harder to stop.” Working with algebraic notation and traditional solution methods, the students
were able to resolve this conflict in their intuitions and determine that the heavier and lighter
block travel the same distance. In this way, the use of algebraic notation helps students in
straightening out issues relating to certain types of intuitive knowledge.
The above programming episode is analogous. The writing of a simulation helped Tim and
Steve in dealing with issues relating to a certain type of intuitive knowledge. Thus, the moral of
the story is the same: The activity of symbol use acts to restructure various parts of physical
intuition. Furthermore the key point here is that programming is more likely to have effects on
intuitions having to do with how the world is broken up into processes; programming picks up a
whole new wing of existing intuitive resources.
Tim and Steve’s work on the Toss simulation was not at all anomalous. In fact, three of the
four other pairs explicitly discussed the issue of whether the simulation should involve one or two
loops. Furthermore, nearly the same issue, in one form or another, arose for all pairs while working
on the Spring simulation. As for the Toss simulation, a single tick procedure and repeat loop can
be used to generate the entirety of the Spring simulation.
But this was far from obvious for the students in my study; pairs frequently discussed various
ways of breaking an oscillation into parts. For example, in the following passage, Ned suggests
breaking an oscillation into three separate pieces: (1) a piece during which the block moves from
the center to its far right location, (2) a piece that takes the block from the far right to the far left,
and (3) a piece that take the block from the far left back to the center. These three pieces would
then be repeated to produce multiple oscillations. Thus, Ned’s suggested solution involved three
repeat loops contained within a broader repeat structure. David countered that this complicated
approach was not necessary, arguing that “it should be able to do it on its own:”
Ned
Okay, repeat it however many times it takes to go here. [g. center to right] And then
change all these again. [g. all the change lines in spring] Inside this program. Repeat it so
399
it goes back here. [g. right to left of g-box] Then change 'em again. [g. spring procedure]
And repeat it so it goes back here. [g. left to center] And then repeat the whole thing. [g.
circle around whole screen]
David
Right, I don't think we'll need to. It should be able to do it on its own. [g. block in center]
See, cause if we start off with zero acceleration, velocity would be velocity. And it'll go
once. [g. moves finger short distance to right] Position'll be position plus velocity. Force
would- (0.5) position times K. And then acceleration'd be force divided by mass. And then,
it'll use that acc, right, the next time around?
Ultimately, David prevailed in his argument, and the students created a simulation that
employed a single repeat loop.
spring
TICK
setup
change
change
change
change
change
change
repeat
change vel
pos 0.0
vel 10.0
acc 0.0
mass 20.0
force 0.0
k 1.5
20 tick
Doit
vel + acc
Doit
change pos
pos + vel
Doit
change force
pos * k
Doit
change acc
force / mass
Doit
fd vel
dot
Doit
This simulation essentially uses the relation F = -kx to compute the force at each instant, and then
that force is used to compute the acceleration. Thus, as in the motion of a projectile, there is a
sense in which the same local physical mechanism generates the entirety of the motion.
When Ned and David ran this simulation, it worked quite nicely, producing clearly oscillatory
behavior. But, interestingly, Ned and David were a little surprised that their simulation worked:
David
Dude, how did we do this?
Ned
I don't know. [laughs]
[They both watch their simulation for a moment.]
Ned
Seriously, I don't know. I- I was expecting this to just go that far. [g. motion from center to
right]
Ned and David are certainly accomplished students, and we might think that they should know
very well that the same forces and mechanism apply at each point along the spring’s motion. And,
in fact, they did know enough to create a simulation with a single loop on their first attempt. But
the result was still a little surprising to them; there were still relevant intuitions for them to work
through. It seems that Ned and David (and especially Ned) were still strongly inclined toward
seeing an oscillation as composed of several parts, at least in certain circumstances. Writing the
Spring simulation forced them to begin to sort out these intuitions.
400
An important question arises at this point. We have seen that programming forces students to
engage more directly with a certain class of issues, namely issues relating to how motions should
most appropriately be sliced up into parts. But it may not be the case that students enter physics
learning with conceptual difficulties—or conceptualizations, of any sort—in this territory. Instead,
it is possible that in giving students a language that allows the breaking of motions in parts, we are
actually generating new conceptual difficulties. If this is the case, then it may not be very
important to deal with issues relating to processes; rather, these issues might be better avoided.
But, whether or not students enter instructions with conceptualizations relating to how the
world is carved into processes, it may nonetheless be useful for instruction to deal closely with
these issues. In fact, I laid the basis for such an argument when I described how the carving into
processes can relate, in some cases, to recognizing a single underlying mechanism.
In addition, at least one researcher, Kenneth Forbus, has hypothesized a central role for
processes in the physical reasoning of experts and novices. In his extensive manuscript entitled
Qualitative Process Theory, Forbus introduces two terms (and corresponding hypotheses
concerning intuitive knowledge) that are relevant to this discussion (Forbus, 1984). The first term
that he introduces, following Hayes (1979, 1984) is the concept of an object “history.” A history is
an object-centered description of a motion, bounded in space in time. It focuses on a particular
object and tells us how the various parameters associated with the object, such as its location, vary
over some finite time interval. Furthermore, for Forbus and Hayes, reasoning about the physical
world is largely about fitting together these histories. Hayes states: “Reasoning about dynamics
resembles a process of fitting together a jigsaw of historical pieces in an attempt to fill out
spacetime … .”
The second relevant term introduced by Forbus is “process.” Simplistically, processes act to
cause the changes that happen in a history—they make the history happen. The point is that
histories are, in a sense, only descriptions of what happens in the physical world. Processes provide a
certain type of explanation of these described events.
Forbus’ terms can easily be used to talk about the issues we have been dealing with here. The
question of whether a toss is treated as having one part or two can be phrased as whether we see it
as consisting of one or more histories. And our observation that the motions in the toss and spring
tasks can each be generated by a single underlying mechanism can be described, in Forbus’ terms,
by saying that a single process drives the entirety of these motions.
401
The big point here is that Forbus is essentially claiming that histories and processes play a
major role in how people understand the physical world prior to any physics instruction. If we
accept Forbus’ claim, then this makes it less likely that the conceptual difficulties observed here
are purely an artifact of the use of programming, and it increases the apparent importance of
dealing with related issues in physics instruction. And my argument here is that programming is
better suited for engaging these issues and helping students to get them straight.
Directedness and causation
Another of my informal hypotheses was that there would be consequences of the fact that
programming employs directed statements, rather than symmetric equality relations. This was
born out partly by the fact that BALANCING did not appear in programming. But another element
of my original hypotheses was that this directed aspect of programming statements would lead to
a greater role for causal intuitions in programming. Let us consider the extent to which this
hypothesis is supported by the results.
Since there is no symbolic form that I labeled the “cause” form, the connection of causal
intuitions to my analysis is not immediately obvious. Among the forms that I identified, the ones
that I believe are most closely related to causal relations are the Dependence forms. A little
explanation is required here to make this relationship between dependence and causation clear.
First note that causal relations are asymmetric in the sense that when I say that X causes Y, this
is not equivalent to saying that Y causes X. Furthermore, there is a family of relations that are
asymmetric in this sense. Dependence is perhaps the weakest of these relations; when we say that
one quantity depends on another, all we are saying is that changing the value of the second
quantity will alter the value of the first quantity, all other things being equal. A slightly stronger
asymmetric relation between quantities is what I call “determining.” If quantity X determines
quantity Y, this means that knowing the value of X (perhaps together with some other quantities
that we know about) is enough to uniquely specify the value of Y. Finally, following Horwich
(1987), causal relations can be understood to be like determining relations, but with an extra
“ingredient” added, such as a requirement of time-ordering.
The point of the above story is not to provide a rational accounting of dependence and
causation. Instead, I simply want to argue that observations pertaining to the Dependence forms
are relevant to what I informally called “causal intuitions” in my original conjectures. More
402
specifically, I want to claim that “Dependence” can be taken as a name for a weaker and more
encompassing category of relations that includes causal relations.
So, at least tentatively, we can rephrase the questions here as follows: Are there differences
between programming-physics and algebra-physics relating to Dependence forms? Did these
forms appear more often in programming or was the sense of dependence somehow stronger?
To answer these questions, first recall that, when only forms relating to within-line structure
were included in the programming analysis, the distribution obtained was very similar to that
obtained for algebra-physics. To some extent this indicates that the mere fact that programming
statements are directed does not have much of an effect. Although it precludes seeing symmetric
relations like BALANCING , directedness does not seem to increase how frequently dependence
relations are seen, as compared with the equations of algebra. Furthermore, whether or not we
choose to trust a detailed comparison of the frequency results, we have the observation that
DEPENDENCE
was one of the most common forms in algebra, and it appeared across a variety of
tasks.
This is an extremely important observation. Although algebraic equations are always, in a
certain sense, symmetric, it does not seem to be overly difficult for students—at least the
relatively advanced students in this study—to read asymmetries into these statements. For
example, although the equal sign in F=-kx specifies a symmetric relation, it is not especially
difficult for students to read this asymmetrically as “force is dependent on position.”
Of course, there may be nuance here that I am missing. It may be, for example, that there
really are separate symbolic forms associated with dependence, determining, and causation. If this
were the case, then an analysis that included these distinctions might reveal important differences
in how individual algebraic and programming statements are understood. However, I looked hard
for this variety and could not justify any such diversity, given the analysis principles I laid out in
Chapters 6 and 10. What was most telling was that students employed an extremely limited
vocabulary in talking about the relations I subsumed under DEPENDENCE. In almost all cases
students used one of a small set of stock phrases, such as “depends on” and “is a function of”
(refer to Appendix D).
This null result is quite interesting: It is an important observation that it was not noticeably
more difficult for my students to see an equal sign as specifying an asymmetric relation. However,
this null result is not the bottom line. I still want to argue that there is, in fact, an important sense
in which programming-physics draws more strongly on causal intuitions. To see this, we need to
403
now include the programming forms that were associated with line-spanning structure. In
particular, remember that there was a new form added to the Dependence Cluster that I called
SERIAL DEPENDENCE .
When this form is included in the analysis, then the Dependence Cluster
becomes proportionally more common in programming than it was in algebra. (Refer to Figure
10-5 and Figure 10-6). So what can we make of this? Does this mean that programming is more
causal or, at least, that it draws more strongly on this family of asymmetric quantity relations?
What I want to argue here is that the existence of the SERIAL DEPENDENCE form points to a
sense in which programming-physics really is more causal. To begin this argument, I will say a
little about some relevant episodes that I touched on in Chapter 9. Recall that issues relating to
SERIAL DEPENDENCE
came up frequently when students were working on the Spring simulation.
Ned and David were quite typical in this regard. From their experience with algebra-physics, they
essentially knew that they had available two “relations for force”—F=-kx and F=ma—and they
knew that they would probably need to use both of these relations in their program. However, in
order to use these relations in a program, Ned and David needed to decide how they should be
ordered. The way they phrased the issue for themselves was to essentially ask: Should we use the
position to find the force using F = -kx and then the force to find the acceleration via F=ma, or
should we do this the other way around, finding acceleration and then position? Here’s the tail
end of their discussion:
Ned
Does the force change because the acceleration changes?
David
The force changes cause X changes. And so that, and so=
Ned
That causes acceleration to change.
So, ultimately, Ned and David decided to use the position to find the force and then the force to
find the acceleration. This was reflected in the final version of their tick procedure:
TICK
change vel
vel + acc
Doit
change pos
pos + vel
Doit
change force
-1 * pos
Doit
Doit
change acc
force / mass
Doit
fd vel
dot
Doit
404
* k
Nearly the same issues arose for most pairs while they worked on the Spring simulation. For
example, as we saw in Chapter 9, Adam and Jerry came to see all of the various changes as flowing
from changes in the position of the oscillating block:
TICK
change force
k * pos
Doit
change acc
force / mass
Doit
change vel
vel + acc
Doit
change pos
fd vel
dot
pos + vel
Doit
Doit
Adam
Well basically, the position is ba- um, the acceleration, everything,
Bruce
Hm-mm.
Adam
is based on where this is, the position, [g. pos in change acc line]
…
Adam
And the position is the most important thing here. If we made it the other way around, we
don’t care about the acceleration, we care about where it is.
Jerry
Hm-mm.
Adam
See acceleration [g. change acc line] derives from the position. [g. change force line]
What we are seeing here is that programming forces students to engage with another unique
set of issues—issues pertaining to which quantities change first and which quantities determine
which other quantities. The importance of this observation is driven home if we think about
algebra-physics. When working within algebra-physics, we typically adopt the stance that many
quantities are changing constantly and simultaneously. For example, when we write F = ma, we do
not need to think about whether force or acceleration changes first. In fact, since these and other
quantities are changing constantly, this question does not even make sense within algebra-physics.
In contrast, programming forces you to take changes that happen simultaneously and break them
up into ordered operations. And, in doing so, you have to decide which operation happens first.
This is a step toward an ordered, causal world.
The point here is that, when line-spanning structure is included, it appears that dependence
structure plays a stronger role in programming than in algebra-physics. When we just looked
within individual programming statements, the directedness of these statements did not appear to
be especially consequential for seeing asymmetric relations. However, when these directed
operations are placed within a larger ordered process, then it appears that there are interesting
consequences: In total, Dependence forms are coded more frequently, and new kinds of
405
dependence structure are seen. Another way to say this is that, although the directedness of
individual programming statements may not be especially consequential, it is consequential that
programs involve a series of directed statements. I have argued that students can learn to see chains
of dependent quantities in a series of such directed statements, and simultaneous events get spread
out and ordered in a way that they do not in algebra.
Are causal intuitions problematic for physics learning?
A number of interesting questions arise at this point. Suppose we accept that programming
tends to draw on causal intuitions in a certain manner, and that it forces programmers to break up
simultaneous changes into sequential operations. Then we have to ask: Are these properties of
programming desirable from the point of view of physics instruction?
One reason we might be prompted to ask this question is that there are actually reasons to
believe that the use of causal intuitions in physics may be dangerous or even incorrect. (See diSessa
& Sherin, 1995, for a somewhat extended version of this discussion.) To illustrate this point, let us
think about the equation F=ma. In my discussion above I described how, when programming the
Spring simulation, students were forced to think about which quantity changes first, force or
acceleration. And, in all cases, the students decided that changes in force should lead to changes in
acceleration, a decision that I believe is appropriate.
However, an algebra-physics traditionalist might argue that it is incorrect to say that forces
cause accelerations. The equation F=ma is symmetric in this regard; it doesn’t say that forces cause
accelerations any more than it says that accelerations cause forces. Thus, in introducing causation
or other asymmetric relations here, we may be adding information that is not really part of a
“correct” physical description of the world. Furthermore, adding information may be dangerous
and could ultimately confuse students as they try to learn physics.
There a number of responses to this objection. First, we must keep in mind that a concern with
how someone understands physics is not the same thing as a concern with what is correct physics.
Even the best physicists may sometimes think of forces as causing accelerations, even if they
somehow know that this is not strictly correct. If it is the case that physicists think in this way,
then it may not be a problem to teach students to think causally.
Furthermore, I believe that it is immensely plausible that physicists draw on causal and related
intuitions. One reason that I believe that it is plausible is that I can think of many important uses
for causal intuitions. For example, notice that, although it may not be right to say that forces cause
406
accelerations, it is not really true that there is no asymmetry in the relation between the concepts
of force and acceleration. At the least, there is an ontological asymmetry—forces and acceleration
are two different sorts of entities.
In addition, students do not always keep this ontological asymmetry straight. One way that
this problem manifests itself is in misinterpretations of the equation F=ma. Sometimes students
look at F=ma and see it as a statement that two forces are in balance. (Recall Alan’s inappropriate
BALANCING
interpretation I discussed in Chapter 4). This is incorrect because only the F in this
equation is directly associated with a force or forces that exist in the world. Thus, there is an
asymmetry here that causal intuitions could help students to keep straight: If you have in mind
that forces cause accelerations, then you are less likely to misinterpret this equation. This is the sort
of place that causal/dependence intuitions are useful. Instead of writing F=ma and seeing it in
terms of BALANCING , it is likely better for students to write a=F/m and recognize the
DEPENDENCE
form, with acceleration dependent on force. (See diSessa, 1993, for a similar argument.)
Here’s another way to think about these issues. To a physicist, the equations that are taught in
an introductory physics course do not constitute a haphazard collection; there is a structure to
Newtonian theory, and equations play different roles within that structure. For example, in
talking about Ned and David’s work on the Spring simulation, I said that they thought of F=-kx
and F=ma as two “relations for force.” But, to a physicist, these relations play very different roles
within the structure of physical theory. The equation F=-kx is a “force law;” it tells you how to
compute the force in a particular class of situations. In this case, it tells you how to find the force
exerted by a spring. In contrast, F=ma is the central dynamic constraint relation of Newtonian
physics; it tells you how to find the acceleration once you know the total force acting on an object.
Throughout the experimental sessions I saw evidence that this simple structure—force laws
that provide forces to plug into F=ma—was not evident to the students in my study. Once again,
this is a place that programming (and causal intuitions) could help. In reflecting on how F=-kx and
F=ma should be ordered within a program, students were engaging with relevant issues. Notice
how the two relations were ordered by students in the above Spring simulations; they first
computed the force and then used the resulting force to compute an acceleration. In an important
sense, this ordering reflects the structure of Newtonian physics.
Of course, the same structure could be reflected in the ordering of algebraic statements. For
example, we could encourage students to wonder whether it makes more sense to write:
407
F = − kx
F
a=
m
or
F = ma
−F
x=
k
But I did not see students using line ordering in this way, and I believe it is not generally done in
algebra-physics. Certainly, I did not observe students engaging in extended discussions
concerning how algebraic statements should be arranged, as they did in programming. I believe
that it is fairly clear why this is the case. First, to the extent that students might choose to be
careful in ordering the above two expressions, they would not want to order them to reflect some
physical relations among the quantities; instead, they would more likely want the ordering to
reflect a path from what is given in the problem to what they want to find. Furthermore, algebra
does not necessitate clear decisions—of any sort—about ordering, to the degree that
programming does. Students can just write F=-kx and F=ma, then manipulate as necessary. For
example, a student could observe that, since these two relations involve force, it is possible to
eliminate force and write -kx=ma.
F = ma
F = − kx
− kx = ma
In this case, there is not necessarily any ordering between these two “expressions for force.”
Before wrapping up this discussion, I want to say a little about the fact that programming
forces you to take instants and smear them out in time. Even more than the use of causal
intuitions in physics, this ordering of what are actually simultaneous changes may be thought to be
questionable from the point of view of standard “correct” physics. At the least, we must certainly
worry that students will be confused and will think that these things really do, in some sense,
happen one after the other, in real time.
My argument that this smearing out of instants is not a problem takes the same form as my
argument that it is acceptable to teach students to draw on causal intuitions. Even though it may
not be strictly correct to think of these simultaneous changes as ordered, it is, in many cases,
extremely useful and powerful. I call this strategy “reasoning through pseudo-time.” I do not want
to argue extensively for this here, but I will note that I am not alone in believing that this smearing
through pseudo-time is a powerful resource for reasoning about physical systems, and one that
experts routinely employ. Most notably, deKleer and Brown (1984) introduced the term
408
“mythical time” for what I call pseudo-time, and their accounts of how people reason about
physical systems involve reasoning through mythical time.
In summary, I did not uncover evidence that individual programming statements are more
likely to be given cause-related interpretations than individual equations. However, when linespanning structure is included in the analysis of programming, then there is evidence that
programming tends to more strongly draw on causal intuitions. Questions of “what quantities
change first” and “what quantities determine which others” are issues that come up naturally and
persistently in programming physics, and they are issues that are strongly related to causal
reasoning.
A Comparison Based In Representational Devices
Now it is time for this discussion to turn to a comparison of the representational devices of
programming and algebra. My exposition here will be somewhat less extensive than my
comparison of the forms in each practice, in part because there is significant overlap in the issues
involved. Each of the practices—its forms, devices, and notational systems—is an integrally
related whole.
In this section I will begin by discussing how the results of the device analysis indicate a shift
toward what I will call “physical” devices. Then I will address the implications of the extremely
important fact that, much more than algebra, programming forces an embedding of expressions
in the particulars of a physical circumstance. Lastly, I will conclude with some final musings on the
topic of whether programs are easier to interpret than equations.
A shift to time-varying or ÒphysicalÓ devices
At the level of classes, the most significant difference revealed by the systematic analysis of the
programming corpus was a shift from the Static Class to the Narrative Class. In comparison to the
algebra-physics corpus, Static devices were proportionally less common and Narrative devices
more common. As I have noted, this shift might only be an artifact of individual variation and the
limited task selection. However, I believe that this is a plausible result, in part because it is
consistent with the image of differences that I have been building in this chapter. In particular, this
shift can be understood to be part and parcel of the changes described in my discussion of forms:
409
In moving from algebra-physics to programming-physics there was less of a focus on situations of
equilibrium, and more of a focus on time-varying systems.
But, although these observations are closely tied, we can still learn more and further
understand these differences by looking closely at individual devices. First, recall that the
prevalence of the Narrative Class in programming had a lot to do with
TRACING ,
the
programming-specific Narrative device. TRACING was quite common in programming, accounting
for 20% of all device codings and nearly a third of Narrative codings. Furthermore, when the
TRACING
device was eliminated from the analysis, we saw that the increased prevalence of
Narrative devices in programming disappeared.
This last result seems to suggest that the only difference in the Narrative devices of
programming was the addition of TRACING . However, I believe there are some other important
differences. Table Chapter 11. -1 displays the results relating to Narrative devices for both algebra
and programming, with the TRACING device excluded. A close examination of these tables suggests
that the PHYSICAL CHANGE device was proportionally more common in programming-physics, as
compared to CHANGING PARAMETERS and CHANGING SITUATIONS
Algebra
Narrative
CHANGING PARAMETERS
CHANGING SITUATIONS
PHYSICAL CHANGE
Programming
Count
Percent
128
79
21
28
54%
33%
9%
12%
Narrative
CHANGING PARAMETERS
CHANGING SITUATIONS
PHYSICAL CHANGE
Count
Percent
135
73
18
44
54%
29%
7%
18%
Table Chapter 11. -1. Narrative Devices in algebra and programming.
Taken together, the addition of TRACING and the increased frequency of PHYSICAL CHANGE can
be seen as part of a systematic difference in the nature of Narrative devices in programming.
Recall how the various devices in this cluster differ. In PHYSICAL CHANGE , the narrated changes
follow the motion under study. For example, if we are examining a ball falling under the influence
of air resistance, then we might narrate how the velocity and the air resistance force change as the
ball drops. In contrast, the changes narrated in a CHANGING PARAMETERS interpretation are of a
different sort. Rather than describing changes that occur through the time of the motion under
study, a CHANGING PARAMETERS interpretation describes changes that could be understood to be
occurring across instances of a motion. For example, again considering the case of a ball dropped
with air resistance, a CHANGING PARAMETERS interpretation might describe how the motion would
410
differ if the ball was heavier. Similarly, the CHANGING SITUATIONS device also does not describe
changes that occur through the time of a motion. Instead, interpretations involving this device
compare motions in physically distinct circumstances.
Given these descriptions of the three Narrative devices of algebra-physics, we can see that there
is a sense in which TRACING and PHYSICAL CHANGE go together: Both of these devices involve
changes through the time (or pseudo-time) of the motion under study. What this suggests is that
programming tends to have more interpretations that follow the time of a motion. Furthermore,
as we have just seen, this is true not only because of the addition of a substantial number of
TRACING
interpretations, it may also be reflected in an increase in codings of PHYSICAL CHANGE .
Thus, what we are seeing here is a general shift toward what I call “physical” narratives—narratives
that follow the motion under study.
This shift toward time-varying devices was also manifested as a decrease in Static
interpretations in my programming corpus. The decrease in Static devices is part and parcel of the
shift away from a focus on equilibrium situations; it generally reflects the fact that programmingphysics tends to deal with time-varying situations, and the fact that programming tends to
support “physical” narrations. Interestingly, it turns out that the shift toward time-varying
interpretations is even reflected in individual Static interpretations that appear in programming.
Recall that there was a significant addition to the Static Class, the RECURRENCE device, which
accounted for 40 of the 67 codings of Static devices in programming. Thus, to the extent that the
RECURRENCE
device is unlike other Static devices, Static interpretations in programming will differ
from those of algebra-physics.
In fact, there may be a sense in which RECURRENCE is somewhat unique among Static devices.
Recall that the RECURRENCE device is closely related to GENERIC MOMENT . While GENERIC MOMENT is
associated with the stance that an equation is true at any time, RECURRENCE interpretations involve
adopting the stance that a relation tells you what transpires on each tick or at each instant. Here are
two examples reproduced from Chapter 9:
411
drop
setup
change
change
change
repeat
vel 0
air 0
mass 20
200
change force mass * acc - vel
change acc force / mass
tick
Doit
Doit
Fred
No, we don’t change it there. Air is - the air term is gonna be something that changes
during the calculation or during the fall. Remember that you said it’s dependent on
velocity, right?
…
Fred
So it’s actually gonna change the force each point, each tick.
Greg
Hm-mm.
Fred
So we’re gonna have to put change something.
_______________
change vel
vel + acc
Doit
Tim
So, every time - every tick it gains nine point eight. The acceleration increases the
velocity by nine point eight.
I do believe that the RECURRENCE device belongs in the Static Class. It is certainly not like
devices in the Narrative Class, which involve describing how a particular parameter changes with
some imagined process. Instead, like other Static devices, it is a particular way of projecting an
expression into a moment in a motion. However, there is a sense in which RECURRENCE
interpretations are not “static.” When we see an expression in terms of RECURRENCE, we are seeing
the expression as describing a local event, some localized change. Thus, the dynamic nature of
programming even shows up in a change in how expressions are projected into moments—we see
statements as little local stories about change, rather than as static relations that hold.
Programming and embedding in particulars
Recall that the full systematic analyses of programming and algebra found that Special Case
devices were roughly as common in each practice. However, when devices associated with seeing
line-spanning structured were excluded from the programming-analysis (i.e., the
TRACING
device),
the frequencies suggested that the Special Case Class was somewhat more common in
programming (refer to Table Chapter 10. -41). Notice that, in Table Chapter 10. -41, Special
Case is not only increased relative to Static interpretations, it is also proportionally greater as
compared to Narrative interpretations.
412
Algebra
Narrative
Static
Special Case
54%
35%
11%
Programming
54%
27%
18%
Table Chapter 11. -2. Comparison of devices with TRACING eliminated.
This is another result that I believe is quite plausible given the differing natures of
programming and algebra. In fact, this difference can be seen to reflect one of the most basic
differences between programming-physics and algebra-physics. (See Sherin et al., 1993, for a
similar argument.) In algebra-physics, it is possible to use equations to solve problems in which the
physical situation is only weakly specified. For example, in the Mass on a Spring task that I gave to
my algebra subjects, the students were asked to determine how the amount of stretch depends on
the mass of the object and the spring constant. However, they were not given any specific values
for these parameters. Instead, they simply derived the equation x=mg/k and then discussed the
properties of this expression. In this way, they were able to answer the question without specifying
the numerical particulars of the physical situation.
Furthermore, you can use algebraic expressions in a manner that is even further removed from
specific physical circumstances. In many cases, we can use equations to answer questions without
even having any physical circumstance in mind. For example, it is possible to use Maxwell’s
equations (a set of very general relations that describe phenomena relating to electricity and
magnetism) to derive the fact that waves can propagate in electric and magnetic fields. In that
case, we are deriving a very general feature of electric and magnetic phenomena; we do not need
to be talking about any particular physical circumstance.
In contrast, programming requires us to embed statements in particulars. For example, when
creating their Spring simulations, the students in my study were forced to invent numerical values
for all of the parameters that appeared in their programs. If they did not invent these numerical
values, then they could not run their programs.
This is a simple explanation of the increased prevalence of Special Case interpretations in
programming. Because programming already requires you to deal with equations in a particular
numerical regime, Special Case interpretations may be more natural.
There is still more to this beyond the fact that programming requires you to pick specific
values. When you write a statement in programming, that statement must be embedded within a
program, and it must be embedded in a specific manner. You cannot have the programming
413
statement change acc force / mass written just anywhere on the computer display, it must be
embedded within a program, doing a particular job. This cannot help but suggest that the
statement be interpreted in a specific manner.
Are programs easier to interpret than equations?
So, what’s the bottom line here? I originally began this work with the hypothesis that programs
would be easier to interpret because they can be “mentally run” to generate motion phenomena.
Was this born out in the data? Are programs really easier to interpret than equations?
The answer to this question is: Yes and no. Given the observations I have presented, there are
reasons to conclude that programs are easier to interpret, and reasons to conclude that they are
trickier to interpret. And it’s a little hard to draw conclusions concerning how these factors balance
out. I did find that my programming subjects interacted meaningfully with (constructed or
interpreted) a larger percentage of the expressions that they wrote, a result that could be taken as
suggesting that programs are easier to interpret than equations. But, as I discussed in Chapter 10,
there are a number of other possible causes for this result, such as the fact that algebra-physics
tends to include a large number of derived expressions.
Since there is no way to draw a definitive conclusion concerning which type of expressions are
easier to interpret, I will instead go through and discuss the various factors on each side. To begin,
an important observation that we must keep in mind is that the interpretation of programs was not
always very different than the interpretation of algebraic expressions. In many cases, individual
programming statements and collections of statements were interpreted much like their algebraic
counterparts, with the rest of the program fading into the background. This observation seems to
suggest that programs are no easier to interpret, and maybe even a little harder, since the lines to
interpret in a program must first be selected. In fact, there was often some cleverness involved in
the isolation of a few programming statements for the purposes of interpretation.
However, on the other side of this debate is the observation that the backdrop provided by the
rest of a program and the programming context may help by providing an orientation to the lines
that are focused on. This is essentially the point I made when discussing Special Case devices just
above. The fact that programming statements are always embedded within a program in a
particular manner provides a leg-up on interpreting those statements. Less has to be done to invent
a stance, at least once the program is written, since the interpretation is already partly suggested
by the role that the statement plays within the program.
414
This observation is actually a corollary of a more fundamental point. In general, interpreting
any specific symbolic expression will be easier when there is more support lying around in the
context, and a person thus has to do less to invent an interpretive framing. Furthermore, this is
true not only of individual expressions, but also of classes of expressions: Any class of symbolic
expressions will be easier to interpret to the extent that it is embedded in practices that provide
natural modes of interpretation.
Yet another way to say this is to state that, to some extent, the ease of interpretation of
symbolic expressions depends on the existence of naturally available and useful representational
devices. And there actually is some evidence that programming has an edge in this regard. First,
the fact that the interpretation of a programming statement can be framed by the running
program is an example of this sort of natural support. Because of the fact that programs can be
run, there is a naturally available representational device—the TRACING device. Although this was
not at all the only device in programming, it accounted for 20% of all interpretations, a quite
healthy percentage.
Furthermore, the fact that programs can be run may actually help to make other devices more
natural. Most notably, we saw that PHYSICAL CHANGE interpretations were more common in
programming-physics than algebra-physics. This may be partly because the orientation provided
by the running program provided natural support for this mode of interpretation.
In conclusion, there are points on both sides of this debate. In some respects the interpretation
of programs is not altogether different than the interpretation of equations, and some cleverness is
required in selecting lines and pushing other lines into the background. However, because
programming statements are embedded in particulars there is more support for interpretation.
Furthermore, there are some naturally available and very powerful representational devices that
simulation programmers can fall back on in interpreting programs.
Channels of Feedback
Throughout this chapter and this entire manuscript, I have been focusing on getting at a very
specific class of differences between programming-physics and algebra-physics. I have been
interested in showing how physics understanding will be altered in a certain fundamental sense,
and, in particular, I have been interested in showing how a different “conceptual understanding”
of physics will flow from the specific notational systems employed in each practice.
415
Of course, there are differences between programming-physics and algebra-physics that are
not captured by this analysis. As I have mentioned, I could have paid more attention to what each
of these notational systems allows. For example, algebraic notation allows you to manipulate
expressions to derive new expressions, and programs can be run to generate dynamic displays.
These differences in what is allowed are extremely important for several reasons. First, they in part
dictate what you can do with each of the notational systems. But, even more important from our
point of view, they will also have effects on students’ understanding. This sort of second-order
effect on understanding is not directly captured by my form-device analysis.
Although I will not come close to giving these other issues their due, I do not want to ignore
them altogether. The fact that programs can be run to produce dynamic displays is such an
important feature of programming-physics that it simply must be addressed, if only briefly. And,
in fact, it turns out that there are some comments that I can make without straying too far from
the theoretical framework I have laid out in this document.
To get at these additional issues, I am going to take a very particular rhetorical approach.
Rather than building an orthogonal body of theory, I am going to take the theoretical work that I
have already done—the Theory of Forms and Devices—and I will embed it in a somewhat
broader view. At the heart of this broader view is a new notion that I call “channels of feedback.”
Numerical Channel
Form-Device Channel
Numerical Channel
Form-Device Channel
Graphical Channel
Figure Chapter 11. -3. Channels of feedback in programming and algebra.
To explain channels of feedback I want to begin with an image of two students (refer to
Figure Chapter 11. -3). One of these students is looking at a sheet of equations that he generated
on the way to solving a problem. The other student is looking at a simulation program that she
just wrote. Now I want to ask the same question about both of these two students: How will each
of them decide if what they have just done is correct or “makes sense?”
416
Actually, both of these students have a number of possible approaches they can take in
deciding if they are satisfied with what they have done, and each of these alternative approaches
involves a different “channel of feedback.” Let’s begin with the algebra student. One way that the
algebra student can draw conclusions concerning the sensibility of his work is to plug some
numbers into one of the equations, and then attempt to judge the reasonableness of the numerical
result produced. We saw an example of this behavior all the way back in Chapter 2. In that
chapter, I described an episode in which Jon and Ella were working on a problem in which a ball is
thrown straight upwards at 100 mph. In their work, they ultimately derived the final expression:
vo2
=x
2a
Then, once they had this expression, they plugged in values for vo and a, finding that the ball
reached an altitude of 336 feet. With this result in hand, they mused for a while on whether it
seemed reasonable. In the end they decided that 336 feet was at least the “right order of
magnitude.”
Jon
Ella
Jon
A hundred miles an hour is,, (1.0) If you were in a car going a hundred miles an hour,
[laughs] straight up into the air.
straight up into the air, would you go up a football field? (0.5) I have no idea. Well it doesn’t
sound - it’s on the right order of magnitude maybe.
I call this channel of feedback the “numerical” channel. One way to judge the acceptability of
expressions is to plug in numbers and then judge the reasonableness of those numbers, a procedure
that essentially gives indirect evidence as to whether the expression itself is also reasonable. The
thing to notice about the numerical channel is that it allows judgments to be made concerning the
sensibility of an expression without any attention being paid to the actual structure the expression.
Of course, there are other ways for algebra students to judge the sensibility of their
expressions. In fact, a large part of my work has been to show that students do have the ability to
judge expressions by looking directly at their structure. Here is an example we have encountered
on numerous occasions, including as early as Chapter 2. After deriving a final expression for the
Mass on a Spring problem, Jim commented:
x=
Jim
mg
k
Okay, and this makes sort of sense because you figure that, as you have a more massive
block hanging from the spring, [g. block] then you're position x is gonna increase, [g.~ x in
diagram] which is what this is showing. [g.~ m then k] And that if you have a stiffer spring,
417
then you're position x is gonna decrease. [g. Uses fingers to indicate the gap between the
mass and ceiling] That why it's in the denominator. So, the answer makes sense.
So, Jim did not need to plug in any numbers here, he was able to judge the sensibility of the
expression itself. This, of course, is the sort of behavior that I accounted for in terms of forms and
devices. I call this channel the “form-device” channel.
Thus, we can now start to see where my earlier work fits within this larger picture that I am
building. Forms and devices are the knowledge associated with one of several channels by which
students get feedback on the reasonableness of their work with symbolic expressions. And, more
specifically, it’s the channel associated with seeing meaning in the structure of expressions.
These same two channels—numerical and form-device—are also available to the programming
student in Figure Chapter 11. -3. The preceding chapters can be taken as an argument that forms
and devices play a role in how programs are understood. In addition, the students in my study
did often look at the values in their variables in order to decide if their simulations were correct.
However, there is some subtlety here. In writing their programs, my subjects generally were not
careful to work with meaningful units, so there was no easy way for them to draw on their world
experience to judge the reasonableness of the values produced. Still it was possible for them to
make rough judgments given the relative size of other values that appeared in the simulation. And,
in some cases, it was enough to see that a value was in a certain qualitative range, such as greater
than zero or less than zero.
Now, the central observation here is that programming has an important channel of feedback
that is not available in algebra-physics, what I call the “graphical” channel. This is just the point
that I began this discussion with, that programs can be run to generate dynamic feedback in the
form of a moving graphical display. This constitutes a channel of feedback because, when a
program is run, a student can look at the display and judge if the simulated motion appears
reasonable.
Furthermore, programming’s graphical channel has a number of very nice properties. One of
these properties is that, in certain respects, looking at the dynamic graphical display is a lot like
looking at a real motion in the physical world. This means that, to judge whether these displays
are correct, a student can draw on some of the same capabilities that they use in recognizing and
judging the plausibility of physical phenomena. Compare this to the form-device channel, which
requires a substantial amount of specialized learning.
However, judging the reasonableness of what one sees in the graphical display is not always as
easy as it might first seem. For example, when the motion generated does not fit in a graphics box
418
in Boxer, then the motion “wraps.” This means that if a moving object disappears off the top of
the graphics box, it will reappear at the bottom. On some occasions, this can be quite confusing.
Furthermore, there are difficulties caused by the fact that simulation programs are discrete. And
when these two problems combine—wrapping and discreteness—the situation can become very
difficult. Suppose that a program is causing the sprite to jump in discrete hops that are many
times the size of a graphics box. In that case, the sprite’s location within the graphics box on any
particular jump may appear nearly random.
In fact, this is precisely what happened when students worked on the Air Resistance
simulation, since that task required students to run their simulation for an extended period in
order to see the approach to terminal velocity. Thus, in that case, the dynamic display was not
useful for judging the correctness of the simulation. Instead, students had to watch how the
numerical values of various parameters changed through the time of the motion.
Thus, some learning is required in order to interpret what one sees in a graphics box, and to
know when it is useful to look closely at the graphical display. In practice, only when the
simulation produces smooth and uniform motion that is largely contained within the graphics box
can students draw on their existing world resources to judge the motion. Still the fact that, in
some cases, the motion in a dynamic display can be judged much like a motion in the physical
world is one of the great strengths of programming-physics.
Finally, I just want to add one other piece to this image of channels of feedback. We can think
of algebra and programming as each having some generative machinery. In algebra-physics you can
manipulate existing expressions to generate new expressions, and you can substitute numbers to
obtain results. In programming-physics you can run programs to generate a simulated motion in a
dynamic display. So, one way to think of the broader image I am building here is to imagine a
core of generative machinery with feedback at different places.
The only point I want to make here is that these differences in the generative machinery that
exists in each practice are hugely important—they are at the heart of many of the major
dichotomies in what is allowed by each notational system. And the implications of these
differences are not thoroughly captured by my analysis. For an analysis of these implications we
must look to other researchers and to future work.
419
The Bottom Line
The goal of this entire project has been to say something about how the understanding
produced by programming-physics instruction would differ from that produced by algebraphysics instruction. To begin to get at this, in Chapter 1 I contrasted “direct” and “indirect”
effects of symbol use, and I stated that I was going to be concerned with the direct effects on
understanding that follow from the use of a symbol system. Furthermore, my focus throughout
this work has been specifically on effects that flow from the nature of the notational systems
themselves.
In this work, I have actually discussed two related types of residual effects of symbol use on
the knowledge of individuals. First, I have argued that existing cognitive resources are adapted by
and for symbol use. Here this means that certain of the cognitive resources that students bring
with them to physics learning develop during instruction. For example, as I discussed extensively
in Chapter 5, I believe that the sense-of-mechanism is altered through the use of symbolic
expressions. The fundamental hypothesis there was that resources that get exercised are reinforced
and thus play a more central role in expert intuition.
The second type of residual effect is the development of new knowledge. Here, I have been
principally concerned with the development of one particular new resource, the form-device
system. A key point about this knowledge system is that, although it is inextricably tied to symbol
use, it is not especially “formal” or “abstract” knowledge. Rather, this cognitive resource can be
thought of as possessing the character of physical intuition. And I believe that once this new
knowledge system has developed, it should be considered to be a new component of an expert’s
“physical intuition.”
Altogether, an analysis of the tuning of existing resources and of the new form-device system
tells us about the character of expert physical intuition. With this in mind, what final conclusions
can we draw? Are there any high-level generalizations to be made about how the physical intuition
of a programming-physics students will be different?
First, an important preliminary comment must be made. Since forms and devices are
inextricably tied to symbolic experience, there is a sense in which this entire component of
physical intuition—the form-device system—is fundamentally different in each of the two
practices. The “physical intuition” of programming-physics is rooted in a somewhat different class
of experiences. (Refer to Chapter 7 for a more thorough discussion of related issues.)
420
But, there are generalizations that can be made at a higher level. Given the discussion in this
chapter, I believe that it is appropriate to describe algebra-physics as a physics of balance and
equilibrium, and programming-physics as a physics of processes and causation. Algebra-physics really
trains students to seek out equilibria in the world. Programming encourages students to look for
time-varying phenomena, and supports certain types of causal explanations as well as the
segmenting of the world into processes.
It is important to keep in mind that these are differences in degree only. Within algebraphysics, for example, it is certainly possible for students to talk about processes, even though these
accounts may not be as directly reflected in representational forms. Nonetheless, these are still
important and fundamental differences—these high-level generalizations reflect real differences in
the physical intuition that will result from learning physics through the practice of algebra-physics,
or through the practice of programming-physics.
421
Chapter 12. Summary and Implications
To college and high school students taking their first physics course, it may appear as though the
world of the physicist is populated largely by symbols and equations. In fact, much student
activity in introductory physics does involve manipulating equations and solving textbook
problems. Given the current emphasis in educational reform on “teaching for understanding,” this
might seem to be an undesirable situation. Is time spent by students in this abstract world of
symbols time well-spent? Time spent in this manner might help students to become better
problem-solvers, but does it have any deep import for students’ “conceptual understanding” of
physics?
As we saw in Chapter 1, for physicist Richard Feynman, the answer to this last question is a
resounding “yes.” Recall that Feynman expressed the intuition that a full appreciation of the
“beauty of nature” requires some experience of the physicist’s world of symbols. To know physics
as a physicist knows it, a student must have the experience of writing the laws of physics as
symbolic expressions, and manipulating expressions to get from one expression to another.
In this manuscript, I have adopted this view and attempted to give it precision, given the
vocabulary and tools of cognitive science. I have tried to establish that for students—at least for
moderately advanced students—a page of expressions is experienced as highly structured and
meaningful, with meaning that is not purely formal and divorced from interpreting the world.
Furthermore, I argued that, because symbolic experience has this character, it can contribute in a
substantial way to the development of expert physical intuition.
But an examination of the traditional symbolic world of physics, which is based in algebraic
notation, was only half of this project. As a way of getting at the inter-relatedness of knowledge
and external representations, I compared the standard practice of algebra-physics with a practice
of physics symbol use based around an alternative symbol system: a programming language.
Throughout this work I adopted the stance that a programming language could, in a strong sense,
replace algebraic notation and function as a bona fide symbol system for physics. Then I worked
to determine, empirically, how a student’s physical intuition would differ if they learned physics
through this alternative practice. The analysis uncovered some interesting differences, both in
detail and character, as well as some surprising similarities.
In this final chapter, I will begin with a summary of the thesis, then I will conclude with a
discussion of the various instructional implications of this work. In addition to laying the
422
groundwork for future application of this work in instruction, this discussion of instructional
implications will also provide an opportunity to revisit some of the central themes from a new
angle.
The Thesis, In Review
The chief goal of this work was the comparison of the knowledge associated with
programming-physics and algebra-physics. Understood in the most general possible terms, this
comparison constitutes a huge project. Thus, although I attempted to structure this work so that I
could make quite broad and interesting claims, the study was of necessity focused and limited in
several important respects. First, I concerned myself only with what I called the “direct” effects of
symbol use in physics, the changes in an individual’s knowledge that occur through actual
participation in a symbolic practice. Furthermore, I was primarily interested in direct effects that
are more or less traceable to features of the notational systems themselves, rather than to what
these notational systems allow. For example, I did not do much to study the effects that flow
from programming’s generation of a dynamic display, except to the extent that this played out in
how students interacted with programming notations.
In addition, to perform the comparison at the center of this work, I framed my task in a very
particular manner, as a program in cognitive science. In general terms, the project was thus to
build a model, in the manner of cognitive science, of the knowledge of individuals associated with
each practice. Again, the degree to which I actually attained this very general goal was limited in
certain respects. In point of fact, I only looked closely at a particular cognitive resource associated
with seeing a certain type of meaning in symbolic expressions. The argument was that this
particular knowledge system was a good place to examine how the deep nature of understanding
would differ if we change symbolic practices.
The model of this cognitive resource—the form-device system—constitutes the central result
of this research and was the key to getting at the goals of this inquiry. Because it is so central to
this work, I will take the time here to present a moderately extensive summary of my account of
the knowledge comprising this model and how that knowledge develops:
Registrations. First I introduced the term “registration,” borrowed from Roschelle (1991), which
refers to how people “carve up their sensory experience into parts.” Each type of registration is a
feature of symbolic expressions to which physicists are sensitive in the sense that these structures
423
are consequential for their behavior. I listed several kinds of registrations that are relevant to
symbol use in physics: individual symbols, numbers, whole equations, groups of symbols, and
symbol patterns.
A vocabulary of symbolic forms. Next I introduced the first of the two major theoretical
constructs in this work, what I called “symbolic forms.” A symbolic form consists of two
components: (1) a symbol pattern and (2) a conceptual schema. Taken together the set of forms
constitutes a conceptual vocabulary in terms of which expressions are composed and understood.
The forms in algebra-physics fell into six groups that I called “clusters:” Proportionality,
Competing Terms, Dependence, Terms are Amounts, Coefficient, and Multiplication. The forms
in programming-physics fell into these same six clusters as well as into a seventh cluster, the
Process Cluster.
Of these clusters, the three most common in both programming and algebra were
Proportionality, Competing Term, and Dependence. I will now say a little about each of these
clusters to help remind the reader of the sort of claims I made and examples I provided. First, the
Proportionality Cluster is associated with sensitivity to a particular aspect of the structure of
expressions, whether a given symbol lies above or below the division sign in a ratio. The most basic
form in this cluster is PROP+, in which the symbol appears above the ratio line and the expression is
thus directly proportional to the quantity in question.
PROP+
…x… 
 … 
PROP+ was the most common form in both my algebra and programming data, and it was
employed over a wide variety of circumstances. One common circumstance in which it appeared
was in interpreting derived relations in algebra-physics:
v=
Alan
mg
k
So now, since the mass is in the numerator, we can just say velocity terminal is
proportional to mass [w. Vt∝m] and that would explain why the steel ball reaches a faster
terminal velocity than the paper ball.
The next most common cluster was the Competing Terms Cluster. Forms in this cluster are
associated with seeing the terms in an expression as influences in competition. Two of the forms in
this cluster were among the most common individual forms in algebra-physics,
and BALANCING .
424
COMPETING TERMS
COMPETING TERMS
±
BALANCING
=
± …
The BALANCING form, for example, appeared frequently in student work on the Buoyant Block
task:
Fb = Fg
Jack
Um, so we know the mass going - the force down is M G and that has to be balanced by the
force of the wa=
Jim
=the displaced water
Finally, the third most common cluster was Dependence. Forms in this cluster are associated
with sensitivity to a very basic aspect of the structure of expressions, the mere presence or absence
of a symbol. Both the DEPENDENCE and NO DEPENDENCE forms played a role in Mike and Karl’s
construction of an expression for the coefficient of friction:
µ
µ = µ1 + C 2
m
Karl
Well yeah maybe you could consider the frictional force as having two components. One
that goes to zero and the other one that's constant. So that one component would be
dependent on the weight. And the other component would be independent of the weight.
In arguing for the existence of symbolic forms in Chapter 3, I attempted to establish that, by
positing forms, we can account for certain phenomena that are not well captured by other existing
theories of the knowledge employed in physics problem solving. Among these phenomena,
perhaps the most important was the apparent ability of students to construct novel expressions
from scratch. Other important phenomena related to the details of how students write and
interpret expressions. For example, one interesting observation was the “qualitative equivalence” of
certain expressions for students. This was illustrated by Mike and Karl’s inability to choose
between two possible expressions for the spring force:
F = kx ; F = 1 2 kx 2
Karl
Okay, now, qualitatively, both K X and half K X squared do come out to be the same
answer because as (…) looking at it qualitatively, both half - both half K X squared and K
X, um, you know, increase as X increases.
The phenomenon of qualitative equivalence was also illustrated by students’ assertions that it was
not possible to easily decide between the expression v = vo + at and other incorrect alternatives.
Mark, for example, thought that the following expression might be correct:
425
v = vo + 12 at 2
Mark
Cause we have initial velocity [w. circles vo term] plus if you have an acceleration over a
certain time. [w. circles 1/2at2] Yeah, I think that's right.
Symbolic forms account for these observations by specifying the particular aspects of expression
structure that are significant for students.
Forms also provided the major categories for comparing programming-physics and algebraphysics. I developed a method for systematically associating forms with events in each data
corpus, and then I obtained results for the frequency with which each type of form appeared. In
both practices, there were no clear winners amongst the various clusters, but a few individual forms
accounted for a large percentage of the codings. In algebra-physics, the most common forms were
PROP+, DEPENDENCE, COMPETING TERMS,
and BALANCING . In programming-physics, this ranking was
essentially duplicated, but with BALANCING replaced by BASE ±CHANGE .
Some of the most important differences between programming-physics and algebra-physics
were reflected in the forms that appeared exclusively in one practice or the other. Notably, the
BALANCING
form, one of the most common forms in algebra-physics, did not appear at all in
programming-physics. There were also a number of new and quite common forms in
programming, including a whole new cluster of forms, the Process Cluster. Process forms involve
seeing a program as a series of ordered processes. In addition, there was a new form added to the
Dependence Cluster, the
SERIAL DEPENDENCE
form. This form involves recognizing chains of
dependent quantities in programs, and it is tied to the ordered and directed properties of
programming statements.
Representational devices and the interpretation of symbolic expressions. In addition to symbolic
forms, I argued that physics initiates possess a complex repertoire of interpretive strategies that I
called “representational devices.” Representational devices are somewhat closer to observable
aspects of behavior than symbolic forms and, in fact, I maintained that they are directly reflected
in the structure of interpretive utterances. For this reason, the project to identify representational
devices involved, in part, the collecting and categorizing of interpretive utterances.
The observed devices fell into three classes: Narrative, Static, and Special Case. In Narrative
devices, some quantities or other aspect of the circumstance are imagined to vary and the result of
this variation is described. The most common Narrative device—and the most common device
overall in both programming and algebra—was
CHANGING PARAMETERS.
In this device a particular
quantity that appears in an expression is selected to vary while all others are held fixed:
426
x=
Jim
mg
k
Okay, and this makes sort of sense because you figure that, as you have a more massive
block hanging from the spring, [g. block] then you're position x is gonna increase, [g.~ x in
diagram] which is what this is showing. [g.~ m then k]
In the second class of devices, the Static Class, an equation is projected into an instant in the
motion under consideration. For example, in the SPECIFIC MOMENT device, a student adopts the
stance that an expression is true at one particular instant:
Cv 2 = mg
Ella
The terminal velocity is just when the - I guess the kind of frictional force from the air
resistance equals the gravitational force?
Finally, in Special Case devices a student considers a special case out of the range of possible
circumstances under which an expression might hold. For example, in the LIMITING CASE device, a
student considers a circumstance in which one quantity takes on an extreme value, such as zero or
infinity:
Ae −v
Jack
So, you could say the probability if you're traveling at infinite velocity is very extremely
small.
In addition to allowing me to enumerate these interpretive strategies, the project of
categorizing interpretive utterances also provided insight into the nature of these utterances. I
argued that it was appropriate to characterize these interpretations as involving “embedding”
rather than correspondence. This meant that the equation was set within a context or framework
that was larger than any listing of the entailments of the equation. And I also observed that the
“meaning” embodied in these utterances was not a simple function of the arrangement of marks
and, instead, was strongly context dependent. For example, we saw that the interpretations that
students gave were sometimes directly related to the question they had been asked.
As part of my comparison of programming-physics and algebra-physics, I compiled a list of
the devices observed in each practice and measured their frequency of use. The analysis revealed
two devices in programming that did not appear in algebra. One of these new devices, TRACING ,
was the second most common device overall in programming. TRACING interpretations are closest
to what we might think of as the “mental running” of a programming. The other new
programming device, the
RECURRENCE
device, was an addition to the static class in which a
statement is seen as describing what transpires at a particular instant of a motion. RECURRENCE
427
interpretations were frequently—though not always—associated with true recurrence relations,
such as the line that computes velocity in the Tick Model: change vel vel+acc.
In both practices, devices in the Narrative Class were most common, with Static devices
second and Special Case devices appearing only rarely. I argued that this result is quite interesting
because, while Special Case and Static interpretations bear some relation to relatively explicit
aspects of practice, Narrative devices seem to be altogether tacit. Students may occasionally be
taught that it is a useful heuristic to “consider a special case,” but they are rarely or never explicitly
taught that it is useful to select one quantity in an expression and imagine that it varies.
Although the clusters were ranked the same way in each practice, the distribution of devices
was nonetheless somewhat different. The results suggested a shift from Static devices to the other
two classes in programming, a shift that can be seen as relating to the general shift from static to
time-varying situations. In addition, there was a shift in the distribution of devices within the
Narrative class toward what I call “physical devices”—devices that embed the expression in a
process that follows the motion under study.
The refinement of physical intuition. In addition to describing the form-device system, I
speculated, mostly in Chapter 5, about the origins of this knowledge during physics learning. I
argued that this cognitive resource develops out of existing resources, including intuitive physics
knowledge and knowledge relating to the solving of algebra problems, and that these previously
existing resources continue to exist and to be refined. Furthermore, I took the stance that “expert
intuition” consists of these old resources— especially the sense-of-mechanism—somewhat refined
for use in expert practice, as well as some new resources, including symbolic forms. I described
how the refinement of the sense-of-mechanism and the development of symbolic forms could, in
part, be traced to experiences of symbol use.
Taken together, registrations, forms, and devices, as well as speculations about the
development of this knowledge, constitute what I have called The Theory of Forms and Devices.
I must again emphasize that, perhaps even more than specific results concerning differences
between algebra and programming, this framework is the central result of this work. For that
reason, I will conclude this summary with a couple of more general comments about the
framework.
First, the mere fact that the phenomena I have identified exist deserves extra emphasis. The
central claim of this work is that students—at least the moderately advanced students in this
428
study—have the ability to see a certain type of meaningful structure in expressions. Equations
aren’t only memorized, and they aren’t only associated with physical principles; the arrangement
of symbols in an expression “says something” to students. For this reason, students can build new
expressions from scratch. And, because the experience of symbol use is meaningful, this adds to
the plausibility of deep effects on the nature of physics understanding.
This last observation takes us to the most important point of all: The differences between
programming-physics and algebra-physics captured here constitute deep and fundamental
differences in individual knowledge. When students enter instruction, existing knowledge is
refined and adapted for symbol use, and new knowledge develops that is inextricably rooted in
experience with symbols, knowledge that I have argued should be considered a component of an
expert’s “physical intuition.” Thus, the real nature of developing physics intuition will be different
if you learn physics through one practice of symbol use or another, through algebra or
programming. In the first case, the result is an intuition suited to seeing a physics of balance and
equilibrium; in the other, the result is an intuition suited to a physics of processes and causation.
Though these must be understood to be differences of degree, they are nonetheless deep and
fundamental differences.
Instructional Implications
As the final piece of this dissertation, I will reiterate some of my major themes from a
specifically instructional orientation. Throughout this work I have occasionally made comments
concerning instructional implications, but here I will explicitly take up this focus and draw these
observations together. This discussion will include implications for traditional physics instruction,
as well as some final comments concerning what we can conclude about the possibilities for
instruction based on programming-physics.
Implications for traditional algebra-physics instruction
A full half of this manuscript was devoted to producing a description of algebra-physics that
could be compared with the account of programming-physics. But this description of algebraphysics is not only useful for what it tells us about programming, it is also useful in its own right.
In fact, with my account in hand, we are in a position to think about some of the most important
problems faced by algebra-physics instruction. Among these problems is the observation that
students often appear to write and manipulate equations without understanding. Almost certainly,
429
it would be worthwhile to work toward helping students to use equations with understanding, at
least more often than they do currently. However, this will not be an attainable goal if we do not
begin with some knowledge of what it means to use equations “with understanding.”
In this manuscript I have essentially described some aspects of meaningful symbol use in
physics, which can now give us more of an idea of the sort of behavior we are trying to attain. I
have pointed to some phenomena of interest, so we know what it looks like and where it happens.
I have described interpretations of equations; so now we know that there may be no single way of
associating meaning with equations, and that interpretation is a complex and context sensitive.
And I have characterized one knowledge system that allows students to see meaning in equations.
When interpretation fails
This work can also help us to be recognize a new class of student difficulties. Now that we
know where to look for understanding of equations, we can be ready to identify cases where this
understanding goes awry. This has not been a major focus of this work, but I did, on many
occasions, observe students giving weak or incorrect interpretations of equations. There were a few
examples of this behavior in Chapter 4, when I discussed what I called inconclusive and
inappropriate interpretations. In an inconclusive interpretation, a student sets out to produce a
certain type of interpretation of an expression, but the interpretation fails to produce a useful
result. An example I gave is the following excerpt taken from Jon and Ella’s work on the Running
in the Rain task, in which they try to interpret their expression for the total rain. The problem with
this interpretation is that, when they imagine that vR increases, this implies that one of the two
terms increases, but the other decreases:


h
w

#
drops


+
=
wx
vd2 v R 
 A⋅s  
 v d 1+ 2

vR


Ella
Okay, V R is bigger. Then (10.0),, This is gonna get smaller, [w. arrow pointing to second
term] but this is gonna get bigger. [w. arrow pointing to first term] I think.
Jon
Yeah, I think - that gets smaller [g. second term] cause V walking is smaller // than V
running.
In contrast, in an inappropriate interpretation, a student makes a statement that is, strictly
speaking, incorrect. These inappropriate interpretations included Alan’s BALANCING interpretation
from the Shoved Block task, which I have mentioned on numerous occasions. This interpretation
430
is incorrect because, among other things, it implies the existence of two horizontal forces where
there is only one:
µmg=ma
Alan
What Bob didn't mention is that this is when the two forces are equal and the object is no
longer moving because the force frictional is equal to the force it received.
The fact that the interpretation of an expression can be difficult and problematic is extremely
important from the point of view of instruction. Throughout this work I have primarily
emphasized the positive—I have tried to draw attention to the fact that students do have some
resources for using expressions with understanding. But, once we are clear on the fact that these
resources exist, we should begin to look for limitations in these resources, as they are possessed by
students.
Because limitations in students’ form-device knowledge is an important topic, and because I
have not emphasized this topic in the other chapters of this manuscript, I want to take a moment
to discuss one persistent difficulty that I observed. This difficulty has to do with how students
understood (and didn’t understand) the equation v = vo +at. This equation is interesting from the
point of view of this work for a number of reasons. To begin with, this is one of the first equations
that students learn in an introductory physics course.
In addition, to even the most mildly expert physicist, the expression v = vo +at is extremely
simple and straightforward. Here’s why: This equation gives the velocity as a function of time for
motion with constant acceleration. The first term, vo , is the initial velocity of the object, the
velocity it has at time t=0, and the second term adds on the amount that the velocity changes in a
time t. Furthermore, this change is pretty easy to compute since acceleration is defined to be the
change per unit time of the velocity and, in this case, the acceleration is presumed to be constant.
Thus, we just multiply the acceleration by the time to find out how much has been added to the
velocity. Another way to understand this is to think of the acceleration as the rate at which the
velocity changes. To get the total change, we multiply this rate by the time to give at.
As we have already seen in Chapter 3, this equation is not quite so straightforward for
students, even the moderately advanced students in my study. While working on the Vertical
Toss problem, students often struggled to regenerate this expression, sometimes writing
expressions that were incorrect. For example, Mark and Roger wrote the following equation and
commented:
431
v = vo + 12 at 2
Cause we have initial velocity [w. circles vo term] plus if you have an acceleration over a
certain time. [w. circles 1/2at2] Yeah, I think that's right.
Mark
As I argued in Chapter 3, this statement suggests that Mark is interpreting this expression up to
the level of the BASE +CHANGE form. He knows that there should be a term corresponding to the
initial velocity and a term that describes how the velocity changes over time due to the
acceleration. However, since his specification is only at the level of BASE +CHANGE (plus some
DEPENDENCE
on acceleration and time), he is not capable of stating precisely what should be in the
second term. Instead, he must resort to just plugging in a piece of a remembered expression that
happens to include acceleration and time.
This is not meaningless symbol use, and the use of BASE +CHANGE here is not incorrect. Rather,
the problem is that Mark and Roger’s understanding is much weaker than it could be. To a
physicist, the precise form of the at term is obvious, given the definitions of velocity and
acceleration. Certainly we would like students to approach the point where this expression makes
as much sense to them as it does to a physicist. In terms of my framework, this means
understanding at in terms of the INTENSIVE •EXTENSIVE form, or in terms of the related rate•time
template. (Refer to Chapter 6 for a discussion of this template.)
Mark and Roger’s situation was entirely typical. At best, students were only able to give
BASE +CHANGE
interpretations of v = vo +at. Furthermore, when pressed, every student said this
expression would have to be remembered or derived from some other remembered equation, such
as the following expression for how the position varies with time in cases of constant acceleration:
x = xo + vot + 12 at 2
Jack
Well you have to remember something to know how to approach the problem, it's just a
question of which um equation you remember of motion. Um, given (1.5) Um given, if you
remember this one [g. equation] you can pretty much derive them all.
Because I felt that my subjects’ inability to see v = vo +at as obvious was quite interesting, I
asked a number of related questions during the Quick Task component of the study. One of these
related questions was the following, rather bald, prompt:
Some students say that the equation v = vo +at is “obvious.” What do you think?
Given what I observed during the earlier portions of the study, student responses to this question
were not too surprising. For example, here is a portion of Mike and Karl’s comments.
432
Mike
Well yeah it is obvious because, well velocity will equal to V naught if it's not being
disturbed. But if it's being acc- If there's a f- acceleration action on it, then uh- and that's
constant, you know then velocity will be decreasing as time goes on. Or increasing,
whatever it works, I mean whichever it does. So, it's like whatever it is and then plus a
correction on the acceleration. So yeah it makes sense. It's obvious, yes it is.
…
Karl
What's obvious to me is that you have the final velocity is obviously going to be equal to
the initial velocity plus however much, however faster it gets. That's what's obvious to me.
What's not necessarily positively obvious is that the amount, the amount that it gets
faster is the acceleration times the time.
Here, Mike and Karl do seem to agree that v = vo +at is “obvious,” but only in a certain limited
sense. Mike says it is obvious because there should be a “V naught if it's not being disturbed” plus
a “correction” that has to do with the fact that it’s accelerating over time. This is an interpretation
at the level of BASE +CHANGE . But Karl is very clear in stating that, even with this interpretation, the
precise form of the second term is not entirely obvious: “What's not necessarily positively obvious
is that the amount, the amount that it gets faster is the acceleration times the time.”
These observations become even more interesting in light of student answers to another
question that I asked during the Quick Tasks:
5.
Imagine that we’ve got a pile of sand and that, each second, R grams of sand are added to the pile.
Initially, the pile has P grams in it. Write an expression for the mass of the sand after t seconds.
The idea here is that there is a pile of sand whose total mass grows linearly with time at a rate R,
starting with an initial mass of P. Thus, the correct answer to this task is the expression M = P+Rt.
Notice that this expression has precisely the same structure as v = vo +at, both involve an initial
value and a term that grows with time at a fixed, linear rate.
Now, the interesting observation is that students found the Pile of Sand question to be utterly
trivial. In every case, including Mike and Karl, the pair of students read the question, then one of
them picked up a marker and wrote an expression equivalent to M = P+Rt, without a single
comment. Furthermore, they generally seemed perplexed that I had even asked such a simple
question.
The point is that it appears that these students have nearly all the resources they need to see v
= vo +at as trivial and obvious, just as it is obvious to a physicist. But, somehow these resources are
not getting engaged here; students do not seem inclined to understand v = vo +at in the same way
that they understand M = P+Rt. For students, there is some mystery left in v = vo +at.
433
Reforming algebra-physics instruction
What’s the moral of this story for algebra-physics instruction? Well, we could take this as just
another example of how students do not really understand physics expressions, even the simplest
ones. Then, the moral might be that work with equations is never going to be a good way for
students to develop a “conceptual understanding” of physics.
But this is more than a story about failure. First of all, although their understanding may not
have been up to the standards of what we ultimately desire for instruction, the students in my
study did have things to say about the meaning of v = vo +at. They did not only write it from
memory, they also understood it as expressing a base plus a “correction” due to the acceleration.
Furthermore, consider the observation that the Pile of Sand task was very easy for students. This
(along with the rest of my observations) constitutes evidence that students do have many of the
resources that are required to have a more thorough understanding of v = vo +at. It just seems that,
for some reason, these resources are not being engaged in this case. I therefore believe that the
moral of the v = vo +at story is this: Although students may fail to use expressions with as much
understanding as we would like, they possess many of the necessary resources; thus, there may be
some plausible steps we can take to get these resources engaged more frequently and at the right
times.
Even if students possess many of the necessary forms and devices, getting these elements
engaged in the right places may not be an easy instructional goal. The form-device system is not
just an unstructured collection of elements; these elements are tied to each other, to other
cognitive resources, and they are tuned to be cued in certain contexts and not others. Thus,
learning related to the form-device system is not simply a matter of learning individual forms and
devices. A big part of what students must learn is to adopt particular stances to individual physics
expressions such as v = vo +at, and this learning is likely to be tied up with other difficult
conceptual issues. For example, it may not be easy for students to learn to treat acceleration as a
“rate.”
But, presuming that physics instructors accept the task of getting students to understand
expressions in a more useful and appropriate manner, what should they actually do? How,
specifically, should instruction be changed?
First, note that the observations presented here really can help, since just being able to
recognize and understand the problems to be addressed is important. Once we have understood
that there is a problem with how students understand v = vo +at and we have undertaken to
434
improve that understanding, then it is at least possible that there are some relatively
straightforward steps we can take. For example, I did not try this in my study, but it seems
plausible that pointing out the parallel between M = P+Rt and v = vo +at could have been a useful
component of instruction designed to help Mike and Karl better understand the latter expression.
In addition, knowing about specific forms and devices may help. We could, for example,
consider the interpretive strategies embodied in representational devices to be a target of
instruction. Of course, it almost certainly will not be sufficient to just tell students about
representational devices; in the long run, the design of new instructional strategies that takes my
theoretical results into account is going to require careful thought and experimentation. However,
there are some reasonable and modest first steps that are not at the level of “just tell them.” One
way that students learn is by watching instructors as they “model” appropriate behavior. For
example, physics instructors frequently model how to solve various types of physics problems by
standing at a blackboard and working through some problems while students simply watch. In a
similar manner—perhaps as part of modeling problem solving—instructors could strive to model
the interpretation of expressions. This means simply that, in addition to writing and manipulating
equations while students watch, they can point to equations and say something about what those
equations mean. Of course, many instructors probably do this to some extent already, but a little
more frequency and emphasis would likely help.
Ultimately, if future research confirms the observations presented here, we will probably want
to do better than just an increased attention to modeling this behavior; we will want to redesign
instruction to take into account the theory and observations I presented. A possibility that I believe
deserves serious consideration is the creation of a new course in mathematical modeling, which
would precede physics instruction in the curriculum. The purpose of this course would be to teach
students to invent simple types of mathematical models, and to use equations to express the
content of these models, not necessarily within the domain of physics. Although one of the aims
of this course would be to help students to develop the requisite form and device knowledge, this
does not mean that the forms and devices described in this document would be explicitly
mentioned in the course, or even that there would be a recognizable one-to-one mapping between
particular knowledge elements and course material. Instead, a successful course would likely focus
on various heuristic strategies and formally recognized categories, such as the distinction between
linear and quadratic relationships.
435
The underlying presumption here is that a mathematical modeling course would give students
a leg-up when they got to physics instruction. For a student with some math modeling
background, each physics equation would not present a new and individual challenge; instead, the
student would be able to see physics expressions as instances of particular types of models. For
example, Mike and Karl would learn to see v = vo +at as an instance of a model with linear growth
in time.
From a certain perspective, this is really what physics is all about, building a certain class of
models of the physical world. When we worry about student misconceptions and related
qualitative understanding, we are essentially worrying about the particulars of certain particular
models, albeit some very important ones. But, it is open to question whether this should be
considered to be the most important target of physics instruction.
In fact, I am not alone in arguing that we should think of learning to build mathematical
models of the physical world as a central goal of physics instruction. For example, Hestenes
(1987) adopts a similar view, arguing that the structure of the models inherent in physical
principles and equations should be made explicit for students. Furthermore, this view can be seen
as part of a more general movement that takes models, not necessarily formulated as equations, as
the targets of physics instruction (e.g., White, 1993a, 1993b; White & Frederiksen, in press). My
contribution to these movements is twofold: (1) the proposal to teach a course in mathematical
modeling, understood more generally, prior to physics instruction, and (2) some new theoretical
resources for understanding the goals of a course in mathematical modeling, namely the learning
of forms and devices.
Some final thoughts on ÒConceptual PhysicsÓ
As instructors, we can see ourselves as faced with a choice: either make student symbol use
more meaningful, or use methods that do not spend as much time with equations. In proposing
the teaching of mathematical modeling, I am implicitly choosing the former alternative. However,
as I discussed in Chapter 1, there are approaches to physics instruction that have made the
opposite choice. These “Conceptual Physics” approaches do not use equations or, at least, do not
use them very much. The rational behind these alternative approaches is clear: Do not spend as
much time on symbol use. Instead, take time to attack the difficult conceptual issues of physics.
In Chapter 7, I discussed how the understanding produced by Conceptual Physics instruction
would differ from more traditional, equation-based instruction. Here, I will focus on summarizing
436
some instructional considerations. First, with regard to tuning physical intuition and addressing
certain types of qualitative understanding, Conceptual Physics instruction may well have some
advantages. Recall that, in traditional physics instruction, physical intuition develops to support
and complement symbolic methods. Thus, when the “crutch” of symbol use is removed, a more
thoroughly functional physical intuition could result, as opposed to a physical intuition that just
“fills in the gaps” in a symbolic practice.
Of course, something is lost if we eliminate symbol use from instruction. The most obvious
losses have to do with the power of symbolic methods. Equations are compact and precise
statements of physical laws that are relatively easy to remember, and that can be manipulated to
derive new expressions. In addition, the use of equations is very powerful for obtaining numerical
results.
But there would also be some more subtle losses. I argued that an expert’s physical intuition
includes a component that is rooted in symbolic experiences. This symbolic component of
physical intuition is a real part of an expert’s physical sense. If physics is taught without much
symbol use then this component of physical intuition will simply be lost.
There is no escaping the fact that Conceptual Physics instruction teaches something different
than more traditional instruction. It leads to students having different capabilities and results in a
divergent sense of the physical world. But this observation that Conceptual Physics instruction is
not just an alternative route to the same end does not necessarily indicate a problem. I believe that
it is acceptable for us to be flexible in our instructional goals; learning about qualitative features of
the world has value, just as does learning to mathematize the world. Furthermore, Conceptual
Physics instruction likely has substantial value as a way to prepare students for more traditional,
algebra-physics, instruction.
This viewpoint really lies at the heart of the orientation that drives this work. My purpose in
arguing that understanding will be altered in a fundamental respect is not to deflect us from
designing alternative approaches; in fact, my purpose is quite the opposite. I want to free
educators, especially those with some domain expertise, to redesign subject matter for instruction.
Toward that end, I have tried to develop theoretical resources that can help us to understand the
consequences of alterations to subject matter, like that embodied in Conceptual Physics.
437
Prospects for programming-physics
I have been especially interested in understanding the strengths and limitation of a particular
proposal, the use of programming in physics instructions. In this regard, this work is still very
preliminary, though I hope to have made some progress. First, I have tried to at least give the
reader a taste for a true practice of programming-physics: We have seen what the activities could
be like, at least within a limited domain.
In addition, I have begun to outline the deep ways in which programming-physics is a
different physics than algebra-physics. I argued that each practice tends to focus on different sorts
of phenomena and to draw on different aspects of physical intuition. The bottom line here was
my characterization of programming-physics as “a physics of processes and causation,” and
algebra-physics as “a physics of balance and equilibrium.” These descriptions do not necessarily
argue for one instructional practice or the other, but they do help us to understand the
consequences of each approach.
Even with only my descriptive comparisons in hand, there may be some real arguments to be
made for or against the use of each practice. A big issue here is the use of expressions “with
understanding.” Although I have taken pains to argue that students do have a surprising degree of
ability for seeing meaning in expressions, this does not mean that students do not frequently fall
into meaningless symbol manipulations, particularly early in the instructional cycle. Thus,
programming-physics might have an important advantage if it turned out that it was easier to use
programming expressions with understanding. I did not draw strong conclusions concerning this
issue, but we did encounter some reasons to believe that programming statements might be easier
to interpret. This boiled down to the observation that the programming context provides support
for the interpretation of expressions. Specifically, this support derived from the fact that
programming statements are always embedded in a program that includes many specifics of the
circumstance under study, and from the fact that the programming context may include more
naturally available devices.
But, many of the important pros and cons of programming-based instruction were not wellcaptured by my analysis. Chief among the elements left out by my examination was some analysis
of what is allowed by each symbol system. With programming, a student can tackle some
phenomena — particularly time-varying phenomena — that are not easily examined with algebra.
Furthermore, the same, core numerical methods can be applied to a wide range of these
phenomena, thus tending to emphasize and make plain the core simplicity of Newtonian theory.
438
In addition, as I discussed in Chapter 11, the use of a programming language allows the
generation of a dynamic display, which provides a very powerful channel of feedback.
Of course there is a flip-side to this story; there are activities and analyses allowed by algebra
that are not allowed by programming. Some phenomena are more easily tackled in algebraphysics. And algebra allows you to manipulate expressions to derive new expressions.
It seems that both programming-physics and algebra-physics have their strong points. Thus,
one conclusion that we can draw is that optimal instruction would draw on both of these practices,
mixed together in a manner that emphasizes their strengths. Given this conclusion, what are some
strategies for instruction? How might we draw on these two practices, would we teach them
together? Or would it be best to teach one first and then the other?
There is certainly no definitive answer to this question, but I do want to emphasize one
strategy that is especially promising. I believe that programming-physics could provide a useful
entry into physics learning, especially for younger students who do not know algebra or are not
yet adept in its use. In fact, the Boxer Research Group at UC Berkeley has implemented
programming-physics instruction with some younger students, including a course for sixth graders
and a pre-physics course for high school students (diSessa, 1989; Sherin et al., 1993). In both
courses, students were able to engage readily in the construction of simulations. Even the sixth
grade students created many simulation programs and these programs did useful work for them,
functioning as precise statements of models of physical phenomena.
To give a feel for how this worked, I will tell a brief story from our high school course. (This
story is more fully recounted in diSessa and Sherin, 1995.) Our high school physics course was
preceded by a one month introduction to Boxer programming. Then, on the first day of the
physics course, the teacher conducted an open discussion of a number of example motion
phenomena. Included in these phenomena was the motion of a ball that is held above the ground
and then simply released. Next, on the second day, the students were asked to program
simulations of some of the motion phenomena. They were not given much prompting, other than
what had already come out in the discussions of the previous day. Nonetheless, the students did
not complain and they went right to work. Over the course of this single classroom session, quite a
variety of interesting simulations were produced, including a number of distinct simulations of the
dropped ball motion.
To make a complicated story simple, the students essentially produced three types of
simulations of this motion: (a) a program consisting only of a series of forward commands and
439
with no particular pattern in the step increments, (b) a simulation in which the amount that the
sprite steps forward on each iteration is multiplied by a constant factor, and (c) a simulation in
which a constant value is added to the amount that the sprite steps. Examples of each of these
three types of simulation are shown in Figure Chapter 12. -1.
ho
fd
fd
fd
fd
fd
fd
fd
fd
fd
fd
fd
fd
fd
fd
(a)
Doit
1 dot
3 dot
5 dot
7 dot
9 dot
11 dot
13 dot
14 dot
15 dot
16 dot
17 dot
18 dot
19 dot
20 dot
n
pace
D1
112.455
20
21
Data
Data
Data
go
go
repeat 18
repeat
fd
n * 1.3
dot
fd
D1
dot
Doit
Doit
Doit
change n
pace
change D1
n * 1.3
(b)
D1 + 1
Doit
Doit
Doit
Doit
(c)
Doit
Doit
Figure Chapter 12. -1. Examples of the three types of simulations of a dropped ball.
Over the next few weeks, the students worked on better understanding each of these
simulations and, with experimentation, they slowly made progress toward deciding which of the
models more properly describes the physical world. In this way, programming provided a way for
students to do some real physics in the very first days of instruction. They developed their own
models and then expressed them in a formal language. Furthermore, once these programs were
written, they functioned in future classroom discussions as precise statements of the associated
models. Thus, programming can provide an entry into physics that does not use algebraic
notation, but which still preserves many of the benefits of symbol use, such as the ability to make
precise and formal statements of models. For this reason, programming has great potential as a
way for younger students, even sixth graders or younger, to learn physics within a practice that
includes many of the important properties of instruction based on algebraic notation.
The Symbolic Basis of Physical Intuition
As humans, we populate our environment with our own inventions. The very nature of our
experience in the world depends on what we have put there, as well as socially and culturally
determined ways of interacting. How can this not leave an indelible stamp on an individual’s
knowledge? Think of some basic examples. We know much about the artifacts in the world — we
know how to operate our computers, how to drive our cars, and how to play the piano. Clearly,
there is no doubt that our minds are cluttered with knowledge relating to human inventions.
440
The only question is how deep this all goes. Are there only effects at the level of details? Is our
knowledge of human artifacts aptly described as just a database of facts added on to a more pure
and timeless core? Or, are there more deep effects that get down to the character of thought? Do
the very categories that we have for understanding the world depend on our inventions?
In the case of some artifacts, it seems unlikely that there are deep and far-reaching effects on
the character of human taught. For example, consider a particular invention, the automobile. How
does the existence of cars effect what people know? Well, there are some obvious additions to the
knowledge of individuals. For example, many people know the details necessary to operate a car.
But there are likely some more subtle effects. For example, we may in part understand some other
domains through metaphor to our experience with cars. Just a moment ago I looked at my old
computer and thought: “Boy that computer has really got a lot of miles on it.” And there are
probably some even more subtle effects. Because there are cars, our experience is altered since we
can and do travel to a greater variety of locations in the world.
So it seems that the existence of cars may have many effects on what we know. But, while
these effects may be spread through many aspects of our knowledge, they likely do not constitute
any major change in the character of knowledge. It seems relatively certain that, in a world
without cars, the basic nature of human knowledge and thought processes would be quite similar,
at least in domains that are not closely related to cars themselves.
But researchers have posited more profound effects for some other types of artifacts. The class
of human artifacts I have been calling external representations seems especially likely to have deep
and far-reaching influence (for some reasons discussed in Chapter 7). In this regard, perhaps the
most highly touted artifact is natural language. To some researchers, natural language is much
more than a tool for communication, it is the very basis of thought. Recall, for example,
Vygotsky’s statement, cited in Chapter 1, that natural language allows an infant to “divide his
formally undifferentiated thought (Vygotsky, 1934/86, p. 219).” In fact, for Vygotsky, language
plays a special role in giving human thought its unique character. Furthermore, as I discussed
extensively in Chapter 1, there has been a wide range of speculation concerning the effects of
literacy on human knowledge (Goody and Watt, 1968; Olson, 1994).
In this work, I did not look at spoken or written language, I examined th
Download