Point and Set Paradigms in Students` Handling of Experimental

advertisement
Paper in “Science Education: Past, Present and Future” edited by Reinders
Duit et. al. (Kluwer Academic Publishers: Dortrecht, 2001) 331-336
Point and Set Paradigms in Students’ Handling of
Experimental Measurements
Saalih Allie and Andy Buffler
Department of Physics, University of Cape Town, South Africa
Fred Lubben and Bob Campbell
Department of Educational Studies, University of York, UK
Abstract
The procedural understanding of first year university students before and after instruction
has been investigated in the context of experimental work in physics. A written instrument
was used to probe the students’ ideas about data collection, processing and comparison.
The responses of the students are analysed in terms of “point” and “set” paradigms which
are proposed as a framework for evaluating the effectiveness of laboratory curricula.
Introduction
Experimentation and measurement are fundamental to knowledge production in
the natural sciences. Meaningful engagement by students in scientific activities that
are experimentally based requires an understanding of the reasons for the procedures
that are followed. However, a growing body of work indicates that students at both
school (Coelho and Séré, 1998) and university (Evangelinos et al., 1999) carry out
the procedures without such a basic understanding. In a study of undergraduate
physics students, for example, a clear disjuncture between virtuosity in applying the
formalism of data analysis and the level of understanding were observed (Séré et al.,
1993). The current work forms part of a research programme in which procedural
understanding of experimental work of entering undergraduate science students at
the University of Cape Town is being documented. The ultimate aim of the exercise
is to use the findings to inform the development of an introductory physics
laboratory curriculum. This paper reports on procedural understanding in the context
of experimental work in physics of a group of students both at entry to the university
and after a twelve-week laboratory course. The laboratory course in question is part
of the physics component of the Science Foundation Programme (SFP) which was
set up at the University of Cape Town to assist educationally disadvantaged
students. Details of the SFP and the laboratory course are contained in Allie and
Buffler (1998).
Since most of the SFP students have little or no first hand practical experience, a
prime aim of the laboratory course is to develop the notion of measurement. Aspects
of data collection and data processing are addressed by exercises such as drawing up
tables, taking several measurements of a quantity, plotting graphs, fitting straight
lines, and calculating the mean and the standard deviation from the statistical
formulae as well as graphically from a Gaussian curve. The idea of spread in data is
introduced by getting the class to measure the time of travel of a sound pulse over a
given dis tance. The readings are processed to form a distribution (a Gaussian curve
results) from which the key ideas of mean and uncertainty are introduced. The
laboratory course consists of weekly three-hour sessions over twelve weeks. About
half of this time is spent in the laboratory carrying out experiments while the
remainder is used for the exercises described.
Methodology
The research instrument comprised a set of nine written questions (probes) based
on those developed for a previous study (Allie et al., 1998). All these probes related
to the same posited experiment which was presented as follows, together with a
detailed diagram (not shown here). “An experiment is being performed in the
Physics laboratory. A wooden slope is clamped near the edge of a table. A ball is
released from a height h above the table as shown in the diagram. The ball leaves
the slope horizontally and lands on the floor a distance d from the edge of the table.
Special paper is placed on the floor on which the ball makes a small mark when it
lands. The students have been asked to investigate how the distance d on the floor
changes when the height h is varied. A metre stick is used to measure d and h”.
The situation was also demonstrated using a large-scale model. The probes focused
on decisions to be made while collecting data, processing data and comparing two
different sets of measurements of the same quantity. Each probe was of the same
form as the example shown in fig. 1 below. Thus, a situation was presented where a
procedural decision was required and a number of alternative actions (A, B, C) were
suggested. Most importantly, a detailed reason for each choice was requested. The
probes were answered under formal examination-type conditions, strictly in the
sequence presented. As each answer sheet was completed it was placed into an
envelope and never reviewed by the student. The instrument was completed by 70
students before and after they had completed the laboratory course.
Figure 1. Example of a probe: in this case the reasons for repeating measurements.
The students work in groups on the experiment. Their first task is to determine d when
h = 400 mm. One group releases the ball down the slope at a height h = 400 mm and, using
a metre stick, they measure d to be 436 mm.
The following discussion then takes place between the students.
Why? We’ve got the
result already. We do
not need to do any
more rolling.
I think that we should
roll the ball a few more
times from the same
height and measure
d each time.
I think we should
roll the ball down
the slope just one
more time from the
same height..
A
B
With whom do you most closely agree? (Circle ONE):
Explain your choice below.
C
A
B
C
Analysis
The analysis of the probes consisted of categorizing the student responses
according to the answer choice (A, B, C) together with the different types of
reasoning put forward by the students. The coding of the responses was undertaken
using an alphanumeric scheme which was developed and tested previously (Allie et
al., 1998). This enabled the underlying reasoning to be identified for each student.
Earlier work (Buffler et al., 1998) suggested that the actions and reasoning
employed by students could be classified into two groups by defining two
constructs, namely a “point” paradigm and a “set” paradigm as discussed below.
Using this framework, the pre- and post-test results were analysed by looking for
patterns across the three areas that the probes addressed, namely, data collection,
data processing and data set comparison.
The point paradigm is characterised by the notion that each measurement
results in a single, “point-like” value which could in principle be the true value. As
a consequence each measurement is independent of the others and the individual
measurements are not combined in any way. In its most extreme form, this way of
thinking manifests itself in the belief that only one single measurement is required to
establish the true value, as indicated in the work of Séré et al. (1993). Responses
were coded as being associated with the point paradigm when, for example: (a) it
was stated that measurements are repeated in order to find a recurring value or to
perfect the measuring skill in order to finally take one ‘perfect’ measurement; (b) a
specific measurement was selected (e.g. the highest, the recurring, the first or the
last) to represent a series of numerical readings; (c) specific points (such as the
origin, the extreme points or any three aligned points) were chosen through which to
draw a straight line to represent a collection of plotted points; or (d) two sets of data
were contrasted either by comparing individual measurements in the sets, or by
treating the mean values of the data sets as points to be compared.
The set paradigm is characterised by the notion that each measurement is only
an approximation to the true value and that the deviation from the true value is
random. As a consequence, a number of measurements are required to form a
distribution that clusters around some particular value. The best information
regarding the true value is obtained by combining the measurements using
theoretical constructs in order to describe the data collectively. The operational tools
that are available for this purpose include the formal mathematical procedures that
can be used to characterise the set as a whole, such as calculating the mean and the
standard deviation. In turn, these quantities become tools for making comparisons
with other data-sets or theory. Responses were coded as being associated with the
set paradigm when, for example: (a) it was stated that repeating measurements was
aimed at taking a mean; (b) the mean and the spread were calculated to represent the
data; (c) a ‘line of best fit’ that took account of all points was drawn for plotted data;
or (d) different sets of measurements were contrasted by comparing the degree of
overlap of the intervals defined by the mean and some measure characterising the
spread of the data.
Results
Table 1 summarises the results from the pre- and post-tests with regard to
students’ understanding about repeating measurements during data collection (3
probes). It shows that before instruction the large majority of students (76%)
subscribed to the point paradigm while after instruction there appeared to be a large
shift (16% to 71%) towards the set paradigm. However, it is not clear that these
students have embraced the set paradigm as a whole. For example, many students
indicated that the purpose of repeating measurements is to allow for a mean to be
generated (rather than a mean being a way of dealing with the inherent scatter in the
data). This suggests there is a strong possibility that elements of the set paradigm are
being used by rote or on an ad hoc basis. The degree to which this is the case
requires the combined analysis of the other probes.
Table 1.
Students’ use of paradigms for
data collection.
Paradigm
before
instruction
Point
paradigm
Set
paradigm
Not
codeable
Total
Paradigm after instruction
Point
Set
Not
paradigm
paradigm
codeable
9
40
4
(13%)
(57%)
(6%)
4
6
1
(6%)
(9%)
(1%)
2
4
0
(3%)
(5%)
(0%)
15
50
5
(21%)
(71%)
(7%)
Total
53
(76%)
11
(16%)
6
(8%)
70
(100%)
Table 2 summarises the pre- and post test findings for 3 probes dealing with the
comparison between two data-sets. The first of the three probes required students to
compare two sets of data with the same mean but different scatter, while the second
probe provided two sets of data with different means but the same (overlapping)
spread. The third probe presented two data-sets with different means and different
but overlapping spreads. In Table 2 students are grouped according to whether or not
their responses across the three probes were consistent with the set paradigm. As
expected from the background of the students, none were classified as using the set
paradigm consistently prior to instruction. After instruction only 26% responded
consistently in terms of the set paradigm while more than two thirds (70%) resorted
to both paradigms, possibly indicating either rote or ad hoc application of the
elements associated with the set paradigm.
Table 2.
Students’ use of paradigms
when comparing data-sets.
Paradigm
before
instruction
Mixed
paradigms
Consistent set
paradigm
Not
codeable
Total
Paradigm after instruction
Mixed
paradigms
48
68%)
0
(0%)
1
(1%)
49
(70%)
Consistent
set paradigm
18
(26%)
0
(0%)
0
(0%)
18
(26%)
Not
codeable
3
(4%)
0
(0%)
0
(0%)
3
(4%)
Total
69
(98%)
0
(0%)
1
(1%)
70
(100%)
Tables 3 and 4 compare various aspects of the post-probes. In Table 3 the post-
probe results of Table 2 (data-set comparison) are contrasted with the combined
results of the post-probes dealing with data collection (one probe) and data
processing (two probes). One of the data processing probes required a mean to be
calculated from a set of numerical data, while the other required a straight line to be
drawn to a set of graphical data. It is interesting to note that there appears to be a
strong link between the paradigms used for these two probes. For example, students
who joined individual data points (i.e. did not fit a straight line to the data as a
whole) often evidenced point paradigm use in the other probe by choosing the
recurring value to represent the data rather than calculating a mean. (Space
limitations preclude showing the evidence in detail). Table 3 shows that only about a
fifth (21%) of the students based their responses on the set paradigm for all of data
collection, data processing and data-set comparison and that the largest group (37%)
were inconsistent in their use of the paradigms.
Table 3.
Students’ use of paradigms
for data collection /
processing and data-set
comparison after instruction.
Paradigms
used in
data-set
comparison
Mixed
paradigms
Consistent
set paradigm
Not
codeable
Total
Paradigms used in data collection / processing
Consistent
point
paradigm
9
(13%)
0
(0%)
0
(0%)
9
(13%)
Mixed
paradigms
25
(37%)
3
(4%)
1
(1%)
29
(42%)
Consistent
set
paradigm
13
(18%)
15
(21%)
2
(3%)
30
(43%)
Not
codeable
Total
2
(3%)
0
(0%)
0
(0%)
2
(3%)
49
(70%)
18
(26%)
3
(4%)
70
(100%)
Table 4 shows results for students who have been classified on the basis of all
the probes discussed thus far, together with the results of the final probe (mean/sd
probe), in which students were asked to compare two data-sets described in the
formal manner of a mean and a standard deviation of the mean.
Table 4.
Students’ action and reasoning for data
collection, data processing and data-set
comparison after instruction.
Classification
of student
reasoning from
all previous
probes
Consistent point
paradigm reasoning
Inconsistent paradigm
reasoning
Consistent set
paradigm reasoning
Not
codeable
Total
Point or set action for
mean/sd probe
Point
Set
Not
paradigm
paradigm
codeable
action
action
3
6
0
(4%)
(9%)
(0%)
11
17
1
(17%)
(24%)
(1%)
14
16
0
(20%)
(23%)
(0%)
1
1
0
(1%)
(1%)
(0%)
29
40
1
(42%)
(57%)
(1%)
Total
9
(13%)
29
(42%)
30
(43%)
2
(3%)
70
(100%)
From Table 4 it is clear that although more than half the students (57%) carried
out an action associated with the set paradigm, fewer than half of this group (23% of
the sample) provided a reason that was also consistent with this paradigm. In other
words 33% of the students (57%? 23%? 1%) appear to have used the correct set
paradigm action either by rote or in an ad hoc way. In summary, only a quarter of
all the students (100%? 42%? 33%) can be regarded as having completely embraced
the set paradigm.
Conclusions
In terms of the constructs of the point and set paradigms, the purpose of
laboratory instruction can be regarded as attempting to shift students’ actions and
reasoning away from the point paradigm to those commensurate with the set
paradigm. Both the present study and the previous work (Buffler et al., 1998)
strongly suggest that students come from school firmly located within the point
paradigm, and that any set paradigm actions (such as calculating a mean) are most
often a rote response. Even after a six month laboratory course, only about one
quarter of the students seem to have reached the required instructional goals. The
present probes and the analysis framework appear to offer useful research tools that
can be used to evaluate the effectiveness of any laboratory curriculum that aims to
address procedural understanding in the context of experimentation.
References
Allie, S. and Buffler, A. (1998). A course in tools and procedures for Physics 1. American
Journal of Physics, 66 (7), 612-623.
Allie, S., Buffler, A., Kaunda, L., Campbell, B. and Lubben, F. (1998). First year physics
students’ perceptions of the quality of experimental measurements. International Journal
of Science Education, 20 (4), 447-459.
Buffler, A., Allie, S., Campbell, B. and Lubben, F. (1998) The role of laboratory experience
at school on the procedural understanding of pre-first year science students at UCT. In N.
A. Ogude and C. Bohlmann, Eds., Proceedings of the 6th Annual Meeting of the Southern
African Association for Research in Mathematics and Science Education, (pp. 495-502).
Pretoria, South Africa: University of South Africa Press.
Coelho, S. M. and Séré, M.-G. (1998). Pupils’ reasoning and practice during hands-on
activities in the measurement phase. Research in Science and Technological Education,
16 (1), 70-96.
Evangelinos, D., Valassiades, O. and Psillos, D. (1999). Undergraduate students’ views about
the approximate nature of measurement results. In M. Komorek et al., Eds., Proceedings
of the Second International Conference of the European Science Education Research
Association Volume 1, (pp. 208-210). Kiel, Germany: IPN.
Lubben, F. and Millar, R. (1996). Children’s ideas about the reliability of experimental data.
International Journal of Science Education, 18 (8), 955-968.
Séré, M.-G., Journeaux, R. and Larcher, C. (1993). Learning the statistical analysis of
measurement error. International Journal of Science Education, 15 (4), 427-438.
Acknowledgements
The research reported in this paper was supported by the British Council (DFID), the
University of Cape Town and the University of York. We also thank Indresan
Govender and Fiona Gibbons for their assistance.
Download