spiral assessments

advertisement
SPIRAL ASSESSMENTS:
A STUDY IN AN ACCELERATED MATHEMATICS CLASSROOM
Except where reference is made to the work of others, the work described in this project
is my own or was done in collaboration with my Advisor. This project does not include
proprietary or classified information.
____________________________________________________________________
Lisa Morgan Skinner
Certificate of Approval:
______________________________
Donald R. Livingston, Ed.D.
Associate Professor and Project Advisor
Education Department
______________________________
Sharon M. Livingston, Ph.D.
Assistant Professor and Project Advisor
Education Department
Spiral Assessments ii
SPIRAL ASSESSMENTS:
A STUDY IN AN ACCELERATED MATHEMATICS CLASSROOM
A project submitted
by
Lisa Morgan Skinner
to
LaGrange College
in partial fulfillment of
the requirement for the
degree of
SPECIALIST IN EDUCATION
in
Curriculum and Instruction
LaGrange, Georgia
July 18, 2011
Spiral Assessments iii
Abstract
The purpose of this study was to research the effectiveness of spiral assessments
for mathematics students. The study compared one year of Accelerated Math 2 students
who received no spiral assessments throughout their course to the following year of
Accelerated Math 2 students who received numerous spiral assessments. In addition, the
spiral group was given appropriate feedback about their assessments as well as positive
changes to their learning environment based on the assessments. Students were surveyed
about whether or not spiral assessments led to improved attitudes about mathematics and
if the practice raised confidence in their mathematics abilities. Finally, this study used a
focus group of ten math teachers and interviews of two administrators to determine if a
spiral assessment practice was feasible for broader implementation and if it would be
supported by the faculty and administration at the school. Overall, this research study
produced positive results with regards to content retention and student confidence in
mathematics.
Spiral Assessments iv
Table of Contents
Abstract ……………………………………………………………………………….. iii
Table of Contents ……………………………………………………………………... iv
List of Tables …………………………………………………………………………... v
Chapter 1: Introduction ………………………………………………………………… 1
Statement of the Problem ……………………………………………………… 1
Significance of the Problem …………………………………………………… 1
Theoretical and Conceptual Frameworks …..…………………………………. 2
Focus Questions ………………………………………………………………... 4
Overview of Methodology …………………………………………………….. 5
Human as Researcher ………………………………………………………….. 6
Chapter 2: Review of the Literature …………………………………………………… 8
Spiral Curriculum ……………………………………………………………… 8
Formative Assessments ………………………………………………………... 9
Student and Teacher Attitudes ………………………………………………... 12
Organizational Change ……………………………………………………….. 14
Chapter 3: Methodology ……………………………………………………………… 18
Research Design ……………………………………………………………… 18
Setting ………………………………………………………………………… 19
Subjects and Participants ……………………………………………………... 19
Procedures and Data Collection Methods ……………………………………. 20
Validity and Reliability Measures ……………………………………………. 23
Analysis of Data ……………………………………………………………… 26
Chapter 4: Results ……………………………………………………………………. 30
Chapter 5: Analysis and Discussion of Results ………………………………………. 47
Analysis of Results …………………………………………………………… 47
Discussion ……………………………………………………………………. 55
Implications …………………………………………………………………... 57
Impact on School Improvement ……………………………………………… 59
Recommendations for Future Research ……………………………………… 60
References ……………………………………………………………………………. 62
Appendices …………………………………………………………………………… 66
Spiral Assessments v
List of Tables
Table 3.1
Data Shell ………………………………………… 21
Table 4.1
Independent t-test ………………………………… 31
Table 4.2
Independent t-test ………………………………… 32
Table 4.3
Independent t-test ………………………………… 33
Table 4.4
Independent t-test ………………………………… 34
Table 4.5
Independent t-test ………………………………… 35
Table 4.6
Independent t-test ………………………………… 35
Table 4.7
Dependent t-test ………………………………….. 36
Table 4.8
Chi-Square Statistic ………………………………. 38
Spiral Assessments 1
CHAPTER 1: INTRODUCTION
Statement of the Problem
In many classrooms throughout the nation, students are memorizing facts and
regurgitating those facts on their assessments. Little is being done by the student, or
perhaps the teacher, to ensure that content knowledge is being retained from year to year.
Johnson (2003) eloquently explained, “we all know that the way most secondary schools
work is that students spend about 179 days preparing for a three-hour Brain Dump in
some gymnasium in June. . . we are relatively sure that one year later they will have
forgotten just about everything from the year before” (p. 8).
In addition, with the new “Failure is Not an Option” [FNO] and retest policies of
some school systems, many students are waiting until the last minute to study for
assessments and many students often resort to just memorizing important content. As a
result, knowledge is not being constructed in a manner that guarantees “learning” has
actually occurred. Johnson (2003) explained that when students score well on a final
exam, teachers naturally conclude they ‘know’ the material and have ‘learned’ that
subject; when, in reality, this assumption may be further from the truth.
Significance of the Problem
According to Zemelman, Daniels, and Hyde (2005), “many [students] come to
believe that they are incapable of doing math. As they progress through the grades, fewer
and fewer students understand and enjoy math, leaving only a handful to continue with
capability and confidence” (p. 106). In addition, while in college, many students “take
only the minimum [math] courses required, despite the fact that many careers depend
upon mathematical knowledge” (Zemelman et al., 2005, p. 106). According to the
Spiral Assessments 2
National Council of Teachers of Mathematics [NCTM] (2000), “the need to understand
and be able to use mathematics in everyday life and in the workplace has never been
greater and will continue to increase” (p. 4). Usually, when students learn for the
moment and do not achieve a true understanding, negative results occur. With
mathematics being a subject that builds on mastery of prior topics, many students find
themselves “lost” and disliking the subject. NCTM (2000) explained that this lack of
mathematical understanding keeps the doors closed for many enhanced opportunities.
The purpose of this research was to find a way to encourage students to become
better learners. With such an emphasis being placed on assessments in schools, this
research aimed at making assessments become a more productive part of education. In
addition, with prior content being an integral aspect of the mathematics classrooms, it
seemed obvious to include this in the assessment process. Therefore, this research
attempted to answer the question, will spiral assessments have a positive effect on student
learning in the mathematics classroom?
Theoretical and Conceptual Frameworks
The LaGrange College Department of Education (2008) along with its teacher
candidates, strongly support a constructivist approach to learning. The philosophy of
constructivism is founded on the premise that individuals learn by reflecting on their own
experiences and by adjusting “mental models” to accommodate new experiences. In
particular, this research project embedded the theory of social constructivism throughout.
Bruner (1996) explained, “you cannot understand mental activity unless you take into
account the cultural setting and its resources, the very things that give mind its shape and
scope” (p. x). Bruner also explained that students have a greater use for knowledge
Spiral Assessments 3
which has been acquired through discovering and making connections to prior
experiences (p. xii). Fosnot (2005) further explained, “rather than behaviors or skills as
the goal of instruction, cognitive development and deep understanding are the foci; rather
than stages being the result of maturation, they are understood as constructions of active
learner reorganization. Rather than viewing learning as a linear process, it is understood
to be complex and fundamentally nonlinear in nature” (pp. 10-11).
The theoretical framework for this study was guided by LaGrange College
Education Department’s (2008) Conceptual Framework. Based on Tenet 2 from the
framework, ‘exemplary professional teaching practice’ students should be active
participants in the learning process, while teachers serve as mere facilitators. In addition,
teachers are expected to pull from a variety of resources in order to be effective in the
diverse classrooms of today. Active learning environments are required to help students
become active participants (LaGrange College Education Department, 2008).
This research was specifically connected to Competency Cluster 2.3 Assessment
Skills of the LaGrange College Education Department’s (2008) Conceptual Framework.
Teachers should understand and use formal and informal assessment strategies to
evaluate and ensure continuous intellectual, social, and physical development of students.
Secondly, teachers should involve students in self-assessment that helps them become
aware of their strengths and needs and should encourage students to set personal goals for
learning. Finally, teachers should monitor and adjust strategies in response to student
feedback. With the application of spiral assessments, teachers should be able to satisfy
all parts of cluster 2.3.
Spiral Assessments 4
The National Board for Professional Teaching Standards [NBPTS] Core
Proposition Three closely related to this research. Proposition Three stated that teachers
are responsible for managing and monitoring student learning. In addition, Element 1D
(Student Learning for Teacher Candidates) of the National Council of Accreditation of
Teacher Education [NCATE] Principles was directly related to this research. Both of
these standards adhere to the belief that teachers are responsible for assessing the learning
of students. This research used spiral assessments as a means by which teachers could
reflect on their students’ knowledge and as a result lead teachers to make positive
changes regarding their curriculum, learning environment, and future assessments.
Focus Questions
To determine if spiral assessments had a positive effect in the math classroom,
further questions were developed which included more specifics. The first focus question
addressed the math content that needed to be learned by the students. The second focus
question was an affective assessment of how students felt about incorporating spiral
assessments into their math class. Lastly, the third focus question addressed the area of
school improvement and whether spiral assessments could feasibly be adopted with the
school. As a result, the following three focus questions were developed to guide this
entire research process.
1. Will the use of spiral assessments, coupled with proper feedback, improve
students’ achievement on the Math 2 End of Course Test?
2. How will spiral assessments affect students’ feelings towards math, and will such
assessments improve student confidence?
Spiral Assessments 5
3. Will a system which employs the use of spiral mathematics assessments be
received positively by the math department and will the use of such a system be
supported by the administration?
Overview of Methodology
This action research occurred in a Georgia rural high school. The study took place
throughout the 2010 school year, initially using 83 tenth-graders from five Accelerated
Math 2 classes. The 2010 data were compared to data collected in the 2009 school year
with an initial 65 tenth-graders from four classes of Accelerated Math 2. Both
quantitative and qualitative data were used throughout the research.
In order to determine if spiral assessments would help students’ achievement on
the Math 2 EOCT (focus question one), several independent t-tests were run. First, an
independent t-test on eighth grade CRCT z-scores was conducted to see if there was
significant difference between the ability levels of the 2009 students and the 2010
students. When a significant difference was found in the two groups, the z-scores were
then screened and matched to equalize the ability levels of the two groups. The initial 83
students from 2010 were reduced to 49 and the 2009 initial 65 students were also reduced
to 49. Next, independent t-tests were conducted on two midterm exam scores for both
school years. The 2009 students were used as the control group, receiving no spiral
assessments throughout their course while the 2010 students received numerous spiral
assessments throughout the year, as well as adjustments to their curriculum. Finally, an
independent t-test was conducted on the Math 2 EOCT scores for both groups to ascertain
significant difference.
Spiral Assessments 6
To address the second focus question, “How will spiral assessments affect
students’ feelings towards math, and will such assessments improve student
confidence?”, surveys and reflections were administered throughout the year. A sevenpoint Likert Scale was used on the surveys and each question was analyzed using the Chi
Square method. Student reflections focused on changes in attitudes and confidence levels
toward mathematics and provided qualitative data for the research project.
The third focus question, “Will a system which employs the use of spiral
mathematics assessments be received positively by the math department and will the use
of such a system be supported by the administration?”, was addressed using a focus
group and two interviews. At the end of the study, math teachers participated in a focus
group discussion about spiral assessments and qualitative data were collected.
Additionally, two administrators were interviewed to ascertain attitudes about spiral
assessments and the feasibility of implementing a spiral assessment program within the
school.
Human as Researcher
Over the course of 17 years of teaching mathematics, I have learned that my
students tend to “memorize for the test” and then quickly lose the knowledge that was
assessed. As a student, I too, exhibited this behavior. Although math was my favorite
subject, it wasn’t until I taught the subject that I truly grasped what I was doing and why I
was doing it.
Last year I was faced with teaching the new Georgia Accelerated Math 2 course
where my students were required to take the regular Math 2 end-of-course test. Over half
of the content assessed on the Math 2 EOCT was taught the previous year in Accelerated
Spiral Assessments 7
Math 1. Although the students in Accelerated Math 2 are typically stronger in their math
ability, I quickly realized there was a ton of content knowledge they lost from the
previous year. As a result, I spent a month in April reviewing content they were expected
to already know. With this research, I hoped to find that a spiral assessment approach,
when implemented throughout the year, would help my students retain content longer and
make them more successful on their Math 2 EOCT.
Spiral Assessments 8
CHAPTER 2: REVIEW OF THE LITERATURE
A review of current and past literature provided a justification for this action
research study. The literature review presented evidence for each of the three focus
questions in the study. The literature review included information regarding spiral
curriculums, formative assessments, student and teacher attitudes, and organizational
change.
Spiral Curriculum
Most people have heard the following phrases on more than one occasion:
“practice makes perfect” and “use it or lose it.” Whether it’s practicing the piano or
perfecting a golf swing, people recognize the importance of practice. Perhaps the most
applicable place for practice is inside a mathematics classroom. Even math teachers who
go a few years without teaching a certain concept have to spend time reviewing the
material before they can explain it successfully to their students.
Bruner (1960) introduced the idea of a spiral curriculum and the importance of
revisiting concepts throughout one’s education. He explained that “a curriculum as it
develops should revisit these basic ideas repeatedly, building upon them until the student
has grasped the full formal apparatus that goes with them” (p. 13). Bruner (1960)
suggested that students initially learn a general idea which can then be used as the “basis
for recognizing subsequent problems as special cases of the idea originally mastered” (p.
17). Doll (1993) explained that iteration is the process of repeating itself over and over
again and stated, “nothing is more central to this new beginning than the concept and
practice of iteration” (p. 97). Doll also stated that Bruner’s “spiral curriculum is worth
looking at again and reframing in light of recursion theory” (p. 102). Doll further
Spiral Assessments 9
explained that “it is worth constructing a curriculum where students revisit with more
insight and depth what they have done” (p. 102).
Harden and Stamper (1999) supported the concept of a spiral curriculum, but
added that “a spiral curriculum is not simply the repetition of a topic taught. It requires
also the deepening of it, with each successive encounter building on the previous one” (p.
141). Snider (2004) also cautioned that a spiral curriculum often limits the depth of
knowledge that students attain. Snider (2004) explained, “in a spiral curriculum, many
topics are covered but only briefly…The result of teaching for exposure is that many
students fail to master important math concepts” (p. 31).
Jensen (1990) also cautioned that spiral curriculums are hindering our efforts to
improve the educational system in our country because they “rob both students and
teachers of the excitement and motivation that is inherent in anticipating learning
something new” (p. 4). According to Jensen (1990), this lack of curiosity prevents
students from learning any topic in depth and keeps students from reaching a level of
meaningful understanding. Instead, students scratch the surface of numerous topics and
revisit those topics without any further depth. Jensen (1990) also explained that countries
with high mathematics achievement have curricula without a high degree of repetition.
Formative Assessments
Now, more than ever before, schools, administrators, and educators are being held
to higher accountability standards for student achievement. Whether one supports or
opposes today’s high-stakes tests, there is no denying the importance of them. With the
emergence of these tests, educators are being forced to closely examine their curriculum,
their teaching practices, and their own classroom assessments. Popham (2001) explained
Spiral Assessments 10
that “classroom assessment has a higher calling – to improve the caliber of classroom
instruction” (p. 117).
At the forefront of discussions about classroom assessments was the importance
of using tests to gather information about students’ learning and to make decisions about
future instruction. Dwyer (2008) stated that “the central purpose of testing will be to
inform and improve teaching and learning” (p. 5). Garcia, Spalding, and Powell (2001)
defined formative assessments as those assessments used to gather information “while
work is still in progress in order to influence the final outcome” (p. 303). Today, there is
a greater push for the use of formative assessments within the classroom, as opposed to
the traditional summative assessments that occur at the end of a chapter or unit.
Andrade and Cizek (2010) believed that formative assessments “offer great
promise as the next best hope for stimulating gains in student achievement” (p. 3). While
formative assessments can take many different forms (observations, oral questioning,
class discussions, projects, portfolios, homework, group work with peer feedback, student
self-assessment), the primary goal of such assessments is to provide information for the
purpose of making adjustments within the classroom. These formative assessments have
the potential to “provide missing linkages between classroom practice and large-scale
assessments” (Andrade & Cizek, 2010, p. 4).
Jones, Carr, and Ataya (2007) believed that using a variety of assessments to
provide continuous feedback will nourish teaching and learning. They cautioned against
using one test score to provide a picture of what a student does or does not know and that
a test score is simply a snapshot in time and subject to error. Jones et al. (2007)
explained, “a teacher who evaluates student learning and instructional practices solely on
Spiral Assessments 11
the basis of test scores is missing valuable information . . . The more information a
teacher collects, the more valid the inferences based on that information” (p. 74). The
National Council of Teachers of Mathematics [NCTM] (2000) stated, “assembling
evidence from a variety of sources is more likely to yield an accurate picture” (p. 24).
Teachers who implement a variety of testing techniques have a better picture of their
students and their classroom and can make more effective decisions regarding both.
In a mathematics classroom, successful completion of problems depends greatly
on successful completion of prior problems. Jones et al. (2007) explained the
importance of formative assessments “when correct procedure is crucial to later success”
(p. 77). Waiting for a summative assessment to discover that a student missed important
information early in the learning process is extremely frustrating to the teacher as well as
the student. Formative assessments must be ongoing and must provide “feedback and
direction to students as they proceed toward a goal” (Jones et al., 2007, p. 77).
The National Council of Teachers of Mathematics [NCTM] (2000) strongly
believed that assessment should “support the learning of important mathematics and
furnish useful information to both teachers and students” (p. 22). An assessment should
no longer just come at the end of instruction, but should be something administered
throughout instruction. NCTM (2000) stated, “assessment should not merely be done to
students; rather, it should also be done for students” (p. 22). In addition, NCTM (2000)
explained that “assessment and instruction must be integrated so that assessment becomes
a routine part of the ongoing classroom activity rather than an interruption” (p. 23).
Zemelman et al. (2005) explained that teachers in ‘Best Practice’ classrooms use
assessments for more than just providing a grade on a report card. To gain a deeper
Spiral Assessments 12
understanding of student learning, progressive teachers “monitor students’ growth in
richer and more sophisticated ways” (Zemelman et al., 2005, p. 252). According to
Zemelman et al. (2005), teachers should use gathered information to guide instruction, to
make critical daily decisions about helping students grow, and most importantly, to help
students set goals, monitor their own work, and evaluate their efforts.
Student and Teacher Attitudes
To quote Winston Churchill (n.d.), “Attitude is a little thing that makes a big
difference.” When taken to heart, what a difference this statement could make in one’s
life. For students and teachers, nowhere is this statement more applicable than inside the
classroom.
Teachers’ attitudes affect the way they interact with their students, the way
they teach, and the way they assess. Likewise, students’ attitudes affect the way they
interact with their teachers, the way they behave, the way they learn, and how well they
perform. When looking at improving student achievement, student and teacher attitude is
an area that must be examined.
There was plenty support in the literature connecting positive attitudes in
mathematics to positive student achievement (Farooq & Shah, 2008; Marzano, 2003;
McMillan, Cohen, Abrams, Cauley, Pannozzo, & Hearn, 2010; Wigfield & Meece,
1988). According to Farooq and Shah (2008), “students’ success in mathematics depends
upon attitude towards mathematics. It also influences the participation rate of learners”
(p. 75). Marzano (2003) also explained the importance of motivation to classroom
success. He stated, “if students are motivated to learn the content in a given subject, their
achievement in that subject will most likely be good” (p. 144).
Spiral Assessments 13
Regarding attitudes towards a spiral curriculum, there were mixed reviews in the
literature. According to Caldwell (2008), when a tenth grade student was asked about
repetition in her math class and why teachers revisited the same topics from year to year,
the girl responded, “so that it’s, like, branded in our brains, so that we know it forever”
(p. 3). Caldwell further explained that discussions with other students revealed feelings
of boredom and frustration towards repetitious work and that students did not appreciate
being reminded that they had covered the work before.
Student and teacher attitudes surrounding formative assessments also impacted
learning. There were several references which supported the use of formative
assessments in the classroom and most revealed a positive correlation between their use
and resulting positive attitudes, both in teachers and in students. McMillan et al. (2010)
explained, “the statistically significant positive relationships between overall formative
practices and class averages of student motivation, suggest an association between at
least some formative practices and student motivation” (p. 10). They also reported that
when teachers employed several types of formative assessments, that students reported
higher levels of motivation (McMillan et al., 2010).
Although there appeared to be a positive connection between the use of formative
assessments and an increase in student motivation, McMillan et al. (2010) discovered that
many secondary teachers “do not use formative assessment practices nearly as much as
what is suggested in the literature, and report that several critical components of
formative assessment, especially instructional correctives, are not widely used” (p. 11).
A possible reason cited for the discrepancy in teacher use of formative assessments was
that teachers may “perceive formative assessment practices as impractical or time-
Spiral Assessments 14
consuming” (McMillan et al., 2010, p. 11). Additionally, in a study conducted by
Volante, Beckett, Reid, and Drake (2010), while teachers acknowledged the importance
of feedback, many teachers found that their feedback was useless and a waste of time if
the students did not even bother to look at it. Finally, although teachers noted difficulties
in utilizing self- and peer assessment within their classrooms, they did concur that
“involving students in the assessment process is vital to student learning” (Volante et al.,
2010, p. 13).
Organizational Change
It is impossible to live in today’s society and not recognize the rate at which
things are changing. Cameron and Green (2009) noted, “the rate of change and discovery
outpaces our individual ability to keep up with it” (p. 1). Lieberman (2005) explained
that the changes educators have to deal with occur one on top of another at “an
increasingly frenetic speed” (p. x). If schools expect to stay current in the lives of their
students, they must be willing to change with the world in which they operate. The
question then becomes, how do schools most effectively bring about change?
In terms of implementing change in any organization, Cameron and Green (2009)
recognized the importance of the people on the receiving end of the change. They
explained, “Whatever the level or degree of organizational change, the people on the
receiving end are individual human beings. It is they who will ultimately cause the
change to be a success or a failure. Without looking at the implications of change on
individuals we can never really hope to manage large-scale change effectively” (p. 3). In
a school system, individual school, or even a classroom, change will not be successful
unless the teachers and students are considered an important piece of the puzzle.
Spiral Assessments 15
One important organizational change strategy supported in the literature was
increased teacher involvement in leadership roles. Gabriel (2005) explained that with No
Child Left Behind and the increased importance of high-stakes tests, many principals are
turning to teachers to provide more effective organizational behavior. Gabriel (2005)
argued, “the only leadership that will make a difference is that of teachers. They alone
are positioned where all the fulcrums are for change. They alone know what the day-today problems are and what it takes to solve them” (p. 1). Gabriel (2005) also stated,
“inviting teachers to participate in the decision-making process by elevating them to
leadership roles should be viewed as a means to accomplish significant change in the
field of education” (p. 156). He also recommended that schools encourage leadership in
teachers and equip them with the skills to become productive leaders in today’s changing
society.
Marzano, Pickering, and Pollock (2001) stressed the importance of teacher desire
and commitment to bring about change within a school. They explained that “a small
group of educators within a school who are enthusiastic about a particular innovation can
‘infect’ an entire staff with that enthusiasm” (pp. 157-158). Cameron and Green (2009)
also pointed out, “individual change is at the heart of everything that is achieved in
organizations. Once individuals have the motivation to do something different, the whole
world can begin to change” (p. 9). Based on the literature, teacher leaders are excellent
change agents within a school; however, teacher leaders must have the desire and will to
see the change occur for there to be any real hope for the change to be a success.
In addition to utilizing teachers as leaders within a school, professional
development was mentioned as a means of bringing about change. Unfortunately, the
Spiral Assessments 16
professional development literature raised some concerns about successful
implementation. Volante et al. (2010) pointed out that “teachers often begrudge topdown, mandated professional development and do not hold value in its execution” (p.
16). Marzano (2003) also claimed that the regularly scheduled staff development in most
schools “is not necessarily meaningful or useful in terms of impacting student
achievement” (p. 65). Fullan (2005) recognized the importance of professional
development in any change strategy but cautioned that the impact of such development
has been minimal through the years.
Since professional development was recognized as an important part of
organizational change, it made sense to find ways to make it more effective. Marzano
(2003) noted that professional development “must be constructed in ways that deepen the
discussion, open up the debates, and enrich the array of possibilities for actions” (p. 66).
According to Marzano, an extensive study on staff development conducted by Michael
Garet and his colleagues revealed three important features of staff development: (1)
focus on content knowledge, (2) opportunities for active learning, and (3) overall
coherence of staff development activities. Marzano summarized Garet’s findings and
explained that staff development should be specific and relevant to a teacher’s subject
area, teachers should be able to apply and test strategies within their classrooms, and staff
development should be coherent and build on one another. Additionally, Volante et al.
(2010) explained that staff development that involved teacher input or staff-development
that was self-directed, tended to lead to more sustained changes in classroom practice.
Based on the literature review conducted for this study, there appeared to be a
natural connection between a spiral curriculum and the use of formative assessments in
Spiral Assessments 17
the classroom. While no current literature exists on ‘spiral assessments’, the study tied
the two concepts together in hopes of improving student achievement. The assessments
in the study revisited concepts previously covered in the curriculum and were considered
‘spiral’ in nature. Spiral assessments as well as daily warm-up exercises in the study
were ‘formative’ in nature and were used to inform students and teachers of progress
being made in the classroom. The literature review also revealed the importance of
teacher leaders and proper staff development within schools if sustained change was
expected to occur.
Spiral Assessments 18
CHAPTER 3: METHODOLOGY
Research Design
This study was conducted using both action and evaluation research and consisted
of both quantitative and qualitative data. Tomal (2003) explained that action research is
“one of the most practical and efficient methods of conducting research by educators” (p.
vii). He further explained that research is not finding a known solution to a problem,
rather it entails “a careful undertaking to discover new facts and relationships concerning
a solution that is unknown” (Tomal, 2003, p. 1). Educators are constantly looking for
solutions to existing problems in their classroom. Action research provides educators a
“systematic process of solving educational problems and making improvements” within
their classrooms (Tomal, 2003, p. 5). In addition, since the goal of action research is to
solve a problem and make improvements, researchers rely less on scientific inquiry and
inductive reasoning, and more on the “practicality and feasibility” of solving a given
issue (Tomal, 2003, p. 9). McNiff and Whitehead (2006) explained that, “action research
can be a powerful and liberating form of professional enquiry because it means that
practitioners themselves investigate their own practice as they find ways of living more
fully in the direction of their educational values. They are not told what to do. They
decide for themselves what to do, in negotiation with others” (p. 8). McNiff and
Whitehead (2006) also supported using action research because anyone and everyone can
do it: “all you need is curiosity, creativity, and a willingness to engage” (p. 13).
Supplementing the action research methodology, this study was conducted using
evaluation research. Charles and Mertler (2002) explained that evaluation research “is
done to determine the relative merits of various products and approaches used in
Spiral Assessments 19
education” (p. 311). The purpose of evaluation research is to assess the product or
program developed in action research (Charles & Mertler, 2002). While action research
was used to seek a solution to a problem in the classroom, evaluation research was used
to determine the effectiveness of the proposed solution. Was student achievement
increased and was the program accepted by the teachers and students?
Setting
This study took place in a rural high school in west Georgia. The school’s student
population was approximately 1300, of which 39% were minority, 54% White, and 7%
other. Of the student population, 47% were economically disadvantaged and 6% were
classified as students with disabilities. This setting was chosen for the study because, as
a mathematics teacher working at the high school, I had easy access to the subjects’ test
records. In order to conduct research in the classroom, permission was obtained from the
high school principal, the county school system, and the institutional review board at
LaGrange College.
Subjects and Participants
There were 148 subjects included in this study. All of the subjects were tenth
grade students enrolled in an Accelerated Math 2 course during the 2009-2010 and 20102011 school years. All of the students were 15-16 years old. Of the 65 students in
Accelerated Math 2 during 2009-2010, 35 were males and 30 were females. Of the 83
students in Accelerated Math 2 during 2010-2011, 44 were male and 39 were female.
In addition to the 148 subjects participating in the study, there were ten math
teachers who participated in a focus group and two administrators who were interviewed.
Five of the ten math teachers participating in the study had less than five years teaching
Spiral Assessments 20
experience, two teachers had more than ten years experience, and three teachers had more
than twenty-five years experience. Both administrators interviewed in the study had
more than twenty-five years experience in education. One administrator had been the
principal at the high school for fifteen years and was an elementary principal for five
years prior to that. The second administrator had been the registrar and math department
supervisor at the high school for the last five years and was a middle school assistant
principal for 5 years prior to that.
Procedures and Data Collection Methods
In order to answer the study’s focus questions, procedures were followed, data
were collected, and statistical tests were used for analysis. Table 3.1 provides an
explanation of the procedures and statistical tests used throughout the study.
Spiral Assessments 21
Table 3.1. Data Shell
Focus Question
Literature
Sources
Type: Method,
Data, Validity
How data are
analyzed
Rationale
Will the use of
spiral
assessments,
coupled with
proper feedback,
improve students’
achievement on
the Math 2
EOCT?
Bruner
(1960)
Doll (1993)
Dwyer
(2008)
NCTM
(2000)
Method:
Midterm
assessments;
End-of-course
assessment
Quantitative:
Independent
t-test
Quantitative:
Determine if
there are
significant
differences
Quantitative:
Dependent
t-test
Chi Square
Cronbach’s
Alpha
Quantitative:
Determine if
there are
significant
differences
How will spiral
assessments affect
students’ feelings
towards math, and
will such
assessments
improve student
confidence?
Caldwell
(2008)
Jensen
(1990)
McMillan
et al.
(2010)
Reeves
(2007)
Data:
Interval
Validity:
Content
Method:
Math Anxiety
Questionnaire,
Student
Reflections
Data:
Nominal;
Qualitative
Validity:
Construct
Will a system
which employs
the use of spiral
mathematics
assessments be
received
positively by the
math department
and will the use of
such a system be
supported by the
administration?
Fullan
(2005)
Gabriel
(2005)
Marzano
(2003)
Method:
Focus Group
Interview
Data:
Qualitative
Validity:
Construct
Qualitative:
Coded for
themes:
Recurring
Dominant
Emerging
Qualitative:
Coded for
themes:
Recurring
Dominant
Emerging
Qualitative:
Look for
categorical and
repeating data
that form
patterns of
behaviors
Qualitative:
Look for
categorical and
repeating data
that form
patterns of
behaviors
Spiral Assessments 22
To determine whether or not spiral assessments could improve student
achievement on the Math 2 End-of-Course Test [EOCT], two groups of students’ test
scores were compared after the implementation of such assessments. The 65 students
from the 2009-2010 Accelerated Math 2 classes were used as the control group and
received no spiral assessments throughout their course. The 83 students from the 20102011 Accelerated Math 2 classes received spiral assessments for each mid-unit and final
unit test throughout their course. There were a total of ten spiral assessments
administered throughout the 2010-2011 school year. In addition, the 2010-2011
experimental group received daily spiral reviews at the beginning of class and received
immediate feedback on the reviews.
When looking at whether spiral assessments could improve students’ attitudes
towards, and confidence in, mathematics, the Math Anxiety Questionnaire [MAQ] (see
Appendix A) was administered as a pre- and post- implementation. The questionnaire
was given to the experimental group prior to spiral assessments and again at the
conclusion of the implementation. The scores were analyzed to determine if students’
math anxiety changed following the use of spiral assessments in the classroom.
To determine if spiral assessments would be received positively by the math
department, as well as the administration, a focus group and two interviews were
conducted at the conclusion of the study. In May, 2011, ten math teachers participated in
a focus group to discuss the general purpose of assessments and to discuss whether or not
spiral assessments could be feasibly administered in their classrooms (see Appendix B).
Lastly, following a comparison of the Math 2 EOCT scores for the experimental and
control groups, two administrators from the participating high school were interviewed
Spiral Assessments 23
about the use of spiral assessments, the feasibility of such assessments, and the
organizational change process for the school (see Appendix C).
Validity, Reliability, Dependability, Bias, and Equity
To determine if the use of spiral assessments could improve students’
achievement on the Math 2 EOCT, teacher-made midterm exams and a state required
end-of-course test were used to gather interval data about students’ content knowledge of
Math 2. Each assessment was examined and found to be valid, reliable, and free from
bias. According to Popham (2009), “assessment reliability refers to the consistency with
which a test measures whatever it’s measuring” (p. 25). Additionally, Salkind (2010)
explained, “a valid test is a test that measures what it is supposed to” (p. 151). The
midterm exams, as well as the Math 2 EOCT, exhibited content validity because they
were all aligned with the Georgia Performance Standards and covered content that was
taught in the classroom. Questions from each assessment were also checked and found to
be free of unfair or offensive bias and were also free of disparate impact.
To determine if spiral assessments would affect students’ feelings towards math
and improve student confidence, both nominal and qualitative data were gathered through
student surveys and student reflections before and after the implementation of spiral
assessments. Measures were taken with both instruments to assure validity, reliability,
and dependability. The Math Anxiety Questionnaire exhibited construct validity and was
considered reliable because it was constructed and tested by a reputable group of
statisticians, including Meece (1981) from the University of North Carolina. According
to Popham (2009), “…internal consistency reliability requires only a single testadministration to a single group of students. This is because it represents the degree to
Spiral Assessments 24
which a test’s items appear to be doing the same consistent measurement job” (p. 25). To
determine internal consistency reliability, Cronbach’s Alpha was run on both surveys.
In order to ensure dependability of the surveys and the student reflections, several
steps were taken. The surveys and reflections were administered at the beginning of each
class period on the same school day, and subjects were consistently given the same
directions throughout the day. There were an adequate number of subjects (eighty-three)
completing the surveys and reflections, and all subjects were given the option of
anonymity. The amount of time allowed between the first and second survey was
approximately seven months, following the use of approximately ten spiral assessments.
Finally, the student reflections were absent of bias because the subjects were simply
asked to reflect on their feelings towards math and towards spiral assessments. The
language used to question the subjects was fair and inoffensive and showed no sign of
disparate impact.
To determine whether or not spiral assessments would be received positively by
the math department and supported by the administration, a teacher focus group and two
administrator interviews were used to collect qualitative data. The data obtained from
these two sources were found to be valid, reliable, and dependable. Focus group
questions, as well as interview questions, exhibited construct validity because they
accurately measured what the focus question asked. All questions were submitted to, and
approved by, a LaGrange College advisor and were found to be reliable and free from
bias. The questions contained no unfair or offensive language and were found to be free
of disparate impact. There were an adequate number of participants in the teacher focus
group, and the participants represented a diverse sample of the high school’s math
Spiral Assessments 25
department. The focus group, as well as the administrator interviews, occurred at the end
of the spiral assessment implementation and all participants were given accurate results
from the student data.
In addition to checking for validity, reliability, dependability, and bias when
answering the study’s focus questions, an equity audit was performed by the researcher in
the school where the study was being conducted. Skrla, McKenzie, and Scheurich (2009)
explained, “equity audits are a systematic way for school leaders . . . to assess the degree
of equity or inequity present in three key areas of their schools or districts: programs,
teacher quality, and achievement” (p. 3). The purpose of the equity audit was to ensure
that all students, regardless of age, gender, race/ethnicity, national origin, or disability,
were receiving equal educational opportunities.
At the beginning of the study, the equity audit revealed that several efforts were
currently being taken to increase equity for students. Specifically, in the math
department, high quality teachers with varying levels of teacher education and experience
were distributed to all levels of classes. Additionally, great emphasis was placed on
collaborative planning for teachers, which provided students with equitable lessons and
common assessments. Lastly, while the equity audit revealed that achievement equity is
still an area in need of improvement; it also revealed that the school’s primary focus was
to close this achievement gap. The school was providing the students with several
opportunities to increase their achievement in classes and on standardized tests, and as a
result hoped to decrease the dropout rate.
Spiral Assessments 26
Analysis of Data
The data collected in this action research were centered on answering the three
focus questions presented in the study. When looking at whether or not spiral
assessments could improve Math 2 EOCT test scores, several independent t-tests were
run. To determine if there was a significant difference in the starting ability levels of the
experimental and control groups, an independent t-test was conducted on the eighth grade
CRCT z-scores. When a significant difference was found in the two groups, the z-scores
were then screened and matched to equalize the ability levels of the two groups. In
October, 2010, the experimental group took a mid-term exam and their scores were
compared to last year’s October midterm scores of the control group. Another
independent t-test was conducted to determine if a significant difference existed. In
March, 2011, the experimental group took a second midterm exam and their scores were
again compared to last year’s March midterm scores of the control group. Another
independent t-test was used to determine significance. Lastly, in May, 2011, the
experimental group’s Math 2 EOCT scores were compared to the control group’s Math 2
EOCT test scores and one last independent t-test was run. For all of the independent ttests, the decision to reject the null hypothesis was set at p<.05. Additionally, the effect
size was measured using Cohen’s d. A small effect size ranged from 0.0 to .20; a
medium effect size ranged from .20 to .50; and a large effect size was set for any value
above .50.
To determine whether spiral assessments affected students’ feelings towards
math, and whether such assessments improved students’ math confidence, the Math
Anxiety Questionnaire [MAQ] was administered to the experimental group two times
Spiral Assessments 27
during the study. The questionnaire was given at the beginning of the study and again at
the conclusion of the spiral assessment implementation. To determine if there was a
significant difference between the pre and post questionnaire scores, a dependent t-test
was run on the data and the decision to reject the null hypothesis was set at p < .05. The
questionnaire used a seven-point Likert scale and was analyzed using Cronbach’s Alpha
to determine internal consistency reliability. Additionally, each question was analyzed
using the Chi Square method to determine significance. The significance level was
reported at the p<.05, p<.01, and the p<.001 levels.
In addition to using the Math Anxiety Questionnaire to analyze students’ attitudes
about math, student reflections were collected towards the end of the study. Students
were asked to reflect on their feelings about math and also their feelings about the use of
spiral reviews and assessments throughout the year. The reflections were then examined
and coded for themes. Highlighters were used to color code recurring, dominant, and
emerging themes that surfaced in the reflections. Coded themes were used to look for
categorical and repeating data that formed patterns of behaviors.
To determine whether or not spiral assessments would be positively received by
the math department and supported by the administration, a focus group of math teachers
and interviews of administrators were conducted at the end of the study. The focus group
discussion and the interviews were all recorded and transcribed. The transcriptions were
then examined and coded for themes. As in the student reflections, highlighters were
used to color code recurring, dominant, and emerging themes found in the data.
Spiral Assessments 28
While data were collected and analyzed to answer three focus questions, the study
was also analyzed holistically to ensure it had validation, credibility, transferability, and
to ensure that it was transformational.
Validation
This study exhibited ‘consensual validation’ because it was approved, reviewed,
and supported by the LaGrange College faculty. At the conclusion of the study, findings
were presented and the study was defended to the LaGrange College faculty. The study
also contained ‘epistemological validation’ because data results were compared to the
review of the literature to determine whether or not the literature supported the findings.
Credibility
Credibility of the study was ensured through the use of multiple data sources. To
answer the three focus questions, data were collected and analyzed from teacher-made
assessments, state assessments, student questionnaires, student reflections, a teacher
focus group, and administrator interviews. Eisner (1991) calls this process ‘structural
corroboration,’ where a confluence of evidence comes together to form a compelling
whole. Within this concept of structural corroboration are embedded the concepts of
‘fairness’ and ‘rightness of fit’. Fairness was achieved in the review of the literature by
presenting opposing points of view regarding the concept of a spiral curriculum. In order
to achieve precision and rightness of fit, extensive research of the literature was
conducted to present a tight argument and a coherent case for the use of spiral
assessments. Additionally, great care was taken when analyzing the data, both
quantitatively and qualitatively, in order to provide strong evidence for judgments made.
Spiral Assessments 29
Transferability
This study was transferable because it can be used by others and applied in
other classroom settings. Eisner (1991) calls this process ‘referential adequacy’ and
explained how perception and understanding by others will increase because of the
research conducted during this study. This study can be easily replicated and can be used
by others for future research to help increase student retention of any content.
Transformational
This study was transformational and contained ‘catalytic validity’ (Lather as cited
by Kinchloe & McLaren, 1998) because of the positive change it brought about in the
researcher, the students, and other math teachers. Students, as well as teachers,
discovered the positive effects that spiral assessments had on content retention. With this
discovery, teachers hoped to find additional ways to include spiraling within the
classroom.
Spiral Assessments 30
CHAPTER 4: RESULTS
The following quantitative and qualitative data were collected during this action
research study. The results of the data were organized around the three focus questions.
To answer focus question one and determine whether or not spiral assessments
would improve student test scores on the Math 2 EOCT, several independent t-tests were
run throughout the study. The purpose of the t-tests was to compare the means of two
different groups (control and experimental), and to determine if there was any
significance between the differences in the means (Salkind, 2010). Additionally, to
determine the magnitude of the difference, the effect size was measured for each t-test
using Cohen’s d. A small effect size ranged from 0.0 to .20; a medium effect size ranged
from .20 to .50; and a large effect size was set for any value above .50.
First, to determine if there was a significant difference in the ability levels of the
experimental and control groups, an independent t-test was conducted on the eighth grade
CRCT z-scores (see Table 4.1). The null hypothesis was there will be no significant
difference between the ability levels of the 2009 control students and the 2010
experimental students.
Spiral Assessments 31
Table 4.1. Independent t-test: Eighth Grade CRCT z-scores
t-Test: Two-Sample
Assuming Unequal
Variances
Mean
Variance
Observations
Hypothesized Mean
Difference
df
t Stat
P(T<=t) one-tail
t Critical one-tail
P(T<=t) two-tail
t Critical two-tail
t(130) = 2.464, p < .05
z-score 2009
5.444
7.916
65
z-score 2010
6.708
9.467
67
0
130
-2.464
0.008
1.657
0.015
1.978
The data from the eighth grade CRCT z-scores showed the mean z-score for the 2009
students was 5.44 and the mean z-score for the 2010 students was higher at 6.71. Since
the obtained value, 2.464, was greater than the critical value, 1.657, the null hypothesis
was rejected. The results of the eighth grade CRCT z-scores showed that there was a
significant difference between the ability levels of the two groups. A Cohen’s d value of
0.431 indicated a medium effect size in the difference.
Following this independent t-test, the CRCT z-scores from the 2010 group were
matched with the closest CRCT z-scores from the 2009 group. The matching of CRCT zscores left 98 students in the study with equal ability levels, 49 students from each group.
To verify no significant difference in the ability levels existed, another independent t-test
was conducted on the eighth grade CRCT z-scores (see Table 4.2). The null hypothesis
was there will be no significant difference between the ability levels of the 2009 control
students and the 2010 experimental students.
Spiral Assessments 32
Table 4.2. Independent t-test: Equalized Eighth Grade CRCT z-scores
t-Test: Two-Sample
Assuming Equal
Variances
Mean
Variance
Observations
Pooled Variance
Hypothesized Mean
Difference
df
t Stat
P(T<=t) one-tail
t Critical one-tail
P(T<=t) two-tail
t Critical two-tail
t(96) = 0.016, p > .05
z-scores 2010
6.274
8.641
49
7.745
z-scores 2009
6.283
6.850
49
0
96
-0.016
0.494
1.661
0.987
1.985
The equalized data from the eighth grade CRCT z-scores showed the mean zscore for the 2009 students was 6.28 and the mean z-score for the 2010 students was
slightly lower at 6.27. Since the obtained value, 0.016, was less than the critical value,
1.661, the null hypothesis was accepted. The results of the equalized eighth grade CRCT
z-scores verified there was no significant difference between the ability levels of the two
groups. A Cohen’s d value of 0.004 indicated a minimal effect size in the difference.
In October, 2010, the experimental group took a mid-term exam and their scores
were compared to last year’s October midterm scores of the control group. Another
independent t-test was conducted to determine if a significant difference existed between
the two groups (see Table 4.3). The null hypothesis was there will be no significant
difference between the October midterm scores of the 2009 control students and the 2010
experimental students.
Spiral Assessments 33
Table 4.3. Independent t-test: October Midterm Scores
t-Test: Two-Sample
Assuming Equal
Variances
Mean
Variance
Observations
Pooled Variance
Hypothesized Mean
Difference
df
t Stat
P(T<=t) one-tail
t Critical one-tail
P(T<=t) two-tail
t Critical two-tail
t(96) = 1.8295, p < .05
Oct Midterm 2010
82.367
98.446
49
166.929
Oct Midterm 2009
77.592
235.413
49
0
96
1.830
0.035
1.661
0.070
1.985
The mean score of the 2009 students was 77.59, which was lower than the mean
score of the 2010 students at 82.37. Since the obtained value, 1.830, was greater than the
critical value, 1.661, the null hypothesis was rejected. The results of the October
midterm scores revealed there was a significant difference between the midterm scores of
the 2010 experimental group and the 2009 control group. A Cohen’s d value of 0.370
indicated a medium effect size in the difference.
In March, 2011, the experimental group took a second midterm exam and their
scores were again compared to last year’s March midterm scores of the control group.
Another independent t-test was used to determine significance (see Table 4.4). The null
hypothesis was there will be no significant difference between the March midterm scores
of the 2009 control students and the 2010 experimental students.
Spiral Assessments 34
Table 4.4. Independent t-test: March Midterm Scores
t-Test: Two-Sample
Assuming Equal
Variances
Mean
Variance
Observations
Pooled Variance
Hypothesized Mean
Difference
df
t Stat
P(T<=t) one-tail
t Critical one-tail
P(T<=t) two-tail
t Critical two-tail
t(96) = 5.329, p < .05
March Midterm
2010
89.265
109.074
49
146.236
March Midterm
2009
76.245
183.397
49
0
96
5.329
3.25574E-07
1.661
6.51149E-07
1.985
Again, the mean score of the 2009 students, 76.25, was lower than the mean score
of the 2010 students at 89.27. Since the obtained value, 5.329, was greater than the
critical value, 1.661, the null hypothesis was rejected. The results of the March midterm
scores revealed there was a significant difference between the test scores of the 2010
experimental group and the 2009 control group. A Cohen’s d value of 1.078 indicated a
large effect size in the difference.
Lastly, in May, 2011, the 2010 experimental group’s Math 2 EOCT scores were
compared to the 2009 control group’s Math 2 EOCT test scores and one last independent
t-test was run (see Table 4.5 and Table 4.6). The null hypothesis was there will be no
significant difference between the Math 2 EOCT scores of the 2009 control students and
the Math 2 EOCT scores of the 2010 experimental students.
Spiral Assessments 35
Table 4.5. Independent t-test: Math 2 EOCT Raw Scores
t-Test: Two-Sample
Assuming Equal
Variances
Mean
Variance
Observations
Pooled Variance
Hypothesized Mean
Difference
df
t Stat
P(T<=t) one-tail
t Critical one-tail
P(T<=t) two-tail
t Critical two-tail
t(96) = 0.513, p > .05
eoct(math2) 2010
448.510
669.880
49
549.761
eoct(math2) 2009
450.939
429.642
49
0
96
-0.513
0.305
1.661
0.609
1.985
Table 4.6. Independent t-test: Math 2 EOCT Percentile Scores
t-Test: Two-Sample
Assuming Equal
Variances
Mean
Variance
Observations
Pooled Variance
Hypothesized Mean
Difference
df
t Stat
P(T<=t) one-tail
t Critical one-tail
P(T<=t) two-tail
t Critical two-tail
t(96) = 1.035, p > .05
eoct(math2) 2010
86.429
28.792
49
25.773
eoct(math2) 2009
87.490
22.755
49
0
96
-1.035
0.152
1.661
0.303
1.985
Although the mean percentile score of the 2010 experimental group, 86.43, was
lower than the mean percentile score of the 2009 control group, 87.49, there was not a
Spiral Assessments 36
large enough difference to make it significant. Since the obtained value, 1.035, was less
than the critical value, 1.661, the null hypothesis was accepted. The results of the Math 2
EOCT scores revealed there was no significant difference between the test scores of the
2010 experimental group and the 2009 control group. A Cohen’s d value of 0.209
indicated a medium effect size in the difference.
To answer focus question two and determine if spiral assessments would affect
students’ feelings towards math, and improve student confidence, a Math Anxiety
Questionnaire was administered to the 2010 experimental students at the beginning and at
the end of the study. To determine if there was a significant difference between the preand post-questionnaire scores, a dependent t-test was run on the data (see Table 4.7). The
null hypothesis was there will be no significant difference between the scores on the
Math Anxiety Questionnaire administered at the beginning and the end of the study.
Table 4.7. Dependent t-test: Pre- and Post-Math Anxiety Questionnaire
t-Test: Paired Two Sample
for Means
Mean
Variance
Observations
Pearson Correlation
Hypothesized Mean
Difference
Df
t Stat
P(T<=t) one-tail
t Critical one-tail
P(T<=t) two-tail
t Critical two-tail
t(82) = 18.493, p<.05
PreTotal
39.422
158.320
83
0.993
0
82
18.493
2.71382E-31
1.664
5.42764E-31
1.989
PostTotal
36.157
174.353
83
Spiral Assessments 37
The result of the dependent t-test revealed that the mean score for the prequestionnaire was 39.42 and the mean score for the post-questionnaire was 36.16,
showing a decrease in the average survey score from pre to post. Since the obtained
value, 18.493, was greater than the critical value, 1.664, the null hypothesis was rejected.
The results of the dependent t-test showed there was a significant difference between the
totals obtained from the pre-Math Anxiety Questionnaire and the post-Math Anxiety
Questionnaire. An effect-size r value of 0.125 indicated a small effect size in the
difference.
With each questionnaire, the chi-square test statistic was calculated to compare
what was observed on the questionnaire to what would be expected by chance (Salkind,
2010). Table 4.8 revealed the results of the chi-square tests for both the pre- and postquestionnaires. Since the questionnaire used a seven-point Likert scale, the degrees of
freedom for the chi square test was 6 and, at a .05 level of significance, the critical value
was 12.59. The obtained value for each question are listed in the table and the
significance level was reported at the p<.05, p<.01, and the p<.001 levels.
Spiral Assessments 38
Table 4.8. Chi-Square Statistic for Math Anxiety Pre- and Post-Questionnaire
Survey
Items
n=11
Item 1
Survey Question
PreQuestionnaire
χ2
When the teacher says he/she is going to ask 30.6667 ***
you some questions to find out how much
you know about math, how much do you
worry that you will do poorly?
When the teacher is showing the class how
to do a problem, how much do you worry
that other students might understand the
problem better than you?
When I am in math class, I usually feel at
ease and relaxed.
When I am taking math tests, I usually feel
nervous and uneasy.
Taking math tests scares me.
Item 2
Item 3
Item 4
Item 5
Item 6
Item 7
I dread having to do math.
Item 8
Item 9
Item 10
Item 11
It scares me to think that I will be taking
advanced high school math.
In general, how much do you worry about
how well you are doing in school?
If you are absent from school and you miss
a math assignment, how much do you worry
that you will be behind the other students
when you come back to school?
In general, how much do you worry about
how well you are doing in math?
Compare to other subjects, how much do
you worry about how well you are doing in
math?
PostQuestionnaire
χ2
43.5833 ***
50.3333 ***
72.1667 ***
65.8333 ***
55.3333 ***
7.5
22.5 ***
35.5 ***
45.3333 ***
99.8333 ***
48.5 ***
46.1667 ***
122.8333 ***
17.6667 **
5.5
8.6667
9.3333
5.3333
7.6667
14.1967 *
11.1667
* p<.05, **p<.01, ***p<.001
The results of the chi-square statistic for the Math Anxiety pre-Questionnaire
highlighted several significant items. Questionnaire items 1, 2, 3, 5, 6, 7, 8, and 11 were
all found to be highly significant when p < .05, .01, and .001, meaning that there were a
high percentage of students that answered a certain way on these questions. However,
items 4, 9, and 10 were not significant at all, which means that there was no significant
difference on these questions between what was observed in the answers and what would
have been expected to occur by chance.
Spiral Assessments 39
The results of the chi-square statistic for the Math Anxiety post-Questionnaire
also highlighted several significant items. Questionnaire items 1, 2, 3, 4, 5, 6, and 7 were
all found to be highly significant when p < .05, .01, and .001, meaning that there were a
high percentage of students that answered a certain way on these questions. However,
items 8, 9, 10, and 11 were not significant at all.
To determine the internal consistency reliability of the items on the Math Anxiety
pre- and post-Questionnaires, the Cronbach’s Alpha test was conducted using the
questionnaire responses. According to Salkind (2010), “[internal consistency reliability]
is used when you want to know whether the items on a test are consistent with one
another in that they represent one, and only one, dimension, construct, or area of interest”
(p. 147). For the Math Anxiety pre-Questionnaire, the computations gave a Cronbach’s
Alpha of 0.98 and for the Math Anxiety post-Questionnaire, the Cronbach’s Alpha was
0.99. Therefore, both of these surveys showed a high level of internal consistency
reliability using the results of the Cronbach’s Alpha test.
In addition to using the Math Anxiety Questionnaire to analyze students’ attitudes
about math, student reflections were collected towards the end of the study and
qualitative data were collected. Students were asked to reflect on their feelings about
math and also their feelings about the use of spiraled warm-up exercises and “blast from
the past” questions on their assessments. Out of 82 reflections, 72 students expressed
positive feelings towards math and 68 students expressed positive feelings towards
warm-ups and spiral assessments. A struggling student explained, “I like the warm-up
reviews and the ‘blast from the past.’ I forget things very quickly and need to be
reminded of formulas and how to work problems. Sadly, I think my grade might be even
Spiral Assessments 40
worse if we didn’t do these.” Another student stated, “I understand math easily enough, I
just don’t always remember it. This [blast from the past] will help me on the EOCT not
to forget how to work the problems.” An additional student commented, “Coming into
class and having warm-ups helps me remember how to do previous things I’ve learned.
I’ve always picked up on math quickly, but sometimes I don’t always hold onto it.”
Another student who is a big fan of the reviews stated, “We NEED warm-ups and ‘blast
from the past.’ Do NOT get rid of them! I have a bad habit of memorizing a formula or
a way to solve a problem for the test instead of learning and understanding the concept.
Those review problems help refresh my memory and show me what I need to focus on
for the EOCT.”
Although the majority of students agreed that revisiting content would help them
retain information and help their scores on the EOCT, some students suggested that the
“blast from the past” questions should count as bonus points on their tests. One student
explained, “Warm-ups and reviewing helps me out a lot and I think the more I see it, the
more I can remember it. Same goes for ‘blast from the past’; I just hate how it brings my
grade down sometimes.” This student’s sentiment was echoed on several reflections.
Another student commented, “I like the ‘blast from the past’ but it should be bonus.
Sometimes you don’t know what the blast from the past is and sometimes it’s not fresh
on your memory. The last test we took, ‘blast from the past’ is what killed me.”
Offering bonus points for “blast from the past” questions appeared on 23 different
reflections.
To answer focus question three and determine whether or not spiral assessments
would be received positively by the math department and supported by the
Spiral Assessments 41
administration, a teacher focus group and administrator interviews were used to collect
qualitative data. During the teacher focus group, six questions were asked regarding the
purpose and types of assessment, as well as the prospect of implementing spiral
assessments. Throughout the focus group discussion, the following three themes
emerged: (1) assessments are important sources of information in the beginning, middle,
and end of a unit, (2) students retain information better with more exposure, and (3) time
is a major issue.
When asked about the purpose of classroom assessments, teachers listed several
uses. Assessments are used to check for student understanding, to determine if the
students are ready to “move on”, to allow students to self-evaluate, and to provide
information for grades. Secondly, teachers explained that the best time to administer
assessments was in the beginning, middle, and end of a unit. Although Teacher 1 stated
the importance of pre-testing a unit to show student gain, other teachers agreed that there
was not enough time to add yet another assessment to the many others currently being
administered in their classrooms. Teacher 3 said the time for testing was “continuously.”
Next, teachers were asked what different forms of evaluation were currently being
utilized in their classrooms and if there were any forms they liked better than others. The
types of assessments being used fell into two categories, traditional and formative. The
type, and amount, of technology teachers had in their classrooms also affected what they
were able to do with assessments. Teacher 1 explained how she used graphing
calculators to conduct “quick polls” throughout the unit to provide immediate feedback to
her students. Teachers without this technology have never had the opportunity to
administer quick polls, but did express an interest in doing so. The majority of teachers
Spiral Assessments 42
listed traditional forms of assessment such as homework, bellwork, quizzes, and tests. A
couple of the teachers also listed formative types of assessment such as, “tickets-out-thedoor”, individual student white boards, group work, and classroom questioning.
Finally, following an explanation of spiral assessments, teachers were asked
whether or not spiral assessments were valid forms of assessments, if they were feasible
in their classrooms and what reservations they had with such assessments. Teacher 3 said
formative assessments were “absolutely” valid, “absolutely” feasible, and that he had no
reservations with such assessments. Teacher 3 also explained that he had been using
spiral assessments in his classroom for the last thirty years and planned to continue doing
so until he no longer taught. Teacher 4 stated, “Math is all about spiraling. Students
have to know previous material in order to master future material.” Teacher 5 added,
“Just like us, when students don’t use it, they lose it.”
Despite these positive comments and the consensus that spiral assessments were
valid and worthwhile, the overall opinion of the group was that spiral assessments were
not very feasible and would be difficult to implement in their classrooms. The majority
of teachers in the focus group said the biggest deterrent to using spiral assessments was
time. Teacher 6 expressed her concern that students would complain, “you didn’t tell us
this was going to be on the test.” She explained that students always want to know
exactly what is on the test with “no surprises”, and because of that, students would expect
her to keep reviewing material and “[she] would never have time to get done what [she]
needed to get done.” Other teachers expressed a concern that preparing spiral
assessments would take extra time and Teacher 7 added, “It will take more time to assess
students because you have to add more questions to every test.”
Spiral Assessments 43
Following the arrival of the Math 2 EOCT scores, interviews were conducted
with the principal and an assistant principal of the high school to share the results of the
study and to determine if the administration would support the use of spiral assessments.
The first six questions of the interview mimicked those of the teacher focus group and the
answers provided by the administrators were very similar to the answers provided by the
teachers. Two additional administrator questions provided suggestions for implementing
change in the classroom and in the school as a whole. During the administrator
interviews, the following three themes emerged: (1) formative and summative
assessments provide valuable information to students, parents, and teachers, (2) practice
makes perfect, and (3) change is a gradual process.
When asked about the purpose of classroom assessments, both administrators
listed reasons similar to those listed by the teachers. Administrators expected teachers to
use assessments throughout a unit to ascertain what their students have mastered and to
make decisions regarding their instruction. They also valued assessments as a source of
documentation and as a device for reporting progress to both parents and students.
Administrator 1 explained, “Teachers and students should use assessments to gather
information about content knowledge. Assessments not only show what the student has
learned, but also how the teacher has taught.” Administrator 2 explained, “Formative
assessments are just as important, if not more important, than the summative ones.
Teachers can’t change what they’re doing if they don’t know what their students are
learning.”
Due to the important role assessments play in teachers’ classrooms, both
administrators agreed that a variety of assessments should be utilized everyday in every
Spiral Assessments 44
classroom. Administrator 1 stated, “Whether formally or informally, students should be
assessed everyday throughout the entire school year.” Administrator 2 commented,
“Assessments should be an on-going, continuous part of every class.” When asked about
the types of assessments observed most in classrooms, Administrator 1 distinguished
between formal and informal assessments. In terms of formal assessments, all teachers
used quizzes, unit tests, benchmarks, midterms, and final exams as part of their grading
practices. Regarding informal assessments, the use of student questioning was observed
the most. Some teachers incorporated “tickets-out-the-door” and other quick feedback
activities; however, the majority of teachers simply asked questions to gather information
about their students. Administrator 1 explained, “Some teachers are more skilled
questioners than others. Some teachers have a natural ability to gather information from
their students, simply by the questions they ask.”
Following general assessment questions, administrators were asked about the
validity and feasibility of spiral assessments in the classroom. Similar to focus group
responses, administrators believed spiral assessments were valid and extremely useful in
the classroom. Administrator 2 stated, “Students wouldn’t be allowed to forget what they
had learned. They would know they were responsible for the content over and over and
over again.” Administrator 1 recounted a statistics class taken in college where spiral
assessments were administered throughout the entire course. She claimed to have learned
more in that course than in any other and said that at the end of the course, “it all fit
together.” Administrator 1 credited this statistics class, and the assessments used in the
class, for her own implementation of cumulative tests when she was a classroom teacher.
Spiral Assessments 45
Although she administered a separate unit test and then a cumulative review and then a
cumulative test, the spiral concept was there.
When asked about concerns or potential problems with spiral assessments,
Administrator 1 was quick to say, “Some teachers just won’t do it…not because they
don’t believe in it, but because they’re too lazy to put forth the effort.” Changing to
spiral assessments would mean teachers would have to re-create tests they’ve been using
for years. Some teachers would be willing to make the change, others would not.
Administrator 2 also noted the problem with teachers not taking corrective action once
spiral assessments were administered. He explained, “If a teacher doesn’t do anything to
correct a problem a student is having, then that student is going to get penalized over and
over again for the same gap in knowledge.”
At the conclusion of the interviews, both administrators were asked whether or
not they thought teachers would be open to administering spiral assessments in their
classrooms and what was the best way to bring about change within a school. Again,
Administrator 1 explained that some teachers would resist the change, “no matter how
good the change was.” Administrator 2 commented, “Some teachers come on board
immediately, some wait until they see others do it, and others never come on board.”
Both administrators suggested a gradual change policy when implementing something
new in the school. Similar to changes that take place now, research and data should be
collected to ensure spiral assessments are worth doing in the first place. The faculty
should then be presented the research and data, and given time to study and understand
the change. When the time comes to move to spiral assessments, Administrator 1
suggested each teacher choose one class to start the process with instead of trying to do
Spiral Assessments 46
all of their classes at one time. The next year each teacher could pick up another class
and keep going from there. Administrator 2 also recommended working with other
teachers in the same subject to collaboratively create the spiral assessments. He added,
“Working with someone else to lessen the burden always makes change easier.”
Spiral Assessments 47
CHAPTER 5: ANALYSIS AND DISCUSSION OF RESULTS
Analysis
The purpose of this study was to determine the effectiveness of spiral assessments
on content retention and student attitudes in a mathematics classroom. To determine the
effectiveness of spiral assessments, both quantitative and qualitative data were collected
throughout the study and used to answer three focus questions.
To answer focus question one, “will the use of spiral assessments, coupled with
proper feedback, improve students’ achievement on the Math 2 End of Course Test?”,
data were collected on two teacher-made midterm exams and a state-made Math 2 EOCT,
for a 2009 control group and a 2010 experimental group. The control group received no
spiral assessments, while the 2010 received numerous spiral assessments. The teachermade midterm exams were identical for the 2009 and 2010 groups, while the state-made
Math 2 EOCT was created by the state with no information on its consistency from one
year to the next. Additionally, 8th grade CRCT data were used in the study to screen zscores in an attempt to equalize the control and experimental groups.
Following an equalization of ability levels in the control and experimental groups,
data were collected on the midterm and EOCT test scores, and three independent t-tests
were run to determine whether or not a significant difference existed. The results of the
October midterm revealed a mean score of 77.89 for the 2009 control group and a higher
mean score of 82.37 for the experimental group. The independent t-test revealed that a
significant difference existed in the groups and a Cohen’s d value determined a medium
effect size. From the results of the October midterm, it appeared that spiral assessments
were having a positive effect on content retention for the 2010 experimental group.
Spiral Assessments 48
The March midterm revealed more promising data on spiral assessments. While
the 2009 control group only averaged 76.24 on the test, the 2010 experimental group
boasted an 89.27 average. Again, an independent t-test revealed a significant difference
in the data with a large Cohen’s d effect size. From the results of the March midterm, it
appeared that, given more time, spiral assessments had an even greater positive effect on
content retention for the 2010 experimental group.
Unfortunately, the Math 2 EOCT results did not echo the same promising data of
the midterm data. The 2009 control group averaged a raw score of 450.94, percentile
score of 87.49, whereas the 2010 experimental group averaged lower scores of 448.51
and 86.43 respectively. If the results of the Math 2 EOCT are viewed in isolation, it
would appear that spiral assessments had a negative effect on content retention for the
2010 experimental group. However, since the EOCT data were in direct contrast to data
collected from both midterms, one has to question whether or not the state altered their
versions of the Math 2 EOCT from one year to the next.
Although the results in the study contradicted each other from teacher-made to
state-made tests, a possible reason for the contradiction was supported in the review of
literature. Jones et al. (2007) cautioned against using one test score to provide the whole
picture. They explained, “A test is a snapshot in time, affected by numerous sources of
error . . . The more information a teacher collects, the more valid the inferences based on
that information” (Jones et al., 2007, p. 74). Therefore, when interpreting results
produced in the study, it is important to not just look at the end result, but also to look at
what happened along the way.
Spiral Assessments 49
In the review of the literature, Bruner (1960) and Doll (1993) emphasized the
importance of revisiting content in order to increase student mastery. Schulman (1998)
also supported the spiral concept and added, “by revisiting the same questions throughout
the school year, children are encouraged to amplify their mathematical thinking” (p. 72).
Schulman (1998) further explained how teachers could focus their instruction based on
gathered information from their student “revisits” (p. 72). Based on the literature and the
results from the teacher-made midterms, revisiting problems throughout the year helped
the 2010 students with content retention.
Also supported in the literature were that formative assessments contributed to the
positive results of the teacher-made midterms. Throughout the 2010 school year, greater
attention was placed on using assessments as an informational tool rather than just a
means for providing a student grade. Reeves (2007) warned, “as long as we use
[assessments] only as a means to rank schools and students, we will miss out on their
most powerful benefits” (p. 27). He further stated, “when teachers’ classroom
assessments become an integral part of the instructional process and a central ingredient
in efforts to help students learn, the benefits of assessment for both teachers and students
will be boundless” (p. 28). Throughout the study, daily warm-ups, quizzes, white board
activities, and summative assessments, all spiral in nature, were used to guide
improvements in instruction. The increased use of formative assessments was heavily
supported in the literature and appeared to have a positive effect on content retention.
To address the second focus question, “how will spiral assessments affect
students’ feelings towards math, and will such assessments improve student
confidence?”, surveys and reflections were administered throughout the 2010 school year
Spiral Assessments 50
to the experimental group. At the beginning and end of the study, quantitative data were
collected on the Math Anxiety Questionnaire [MAQ] using a seven-point Likert Scale. A
student answering with extreme anxiety on each of the questions would have scored a 71.
Both the pre- and post-questionnaire mean scores were well below this extreme case.
The pre-questionnaire mean score was only a 39.42 with the post-questionnaire mean
score not far away with 36.16. Although the Accelerated Math students that participated
in the study exhibited low levels of anxiety even at the beginning, they did show a
decrease in their levels at the end. The dependent t-test run on the pre- and postquestionnaire data revealed that there was a significant difference in the two scores. The
drop in pre and post scores revealed students had lower anxiety following the use of
spiral assessments. Internal consistency reliability was proven with a Cronbach’s Alpha
score of a .98 and .99 for pre- and post-questionnaire respectively.
Following the dependent t-test, each question was analyzed using the Chi Square
method to determine whether or not students answered significantly in one direction.
Item 1, “When the teachers says he/she is going to ask you some questions to find out
how much you know about math, how much do you worry that you will do poorly?”, and
Item 2, “When the teacher is showing the class how to do a problem, how much do you
worry that other students might understand the problem better than you?”, were both
significant at the p<.001 level, with students predominantly answering on the low end of
the scale, with a 1 “not at all”. On the pre-questionnaire, 63% answered between 1 and 3
on Item 1 and 75% on Item 2. These percentages grew to 76% on Item 1 and 80% on
Item 2 on the post-questionnaire which indicated more students were less worried at the
end of the study.
Spiral Assessments 51
Another item of interest was Item 3, “When I am in math class, I usually feel at
ease and relaxed”. On the pre-questionnaire, only 18% responded with a 7 “very much at
ease”; however, 31% responded with a 7 on the post-questionnaire. This increase from
18% to 31% indicated that more students felt comfortable in math class at the end of the
study. Item 4, “When I am taking math tests, I usually feel nervous and uneasy”, did not
show any significance on the pre-questionnaire because students’ responses were evenly
spread along the scale. However, following spiral assessments, the question became
significant at the p<.001 level, with students predominantly choosing on the low end of
the scale with a 1 “not at all nervous”.
Item 5, “Taking math tests scares me”, and Item 6, “I dread having to do math”,
produced similar results to Items 1 and 2. Both questions were significant at the p<.001
levels because students predominantly chose numbers on the lower end of the scale with
a 1 “never”. Again, pre and post totals revealed an increase in percentages of students
choosing lower numbers on the post-questionnaire than the pre-questionnaire.
Item 7, “It scares me to think that I will be taking advanced high school math”,
was the most significant question in terms of chi square numbers, but the least significant
question in terms of the study. Accelerated Math students have already chosen the
advanced math path and therefore chose mostly 1’s and 2’s for this question. However,
even with this extreme case, the percentage of students choosing a 1, “not at all”, grew
from 42% on the pre-questionnaire to a 54% on the post-questionnaire.
Items 8 through 11 did not offer much evidence for this study. Item 8, “In
general, how much do you worry about how well you are doing in school?”, did show
significance at the p<.01 level with students choosing on the higher end of the scale.
Spiral Assessments 52
However, this question addressed anxiety towards school and not specifically towards
math. Also, students appeared to be more worried about school at the beginning of the
year than they did at the end of the year. Item 9, “If you are absent from school and you
miss a math assignment, how much do you worry that you will be behind the other
students when you come back to school?”, was probably not significant because
Accelerated Math students, in general, have good attendance rates. Item 10, “In general,
how much do you worry about how well you are doing in math?”, was also not
significant. If students thought this question was referring to their grade in math class,
then this lack of significance could be attributed to the fact that most Accelerated Math
students have high math grades. Lastly, Item 11, “Compared to other subjects, how much
do you worry about how well you are doing in math?”, exhibited a small significance in
the pre-questionnaire and no significance in the post-questionnaire. Again, students may
have been a little more worried at the beginning of the year, but at the end they knew
their class average and knew how their math grade compared to other subjects.
In addition to the pre- and post-Math Anxiety Questionnaire, student reflections
were administered at the end of the study. Student reflections provided qualitative data
for the second focus question and focused on attitudes about math and the use of spiraled
reviews and assessments. Because the study included subjects that were enrolled in
Accelerated Math, it was not surprising that the majority of students expressed positive
feelings towards math in their reflections. However, I was a little surprised that students
reflected so positively on the spiraled warm-ups and assessments. Most students
recognized the importance of practicing content in order to master it. Additionally, most
students were in favor of the “blast from the past” test questions and believed the
Spiral Assessments 53
questions would help them perform better on the Math 2 EOCT. The student’s comment,
“Do NOT get rid of them”, demonstrated how important he thought the spiraled content
was.
Although the students in this study exhibited less anxiety and more positive
feelings towards math from the beginning, their decrease in anxiety demonstrated in the
MAQ results and positive comments in the student reflections were supported in the
literature. Reeves (2007) explained, “when students are involved in the classroom
assessment process, they are more engaged and motivated, and they learn more” (p. 31).
McMillan et al. (2010) explained, “…overall formative practices showed a positive
relationship with student motivation” (p. 11). Throughout the 2010 school year, students
contributed to the creation of the spiraled warm-ups and were provided immediate, daily
feedback. Reeves (2007) added, “students are highly motivated when they have more
choice during the learning process and receive more quality feedback” (p. 36).
To answer the third focus question, “will a system which employs the use of
spiral mathematics assessments be received positively by the math department and will
the use of such a system be supported by the administration?”, a focus group of ten math
teachers and interviews of two administrators were conducted at the conclusion of the
study. Qualitative data were collected from the focus group discussion and the interviews
and then coded for themes.
As discussed in Chapter 4, the math teacher focus group revealed the following
three themes: (1) assessments are important sources of information in the beginning,
middle, and end of a unit, (2) students retain information better with more exposure, and
(3) time is a major issue. Basically, teachers agreed that student retention is an ongoing
Spiral Assessments 54
problem and a problem that they have not been able to remedy over the years. They also
agreed that the more a student practices a concept, the more they understand it.
Additionally, teachers viewed assessments as a major component of their classroom and
agreed that assessments provided them a wealth of information.
The discouraging data that emerged from the focus group was that, although
teachers supported the spiraling concept, they were leery to make a change because of the
time and effort it would take on their part. I believe this lack of teacher willingness
resulted from enduring years of constant change. Each year, teachers are forced to
participate in school improvement efforts which they may or may not believe in. Further
frustration arises when the improvement effort is short lived and another change is
implemented before their prior efforts have time to work.
Similar to the teacher focus group, the administrator interviews revealed the
following three themes: (1) formative and summative assessments provide valuable
information to students, parents, and teachers, (2) practice makes perfect, and (3) change
is a gradual process. In general, teachers and administrators viewed assessments in the
same light and stressed the important role assessments play in classroom instruction.
Additionally, both administrators expressed an expectation for teachers to use
assessments frequently and to use a variety of assessments to make informed decisions in
their classrooms. Administrators also recognized the importance spiraling content
because it allowed students to practice content over and over. Lastly, administrators
recommended a gradual time frame for implementing a change with teachers.
Although both administrators were in full support of using spiral assessments in
the classroom, neither was ready to jump on board with a school-wide implementation. I
Spiral Assessments 55
believe this lack of willingness to bring about school-wide change also stemmed from
tolerating a multitude of changes over the last few years. I believe administrators are as
frustrated as teachers when it comes to school improvement. Some changes are forced
upon them by the state or county, and other changes result from their own efforts.
Regardless of where the change originates, I believe administrators recognized teacher
frustration and feared what another change may do to the morale of the school.
The ability of teachers and administrators to greatly impact change reform in
school was heavily supported in the literature (Fullan, 2009; Gabriel, 2005; Marzano et
al., 2001). The concerns and reluctance of teachers and administrators to implement
spiral assessments were also evident in the literature. Marzano et al. (2001) commented,
“busy teachers who have been doing things the same way for a fair amount of time will
have many valid reasons for not trying a new strategy. What is clearly required to alter
the status quo is a sincere desire to change and a firm commitment to weather the
inevitable storms as change occurs” (pp. 157-158). They further commented,
“Administrators and classroom teachers are often overwhelmed by the sheer amount of
change attempted and the work involved” (Marzano et al., 2001, p. 159). Lieberman
(2005) agreed with the amount of changes schools attempt and referred to those change
numbers as “staggering” (p. vii). It was clear in the literature that teachers and
administrators across the nation feel as frustrated as the teachers and administrators in the
study when it comes to change implementation.
Discussion
Overall, this research study produced positive results with regards to content
retention and student confidence. Quantitative data showed spiraled students with an
Spiral Assessments 56
advantage over non-spiraled students on teacher-made midterms. Quantitative data on
the Math Anxiety Questionnaire also revealed spiraled students with less math anxiety
following the implementation of spiral assessments.
Although quantitative data were inconsistent between teacher-made and statemade tests, qualitative data were consistent among students, teachers, and administrators.
Qualitative data revealed a general consensus that students lose content knowledge when
they do not use the knowledge over and over. Students, teachers, and administrators also
recognized the importance of assessments in the classroom, and in particular, the positive
effect spiral assessments could have over student retention and student confidence.
Students’ not only made positive comments about ‘blast from the past’ questions, but also
about feeling prepared for their EOCT.
This study was successful in highlighting the importance of using meaningful
assessments in the classroom to make informed decisions about students’ content
knowledge and to provide students with effective feedback. This study also focused on
the importance of revisiting concepts throughout a course in order to ensure student
mastery and produced results that quantitatively and qualitatively supported the spiral
concept.
This study achieved structural corroboration and credibility through the use of
multiple data sources (Eisner, 1991). Quantitative data were collected through teachermade midterms, a state-made EOCT, and a Math Anxiety Questionnaire. Additionally,
qualitative data were collected through student reflections, a teacher focus group, and
administrator interviews. Fairness was achieved by the presentation of opposing views in
the review of literature, as well as the presentation in the discrepancy of student scores in
Spiral Assessments 57
the teacher versus state-made tests. Finally, with an extensive research of the literature
and a careful analysis of the data, this study attained a tight argument, a coherent case,
and rightness of fit.
Implications
Since statistical significance was not found using the Math 2 EOCT scores, the
results of this study cannot be generalized to a larger population. However, the results of
the teacher-made midterms did show statistical significance and were not affected by the
possibility of varying difficulty levels in the state-made test. Being able to control the
consistency of state tests would greatly benefit future research and would hopefully yield
results that could be generalized to a larger population.
While the quantitative data did not produce results that could be generalized to a
larger population, the qualitative data provided great insight into the benefits of spiral
assessments. The qualitative data from students revealed an overwhelming support for
the use of spiral assessments. Students recognized their inability to retain information if
they were not provided with opportunities to practice the content. Student reflection
comments supported the use of daily spiral reviews and unit spiral assessments. In
addition to student comments, teacher comments reflected a similar support of spiral
assessments. Teachers also recognized their students’ difficulty retaining information
and supported the concept of spiraling. The qualitative results of the study revealed an
overwhelming support of the spiral concept and could be generalized to a larger
population.
As mentioned above, this study was transferable and had ‘referential adequacy’
(Eisner, 1991). Content retention is a problem experienced by teachers throughout the
Spiral Assessments 58
county, state, and nation. Teachers from any discipline can apply this study to their
classroom in an attempt to increase content retention. Teachers can easily replicate and
expand this study to include spiral assessments, as well as other activities that are spiraled
in nature.
This study was transformational and contained ‘catalytic validity’ (Lather as cited
by Kinchloe & McLaren, 1998) because of the positive change it brought about in the
students, other math teachers, and me. As a whole, students revealed in their reflections
that spiral assessments helped them remember information they normally would have
forgotten. Students recognized the importance of practicing content over a period of time
and believed this practice would help them perform better on the Math 2 EOCT.
Hopefully, this study encouraged students to apply the saying “practice makes perfect” to
all of their classes.
In the math teacher focus group discussion, teachers expressed an overwhelming
support for spiraling content. Every teacher expressed a concern that their students
memorize information to do well on a test and then quickly lose that knowledge.
Although teachers were concerned about the time involved implementing spiral
assessments, they also appeared very open to finding ways to change. Hopefully, this
study made teachers aware of the potential spiral assessments have to bring about
positive outcomes in their classrooms.
While I believe this study had a positive effect on students and teachers, I believe
it had the greatest impact on me. This study not only led me to recognize the importance
of spiraling my assessments, but also the need to spiral content on a daily basis.
Additionally, I recognized that spiraling content is not only important for success on state
Spiral Assessments 59
tests, but more importantly for preparing students for future math classes. This study
brought about a determination to use spiral assessments throughout the remainder of my
teaching career. This study also brought about a desire to research vertically what
content future math teachers need my students to be experts in. Not only do I plan to use
spiral assessments to prepare my students for state tests, but with this vertical knowledge,
I plan to better prepare students for future success.
Impact on School Improvement
According to Tomal (2003), action research provides a means for educators to
solve problems and make improvements within their classrooms. Hopefully, with
classroom improvements come school-wide improvements. Marzano et al. (2001)
insisted that it only takes a few enthusiastic educators to “infect an entire staff with
enthusiasm” (pp. 157-158).
Following this research study, I personally witnessed two teachers within my
department discussing ways to change their upcoming tests to include questions from
previous units. I also watched as they divided up the tests to start changing over the
summer. With “time” emerging as a major deterrent for implementing spiral assessments
during our focus group discussion, I was pleased to see these teachers finding a way to
overcome this problem.
If nothing else, this study has lead to many discussions regarding classroom
assessments and how assessments can be better utilized by teachers. Not only has the use
of spiral assessments been greatly discussed, but also the importance of formative
assessments and teacher feedback to students. Not only have such discussions occurred
within my department, but also between members of other departments. It is my hope
Spiral Assessments 60
that teachers become increasingly aware of how they are using assessments within their
classrooms and that teachers start looking for ways to embrace the spiral concept.
Recommendations for Future Research
The greatest obstacle this research study contended with was the frequent
changing of the math requirements from the State Department of Education. The 2009
control group was considered the “guinea pig” group because, each year they moved up,
starting from the 6th grade, the new GPS curriculum rolled in and their teachers were
novices to the new standards. Additionally, in the middle of the study, the state
department decided to drop the GHSGT requirement and move to an EOCT requirement.
Therefore, it is uncertain whether the state department changed the 2010 Math 2 EOCT
from the 2009 Math 2 EOCT and whether a change affected test scores from one year to
the next.
As a result of the numerous math changes over the last few years from the state
department, I don’t feel confident in the state test scores that resulted during the study. I
strongly recommend the study be conducted in the future, after things are more settled
and consistent within the state department. Otherwise, the study needs to be conducted
without the use of state tests.
One way to ensure that a change in the state EOCT does not affect test scores
from one year to the next, is to use a control and experimental group from the same
school year. A couple of classes could receive spiral assessments throughout the year,
while a couple of other classes would not. Therefore, all four classes would take the
exact same EOCT and there would be no doubt which students performed better. The
Spiral Assessments 61
problem with this method is, if spiral assessments improve student achievement, then the
control group would be getting the raw end of the deal.
Additionally, I would like to see this action research study conducted on nonaccelerated math students. It became apparent in the MAQ results and the student
reflections that most of my accelerated math students were already confident in their
math ability, and most of my students already liked math. This makes sense because
students taking accelerated math are doing so because they are better at math than their
peers. Therefore, when trying to ascertain whether or not spiral assessments improve
student attitudes and confidence, many of these students did not have much need for this
improvement. I believe surveys and reflections from non-accelerated students would
reveal more information towards focus question two.
Spiral Assessments 62
References
Andrade, H. & Cizek, G. (2010). Handbook of formative assessment. New York, NY:
Routledge Taylor & Francis Group.
Bruner, J. (1960). The process of education. Cambridge, MA: Harvard University Press.
Bruner, J. (1996). The culture of education. Cambridge, MA: Harvard University Press.
Caldwell, S. (2008). It’s branded in our brains. Mathematics Teaching Incorporating
Micromath, 208, 3-5.
Cameron, E. & Green, M. (2009). Making sense of change management: A complete
guide to the models, tools & techniques of organizational change (2nd ed.).
Philadelphia, PA: Kogan Page Ltd.
Charles, C. & Mertler, C. (2002). Introduction to education research (4th ed.).
Boston, MA: Allyn and Bacon.
Churchill, W. (n.d.). BrainyQuote.com. Retrieved July 10, 2010, from BrainyQuote.com
Web site: http://www.brainyquote.com/quotes/quotes/w/winstonchu104164.html
Doll, W. (1993). A post-modern perspective on curriculum. New York, NY: Teachers
College Press.
Dwyer, C. (2008). The future of assessment: Shaping teaching and learning. New
York, NY: Lawrence Erlbaum Associates Taylor & Francis Group.
Eisner, E. (1991). The enlightened eye. New York: MacMillan.
Farooq, M. & Shah, S. (2008). Students’ attitude towards mathematics. Online
Submission, Pakistan Economic and Social Review, 46(1), 75-83.
Fosnot, C. (2005). Constructivism: theory, perspectives, and practice (2nd ed.). New
York: NY: Teachers College Press.
Spiral Assessments 63
Fullan, M. (Ed.) (2005). Fundamental change: International handbook of educational
change. AA Dordrecht, The Netherlands: Springer.
Gabriel, J. (2005). How to thrive as a teacher leader. Alexandria, VA: Association for
Supervision and Curriculum Development.
Garcia, J., Spalding, E., & Powell, R. (2001). Contexts of teaching: Methods for middle
and high school instruction. Upper Saddle River, NJ: Prentice-Hall, Inc.
Harden, R. & Stamper, N. (1999). What is a spiral curriculum? Medical Teacher,
21(2), 141-143.
Jensen, R. (1990). The spring is wound too tight on our spiral curriculum. Arithmetic
Teacher, 38(1), 4-5.
Johnson, B. (2003). The student-centered classroom handbook: A guide to
implementation. Volume one: secondary social studies/history. Larchmont, NY:
Eye On Education, Inc.
Jones, P., Carr, J. & Ataya, R. (2007). A pig don’t get fatter the more you weigh it:
Classroom assessments that work. New York, NY: Teachers College Press.
Kinchloe, J., & McLaren, P. (1998) Rethinking critical theory and qualitative research. In
N. Denzin & Y. Lincoln (Eds.), The landscape of qualitative research: Theories
and issues (pp. 260 – 299). Thousand Oaks, CA: Sage Publications.
LaGrange College Education Department. (2008). Conceptual Framework. LaGrange,
GA: LaGrange College.
Lieberman, A. (Ed.) (2005). The roots of educational change: International handbook
of educational change. AA Dordrecht, The Netherlands: Springer.
Spiral Assessments 64
Marzano, R., Pickering, D., and Pollock, J. (2001). Classroom instruction that
works: Research-based strategies for increasing student achievement. Alexandria,
VA: Association for Supervision and Curriculum Development.
Marzano, R. (2003). What works in schools: Translating research into action.
Alexandria, VA: Association for Supervision and Curriculum Development.
McMillan, J., Cohen, J., Abrams, L., Cauley, K., Pannozzo, G., & Hearn, J. (2010).
Understanding secondary teachers’ formative assessment practices and their
relationship to student motivation. Online submission. Richmond, VA: Virginia
Commonwealth University.
McNiff, J. & Whitehead, J. (2006). All you need to know about action research.
Thousand Oaks, CA: SAGE Publications Ltd.
Meece, J. (1981). Individual differences in the affective reactions of middle and high
school students to mathematics: A social cognitive perspective. Unpublished
doctoral dissertation, University of Michigan.
National Council of Teachers of Mathematics. (2000). Principles and standards for
school mathematics. Reston, VA: The National Council of Teachers of
Mathematics, Inc.
Popham, W.J. (2001). The truth about testing: An educator’s call to action.
Alexandria, VA: Association for Supervision and Curriculum Development.
Popham, W.J. (2009). Instruction that measures up: Successful teaching in the age of
accountability. Alexandria, VA: Association for Supervision and Curriculum
Development.
Spiral Assessments 65
Reeves, D. (2007). Ahead of the curve: The power of assessment to transform teaching
and learning. Bloomington, IN: Solution Tree.
Salkind, N. J. (2010). Statistics for people who (think they) hate statistics (Excel 2nd ed.).
Thousand Oaks, CA: Sage.
Schulman, L. (1998). A problem worth revisiting. Teaching children mathematics,
5(2), 72-77.
Skrla, L., McKenzie, K., & Scheurich, J. (2009). Using equity audits to create equitable
and excellent schools. Thousand Oaks, CA: Sage.
Snider, V. (2004). A comparison of spiral versus strand curriculum. Journal of Direct
Instruction, 4(1), 29-39.
Tomal, D. (2003). Action research for educators. Lanham, MD: Scarecrow Press,
Inc.
Volante, L., Beckett, D., Reid, J., & Drake, S. (2010). Teachers’ views on conducting
formative assessment within contemporary classrooms. Paper presented at the
Annual Meeting of the American Educational Research Association Denver,
CO.
Wigfield, A. & Meece, J. (1988). Math anxiety in elementary and secondary school
students. Journal of educational psychology, 80(2), 210-216.
Zemelman, S., Daniels, H., & Hyde, A. (2005). Best practice: Today’s standards for
teaching and learning in America’s schools (3rd ed.). Portsmouth, NH:
Heinemann.
Spiral Assessments 66
Appendix A
Math Anxiety Questionnaire (MAQ)
1) When the teachers says he/she is going to ask you some questions to find out how
much you know about math, how much do you worry that you will do poorly?
1
Not at all
2
3
4
5
6
7
Very much
2) When the teacher is showing the class how to do a problem, how much do you worry
that other students might understand the problem better than you?
1
Not at all
2
3
4
5
6
7
Very much
3) When I am in math class, I usually feel at ease and relaxed.
1
2
Not at all at ease
3
4
5
6
7
Very much at ease
4) When I am taking math tests, I usually feel nervous and uneasy.
1
2
Not at all nervous
3
4
5
6
7
Very much nervous
5) Taking math tests scares me.
1
Never
2
3
4
5
6
7
Most of the time
4
5
6
7
Most of the time
6) I dread having to do math.
1
Never
2
3
7) It scares me to think that I will be taking advanced high school math.
1
Not at all
2
3
4
5
6
7
Very much
8) In general, how much do you worry about how well you are doing in school?
1
Not at all
2
3
4
5
6
7
Very much
Spiral Assessments 67
9) If you are absent from school and you miss a math assignment, how much do you
worry that you will be behind the other students when you come back to school?
1
Not at all
2
3
4
5
6
7
Very much
10) In general, how much do you worry about how will you are doing in math?
1
Not at all
2
3
4
5
6
7
Very much
11) Compared to other subjects, how much do you worry about how well you are doing
in math?
1
2
Much less than other subjects
3
4
5
6
7
Much more than other subjects
Spiral Assessments 68
Appendix B
Teacher Focus Group Questions
1. How can assessments be utilized within the classroom? What is the purpose of
classroom assessments?
2. When is the best time for assessing knowledge in the classroom?
3. What other forms of evaluation do you use in your classroom? Do you think some
forms are better than others?
4. Do you think spiral testing is a valid assessment of skills?
5. Are spiral assessments a feasible form of assessment?
6. What problems or reservations do you have with spiral assessments?
Spiral Assessments 69
Appendix C
Administrator Interview Questions
1. How should assessments be utilized within the classroom? What is the purpose of
classroom assessments?
2. When is the best time for assessing knowledge in the classroom?
3. What forms of evaluation do most teachers use in their classrooms? Do you think
some forms are better than others?
4. Do you think spiral testing is a valid assessment of skills?
5. Are spiral assessments a feasible form of assessment for our school?
6. What problems or reservations do you have with spiral assessments?
7. Do you think teachers at our school would be open to changing the type of
assessments they administer in their classrooms?
8. What is the best way to educate our teachers about spiral assessments and how is the
best way to bring about change in our school?
Download