THE LEARNING MODEL AT BYU-IDAHO: EVIDENCE FROM

advertisement
THE LEARNING MODEL AT BYU-IDAHO: EVIDENCE FROM
COURSE EVALUATIONS
By
SHAWN V. JENSEN
A dissertation submitted in partial fulfillment of
the requirements for the degree of
DOCTOR OF EDUCATION
WASHINGTON STATE UNIVERSITY
Department of Educational Leadership and
Counseling Psychology
DECEMBER 2010
To the Faculty of Washington State University:
The members of the Committee appointed to examine the dissertation of SHAWN
V. JENSEN find it satisfactory and recommend that it be accepted.
____________________________________
Paul Pitre, Ph.D., Chair
____________________________________
Paul Goldman, Ph.D.
____________________________________
Michael Pavel, Ph.D.
ii
ACKNOWLEDGEMENT
Throughout this four year journey in pursuit of a terminal degree, I have
recognized how blessed I have been to have the support and guidance of kind and loving
individuals. First and foremost, I acknowledge the grace and tender mercies of my
Savior and God. I have felt his direction in my life. I know that he loves me and wants
me to be successful in all virtuous endeavors I embark in. Second, I acknowledge the
blessing it is to have a living prophet. For the words of President Gordon B. Hinckley to
“get as much education as possible” has motivated me to keep going when it seemed too
difficult to continue. Third, my family has been ever constant in their support and
encouragement. Hilary, my sweet wife, never complained about all of the evenings she
spent as a “single mom” while I was away taking classes, or at summer institutes, or
working on my dissertation. She has been the best blessing of my life. I also
acknowledge the sacrifice of my five children as they have given up hours of play time
with their daddy over the past four years to allow me time to finish. Last but not least, I
acknowledge my committee; Dr. Pitre, Dr. Goldman, and Dr. Pavel. I appreciate all of
their direction and guidance throughout this process. Without the support and help of
these kind individuals, this dissertation would not be possible.
iii
THE LEARNING MODEL AT BYU-IDAHO: EVIDENCE FROM
COURSE EVALUATIONS
Abstract
by Shawn V. Jensen, Ed.D.
Washington State University
December 2010
Chair: Paul Pitre
The purpose of this study is to examine the relationship between the Learning
Model (LM) compared with end-of-course ratings and student perceived learning per
existing course evaluation data at Brigham Young University-Idaho. The LM is based on
a broadly accepted pedagogy associated with improving student learning by way of
increasing student engagement. Students are asked to take more responsibility for their
own education and learning by using LM principles and applying related process steps.
The three main process steps of the LM include: Prepare, Teach-One-Another, and
Ponder/Prove. This study represents an attempt to measure the effects of implementation
of the LM on course evaluations. The change in course ratings and student perceived
learning scores were measured. The correlation between LM scores and perceived
learning scores were likewise evaluated. LM scores were derived from the new
evaluation instrument beginning Winter 2009 semester. These scores reflect the level of
student and faculty participation of LM process steps for a given course. A total of eight
semesters were used in this study. A comparison of means analysis was used to compare
iv
the changes between course evaluation score from the Fall 2007 semester to the Winter
2010 semester. Also a correlation matrix was employed to measure the relationship
between LM scores and student perceived learning scores from the time the new
evaluation instrument was implemented (Winter 2009) to the Winter 2010 semester.
Once data were collected, disaggregated, and analyzed as mentioned, further
investigation of breakout variables based on class size, faculty status, and types of
courses were compared to one another. The findings suggest that course ratings
significantly increased from the time the LM was introduced in the Fall 2007 to the
Winter 2010 semester. Interestingly, student perceived learning scores did not seem to
improve significantly over the same time period for the 33 courses identified in this
study. Finally, conclusions and implications were inferred regarding the significance of
applying the three LM process steps on course ratings and student perceived learning.
The implications of these findings can have far reaching consequences as other
institutions consider implementing and/or measuring a campus wide learning model.
v
TABLE OF CONTENTS
Page
ACKNOWLEDGEMENTS .......................................................................................iii
ABSTRACT...............................................................................................................iv
LIST OF TABLES .....................................................................................................ix
LIST OF FIGURES ...................................................................................................x
CHAPTER
1. INTRODUCTION .........................................................................................1
Background ........................................................................................1
Research Questions ............................................................................2
The Learning Model ..........................................................................2
Prepare ..................................................................................5
Teach-One-Another ..............................................................9
Ponder/Prove .........................................................................10
Application .........................................................................................11
The New Evaluation Instrument ........................................................12
2. REVIEW OF LITERATURE ........................................................................18
A Change from the Traditional ..........................................................18
A New Model for Learning................................................................24
Preparing for an Active Learning Environment.....................25
Active Learning Pedagogy .....................................................26
Critical Reflection ..................................................................30
Course Evaluations ............................................................................31
vi
Measuring Effective Instruction ............................................33
3. METHODOLOGY ........................................................................................35
The Two Evaluation Instrument ........................................................37
LM Scores ..............................................................................38
Research Questions and Methods ......................................................41
Research Question 1 ..............................................................42
Research Question 2 ..............................................................46
Research Question 3 ..............................................................49
Limitations .........................................................................................51
4. FINDINGS…. ................................................................................................53
Research Question 1: A Change in End-of-Course Scores ................54
Research Question 2: A Change in Perceived Learning Scores ........65
Research Question 3: The Relationship between LM and
Perceived Learning Scores .................................................................72
The Effects of Class Size .......................................................80
The Effects of Semester Enrollment ......................................81
The Effects of Instructor Employment Status ........................83
Summary of Findings .........................................................................85
5. CONCLUSIONS............................................................................................86
Results of Findings ............................................................................86
The Change in Course Ratings Over Time ............................87
The Change in Perceived Learning Over Time .....................89
vii
The Relationship between LM Scores and Perceived
Learning Scores .....................................................................92
Possible Implications .........................................................................96
Suggestions for Future Research .......................................................98
REFERENCES ..........................................................................................................100
APPENDIX I - INITIAL RESEARCH FINDINGS BY SCOTT BERGSTROM ....109
APPENDIX II - OLD EVALUATION INSTRUMENT ...........................................113
APPENDIX III - NEW EVALUATION INSTRUMENT.........................................116
APPENDIX IV – CORRELATION MATRIX FOR COURSE RATINGS ..............119
APPENDIX V – TWO SAMPLE T-TEST MEASURING SIGNIFICANCE ..........120
APPENDIX VI – CORRELATION MATRIX FOR PERCEIVED LEARNING ....121
APPENDIX VII – PARAMETER ESTIMATES COMPARING PERCEIVED
LEARNING SCORES WITH EMPLOYMENT STATUS .......................................122
APPENDIX VIII – MEAN LM SCORES BY SEMESTER BASED ON
CLASS SIZE .............................................................................................................. 123
APPENDIX IX – LM SCORES COMPARED TO PERCEIVED LEARNING
SCORES BY CLASS SIZE .......................................................................................124
viii
LIST OF TABLES
1. Course Ratings of Qualifying Courses for T1Semesters ...............................56
2. Course Ratings of Qualifying Courses for T2 Semesters ..............................59
3. ANOVA: End-of-Course Ratings for T1& T2 Semesters .............................61
4. Perceived Learning Scores for T1 by Semester .............................................67
5. Perceive Learning Scores for T2 by Semester ...............................................69
6. ANOVA: Student Perceived Learning Scores for T1& T2 Semesters ..........71
7. Mean Perceived Learning Score by Semester Based on Instructor’s
Employment Status ........................................................................................75
8. Mean LM Scores by Semester Based on Instructor’s Employment Status ...77
9. Mean Perceived Learning Scores by Semesters Based on Class Size ...........79
10. Correlation Matrix of LM Scores and Perceived Learning Scores by
Class Size .......................................................................................................81
11. Correlation Matrix of LM Scores and Perceived Learning Scores
by Semester ....................................................................................................82
12. Correlation Matrix of LM Scores and Perceived Learning Scores by
Instructor Employment Status........................................................................84
ix
LIST OF FIGURES
1. The traditional teaching process of higher education ....................................7
2. BYU-I’s LM process......................................................................................8
3. The three process steps of the LM .................................................................8
4. Questions Regarding “Prepare” from both Evaluation Instruments ..............14
5. Questions Related to “Teach-One-Another” from both
Evaluation Instruments ..................................................................................15
6. Question Regarding “Ponder/Prove” from both Evaluation Instrument........16
7. LM Question Set from the New Course Evaluation Instrument ....................39
8. LM Score Totals ............................................................................................41
x
CHAPTER 1: INTRODUCTION
Background
Over the last decade Brigham Young University-Idaho (BYU-I) has seen many
changes. Perhaps the most significant change occurring in June 2000 when the Board of
Trustees announced an expansion of the school from a two year junior college (Rick’s
Junior College) to a four year university (BYU-I) (see General History of BYU-Idaho).
This announcement brought about opportunities to implement, enhance, and introduce
new programs, curriculum, and courses. In August 2001 the school began its first
semester as a four year university.
More recently, another significant event occurred on June 6, 2005 when Dr. Kim
B. Clark, former dean of Harvard Business School, was announced as the 15th President
of BYU-Idaho (Rydalch, 2005). From day one President Clark’s mantra for BYU-I has
been, “rethinking education” (Clark, 2009). In his inaugural address, President Clark
encouraged a more progressive and active learning approach to teaching when he stated:
The challenge before us is to create even more powerful and effective learning
experiences in which students learn by faith…. Students need opportunities to
take action…where prepared students, exercising faith, step out beyond the light
they already possess, to speak, to contribute, and to teach one another. (As cited
by Gilbert, et al, 2007, p. 98)
1
Research Questions
With the implementation of the LM at BYU-I—which denotes the need for
students to take more responsibility for their learning—and with an understanding of the
three process steps of the LM, one is left to ask the following questions:
1. How have end-of-course ratings changed from the Fall 2007 semester (pre-LM
implementation) to the Winter 2010 semester, as the LM was introduced on
campus?
2. How have student perceived learning scores changed based on course evaluation
data collected from Fall 2007 to the Winter 2010 semesters?
3. Is there a relationship between LM scores and student perceived learning based on
the new evaluation instrument (beginning Winter 2009 semester)?
The Learning Model
In the summer of 2007, President Clark introduced to the faculty and students a
new model of teaching and learning that was simply termed, the Learning Model (LM)
(Learning Model, 2008). The LM was introduced as a pedagogical approach to
classroom teaching and learning. This new model reflected President Clark’s philosophy
of active student engagement and by implementation of said model, faculty were invited
and encouraged to rethink and evaluate their instructional pedagogy. Active student
engagement plays an essential role in the effective use of the LM. By creating a more
participatory learning environment students are able to be more active and engaged in
their own learning. Faculty members were encouraged to evaluate their teaching
methodology and evaluate how they could facilitate a more active learning environment
2
in their course(s). The LM was implemented to foster student growth and development,
both academically and spiritually through active student engagement (Learning Model,
2008). As President Clark stated:
BYU-Idaho has embarked on a journey of growth and progress that began with
the decision to transform Ricks College into a four-year university. From that
change, BYU-Idaho has implemented unique academic offerings, creative
calendaring, and programs focused on the development of students. (Clark, 2009,
p. 1)
In this study, the researcher will compare the effects of using the newly
implemented Learning Model (LM) on course ratings, instructor ratings, and student
perceived learning scores as found on course evaluations at BYU-I. This LM consists of
three process steps which students are encourage to incorporate and apply as a key
component to learning, and are explained in further detail below. The manner in which
the LM has affected course ratings, instructor ratings, and student perceived learning
scores over a two and ½ year period as found on end-of-course evaluations is the focus of
this study.
It is important to point out however, that learning models are typically a
compilation of proven practices in teaching and learning (Thurgood & Gilbert, 2008).
Likewise, BYU-I’s LM is not proprietary to BYU-I. In general, educational theorists and
cognitive psychologists have endorsed a more active learning environment for students
for some time now (Thurgood & Gilbert, 2008). Also, active learning methodology has
been a focus of study for many researchers throughout the recent years. For example,
methods for engaging students that have been studied include peer instruction (Mazur,
3
1997), cooperative learning (Slavin, 1995), and mutual peer tutoring (King, et al, 1998),
just to name a few. BYU-I’s LM also endorses a more active learning environment for
students, but the implementation of this LM is unique in that instructors and students
campus wide are encouraged to apply the three process steps of this model in their
learning and teaching, respectively. The level at which students perceive the LM being
implemented into their respective classes may be reflected in end-of-course evaluation
scores.
Obviously, the LM is not structured to be a prescriptive “fix all” to passive
learning. The ways in which an instructor designs his/her course(s) to facilitate student
activity and learning, as it relates to the LM, can vary considerably (Gilbert, 2008). In
fact, instructors are encouraged to incorporate LM strategies that best fit the subject they
teach (Gilbert, 2008). For example, a physics course may have different options to
choose from, in terms of motivating students to actively participate, then does an English
course or computer science course. In each case, the instructor has the latitude to
consider which instructional tool will be most effective in actively conveying the
information to the students. Thus, instructors have a wide range of autonomy in choosing
which teaching technique to incorporate that best fits their teaching style as it relates to
the three process steps of the LM.
As the LM was introduced, both instructor and student were encouraged to
become familiar with the principles and process steps related to the model (Clark, 2007).
New freshman entering their first semester at BYU-I are introduced to the processes and
principles outlined in the LM in almost all of their required freshman level courses
(Learning Model, 2008). The University’s website contains information regarding this
4
model which can easily be accessed. The following was taken from the above mentioned
website and gives a brief definition of the LM:
The Learning Model was created to deepen the learning experiences of students at
BYU-Idaho. The Learning Model enables students to take greater responsibility
for their own learning and for teaching one another. The model is structured
around three process steps. (Learning Model, 2008)
The three process steps of the LM are designed to help students become more
responsible for their own learning, as mentioned above, and include: Prepare, Teaching
One Another, and Ponder/Prove. In other words, the LM endorses student preparation
before attending class (i.e., reading the assigned chapters or literature) to deepen class
discussion, during class time through actively participating in discussions, and after class
by reflecting on the material learned to help extend their understanding of the information
being presented in class.
Prepare. Beichner et al. (2007) states that in order to create an active learning
environment in the classroom the foremost element is student preparation. In this study,
the LM used at BYU-I included a preparation step. In fact, preparation is the first process
step in the LM and contains three parts (Learning Model, 2007). There is first an
ongoing, underlying spiritual preparation for each student. That spiritual preparation is
critical to learning by faith and being guided and taught by the Holy Spirit. It requires
daily attention to prayer and scripture study, and a commitment to obedience (Learning
Model, 2007).
The second part of preparation is individual. Here the student prepares to learn
and to teach by organizing required readings or assignments; by reading and studying and
5
writing or solving problems; and by thinking about questions the material raises, and how
to answer those questions (Learning Model, 2007).
The third part of preparation involves group activities. For example, the process
assumes that students have been organized into small learning teams or study groups (5-7
members). These groups “meet” for a short period (e.g., 30 minutes) on a regular basis
(before class—daily or multiple times a week) either online or face-to-face to discuss
assignments, to try out ideas, and to test understanding. The groups may also be required
to complete a group assignment or project. With this kind of preparation, students come
to class ready to learn and to teach (Learning Model, 2007).
The Prepare step is not solely the responsibility of the student. The instructor
must organize, assign, and hold students accountable for being prepared. Information
regarding the Prepare process step of the LM can be found on the BYU-I’s teaching and
learning website and reads as follows:
Student preparation work is designed, guided and aided by the instructor, but the
impetus for actually doing the work is on the student. Students rely on the help of
the instructor to show them how to successfully prepare. The instructor defines
the questions framing the assignment, provides suport materials like worksheets,
reading questions or the like, and the instructor defines both the way in which
students are expected to engage the new material and how it will be assessed.
(BYU-I, 2009)
The LM places more responsibility on the student to prepare themselves before
class (Learning Model, 2007). By doing so, the student can absorb the new material
based on their preferred learning style and at their own pace. Bonwell & Eison (1991)
6
suggest that “cognitive research has shown that a significant number of individuals have
learning styles best served by pedagogical techniques other than lecturing” (p. 46). In
addition, “several studies have shown that students prefer strategies promoting active
learning to traditional lectures” (Bonwell & Eison, 1991, p. 45). Gilbert (2008) defines
the traditional teaching process as following this sequence; students attend a class lecture,
followed by homework assignment to review the material covered in class, followed by
another lecture (see fig. 1).
Lecture
Homework
Lecture
Homework
Lecture
Homework
Figure 1. The traditional teaching process of higher education as defined by Gilbert et al
(2008).
The BYU-Idaho LM (2009) suggests that students prepare before class so they
can better benefit from and contribute to in-class learning and teaching experiences (see
fig. 2). This model places more responsibility on the student to come prepared to
participate in classroom discussion. In fact, the preparation element is critical to
classroom participation.
7
Preparation
Class
Preparation
Class
Preparation
Class
Figure 2. BYU-I’s LM process.
This approach places a responsibility for students to prepare before class to allow
for richer conversations and higher level comprehension (Gilbert et al, 2008). With this
approach “homework” becomes preparation work for future class discussion. More time
can then be used during class to facilitate a more active and engaged classroom
environment. Using this method of instruction, students study before class to prepare in
order to better participate during class. Requiring more of students early in the process
allows the instructor to spend more time deepening understanding rather than dispensing
information (BYU-I, 2009). Prepare is a key element of the LM. When students fail to
prepare for class the level of active participation during class is hampered. Each process
step of the LM is connected and leads to the next (see figure 3).
Prepare
Teach One
Another
Figure 3. The three process steps of the LM.
8
Ponder/
Prove
Student preparation cannot be overlooked nor neglected for the LM to be
effective (Gilbert, 2008). As shown from the figure above, each of the three process
steps of the LM lead to the next step with the third process step, ponder/prove,
connecting back to prepare.
Teach One Another. Preparing before class to actively contribute during class
time naturally lends itself to a more participatory learning environment, given that the
instructor encourages peer-to-peer discussion (Gilbert et al, 2008). The second process
step of BYU-I’s LM is called “Teach One Another”. This step includes group projects,
peer evaluations, teams, presentations, and any other means to facilitate a more active
learning environment. These techniques and teaching strategies are, in essence, what
makes up the Teaching One Another process step of the LM. Student are encouraged to
come to class prepared to participate and contribute to classroom learning via any of the
above mentioned techniques and strategies. BYU-Idaho Boardmember, Richard G. Scott
(2005) stated the importance of having an engaged and active learning environment in the
classroom when he stated:
Never, and I mean never, give a lecture where there is no student participation. A
“talking head” is the weakest form of classroom instruction…. Assure that there
is abundant participation because that use of agency by a student authorizes the
Holy Spirit to instruct. It also helps the student retain your message. As students
verbalize truths, they are confirmed in their souls and strengthened in their
personal testimonies. (Scott, 2005)
9
In addition, President Clark has taught that “overtime it will become apparent that
the most powerful way for students to learn is for them to teach… they will teach to
learn” (BYU-I, 2009).
Ponder/Prove. The third process step of the LM includes two elements;
pondering and then proving what is being taught. The “ponder” element includes taking
time to reflect on what is being learned to help solidify and relate the new concepts,
ideas, or perspectives being taught. This can be accomplished in a variety of ways.
Reflecting individually through journals, notes, blogs, and/or by simply telling another
person what you have learned can help solidify new learning. Group discussion and
reflection can also play a role in helping individuals ponder ideas and concepts being
taught (Learning Model, 2009). Research has demonstrated, for example, that if an
instructor allows students to ponder what is being taught and allow students time to
consolidate their notes by pausing three times for two minutes each during an hour long
lecture, they will learn and retain significantly more information (Ruhl, Hughes, and
Schloss 1987).
“Prove” includes, but is not limited to, the assessment and evaluation element of a
course (Learning Model, 2009). The Ponder/Prove process of the LM might include
tests, exams, and even thought provoking questions, where students must apply what they
have learned in order to solve a problem or in answering a question. These process steps
are grounded in current literature surrounding active learning and student engagement.
Research has consistently shown that traditional lecture methods, where professors
lecture and students listen and write, are less effective compared with active learning
10
methodology (Prince & Felder, 2006; Prince, 2004; Bonwell & Eison, 1991). Students
must be given an opportunity to actively prove their learning from time to time.
Active learning and student-centered instruction encourages students to be
actively engaged in reading, writing, discussions, or problem solving activities (Prince &
Felder, 2006; Draper & Brown, 2004; Chickering & Gamson, 1987). These types of
learning experiences must engage students in higher order thinking tasks, such as
analysis, synthesis, and evaluation. The LM was introduced as a means of encouraging
more participation in these proven teaching methods and techniques.
Application
Effectively applying the three process steps of the LM by both students and
instructors may require a change in either the instructor’s teaching methodology or the
students’ perception of what is expected of them, and sometimes both changes are
requisite (Felder & Brent, 1996). As instructors actively apply these process steps of the
LM in their courses some students might struggle with the added responsibilities placed
on them to be active participants in the classroom. Woods (1994) observes that students
forced to take major responsibility for their own learning go through some or all of the
steps psychologists associate with trauma—from shock and denial to resistance and then
finally integration. Students may not be used to taking more responsibility for their own
learning and students may push back when an instructor takes this approach. This type of
instruction may impose steep learning curves on both instructors and students, and initial
instructor awkwardness and student hostility are both common and natural (Felder &
Brent, 1996). Students who are accustom to the traditional method of teaching—which
11
includes note taking during a lecture followed by some form of independent
homework—can have a difficult time initially adjusting to a more participatory or active
model for learning. Felder and Brent (1996) gave the following example:
When confronted with a need to take more responsibility for their own learning,
[students] may grouse that they are paying tuition-or their parents are paying
taxes-to be taught, not to teach themselves…They may gripe loudly and bitterly
about other team members not pulling their weight or about having to waste time
explaining everything to slower teammates (p. 7).
As instructors begin to hear this type of resistance some may feel awkward and
want to return to the more familiar lecture methodology. What’s more, Felder and Brent
(1994) have found that an instructor who uses a more participative and active teaching
methodology (i.e., as in the three process steps of BYU-I’s LM) can initially experience a
drop in their course-end ratings.
The New Evaluation Instrument
BYU-I has recently adapted their end-of-course evaluation instrument (Winter
Semester, 2009) to reflect the three process steps of the LM, which emphasizes the need
for students to take more of a responsibility for their own learning. The means by which
to track how effective students are in taking more responsibility for their learning is
through questions asked on the end-of-course evaluations regarding LM process steps.
This new evaluation instrument has questions built-in that directly relate to the level at
which LM process steps are being implemented, based on students’ perceptions
12
(Bergstrom, 2009). Students are asked to rate how effectively they and their instructor(s)
used the three process steps of the LM throughout the course (see Appendix III).
The changes between the two end-of-course evaluation instruments are stark. For
example, in the old evaluation instrument questions were general to the course and/or to
the instructor’s teaching abilities. However, in the new evaluation instrument students
are asked questions related to their own preparation, participation, and personal reflection
in addition to how well they perceived the LM was implemented by their instructor and
peers. Using a side-by-side comparison of the two end-of-course evaluation instruments,
one can detect the difference in the question sets as they relate to students taking
responsibility in each of the three process step of the LM. The change in end-of-course
evaluation instruments is important to keep in mind throughout this study as the
methodology is affected as a result of the differences between the two evaluation
instruments. Therefore, questions that address a similar concept (i.e., the Prepare process
step of the LM) are compared and contrasted using the following tables. The first sideby-side comparison shows the difference in questions students are asked to answer as
they relate to the Prepare process step of the LM (see Figure 4).
13
Old Evaluation Instrument
Assigned homework is not just busywork.
New Evaluation Instrument
I was prepared for each class.
Text(s) and other materials have helped
The instructor held me accountable for
me understand course content.
coming to each class prepared.
The approximate number of hours per
week that I have spent in outside
preparation for this class is…
The approximate number of hours per
week that I have spent in outside
preparation for this class is…
I arrived at class on time
Figure 4. Questions Regarding “Prepare” from both Evaluation Instruments
Notice the two questions directly related to student accountability found in the
new evaluation instrument, “I was prepared for each class” and “I arrived at class on
time”. Essentially, this means that course and instructor end-of-course evaluation scores
are being influenced by student participation in the LM.
A significant change is noticed in the amount of questions asked regarding the
second process step of the LM—the Teach One Another step—between the two end-ofcourse evaluation instruments in question. The Teach One Another process step was
heavily emphasized on the new evaluation instrument, as shown in the side-by-side
comparison in Figure 5. One will quickly notice that no questions were found in the old
evaluation instrument that relate to the Teach One Another process step of the LM while
the new evaluation instrument contains six related questions. This change in course
evaluations highlights the added focus and attention administrators at BYU-Idaho are
14
giving to the LM, particularly the Teach One Another process step. These added
questions focused on peer-to-peer teaching may be used as a motivational tool to
encourage student and faculty to participate in developing and maintaining a more
participatory learning environment.
Old Evaluation Instrument
New Evaluation Instrument
I was an active participant in online or
face-to-face class discussion.
I sought opportunities to share my learning
with others outside of class.
I feel that I made important contributions
to the learning and growth of fellow
classmates.
The course provided opportunities to learn
No related questions were found.
from and teach other students.
Group work, if assigned, was beneficial
and meaningful.
Students were actively involved in this
class through discussions, group work, &
teaching.
The instructor provided appropriate
opportunities to participate in the class.
Figure 5. Questions Related to “Teach-One-Another” from both Evaluation Instruments
15
An increase in questions asked regarding the third process step of the LM—the
Ponder/Prove step—was also noticed when comparing the new evaluation instrument
with the old. The evaluation instrument that was in place previous to the implementation
of the LM contained one question related to pondering and/or proving new learning,
whereas the new evaluation instrument contained three. Likewise, the new evaluation
instrument contains questions that emphasis student taking more responsibility for their
own learning as it relates to the third process step of the LM (see Figure 6).
Old Evaluation Instrument
New Evaluation Instrument
The instructor provided opportunities to
reflect upon my learning and experiences
Exams are good measures of my
in the class.
knowledge, understanding, or ability to
I sought opportunities to reflect on what I
perform.
had learned in the class.
The course as a whole has produced new
knowledge, skills, and awareness in me.
Figure 6. Question Regarding “Ponder/Prove” from both Evaluation Instrument
This comparison between the old evaluation instruments with the new evaluation
instrument shows the evolution of the question sets and how the new evaluation
instrument has a strong LM emphasis (see Appendix II). These differences may
contribute to a change in an instructor’s course rating as students are asked to consider
additional or new elements when rating their instructor(s). For example, an instructor
16
might have had a class rating of 5.0 (out of 7 possible) based on the old evaluation
instrument, but on the new evaluation instrument the same instructor’s course rating
could now be positively or negatively affected by student perceptions of how effectively
they and their instructor applied the LM process steps throughout the course. This added
emphasis on students taking more of a role for their own learning and being held
accountable for such, surely may have an effect on end-of-course evaluations, but to what
extend? This is the purpose of this study, to explore the implications of introducing a
campus wide learning model on course ratings and student perceived learning as found
on end-of-course evaluations.
17
CHAPTER 2: REVIEW OF LITERATURE
The focus of this study is based on the implementation of the three process steps
inherent in the LM and the effects of said implementation on end-of-course student
evaluations. The researcher has chosen to focus on literature that outlines the rationale
behind the need for a change in methodology. Additionally, literature that relates to the
three process steps of the LM—Prepare, Teach One Another, and Ponder/Prove have
been examined. Research collection and analysis of data using course evaluations are
also a significant aspect to this study as the finding are based on students’ assessments of
courses they have taken. Therefore, the subsequent review of literature is divided into
three main sections; A Change from the Traditional, A New Model for Learning, and
Course Evaluations.
The first section, A Change from the Traditional, includes a review of literature
that has been published regarding the traditional teaching methodology in higher
education (specifically lectures) and reviews relevant studies regarding the significance
of a more active learning approach. The second section, A New Model for Learning,
consists of a compilation of studies related to the three process steps embedded in the
LM; i.e., prepare, teach-one-another, and ponder/prove. The third section, Course
Evaluation, considers the viability of using this type of method to evaluate effective
teaching and reviews literature that supports its use.
A Change from the Traditional
Traditionally, higher education courses have been taught using some form of a
lecturing method, mostly to accommodate for larger audiences. Recent studies have
18
shown that lecturing is not always the best methodology for disseminating information
from instructor to the students (Goyak, 2009; Johnson et al, 1998; Felder & Brent 1996).
One author, Woods (1994), describes lecture based instruction as a waste of everyone’s
time. He states that it is neither teaching nor learning and suggests that it is more a
practice in stenography. “Instructors recite their course notes and transcribe them onto the
board, the students do their best to transcribe as much as they can into their notebooks,
and the information flowing from one set of notes to the other does not pass through
anyone's brain” (Felder & Brent, 1996, p. 44). Instructors have the best intentions of
helping their students understand the material they are presenting but many times the
learning is minimal.
One problem with this form of teaching is that the instructors are largely
responsible for the dissemination and explanation of course content instead of placing
more responsibility to the students for their own learning (Crouch & Mazur, 2001;
Bonwell & Eisen, 1991). Most often instructors teach based on how they were taught.
However, if instructors were to shift responsibility of learning more to students then
learning would be transposed based on the students’ preferred learning methodology.
American educational philosophers Dewey (1916), Piaget (1928), and Vygotsky (1978)
each argues that learning is more powerful when learners are actively engaged in the
knowledge creation process. “Many proponents of active learning suggest that the
effectiveness of this approach has to do with student attention span during lecture”
(Prince, 2004). Wankat (2002) cites numerous studies that suggest “student attention
span during lecture is roughly fifteen minutes.” Earlier, Hartley and Davies (1978) found
that after fifteen minutes of lecturing the number of students paying attention drops
19
quickly and their ability to retain the information likewise decreases with each passing
minute. They further found that immediately following a lecture, students retained 70
percent of information given during the first ten minutes of the lecture and only 20
percent of information was retained from the last ten minutes of the lecture. These
finding suggest that breaking up the lecture may be beneficial because students’ minds
start to wander and activities provide the opportunity to start fresh again, keeping
students engaged (Prince, 2004).
In the mid-1980s, Halloun and Hestenes (1985) conducted a series of studies
measuring knowledge gains of conceptual material for physics students. They
demonstrated that physics students performed poorly on pre- and post-tests when even
though their grades suggested otherwise. Students could memorize the formulas, but did
not seem to understand the underlying concepts (Halloun & Hestenes, 1985). This led to
several different studies experimenting with peer instructional techniques including
concept tests designed by Eric Mazur (2001) and others. Internal departmental studies
began to show dramatic improvements in conceptual understanding (Mazur, 2001). Large
sample studies were next to be studied, of which the Hake (1998) study was the most
significant.
There has been continued growth in regards to the empirical literature that
encourages and promotes a more active learning environment (Gilbert et al, 2007). For
example, Karabenick and Collins-Eaglin (1997) conducted a study of more than 1,000
students across nearly 60 different classes and found that students in courses that
emphasized collaborative learning were more likely to develop higher-order learning
skills, such as comprehension and critical thinking. Another study by Laws et al (1999)
20
shows how a more active learning environment can greatly enhance student
understanding of basic physics concepts compared with traditional teaching techniques.
In addition, Alison King and her colleagues (1998) found that students who used
structured questioning techniques to teach and learn from their peers had a better
understanding of the information than did students who were taught by traditional
methods. Richard Hake (1998) studied more than 6,000 students in 62 introductory
physics courses and found that “students who used peer instruction techniques showed
almost two standard deviations greater average normalized gain on pre- and post-tests
than students who were taught through traditional methods” (as cited by Gilbert et al,
2007, p. 102). This is a significant finding for those advocates of active learning, in that
it shows evidence of student learning when compared with traditional lecturing
techniques.
The use of these techniques in the classroom is vital because of their powerful
impact upon students' learning. Research has demonstrated, for example, that if a faculty
member allows students to consolidate their notes by pausing three times for two minutes
each during a lecture, students will learn significantly more information (Kellogg, 2008;
Boyle, 2001; Ruhl, Hughes, and Schloss 1987). The reason why students learn more
information is due to the fact that they are given time to actively process the new
information and compartmentalize new ideas more effectively as they are being taught
(Kellogg, 2008). Other research studies evaluating students' achievement have
demonstrated that many strategies promoting active learning are comparable to lectures
in promoting the mastery of content but superior to lectures in promoting the
development of students' skills in thinking and writing (Smith et al., 2005; Prince, 2004;
21
Bonwell & Eison, 1991). However, students have a variety of learning styles and no
instructional approach can be optimal for everyone (Grasha, 1994; Felder, 1993; Claxton
& Murrell, 1987). Therefore, the importance to differentiate instruction becomes that
much more important as it relates to student learning.
Felder and Brent (1996) discuss some of the broad approaches to differentiating
instruction that could lead to more student engagement:
•
Substituting active learning experiences for lectures.
•
Holding students responsible for material that has not been explicitly
discussed in class.
•
Assigning open-ended problems and problems requiring critical or
creative thinking that cannot be solved by following text examples.
•
Involving students in simulations and role-plays.
•
Assigning a variety of unconventional writing exercises.
•
Using self-paced and/or cooperative (team-based) learning.
Other active learning pedagogies include cooperative learning, debates, drama,
role playing and simulation, and peer-to-peer teaching. In short, the published literature
on alternatives to traditional classroom presentations provides many different approaches
faculty could use as instructional tools (Bonwell & Eison, 1991).
Student engagement and active participation has shown to be helpful in promoting
long-term retention of information, motivating students toward further learning, allowing
students to apply information in new settings, and in developing students' thinking skills
(McKeachie et al., 1986). Research has suggested, however, that to achieve these goals
faculty must be knowledgeable of alternative techniques and strategies for questioning
22
and discussion (Hyman, 1980) and must create a supportive intellectual and emotional
environment that encourages students to take risks (Lowman, 1984). In other words,
faculty must be willing to try new methodology to engage students.
So why doesn’t every higher educational instructor strive to improve student
engagement in terms that were previously outlined? Bonwell and Eison (1991) speak of
some common barriers to instructional change. Included in their list are the powerful
influence of educational tradition, faculty self-perceptions and self-definition of roles, the
discomfort and anxiety that change creates, and the limited incentives for faculty to
change. However, there are certain specific obstacles that are associated with the use of
active learning. These include limited class time, a possible increase in preparation time,
the potential difficulty of using active learning in large classes, and a lack of needed
materials, equipment, or resources (Bonwell & Eison, 1991).
Perhaps the most significant barrier for instructors not being willing to modify
their instruction to fit a more participatory learning environment involves risk. They may
feel the risk that students will not respond to the active learning requirements that
demand higher-order thinking and open participation of students. In other words,
instructors who are trying to change their teaching methodology to a more active learning
environment take a risk that students will resist the change and not respond to the added
responsibilities associated with active learning. Some instructors may feel a loss of
control as they implement active learning strategies like peer-to-peer instruction. What’s
more, some instructors may even feel a lack of necessary skills required in fostering an
active learning environment. However, each of these barriers can be overcome with
23
careful planning, direction, and preparation (Bonwell & Eison, 1991). The LM at BYU-I
is designed to help faculty and students overcome some of these barriers (Gilbert, 2008).
A New Model for Learning
This section of the literature review is focused on the relevant literature
surrounding the process steps of the BYU-I’s LM—Prepare, Teach-One-Another, and
Ponder/Prove. For the most part, the LM is based on student engagement and active
participation. Facilitating an active learning environment is possibly the most effective
way to increase student engagement (Kuh, 2001). All three process steps focus on ways
of improving activity and engagement of students to increase their learning. The
literature review sections that relates to the LM are termed; Preparing for an Active
Learning Environment, Active Learning Pedagogy, and Critical Reflection.
Empirical literature is building in each area of the LM. One example of this
would be active learning, which has been studied by many authors, and is the basis of the
Teach One Another process step of the LM (Borman et al., 1993; Brockway, Mcneely
and Niffennegger, 1992; Dev, 1990; Hunt et al., 1985; Mallor, Near, & Sorcinelli, 1981).
The natural question that follows would be how one could implement this style of
teaching into their regular classroom instruction. Penner (1984) gives an answer to this
question when he suggests that the modifications of traditional lectures are ways to
incorporate this form of active learning in the classroom. The following literature will
outline the most relevant methodology that supports modification of lectures to make
room for active student engagement.
24
Preparing for an active learning environment. The LM’s first process step is
“Prepare”. This step is foundational to the integrity and efficacy of the LM. Beichner et
al. (2007) agree that student preparation is foremost in facilitating a cooperative learning
environment where students freely collaborate, question, and teach one another. Most
Universities preach the unwritten law of “two for one” as their academic mantra, meaning
“undergraduate students should spend at least two hours preparing for every class
hour…Unfortunately, most students spend only about half that amount of time” (Kuh,
2003, p. 27). Additionally, only a fifth of the first-year students through seniors
“frequently” come to class with any preparation. Meaning, that only 20% of the student
body come to class prepared to ask questions, discuss meanings, and/or contribute to the
learning environment by sharing experiences, stories, ideas, and thoughts that all have to
potential of increasing the level of learning for others in the classroom. What’s more,
these students say “their institutions give little emphasis to studying and spending time on
academic” preparation before class (Kuh, 2003, p. 27).
One pedagogical approach used to encourage student preparation is by inviting
students to begin learning about a topic before it is presented in class by reading
information that pertain to the topic. “Students are asked to focus on the key ideas and
express them in their own words before completing an assignment based on the reading.”
(Beichner et al, 2007, p. 5). Another methodology used to encourage pre-class
preparation is the use of online reading quizzes. These quizzes are designed to encourage
students to read the material before it is covered in class. In this way, students come to
class already prepared to participate in discussions and other activities (Beichner et al,
2007).
25
Active learning pedagogy. The second process step in The LM is termed TeachOne-Another. This type of teaching methodology is not new. In fact, “first-century
Roman rhetorician, stated that students could benefit by teaching one another” (Goyak,
2009, p. 31). Likewise, Johann Amos Comenius, a prolific reformer of pedagogy of the
early 1600s, believed that peer-to-peer teaching was beneficial to student learning
(Johnson et al., 1998). The concept of teaching one another is not new but it has taken
many different names and forms throughout the years. Forms of teaching one another
have been termed as active learning, peer-to-peer instruction, collaborative learning,
cooperative learning, inductive learning, and student-centered learning (Felder & Brent,
2006).
Active learning incorporates activities that engage students in doing something
besides listening to a lecture and taking notes to help them learn and apply course
material. Students may be involved in talking and listening to one another, or writing,
reading, and reflecting individually (Felder & Brent, 2006). “Active learning strategies
reach all types of learners in the visual, auditory, read/write, kinesthetic, and tactile
schemes. With active learning strategies, visual learners are targeted by the presence of
models and demonstrations” (Lujan & DiCarlo, 2006, p. 18). Auditory learners are
reached through discussion during peer instruction (Cortright et al, 2005; Rao & DiCarlo,
2000), debates (Scannapieco, 1997), games (Bailey et al, 1999; Mierson, 1999; Moy et al,
2000; Patil et al, 1993), and answering questions. Manipulating models (Rodenbaugh et
al, 1999; Silverthorn, 1999; Chan et al, 1991) and role playing (Kuipers & Clemens,
1998) satisfy kinesthetic and tactile learners. It is generally thought that students have
better retention and understanding of knowledge when taught by active as opposed to
26
passive methods (Matlin, 2003; Rao & DiCarlo, 2001; McKeachie, 1994; McDermott,
1993; Michael, 1993).
Lymna's (1981) Think-Pair-Share and Mazur's (1997) Peer Instruction are “peer
instruction activities that provide opportunities for students to be actively engaged in the
reasoning and application of concepts” (as cited by Lujan & DiCarlo, 2006, p. 20).
Studies have “documented that pausing two to three times during a 50-min class to allow
peer instruction of concepts enhanced the students' level of understanding and ability to
synthesize and integrate material” (Lujan & DiCarlo, 2006, p. 20). Specifically, peer
instruction enhanced the mastery of original material and meaningful learning (Lujan &
DiCarlo, 2006).
Collaborative learning is a subset of active learning in which students interact
with one another while they learn and apply course material (Felder & Brent, 2006).
Collaborative testing is an effective cooperative learning teaching strategy (Cortright et
al, 2003; Rao et al, 2002). To conduct a collaborative testing experience, students first
complete a quiz individually. Once the quiz is completed individually, students complete
the same quiz in groups. Specifically, students complete the quiz individually for the first
50 minutes immediately after completing the quiz individually, students are assigned to a
group of two or three and work as a team to answer the original questions; 80% of the
final quiz score is based on individual results, and 20% of the final quiz score is based on
group results. This collaborative group testing has been shown to increase student
learning (Rao et al, 2002) as well as student retention of previously learned material
(Cortright et al, 2003). For this study, the second process step of the LM ties in naturally
to collaborative groups and other cooperative learning techniques.
27
Cooperative learning is a form of collaborative learning in which students work
together on structured assignments or projects under conditions that assure positive
interdependence, individual accountability, periodic face-to-face interaction, appropriate
development and use of interpersonal skills, and regular self-assessment of group
functioning (Felder & Brent, 2006). When students learn with others, they have the
emotional and intellectual support that allows them to go beyond their present knowledge
and skills and accomplish shared goals (Silberman, 1996). In the cooperative learning
setting, students become more engaged in learning by discussing material, facilitating
understanding and encouraging hard work (Lujan & DiCarlo, 2006). In addition,
cooperative learning has positive effects on race relations, self-esteem, and a willingness
to cooperate in other settings (Slavin, 1983).
Sapon-Shevin and Schniedewind (1992) promoted the concept and purpose of
cooperative learning in the following statement:
Cooperative learning [is] more than a teaching strategy, more than an instructional
technique. Cooperative learning is an entirely different way of viewing the
educational process of schools, reshaping them into communities of caring in
which individual students take responsibility for the learning of their classmates
and respect and encourage each other’s diversity. Cooperative learning has the
potential to completely transform all aspects of your classroom and of your school
so as to promote the sharing of power, responsibility, and decision-making
throughout. (p. 16)
Cooperative learning argues that it is not enough merely to work with others or
just to work in groups within the class. Cooperative learning demands that all students
28
within a class actively and meaningfully engage in a discovery process of learning
(Johnson & Johnson, 1999).
Goyak (2009) suggests that the reward structure of a classroom can also be very
different from a traditional setting when viewed in light of cooperative learning
principles. He further states that an interpersonal reward structure highlights rewards by
how they are linked to students. “In a negative reward interdependence system,
competition in the classroom is emphasized. Competition in this framework refers to one
student’s success necessitating another student’s failure. Conversely, in a positive reward
interdependence system, cooperation within groups is emphasized” (Goyak, 2009, p. 9).
Goyak (2009) explains that one student’s success allows all group members to experience
success in the classroom.
Inductive learning has been referred to as problem-based learning, project-based
learning, guided inquiry, or discovery learning (Felder & Brent, 2006). These new
approaches to instruction present challenges to both teachers and students (Marx et al,
2004). For teachers using instructional methods based on recitation and direct instruction,
inductive teaching strategies challenges them to develop new content knowledge,
pedagogical techniques, approaches to assessment, and classroom management
(Blumenfeld, Marx, & Soloway, 1994; Edelson, Gordin, & Pea, 1999; Marx, Blumenfeld,
Krajcik, & Soloway, 1997)
Felder and Brent (1996) studied student-centered instruction (SCI) and found that
SCI differs from traditional instruction. In traditional instruction, the teacher's primary
functions are lecturing, designing assignments and tests, and grading; in SCI, the teacher
still has these functions but also provides students with opportunities to learn
29
independently and from one another and coaches them in the skills they need to do so
effectively. More of a focus is centered on peer-to-peer instruction and student’s
participation. In recent decades, the education literature has described a wide variety of
student-centered instructional methods and offered countless demonstrations that
properly implemented SCI leads to increased motivation to learn, greater retention of
knowledge, deeper understanding, and more positive attitudes toward the subject being
taught (Bonwell and Eisen 1991; Johnson Johnson and Smith 1991a,b; McKeachie 1986;
Meyers and Jones 1993).
Critical reflection. Critical reflection is a term used in academia that relates to
the third process step of the LM, the ponder/prove process step. Critical reflection,
sometimes grouped with transformative learning, involves the critical examination and
self-reflection of held personal beliefs to assess their validity, bias, and limitations in a
given context (Wolf, 2009). Mezirow (1990) proposes that critical “reflection can take
place within both instrumental and communicative learning realms” (p. 7).
Ponder/Prove, occurs after class and may involve individual or group activities. In
this step students ponder what they have learned in class. They keep a learning journal
and write down impressions and insights they have received. Students pursue unanswered
questions and discuss the class with their study group. This process step naturally
transitions back to the initial process step of preparing before class (BYU-I, 2007).
The process step is not only functional and helpful as a way of increasing student
learning, it is also a good practice for instructors. Further, each faculty member should
engage in self-reflection, exploring his or her personal willingness to experiment with
alternative approaches to instruction (Bonwell & Eison, 1991).
30
Course Evaluations
Course evaluations have been used for many years in higher education to measure
student perceptions regarding the course design and the instructors’ strengths and
weaknesses. Studies have also been conducted on the implementation, usefulness,
design, and application of course evaluations for a number of years. In the mid-eighties,
Marsh (1987) conducted a thorough review of the massive literature on the use of student
evaluations and their relation to teaching effectiveness. He identified nine dimensions
critical to effective instruction at the university level; workload, teachers' explanations,
empathy (interest in students), openness, and the quality of assessment procedures
(including quality of feedback) (March, 1987). He also found that lecturers rated
themselves higher than students did regarding good teaching at the university level.
A decade earlier, Feldman's (1976) developed a rubric of 19 categories of
instructional effectiveness that included; stimulation of student interest, teacher
sensitivity to class level and progress, clarity of course requirements, understandable
explanations, respect for students, and encouraging independent thought.
Another studies based on evaluation of course improvement rather than individual
instructor ratings have emerged. Entwistle & Tait (1990), for example, found that those
factors that best improve a course include; the provision of clear goals, appropriate
workload and level of difficulty, assignments providing choice, quality of explanations,
level of material and the pace at which it is presented, enthusiasm, and empathy with
students' needs.
Literature on student evaluation such as those of Marsh (1987), Feldman (1978),
and Roe & Macdonald (1983), together with the authoritative summaries of key findings
31
of validity studies (e.g. McKeachie, 1983; Centra, 1980) attest to the usefulness and
accuracy of student evaluation of instruction in comparison with other measures such as
peer evaluations (Ramsden, 1991, p. 135). By the time students find themselves in a
higher educational classroom, most have seen a great deal of teaching and many
understand which teaching approaches best maximize their own learning. As Ramsden
(1991) puts it, “nonexperts in a subject are uniquely qualified to judge whether the
instruction they receive is helping them to learn” (p. 137). Marsh (1987) reports students
are rarely misled into confusing good entertainment or performance with effective
teaching. Some instructors are good performers or entertainers, yet fail to teach
effectively the concepts and objectives of the course. Student evaluations are the means
whereby they evaluate effective teaching, and rarely are students mislead into thinking an
entertaining performer is always an effective teacher (Ramsden, 1991). However, an
effective teacher can also be a great entertainer. A fair and effective evaluation based on
performance and designed to encourage improvement in both the teacher being evaluated
and the college or university is a basic need of a good teacher evaluation system (Stronge,
1997). A comprehensive teacher evaluation system should be outcome oriented in that it
contributes to the personal goals of the teacher, to the mission of the program, to the
school, and to the educational institution, and should provide a fair measure of
performance (Stronge, 1997).
Studies have shown a positive relation between student evaluation and student
achievement, supporting the validity of student ratings. The many sources of potential
bias in student evaluations are minimized if exacting controls over methods of
administration are imposed and if students are asked only about those aspects of teaching
32
which they are qualified to comment upon. The general consensus is that there is no other
single measure of teaching performance which is as potentially valid (Ramsden, 1991).
Measuring effective instruction. Studies have shown aggregate-level
associations between course quality and perceived student learning using the instrument
of course evaluations (Entwistle & Tait, 1990; Ramsden & Entwistle, 1981; Ramsden et
al., 1989). These findings point to the usefulness and reliability of student evaluations in
comparison to other measures such as peer evaluations (Marsh, 1987; Feldman, 1978;
Roe & Macdonald, 1983). Ramsden (1991) also stated that because students see a great
deal of teaching, they are in an unrivalled position to comment on its quality. Moreover,
it shows that non-experts in a subject (i.e., students) are uniquely qualified to judge
whether the instruction they receive is helping them to learn. After all, students know if
they are benefiting from and learning as a result of a particular method or style of
teaching.
In summary, there is much literature available that supports the concepts
presented in the LM. Preparation before class is essential to creating a more active
learning environment in the classroom. Incorporated into the Teach One Another step of
the LM are a variety of teaching pedagogy and strategies that could be employed.
Cooperative learning, peer-to-peer instruction, and group work—just to name a few—
have been found to be effective instructional teaching strategies when correctly
implemented. Likewise, the third process step of the LM—the Ponder/Prove process
step—also has scholarly roots tied to it. In addition, literature also is prevalent in the area
of using end-of-course evaluations to measure effective teaching. This study will
33
investigate the implications of implementation of the three process steps of the LM on
end-of-course evaluations.
34
CHAPTER 3: METHODOLOGY
As mentioned previously, the purpose of this study is to examine the effects of the
LM on course ratings and student perceived learning as found on end-of-course
evaluations. The methodology used to collect the desired data for this study was
predominately through the compilation and interpretation of end-of-course student
evaluations. Course evaluations are archived at BYU-I and are considered public
information; they may also be viewed upon request. This chapter first details the
methods used to collect and compile data from end-of-course evaluations for both the old
evaluation instrument and the new evaluation instrument, both used to measure course
ratings and student perceived learning. The second section of this chapter specifies the
methodology and procedures used to address each of the three research questions
governing this study.
The methodology used in this study was determined based on the three research
questions, as previously outlined. The first research question asks if a change has
occurred in end-of-course ratings from the Fall 2007 semester (pre-LM implementation)
to the Winter 2010 semester, as the LM was introduced on campus? The purpose in
asking and addressing this question was to learn how end-of-course ratings have changed,
if at all, from the semester just prior to when the LM was introduced (Fall 2007) to the
latest semester in which data were collected for this study (Winter 2010). Remember, the
LM was first introduced during the Fall 2007 semester, but was not fully endorsed and
encouraged as a campus wide model for teaching and learning until the Winter 2008
semester. The purpose in asking research question one was to learn if the implementation
of the LM at BYU-I has had any significant affect on course ratings over time, from a
35
general perspective. Now admittedly, many variables can be contributed to the inflation
or deflation of end-of-course ratings from one semester to another. This chapter will
discuss such variables and will also address the methods used to control for confounding
variables associated with each research question.
Research question two states, “How have student perceived learning scores changed
based on course evaluation data collected from Fall 2007 to Winter 2010 semesters?”
The methodology used in examining this research question is similar to the methodology
used to address research question one, as mentioned above, with the exception of student
perceived learning scores being the focus. In other words, research question two
examines the changes that may have occurred in perceived student learning scores over
time.
Both the old evaluation instrument along with the new evaluation instrument
(introduced Winter 2009 semester) asks the same, identical question regarding student
perceived learning. Even though the two course evaluation instruments contain the same
question verbatim, regarding perceived learning, the questions preceding it are different.
The differences in question sets from the old evaluation instrument to the new evaluation
instrument could bias the way students answer the perceived learning question. Methods
to control for the difference in the two evaluation instruments are discussed further in this
chapter.
The third research question governing this study asks, “Is there a relationship between
LM scores and student perceived learning based on the new evaluation instrument
(beginning Winter 2009 semester)?” This question asks how strongly correlated are LM
scores with perceived learning? Remember, LM scores are derived from the new course
36
evaluation instrument which was introduced as a means of collecting data Winter 2009
semester. The LM had already been introduced campus wide for one year when the
evaluation instrument changed to reflect the three process steps of the LM. In this new
evaluation instrument, questions are asked and scores given based on students’
perceptions of the level at which the LM was applied throughout the course. LM scores
are then compiled based on this information. This third research question addresses the
relationship between LM scores and student perceived learning scores. Thus a
correlation methodology was employed to help answer this question. Also included in
this chapter are the control variables used to limit outside influences that may have biased
the study.
The Two Evaluation Instruments
For this study, it is important to understand that course evaluation data spans two
different evaluation instruments (see Appendix II and III). The differences of these two
evaluation instruments will be evaluated in this section. The purpose in doing so is
critical to the integrity of this study as research questions one and two involve a
comparison of data that were generated from both instruments. For research question one
and two, both evaluation instruments were employed as a means of tracking changes in
course ratings and student perceived learning scores. The differences between the two
evaluation instruments are important to consider when comparisons are being made.
As mentioned previously, the new course evaluation instrument was first used to
measure course ratings and student perceived learning for the Winter semester of 2009.
The evaluation instrument was changed in an attempt to emphasize the LM and to
37
highlight the importance of students taking personal responsibility for their own learning.
As a result, questions were changed and added to reflect this initiative, as explained in
Chapter 1. This is important to consider as a methodological parameter as it has played a
role in determining how the variables will be studied.
LM scores. Questions that were added to the new evaluation instrument were
specific to the way either the student or instructor had applied aspects of the LM
throughout the course. Questions were either rewritten or new questions were added that
specifically addressed the level of implementation of the three process steps of the LM—
Prepare, Teach One Another, and Ponder/Prove (see Appendix III). As a result, LM
scores were computed. These results are made available for the course instructor. These
scores reflect the level at which the LM was applied in a course as students perceived it
to be. For example, if the collective student population for a specific course rated both
themselves and the instructor as moderately implementing the Prepare aspect of the LM
throughout the semester the instructor would receive around a 70% LM score for the
course. LM scores are critical to this study as they will be used in determining the
relationship between LM scores and perceived learning scores. Specific LM scores are
only found in the new evaluation instrument, employed Winter 2009. For this reason,
research question three, which addresses the relationship between the two variables, only
uses data from the Winter 2009 semester to the Winter 2010 semester.
The questions that make up the LM scores are listed in Figure 7. Each question
makes up the overall LM score for that particular LM process step, as shown. For
example, there are four questions that relate to the prepare process step of the LM. The
38
prepare LM score for a course is based on how students answer each of the four questions
respectively.
Prepare
I was prepared for each
class.
I arrived at class on
time.
Teach-One-Another
Ponder/Prove
I was an active participant in
I sought opportunities to
online or face-to-face class
reflect on what I had
discussions.
learned in the class.
I sought opportunities to share
my learning with others outside
of class.
The course as a whole
has produced new
knowledge, skills, and
awareness in me.
I worked hard to meet
I feel that I made important
the requirements of this
contributions to the learning and
class.
growth of fellow classmates.
The instructor held me
The course provided
accountable for coming
opportunities to learn from and
to each class prepared.
teach other students
The instructor provided
opportunities to reflect
upon my learning and
experiences in the class.
Group work, if assigned, was
beneficial and meaningful.
Students were actively involved
in this class through discussions,
group work, and teaching.
The instructor provided
appropriate opportunities to be
an active participant in the class.
Figure 7. LM Question Set from the New Course Evaluation Instrument
39
For the third research question, LM scores serve as the independent variable. As
mentioned earlier, these scores are derived from the questions students are asked to
evaluate based on their own performance in, and instructor’s level of, implementing the
three LM process steps throughout a course. In other words, students are asked to rate
the level at which they perceived the presence of the LM both in their effort and
throughout the course as designed by the instructor. They may choose from the
following categories: Very Strongly Disagree (VSD), Strongly Disagree (SD), Disagree
(D), Agree (A), Strongly Agree (SA), and Very Strongly Agree (VSA). Points are
assigned to their choice with the following scale: 1=VSD, 2=SD…7=VSA (see Figure 8).
The way the LM scores are compiled is by taking the average score for each process step
based on student responses. For example, suppose ten students enrolled for a particular
course one semester, and for each of the Prepare questions found on the end-of-course
evaluation they each answered VSA. In this case, 7 points would be added to each of the
four questions pertaining to the Prepare element of the LM for this particular course.
Thus, the average would be 28 points out of a maximum of 28 points. The end-of-course
evaluation LM score for Prepare would be 100% since each student selected VSA for this
element.
See the following figure for a breakdown of all three LM process steps:
40
Number of
Questions
4
Maximum Points
From Evaluation
7
Teach-One-Another
7
7
49 Points
Ponder/Prove
3
7
21 Points
Process Step of the LM
Prepare
Total Points Possible
28 Points
Figure 8. LM Score Totals
Another important aspect of course evaluations is the manner in which they are
administered at BYU-I. Courses are evaluated based on instructor’s employment status.
At BYU-I, faculty are hired as adjunct (ADJ), one-year appointments (1YR), or PreContinuing Faculty Status (CFS). Faculty, once in the CFS track, can work towards
continuing faculty status (VET) or in other words a “tenured” faculty member. This
process takes 4 years to successfully earning VET status. To earn this status, one factor
is largely dependent on end-of-course ratings. ADJ, 1YR, and CFS are evaluated each
semester. Later in this chapter the methodology used to control for differences in
employment status will be addressed for each research question.
Research Questions and Methods
Each research question has a slightly different methodological approach and will
therefore be addressed one at a time. It is helpful to remember that questions one and two
of this study compare the differences between course evaluation scores and perceived
learning scores, respectively, over a specific time period. The semesters included for this
analysis are the Fall 2007 semester through to the Winter 2010 semester, a total of eight
41
individual semesters. For both previously mentioned questions the purpose in comparing
the two variables to an earlier point in time is to learn if a change has occurred in course
ratings or perceived learning scores as the LM has been implemented campus wide. The
methods and procedures used to collect data and to control for confounding variables are
outlined below.
Again, the purpose in asking the first research question is to establish whether end-ofcourse ratings have been affected during the time period when the LM was introduced
across campus. The second research question asks how student perceived learning scores
have been affected, if at all, by the implementation of the LM. The third research
question delves deeper into the strength of correlation between the LM scores derived
from course evaluations and perceived learning scores. Again, each question has its own
set of variables and methodology in gathering and analyzing the data. Each of the three
research question in this study will address the following: the purpose in asking each
research question, the variables measured and controlled for, the data collection
methodology employed to aid in answering each respective question, and the analysis
procedure used for each research question.
Research Question 1. The first research question asks, “How have end-of-course
ratings changed from the Fall 2007 semester (pre-LM implementation) to the Winter
2010 semester, as the LM was introduced on campus?” The purpose in asking this
question is to find out if the LM has had an effect on end-of-course ratings over time, as
the LM has been implemented by the University to further student learning and to
encourage a more participatory learning environment. Once again, the LM was
introduced to students and faculty beginning Winter 2008 semester. From that point on,
42
did course ratings increase, decrease, or stay the same? One might conclude that due to
the added emphasis on students taking more responsibility for their own learning, course
ratings might be adversely affected by students who prefer a more traditional approach in
receiving instruction (Felder & Brent, 1996). Thus begs the question, does application
of the LM positively or negatively affect end-of-course ratings?
As mentioned above, questions were added to the new evaluation instrument in
the Winter semester of 2009. The change in the evaluation instrument has impacted the
methodology used to address research question one. To control for the difference in
evaluation instruments the researcher collected data regarding course ratings for the four
semesters preceding Winter 2009 semester and data found in the four semesters after the
new evaluation instrument was employed. This was done to control for the differences in
the two evaluation instruments. Therefore, the four semesters previous to the Winter
2009 semester were evaluated independent of the four semesters following this time.
Again, the first four semesters, beginning with the Fall 2007 semester, represents the time
period where the old evaluation instrument was employed to survey students. This first
time period will be referred to as T1 from this point on. The second time period,
beginning with the Winter 2009 semester, represents the four semesters following the
change in the course evaluation instrument, as previously explained. This time period
will be referred to as T2.
In addition, a second variable addressed in this study was for new instructors; or
in other words, the natural change that occurs when an instructor teaches a course more
than one time. In general, instructors have a natural propensity, to change, adapt, and
adjust the content and approach of a course each time they teach it (Kagan, 1992). This
43
can be especially true when an instructor teaches a course for the first or second time. No
doubt these adjustments, no matter how subtle they may be, could affect end-of-course
ratings. In other words, instructors experience learning curves in their first few attempts
at teaching a course, which can affect end-of-course ratings. To control for this variable
the researcher collected data from only those faculty members who had taught each
respective course more than three times previous to their teaching it the Fall of 2007
semester. Comparing data in this manner will reduce the chance of inflation of course
evaluations due the learning curve that can be present with instructors teaching a course
in their first few attempts (Kagan, 1992).
Another consideration in addressing research question one is whether to compare
end-of-course evaluation scores for specific instructors who taught the same course in
specific semesters or to include data from all instructors over a broader time period. In
an attempt to control for conflicting variables, the researcher has chosen to look
specifically at instructors who taught the same course during the specific semesters in
question (i.e., from the Fall 2007 semester to the Winter 2010 semester). In this way, the
same instructor teaching the same course Fall of 2007 is compared with himself/herself
teaching the same course Winter of 2008, etc…. The reasoning behind this
methodological approach being that the change in scores will more closely reflect the
change in the use of the LM by the course instructor rather than other influences.
Another variable that was considered being a contributing factor in end-of-course
scores was class size. For instance, if a class increased in student count from the Fall
semester compared to the Winter semester or vice versa, the change may have affected
course ratings for that particular course. Class size can and does have an effect on
44
learning outcomes as shown in a study by Krueger (1999). Krueger measured gains on
standardized tests for more than 11,600 students in Tennessee. In his study Krueger
categorized students into small and regularly sized classrooms based on the number of
students in each. He found a significant gain in test scores for those students in a smaller
classroom compared with those in regular sized classroom.
To control for class size the researcher has disaggregated the data (based on
Krueger’s 1999 study) to the following class sizes; small, medium, large, and very large.
For this study, classes that had less than 10 students enrolled were categorized as a small
class, classes with 10-20 students were categorized as a medium class, classes with 20-30
students enrolled were categorized as a large class, and classes with more than 30
students enrolled were categorized as a very large class. For research question one and
two, only those instructors who consistently had the same amount of students throughout
the eight semesters were included in the study.
The statistical tool chosen to analyze the change in end-of-course ratings between
semesters was through a comparison of means and ANOVA. With the use of ANOVA
the difference in means of course ratings were measured and compared one with another.
A paired, two sample t-test was a second statistical tool that was also employed for this
research question. In addition, a significance value (P-value) was also calculated to
determine if the change in end-of-course ratings were considered to be statistically
significant. The significance level was pre-determined to be .05, as it was calculated
from the two-sample two-tailed t-test. Meaning, if a change in scores had occurred
between the two semesters and was calculated to be less than the .05 percentile, it was
45
considered to be a significant change. This analysis tool was used to measure the
difference in end-of-course ratings for both T1 and T2, respectively.
In summary, the focus of research question one relates to the change in course
ratings as the LM has been implemented. As mentioned, the variables that will be
controlled for are; the change to the evaluation instruments, learning curve of new
instructors, inconsistencies in different instructors teaching a particular course, and class
size. To control for the confounding variables associated with this question, the
researcher collected data from those instructors who have taught the same course each of
the eight semesters, from the Fall 2007 to the Winter 2010 semesters. Instructors who
had not taught the course at least three times previous to the Fall 2007 semester were
eliminated. Due to the change in evaluation instruments, the first four semesters will be
analyzed independent of the last four semesters. In other words, T1 represent the period
when the old evaluation instrument was being used to track course ratings. T1 was
examined independent of T2, representative of the period when the new evaluation
instrument was employed to measure course ratings. Class size was also factored in,
meaning only those instructors who taught small, medium, large, or very large classes
consistently over the four semester time frame were included. As a result of controlling
for these variables, the data set was reduced from 850 courses down to 33 courses. The
following chapter will address the findings for research question one as outlined above.
Research Question 2. The second research question is much like the first in that
data was collected in an attempt to learn how the implementation of the LM affected, if at
all, student perceived learning. In this case, the second question specifically addresses
the change in student perceived learning over time. The question is stated in this way,
46
“How have student perceived learning scores changed based on course evaluation data
collected from Fall 2007 to the Winter 2010 semesters?” The purpose in asking this
question is to learn if students, in general, perceive that they have learned more in their
courses—compared to other college courses they have taken—as the LM has been
introduced at the University level.
In both the new course evaluation instrument and the old evaluation instrument
the perceived learning question remained the same; “Compared to other college courses
you have taken, would you say that you have learned….” The students then select either
“a great deal less”, “a little less”, “about the same”, “a little more”, or “a great deal more”
based on their experience. On both evaluation instruments students are asked to rate their
learning as it compares with other classes taken. This information was used to determine
the change, if any, in student’s perceived learning between the Fall 2007 and the Winter
2010 semesters. The student perceived learning score is derived from the above
mentioned question found on both course evaluation instruments (see Appendix II & III).
If a student answers that they have learned about the same amount when compared with
other college courses taken then the instructor would receive a 0 score, per that students’
answer. All students perceived learning scores are averaged for each course. In other
words, if an instructor received a 1.49 student perceived learning rating for a particular
course this would mean that the average student perceived learning for that course was
between “a little more” and “a great deal more” as compared with other college courses
taken.
As mentioned, both evaluation instruments contain the same question regarding
student perceived learning. Yet, the questions leading up to this one are different on the
47
pre-LM instrument compared with the new evaluation instrument, and could affect how
the latter is answered. More specifically, the new course evaluation includes questions
related to the level at which students were responsible for their own learning; “I was
prepared for each class”, for example. These questions may influence the way students
answer the perceived learning question found at the end of the evaluation. To control for
this variable, the researcher collected student perceived learning for each semester,
beginning with the Fall 2007 semester and each semester thereafter, up to the Winter
2010 semester. Attention was given to the difference in student perceived learning scores
between semesters. However, the first four semesters were compared one to another
independent of the last four semesters, and vice versa. This was done to reflect the
semesters when the old evaluation instrument was used compared with the semesters
when the new evaluation instrument was used to survey students.
Similar to the methodology employed for the previous research question, the
researcher compared only those semesters together in which the evaluation instrument
was unchanged. Again, T1 semesters were compared together, as were T2 semesters.
Thus, eliminating the possible effects associated with the use of two different course
evaluations instrument to measure student perceived learning scores. The same variables
were controlled for to measure the change in perceive learning scores as were used to
measure the change in course ratings. This resulted in the same data set of 33 courses
evaluated regarding changes in perceived learning scores over the two time periods—T1
and T2.
A comparison of means and ANOVA were also employed as the statistical
method for analyzing the change in mean scores for student perceived learning. The
48
differences between the means were explained. Likewise, a p-value was employed to
measure the level of significance between the differences. A small p-value indicates that
the probability of the difference in end-of-course perceived learning scores changing due
to chance unlikely. A t-test was then employed to measure the level of significance
between the difference of means for T1 semesters, and again for T2 semesters. The
significance level was pre-determined to be .05 or less. In other words, the change from
the time the LM was first introduced to the time the course evaluation was changed will
be analyzed independent of the time the new course evaluation was introduced to the
most recent semester.
Research Question 3. The third research question states, “Is there a relationship
between LM scores and student perceived learning based on the new evaluation
instrument (beginning Winter semester 2009)?” Another way to ask this question would
be to say, “Does evidence of LM application in a course help predict student perceived
learning scores for a particular course?” The purpose in asking this question is to identify
the relationship of the LM on perceived learning scores. Do scores drop, increase, or stay
the same as the level of implementation of the LM changes, for example. Perhaps, the
most telling information that can be gained in asking this question is whether or not
students perceive themselves learning more when LM process steps are employed by an
instructor. In other words, do students perceive themselves as learning more information
when an instructor employs the three process steps of the LM? This is the purpose of
research question three, to learn the relationship between LM scores and perceived
learning scores.
49
Perceived learning scores can be influenced by a variety of variables, as mentioned
previously. However, for this question LM scores are used as the independent variable
whereas perceived learning scores serve as the dependent variable. In other words, the
researcher is interested in learning how the level of implementation of the LM correlates
with students’ perceived learning for a particular course. Furthermore, the researcher has
broken down the data to more specific variables. These variables include the relationship
between LM scores and perceived learning scores as they relate to; (1) class size (number
of students enrolled in the course), (2) semester in which the student was enrolled, and
(3) instructor employment status (i.e., adjunct, one-year appointment, Pre-CFS (nontenured), and CFS (tenured)).
As mentioned previously, LM scores are derived from the new evaluation instrument
and are based on student perceptions of the level at which the three process steps of the
LM were infused throughout the course. The data set for this question comes from
student evaluations of individual courses during T2. The researcher disaggregated the
information from course evaluations to draw out the previously mentioned variables.
All data sets found in T2 (the last four semesters of this study) were incorporated in
the findings. Remember research question three explores the relationship between LM
scores and perceived learning scores given particular variables such as; class size,
semesters taken, and instructor’s employment status. LM scores are a product of the new
evaluation instrument and are therefore only available for the last four semesters of this
study. When the evaluation instrument was changed the LM had already been introduced
to students and faculty for one year (three semesters). All perceived learning scores and
50
their associated LM scores for T2 were used to measure the strength of correlation
between these variables.
To find this relationship between LM scores and perceived learning scores the
researcher first compared mean perceived learning scores by semester and employment
status for all data points in T2. The relationship between the independent variable with
the dependent variables, as previously explained, was evaluated using a correlation
matrix. Data points were collected and analyzed using SPSS as the statistical analysis
tool. R-values and P-values were calculated to show the strength of correlation and the
level of significance for each set of variables. Results and findings are found in the
following chapter.
Limitations
Limitations of this study include the consistency of data and regularity of course
evaluations collected for instructors. Untenured faculty members including adjunct, oneyear appointments, and Pre-CFS are usually evaluated each semester for their first four
years of teaching. Tenured faculty members who have earned CFS are only evaluated
once every three years, unless requested by themselves or by one of their supervisors.
Due to this method for evaluating faculty members, the numbers of qualifying courses for
this study are limited.
Another limitation of this study relates to the fact that the faculty is ever
changing. Faculty members have been hired at different times and at different capacities
(i.e., Full-time, adjunct, or temporary status). As a result of faculty being ever changing,
data will likely reflect this pattern. BYU-Idaho has experienced tremendous growth since
51
the announcement of Ricks Jr. College becoming a four year university. Each year, an
increase in the number of new faculty members hired has gone up, to keep up with the
student growth. This again has affected the methodology for this study as controls have
been put in place, as previously mentioned to control for new instructors.
Faculty members often change their teaching methodology and strategies from
semester to semester based on previous findings, student’s needs, workload, and a variety
of other reasons. Determining the actual reasons for increased end-of-course ratings,
instructor ratings, and perceived learning can be a difficult task. Even though many of
the variable that could affect the viability of this study have been controlled, one
significant limitation must be recognized; students differences from one semester to
another. Students’ attitudes, learning styles, and personalities can change significantly
semester-by-semester. This limitation will not be accounted for in this study; only the
collective perceived answers given by students as it relates to end-of-course evaluations
have been evaluated and compared.
Lastly, but perhaps most significant, is the fact that the student population
changes each semester. New students enter each semester with new learning styles and a
variety of experiences. The demographics of each semester can also change dramatically
from semester to semester. Likewise, class sizes fluctuate between semesters.
As this is an Ex Post Facto study, some of the limitations mentioned are not
possible to measure or control for. However, those variables that are controllable have
been considered and methodology to control for such variables has been previously
mentioned.
52
FINDINGS
In this chapter, data pertaining to each research question were analyzed and
findings were discussed. This study was conducted in an attempt to determine the
significance of implementing and applying the three process steps of the LM (Prepare,
Teach One Another, and Ponder/Prove) relative to course ratings and perceived learning
scores, as measured from course evaluations. Findings associated with each research
question are addressed in this chapter.
Research questions one and two are similar in methodology, in that the researcher
took a large data set and disaggregated it down to a smaller fraction to control for
confounding variables. As a result, 33 courses were used in determining the change in
course ratings and perceived learning scores over time. Remember, research question
one asks, from the Fall 2007 semester up to the Winter 2010 semester, how have course
ratings changed over the eight semesters time period? In the second research question,
the change in student perceived learning for the same time frame was the focus. Both
research question one and two employed a comparison of means, ANOVA, and a two
sample t-test to measure the differences in mean scores.
Research question three looks at a variety of variables that may contribute to the
level of correlation between LM scores and student perceived learning scores. As
mentioned in Chapter 3, the three variables that are measured relative to the third
research question are; class size, semester when courses were taken, and instructor
employment status. By disaggregating the data to class size, semesters, and instructor
employment status the degree at which these variables affect the strength of correlation
53
between LM and student perceived learning scores were evaluated and analyzed
respectively.
Research Question 1: A Change in End-of-Course Scores
The first research question in this study asks if a change was noticed in course
ratings beginning Fall 2007, representing the semester preceding the implementation of
the LM. In other words, the LM was fully introduced to students and faculty before the
second semester (Winter 2008) of this study. Remember, there are a total of eight
semesters examined throughout the study. The first semester represents the time before
the LM was fully implemented, while the remaining seven semesters are reflective of the
period when the LM was introduced and encouraged as a learning and teaching
methodology for student and faculty campus wide.
As mentioned previously, the evaluation instrument used to measure course
ratings was changed beginning Winter semester, 2009. For this reason, course ratings for
T1 will be compared one with another. Likewise, end-of-course ratings for T2 will be
evaluated. Remember, T1 represents the semesters (Fall 2007 to Fall 2008) where the old
evaluation instrument was used to collect data from student evaluations. T2 represents
the semesters (Winter 2009 to Winter 2010) in which the new evaluation instrument was
used for the same purpose. The methodology used to measure the mean differences in
course ratings were primarily through a comparison of means using ANOVA and a
paired, two-sample mean t-test, as previously explained.
Have course ratings change over time? For question one, the researcher collected
end-of-course ratings for all courses from the Fall 2007 semester to the Winter 2010
54
semesters. The data were further disaggregated to only those instructors who had taught
the same course in each semester, consecutively. The researcher found 33 courses that
met the criteria. Course ratings for the 33 courses were compared and analyzed.
A correlation matrix was used to evaluate the mean differences in course ratings
for each semester (see Appendix IV). This revealed that each semester’s course ratings
are closely correlated. In other words, no semesters revealed a significant deviation from
the other semesters.
Furthermore, a breakdown of each individual course by semester was given in a
comparison of means table. Table 1 shows those qualifying courses for this study; each
course is categorized by department. Notice the difference in individual course ratings
for each individual course over the eight semesters. For the most part, increases in course
ratings between semesters were realized over time. When comparing course ratings for
Fall 2007 with Fall 2008 course ratings 21 out of the 33 courses showed gains, one
course stayed equal, and 11 courses showed a drop in courses ratings.
55
Table 1.
Course Ratings of Qualifying Courses for T1Semesters
Departments
Accounting
Agribusiness
Animal Sciences
Arch & Const
Art
Biology
Communications
Computer Science
English
Exercise and
Sports Sciences
Foreign Languages
History
Home and Family
Mathematics
Music
Psychology
Sociology
Inst.
ID
146
146
267
266
727
888
471
471
225
228
743
931
889
717
297
203
767
722
722
722
722
367
202
470
836
508
508
111
823
534
245
245
245
Fall
2007
4.37
4.61
5.68
5.60
6.33
5.00
5.24
5.75
5.63
6.18
5.95
6.22
5.00
4.37
6.17
5.34
4.50
6.14
6.30
6.05
5.81
5.53
5.07
5.95
4.38
5.58
5.33
7.00
4.79
4.76
5.57
4.68
5.46
Course
ACCTG 202
ACCTG 312
AGRON 122
AS 347
CONST 430
ART 130
B 461
B 478
B 466
COMM 150
COMM 230
COMM 130
COMM 100
COMM 316
COMM 102
CS 124
ENG 316
ESS 160
ESS 161
ESS 162
ESS 264
SPAN 101
HIST 201
CHILD 210
CHILD 335
HFED 100
HFED 350
MATH 100A
MUSIC 209
PSYCH 302
SOC 111
SOC 340
SOC 360
56
Winter
2008
4.71
4.29
5.70
6.29
6.09
5.19
5.53
6.10
6.00
6.26
6.47
6.20
4.94
4.67
5.77
5.59
4.87
6.29
6.47
5.93
5.91
5.38
4.94
5.47
4.67
5.74
5.24
6.29
4.54
5.43
5.66
4.43
4.91
Spring
2008
5.15
4.78
6.00
6.80
5.86
5.68
5.90
6.10
6.67
5.86
6.29
6.00
4.58
4.75
5.96
5.82
5.11
6.10
6.20
6.10
5.54
4.89
4.48
5.77
5.41
5.86
5.90
5.50
4.50
4.87
5.06
5.20
5.40
Fall
2008
4.60
4.61
5.54
6.43
5.83
5.22
5.91
6.36
6.42
6.59
5.94
6.52
4.75
4.52
6.23
6.22
5.55
6.18
6.25
6.18
6.25
5.00
5.00
5.80
6.00
6.12
5.38
5.95
4.62
5.12
5.50
4.95
6.00
Mean
Score
4.71
4.57
5.73
6.28
6.03
5.27
5.64
6.08
6.18
6.22
6.16
6.23
4.82
4.58
6.03
5.74
5.01
6.18
6.31
6.07
5.88
5.20
4.87
5.75
5.12
5.82
5.46
6.18
4.61
5.04
5.44
4.82
5.44
Also, notice the mean scores for the 33 courses in question. The two
lowest mean course ratings were ACCTG 312 and COMM 316. The two highest mean
course ratings were ESS 161 and AS 347. Interestingly, the two courses that scored
lowest on course ratings were found to be in the Accounting and Communication
departments, respectively. While the two highest mean course ratings were from the
Exercise and Sports Sciences and Animal Sciences departments. The courses that scored
lowest are typically taken to fulfill a general education requirement, whereas the courses
that scored the highest normally are taken to satisfy a major course requirement.
Other observations noticed as a result of comparing course ratings for T1were the
quantity of courses that showed consistent gains over the four semester period. There
were six courses that showed consistent gains in course ratings from one semester to the
other over time. Those six courses include B461, B478, CS 124, ENG 316, CHILD 335,
and HFED 100. There was only one course found of the 33 total courses in which a
consistent drop in course ratings was noticed over the four semesters, CONST 430. In
other words, six of the 33 courses showed consistent gains in course ratings over the four
semester period, whereas only one course showed consistent losses in course ratings for
the same period.
Once again, T1 represents the first four semesters of this study. In these
semesters the old evaluation instrument was used to collect course rating data. The first
semester (Fall 2007) is representative of the last semester preceding the implementation
of the LM. The next three semesters in T1 represent the first three semesters when the
LM was introduced and application of said model encouraged campus wide.
57
Course rating data for T2 was also collected and organized in a similar manner.
T2 represents the time period when the new evaluation instrument was employed to
gather end-of-course data. Remember, T2 includes the last 4 semesters of this study.
Due to the fact that the two evaluation instruments are somewhat different as it relates to
the question sets, independent analyses where conducted for each time period. Table 2
shows course ratings for each of the 33 courses during T2.
58
Table 2
Course Ratings of Qualifying Courses for T2 Semesters
Departments
Accounting
Agribusiness
Animal Sciences
Arch & Const
Art
Biology
Communications
Computer Science
English
Exercise and Sports
Sciences
Foreign Languages
History
Home and Family
Mathematics
Music
Psychology
Sociology
Inst
ID
146
146
267
266
727
888
471
471
225
228
743
931
889
717
297
203
767
722
722
722
722
367
202
470
836
508
508
111
823
534
245
245
245
Course
ACCTG 202
ACCTG 312
AGRON 122
AS 347
CONST 430
ART 130
B 461
B 478
B 466
COMM 150
COMM 230
COMM 130
COMM 100
COMM 316
COMM 102
CS 124
ENG 316
ESS 160
ESS 161
ESS 162
ESS 264
SPAN 101
HIST 201
CHILD 210
CHILD 335
HFED 100
HFED 350
MATH 100A
MUSIC 209
PSYCH 302
SOC 111
SOC 340
SOC 360
Winter Spring
2009
2009
4.63
4.54
4.25
4.39
5.68
5.29
6.67
6.45
6.07
6.20
6.11
5.24
6.13
5.71
6.63
6.00
5.92
6.00
6.35
6.19
5.67
6.17
6.45
6.08
5.38
4.53
4.76
5.30
6.29
6.06
6.08
6.00
5.16
5.02
6.18
6.31
6.69
6.54
6.26
6.25
6.10
6.38
5.24
6.00
5.33
5.03
6.18
6.07
5.29
4.92
6.07
5.86
5.74
5.93
6.57
5.50
4.52
5.35
4.77
5.56
5.47
5.67
5.16
4.84
5.00
5.93
59
Fall
2009
4.81
4.59
5.96
6.60
6.11
6.00
6.15
6.62
5.62
6.23
6.00
6.02
4.81
4.50
6.20
5.39
5.63
5.83
6.58
6.53
6.70
5.80
5.00
6.13
6.00
6.19
5.89
6.00
5.32
5.75
5.73
5.27
5.41
Winter Mean
2010 Score
4.87
4.71
5.38
4.65
5.70
5.66
6.50
6.56
6.24
6.15
5.43
5.70
5.81
5.95
6.05
6.32
5.75
5.82
6.24
6.26
5.54
5.84
6.21
6.19
5.60
5.08
5.60
5.04
6.26
6.20
6.07
5.89
5.36
5.29
5.77
6.02
6.57
6.60
6.13
6.29
6.55
6.43
5.59
5.66
4.43
4.95
6.10
6.12
6.00
5.55
6.06
6.05
5.60
5.79
6.36
6.11
5.15
5.08
5.48
5.39
5.31
5.54
5.12
5.10
5.08
5.35
During T2, it seems that a leveling off occurred for each of the 33 courses in
question. Only one course (ACCTG 312) showed consistent gains in course ratings for
each of the four semesters in T2. Likewise, there was only one course (HIST 201) that
showed consistent drops in each consecutive semester during the same time period.
Interestingly, ACCTG 312—the one course that showed gains in each of the four
semesters during T2—had the lowest mean score for T1. This being said, the two lowest
mean scores in T2 were found to be ACCTG 202 and ACCTG 312. Even with consistent
gains in ACCTG 312, this course went from having the lowest mean course ratings of the
33 courses being evaluated to next to lowest mean rating. What’s most curious about this
course is the jump in course rating for Winter 2010 semester. What was different about
this semester that caused the change? Was it more use of the LM? Research questions
two and three will examine further these inquiries.
The two highest course ratings were again found for ESS 161 and AS 347. In
fact, mean course rating scores for both ESS 161 and AS 347 for T1 and T2 respectively,
show an increase from one period to another. ESS161 showed a gain in mean score from
T1 to T2 of .29 points. Likewise, AS 347 course showed gains from 6.28 to 6.56 during
the same time, a difference of .28. The high ratings and evidence of continued improved
scores are cause for further investigation, as this study relates to the affects of LM
application on course ratings and student perceived learning. Are LM scores associated
with higher course ratings? Again, question three will address this relationship.
Once individual courses were examined using a comparison of means analysis the
researcher used a one-way ANOVA test to measure the differences in the combined
means for T1 and T2. In other words, an average course rating of all 33 courses were
60
computed for each semester. This was done in an attempt to compare the differences in
mean course ratings for the entire 33 courses by semester. The researcher wanted to learn
if course ratings improved over time when the 33 courses were grouped into one body.
Again, T1 represents the semesters previous to the change in the course evaluation
instrument (from Fall 2007 to Fall 2008 semesters). T2 represents the semesters when
the new course evaluation was employed (from Winter 2009 to Winter 2010 semesters).
Table 3 shows the results of the one-way ANOVA test for T1 and T2. Notice the
differences between the means for each semester.
Table 3
ANOVA: End-of-Course Ratings for T1& T2 Semesters
Evaluation Instrument
Standard
Semester
N
Course Rating
Employed
Deviation
Fall 2007
33
5.46
.45
Old Evaluation
Winter 2008
33
5.51
.41
Instrument (T1)
Spring 2008
33
5.57
.37
Fall 2008
33
5.68
.41
Winter 2009
33
5.72
.46
New Evaluation
Spring 2009
33
5.67
.35
Instrument (T2)
Fall 2009
33
5.79
.34
Winter 2010
33
5.75
.26
*Indicates a significance level at or less than .05 between semesters
61
In five out of the seven opportunities for course ratings to change an increase in
average course ratings was realized. Only in two semesters did the average course rating
drop for the composite body of courses. Although the average means for each semester
were different, the significance level was measured at .55 between each semester. This
indicates that no significant changes in end-of-course ratings were found between
semesters during the T1 time period. Or in other words, no significance was found for
the change in average course ratings from one semester to another (i.e., from Fall 2007
semester to the Winter 2008 semester). Again, for T2 no significant differences between
semesters were found. The P-value for end-of-course ratings for T2 was measured at .86,
clearly showing no significance.
However, notice the increase in course ratings from Fall 2007 semester to Winter
2010 semester for the 33 courses in question. Gradually, the end of course ratings
increased from one semester to another over the time period when the LM was introduced
and implemented campus wide. Remember, the LM was informally introduced just prior
to the Fall 2007 semester. It wasn’t until the next semester, Winter 2008, when the LM
was fully encouraged to be implemented by students and faculty throughout the campus.
The three semesters following the Fall 2007 semester, representing the time period when
the old evaluation instrument was employed, realized a consistent increase in course
ratings.
From the Fall 2007 semester to Fall 2008 semester, average end-of-course ratings
improved from 5.46 to 5.68, more than .2 points increase. However, the improvements in
end-of-course ratings for T2 were not as profound as in T1. The findings reveal a
average course rating for Winter 2009 of 5.72 and a mean course rating for Winter 2010
62
of 5.75, a little more than .03 point difference. A two-sample t-test was used to measure
the level of change and to calculate the significance level of the change for T1 and T2.
The hypothesized change was selected to be zero, meaning that no changes in course
ratings were anticipated.
The results of this t-test showed that the change in course ratings for T1 was
significant (see Appendix V). Remember, the significance level of this study was set at
the .05 level. For T1, the findings reveal that the increase in end-of-course ratings for T1
measured .019. A similar t-test was conducted for T2. Taking the beginning semester
(Winter 2009) and comparing it to the most recent data available for this study (Winter
2010), revealed a P-value of .66 (see Appendix V).
The significance value for T2 is very large at .66, revealing no significance in
the differences between mean end-of-course ratings for this time frame. The findings
show that during the T1 semesters a steady increase was apparent for the 33 courses
being measured. But, in T2 semesters the change in end-of-course ratings showed no
significance between Winter 2009 and Winter 2010 semesters. The change in course
ratings for T1 was found to be significant, whereas the change in course ratings for T2
was not significant.
The difference between end-of-course ratings for Fall 2008 semester and Winter
2009 semester was also evaluated to see if the change in course evaluation instrument had
any significant affect on course ratings. The Fall 2008 semester showed a mean course
rating for the 33 courses under investigation measuring 5.68 points. Whereas, the Winter
2009 semester showed a mean course rating of 5.72 for the same courses. The
differences of mean course ratings between Fall 2008 semester and Winter 2009 semester
63
measured less than .04 points, not a noticeable change. Likewise, using a two sample ttest the significance value between these two semesters measured .57, revealing no
significance (see Appendix V). Remember, the Fall 2008 semester represents the last
semester in which the old evaluation instrument was used to measure end-of-course
ratings. Also, the Winter 2009 semester represents the first semester in which the new
evaluation tool was employed to gather end-of-course ratings, among other data. This
lack of significance between the Fall 2008 and Winter 2009 semester is surprising. The
change in evaluation instruments did not seem to have much of an effect on course
ratings between the two semesters in question.
As a result of finding no difference between course ratings between T1 and T2 a
comparison of means for the beginning semester of this study with the last semester of
this study was conducted. The difference in mean course ratings for the Fall 2007 and
the Fall 2010 semesters were evaluated using the same two sample t-test. The same 33
courses were used in this evaluation to keep with methodological criteria previously
established for this research question. The purpose in comparing the Fall 2007 semester
with the Winter 2010 semester was to learn the significance of the change in course
ratings from the semester before the LM was implemented to the latest semester. In other
words, for the average mean course ratings of the 33 courses in question, what changes
have occurred over the last eight semesters and was this change significant?
When course ratings from the first to the last semester are compared a significant
difference in mean scores is realized. The findings show a significant change between
composite course ratings over time. The significance level for this two-tailed t-test was
measured at the .003 level, meaning that there is roughly a 99.997% chance that the
64
difference between the two end-of-course mean ratings did not occur by chance alone
(see Appendix V). The mean rating for Fall 2007 semester was measured to be 5.46,
compare that with the mean score for Winter 2010 of 5.75. With an n of 33, this
difference becomes significant between the two semesters.
What does this tell us? For one, it shows that some factor has caused or aided this
increase in scores from one semester to another. However, it doesn’t help us understand
what that factor was. Many variables may have played a role in causing this increase in
course ratings and only one of those contributing factors might be the implementation of
the LM. Research Question three deals with data that may help us understand to what
level the LM may have been a contributing factor in raising end-of-course scores for
these instructors.
Research Question 2: A Change in Perceived Learning Scores
The second research question governing this study asks how student perceived
learning scores have changed during the time the LM was introduced and implemented
campus wide. Perceived learning scores are evaluated and calculated with a similar
methodology used to evaluate the previous research question regarding course ratings.
The purpose in this evaluation is to learn if students’ perceived learning scores have
increased, decreased, or if the scores stayed the same in relation to the time the LM was
implemented campus wide (between the Fall 2007 and Winter 2010 semesters).
The data used to find the differences between perceived learning scores for T1
and T2 were compiled using the same methodology as in the first research question.
Instructors were identified who taught the same courses for each of the eight semesters,
65
spanning the Fall 2007 semester through to the Winter 2010 semester. The differences in
class sizes between semesters were controlled for and beginning instructors (those who
have taught the course less than 3 times previous to the Fall 2007 semester) were
eliminated from the data set. In controlling for such variables, the remaining qualifying
data set revealed 33 courses that met all criteria previously address in Chapter 3. Again,
the 33 courses used to answer question two, regarding the change in student perceived
learning over time, are the same courses used in research question one.
In the previous section, the findings show that course ratings increased during T1,
T2, and from Fall 2007 to Winter 2010 semesters, as reported above. But, how has
students’ perceived learning changed, if at all, during T1, T2, and between the Fall 2007
to the Winter 2010 semesters? This section will attempt to compile and compute data
that will help in answering this question.
A correlation matrix was used to quickly compare perceived learning scores for
each of the eight semesters of this study (see Appendix VI). In each case, the correlation
values of perceived learning scores for each semester were found to be significantly
correlated. In other words, none of the semesters showed extreme differences in
perceived learning scores. Each of the perceived learning scores for the eight semesters
was found to be homogeneous one with another.
Individual perceived learning scores were compiled using a comparison of means
table. Table 4 shows the results of individual perceived learning scores for each of the 33
courses for T1. Remember, the scores may range from negative two to positive two.
66
Table 4.
Perceived Learning Scores for T1 by Semester
Departments
Accounting
Agribusiness
Animal Sciences
Arch & Const
Art
Biology
Communications
Computer Science
English
Exercise and Sports
Sciences
Foreign Languages
History
Home and Family
Mathematics
Music
Psychology
Sociology
Inst
ID
146
146
267
266
727
888
471
471
225
228
743
931
889
717
297
203
767
722
722
722
722
367
202
470
836
508
508
111
823
534
245
245
245
Course
ACCTG 202
ACCTG 312
AGRON 122
AS 347
CONST 430
ART 130
B 461
B 478
B 466
COMM 150
COMM 230
COMM 130
COMM 100
COMM 316
COMM 102
CS 124
ENG 316
ESS 160
ESS 161
ESS 162
ESS 264
SPAN 101
HIST 201
CHILD 210
CHILD 335
HFED 100
HFED 350
MATH 100A
MUSIC 209
PSYCH 302
SOC 111
SOC 340
SOC 360
67
Fall
2007
0.64
0.19
1.21
1.89
1.38
1.43
0.81
1.30
1.26
1.38
1.38
1.78
0.62
0.37
1.06
1.11
0.17
0.91
0.77
0.57
0.94
1.41
0.93
0.74
1.00
1.00
0.44
2.00
0.59
0.75
1.00
1.08
1.39
Winter Spring Fall Mean
2008
2008
2008 Score
0.71
1.32 0.73
.85
-0.14
0.68 0.10
.21
1.00
1.55 0.86 1.16
1.71
2.00 2.00 1.90
1.35
1.20 1.17 1.28
1.21
1.64 1.33 1.40
1.20
1.50 1.55 1.27
1.60
1.90 1.68 1.62
1.50
1.89 1.63 1.57
1.07
1.05 1.51 1.25
1.50
1.53 1.13 1.39
1.58
1.53 1.48 1.59
0.58
0.00 0.07
.32
0.25
0.39
-.16
.21
0.87
1.36 1.00 1.07
1.17
0.30 1.50 1.02
0.35
0.44 0.47
.36
1.00
0.71 1.04
.92
0.94
0.48 0.50
.67
0.60
0.60 0.73
.63
1.09
0.92 1.25 1.05
0.79
1.58 1.42 1.30
0.97
0.67 0.46
.76
0.75
1.07 0.91
.87
1.20
1.35 1.50 1.26
1.29
1.37 1.13 1.20
0.48
0.95 0.81
.67
1.75
2.00 1.60 1.84
0.57
1.17 0.50
.71
1.00
1.13 0.97
.96
1.04
0.56 0.88
.87
0.90
1.29 1.23 1.13
1.09
1.38 1.35 1.30
The findings show that the two courses with the lowest mean perceived learning
scores for T1 are the same courses that had the lowest course ratings for the same time
period. Both ACCTG 312 and COMM 316 scored a .21 for perceived learning. This
shows that a correlation between course ratings and perceived learning scores may exist.
Likewise, AS 347 had the highest mean perceived learning score and course rating for
T1.
What about for T2? Table 5 shows the result of perceived learning scores by
semester for T2. Again, AS 347 showed the highest perceived learning score for T2.
Also, ACCTG 312 and COMM 316 accounted for two out of the three lowest perceived
learning scores for T2.
68
Table 5.
Perceive Learning Scores for T2 by Semester
Departments
Accounting
Agribusiness
Animal Sciences
Arch & Const
Art
Biology
Communications
Computer Science
English
Exercise and Sports
Sciences
Foreign Languages
History
Home and Family
Mathematics
Music
Psychology
Sociology
Inst
ID
146
146
267
266
727
888
471
471
225
228
743
931
889
717
297
203
767
722
722
722
722
367
202
470
836
508
508
111
823
534
245
245
245
Winter Spring Fall Winter
Course
2009
2009 2009 2010
ACCTG 202
0.53
0.90 0.73
0.69
ACCTG 312
0.19
0.25 0.24
0.71
AGRON 122
1.03
0.80 1.00
1.09
AS 347
1.89
1.80 1.80
1.80
CONST 430
1.34
1.47 1.33
1.39
ART 130
1.68
1.38 1.74
1.21
B 461
1.69
1.62 1.75
1.38
B 478
2.00
1.57 1.92
1.52
B 466
1.62
1.65 1.08
1.30
COMM 150
1.50
1.29 1.40
1.42
COMM 230
1.20
1.08 1.42
1.15
COMM 130
1.63
1.62 1.43
1.62
COMM 100
0.58
0.15 0.43
1.10
COMM 316
-0.16
0.65
-.17
0.35
COMM 102
0.38
1.08 0.24
1.03
CS 124
1.49
1.61 1.44
1.53
ENG 316
0.61
0.71 0.81
0.49
ESS 160
1.17
0.93 0.78
0.54
ESS 161
0.86
0.34 0.83
0.67
ESS 162
0.40
0.50 0.27
0.13
ESS 264
1.20
1.46 1.60
1.64
SPAN 101
1.13
1.12 1.16
1.18
HIST 201
0.95
0.47 0.78
0.43
CHILD 210
1.24
1.24 1.17
0.56
CHILD 335
1.07
0.92 1.79
1.67
HFED 100
1.05
0.32 0.68
0.49
HFED 350
0.97
1.17 1.02
0.76
1.33 1.00
1.91
MATH 100A 1.67
MUSIC 209
0.03
0.65 0.77
0.55
PSYCH 302
0.62
1.12 1.14
1.20
SOC 111
0.74
1.10 1.10
0.85
SOC 340
1.03
1.17 1.20
1.18
SOC 360
0.75
1.31 0.71
0.46
69
Mean
Score
.71
.35
.98
1.82
1.38
1.50
1.61
1.75
1.41
1.40
1.21
1.58
.57
.17
.68
1.52
.66
.86
.68
.33
1.48
1.15
.66
1.05
1.36
.64
.98
1.48
.50
1.02
.95
1.15
.81
Preliminary evidence shows that a correlation may be present between course
ratings and perceived learning scores. In other words, if perceived learning scores are
high then course ratings will likely be high as well. This intuitive result seems plausible
based on experience. When a student enjoys a course they may have a tendency to rate
the course high. Also, when a student feels he/she has learn a significant amount in a
course they may tend to rate highly all aspects of the course. Could application of LM
process steps be the reason students perceive themselves as learning more in a given
course? Again, research question three will address this in detail.
Individual mean scores for perceived learning were averaged for each semester
for both T1 and T2. This was done to study the average fluctuation of perceived learning
scores by semesters. Table 6 shows the mean perceived learning scores for T1 and T2 by
semesters. As previously defined, perceived learning scores are based on a scale from a
negative two to a positive two. A negative two indicates that all students participating in
the end-of-course evaluation felt they had learned a “great deal less” when compared
with other courses. A combined evaluation score of negative one indicates that students
felt they had learned “a little less” in this particular course as it compares to other courses
they have taken. A zero score indicates students perceived that they had learned “about
the same” as compared to other courses they have taken. A positive one perceived
learning score suggests students felt as if they had learned “a little more” than other
courses taken. Lastly, a score of positive two would mean all students felt they had each
learned “a great deal more” when compared to other college courses taken.
70
Table 6
ANOVA: Student Perceived Learning Scores for T1& T2 Semesters
Evaluation Instrument
Semester
Perceived Learning
Standard
Score
Deviation
N
Employed
Fall 2007
33
1.01
.20
Old Evaluation
Winter 2008
33
0.99
.18
Instrument (T1)
Spring 2008
33
1.13
.27
Fall 2008
33
1.03
.26
Winter 2009
33
1.03
.28
New Evaluation
Spring 2009
33
1.05
.20
Instrument (T2)
Fall 2009
33
1.04
.25
Winter 2010
33
1.02
.22
*Indicates a significance level at or less than .05 between semesters
Notice that the mean scores for student perceived learning is close to positive one,
plus or minus a tenth of a point. What this suggests is that for the 33 courses being
evaluated the average student perceptions of their learning was “a little more” than other
college courses taken. However, no noticeable pattern of an increase in perceived
learning scores are apparent in this initial ANOVA for T1 & T2.
The differences between the Fall 2007 semester and the Fall 2008 semester is
very slight. There seemed to be a spike in perceived learning scores during the Spring
2008 semester, but then scores seem to level off around 1.04. No significance was found
between the differences in perceived learning scores for either T1 or T2. Likewise, the
71
differences in mean scores between Fall 2007 and Winter 2010 were such that almost no
difference was noticed. The significance level was found to be .904, a value close to one,
indicating little to no change.
In summary, the findings show that no significant change in student perceived
learning was noticed for the 33 courses measured. The average mean perceived learning
score indicated that, in general, student felt that they had learned “a little more” in the
courses being studied compared to other college course they had taken. Interestingly, the
findings for research question one indicated that a significant change in course ratings
was realized for this same population being studied. Yet, research question two revealed
no significant change in student perceived learning during the same time period. It seems
that course ratings have been affected by some outside factor, but student perceived
learning has remained unchanged, regardless of the change in course evaluation
instruments. To what degree has the LM contributed to the change in these variables?
The next section will explore the strength of correlation between LM scores and student
perceived learning scores.
Research Question 3: The Relationship between LM and Perceived Learning Scores
The purpose in asking the third research question is to learn how closely LM
scores correlate with student perceived learning scores. If a correlation exists between
these variables one could assume that the independent variable could help predict the
outcome of the dependent variable. In other words, if evidence shows a strong
correlation between variables, then LM scores could be used to predict the general
outcome of student perceived learning scores. Furthermore, one could extrapolate from
72
this assumption that when instructors are effective in implementing LM process steps in a
course an increase in student perceived learning will be the outcome, generally speaking.
For this study, LM scores act as the independent variable and student perceived learning
represent the dependant variables measured.
In addition, by measuring the strength of correlation between LM scores and
student perceived learning scores the implications of applying LM process steps begin to
be revealed. The finding for this section can help answer questions like the following:
“Do student perceived learning scores improve as LM scores improve?” or “When LM
scores are high in a given class do students perceive to learn more in these classes
compared to course where a low LM score was recorded?”
The first step in learning the strength of correlation between LM scores and
perceived learning scores was to compare the means for each variable; this was
completed in the previous two sections. For this research question all data points found
in T2 were used, totally 5,696 courses over the four semester span. These data points
were disaggregated to control for class size, semesters in which the course was taught,
and the instructor’s employment status.
Table 7 shows mean perceived learning scores for each semester based on
instructor’s employment status. Notice the increase in mean perceived learning scores
for CFS instructors when compared to 1YR, ADJ, and VET for each of the four
semesters represented. On average, CFS instructor’s mean perceived learning scores
were found to be at least .15 points higher than 1YR instructor’s mean perceived learning
score and .21 points higher than VET instructor’s scores. Also, notice the standard
deviation for CFS is lower than the other three variables with an average of .46 points.
73
Less variation from the mean perceived learning score was found for CFS instructor’s
than for 1YR, ADJ, and VET. VET instructor’s had the highest deviation from the mean
perceived learning score with an average deviation of .59 points.
74
Table 7.
Mean Perceived Learning Score by Semester Based on Instructor’s Employment Status
Semester
Fall
2009
Spring
2009
Winter
2009
Winter
2010
Mean
Instructor’s Employment
Status
1 Year Appointment
Adjunct
Continued Faculty Status
Veteran
Total
1 Year Appointment
Adjunct
Continued Faculty Status
Veteran
Total
1 Year Appointment
Adjunct
Continued Faculty Status
Veteran
Total
1 Year Appointment
Adjunct
Continued Faculty Status
Veteran
Total
1 Year Appointment
Adjunct
Continued Faculty Status
Veteran
Total
Mean
.93
.83
1.09
.87
.94
.91
.94
1.07
.86
.96
.88
.89
1.10
.89
.96
.99
.85
1.07
.86
.93
.93
.88
1.08
.87
.95
N
122
311
488
564
1485
111
233
458
402
1204
102
405
533
578
1618
128
306
404
551
1389
463
1255
1883
2095
5696
Std. Deviation
.53
.54
.45
.58
.54
.51
.53
.46
.63
.55
.53
.53
.48
.59
.55
.45
.52
.47
.55
.52
.50
.53
.46
.59
.54
Next, the researcher likewise compared mean LM scores by semester based on
instructor’s employment status (see Table 8). Similar results were noticed with CFS
instructors having the highest mean LM score per semester over 1YR, ADJ, and VET
instructors. The difference between average LM scores was not as stark of a contrast
75
when compared to perceived learning scores. On average, only .03 points separated the
top score from the bottom score. A similar pattern was noticed in standard deviation, as
CFS instructors showed lowest deviation from the mean of .06 points and VET
instructors had the highest standard deviation score of .10 points.
76
Table 8.
Mean LM Scores by Semester Based on Instructor’s Employment Status
Semester
Fall
2009
Spring
2009
Winter
2009
Winter
2010
Mean
Instructor Employment
Status
1 Year Appointment
Adjunct
Continued Faculty Status
Veteran
Total
1 Year Appointment
Adjunct
Continued Faculty Status
Veteran
Total
1 Year Appointment
Adjunct
Continued Faculty Status
Veteran
Total
1 Year Appointment
Adjunct
Continued Faculty Status
Veteran
Total
1 Year Appointment
Adjunct
Continued Faculty Status
Veteran
Total
Mean
.83
.83
.85
.82
.83
.84
.84
.85
.81
.83
.82
.83
.85
.82
.83
.84
.83
.85
.83
.84
.83
.83
.85
.82
.83
N
122
311
488
564
1485
111
233
458
402
1204
102
405
533
578
1618
128
306
404
551
1389
463
1255
1883
2095
5696
Std. Deviation
.07
.09
.06
.10
.08
.06
.07
.07
.12
.09
.06
.09
.06
.10
.09
.05
.10
.07
.07
.08
.06
.09
.06
.10
.08
A similar pattern was noticed for both perceived learning scores and LM scores
when controlling for instructor’s employment status. For each semester, CFS instructors
77
scored the highest for both perceived learning scores and LM scores. Likewise, VET
instructors had the lowest score for most of the semesters. Standard deviation scores
were low for CFS and high for VET instructors. Compared with VET instructors, in
general CFS instructors will have a perceived learning score that is .101 points higher
(see Appendix VII). Does this trend hold true when perceived learning scores and LM
scores are disaggregated by class size? Does class size affect this trend?
Perceived learning scores were categorized by semester and class size to compare
the differences in means (see Table 9). The results of the comparison show a difference
in mean perceived learning scores between small classes and very large classes. Small
sized classes have the highest perceived learning score, whereas very large sized classes
tend to have the lowest perceived learning scores. However, perceived learning scores
tend to have the highest deviation from the mean and very large classes have the lowest
variance from the mean score, generally speaking.
78
Table 9.
Mean Perceived Learning Scores by Semesters Based on Class Size
Semester
Fall 2009
Spring
2009
Winter
2009
Winter
2010
Mean
Class Size
Mean
N
Std. Deviation
0-10
1.04
102
.74
11-20
21-30
Over 30
Total
0-10
11-20
21-30
Over 30
Total
0-10
11-20
21-30
Over 30
Total
0-10
11-20
21-30
Over 30
Total
0-10
11-20
21-30
Over 30
Total
1.0
.96
.87
.94
1.13
1.05
.97
.88
.96
1.12
.96
.97
.89
.96
1.02
1.01
.94
.87
.93
1.10
1.02
.96
.88
.95
263
417
703
1485
126
238
308
532
1204
240
332
449
597
1618
97
268
435
589
1389
565
1101
1609
2421
5696
.60
.51
.48
.54
.67
.50
.53
.53
.55
.74
.59
.47
.47
.55
.65
.64
.52
.42
.52
.71
.59
.51
.48
.54
LM mean scores were also compared by semester based on class size (see
Appendix VIII). No differences in mean LM scores were noticed between large,
medium, and small classes, each scoring on average .84 points. Very large class sizes
showed a slight drop in mean LM score with an average score over the four semesters of
.82 points. For the most part, there were virtually no differences in LM scores from one
79
semester to another. In other words, mean LM scores for class size based on the semester
year in which the course was taught showed little to no difference between categories or
semesters.
Preliminary evidence shows that perceived learning scores tend to be affected by
class size. Yet LM scores show little to no effects from the change in class size from one
semester to the next. Likewise, LM scores tend to be similar regardless of a small,
medium, or large class. There seems to be some variability for very large classes,
however. Very large classes show a drop in LM scores for most of the semesters in
which it was taken.
The last section for question three addresses the relationship between the LM and
student perceived learning. This portion of the third research question is most revealing
and telling about the effectiveness of the LM and the importance of its use in the
classroom. Implications could be significant if students feel they learn more when the
LM process steps are applied throughout a course.
Again, this section is broken down into three areas of study in an attempt to
evaluate the strength of correlation between the LM and student perceived learning, given
different variables. First an overview of the average correlating values will be looked at,
followed by a breakdown of variables as previously mentioned.
The effects of class size. Class size can have an effect on both LM scores and
student perceived learning scores, as show previously. The following section will look at
the correlating effects of class size and evaluate how the difference in class sizes might
affect the strength of correlation between the two variables. Table 10 shows how average
LM scores, perceived learning scores, and class size correlate one with another.
80
Table 10.
Correlation Matrix of LM Scores and Perceived Learning Scores by Class Size
Average LM
Perceived
Class Size
Scores
Learning Scores
Average LM Scores
1
Perceived Learning Scores
.518
1
Class Size
-.091
-.142
1
Each of the correlating values was found to be significant at the .05 level. In
other words, evidence suggests when LM scores increase perceived learning scores do
likewise, generally speaking. Another significant finding was noted in this correlation
matrix regarding class size. Notice the negative correlation between class sizes and both
LM scores and perceived learning scores, respectively. This shows that when class sizes
are increased, LM scores and perceived learning scores decrease. Remember, from
earlier findings very large sized classes tend to have the lowest LM scores and perceived
learning scores when compared to other sized classes. This begs the question, could
using the LM in very large sized courses produce greater change in perceived learning
scores compared with other sized classes?
The effects of semester enrolled. This next section focuses on the effects course
fulfillment status may have on LM scores and student perceived learning scores. As
mentioned from previous sections, semesters are divided into four separate categorizes;
81
Winter 2009, Spring 2009, Fall 2009, and Winter 2010. Table 11 shows the correlation
between LM scores and student perceived learning scores by semesters.
Table 11.
Correlation Matrix of LM Scores and Perceived Learning Scores by Semester
Semester
Winter 2009
Spring 2009
Fall 2009
Winter 2010
Variables
LM Scores / Perceived Learning
Score
LM Scores / Perceived Learning
Score
LM Scores / Perceived Learning
Score
LM Scores / Perceived Learning
Score
LM Scores /
Perceived Learning
Scores
.61
N
1485
.56
1204
.62
1618
.64
1389
To summarize this section of the study regarding the semester when students
enrolled, the pattern of a positive correlation between LM scores and student perceived
learning scores are consistent throughout. Similarly to previous sections of this study, a
correlation seems to be present between the two variables when compared one to another.
Courses taken during the Winter 2010 semester showed the strongest correlation between
LM scores and student perceived learning scores. In other words, the LM scores of those
students enrolled in a course during the Winter 2010 semester could be used to predict
perceived learning score more accurately than the other three semesters. However, this is
not conclusive. Not every winter semester will bring similar result. Likewise, in each
semester, the correlation values are positive and fairly strong. In other words, the
82
semester in which a course was taken seems to have little to no affect on the strength of
correlation between LM scores and perceived learning scores.
The effects of instructor employment status. The last portion of this section,
regarding the relationship between LM scores and perceived learning scores, looks at the
effects that instructor employment status might have on the strength of correlation
between these two variables. In other terms, will the strength of correlation between LM
scores and perceived learning scores be affected by whether the instructor is tenured
(VET), working towards tenure (CFS), adjunct (ADJ) or hired on as a temporary oneyear appointment (1YR)?
The first to be considered are instructors who are working towards tenure (CFS).
It is assumed that most of instructors in this category are working towards and are hoping
to receive tenure status. One requisite in receiving tenure at BYU-Idaho are good student
evaluations. Likewise, instructors who use good teaching strategies, like those inherent
in the LM, will have more of the advantage receiving tenure status. Table 12 represents
the relationship between LM scores and student perceived learning scores for each
respective instructor. Notice again the strong correlation between the two variables in a
positive direction. This value was calculated to be at the .000 significance level, meaning
that the probability of this correlation between LM scores and perceived learning scores
happening by chance is less than .001%. Therefore, not only does this tell us that a
positive correlation exists between these two variables, but that the correlation itself is
significant.
83
Table 12.
Correlation Matrix of LM Scores and Perceived Learning Scores by Instructor
Employment Status
Instructor
Employment
Status
1YR
ADJ
CFS
VET
Variables
LM Scores / Perceived Learning
Score
LM Scores / Perceived Learning
Score
LM Scores / Perceived Learning
Score
LM Scores / Perceived Learning
Score
LM Scores /
Perceived Learning
Scores
.76
N
463
.67
1255
.60
1883
.55
2095
It is interesting to see that the strongest correlation between LM scores and
student perceived learning was found for 1YR instructors. The lowest correlation
between the two variables was found to be with VET instructors. In other words,
instructors on a one year appointment have a better chance of predicting perceived
learning scores based on LM score results over courses taught by VET instructors. This
might be due to teacher reputations. A VET instructor may already be known to
incoming students. This previous bias may play a role in the difference in strength of
correlation between these variables.
The findings for employment status seem to be consistent in showing a positive
correlation between LM scores and student perceived learning scores. Some correlations
are stronger than others, but all seem to be positively correlated, meaning LM scores are
good predictors of student perceived learning, regardless of class size, semester enrolled,
or instructor employment status.
84
Summary of Findings
The findings for each of the three research questions reveal differing results. For
research question one, findings show a significant increase in end-of-course ratings from
the Fall 2007 semester and the Fall 2008 semester. Likewise, a significant increase in
end-of-course ratings from the Fall 2007 and Winter 2010 semesters were measured for
the 33 courses studied. Yet, no significant difference in mean end-of-course ratings were
noticed between the Winter 2009 and Winter 2010 semesters, the semesters reflective of
the new course evaluation instrument.
For question two, no significant change in perceived learning scores was revealed
for the 33 courses under investigation. Somewhat surprising was the fact that little to no
difference in student perceived learning scores was measured between the Fall 2007 and
the Winter 2010 semesters. Overall, student perceived learning scores stayed relatively
constant at a rating of around 1.04.
The third research question was a study of the relationship between LM scores
and student perceived learning scores for semesters when the new course evaluation
instrument was used to collect student data. Furthermore, the findings were
disaggregated into class size, semester enrolled in course, and instructor employment
status. In all cases, a significant correlation was found between these variables.
The question remains, what do these findings mean? Chapter 5 will focus on the
possible implications and future impact these findings may have on curriculum
development, student study habits, course design, among other related areas.
85
CONCLUSIONS
The purpose in conducting this study was to determine if the implementation of
LM principles and process steps has affected (either negatively or positively) course
scores and/or student perceived learning scores as found on end-of-course evaluations.
Furthermore, the researcher inquired to know what effects class size, semesters enrolled,
and instructor employment status had, if at all, on the strength of correlation between LM
student perceived learning, respectively. Data was collected, based on course evaluations
from the Fall 2007 through Winter 2010 semester. The data were disaggregated to
control for the different variables, as previously mentioned, and statistical analyses were
employed to reveal the findings. This Chapter will summarize the revealed findings and
draw conclusions related to the implementation of the LM and the associated affects on
course scores and student perceived learning. Also, possible implications of these
findings along with further research possibilities and suggestions are explored.
Results of Findings
The purpose in asking research question one was to learn if end-of-course scores
had changed beginning with the Fall 2007 semester to the Winter 2010 semesters. The
difference between the two semesters is roughly two and ½ years. Shortly following the
Fall 2007 semester the LM was formally introduced to students and faculty as a campuswide model to follow. The researcher wanted to learn if the implementation of the LM
has had an effect on course scores over time.
The second research question was asked in an attempt to learn if student
perceived learning scores changed from the Fall 2007 to the Winter 2010 semesters. The
86
claim for implementing the LM at BYU-Idaho was to increase student learning and
understanding of the material being presented in class. If students perceived themselves
as learning more from one semester to another then perhaps something might be
contributing to this change. This was the intent in asking the second research question.
The third research question looks specifically at the semesters after the course
evaluation instrument was changed to measure the level at which the three process steps
of the LM were being employed by the instructor and students. This question was asked
in an attempt to evaluate the strength of correlation between LM scores and student
perceived learning scores. Furthermore, data were disaggregated down to subcategories
of class size, semester when taken, and instructor employment status. These
subcategories were evaluated to see what effects they might have on the strength of
correlation between LM scores and perceived learning scores.
Change in course ratings over time. For research question one, the researcher
first looked at course ratings of those classes that were taught by the same instructor for
the Fall 2007 semester to the Winter 2010 semester. The researcher used a comparison of
means and ANOVA to measure the differences in mean course ratings for the 33 courses
identified as matching the criteria previously determined for this question. A two-paired
t-test was used to measure the significance between mean scores for T1, T2, and the Fall
2007 and Winter 2010 semesters. This analysis revealed a significant difference between
the mean course ratings for the Fall 2007 semester compared with the Winter 2010
semester. This finding suggests that the change in mean course ratings from Fall 2007 to
the Winter 2010 semester did not happen by chance. In other words, during the two and
½ year time frame between the Fall 2007 and Winter 2010 semesters something occurred,
87
or a variety of things occurred, to bring about the change in end-of-course scores from
one semester to the other.
Remember, the course evaluation instrument was changed in the Winter 2009
semester. For this reason, the collection of data was divided into two parts, T1 and T2.
T1 represents the semesters in which the old evaluation instrument was used to collect
student data (Fall 2007 – Fall 2008), and T2 represented the semesters in which the new
course evaluation instrument was employed for the same reason (Winter 2009 – Winter
2010). The findings revealed a significant difference in mean course ratings for T1, but
not for T2. Also shown, was the fact that each consecutive semester’s mean course rating
improved for both T1and T2 time periods. With the implementation of the LM occurring
after the Fall 2007 semester, one may question “how has the LM process steps attributed
to this increase in course ratings?” Definite conclusions cannot be made at this point
regarding the cause of the increasing in course ratings during T1 through to the end of
T2. However, evidence shows that course ratings have been increasing, generally
speaking, from the Fall 2007 to the Winter 2010 semesters.
The change in mean course ratings for the Fall 2008 and the Winter 2009
semester was also measured. The purpose in doing so was to learn what affects the new
course evaluation instrument may have had on end-of-course ratings for the 33 courses
being evaluated. The finding revealed no significant change in course ratings between
the two semesters. Likewise, the increase in course ratings from the Fall 2008 to the
Winter 2009 semester was equal to the increases of previous semesters. What this
suggests is that the change in evaluation instrument seemed to have had little to no affect
on course ratings for the 33 courses studied. With this in mind, the researcher expanded
88
the study to span both T1 and T2. In other words, with no significant change in course
ratings between the Fall 2008 and Winter 2009 semesters, the researcher conducted a ttest on the mean differences in course ratings between the Fall 2007 and the Winter 2010
semesters.
The differences between mean course ratings between Fall 2007 and Winter 2010
had a positive difference, meaning that the average scores increased from one semester to
the other. On average, the scores did not drop between the two semesters. In fact, the
average mean scores for this group of instructors increased significantly from the
beginning semester to the end semester in this study. This is important in learning that
the implementation of the LM did not have a negative effect on course scores, generally
speaking.
In conclusion, for the 33 courses evaluated course ratings seemed to have
improved between the first semester and the last semester of this study. Still the question
remains: Could the LM be a reason for the increase? One of the probable causes of this
increase may be due to the application of LM process steps by instructors, but that
conclusion cannot be made at this time. Research question three addresses the strength of
correlation between these two variables in question, and will be addressed later in this
section.
The change in perceived learning over time. Findings from research question
one shows a definite change in course ratings from the semester before the
implementation of the LM to the Winter 2010 semester, representing the most recent data
available. Likewise, perceived learning scores were evaluated using the same
methodology as was used to address research question one. However, results for
89
perceived learning ratings for the 33 courses evaluated revealed a much different
outcome than was found for course ratings. No significant changes in perceived learning
scores were noticed for the 33 courses evaluated. T1 and T2 showed no significant
difference in perceived learning scores. In fact, the correlating value approached 1,
meaning the perceived learning scores between semesters were nearly the same.
Likewise, when the study was expanded to include the beginning semester (Fall 2007) to
the last semester in this study (Winter 2010) a similar result was realized. For the 33
courses measured, no significant differences in perceived learning scores were found.
Perhaps the reason for the relative similarity in perceived learning scores between
semesters over time is due to the structure of the question. The question used on both
evaluation instruments to measure perceived learning scores asks, “Compared to other
college courses you have taken, would you say that you have learned…” then students
select one of the following; “much less”, “a little less”, “about the same”, “a little more”,
or “much more”. The findings show that the mean perceived learning score is close to
one, which correlates with the response, “a little more”. In other words, for the 33
courses evaluated, students felt that, “compared to other college courses [they] have taken
… [they] have learned a little more.” One issue with this question is that students are
basing their answers on other college course they have taken. Does this give a real
representation of what students have learned in the particular class being evaluated or is it
merely a reflection of how much more or less a student has learned in a particular course
compared with other college courses taken? What if a student is evaluating a course in
which he/she has no experience in? Would the perceived learning score be higher in this
course compared with a course in which the student has a fundamental background
90
knowledge. For example, a student who is a native Chinese speaker may feel she learns
less in an Introduction to Chinese course compared with a non-Chinese speaking student,
simply due to past experiences. Certainly, perceived learning scores would be affected in
this scenario. A more accurate measure of perceived learning scores would be to
measure students’ competency level at the beginning of a course and again at the
conclusion of the same course.
Another interesting finding noticed when comparing mean perceived learning
scores with mean course ratings per semester was the fact that both scores seemed to
follow a similar trend. What was noticed was those courses which had the highest
perceived learning scores also had the highest course ratings. Likewise, those courses
that scored lowest in perceived learning also has the lowest course ratings. This
preliminary evidence shows that course ratings and perceived learning scores are
correlated. In other words, when course ratings go up it is likely that perceived learning
scores would also improve. Research question three would measure the strength of this
correlation.
In conclusion, the findings show no significant change in perceived learning
scores for the 33 courses evaluated during T1 or T2. However, the fact that perceived
learning scores measured close to a one for each respective semester suggests that in
general they felt they had learned “a little more” in these courses compared to other
courses they had taken. Does this suggest that these courses are at a higher level than
other courses? There are still too many variables to say what factors help predict student
perceived learning scores. More research is needed related to the variables that affect
91
student perceived learning scores. However, research question three of this study
examined the effects of LM scores on this variable and others.
The relationship between the LM scores and perceived learning scores. The
third research question looks particularly at the relationship or correlation between LM
scores and student perceived learning scores. The purpose in evaluating the correlation
between these two variables is to learn if an instructor whom implements the process
steps of the LM would realize a higher student perceived learning score. The findings
point to the fact that there is a correlation between LM scores and student perceived
learning scores.
In an attempt to measure the strength of correlation between each factor, LM
scores and perceived learning scores for T2 were disaggregated to class sizes, semesters
when the courses were taught, and instructor employment status. In each case, the
findings showed a positive correlation between LM and student perceived learning
scores. This suggests that the higher the LM score is for an instructor the higher their
student perceived learning score will be. The correlation between LM scores and
perceived learning scores remained strong regardless of class size, semesters when course
was taken, or instructor employment status.
Furthermore, when data were disaggregated to class size a difference in perceived
learning scores were realized between small and very large classes. Small classes
showed a composite mean perceived learning score of 1.10. Compare this score to the
composite mean perceived learning score for very large sized classes, measuring .88, and
one can quickly notice a significant difference between the two scores. In other words,
students perceive themselves as learning more in small sized classes than they do in very
92
large sized classes, generally speaking. This is congruent with Krueger’s (1999) study of
gains in test scores for a large population of students, as mentioned in Chapter 3.
Also noted was the difference in LM scores between small and very large sized
classes. Small sized classes had a higher LM score than did courses that were considered
to be very large in size. This difference between small and very large classes tells us one
of two things. Either, those instructors who are teaching small classes are implementing
and practicing the LM more effectively than instructors of very large classes, or
instructors of very large classes simply are not using the LM in their respective classes.
A correlation matrix was employed to measure the strength of correlation and
direction between LM scores, perceived learning scores, and class size. Evidence was
found suggesting that LM scores and perceived learning scores are positively correlated.
The strength of this correlation was found to be .518, a fairly strong result. In other
words, when LM scores increase student perceived learning scores will likely increase as
well. This is a significant finding considering initial literature suggests that instructors
are resistant to changing their teaching methodology for fear of lower course ratings and
student perceived learning scores (Bergstrom, 2009). This is evidence of just the
opposite. When the LM is practiced more in a course, chances are course ratings and
student perceived learning will improve.
Also noted from the correlation matrix measuring LM scores, perceived learning
scores, and class size, were the effects class size had on LM scores and perceived
learning scores. A negative correlation was found between these variables, meaning
evidence points to the fact that when class size increases LM scores decrease. Likewise,
when class size increases perceived learning scores decrease as well. In other words, the
93
bigger the class size the lower the LM score and perceived learning score will be,
generally speaking.
However, the strength of correlation for class sizes showed growth between LM
scores and perceived learning scores as the size of classes increase (see Appendix IX).
The larger the class size the stronger the correlation between the two variables. In other
words, the confidence in predicting perceived learning scores based on LM scores would
be higher for very large sized classes compared to small sized classes, generally
speaking. In fact, very large sized classes revealed a very strong correlation between LM
scores and perceived learning scores of .71. What this suggests is that in courses of very
large size, instructors who have a high LM score will likely have high student perceived
scores as well. Since this correlation between the two variables is so strong, one must ask
if instructors of very large sized classes would experience the most improvement in
course ratings and student perceived learning by simply using more of the LM process
step in their teaching methodology. This inquiry demands further study and investigation
to give a definite answer to this question.
Interestingly, when data were disaggregated to semester, no significant changes
were found between LM scores and perceived learning scores. In other words, it is
irrelevant when a course is taken as it relates to LM scores and perceived learning scores.
The time of year when a course was taken showed no differences between LM scores and
perceived learning scores, respectively. Fall, Winter, and Spring semesters all had
similar scores.
Conversely, significant changes in LM scores and perceived learning scores were
found when data were disaggregated to employment status. CFS instructors showed the
94
highest mean LM and perceived learning scores, whereas VET instructors showed the
lowest respective scores. Remember, CFS instructors are those who are hired fulltime
but not yet tenured. BYU-Idaho’s tenured track system is a four year process. VET
instructors represent that portion of the faculty who are tenured and have been employed
at BYU-Idaho for a minimum of four years. It’s not surprising to learn that those
instructors who are working towards tenured status are implementing the LM at a higher
level than those instructors who are not. However, it is somewhat discouraging to find
that VET instructors have the lowest perceived learning scores compared to the rest of the
faculty population. One might assume that an instructor who is tenured and has been
teaching for at least four years would have at least comparable student perceived learning
scores with their colleagues. But, evidence shows that VET instructors have significantly
lower LM and perceived learning scores compared with CFS instructors.
One factor that may be contributing to the differences in LM and perceived
learning scores between VET and CFS instructors could be due to variance in scores.
CFS instructors showed the lowest deviation from mean for both LM and perceived
learning scores. Conversely, VET instructors were found to have the highest deviation
from the mean for both variables. What this reveals is the fact that more CFS instructors
are receiving scores that are close to the mean score than VET instructor. In other words,
more VET instructors are receiving a wider range of scores than are CFS instructors.
More VET instructors are receiving very low LM and perceived learning scores
compared to CFS instructors.
However, when comparing the strength of correlation between LM and perceived
learning scores for each of the four categories of employment status, 1YR instructors
95
received the strongest correlation value of .76. VET instructors measured the lowest
correlation strength of .55. Both of these correlation values are strong, suggesting that
LM scores are a good predictor of perceived learning scores. The difference in
correlation strength between each employment category is interesting. Perhaps one
reason for the difference in correlation strength is due to reputation? Notice, those
instructors who are hired as a one-year appointment are relatively unknown to student
prior to their teaching assignment. Compare them to VET instructors who have been
teaching for at least four years and have likely built some type of reputation with
students. An instructor who is unknown to the student and is without prior reputation
seems to have the strongest chance of improving student perceived learning scores by
implementing more of the LM. However, more research is needed in this area to make a
conclusive statement.
To conclude, evidence suggests that when the LM is practiced at a higher level of
implementation in the classroom student perceived learning scores will likely improve,
generally speaking. Also, regardless of class size, the semester when the course is taken,
or the employment status of an instructor, when LM scores are high student perceived
learning scores will likely also be high.
Possible Implications
Scott Bergstrom, BYU-I’s Institutional Researcher, writes, “We are aware that
some instructors are concerned that their course evaluations might be adversely affected
by incorporating elements of the Learning Model into their teaching.” Then he adds,
“The new course evaluation instrument, which has a strong Learning Model orientation,
96
gives us some preliminary information that might allay this concern” (Bergstrom, 2009).
Bergstrom found that in courses where students percieved the highest occurance of LM
applications, instructors likewise received equally high marks. Similarly, this expanded
study of more than 5,600 courses also corroborate Bergstroms findsings. This suggests
that instructors who effectively impliment the LM process steps may not see a drop in
their course ratings nor in student perceived learning scores.
The implications of these findings may help alleviate concerns regarding
instructors changing teaching techniques to mirror LM practices. In turn, if instructors
realize the positive effects of implementing LM practices in their teaching curriculum,
student learning will be enhanced and the level of learning will increase campus wide.
In this study, the researcher compared data previously collected from the old
course evaluations with data collected from the new course evaluation instrument,
presently being used on campus. What was discovered was that, in general, the
implementation of the LM has had no ill effects on course ratings or student perceived
learning. What’s more, the findings point to evidence that shows an increase in course
ratings and student perceived learning when the level of implementation of the LM is
increased. The implications of such a study could be useful for administrators,
instructors, and students alike as they consider what type of teaching methodology could
best facillitate learning in the classroom.
On a broader scale, researching the significance of implimenting a model for
teaching and learning in the arena of higher education could have far reaching
implications. Results from this study could be used in developing an accountability
model for students, much like how the end-of-course evaluations are designed in this
97
study. Institutions with a desire to emphasis the importance of students taking more
responsibility for their own learning could also profit from this study.
Fears regarding large scale implimentation of a learning model at a University
with similar demographic may be alliviated as a result of these findings. What’s more,
the findings in this study show a strong correlation between teaching practices that
encourage active participation and engagement of students and perceived student learning
scores. Meaning, students learn more when they actively prepare for and participate in
class. The finding is not new, but the way to encourage students and faculty to apply
good teaching practices is. The process steps inherent in the LM at BYU-Idaho
incorporates elements of sound, effective teaching strategies. Another University or
college could, in effect, incorporate the idea of a model for learning and teaching—
founded in effective teaching strategies. This study highlights one method of tracking
good teaching through the use of course evaluations. Having a LM score reflective of the
level of active participation in and implementation of effective teaching can be benefitial
to other orgainizations.
Suggestions for Future Research
This study has revealed many avenues for future study. As the study progressed
over time, additional questions arose and caused the researcher to take note. Some of the
possible future research ideas involve specific techniques while other questions are
general in nature. This study was the first attempt at measuring the effects of the LM at
BYU-Idaho. A more detailed look into specific teaching techniques, instructor skills, and
student demography, along with many other variables are still waiting to be explored.
98
One possible research idea might include a qualitative analysis of an effective
instructor and an ineffective instructor and determining the differences between the two.
Also, one might focus their study on an instructor that has a low LM score but has high
course ratings, instructor ratings, and high student perceived learning scores and compare
their teaching style to an instructor who scores low on their course ratings, instructor
ratings, and student perceived learning scores although their LM score is high.
Another possible action research idea would be to introduce this LM to an
instructor who’s teaching technique is more of a traditional style and measure whether
their end-of course ratings, instructor ratings, and student perceived learning scores
increase as the instructor increases the level of implementation of the LM. This could
also be extended to a group of instructors or even to another University.
Again, there are many avenues that one could pursue in conducting further
research in this area of study. The findings of this study are significant enough that they
do demand further study and investigation. After all, the purpose in education is to
further knowledge and increase learning. This study suggest that there is evidence in
doing so, which should cause all true educators to sit up and take notice.
99
REFERENCES
Bailey, C.M., Hsu, C.T., & DiCarlo, S.E. (1999). Educational puzzles for understanding
gastrointestinal physiology. Advanced Physiology Education, 21, 1–18
Beichner, R.J., Saul, J.M., Abbott, D.S., Morse, J.J., Deardorff, D.L., Allain, R.J.,
Bonham, S.W., Dancy, M.H., & Risley, J.S. (2007). The Student-Centered
Activities for Large Enrollment Undergraduate Programs (SCALE-UP) Project.
In Research-based reform of introductory physics, ed. E. Redish. Available online
by searching PER Central at www.compadre.org/per.
Bergstrom, S. (2009, March). Learning and Teaching. Retrieved March 20, 2009, from
Brigham Young University-Idaho:
http://web.byui.edu/LearningAndTeaching/post/2009/03/The-Impact-of-LearningModel-Implementation-on-Student-Course-Evaluations.aspx
Blumenfeld, P., Krajcik, J.S., Marx, R.W., & Soloway, E. (1994). Lessons learned: How
collaboration helped middle grade science teachers learn project-based
instruction. The Elementary School Journal, 94, 539–551.
Bonwell, C.C. & Eison, J. (1991). Active Learning: Creating Excitement in the
Classroom. ED 340 212.
Borman, W.C., Hanson, M.A., Oppler, S.H., Pulakos, E.D., & White, L.A. (1993). Role
of Early Supervisory Experience in Supervisor Performance. Journal of Applied
Psychology, 78(3), 443-449.
100
Boyle, J.R. (2001). The Effects of Strategic Notetaking on the Recall and
Comprehension of Lecture Information for High School Students with Learning
Disabilities. Learning Disabilities Research & Practice, 16(3), 133-141.
Brockway, G., McNeely, S., & Neffennegger, P. (1992). A point-counterpoint classroom
exercise. Marketing Education Review, Summer, 34-36.
BYU-I. (2007, September). LM. Rexburg, Idaho, USA.
Chan, V., Pisegna, J.M., Rosian, R.R., & DiCarlo, S.E. (1991). Construction of a model
demonstrating neural pathways and reflex arcs. Advanced Physiology Education,
271, 14–42.
Clark, K. B. (2009). About BYU-Idaho. Retrieved April 22, 2009, from BYU-Idaho:
http://www.byui.edu/aboutbyuidaho/
Chickering, A. W., & Gamson, Z. F. (1987). Seven Principles for Good Practice. AAHE
Bulletin 39: 3-7. ED 282 491. 6 pp. MF-01; PC-01.
Cochran, Leslie H. 1989. Administrative Commitment to Teaching. Cape Girardeau,
Mo.: Step Up, Inc.
Cortright, R.N., Collins, H.L., & DiCarlo, S.E. (2005). Peer instruction enhanced
meaningful learning: ability to solve novel problems. Advanced Physiology
Education, 29, 107–111.
Cortright, R.N., Collins, H.L., Rodenbaugh, D.W., & DiCarlo, S.E. (2003). Student
retention of course content is improved by collaborative-group testing. Advanced
Physiology Educational, 27, 102–108.
Crouch, C.H. & Mazur, E. (2001). Peer Instruction: Ten Years of Experience and
Results. American Journal of Physics, 69 (9), 970-977.
101
Dev, C.S. (1990). Measuring the Value of Experiential Learning. The Cornell H.R.A.
Quarterly. August: 105-07.
Dewey, J. (1916). Democracy and Education. New York, NY: Macmillan.
Draper, S.W. & Brown, M.I. (2004). Increasing Interactivity in Lectures Using an
Electronic Voting System. Journal of Computer Assisted Learning, 20, 81-94.
Edelson, D.C., Gordin, D.N., & Pea, R.D. (1999). Addressing the challenges of inquirybased learning through technology and curriculum design. Journal of the Learning
Sciences, 8, 391–450.
Felder, R.M., & Brent, R. (1996). Navigating the bumpy road to student-centered
instruction. College Teaching, 44(2), 43–47.
General History of BYU-Idaho. (n.d.). Retrieved November 10, 2009, from
http://www.byui.edu/PR/General/BYUIHistory2.htm
Gilbert, C.G., Hunsaker, S., & Schmidt, B. (2007). Peer Instruction: Faculty as
Architects of Peer Learning Environment. Retrieved March 19, 2010, from
http://www.byui.edu/perspective/v7n2pdf/v7n2_gilbert_et_al.pdf
Goyak, A.. The effects of cooperative learning techniques on perceived classroom
environment and critical thinking skills of preservice teachers. Ed.D. dissertation,
Liberty University, United States -- Virginia. Retrieved July 9, 2009, from
Dissertations & Theses: Full Text.(Publication No. AAT 3352768).
Hake, R.R. (1998). Interactive Engagement Versus Traditional Methods: A SixThousand-Student Survey of Mechanics Test Data for Introductory Physics
Courses. American Journal of Physics, January, 64-74
102
Hartley, J., & Davies, I. (1978). Note Taking: A Critical Review. Programmed Learning
and Educational Technology, 15, 207-224.
Hunt, T., Lederman, L., Kreps, G., & Ruben, B. (1985). Instructional simulation helps
public relations students to understand roles in organizational communication.
ERIC, ED 259 365.
Hyman, Ronald T. (1980). Improving Discussion Leadership. New York: Columbia
Univ., Teachers College Press.
Johnson, D. W., & Johnson, R. T. (1999). Making cooperative learning work. Theory
Into Practice, 38(2), 67-73.
Johnson, D. W., Johnson, R. T., & Smith, K. (1998). Cooperative learning returns to
college. Change, 30(4), 26-35.
Kagan, D. M. (1992). Professional Growth Among Preservice and Beginning Teachers.
Review of Educational Research, 62,129-169.
Karabenick, S. & Collins-Eaglin, J. (1997). Relation of Perceived Structural Goals and
Incentives to College Students’ Use of Learning Strategies. Journal of
Experimental Education, 65, 331-341.
Kellogg, R.T. (2008). Training Writing Skills: A Cognitive Developmental Perspective.
Journal of Writing Research, 1(1), 1-26.
King, A., Staffieri, A., & Douglas, A. (1998). Mutual Peer Tutoring: Effects of
Structuring Tutorial Interaction to Scaffold Peer Learning. Journal of
Educational Psychology, 134-152.
Krueger, A.B. (1991). Experimental Estimates of Education Production Functions. The
Quarterly Journal of Economics, 114, (2), 497-532.
103
Kuh, G. D. (2001). Assessing What Really Matters to Student Learning: Inside the
National Survey of Student Engagement. Change, 33, 10-17.
Kuh, G.D. (2003). What We’re Learning About Student Engagement from NSSE:
Benchmarks for Effective Educational Practices. Change, 35, 24-32.
Kuipers, J.C. & Clemens, D.L. (1998). Do I dare? Using role-play as a teaching strategy.
Journal of Psychosocial Nursing Mental Health Services, 36, 12–17.
Laws, P., Sokoloff, D., & Thornton, R. (1999). Promoting Active Learning Using the
Results of Physics Education Research. UniServe Science News, 13.
LM. (2008, January). Retrieved April 20, 2009, from BYU-I:
http://www.byui.edu/learningmodel/src/default.htm
Lowman, Joseph. 1984. Mastering the Techniques of Teaching. San Francisco: JosseyBass.
Lujan, H. L. & DiCarlo, S. E. (2006). Too much teaching, not enough learning: what is
the solution? Advances in Physiology Education, 30, 17-22.
Lymna, F. (1981). The responsive classroom discussion. In: Mainstreaming Digest,
edited by Anderson AS. College Park, MD: Univ. of Maryland College of
Education.
Mallor, J.P., Near, J.P., & Sorcinelli, M.D. (1981). Increasing student involvement in
large lecture courses: two case studies. Journal of Experiential Learning and
Simulation, 3(4), 191-196.
Marx, R. W., Blumenfeld, P. C., Krajcik, J. S., Fishman, B., Soloway, E., Geier, R., &
Tal, R. T. (2004). Inquiry-Based Science in the Middle Grdes: Assessment of
104
Learning in Urban Systemic Reform. Journal of research in Science Teaching,
41 (10), 1063-1080.
Marx, R.W., Blumenfeld, P., Krajcik, J.S., & Soloway, E. (1997). Enacting project-based
science. Elementary School Journal, 97, 341–358.
Matlin, M.W. (2003). Cognition. Hoboken, New Jersey, (101), 176-178.
Mazur, E. (1997). Peer Instruction: A User’s Manual. Upper Saddle Valley, New Jersey:
Prentice Hall.
Mazur, E. (1997). Peer instruction: getting student to think in class. In: The Changing
Role of Physics Departments in Modern Universities, edited by Redish EF and
Rigden JS. Buenos Aires, Argentina: Proceedings of ICUPE 399, 981–988.
McDermott, L.C. (1993). How we teach and how students learn. Ann NY Acad Sci 701,
9–20.
McKeachie, W.J. (1994). Learning and Cognition in the College Classroom. Teaching
Tips: Strategies, Research and Theory for College and University Teachers.
Lexington, MA.
McKeachie, Wilbert J., Paul R. Pintrich, Yi-Guang Lin, and David A.F. Smith. 1986.
Teaching and Learning in the College Classroom: A Review of the Research
Literature. Ann Arbor: Regents of The Univ. of Michigan. ED 314 999. 124 pp.
MF-01; PC-05.
Michael, J.A. (1993). Teaching problem solving in small groups. Ann NY Acad Sci, 701,
37–48.
Mierson, S. (1999). Skits and games to enhance students' learning of physiology.
Advanced Physiology Education, 277, 283–284.
105
Moy, J.R., Rodenbaugh, D.W., Collins, H.L., & DiCarlo, S.E. (2000). Who wants to be a
physician? An educational tool for reviewing pulmonary physiology. Advanced
Physiology Education, 24, 30–37.
Odenweller, C.M., Hsu, C.T., & DiCarlo, S.E. (1998). Educational card games for
understanding gastrointestinal physiology. Advanced Physiology Education, 20,
78–84.
Penner, J.G. (1984). Why Many College Teachers Cannot Lecture. Springfield, Ill.:
Charles C. Thomas.
Piaget, J. (1928). The Development of Thought: Equilibration of Cognitive Structure.
New York, NY: Basic Books.
Prince, M.J. (2004). Does Active Learning Work? A Review of the Research. Journal of
Engineering Education, 93(3), 223-31.
Prince, M.J & Felder, R.M. (2006). Inductive Teaching and Learning Methods:
Definitions, Comparisons, and Research Bases. Journal of Engineering
Education, 95(2), 123-138.
Rao, S.P., Collins, H.L., & DiCarlo, S.E. (2002). Collaborative testing enhances student
learning. Advanced Physiology Education, 26, 37–41.
Rao, S.P. & DiCarlo, S.E. (2000). Peer instruction improves performance on quizzes.
Advanced Physiology Education, 24, 51–55.
Rao, S.P. & DiCarlo, S.E. (2001). Active learning of respiratory physiology improves
performance on respiratory physiology examinations. Advanced Physiology
Education, 25, 55–61.
106
Rodenbaugh, D.W., Collins, H.L., & DiCarlo, S.E. (1999). Construction of a model
demonstrating cardiovascular principles. Advanced Physiology Education, 277,
67–83.
Ruhl, Kathy L., Charles A. Hughes, and Patrick J. Schloss. Winter 1987. "Using the
Pause Procedure to Enhance Lecture Recall." Teacher Education and Special
Education 10: 14-18.
Rydalch, B.J. (2005). Inauguration set for BYU-Idaho President Kim B. Clark. BYUIdaho News Release, October 3. Retrieved from
http://www.byui.edu/News/NewsReleases2005/051003Innauguration.htm
Sapon-Shevin, M., & Schniedewind, N. (1992). If cooperative learning's the answer,
what are the questions? Journal of Education, 174(2), 11-37.
Scannapieco, F.A. (1997). Formal debate: an active learning strategy. Journal of Dental
Education, 61, 955–961.
Scott, R.G. (2005). To Understand and Live Truth. Address, CES Religious Educators at
the Jordan Institute of Religion, February 4.
Silberman, M. (1996). Active Learning: 101 Strategies to Teach Any Subject.
Philadelphia, PA: Allyn and Bacon.
Silverthorn, D.U. (1999). Using demonstrations to uncover student misconceptions: the
Law of LaPlace. Advanced Physiology Educational, 277, 281–282.
Slavin, R. E. (1995). Cooperative Learning. Boston: Allyn and Bacon.
Slavin, R.E. (1983). When does cooperative learning increase student achievement?
Psychology Bull, 94, 429–445
107
Smith, K.A., Sheppard, S.D., Johnson, D.W., & Johnson, R.T. (2005). Pedagogies of
Engagement: Classroom-Based Practices. Journal of Engineering Education,
January, 1-15.
Stronge, J. H. (1997). Improving schools through teacher evaluation. In J. H. Stronge
(Ed.), Evaluating teaching: A guide to current thinking and best practices (p. 123). Thousand Oaks, CA: Corwin Press, Inc.
Thurgood, L. & Gilbert, C. (2008). What is Unique about the BYU-Idaho Learning
Model? Learning and Teaching Website. Retrieved from
http://web.byui.edu/learningandteaching/post/2008/10/What-Is-Unique-about-the-BYU-Idaho-Learning-Model.aspx
Vygotsky, L. (1978). Mind in Society: The Development of Higher Psychological
Processes. Cambridge, Massachusetts: Harvard University Press.
Wankat, P. (2002). The Effective Efficient Professor: Teaching, Scholarship and Service.
Allyn and Bacon: Boston, MA.
Wiggins, G. (2004). Assessment as feedback. Retrieved May, 2009 from
http://www.newhorizons.org/strategies/assess/wiggins.htm.
Wolf, P.. Fostering critical reflection in higher learning: The self study of a practitioner
and coach. M.Ad.Ed. dissertation, St. Francis Xavier University (Canada),
Canada. Retrieved July 9, 2009, from Dissertations & Theses: Full
Text.(Publication No. AAT MR27250).
108
APPENDIX I - INITIAL RESEARCH FINDING BY SCOTT BERGSTROM
Using questions from the course evaluation instrument, I created five scales to
correspond to the five elements of the LM architecture – Outcomes, Architecture,
Prepare, Teach One Another, and Ponder/Prove. Space does not permit me to describe
each of these scales so I will use the Teach One Another scale as an example. Student
responses to seven questions from the instrument were used to create the Teach One
Another scale:
o I was
an active participant in online or face-to-face class discussions.
o I sought
o I feel
opportunities to share my learning with others outside of class.
that I made important contributions to the learning and growth of fellow
classmates.
o The course provided
o Group
opportunities to learn from and teach other students.
work, if assigned, was beneficial and meaningful.
o Students
were actively involved in this class through discussions, group work, and
teaching.
o The instructor
provided appropriate opportunities to be an active participant in
the class.
The scale score derived from these seven questions serves as an indicator of the extent to
which this element of the LM is present in a course.
Using the scores on the Teach One Another scale, I divided all courses that were
evaluated in Fall 2008 into four quartiles representing the degree to which the Teach One
Another element of the LM was present in the course. The first quartile represents
109
courses in the lowest 25% of the Teach One Another scores, the second represents the
next 25% of courses, and so on. I computed the average overall instructor rating1 and the
average level of perceived learning2 for each quartile group of courses. The results of
this analysis are in Figures 1 and 2.
Figure 1. The average overall instructor rating for each quartile of the Teach One Another
scale.
110
Figure 2. The average level of perceived learning for each quartile of the Teach One
Another scale.
As shown in Figure 1, the overall instructor rating rises steadily from the 1st to
the 4th quartile. The difference between the bottom and top quartiles is 1.34 points, a
rather large (and statistically significant) difference. As shown in Figure 2, the level of
perceived learning rises sharply from the 1st to the 2nd quartile and then steadily upward
thereafter. The top quartile average level of perceived learning is more than three times
larger than the bottom quartile average!
The pattern seems clear for both measures. The higher the presence of the Teach
One Another element in a course, the higher the overall instructor rating and the greater
the level of perceived learning. My analysis showed that this same pattern holds true for
all five elements of the LM: the greater the presence of a LM trait in a class, the
higher the overall instructor rating and the higher the level of perceived learning.
This is a very preliminary analysis and more work needs to be done before we can
draw firm conclusions. In the first place, this [is] a correlational analysis. We do not
know if there is a causal relationship at work. Do LM activities make effective teachers
or are effective teachers more likely to use LM activities? Second, there may be other
factors and traits that are more influential on the overall instructor rating or on the level
of perceived learning. Is there a halo effect in play here wherein teachers who are
popular and courses that are well-liked yield high scores on all questions no matter
what? We simply do not know from this preliminary analysis how influential LM traits
are on these selected measures. The presence of such a pronounced effect connected with
111
LM practices compels further examination. Third, the power of the indicators being used
remains to be determined. For example, do the seven items that make up the Teach One
Another scale really provide a good measure of the presence of that trait in a class? Are
there better indicators of the Teach One Another trait?
In the spirit of the scholarship of teaching and learning, I am hopeful that some
instructors will experiment with a well-defined LM practice and see if there are any
changes in the course evaluations or, more importantly, in overall student learning and
growth. For example, several members of the Physics faculty have conducted pre- and
post-test student assessments comparing the impact of different peer- and problem-based
teaching innovations. Such experiments may provide us with compelling evidence of
LM effects, beyond what we can derive from course evaluation data.
At the very least, it is exciting that this first empirical look at LM practices
implies a positive relationship with the overall instructor rating and the level of perceived
learning. It also suggests that, in general, teachers will not be penalized by students for
using elements of the LM in their classes. (Bergstrom, 2009)
112
APPENDIX II – OLD EVALUATION INSTRUMENT
BYU-IDAHO COURSE EVALUATION
Please evaluate the following instructor and course. If you feel an item is not applicable to your
course, then leave it blank. When you are finished, click on the SUBMIT button at the bottom of the
page. Your identity is completely anonymous. Please be as thorough and as accurate as possible. Your
feedback is highly valued. It is used by your instructor and the school's administration to improve
teaching.
Instructor:
CHECKETTS, MAX L
Course :
REL 233
CHURCH HISTORY
Section:
9
Overall Rating: CHECKETTS MAX L - REL 233
very poor
poor
fair
good
very good excellent
exceptional
What is your overall rating of this instructor.
What is your overall rating of this course .
Items about the Course: CHECKETTS MAX L - REL 233
Strongly
Disagree
1
Course objectives are clear.
2
Course is well-organized.
3
Student responsibilities are clearly defined.
4
Course content is relevant and useful.
5
Assigned workload is appropriate for credit hours.
6
Assigned homework is not just busywork.
7
T ext(s) and other materials have helped me
understand course content.
8
Exams concentrate on important points of the course.
9
Exams are clearly worded.
10
Exams are good measures of my knowledge,
understanding, or ability to perform.
11
Grading procedure is fair and impartial.
12
Assignments are appropriately distributed
throughout the course of the semester.
13
Course as a whole has produced new knowledge,
skills, and awareness in me.
113
Disagree
Somewhat Somewha
Disagree t Agree
Agree
Very
Strongly
Strongly
Agree
Agree
Items about the Instructor: CHECKETTS MAX L - REL 233
Strongly
Somewhat Somewha
Disagree
Disagree
Disagree t Agree
1
Has an excellent knowledge of the subject matter.
2
Is enthusiastic about the subject.
3
Is well prepared for each class.
4
Makes good use of class time.
5
Gives clear examples and explanations.
6
Makes helpful feedback of my work (e.g., papers,
exams).
7
Clearly explains difficult concepts, ideas, or theories..
8
Responds respectfully to student questions and
viewpoints.
9
Is genuinely interested in helping me understand
the subject matter.
10
Is available to students during regular and
reasonable office hours.
11
Motivates me by his/her example to want to learn
about the subject.
12
Has produced new knowledge, skills, and
awareness in me.
13
Starts/dismisses class at scheduled times.
14
Seldom misses class.
Agree
Very
Strongly
Strongly
Agree
Agree
Agree
Very
Strongly
Strongly
Agree
Agree
Items about Core Values:CHECKETTS MAX L - REL 233
Somewhat Somewha
Strongly
Disagree
Disagree t Agree
Disagree
1
Appropriately brings Gospel insights and values
into secular subjects.
2
Inspires students to develop good character.
3
Helps students prepare to live effectively in
society.
Is spiritually inspiring insofar as the subject
matter permits.
4
Other Information: CHECKETTS MAX L - REL 233
a gre at
de al le ss
a little
less
about
the
same
a little
more
a gre at
deal
more
5
6
7
8
9
Compared to other college courses you have taken, would you say that you
have learned . . .
T he approximate number of hours
per week that I have spent in
outside preparation for this class is
...
1
2
3
114
4
ge ne ral
educatio
n
e le ctive
other
major
minor
90% to
100%
75% to
90%
50% to
75%
le ss
than
50%
neve r
atte nde
d
B
C
D
F
O the r
T his course fills requirements for my . . .
My class attendance has been . . .
A
T he grade I expect from this course. . .
Is there anything about this course and/or instructor that was particularly good? If so, what?
What could be done to improve this course to help you learn more?
115
APPENDIX III – NEW EVALUATION INSTRUMENT
BYU-IDAHO COURSE EVALUATION
Please evaluate the following instructor and course. If you feel an item is not applicable to your
course, then leave it blank. When you are finished, click on the SUBMIT button at the bottom of the
page. Your identity is completely anonymous. Please be as thorough and as accurate as possible. Your
feedback is highly valued. It is used by your instructor and the school's administration to improve
teaching.
Instructor:
CHECKETTS, MAX L
Course :
REL 233
CHURCH HISTORY
Section:
9
Overall Rating: CHECKETTS MAX L - REL 233
very poor
poor
fair
good
very good excellent
exceptional
What is your overall rating of this instructor.
What is your overall rating of this course.
Items about Student Performance in this Course:
Strongly
Disagree
1
I was prepared for each class.
2
I arrived at class on time.
3
4
5
6
I was an active participant in online or face-toface class discussion
I sought opportunities to share my learning with
others outside of class.
I worked hard to meet the requirements of this
class.
I sought opportunities to reflect on what I had
learned in the class.
7
I feel that I made important contributions to the
learning and growth of fellow classmates.
8
T he course as a whole had produced new
knowledge, skills, and awareness in me.
116
Disagree
Somewhat Somewha
Disagree t Agree
Agree
Very
Strongly
Strongly
Agree
Agree
Items about the Course:
1
Course objectives are clear.
2
Course is well-organized.
3
Student responsibilities are clearly defined.
4
Instructional resources--textbook(s), course
guide(s), online material, etc--were useful and
helped me to achieve course objectives.
5
Assessment activities--exams, quizzes, papers,
hands-on demonstrations, presentations, etc.-accurately and fairly measured the knowledge and
abilities I acquired from the course.
6
Class assignments contributed to my learning and
growth.
7
T he course provided opportunities to learn from
and teach other students.
8
Group work, if assigned, was beneficial and
meaningful.
9
Students were actively involved in this class
through discussions, group work, and teaching.
Strongly
Disagree
Disagree
Somewhat Somewha
Disagree t Agree
Agree
Strongly
Disagree
Disagree
Somewhat Somewha
Disagree t Agree
Agree
Very
Strongly
Strongly
Agree
Agree
Items about the Instructor:
1
T he instructor effectively modeled problemsolving and application of subject matter.
2
3
T he instructor
made good
of class time.
When
given, examples
anduse
explanations
were
clear.
4
T he instructor gave helpful feedback of my work.
5
constructively to student questions and
viewpoints.
6
T he instructor was available to me when I
requested assistance, in class or outside of class.
7
8
9
10
11
T he instructor motivated me by his/her
enthusiasm to want to learn about the subject.
T he instructor starts/dismisses class at scheduled
times.
T he instructor held me accountable for coming to
each class prepared.
opportunities to be an active participant in the
class.
T he instructor provided opportunities to reflect
upon my learning and experiences in the class.
117
Very
Strongly
Agree Strongly
Agree
Items about Core Values:
Strongly
Disagree
1
Appropriately brings Gospel insights and values
into secular subjects.
2
Inspires students to develop good character.
3
Helps students prepare to live effectively in
society.
Is spiritually inspiring insofar as the subject
matter permits.
4
Disagree
Somewhat Somewha
Disagree t Agree
Agree
Very
Strongly
Agree Strongly
Agree
Other Information:
a great
de al less
a little
less
about
the
same
a little
more
a great
deal
more
a great
de al less
a little
less
about
the
same
a little
more
a great
deal
more
5
6
7
8
9
Compared to other college courses you have taken, would you say that you
have learned . . .
Compared to other college courses you have taken, would you say that your
satisfaction is . . .
1
2
3
4
per week that I have spent in
outside preparation for this class is
...
general
educatio
n
elective
major
minor
90% to
100%
75% to
90%
50% to
75%
le ss
than
50%
never
attende
d
B
C
D
F
O ther
other
T his course fills requirements for my . . .
My class attendance has been . . .
A
T he grade I expect from this course. . .
118
APPENDIX IV – CORRELATION MATRIX FOR COURSE RATINGS
Fall
2007
1
Winter
2008
Winter
2008
.87
1
Spring
2008
.60**
.76**
1
Fall
2008
.70**
.79**
.82**
1
Winter
2009
.79**
.84**
.78**
.82**
1
Spring
2009
.74**
.82**
.70**
.77**
.72**
1
Fall
2009
.65**
.75**
.71**
.78**
.80**
.76**
1
Winter
2010
.65**
.70**
.64**
.72**
.77**
.66**
.73**
Fall
2007
Spring
2008
**. Correlation is significant at the 0.01 level (2-tailed).
119
Fall
2008
Winter
2009
Spring
2009
Fall
2009
Winter
2010
1
APPENDIX V- TWO SAMPLE T-TEST MEASURING SIGNIFICANCE
Mean Differences in Course Ratings
Period
Semester
Mean
Std.
Std. Error
Deviation
Mean
N
Compared
Fall 2007
5.46
33
.67
.11
Fall 2008
5.68
33
.64
.11
Winter 2009
5.72
33
.67
.11
Winter 2010
5.75
33
.51
.08
Between T1
Fall 2008
5.68
33
.64
.11
& T2
Winter 2009
5.72
33
.67
.11
Semester 1
Fall 2007
5.46
33
.67
.11
&8
Winter 2010
5.75
33
.51
.08
T1
T2
120
APPENDIX VI – CORRELATION MATRIX FOR PERCEIVED LEARNING
Fall 2007
Fall Winter Spring Fall Winter Spring Fall Winter
2007 2008 2008 2008 2009 2009 2009 2010
1
Winter 2008
.87
1
Spring 2008
.70
.67
1
Fall 2008
.79
.83
.73
1
Winter 2009
.73
.81
.58
.84
1
Spring 2009
.64
.66
.59
.83
.73
1
Fall 2009
.55
.68
.51
.78
.83
.72
1
Winter 2010
.67
.68
.51
.71
.71
.68
.73
**. Correlation is significant at the 0.01 level (2-tailed).
121
1
APPENDIX VII – PARAMETER ESTIMATES COMPARING PERCEIVED
LEARNING SCORES WITH EMPLOYMENT STATUS
Parameter Estimates
Dependent Variable: Perceived Learning Score
Parameter
Intercept
Std.
B
Error
-2.13
.05
t
-39.71
Sig.
.00
95% Confidence
Interval
Lower
Upper Partial Eta
Bound
Bound
Squared
-2.23
-2.02
.21
3.65
.06
56.89
.00
3.52
3.77
.36
Employment
Status=1YR
.00
.02
.37
.70
-.03
.05
.00
Employment
Status=ADJ
-.04
.01
-3.08
.00
-.07
-.01
.00
Employment
Status=CFS
.10
.01
7.36
.00
.07
.12
.00
Employment
Status=VET
0a
.
.
.
.
.
.
LM Scores
a. This parameter is set to zero because it is redundant.
122
APPENDIX VIII – MEAN LM SCORES BY SEMESTER BASED ON CLASS
SIZE
Mean LM Scores by Semesters Based on Class Size
Semester
Class Size
Mean
F09
0-10
.82
11-20
.84
21-30
.84
Over 31
.83
Total
.83
S09
0-10
.86
11-20
.85
21-30
.83
Over 31
.82
Total
.83
0-10
.84
11-20
.84
W09
21-30
.84
Over 31
.82
Total
.83
0-10
.84
11-20
.84
W10
21-30
.84
Over 31
.83
Total
.84
0-10
.84
11-20
.84
Total
21-30
.84
Over 31
.82
123
N
102
263
417
703
1485
126
238
308
532
1204
240
332
449
597
1618
97
268
435
589
1389
565
1101
1609
2421
Std. Deviation
.18
.11
.06
.06
.08
.13
.06
.09
.09
.09
.17
.07
.07
.06
.09
.14
.10
.06
.06
.08
.16
.09
.07
.07
APPENDIX IX – LM SCORES COMPARED TO PERCEIVED LEARNING
SCORES BY CLASS SIZE
Correlation Matrix of LM Scores and Perceived Learning Scores by Class Size
Class Size
LM Scores
Small
LM Scores
Perceived Learning
Score
Medium
LM Scores
Perceived Learning
Score
Large
LM Scores
Perceived Learning
Score
Very Large LM Scores
Perceived Learning
Score
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
124
1
565
.50
.00
565
1
1101
.56
.00
1101
1
1609
.67
.00
1609
1
2421
.71
.00
2421
Perceived
Learning
Score
.50
.00
565
1
565
.56
.00
1101
1
1101
.67
.00
1609
1
1609
.71
.00
2421
1
2421
Download