Overview of Buffalo State's Evaluation Effort

advertisement
Evaluating Undergraduate Research:
Measuring Student Outcomes
Jill Singer, Buffalo State College
Bridget Zimmerman, Buffalo State College
Dan Weiler, Daniel Weiler Associates
UNDERGRADUATE RESEARCH AS TRANSFORMATIVE PRACTICE
13th National Conference June 19-22, 2010
Weber State University, Ogden, Utah
Outline
• Motivation for evaluating Buffalo State’s summer
research program
• Process we followed
• Findings
• Modifications to procedure and instruments
• Available resources
• Questions
Motivation: Why Evaluate the Undergraduate
Summer Research Program (USRF)
• All programs on our campus are required to have assessment plans
• The USRF is the largest single program administered by Buffalo
State’s Office of Undergraduate Research
• Program supports research and creative endeavors
• Each award includes a student stipend of $2500 (in 2010 increased
to $2750), faculty stipend of $1000, and $500 for project-related
travel and supplies
• Program length is 8 weeks
• USRF currently in its 12th year
• Program enjoys a high level of prestige and participants are very
supportive of our ongoing efforts to evaluate the impact of
undergraduate research
• Made sense to develop an assessment process that would apply
both to science and non-science disciplines
• Multi-year effort aimed at development and field-testing a
methodology for measuring student learning and related outcomes
(evaluation findings reported are for summers 2008 and 2009)
Purposes of the Evaluation
• Obtain reliable assessment of the USRF’s impact on
participating students that goes beyond student survey
data
• Obtain information from faculty mentors on how the USRF
program has influenced their teaching and view of student
capabilities
• Provide information to participating students that helps
them assess their academic strengths and weaknesses
• Begin the creation of a longitudinal database that can
provide impact data on a range of outcomes for different
types of students and a variety of academic disciplines
Process: Retreat  Pilot

Full Scale
• A two day retreat in summer 2006 attended by faculty
representing a range of disciplines (arts, humanities, social
sciences, and physical sciences) was held to reach consensus on
student learning outcomes to be measured
• In advance of the retreat faculty were asked to read relevant
papers
• First day of the retreat devoted to identifying and confirming
11 outcome categories
• Second day of retreat devoted to refining outcome categories
and identifying outcome components
• The retreat resulted in 11 outcome categories to be measured
and a detailed rubric that describes the specific components
for each outcome
Outcome Categories
Communication
Creativity
Autonomy
Ability to deal with obstacles
Practice and process of inquiry
Nature of disciplinary knowledge
Critical thinking and problem solving
Understanding ethical conduct
Intellectual development
Culture of scholarship
Content knowledge skills/methodology
…. and option for mentor and their student to identify and add more
outcomes
Outcome Categories
• Each outcome category includes list of specific outcome components.
• For “creativity” components are: brings new insights to the problem at
hand; shows ability to approach problems from different perspectives;
combines information in new ways and/or demonstrates intellectual
resourcefulness; and effectively connects multiple ideas/approaches.
• For “critical thinking and problem solving” components are: trouble-shoots
problems, searches for ways to do things more effectively and generates,
evaluates and selects between alternatives; recognizes discipline-specific
problems and challenges established thinking when appropriate;
recognizes flaws, assumptions and missing elements in arguments.
• For “ability to deal with obstacles” components are: learns from and is not
discouraged by set-backs and unforeseen events; shows flexibility and a
willingness to take risks and try again.
• This pattern is followed for all 11 outcome categories.
Process: Retreat  Pilot

Full Scale
• Six student-mentor pairs participated in pilot implementation
during summer 2007 of the evaluation effort
• Focus groups composed of student researchers and faculty
mentors led by Dan Weiler with Bridget Zimmerman taking
notes
• Based on findings, minor modifications made to instruments
and clarification of overall evaluation process
Process: Retreat  Pilot

Full Scale
• Full scale implementation of the evaluation effort for two
years (2008 and 2009) involving 37 students and 37
faculty mentors
• Mentors represented both first time and experienced (‘repeat’)
mentors familiar with the USRF program (some having mentored
students 5 or more times over the program’s 12 years)
• Evaluation data (qualitative and quantitative) compiled and
analyzed by Bridget Zimmerman
• Qualitative parts of the evaluation took advantage of the
already existing mid-summer and final report forms and the
quantitative portions used the newly developed assessment
instruments
Implementation and Data (1)
• Student completes a survey answering general and specific
questions designed to define a ‘starting point’. Questions cover
such areas as: motivation, knowledge and expectations, and their
understanding of their academic strengths and weaknesses.
• Survey has two purposes: (1) provide mentors with insights into student
knowledge and thought processes, as an aid to the mentors in completing
preliminary student assessments; and (2) provide students with structured
opportunity to explore goals, knowledge, and readiness for the program.
• Mentor reviews responses and meets with student to help them
learn enough to formulate preliminary assessment (intended to
overcome challenge of completing pre-research assessment with
limited knowledge about the student).
• Mentors provided with list of candidate questions for this purpose,
showing relationship between each candidate question and relevant
student outcomes.
Implementation and Data (2)
• Both student and mentor complete a pre-research assessment
and meet to share their responses.
• Instrument has 11 outcome categories
• Five-point scale (the student always, usually, often, seldom, or never
displays the particular outcome, or not applicable).
• Given that mentors may not have equally rich information about
their students across every outcome listed, asked to indicate for
each score their level of confidence (very, fairly, somewhat, not
terribly, or not at all confident).
• Student and mentor meet to compare assessments and discuss
reasons for any differences and help students obtain a more
realistic assessment of their strengths and weaknesses.
• Research begins.
Implementation and Data (3)
• Students keep a journal (electronic or paper) and mentors are
encouraged to keep notes about the student’s progress.
• Student and mentors complete mid-summer reports that include
several questions about progress, changes from the original
projects, and plans for the second half of the summer. It also
contains all the questions asked in the pre-research assessment
questionnaire.
• Both student and mentor meet to share their responses.
• Research continues.
• Student and mentors complete final reports at the end of the
summer that include questions about research findings, places
where students might present their work, mentoring process (for
faculty), and also includes the same assessment questions.
Findings (1)
• Evaluation focused on student and mentor responses to
assessments, mid-summer and final project reports, and journals
provided anecdotal information.
• Assessment survey repeated three times (pre-, mid-, and post-) and
contains a base of 68 questions (34 questions and 34 confidence
scores).
• Cronbach’s alpha used to determine how strongly associated the items
of the scale relate and how well items hold together. The mentor
survey (N= 70) had an overall coefficient alpha of .98. The student
survey (N=68 ) had an overall coefficient of .96. Both coefficients
suggest that there is high level of internal reliability.
• Factor Analysis/Principal Components Analysis (PCA) used to
determine if item structure could be characterized by smaller set of
underlying components and make sure all questions add value to
defining program’s impact on participants.
Findings (2)
• Thirty-four of the sixty-eight base questions are related to
confidence in score’s accuracy.
• Students did not seem to fully understand why they were being
asked to rate themselves on their ability to rate themselves. Based on
this finding, the confidence questions were dropped from the student
version of the surveys and replaced with a single question about
their certainty in responding at the end of the questionnaire.
• Mentors understood the rationale for indicating confidence. Their
degree of confidence increased with each survey and increasing
knowledge of their student. To reduce survey fatigue, the confidence
interval was retained for the faculty pre- and post-research
assessments, but dropped from the mid-summer report.
• To determine if program had impact on participants, pre-, midand post-responses were analyzed for significant differences
with Repeated Measures ANOVA.
Findings (3)
• Overall surveys were well received by students and mentors
and compliance was very high (reflects a number of reasons
related to the value of the USRF program).
• Open-ended responses (both students and mentors) are
thoughtful, thorough, and insightful
• Student responses in particular demonstrated significant growth in
disciplinary expertise from the pre- to the post-assessments.
Findings (4)
• Student findings:
• Tended to initially rate themselves slightly higher than their mentor
rated them.
• Reported growth on every survey item.
• Because students often ‘inflated’ their pre-assessment scores and later
scored themselves more realistically, the actual differences between
pre- and post-scores was less frequently significant.
• Mentor findings:
• Initially rated the students lower than the students rated themselves but
by the end of the program rated the students higher than the students’
self-reported scores (higher rate of significant differences).
• A few items with decreased scores indicated that the mentor initially
overrated their student and later learned that the student was not as
strong as originally scored (e.g., writing skills, contribution to discipline).
Findings (5)
• Student comments included remarks that highlighted:
• Improvements in knowledge of discipline, how scholarly activities take
place in the discipline, new methods and tools for conducting research,
and ability to interact and communicate on a scholarly level
• Contributions to the discipline and that they were making discoveries
that had value as well as contributing resources to future students
• Value of their participation in the USRF program and how it would
contribute to future endeavors (going to grad school, competing in job
market, networking, etc)
• Gained knowledge beyond the classroom, their peers that were not
conducting research, and exceeded their own expectations for what
they could accomplish
Findings (6)
• Mentor* comments included remarks that highlighted:
• Balance between ‘teacher’ and ‘mentor’ allowing student to explore,
make mistakes, and grow without constant presence (reflects stages of
mentoring)
• Value of working in a collaborative/collegial way with their student
• Program’s influence on teaching practice. In particular, spoke of desire
to increase opportunities for students to engage in research within a
course and also raised expectations about what students can
accomplish.
* While some mentors initially commented on the redundancy of completing a pre-,
mid-, and post-assessment, at the end of the process many recognized the value of
capturing the growth of their particular student. The reason for this in part stems
from seeing how the pre-, mid-, and post-assessments and subsequent conversations
with their student was invaluable to the student’s self-awareness of their growth,
strengths, and weaknesses and that the student found the process very instructive.
Modifications Made Based on Findings
• Clearer instructions about evaluation process provided to
students and mentors along with summary of rationale for
evaluation methodology and modifications to protocol
• Elimination of ‘confidence’ score from all three student
questionnaires and replaced by a single question at the end of
the survey that asked students how certain they were about
their skill level and responses and why
• Elimination of ‘confidence’ scores from mid-summer report for
mentor (retained in pre- and post-research assessments)
• Simplification of questionnaires so that optional outcomes
identified at end rather than after each outcome category
Resources
• http://www.buffalostate.edu/undergraduateresearch/x561.xml
Links to static versions of the survey instruments are
provided from this url
• CUR Quarterly article (Spring 2009) describes the approach
Download article from above url
Concluding Remarks
• Evaluation process is at a point where it is recognized by
students and mentors as a valuable part of participating in the
USRF program.
• Unanticipated benefit to the evaluation process is improved
mentoring (provided a mechanism for facilitating regular and
semi-structured interactions between student and mentor)
• Recognized by administration and has helped make a case for
continued growth in the program funding.
• Evaluation is ongoing (for summer 2010 twenty-four USRF
awards were made).
• Additional year of data will inform future modifications and
refinements to our instruments and process as well as yield data
documenting the program’s impact on students and mentors.
Questions and Comments?
Download