"Assessing the Effectiveness" Powerpoint

advertisement
Assessing The Effectiveness Of
Your Academic Advising Program
Tom Grites
Assistant Provost
The Richard Stockton College
Tom.Grites@stockton.edu
Audience Poll
What are your expected learning outcomes for this
webinar?
A.
B.
C.
D.
E.
F.
Why do we assess?
What do we assess?
Where do we begin?
What tools do I use?
Who defines our success?
What I need to do tomorrow.
Overview
•
•
•
•
•
Terminology and Rationale
Assessment as a Process
Focus on the Contexts
Non-Assessment
Summary
Terminology
• Assessment
• Evaluation
• Measurement (the tools)
Assessment
(re: academic advising)
“Assessment is the process through which we
gather evidence about the claims we are making
with regard to student learning and the
process/delivery of academic advising in order to
inform and support improvement”
(Campbell, 2008)
Uses/Contexts
Assessment – tends to be more related to
programmatic issues and outcomes
Evaluation – tends to be more related to people
(advisor) skills, performance, and outcomes
It’s OK to use evaluation as part of the
assessment process
Intentions (related to both)
• Formative – more associated with assessment;
includes a wider range of efforts; requires more
analysis; provides a broader perspective; focus
on improvement
• Summative – more associated with evaluation;
more limited effort; focus on “Does it work?” or
“How well was job performed?”
The Rationale
• “…a lack of assessment data can sometimes
lead to policies and practices based on intuition,
prejudice, preconceived notions, or personal
proclivities – none of them desirable bases for
making decisions”
• (Upcraft and Schuh, 2002, p. 20)
More Rationale
“In God we trust; all others bring data.”
“An ounce of data is worth a pound of
opinion.”
(Magoon, c. 1975)
Other Reasons
•
•
•
•
•
Accountability
Effectiveness
Accreditation
Trustees/Regents
Legislators
• Program Improvement (to monitor and
improve student success) – the most important
reason
The Assessment Process:
A Cycle
• Resources:
• Assessment of Academic Advising Package
(3 CDs available from NACADA via www.nacada.ksu.edu)
• Assessment of Academic Advising Institute
(Feb 12-14, 2014 Albuquerque, NM)
Getting Started: Identify Stakeholders
• Complete set of advising constituents (students, staff
and faculty advisors)
• Broad range of key offices (Registrar, Enrollment
Management, similar advising units, certain campus
referral resources, IR office)
• Critics, Antagonists, and Naysayers
• FYIs – Faculty Senate, Deans Council, Retention
Committee, others as appropriate
The Advising Hub
What Do We want To Know or Demonstrate as a
Result of Academic Advising?
• Focus on student learning
• Connect learning to mission, vision, values, goals in
your advising program
– How will your program contribute to student
learning?
– Who, what, where, when, how will learning take
place?
• Define measures of student learning
– Gather evidence, set levels of expected
performance
The Assessment Process/Cycle
• Alignment with institutional and unit missions
• Specify goals and/or objectives
• Identify the outcomes expected (student learning
and/or programmatic)
• Gather evidence (the measurements)
• Share findings, interpretations, and
recommendations
• Begin implementation and re-start the cycle
Mission/Purpose
• A working model…
• Academic advising is integral to fulfilling the teaching and learning
mission of higher education. Through academic advising, students
learn to become members of their higher education community, to
think critically about their roles and responsibilities as students, and
to prepare to be educated citizens of a democratic society and a
global community. Academic advising engages students beyond
their own world views, while acknowledging their individual
characteristics, values, and motivations as they enter, move through,
and exit the institution.
• (Preamble, Concept of Academic Advising, NACADA, 2006)
Goals/Objectives
(how we intend to achieve our mission)
These need to emanate from and reflect the nature of the unit to be
assessed (total institution, Advising Center and its clientele, College
Dean’s Office, etc)
Examples:
• To assist students to become independent and lifelong learners
• To assist students in understanding the relevance of the total
curriculum
• To assist students in making good decisions based on their own
evidence (e.g., selecting a major)
Identify Outcomes
Student Learning Outcomes – examples
• All students will select an appropriate major by the end of their
third semester.
• All students will become engaged in at least one co-curricular
activity each semester.
• All students will be able to identify and will select courses that
enhance their human capital.
• At least 30% of the students will choose to participate in a service
learning course.
• All (CC) students will be able to distinguish among the A.A., A.S.,
and A.A.S. degree programs
A Task For You…re: Course Selection
• How many courses are in your Catalog? (A)
• How many courses are required to earn a
degree from your institution? (B)
• What percentage of what your institution offers
do students actually take in order to earn a
degree? (B/A)
• Now, for each course a student takes, how many
are eliminated?
Outcomes (continued)
• Programmatic/Process Outcomes – examples
• As a result of our advising services, the
retention/persistence rate of first-year students will
increase by 10% in the next 3 years.
• As a result of our intervention strategies, the percentage
of students who are removed from academic probation
will increase by 10% in the next academic year.
• After two advising sessions, all students will come to
their future sessions with a degree audit already run and
with a plan for meeting outstanding requirements
Everybody’s Favorite
All students will be able to understand,
appreciate, and articulate the value of
general education.
Gather Evidence
Mapping the Experience (Maki, 2004)*
• Not all outcomes will necessarily occur as a
direct result of what we do as advisors, so we
need to know what other learning opportunities
exist in order for the students to meet our stated
goals/objectives.
• WHAT learning is to occur?
• WHERE might it be learned?
• By WHEN should it be learned?
*This process can also inform the kinds of evidence that
need to be gathered for appropriate assessment.
The Advising Hub
Types of Measurement and Data
Qualitative – open-ended survey questions; focus groups; indepth responses, but small N
Quantitative – descriptive, structured, numbers and statistics
from surveys, demographics, etc; limited content responses, but
large N
Direct – observations; recorded data; pre-post information
Indirect – perceptions, inferences, even “inclinations”
Use Multiple Measures!!!
Gather (Multiple) Evidence
• Satisfaction Surveys (OK, but not enough)
• Institutional Data (changes of major, drop/add transactions, grades
in gateway courses, retention and graduation rates, use of services
provided elsewhere, advisor : advisee ratios, NSSE, etc)
• Office Data (number of appointments vs. walk-ins, nature of
sessions, results of sessions, transcript analyses, other advisor
tasks/activities; “What did you learn?”)
• Focus groups
measure)
(of clients, of faculty advisors, others – a qualitative
• The Advising Syllabus* can inform what evidence should be
collected
*http://www.nacada.ksu.edu/Clearinghouse/AdvisingIssues/syllabus101.htm
http://intraweb.stockton.edu/eyos/page.cfm?siteID=123&pageID=42#syllabus
Share the Results
Tips…
• Be sure that the stakeholders you identified
earlier are informed throughout the process in
order to enable their support in the decisionmaking for implementation of your
recommendations.
• Academics have a preferred method of review,
so it makes sense to conform to their
expectations.
Sharing the Results
(Format and Content)
These elements are often best provided in a
standard research report or journal format…
 Purpose of the assessment project
 Method of data collection
 Results found
 Interpretation of the results
 Recommendations with timetable for and
anticipated cost of implementation
 Executive Summary or Abstract
How Results Will Inform Decision-Making
• Revise pedagogy or curriculum or policy/procedure
• Develop/revise advisor training programs
• Design more effective programming
– advising, orientation, mentoring, etc.
• Increase out-of-class learning opportunities
• Shape institutional decision making
– planning, resource allocation
Sample Implementation Recommendations
• Redesign the advising effort in the Orientation
Program
• Develop a peer advising/mentoring program
• Streamline office procedures
• Initiate proposals for policy changes
• Improve communication with other service
offices and personnel
• Request/Reallocate resources (human, fiscal,
and/or physical)
You Did It!!
• This will complete the assessment cycle, which
provides the evidence for change and
improvement.
• Completion of the cycle may also provide new
goals and objectives, new assessment
strategies and tools, and other aspects that will
be need to be included in beginning the next
cycle.
(See Darling, 2005 handout)
You’ve Earned a Break
Please take a few minutes to submit any questions
you may have at this point via the chat function.
Back to the Original Contexts
People…
Academic advising, as a teaching and learning
process, requires a pedagogy that incorporates the
preparation, facilitation, documentation, and
assessment of advising interactions. Although the
specific methods, strategies, and techniques may vary,
the relationship between advisors and students is
fundamental and is characterized by mutual respect,
trust, and ethical behavior.
(Concept of Academic Advising, NACADA, 2006)
NACADA Core Values
Academic Advisors are responsible
 to the individuals they advise
 for involving others, when appropriate, in the
advising process
 to their institutions
 to higher education in general
 to their educational community
 for their professional practices and for
themselves personally
Assessment (Evaluation) of Advisors
• SELECTION
• TRAINING
• EVALUATION
• RECOGNITION/REWARD
Selection of Academic Advisors
•
•
•
•
•
•
Use the best
Add from other resources/units
Target specific populations
Cross disciplinary lines
Develop mentors
Use other skills/expertise
Potential Pitfalls
Making a distinction
Faculty Advising (Programmatic;
Assessment)
Faculty Advisors (Personal; Evaluation)
Inappropriate Comparisons
Professional Academic Advisors
Peer Advisors
No Improvement Plan
Training
Faculty vs. Professional Staff
Advisors
• Too often all are expected or
required to advise, but also
teach, publish, seek grants, etc
– no selection
• Training ranges from near
nothing to perhaps a day or 2,
but usually only a few hours
• Evaluation is not systematic
• Recognition/Reward is very
limited in the tenure and
promotion process; mostly
intrinsic; can also be a reverse
structure (better = more)
• They are hired via a search
process and have specific job
descriptions – they are
selected
• Their training is systematic,
intentional, and ongoing; staff
development is expected
• They are evaluated through
annual performance reviews
• They are rewarded with salary
and benefits
ASSESSMENT (Evaluation)
• 37 %
OF ALL INSTITUTIONS HAD NO
PERFORMANCE EVALUATION MEASURES FOR
FACULTY IN THEIR ACADEMIC ADVISING ROLE
• 44 %
in 2 yr public institutions
• 25 %
in 4 yr public institutions
• 39 %
in 4 yr private institutions
(Habley, 2004)
PARAMETERS (faculty advisors)
• Faculty Contract
• List of Responsibilities
• Availability of Resources
• Assignment of Advisees
• Recognition/Reward
Tools for Assessment (and/or Evaluation)
Of Advisors
• Self evaluation
• Student surveys (locally designed)
• Survey of Academic Advising (ACT)
• Academic Advising Inventory (NACADA)
• Student Satisfaction Inventory (Noel-Levitz)
• NACADA Clearinghouse
Back to the Original Contexts
• Program…
• “…a lack of assessment data can sometimes
lead to policies and practices based on intuition,
prejudice, preconceived notions, or personal
proclivities – none of them desirable bases for
making decisions”
• (Upcraft and Schuh, 2002, p. 20)
Other Tools and Strategies
•
•
•
•
•
•
•
•
Satisfaction Surveys
Institutional Data
Office Data
Focus groups
The Advising Syllabus
External Reviews
CAS Standards
Others…
CAS Assessment Worksheet
An Economic Model
• Though not an outcomes-based model per se, this
approach to assessment is a functional analysis
based on the premise that every task an advisor
performs and every operation that an advising unit
conducts has some monetary value related to it.
• The analysis results in a comparison of the fiscal
expenditures required to perform the tasks to the
cost benefits as results.
• The model operates from the perspective of a threat
to the existence of an advising unit, function or
personnel. A quick example…
Determining Your Worth
• Identify every function the unit performs
• Identify all possible alternatives for each
function, if the unit was dissolved
• Determine the cost of those functions that
cannot be replaced and who would perform
them; estimates will sometimes be required
• Determine the cost of those functions that could
be eliminated
(In Markee and Joslin, 2011)
Where are the data?
Bill Gates – “colleges today know more about how
many kids attend basketball games and which
alumni give money than how many students
showed up for economics class during the week…”
(jn review of Academically Adrift).
Where are the Data?
Jeff Selingo – “Think about it. Before we buy a
car, we can find various measures on everything
from gas mileage to results of safety tests. We
can turn to objective sources to check
comparisons of similar vehicles and see which
cars hold their value over time. But when it
becomes to potentially one of the most expensive
purchases in a lifetime, the attitude from colleges
has always been that we should just trust them on
the quality of their product.” (p. 25)
What Are We Not Assessing…And Should We Be?
• Student expectations, intentions
• Whether advising strategies actually can be
attributed to different types of student success
(removed from probation, successful choice of
major, overcome a skills deficiency or harmful social
habit, etc)
• Retention and graduation rates of transfer
students
Expectations vs. Experience
Be Undecided
Change Majors
Fail a course
Extra time to complete degree
Drop out
Transfer institutions
Work while in school
Seek personal counseling
Need tutoring
Seek career guidance
(Habley 2011)
Expect
7%
12
1
8
1
12
36
6
15
5
Experience
20%
65-85
16
60
40
28
60
27
20
25
Non-Assessment (continued)
• Use and value of articulation agreements –
number of students who use them, are they
updated
• Currency of academic policies, e.g., course
repeats, course pre-requisite criteria,
drop/add/withdrawal processes, academic
warning, probation, and suspension policies
• Does advisor training result in better advising?
Summary
• Assessment is a process, not an event
• Collaboration and cooperation are necessary for
productive assessment to occur
• “An ounce of data is worth a pound of opinion”
(Magoon, c. 1975)
Avoid the N of 1 syndrome
• The purpose and results of assessment should always
be used for program and/or advisor improvement in
order to realize maximum student development and
success
Questions?
Tom.Grites@stockton.edu
Download