Assessment Activity Report, Program Review

advertisement
Assessment Activity Report
Spring 2005
Program Review
The first activity of the semester was to revisit the original cycling sheet and determine where
each program was in the review process (see Table 1). Meetings were held with two departments
and eight department/program chairs to assist them in meeting review expectations.
Concurrently, I have met with Steve Hora to discuss the process and expected outcomes of
program reviews as well as plan to complete the pending reviews.
The program review self-study consists of an executive summary, background information
(Mission, Curricular Goals/Program Student Learning Outcomes and Role in UH system and in
UH Hilo mission and strategic plans), program organization and performance, student learning
outcome assessment, current resources, program chair’s evaluation and a statement of future
goals. Based on the first completed program reviews, the Assessment Support Committee revised
the program review guidelines in an attempt to reduce the amount of qualitative narrative in the
description of program organization and performance and increase the amount of quantitative
data and analysis (see tables attached). However, the Congress did not mandate the changes and
instead added the following preamble:
The following is a suggested format for use in developing the self study report
during a program’s seven-year review. The intention of the format is to permit the
presentation of objective data in as clear and concise a form as possible in order to
enhance the readability of the report. Use of the proposed format should be
considered voluntary by the program undergoing review except for the BOR
required table. The Institutional Researcher will provide the data for the required
table. Any part, or all, of the proposed format may be incorporated into the
program’s review.
Beginning in February, Lynne Stamoulis, Randy Hirakawa and I met bi-weekly to analyze the
current guidelines and explore alternatives that might better meet faculty and administrative
needs. The emphasis was on simplifying the process while still providing data and analysis on
which to base program changes, increase or decrease the number of faculty positions and allocate
resources. We examined models from other universities that included indicators of efficiency:
the degree to which a program uses its current resources and manages enrollment to provide both
program specific and general education courses. In addition, we are concerned about the quality
of each academic program as demonstrated by faculty productivity and evidence that students
learning outcomes are being achieved.
Efficiency and Enrollment Management
Currently, the efficiency and enrollment data for programs are combined in the BOR required
data table with the addition of a graduation index (used by the CAS Dean and VCAA for
determining resource allocations). The table includes:
1
Required Data
Number of Majors
Number of degrees earned in major (or # of graduates)
Graduation index
SSH Taught (Student Semester Hrs) (Fall)
FTE Enrollment
Crossover data % own majors
% within college
% all others
Number of classes or sections offered
Student Registrations
Avg. class size
FTE Faculty (Fall)
Fall
Annual
(4*grads / majors)
LD
UD
Grad
Undergrad
(SSH/15)
Grad
(SSH/12)
Fall
Fall
Fall
I-2
I-3
I-4
I-5
FTE Student FTE Faculty Ratio
Majors per FTE Faculty
Budget Allocation
Cost per Credit Hour
Sources of Data for Each Indicator on the Required Data Table (Excluding the Graduation Index)
Indicator:
Data Source:
Management & Planning Support (MAPS) Reports (Institutional
Research Office) http://www.hawaii.edu/iro/maps.htm
Headcount Majors
Course FTE Enrollment
Fall Enrollment Report,
UH Hilo, Fall Semesters
Course Registration Report,
UH Hilo, Fall Semesters
Data given only for
programs under review.
Data not available for
specific degree program(s)
under review, so are given
at the department level.
2
Number of Classes
Course Registration Report,
UH Hilo, Fall Semesters
Data not available for
specific degree program(s)
under review, so are given
at the department level.
Average Class Size
Course Registration Report,
UH Hilo, Fall Semesters
Data not available for
specific degree program(s)
under review, so are given
at the department level.
Data given only for
programs under review.
Degrees Awarded
Degrees & Certificates Earned,
UH Hilo, Fiscal Years
SSH Taught
Course Registration Report,
UH Hilo, Fall Semesters
Data not available for
specific degree program(s)
under review, so are given
at the department level.
% SSH Taken by own
Majors
Academic Crossover Study, UH
Hilo, Computer Report 1419
Data given only for
programs under review.
% SSH Taken by Nonmajors
Academic Crossover Study, UH
Hilo, Computer Report 1419
Data given only for
programs under review.
Analytical FTE Faculty
Departmental Activity &
Instructional Workload Measures,
Data not available for
specific degree program(s)
under review, so are given
at the department level.
UH Hilo, Fall Sem.
Student-Faculty Ratio
Departmental Activity &
Instructional Workload Measures,
UH Hilo, Fall Sem.
Data not available for
specific degree program(s)
under review, so are given
at the department level.
Note that beginning FY
Instructional Unit Cost Study, UH 2000-01, data will no longer
Direct Instructional Cost Hilo,
be available
(Information Technology Services),
Fiscal Years
http://drue.its.hawaii.edu/uhexpend/
Institutional Research Office, University of Hawai‘i, March 2004.
3
There are a number of problems associated with the current table. First, the data are for Fall
semester only which may provide an inadequate and/or incomplete picture of program
efficiency. In addition, the systemwide Institutional Research Office Management & Planning
Support (MAPS) Reports no longer collect and make available all of the required data. While
this data provides an overview of performance in the fall semester, it doesn’t tell enough about
program organization.
Program Quality and Student Learning Outcome Assessment
Universities are being asked to be accountable for student learning. Instead of measuring inputs
like the number of journal subscriptions in the library or the number of faculty members with
terminal degrees, programs are being asked to demonstrate that students have gained skills and
knowledge as a result of program completion. During the first round of program reviews under
the new guidelines, each program was expected to present an assessment plan. These plans
would be implemented and provide student learning outcome data for the next seven year
review. For example, the English Program Self-Study stated:
Preparing to Assess the Major.
English 300, the introduction to the major,
was offered for the first time in Fall 2003, specifically for literature track majors.
In Spring 2004, English 303 Backgrounds to English Studies will be offered. The
two courses will be offered annually. In the second offering of Eng 300 in Fall
2004, students will develop learning portfolios to be reviewed by the faculty for
the purposes of program assessment.
However, programs have been slow to respond. One possible solution would be to ask programs
to either develop and implement an assessment plan or adopt a nationally normed test such as
Excelsior College Examinations (ECE) and the College Level Exam Program (CLEP). For
example, according to the Chemistry self-study:
Standardized ACS exams are utilized by the Department in all areas of the
program. These exams are made available through the ACS with national
norms. Our students have consistently been within these norms. The
General Chemistry norms are in the 60 percentile nationally and our
students are in that range for the mean of the exam. The norms for
Organic Chemistry are in the 57 percentile and our students have scored in
that range.
If other programs adopted such standardized tests, they would report the scores, analyze student
performance and make changes in the program based on the data.
In addition to student performance, an assessment of program quality should include
stakeholders such as program graduates, employers, and successful entrance into graduate
schools. The Office of Institutional Research conducts and reports graduating student, alumni
and community surveys. These can be customized to provide feedback to programs for the
4
purposes of program review. Dean Hirokawa has encouraged Division and Department chairs to
take advantage of this opportunity to use data that is already being collected and analyzed and to
participate more actively in formulating questions central to program improvement. Nationally
normed data is available through the National Survey of Student Engagement that is conducted
and reported each year. Programs are encouraged to compare their program to peer and
benchmark institutions accessible through the Office of Institutional Research.
Tasks Remaining
Collaboration with administration should continue. The focus should remain on developing the
least labor intensive and most simple system that will provide the necessary data for program
improvement and university decision-making. A faculty member needs to monitor, guide and
provide assistance to programs that are preparing self-studies. In addition, professional
development opportunities need to be provided for new faculty members and those responsible
for program review. These should include workshops and individualized instruction.
5
Proposed Tables in the Voluntary Guidelines
Efficiency and Enrollment Management Data Tables
List Degrees (include tracks, options, and areas of specialization), certificates, minors, etc.
offered by your program on the table below. (Please asterisk core courses that every major must
take in order to meet major requirements.)
Specific Courses
Required in the Major
Tracks, Options,
Areas of
Specialization
Certificate Courses
Minor Course
Requirements
List GE courses provided by your program. (See
http://www.uhh.hawaii.edu/uhhl genedfac/courseproposalreview. php for information about the
UH Hilo GE program.)
Course
Hallmark
Frequency of
offering (Sem., Yr.,
Infreq.)
Enrollment
List service courses your program provides for other programs. For example, the English
program provides Eng 225 Writing for Science and Technology for science majors. Include
cross-listed courses where your program provides the course for the cross-listed program.
Service Course
Program
Frequency
Average
Enrollment
List student-faculty collaboration with dates during the past 7 years in the following
categories:
a. Directed Studies
b. Student-faculty research projects c. Student theses/ senior projects
c. Student theses/ senior projects
Show by flowchart or in lists with narrative the sequence of courses that would enable a
freshman to graduate with your major in four years. See http://cs.uhh.hawaii.edu for samples.
Year 1
Fall
Spring
Year 2
Fall
Spring
Year 3
Fall
Spring
Year 4
Fall
Spring
6
Please complete the following faculty productivity table. For a seven years since your last
program review, include the number of refereed publications (RP), book chapters (Ch), books
(B), other publications (OP), grants received (G - number/total dollar amount) for the past seven
years. Please explain the nature of "Other" creative/scholarly activities.
Faculty
Rank
#
SSH Fall
Courses
Taught
Creative/Scholarly Activities
RP
Ch
B
OP
G (#/$)
Other
-5
-4
-3
Please list service to the department, college/university and the community provided by faculty
at each rank below:
Faculty
Rank
1-5
1-4
1-3
#
Department
College & University
Big Island Community
Program Quality and Student Learning Outcome Assessment
Complete the following Opportunity to Learn Matrix. Indicate where each student learning
outcome is introduced (I), practiced (P), and demonstrated (D). Click here to view a sample.
Regularly Offered
Program Courses
Program Student Learning Outcomes
2
3
4
5
1
6
Sample from the Mathematics Program Self-Study
205/6
231/2
300/1
303/4
310
311
317
421/2
431/2
454/5
Understanding different areas of
math and how they interrelate
Classical theorem-proving skills
X
X
x
x
x
X
x
X
X
X
Refined understanding of the
problem solving process
Ability to independently develop
and deliver all pre-college math
Working knowledge of
technology appropriate to field
Skills:
a. Read, write, translate and
articulate mathematics
b. Solve problems algebraically,
numerically, and graphically
c. Make Inferences &
generalizations
x
x
x
x
x
x
X
X
X
X
X
x
x
x
X
X
X
X
X
X
X
X
X
x
X
X
x
X
X
x
X
X
X
X
X
x
x
X
X
X
X
x
x
x
x
7
In each major course, how are students assessed? From these assessments, which ones are used
to provide data for program improvement? Place an asterisk next to each measure that is used for
program improvement. Please choose at least one assessment that program faculty assess to get
feedback about program effectiveness. Click here for an example.
Course Level Assessment of SLO
Measures (Please add or
delete rows as
appropriate)
Objective Tests
Essav Exams
Research Papers
(Library)
Research Papers
(Original data-based)
Pre/Post Test
Comparisons
Group Projects
Portfolios
Simulations
Presentations
Service Learning
Overall Pass Rates
How effectively does this program meet the stakeholder expectations including alumni,
employers and other community members? Please complete the table below adding or deleting
rows as appropriate.
Targeted Outcome
Measurement (data
collection process)
Results
How results used to
improve the program
8
Program Review Schedule 2002 – 2009
(Revised 2005)
2002/2003
Chemistry
Computer Science
Biology
Geology
Mathematics
CAFNRM
Occupational Safety & Health (minor)
Women's Studies
Externally Reviewed Programs
Business Administration
Education
Nursing
2003/2004
Hawaiian Studies
English
Philosophy/Religious Studies
Communication
Linguistics
2004/2005
Anthropology
Art
Japanese Studies
2005/2006
Economics
Physics/Astronomy
History
Political Science
2006/2007
Geography
Administration of Justice
Liberal Studies
Marine Science
Psychology
Sociology
2007/2008
Performing Arts
(Other Programs & Certificates)
Environmental Studies
Honors
International Studies
Natural Sciences
Pacific Islands Studies
9
Download