PROGRAM REVIEW Self-Study Resource Guide 2007-08

advertisement
PROGRAM REVIEW
2007-08
Self-Study Resource Guide
College of the Redwoods
Redwoods Community College District
April 2007
i
Table of Contents
I.
Introduction: What is Instructional Program Review? ............................................1
A.
Purpose and Guiding Principles...................................................................1
B.
How Program Review Integrates with Planning Assessment......................2
C.
Constituents and Responsibilities ...............................................................5
D.
List of Approved Programs..........................................................................9
E.
Instructional Program Review Schedule for Approved Programs .............13
F.
Instructional Program Review Schedule for Disciplines ...........................15
F.
Student Services and Learning Support Program Review Schedule .........16
G.
Program Review Distribution of Activities 2007-08 Calendar..................18
II.
Comprehensive Instructional Program Review .....................................................19
A.
Overview....................................................................................................20
B.
The Comprehensive Program Review Self-Study Document Template ...22
C.
Quality Improvement Plan (QIP) ..............................................................31
D.
Program Review Committee Self-Study Review Guidelines ....................33
E.
Program Review Participant Survey..........................................................40
III.
Annual Instructional Program Review Update ......................................................42
IV.
Appendices for Instructional Program Review......................................................47
A.
Appendix A: Student Learning Outcomes, Course and Program Level ....48
B.
Appendix B: Annual Update Forms ..........................................................54
C.
Appendix C: Assessment and Strategic Planning......................................59
D.
Appendix D: Glossary of Assessment Terms ............................................60
V.
Student Services and Administrative Services Program Review ..........................61
A.
Introduction................................................................................................62
B.
Schedule of Reviews..................................................................................64
C.
Comprehensive Program Review Process Timeline..................................65
D.
Comprehensive Program Review Process Steps........................................69
E.
Appendices for Student Services and Administrative Services
Program Review ........................................................................................71
1.
Appendix A: Comprehensive Program Review Workbook...........71
2.
Appendix B: Program Review Annual Update Report..................76
3.
Appendix C: Program Review Evaluation Forms .........................79
F.
Assessment Glossary and Data Element Definitions.................................81
VI.
Calendars for Program Review Activities .............................................................85
A. Instructional Programs Fall 2007.....................................................................85
B. Student Services and Administrative Services Fall 2007 & Spring 2008 .......91
Based on a compilation of program review models from Riverside Community College District (CA) and Bluegrass Community and Technical
College (KY). Used with permission..
Section I
Introduction--What is Program Review?
Section I
Introduction--What is Program Review?
Purpose
Program review is a means by which the college can systematically and regularly gather and
analyze both qualitative and quantitative data in order to facilitate the continuous improvement
of each program, to guide resource allocation, and to assist the faculty, administration and Board
in making decisions about programs. As the primary process for program-level planning,
program review provides mechanisms for:
•
•
•
•
•
•
•
improving program quality
articulating the program mission and vision with those of the institution
assessing program strengths, weaknesses, opportunities and threats
determining program capabilities
identifying and removing barriers between institutional areas of operation
making use of college processes for fiscal, human, physical, equipment, and
technology resource allocation
assessing compliance with Board policy, WASC, and Ed Code
The state Academic Senate has recognized that “the principal purposes of the review process are
to recognize and acknowledge good performance, to enhance satisfactory performance and help
programs which are performing satisfactorily to further their own growth, and to identify weak
performance and assist programs in achieving needed improvement…” (“Program Review:
Developing a Faculty Driven Process,” Spring 1996).
The Program Review Committee also acknowledges potential inappropriate uses of the
program review process. The District process will not be misused in the ways discussed below.
It is not a process for evaluating individual performance.
It is not a process for eliminating programs. Decisions about discontinuing programs
should be made through an alternative process designed to address discontinuation
issues directly.
It should not be, or appear to be, an adversarial process between faculty and
administration. The process should lend itself to developing more effective programs
through dialogue and collaborative efforts.
It should not be merely an exercise to fill out a document that sits on a shelf. It should
have real impacts on program effectiveness and resource allocation decisions.
•
•
•
•
Guiding Principles
The following principles should guide the design and implementation of the Program
Review self-study process:
•
Flexibility: the process needs to be open and flexible enough to accommodate
1
•
•
•
•
differences among instructional programs
Collegiality: the process should be a collaborative process guided by a spirit of open
and honest inquiry
Relevance: the process should result in answers to important and relevant questions
for programs
Practicality: the process should:
Minimize the reporting burden on programs
Have measures that are few in number, simple, easy to implement and
interpret, and useful for program improvement purposes
Provide the necessary data resources to assist with the self-study process
Effectiveness: the process should result in a clear sense of direction and
accomplishment for participants. To facilitate the implementation of plans and the
accomplishment of goals, resource needs identified through the self-study process
will be linked with budgeting, facilities, and human resource planning.
How Program Review Integrates with Planning and Assessment
Planning – Self-study documents are intended to be key drivers informing the strategic
planning process. Appendix C in Instructional Program Review (p. 59) provides a graphic
representation of how assessment and the results of assessment feed into the planning process in
support of the mission of the College.
Student Learning Outcomes Assessment – In the self-study, programs are asked to
develop a student learning outcomes assessment plan that enables them to examine the degree
to which students have achieved the desired learning outcomes for a course, course sequence, or
program of study. The Program Review Committee will help programs and courses develop
instructional program and course student learning outcomes and assessment plans. Further details
about how to implement student outcomes assessment follow later in this document. (See
Instructional Program Review, Appendix A, p. 48.)
2
Model for Planning and Assessment Process
3
All instructional and student services programs identified annually by the Program Review
Committee and approved by the Academic Senate will undergo self-study as part of a process
that results in a comprehensive assessment of institutional effectiveness. The District Program
Review Committee will forward approved self-studies and action plans to the Office of
Academic Affairs, the Office of Student Services, and designated planning groups. These groups
will review the self-studies and use information to inform resource allocation decisions. Linking
program review and planning is essential; consequently, resource allocations and planning for
your discipline will be directly impacted by this plan.
The most important extrinsic purpose of this document is for you to receive the resources you
need to serve your students. Your program’s priority for faculty or staff, equipment, space and
supply allocations will be dependent upon your submission of this report.
The intrinsic purpose of the self-study process is to help programs clarify and achieve their goals.
In addition, the process aids programs in:
•
•
•
•
Strengthening programs through self-improvement and self-determination;
Strengthening bonds and fostering cooperation within the college community;
Providing systematic feedback on student learning;
Generating continuous and ongoing dialogue about how student learning can be
enhanced through program and service improvements.
There are two parts to the Program Review Process:
• Comprehensive Program Review. Occurring every five years for instructional
programs, an intensive, comprehensive program review will address each program’s
courses, pedagogy and other developments related to the future of the program across
the district. Beginning in 2007, instructional program reviews will follow a five-year
comprehensive cycle with annual updates. As much as possible, comprehensive
instructional reviews will be coordinated with the reviews required for external
accreditation. Student services program reviews will also follow a five-year
comprehensive cycle beginning in 2006.
• Annual Program Review Update allows each program and each discipline to
address its needs for resources (staff/faculty, space, equipment, etc.) and to report on
outcomes assessment activities for the prior year. This annual update will also allow
occupational and technical programs to report mandated criteria on program efficacy.
All student services programs will also provide an annual update.
The Program Self-Study Group (comprised of those individuals writing the comprehensive or the
annual review) uses data provided by the program and by the Institutional Research office to
analyze the program’s
• Relevance to the college’s mission and its intended outcomes
• Responsiveness to community needs and satisfaction of community demand
• Accessibility to students and identification of unnecessary barriers
• Student learning outcomes, including participation and successful completion rates
• Measures of program quality
• Adequacy of resources and efficiency of resource use
4
•
•
Performance, price, and enrollment compared to that of alternate local suppliers
Direct and indirect revenues and costs to the college
The Program Self-Study Group (PSSG) will submit its Self-Study Report (SSR) and a
Quality Improvement Plan (QIP) to the Program Review Committee. The PRC will verify the
information, complete a questionnaire based on this information, assess performance indicators,
and provide feedback regarding the QIP and the Self-Study. The PRC will then forward all
documentation (the Self-Study, the questionnaire and the QIP) to an established planning group.
Constituents & Responsibilities
Program Self-Study Group (PSSG)
• Membership: All (when possible or feasible) of a program’s faculty or unit’s staff led by
the program coordinator or unit administrator (and, whenever possible or feasible, the
Advisory Board and students). For instructional programs with no full-time faculty
member, the division chair, director, or campus vice president shall coordinate and help
facilitate the process of forming the self-study group. For instructional programs with
only one full-time faculty member, the program coordinator will facilitate the process of
forming the self-study group. As much as possible, support staff, advisory groups and
students should be involved in the self-study groups.
• Responsibilities: Generates Self-Study Report; submits Self-Study Report and Quality
Improvement Plan to PRC; reviews the PRC response questionnaire and
recommendations and makes any necessary revisions; submits final Self-Study Report
and QIP to PRC for final approval.
IR
• Responsibilities: On a schedule, collect and disseminate data to programs, review data
with chairs, coordinators and/or unit administrators.
District Program Review Committee
• Membership: The Committee will be comprised of faculty, staff, and administrators from
across the three campuses, as defined in AR122.01.
• Responsibilities: Work with IR to review data; maintain the PR schedule; act as liaisons
to PSSGs; inform programs of upcoming reviews; provide training; recommend either
full or conditional approval of completed program reviews (conditional approval will
require the program to make specified revisions to the report and/or the QIP to gain full
approval); make fully approved program reviews available to the public; and submit these
reviews along with QIPs to Academic Senate (for information purposes) and to
designated planning groups (for planning purposes).
Academic Senate
• Responsibilities: Receives self-study and QIP for informational purposes.
Planning Committee
• Responsibilities: A designated planning committee will incorporate information resulting
from reports into the planning and budgeting process.
5
Board of Trustees
• Responsibilities: Review approved program reviews and recommendations for
institutional improvement.
Publication and Discussion of the Findings
Fully approved self-studies will be placed online and distributed to the Office of Academic
Affairs, the Office of Student Services, and the Board of Trustees. Presentations of self-studies
may be made to the Board of Trustees, Cabinets, and the Academic Senate.
6
COMPREHENSIVE PROGRAM REVIEW PROCESS FLOW CHART
Program Review
Committee (PRC)
~Informs programs of
upcoming reviews
~Trains program selfstudy group members
Institutional
Research
~Provides data to
program
Program Self-Study
Group (PSSG)
~Generates Self-Study
~Submits Self-Study
to Program Review
Committee
PRC
~Reviews
preliminary SelfStudy
~Completes
questionnaire and
returns report,
questionnaire,
and recommendations to PSSG to
do QIP
7
PSSG
~Reviews results of
questionnaire
~Completes QIP
~Submits final
report and QIP to
PRC
PRC
~Collects all final
material QIP
~Approves report
~Submits fully
approved reports to
Senate, Academic
Affairs, Student
services for
information and to
planning group for
action
Planning
Committee
~Submits to Board
of Trustees for
information
~Incorporates
relevant information
into planning,
budgeting and goals
development
ANNUAL UPDATE PROCESS FLOW CHART
Program Review
Committee (PRC)
~Informs programs of
upcoming reviews
~Trains program selfstudy group members
Institutional
Research
~Provides data to
program
PRC
~Reviews Annual
Updates
~Submits fully
approved reports to
Senate, Academic
Affairs, Student
services for
information and to
planning group for
action
Discipline or
Program Faculty
~Generates Annual
Update and submits
to Program Review
Committee
8
Planning
Committee
~Submits to Board
of Trustees for
information
~Incorporates
relevant information
into planning,
budgeting and goals
development
List of Programs
(Approved annually by the Academic Senate. Last approved 11/17/07.)
Academic Instruction Programs
Addiction Studies (CA)
Administration of Justice
Administration of Justice (AS)
Administration of Justice (CA)
Basic Law Enforcement (CA)
Agriculture
Agriculture (AS)
Plant Science (AS)
Plant Science (CA)
Sustainable Agriculture (CA)
Landscape Maintenance (CC)
Nursery Practices (CC)
Automotive Technology
Automotive Technology (AS)
Advanced Auto Tech (CA)
Basic Auto Tech (CC)
Business
General Business (AS)
General Business (CA)
Medical Office Business Skills (CA)
Bookkeeping (CC)
Desktop Publishing (CC)
Income Tax Preparation (CC)
Payroll Clerk (CC)
Word Processing (CC)
Business Technology
Computer Support Specialist (AS)
Office Professional (AS)
Computer Information Systems
Computer Programming (AS)
Computer Programming (CA)
CIS Networking (AS)
CIS Networking (CA)
9
Construction Technology
Construction Technology (AS)
Historic Preservation & Restoration (AS)
Historic Preservation & Restoration (CC)
Fine Woodworking I (CA)
Fine Woodworking II (CA)
Residential Construction I (CC)
Residential Construction II (CA)
Cabinetmaking and Millwork (CC)
Residential Wiring (CC)
Court Reporting (CA)
Diesel Heavy Equipment Technology
Diesel Heavy Equipment Technology (AS)
Diesel Heavy Equipment Technology (CA)
Diesel Engine Overhaul (CC)
Diesel Fuel Injection & Electronics (CC)
Diesel Truck Maintenance & Power Train (CC)
Digital Media
Digital Media (AS)
Digital Media (CA)
Game Art & Animation (CA)
Game Programming (CA)
Animation (CC)
Internet Design & Electronic Commerce (CC)
Drafting Technology
Architectural Drafting (AS)
Architectural Drafting (CA)
Civil Design (AS)
Civil Design (CA)
Mechanical Drafting (AS)
Mechanical Drafting (CA)
Early Childhood Education
ECE-Infant/Toddler (AS)
ECE-Preschool (AS)
ECE-School Age (AS)
ECE-Special Needs (AS)
ECE (CC)
ECE (CA)
Child Development Center
10
Education
Paraprofessional Education (AS)
Forestry and Natural Resources Technology
Forestry & Natural Resources, Transfer (AS)
Forestry Technology (AS)
Forestry Technology (CA)
Geographic Info Systems (AS)
Natural Resources (AS)
Graphic Communications
Graphic Communications (AS)
Graphic Communications (CA)
Health Occupations
Dental Assisting (AS)
Dental Assisting (CA)
Licensed Vocational Nursing (AS)
Licensed Vocational Nursing (CA)
Medical Assisting (AS)
Medical Assisting (CA)
Nursing Assistant (CA)
Registered Nursing (AS)
LVN to RN Upgrade (AS)
LVN to RN Upgrade (CA)
Hospitality, Restaurant, and Culinary Arts
Culinary Management (AS)
Restaurant Management (AS)
Culinary Supervision (CA)
Restaurant Supervision (CA)
Culinary Operations (CC)
Restaurant Operations(CC)
Legal Studies
Paralegal Studies (AS)
Paralegal Studies (CA)
Child Support Specialist (CC)
Manufacturing Technology (AS, CA)
CADD/CAM Manufacturing (AS, CA)
Marine Science Technology
Marine Science Technology (AS)
Marine Science Technology (CA)
11
Natural History (CC)
Real Estate (CA)
Science Major University Prep
Welding Technology
Welding Technology (CC)
Electric Arc & Oxyacetylene Welding (CC)
MIG & TIG Welding (CC)
AA—University Studies
AA—General Studies
AS—General Studies
General Education (CR)
Student Services Programs
Academic Support Center
Athletics
CalSOAP
Career Center/Student Employment
Counseling and Advising
DSPS
Admissions and Records
EOPS
Library
Residential Life and Campus Activities
Student Health Center
Upward Bound
Judicial Affairs
12
Program Review Schedule for Instructional Programs
Program
Addiction
Studies
Administration
of Justice
Agriculture
Automotive
Technology
Business
Business
Technology
Computer
Information
Systems
Construction
Technology
Court
Reporting*
Diesel Heavy
Equipment
Technology
Digital Media
Drafting
Technology
Early Childhood
Education, incl.
CDC
Education*
Forestry and
Natural
Resources
Technology
Graphic
Communications
All Health
Occupations
degrees and
certificates
Hospitality,
Restaurant, and
Culinary Arts
Manufacturing
Technology
2007-08
2012-13
2008-09
2013-14
2009-10
2014-15
2010-11
2015-16
2011-12
2016-17
C
A
A
A
A
A
A
A
A
C
A
C
C
A
A
A
A
A
A
A
C
C
A
A
A
A
A
A
A
A
C
A
A
A
A
C
A
A
A
A
C
A
A
A
A
C
C
A
A
A
A
A
A
A
A
A
A
C
A
A
A
C
A
A
A
A
A
A
C
A
A
A
C
A
A
C
A
A
A
A
A
13
Program
2007-08
2012-13
2008-09
2013-14
2009-10
2014-15
2010-11
2015-16
2011-12
2016-17
Marine Science
Technology
Natural History
Real Estate
Science Major
University Prep
Welding
Technology
AA—University
Studies
AA—General
Studies
AS—General
Studies
General
Education
A
C
A
A
A
A
A
A
C
A
C
A
C
A
A
A
A
A
A
A
A
A
C
A
A
A
A
A
A
C
A
A
A
A
C
C=Comprehensive Program Review A=Annual Update
*These programs were on the approved (11/17/06) list of programs but status of programs may
change in 2007-08 academic year.
14
Program Review Schedule for Instructional Disciplines (not on the program list)
Discipline
2007-08
2008-09
ACC
ANTH
ART
ASTRO
BIOL
CHEM
CINE
CET
COMM
DANCE
DRAMA
ECON
EMT
ENGL
ENGR
ENVSC
FS
FRNC
GS
GEOG
GEOL
GERM
GUID
HE
HIST
HRC
IT
IDS
ITAL
JPN
JOURN
MA
MATH
METEO
MUSIC
NAS
PHIL
PE
PHYSC
PHYS
POLSC
SNLAN
SOC
SPAN
SPCH
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
A
15
Schedule of Reviews for Student Services and Learning Support
C=Comprehensive Program Review
Unit
Admissions &
Records
Residential Life and
Campus Activities
Judicial Affairs
EOPS
Counseling and
Advising
DSPS
CalSOAP
Upward Bound
Career
Development/Student
Employment
Library
Academic Support
Center
Athletics
Student Health
Center
A=Annual Update
Academic Year
2006- 200707
08
C
A
200809
A
200910
A
201011
C
201112
A
201213
A
201314
A
C
A
A
A
C
A
A
A
C
C
C
A
A
A
A
A
A
A
A
A
C
C
C
A
A
A
A
A
A
A
A
A
C
C
C
C
A
A
A
A
A
A
A
A
A
A
A
A
C
C
C
C
A
A
A
A
A
A
A
A
A
A
A
A
C
C
A
A
A
A
A
A
C
C
A
A
A
A
A
A
A
A
C
C
A
A
A
A
A
A
C
C
A
A
A
A
16
Potential data to be provided by Institutional Research
(Data required by each program review will vary by program)
Data will be sorted by fall/spring for four years.
• Curriculum history by semester and time-of-day
• Current and historical course enrollment by headcount & FTES
• Student cumulative GPAs by course
• Retention by course (from census to end of term)
• Retention vs. successful completion data for students who meet the advisory (if one exists) and
for those who do not
• Percentage of successful completion by course by student type
• Attrition and retention of students continuing in a program by semester for two full years (e.g.,
fall 2005, spring 2006, fall 2006, spring 2007)
• Current & historical count of graduates by student type
• Current & historical count of transfers by student type
• Average class sizes
• Percentage of courses taught by full- and part-time faculty
• Ratio of part-time faculty to full-time faculty and the percent of FTES taught per part-time faculty
• Student transfer and transfer success tracking
• Additional information available upon request (e.g., student demographics)
• * Results of continuing & current student questionnaire
• * Results of alumni and graduate surveys (student feedback on career and technical preparation)
• * Employer satisfaction survey results (employer feedback on career and technical preparation)
• * Alumni satisfaction survey results
* Data currently unavailable; April 2007
Other data sources which might be helpful and which IR can help locate (although may not be
responsible for):
• Population Demographics
• Labor Market Data
• Student Program Intention
• Feedback from Technical Advisory Committees
• Student Placement/Wage Data
• Other student tracking needs (tracking students who apply for state credentials, etc.)
• Comparative Data
• Mission statement and strategic plan
17
Comprehensive Program Review Distribution of Activities Calendar for 2007-08
Date
Ongoing
Institutional Research
IR provides the Area/Program Coordinator or Student Services Administrator (when
applicable) with all necessary data to complete the program review
Program Review Committee
Provides orientation and training to the Area/Program Coordinator or Student Services
Administrator
April 2007
Provides Program Review Guide to the Area/Program Coordinator or SS Administrator
Nov. 1-15, 2007 Reviews the Self-Study Report and Quality Improvement Plan (QIP)
Nov. 30
Makes recommendations to Self-Study Group (if necessary)
Jan. 30, 2008
Receives and approves completed Self-Study and QIP; completes the Review Questionnaire
Feb. 15
Monitors the submission of the Program Review Report and the Program’s Quality
Improvement Plan to Academic Senate and Institutional Effectiveness by the appropriate
deadlines
Spring 2008
Posts a record for each program review that includes all key features of the review process
Spring 2008
Compiles surveys and makes necessary adjustments to the process for the following year
April 2007
Area/Program Coordinator/SS Administrator
Apr-Nov.1, 2007 Works with self-study team/program members to prepare report (which should include input
from students and, as applicable, advisory board members)
Nov. 1
Submits report and QIP to the PRC
Nov. 1-15
Provides any supplementary information/documentation needed by PRC
Nov. 30
Begins meeting with program faculty to prepare an action plan addressing any program
recommendations identified in the report, including who is responsible for the action and the
date of expected completion
Feb. 7, 2008
Sends one electronic and one hardcopy of the action plan to the PRC and the Academic
Affairs Office by the appropriate deadline
Program Review Annual Updates Distribution of Activities Calendar for 2007-08
Date
Ongoing
April 2007
April 2007
August 1
Jan 30, 2008
Feb. 15
Spring 2008
Spring 2008
Fall 2007
Oct. 15-Dec. 1
Jan. 2008
Feb. 7, 2008
Institutional Research
IR provides the Area/Program Coordinator or Student Services Administrator (when
applicable) with all necessary data to complete the program review (for fall and spring of
previous year)
Program Review Committee
Provides orientation and training to the Area/Program Coordinator or Student Services
Administrator
Provides Program Review Guide to the Area/Program Coordinator or SS Administrator
Determines IR has previous year’s data for programs and instructional disciplines
Reviews the Annual Updates
Monitors the submission of the Program Review Annual Updates to Academic Senate,
General Education Committee and Institutional Effectiveness
Posts a record for each program review that includes all key features of the review
process
Compiles surveys and makes necessary adjustments to the process for the following year
Area/Program Coordinator/SS Administrator
Works with self-study team/program members to prepare update
Submits Annual Update to the PRC (due date set by PRC)
Provides any supplementary information/documentation needed by PRC
Sends one electronic and one hardcopy of the action plan to the PRC and the Academic
Affairs Office by the appropriate deadline
18
Section II
Comprehensive Instructional Program Review
19
Section II
Comprehensive Instructional Program Review
Overview of the Program Self-Study Group Process
The normal time frame for the Comprehensive Program Review self study is one academic
year. The process consists of two major steps:
1.
2.
Initial Planning
Self-Study Review Process
The Program Self-Study Group will:
• Develop detailed task list and timeline
• Review data packet from Institutional Research
• Discuss self-study guide and questions within the program
• Gather any additional information the program deems necessary
• Update course outlines and submit to Curriculum Committee if needed
• Prepare draft documents resulting from dialogue within the program
• Review and finalize the Self-Study Report
• Submit Self-Study Report to the Program Review Committee (including updated
course outlines)
• Will review Program Review Committee’s completed questionnaire and list of
recommendations for revising the Self-Study Report and for preparing a Quality
Improvement Plan
• Submit revised Self-Study and Quality Improvement Plan to Program Review
Initial Planning & Training
The initial orientation/training meeting will:
•
•
•
•
Familiarize instructional programs with ACCJC/WASC accreditation standards, the
program review process, student learning outcomes assessment, and the role of strategic
planning.
Allow programs to begin the self-study with hands-on training in data interpretation and
assessment techniques.
Assist programs in the development of a project plan with timelines and assignment of
responsibilities.
Generate discussion about the discipline/self-study process, including:
1. Look at the data provided by Institutional Research that have informed your
document, discuss it with your colleagues, and identify major changes or trends
you expect to be of particular relevance to your discipline in the next five years.
2. What will the implications of these trends be for your discipline and what changes
will your discipline need to make to respond to these issues?
20
3. Comment on the enrollment trends and indicate plans as a result of that analysis.
What trends, if any, exist in the data by gender, ethnicity, night vs. day, traditional
vs. non-traditional age, etc.? Why might these trends be occurring? Are there
additional data that would assist your discipline in analyzing these trends?
As many members of the program as possible, including the discipline/area coordinator and the
division chair should attend the training. Each program should select program representatives
who will communicate with Institutional Research and the Program Review Committee and
provide organizational leadership for completion of the self-study.
21
Comprehensive
Instructional Program Review Document
Programs are advised to review their prior Program Review submission (if one exists) and utilize
those portions that remain relevant and appropriate to the current format.
A. Mission and Relationship to the College(s)
Describe how the activities and goals of your program relate to the mission statement and
strategic plan of the college.
B. Program Description, Curriculum, and Information
1. Provide official program description and list program student learning outcomes. Use the
following table to align program student learning outcomes with courses in which each
student learning outcome is addressed. Provide additional information relevant to your
discipline.
a.
b.
Official Program Description
Develop a two-column matrix which presents:
i. Program-Level Student Learning Outcomes (left-hand column)
ii. Course(s) where these program-level SLOs are presented (right-hand
column)
Program Student Learning Outcome Statements
SLO 1
SLO 2
Etc.
Course (s)
Course(s) addressing SLO 1
Course(s) addressing SLO 2
Etc.
2.
Program-Specific Criteria and/or Admissions Guidelines (as applicable to program)
3.
Outline the curriculum as it is being implemented for a full-time student completing a
degree or certificate in this program. The outline should include course number, course
title, units, lecture hours, and lab hours for each semester for the complete curriculum.
Course
Number
Course
Title
Units
22
Lecture
Hours
Lab
Hours
4. As part of your self-study, review and summarize the development of curriculum in the
program. Include recent additions, deletions, or revisions of courses (attach current
course outlines). Evaluate the timing, frequency and coordination of course offerings to
determine the adequacy of course offerings relative to a transfer degree (articulation),
vocational/occupational certificates, and other appropriate aspects of the
district's/campus’/college’s mission.
a) Please ensure that there is a thorough review of the course outlines of record and
course content during this review period. Please indicate on the course outlines the
date on which they were last revised. If the last major curriculum revision occurred
more than five years ago, please indicate when the next major revision is planned.
Please also review course prerequisites, co-requisites, and advisories as well as
obtaining necessary approvals for distance education courses.
b) Send updated course outlines to the Curriculum Committee.
c) Please file the appropriate form with the Curriculum Committee to delete classes that
have not been taught for three or more years unless you plan to teach them in the
future.
d) With respect to updating course outlines of record, list any relevant trends in your
discipline with regard to:
1. Knowledge requirements
2. Skills/student learning outcome requirements
3. Instructional methods
e) Describe the various educational delivery methods currently being utilized by the
program. Examples include but are not limited to traditional in-person classroom
delivery, in-person lab, field studies, online, interactive television, telecourses,
clinical instruction, etc.
f) Identify curricular revisions, program innovations, and new initiatives planned for the
next five years.
5.
If applicable, indicate the program external accreditation/approval status. Include a copy
of the most recent notification of accreditation/approval status from the appropriate
agency. If external accreditation is available but the program has chosen not to seek
accreditation status, please explain.
23
C. Program History
1.
History (update) since last review:
What have been the major developments, activities, changes, and/or projects in
your discipline over the past 2 or 4 years (longer if no recent review exists)?
(This does not need to include curriculum updates, addressed above.)
2.
What were the recommendations from your last program review (if any) and how has
your discipline responded to those recommendations?
3.
If you have goals from your previous program review, please list them along with the
objectives related to your goals, the strategies being used to achieve objectives, and the
documentation or evidence that demonstrates success. If no prior program review exists,
skip #3.
4.
If goals were not achieved, please explain.
5.
Discuss any collaborative efforts you have undertaken with other programs (instructional
or non-instructional) at College of the Redwoods District and offer an assessment of
success and challenges, and potential changes in collaborative efforts.
6.
Discuss any activities or projects you have undertaken with other educational institutions,
the community, or business/industry.
D. Measures of Effectiveness
1.
Quality of Education
a.
Results of certifying, licensing, or registry examinations for each of the last five
years (or list not applicable):
1st Attempt
Year
Number
Attempting
Pass
Percent
20022003
20032004
20042005
24
2nd Attempt
Pass
Percent
National %
Goal
State %
Goal
20052006
20062007
b.
Faculty Qualifications:
(1)
Is there one full-time faculty member whose primary assignment is
responsibility for this program?
Yes
No
If not, explain.
(2)
Are minimum faculty qualifications according to standards set by
accrediting/approval bodies met?
Yes
No
(a)
State Chancellor’s Office?
(b)
External Accrediting/approval organizations
Yes
No
If “No” for either response, explain.
c.
Student Outcomes Assessment: Attach course- and program-level outcomes and
assessment reports for each of the last four years. (See Instructional Program
Review Appendix A, pp. 48-49). Program-level outcomes should include
information from graduate/alumni and employer surveys.
d.
2.
Vitality
a.
Attrition and Retention
Number of students entering program in fall two academic years prior to this
comprehensive program review year:
e.g., Fall 2005
Number of students continuing in program (continuing students: all entering
students except transfers or readmissions)
Spring 2006
Fall 2006
Spring 2007
25
Number Graduating
in 2 years (Fall 2005
Entering Class)
Explain any attrition indicated by these data.
b.
Enrollment and Graduate Projections
NOTE: Explain method used in making these projections. Use census day
enrollments for this figure.
Total Enrollment
(unduplicated
headcount)
Graduates
20052006
N=
20062007
N=
20072008
N=
20082009
N=
20092010
N=
N=
N=
N=
N=
N=
c.
Total number of 2006-2007 graduates by ethnic group and gender.
d.
Indicate the number and percent of graduates from the most recent graduating
class who are employed in positions related to the program major or continuing in
a higher degree program.
NOTE: Indicate sources of information and provide any explanation as necessary.
Total Number
Of Graduates
Employed in
Related Field
N
%
Continuing
Education
N
%
Year
e.
Indicate the beginning mean and median annual salary for graduates of the
program employed full-time, as collected from alumni survey and program
advisory meetings.
26
Mean and Median Beginning Annual Salary by Graduating Class
Mean Beginning Salary
Median Beginning Salary
2000-2000
2000-2000
2000-2000
2000-2000
2000-2000
f. Have there been any significant changes in enrollment, retention, success rates, or student
demographics that impact your discipline? If so, please include data sheets (Excel or Word
format) showing these changes.
3.
Efficiency
a.
Describe current student population in terms of enrollment by ethnic group and
gender.
b.
Composition of enrollment for selective admissions programs:
Fall
2003
Selective Admission
Programs
Number of Applicants N=
Number of
Acceptances
Number of Entries
c.
Fall
2004
Fall
2005
Fall
2006
Fall
2007
N=
N=
N=
N=
N=
N=
N=
N=
N=
N=
N=
N=
N=
N=
Faculty Staffing Pattern
2002-2003
a) Full-time
faculty
b) Part-time
faculty
2003-2004
2004-2005
2005-2006
2006-2007
N=
N=
N=
N=
N=
N=
N=
N=
N=
N=
27
d.
Please rate the resources indicated in the table below with respect to how they
support this program. Feel free to supplement the resource categories in the extra
space provided. The four rating categories are defined as follows:
Adequate:
Minimally
Adequate*:
Inadequate*:
Not
Applicable:
This one resource is adequate to permit the program to
function effectively.
Program functioning at minimally effective level.
The limitations imposed by this one resource
render the program marginal and will require immediate
review for program improvement or continuation.
Does not apply to this program.
Resources
Rating
Adequate Minimally Inadequate
Not
Adequate
Applicable
a) Faculty
Full-time
Associate
b) Support Staff
Clerical
Technical
Instructional Support
Other Personnel
c) Current Expenses
Office Supply Expenses
Instructional/lab supply
Expenses
d) Equipment
e) Library Resources
f) Facilities
g) Professional Development
*If any component is rated Inadequate or Minimally Adequate, please explain. Include
recommendations for reducing or eliminating the limitations.
28
e.
Budgetary Categories--Please provide funding amount budgeted for each category
during the most recent academic year and for each of the previous four years.
(Please replicate this table for each year)
Year:
General Funds
Amount
Non-General Funds
Amount
Full-time Faculty
Part-time Faculty
Other, please
specify
1.
2.
3
4
Sub-Total
Travel
Supplies
Equipment
Sub-Total
TOTAL
Explain any changes in funding and add any comments relevant to the adequacy of resources.
29
Summary and Recommendations
1.
Summary
●
Program Strengths (Please consider how these strengths can be marketed to the
community and provide suggestions to PR department)
●
Program Weaknesses
2.
Recommendations for program improvement generated by self-study. (List by
number and use these numbers in the Quality Improvement Plan.)
3.
List program members who participated in completion of the self-study report.
Vision and Goals
1.
Based upon data from California 2025 and/or data from Institutional Research, briefly
describe how you would like your discipline to evolve in the next five years. In what
ways does your current state differ from your desired state?
2.
What specific goals and objectives would you like to achieve to move you toward your
vision?
3.
What support from the college or district is needed to help you achieve your goals and
objectives?
4.
What documentation/evidence will demonstrate that you are making progress toward
achieving your goals, objectives, and vision?
5.
What changes will make the self-study process more helpful to you?
30
Completing the Quality Improvement Plan (QIP)
The Quality Improvement Plan (QIP) is intended to assist the program in thinking and planning
for a minimum of the next three years. Many factors that influence the implementation of the
program’s plans can and do change over time. Each year, the coordinator, division chair or
administrator, in consultation with other faculty, will be asked to update the QIP to make
adjustments to the plans the program enters this year.
The purpose of the QIP is to provide a format for translating both a) the self study group’s
recommendations in the Instructional Program Review Self-Study and b) the review team’s
recommendations in reviewing that report into actions for improving or maintaining the quality
of the program; it is also a mechanism for administrators to determine input to the planning
process. Program Self -Study teams complete this form after the Program Review Committee has
met and reviewed the instructional program review self-study.
Because this document will be used to inform planning processes, it is very important that all the
requested information be provided. The form has been designed to elicit the information needed
for this process. Each “block” on the form is for a single recommendation; thus, the program
should complete all the fields for each of the recommendations cited in the self-study. To add
blocks, please use the copy and paste functions in Word.
31
Quality Improvement Plan
Program:
Year:
Field
Recommendation # (from self-study report
or from review-team report)
Recommendation
Planned Implementation Date
Estimated Completion Date:
Action/Tasks
Measure of Success/Desired Outcome
Estimated Cost(s)
Who is responsible?
Consequence if not funded
External Accreditation Recommendations
(if applicable)
What to include
Number should have been assigned to each
recommendation in the Instructional
Program Review Self-Study (IPRS).
Use wording from IPRS.
When does the program expect to begin to
implement this recommendation?
When does the program estimate this
recommendation to be fully implemented?
What steps must the program take to
implement the recommendation?
If the recommendation is implemented,
what about the program will be improved?
What difference will the implementation of
this recommendation make in relation to
students, the program’s purpose, and the
college’s mission? How will this
recommendation improve learning and help
meet targeted objectives?
This field is particularly important because
the information the program enters here is
the information that the Institutional
Effectiveness/Planning committee will
consider in its planning processes.
Who will ensure the change is
implemented and tracked?
If this recommendation is not funded, how
will students, the program, the college, or
the community be negatively impacted?
Does an external accrediting body require
that action be taken? If so, what steps
should be take to achieve results? What is
the cost implication of this
recommendation?
32
Program Review Committee Self-Study Review Guidelines
I.
Areas of Focus
The PRC will provide the program with an external perspective concerning all
administrative and educational aspects of the program. This perspective includes a
judgment of program quality and an assessment of the effectiveness of the instructional
program. Although certain information shall be incorporated in all program review
reports, they are not intended to be totally comprehensive.
Using the Questionnaire on the next page, the PRC will focus its review and
recommendations on the following major areas:
1.
Quality of degree programs and student learning
2.
Quality of faculty/staff and program productivity
3.
Quality of recruitment and retention of a diverse faculty, staff, and student body
4.
Quality of management, adherence to policies, and collegial environment
5.
Quality of planning, evaluation and resources maximizing program effectiveness
in terms of allocation
33
II.
Questionnaire
In reviewing the Self-Study Report, the PRC should use this document to answer
questions about each program. Please answer “yes” or “no” for each question and provide
an explanation in the comments section when appropriate (all “no” responses must be
accompanied by an explanation).
Quality of Degree Program and Student Learning
1.
Is the sequence of courses leading to the degree appropriate?
Yes _______
2.
Comments:
No _________
Comments:
Are instructional techniques and policies appropriate for the curriculum and types
of students enrolled?
Yes _______
5.
No _________
Is there evidence that graduates are competent in reading, writing, oral
communication, math and the basic use of computers?
Yes _______
4.
Comments:
Does the curriculum contain at least 18 semester hours of general education
courses (for an Associate’s degree) including at least one course in each of the
following areas identified by California Education Code: humanities, social
science, natural science, writing, oral communications and analytical thinking?
Yes _______
3.
No _________
No _________
Comments:
Are students provided with written information (syllabus) describing course goals,
requirements, content, evaluation methods, etc.?
Yes _______
No _________
34
Comments:
6.
Are results obtained from evaluation of instruction used to improve the quality of
instruction?
Yes _______
7.
No _________
Comments:
No _________
N/A _______
Comments:
No _________
N/A _______
Comments:
No _________
N/A _______
Comments:
Are students’ files adequately maintained to determine students’ progress through
the program?
Yes _______
14.
Comments:
Are results of evaluations of the orientation program used to improve the
orientation program?
Yes _______
13.
N/A _______
Is the orientation program evaluated regularly?
Yes _______
12.
No _________
Is an effective orientation program made available to all full- and part-time
students?
Yes _______
11.
Comments:
Are program length, clock hours or credit hours, tuition and fees appropriate for
the degree awarded?
Yes _______
10.
No _________
Are affiliations with outside agencies (e.g. clinicals in allied health programs)
under the control and supervision of the educational institution?
Yes _______
9.
Comments:
Does instruction take place in an environment conducive to learning?
Yes _______
8.
No _________
No _________
Comments:
Are students satisfied with their educational experiences?
35
Yes _______
15.
No _________
Comments:
Are board exam (e.g. certifying, licensing, and registry) results acceptable?
Yes _______
No _________
N/A _______
Comments:
Quality of Faculty/Staff and Program Productivity
16.
Are the quality and quantity of full- and part-time faculty and staff adequate to
support the program?
Yes _______
17.
Comments:
No _________
Comments:
No _________
Comments:
Are there adequate jobs for program graduates?
Yes _______
25.
No _________
Is program enrollment appropriate?
Yes _______
24.
Comments:
Is graduation rate appropriate?
Yes _______
20.
No _________
Are faculty workloads equitable and reasonable?
Yes _______
19.
Comments:
Is institutional support adequate for faculty/staff professional development?
Yes _______
18.
No _________
No _________
Comments:
Are salaries of graduates appropriate for their level of education?
Yes _______
No _________
36
Comments:
Quality of Recruitment and Retention of a Diverse Faculty, Staff and Student Body.
26.
Is there evidence that the program is involved in recruitment and retention of
diverse faculty and staff?
Yes _______
27.
Comments:
No _________
Comments:
Are admission policies appropriate for this program?
Yes _______
30.
No _________
Does the program have clearly defined admission policies?
Yes _______
29.
Comments:
Is there evidence that the program is involved in recruitment and retention of a
diverse student body?
Yes _______
28.
No _________
No _________
Comments:
Is there evidence that admission policies are followed as published?
Yes _______
No _________
Comments:
Quality of Management, Adherence to Policies and Collegial Environment
31.
Is there an atmosphere of mutual respect among faculty, staff and students?
Yes _______
32.
No _________
Comments:
Is program coordination accomplished in an effective manner?
Yes _______
No _________
37
Comments:
33.
Are management policies relating to program operation administered fairly and
equitably?
Yes _______
34.
No _________
Comments:
Is there evidence of linkages/relationships with other post-secondary educational
institutions?
Yes _______
No _________
Comments:
Quality of Planning, Evaluation, and Resources Maximizing Program Effectiveness in Terms
of Allocation
35.
Are planning and budgeting closely linked?
Yes _______
36.
No _________
Comments:
No _________
Comments:
Are adequate instructional materials and supplies provided for the program?
Yes _______
40.
Comments:
Is adequate classroom and laboratory space provided for the program?
Yes _______
39.
No _________
Are program goals and objectives consistent with college’s mission?
Yes _______
38.
Comments:
Is there broad-based faculty/staff input in the planning and budgeting process?
Yes _______
37.
No _________
No _________
Comments:
Are library resources adequate for faculty, staff and students?
Yes _______
No _________
38
Comments:
41.
Is computer technology adequate for faculty, staff and students?
Yes _______
42.
Comments:
No _________
Comments:
Are educational outcomes assessment results acceptable?
Yes _______
45.
No _________
Are appropriate instruments used to assess educational outcomes?
Yes _______
44.
Comments:
Is there evidence the program advisory committee is utilized effectively?
Yes _______
43.
No _________
No _________
Comments:
Are results of assessment of educational outcomes used to improve the program?
Yes _______
No _________
Comments:
Using the five “Quality” statements as framework, compose a narrative describing the
program’s strengths and weaknesses. Additionally, recommendations for program
improvement must be provided for each weakness identified.
Summary and Recommendations
1.
Summary
●
Program Strengths
●
2.
Program Weaknesses
Recommendations for program improvement generated by self study. (List by
number and use these numbers in the Quality Improvement Plan.)
39
Program Review Survey
This survey should be completed by all faculty working on the Comprehensive SelfStudy Report, the Annual Update, or involved in the Program Review process in
any way.
Thank you for assisting in the improvement of CR’s instructional programs. Please take
a moment to answer the questions below. Your suggestions and comments benefit future
program review processes and participants. Thank you for completing this survey.
Circle One
1. I received proper orientation to understand my role in the program review
process.
Disagree
Agree
Strongly Agree
N/A
Suggestions for improvement:
2. The Program Review Self-Study Guide contained sufficient information to help
me gain a full understanding of the program.
Disagree
Agree
Strongly Agree
N/A
Suggestions for improvement:
3. I feel as though my contribution to the program review process was valued,
appreciated, and respected.
Disagree
Agree
Strongly Agree
N/A
Suggestions for improvement:
4. The program review process was open, honest, and healthy.
Disagree
Agree
Strongly Agree
N/A
Suggestions for improvement:
40
5. For those programs which undergo external accreditation, CR’s program review
process was helpful in completing your external report.
Disagree
Agree
Strongly Agree
N/A
Suggestions for improvement:
6. Overall, the program review process is properly designed to achieve its intended
goal of program improvement.
Disagree
Agree
Strongly Agree
N/A
Suggestions for improvement:
7. Please provide any comments or suggestions on how the Program Review
Committee can improve the process or facilitate your role.
41
Section III
Annual Program Review Update
42
Section III
Annual Program Review Update
Instructions
The Annual Self-Study is conducted district-wide by each program and each academic
area and consists of a) analysis of general changes, staffing, resources, facilities,
equipment and other needs, as well as b) reporting of curricular changes and outcomes
assessment.
The questions on the subsequent pages are intended to assist you in planning for your
program or area. Input should be sought from all campuses. It should be submitted or
renewed every year by the designated date in anticipation of budget planning for the
next fiscal year.
If there is no change from your prior report, you may simply resubmit that report
(or any portion that remains constant) from the prior year with a new date.
The majority of the information you will use for this annual update will be provided by
IR.
Please include pertinent documents such as student learning outcomes assessment reports
(see Appendix A for assessment suggestions), and updated course outlines, as well as
using data analysis to support any requests for new faculty, facilities, equipment, etc.
You are encouraged to use the forms in Appendix B to clarify your requests for
equipment, staffing, funding, etc. to make them easy for large committees to review
quickly.
Please retain this information for your discipline’s use, and submit an electronic copy to
Program Review.
43
Annual Program Review Update
*Be sure to include information from all three campuses.
Program/Discipline: ____________________________
Date:
____________________________
Trends and Relevant Data
1. Has there been any change in the status of your program or area? (Have you
shifted departments? Have new degrees or certificates been created by your
program? Have you added or deleted courses? Have activities in other programs
impacted your area or program? For example, a new nursing program could
cause greater demand for life-science courses.) If not, skip to #2.
Note: curricular changes should be addressed under 12-14.
2. Have there been any significant changes in enrollment, retention, success rates, or
student demographics that impact your discipline? If so, please include data
sheets (Excel or Word format) showing these changes.
3. Occupational programs must review the update of their labor-market data, some
of it provided by Institutional Research, to illustrate that their program:
a. Meets a documented labor market demand,
b. Does not represent duplication of other training programs (in the region), and
c. Is of demonstrated effectiveness as measured by the employment and
completion success of its students.
Other Resources
4. Do you have needs (professional development, library resources, and so forth) not
previously required by the discipline or not previously addressed in budget or
equipment considerations? Please describe.
5. Does your discipline need additional support from Student Services beyond that
previously provided?
Human Resource Needs
6. Complete the Faculty Employment Grids below (please list full- and part-time
faculty numbers in separate rows):
Faculty Load Distribution in the Program
Discipline
Name
(e.g., Math,
English,
Accounting)
Total
Teaching
Load for fall
2006 term
% of Total
Teaching
Load by FullTime Faculty
% of Total
Teaching Load
Taught by PartTime Faculty
44
Changes from
fall 2005
Explanations and
Additional
Information (e.g.,
retirement,
reassignment, etc.)
Faculty Load Distribution in the Program
Discipline
Name
(e.g., Math,
English,
Accounting)
Total
Teaching
Load for
spring 2007
term
% of Total
Teaching
Load by FullTime Faculty
% of Total
Teaching Load
Taught by PartTime Faculty
Changes from
spring 2006
Explanations and
Additional
Information (e.g.,
retirement,
reassignment, etc.)
Do you need more full-time faculty? Associate faculty? If yes, explain why and
be sure to include data sheets justifying the need.
7. Complete the Staff Employment Grid below (please list full- and part-time staff
numbers in separate rows:
Staff Employed in the Program
Assignment
Full-time
Part-time staff
(e.g., Math,
(classified) staff (give number)
English)
(give number)
Gains over
Prior Year
Losses over
Prior Year (give
reason:
retirement,
reassignment,
health, etc.)
Do you need more full-time staff? Part-time staff? If yes, explain why and be sure
to include data sheets justifying the need.
8. If necessary, to clarify your needs, please comment on current available staff and
distribution of FTE's for contract and part-time faculty. Describe strengths and
weaknesses of faculty/staff as appropriate to program's current status or future
development.
Facilities
9. Comment on facilities the program uses, their current adequacy, and any
immediate needs. Have your discipline’s facilities needs changed? If so, how?
Please provide a data-based justification for any request that requires new or
additional facilities construction, renovation, remodeling or repairs.
Equipment
10. Have your discipline’s equipment needs changed? If so, how? Is equipment in
need of repair outside of your current budget? Please provide a data-based
justification for any request that requires a new or additional budget allotment.
45
Learning Outcomes Assessment Update
11. How has your area or program been engaged in student learning outcomes
assessment?
a.
Summarize your results.
b.
What did your program learn from these results that enabled you to
improve teaching and learning in the discipline?
c.
How have part-time faculty been made aware of the need to assess SLOs?
Curriculum Update
(Reminder: Send updated course outlines to the Curriculum Committee.)
12. Identify curricular revisions, program innovations, and new initiatives undertaken
in the last year.
13. Identify curricular revisions, program innovations, and new initiatives planned for
the next year.
14. Complete the grid below
Course
Year Course Outline Year Next Update
Last Updated
Expected
Goals and Plans
15. If you have recently undergone a comprehensive review, attach your Quality
Improvement Plan if applicable.
16. If you do not have a QIP, what goals and plans does your area have for the
coming year?
46
Instructional Program Review
APPENDIX A: Student Learning Outcomes
APPENDIX B: Forms for Annual Update Resource/Staffing Needs
APPENDIX C: Assessment’s Link with Strategic Planning
APPENDIX D: Glossary of Terms
47
Instructional Program Review—APPENDIX A: Course and Program
Outcome Assessment Initiatives
Student Outcomes Assessment at the course level
1.
The Course Outline of Record template requires that each outline contain a list of
student learning outcomes (SLOs) for the course, the core competencies that
instructors expect students to leave the course with. Since all SLOs for every
course will eventually have to be assessed, the Program Review Committee
recommends that disciplines focus on comprehensive student learning outcomes,
so that the typical course will have no more than six SLOs, with fewer than five
entirely possible.
●
●
For courses meeting general education requirements, disciplines are
encouraged to link course SLOs with the institutional general education
(GE) SLOs as much as possible, since this will be a way for us to
assess GE outcomes at the same time we’re assessing the outcomes of
individual courses.
You will be asked by the template to link methods of evaluation and
methods of instruction with specific SLOs, which helps make the SLOs
“operational”—more easily measured.
Evaluate all courses meeting (or potentially meeting) general education requirements
against the institutional general education SLOs, using the Student Learning Outcomes
for General Education. This process should be done collaboratively.
●
CR has agreed on ____basic core student learning outcomes for its general
education program, and all courses meeting GE requirements will link to one (or more) of
those student learning outcomes. But we need evidence that they do. You’ll be asked to
indicate where a particular GE course SLO matches (even in general terms) to an
institution-wide GE SLO.
●
As you revise course outlines, this is a good opportunity to adjust some
course SLOs so that they more explicitly match the overall GE
outcomes. (For example, Ed Code and IGETC requirements mandate that
GE courses have critical thinking and writing SLOs, and now is the
time to make them explicit.) This will permit PRC to assess overall GE
outcomes as your program assesses specific course SLOs.
2.
Do an assessment inventory and write a brief report about assessment efforts
already undertaken in your discipline. Start with what you are already doing,
even if you’re not calling it assessment.
48
Questions to consider:
●
●
●
●
●
3.
What are the expected career/occupational outcomes for students in your
discipline (e.g., transfer, employment)?
Where do faculty members already evaluate student work collaboratively?
Are there written criteria that guide this evaluation? Would such criteria
be useful if they don’t presently exist?
Do faculty members ever report, formally or informally, about strengths
and weaknesses in student work? Does information about student learning
ever become a topic for discussion at discipline or department meetings?
Could it be?
○
Sometimes detailed minutes of division meetings contain
extremely useful assessment information, particularly when faculty
are encouraged to talk about what they see as areas where students
are having trouble.
What changes has your department or discipline undertaken recently to
improve student learning? What beliefs or information led you to make
these changes?
What information regarding student success in your department or
discipline might already exist that might lead you to reconsider teaching
practices?
○
IR may be able to give you invaluable data for strengthening
student learning.
Begin to develop and implement a comprehensive plan to assess course-level
SLOs in either of the ways described below, depending on the particular nature of
your discipline:
a.
Option 1
If at least one course in the discipline is taught by a number of different
instructors, choose such a course and work collaboratively to assess
learning outcomes in the course. You may want to focus on a single SLO
in the initial effort or perhaps several related SLOs. Involve part-time
instructors in the process as much as possible. Focus on an assignment or
examination toward the end of the semester in which students can b e
expected to demonstrate what they’ve gained from the course.
Depending on the discipline, you might develop a rubric by which to
assess learning in sample essays (or other kinds of student work) taken
from many different sections of the course, or—if your discipline
routinely employs objective tests as a way of evaluating students—you
may simply want to embed common questions in such tests and aggregate
results to see where students are having success or difficulty.
49
b.
●
Keep careful records of these course-based assessment projects,
which should include sections on methodology, results, and
analysis of results.
●
Sometimes minutes of department meetings are useful appendices
to show how results are used to improve teaching and learning.
Option 2
If you teach in a discipline with few or no courses taught by multiple
faculty members, you may not find collaboration feasible or desirable. In
that case, ask every instructor (full- and part-time) to choose a course and
develop an individual assessment project for it.
Ask instructors to identify a significant SLO for a course they teach, then
choose a major graded assignment in that class that they believe measures
that SLO. Next, have them develop a rubric (or, for objective
examinations, some other method of analysis) by which achievement of
that SLO can be assessed. Once the course, SLO, and assignment are
chosen and the method of assessment created, have instructors do the
actual assessment.
Finally, ask instructors to report on and analyze the results of the
assessment, in writing. It’s useful to plan some discipline meeting time
for discussion of these results, but even more vital that they be captured in
written form. Make sure that the reports describe how the instructor
would change or improve the teaching of this assignment (and the
assignment itself). Discipline discussion should identify common
problems instructors are having and potential solutions to those problems.
4.
Provide a time line for implementing a plan that would systematically assess
learning in other courses in the discipline.
NOTE: Assessing student learning at the course level is a daunting process. It will not
happen overnight, and no one expects it to. PRC believes it has developed guidelines that
are not overly burdensome, while still enabling disciplines to generate the kind of
information that will be of maximum benefit, to your discipline and to the college as a
whole. Please bear in mind that this must be an ongoing process. For that reason, it’s
vital not to conceive of assessment as something we do only when our discipline is up for
program review. PRC is available to help you with all parts of the assessment process.
50
Program Outcomes Summary
Program Name:
Effectiveness
Indicator
Performance Standard
Measurement/Data
Source
Responsible
for collecting
data
I. Effectiveness
A. Quality of Education
Student
Satisfaction
At least 85% of AS program graduates Exit and Alumni
will indicate they are satisfied with the Surveys
quality of their education.
Professional
Licensure
Program graduates will successfully
Board and/or
pass professional licensure exams at or Certification Exam
above available norms (national or
Scores
state).
Employer Satisfaction
At least 85% of employers will
indicate program graduates are
academically prepared for
employment.
Employer Survey
IR
At least 85% of employers will
indicate program graduates are
competent in general education.
Employer Survey
IR
100% of faculty must meet ed code or
WASC criteria.
Faculty personnel
file
Academic
Affairs
100% of faculty must meet external
accreditation standards
Faculty personnel
file
Program Coordinator
Faculty Credentials
51
IR
Program
Coordinator
(year) (year) (year) (year)
Program Name:
Effectiveness
Indicator
Performance Standard
Measurement/Data
Source
Responsible (year) (year) (year) (year)
for collecting
data
B. Vitality
Job
Placement
Annual Unduplicated
Headcount of Program
Majors
Advisory Committee
Participation
At least 90% of program graduates will Graduate/alumni survey
secure employment in their field of
study within six months of graduation.
At least 12 students enrolled in
program in most recent fall term.
At least 75% of external members will Advisory Committee
attend one meeting within the
Members
academic year.
IR
IR
Program Coordinator
Advisory Committee
At least 85% of Advisory Committee
survey respondents will indicate they
agree or strongly agree with survey
items.
Program Advisory
Committee Survey
IR
Graduates
At least ? graduates per year.
Data
IR
Student Retention
Adequate to ensure program
efficiency.
Fall to fall retention
data excluding transfers
and readmits
IR
52
Program Name:
Effectiveness
Indicator
Performance Standard
Measurement/Data
Source
Responsible (year) (year) (year) (year)
for collecting
data
II. Efficiency
Program Coordinator
Annual Program Costs
Annual Program Report
Space Utilization
Annual Program Report
Average Section Size/Program
Enrollment
Annual Program Report
Date
53
IR
Instructional Program Review—Appendix B: Annual Update Forms
Campus/Program Needs Worksheet
Annual Program Review Update
Resources
1.
2.
3.
4.
54
Low
Moderate
High
List Resources Needed for Academic Year___________________
Not
Approved
Very
High
Please list/summarize the needs of your program on your campus below
This section to be filled out by
Subcommittee
Degree of
Justification (as
Approval
substantiated by
Status
the program
review)
Approved
This section to be filled out by the program at each campus
Campus/Program Needs Worksheet
Annual Program Review Update
Faculty
1.
2.
3.
4.
5.
55
Low
Moderate
High
List Faculty Positions Needed for Academic Year___________________
Not
Approved
Very
High
Please list/summarize the needs of your program on your campus below
This section to be filled out by
Subcommittee
Degree of
Justification (as
Approval
substantiated by
Status
the program
review)
Approved
This section to be filled out by the program at each campus
Campus/Program Needs Worksheet
Annual Program Review Update
This section to be filled out by the program at each campus
This section to be filled out by
Subcommittee
Please list/summarize the needs of your program on your campus below
Approval
Status
List Staff Positions Needed for Academic Year___________________
Approved
Staff
1.
2.
3.
4.
5.
56
Low
Moderate
High
Not
Approved
Very
High
Degree of Justification (as
substantiated by the
program review)
Campus/Program Needs Worksheet
Annual Program Review Update
Equipment
1.
2.
3.
4.
5.
6.
57
Low
Moderate
Approximate
Cost
High
List Equipment or Equipment Repair Needed for Academic
Year___________________
Not
Approved
Very
High
Please list/summarize the needs of your program on your campus below
Approved
This section to be filled out by
Subcommittee
Approval
Degree of
Status
Justification (as
substantiated by
the program
review)
This section to be filled out by the program at each campus
Campus/Program Needs Worksheet
Annual Program Review Update
Facilities
1.
2.
3.
4.
5.
6.
58
Low
Moderate
Approximate
Cost
High
List Facility Needs for Academic Year___________________
(Remodels, Renovations or added new facilities)
Not
Approved
Very
High
Please list/summarize the needs of your program on your campus below
Approved
This section to be filled out by
Subcommittee
Approval
Degree of
Status
Justification (as
substantiated by
the program
review)
This section to be filled out by the program at each campus
Instructional Program Review—Appendix C: Assessment’s Link with
Strategic Planning
The model below designed by Nichols (Nichols, 1997) shows the relationship of assessment to strategic
planning. Positioned at the apex of the triangular model is the mission of the college. Both the development
of strategic planning on the left side of the model and assessment practices on the right side flow from the
core elements articulated in the College of the Redwoods’ mission statement.
The right side of the model represents assessment practices and activities from which the College learns of
ways to improve academic programming through the use of assessment results. The intended results, or
expected educational results referred to in the WASC criteria, appear in the assessment side of the model
along with the means of assessment, assessment results, and uses of results.
The left side of the model represents the strategic planning process. Planning on this side of the model is
guided by the underlying premise that strategic activities in academic, administrative, budget, and facilities
planning are necessary to support the mission of the College.
In concert, the information gleaned from assessment practices and use of those results will drive the
strategic planning of the college to improve the overall mission of the college. The two sides of the triangle
must be linked by budgetary support. When the model is viewed in its entirety it represents institutional
effectiveness.
59
Instructional Program Review—Appendix D: Glossary of Terms
Assessment: the process of defining, selecting, collecting, and using information to
analyze and interpret student learning.
Benchmark: a standard by which performance can be measured or judged.
Criteria, assessment: a standard of judgment; describes the cognitive and affective
outcomes of the educational program; specifies the content knowledge to be mastered.
Evaluation: the judgments applied to assessment efforts.
General Education Program: A core curriculum organized to promote well-educated
individuals who have attained essential, agreed-upon competencies.
Impact Statement: a measurable expression of the outcome of a learning experience.
Indicator, core: a gauge used to monitor essential condition or performance; usually
described in numerical values or indices which together illuminate the condition or
direction of educational outcomes being studied.
Institutional Effectiveness: the process of articulating the mission of the college, setting
goals, defining how the college and community will know when the goals are being met,
and using the data from assessment in an ongoing cycle of goal-setting and planning.
Intended student outcomes; formal statements of the intended effects or consequences
of a course of instruction.
Measure, measurement: a set of questions or an inquiry designed to capture evidence on
student performance.
Stakeholders: persons or groups who have vested interests in the outcomes of the
institution (e.g., faculty, staff, employers, licensing agencies, students).
Student learning: elements of student learning include knowledge, aptitude, skills, and
perceptions.
60
SECTION V
Student Services and Learning Support and Administrative Services
Program Reviews
Introduction
The Program Review Resource Guide developed by the Program Review Committee
serves as the basis for the Student Services and Administrative Services program review
process. Student Services and Administrative Services has representation on the
Committee to ensure that the review timeline, procedures, and evaluative criteria used in
the Student Services and Administrative Services reviews are relevant to the college’s
broad program review effort.
This section of the Program Review Resource Guide provides guidelines for use in
Student Services and Administrative Services program review activities. These
guidelines are indicative of Student Service’s and Administrative Service’s commitment
to provide the information and coordination needed to support the district-wide program
review process.
Purpose and Objectives
Program Review is a time consuming effort; however, it is important to note the positive
impact program review will have on planning, decision-making, transformation, and
stature in the college for student services and administrative services departments. The
program review process gives Student Services and Administrative Services an
opportunity to: 1) assess a department’s current service, organizational effectiveness and
efficiencies, and resources; 2) uncover and validate their strengths; 3) develop program
and/or student learning outcomes; 4) identify strategies and potential transformations to
address the department’s challenges; and 5) provide an important communication tool
within a department and between the department and the administration.
The objectives of Student Services and Administrative Services program reviews are to:
•
•
•
•
•
Assess how well the department is performing in relation to its goals, the college
strategic plans, and the division’s goals.
Encourage strategic thinking about the department’s future plans.
Provide evidence of the department’s effectiveness.
Evaluate the extent to which a department has successfully addressed program
and/or student learning outcomes.
Uncover the challenges that may prevent a department from achieving its goals.
Principles of Program Review in Student Services and Administrative Services
Principle #1: The primary purpose for program review is to make sure that the programs
61
and offerings are connected to the mission, goals, and priorities of the division and the
college.
Principle #2: All student services and administrative services units contribute to the work
of the divisions and therefore will be a part of the program review process.
Principle #3: Broad faculty, staff and student participation in the reviews is essential in
order to gain important insights and ideas for improvement as well as a broad recognition
and understanding of department strengths.
Principle #4: Program reviews are independent of any other type of review but, to ease
workload and reduce unnecessary duplicative efforts, should be scheduled to compliment
other department specific accrediting processes.
Coordination of the Program Review Process
Responsibilities for carrying out reviews will be shared between the Office of the Vice
President for Student Services, the Office of the Vice President, Chief Business Officer,
the Vice President of Human Resources, and the departments under review. Program
review responsibilities handled by the Vice Presidents include:
•
•
•
•
•
•
•
Establishing and maintaining the overall Student Services and Administrative
Services Program Review Schedule of Reviews and coordinate this schedule with
the college’s Program Review Committee.
Managing individual department program review calendars.
Notifying departments scheduled for review.
Planning and conducting department program review orientation meetings with
participation from Institutional Research.
Reviewing department self-studies for completeness of information.
Monitoring improvement plan follow-up reporting, and
Evaluating the review process itself and implement recommendations for
improvements/revisions to the program review process as deemed necessary.
The Program Review Area Coordinator will serve a team leader for the department under
review. The Coordinator’s responsibilities include:
•
•
•
•
•
•
•
Working with program members to prepare report.
Recruiting external and student review team members.
Finalizing team membership.
Arranging initial meetings with review team members and scheduling future
meetings.
Distributing the Program Review Guidelines and the completed Program Review
Report to review team members.
Establishing review team member’s responsibilities/duties.
Leading review team activities and assuming responsibility for team’s preliminary
report.
62
•
•
Providing any supplementary information/documentation needed by review team
members.
Making any factual corrections to the final draft.
Program Review Cycle
Student Services and Administrative Services programs will undergo review on a yearly
basis. The comprehensive self-study program review process will be completed by each
program once every five years. The annual program review update is to be completed by
each program in the years the comprehensive review is not required. The program
review schedule may be revised to align with state and/or federal program review
reporting requirements.
63
Schedule of Reviews for Student Services and Learning Support
C=Comprehensive Program Review
Unit
Admissions &
Records
Residential Life and
Campus Activities
Judicial Affairs
EOPS
Counseling and
Advising
DSPS
CalSOAP
Upward Bound
Career
Development/Student
Employment
Library
Academic Support
Center
Athletics
Student Health
Center
A=Annual Update
Academic Year
2006- 2007- 200807
08
09
C
A
A
200910
A
201011
C
201112
A
201213
A
201314
A
C
A
A
A
C
A
A
A
C
C
C
A
A
A
A
A
A
A
A
A
C
C
C
A
A
A
A
A
A
A
A
A
C
C
C
C
A
A
A
A
A
A
A
A
A
A
A
A
C
C
C
C
A
A
A
A
A
A
A
A
A
A
A
A
C
C
A
A
A
A
A
A
C
C
A
A
A
A
A
A
NA
NA
C
C
A
A
A
A
A
A
C
C
A
A
A
A
64
Schedule of Reviews for Administrative Services
C=Comprehensive Program Review
Unit
A=Annual Update
Academic Year
20072008- 200908
09
10
A
C
A
201011
A
201112
A
201213
C
2013-14
Bookstore
A
A
A
C
A
A
A
Facilities and
Grounds
A
A
C
A
A
A
C
Financial Aid
A
C
A
A
A
C
A
Fiscal Services
C
A
A
A
C
A
A
Food Service
A
A
C
A
A
A
C
Information
Technology
Services
Security
VP, CBO
Human
Resources
Communications
and Marketing
Distance
Education
A
C
A
A
C
A
A
A
A
C
A
C
A
A
A
A
C
A
A
A
A
C
A
C
A
A
A
A
NA
A
A
A
C
A
A
C
A
A
A
C
A
A
Arcata
Instructional
Site
A
Program Review Process and Timeline
The comprehensive program review process consists of six steps. These steps are: 1) prereview planning, 2) unit self-study, 3) external review, 4) writing of the final program
review report, 5) creating the transformation improvement activities, and 6)
implementing the department transformation improvement plan.
The normal timeline required for a department program review includes a pre-review
preparation period that establishes the groundwork for the review and 10-12 months for
the review itself. The 10-12 months schedule creates a timeline that serves to structure
and standardize the review process.
65
Step 1: Pre-Review Planning
Each program should select program representatives who will communicate with
Institutional Research and the Program Review Committee and provide organizational
leadership for completion of the self-study. The pre-review planning meeting will:
•
•
•
•
Familiarize programs with ACCJC/WASC accreditation standards, the program
review process, and program and student learning outcomes assessment.
Allow programs to begin the self-study with hands-on training in data
interpretation and assessment techniques.
Assist programs in the development of a project plan with timelines and
assignment of responsibilities.
Generate discussion about the area self-study process, including:
1. Looking at the data provided by Institutional Research that have informed
your document, discuss it with your colleagues, and identify major changes or
trends you expect to be of particular relevance to your area in the next five
years.
2. Dialoguing about the implications of these trends for your area and what
changes your area will need to make.
3. Commenting on the trends and indicate plans as a result of that analysis.
What trends, if any, exist in the data? Why might these trends be occurring?
Are there additional data that would assist your area in analyzing these trends?
Step II: Department Self-Study
The self-study represents a valuable opportunity for the department to make a candid
assessment of itself and to consider future directions and options that would strengthen it.
Each department will prepare a self-study report using as its framework the
Comprehensive Workbook. The workbook is intended to provide a structure for the
review and should be augmented by whatever information is deemed necessary. The selfstudy should be shared with all staff of the department and should represent consensus.
The timeframe for a typical Comprehensive Program Review self-study review process
is:
Summer
• Prepare draft documents resulting from dialogue within the program
• Develop detailed task list and timeline
• Review data packet from Institutional Research
• Discuss self-study guide
• Gather any additional information the program deems necessary
• Update course outlines and submit to Curriculum Committee if needed
• Discuss, and revise if necessary, department mission, goals, learning
outcomes, and assessment plan
November/December/January/February/March
• Review, revise, and finalize the Program Review document
66
April/May
• Submit to the appropriate review bodies
Step III: External Review
The external peer review procedure should be flexible and may be influenced by the
results of the self-study. Generally, the external reviewer will conduct the review based
on the department’s self-study report. The external reviewer will prepare a summary of
recommendations with emphasis given to strengths and challenges of the department and
how the department can continue to improve the quality of its programs and services. The
summary will be included in the Final Program Review report.
Step IV: Writing of the Department Review Report
The department’s Program Review Coordinator is responsible for writing the final
program review document. The final program review will be forwarded to the Vice
President for Student Services and Learning Support, Vice President, Chief Business
Officer, or the Vice President for Human Resources for review and comment. The Vice
President will forward the report to the college’s Program Review Committee. Following
receipt of the report, a meeting is scheduled with the Program Review Committee, Vice
President of Student Services, Vice President, Chief Business Officer, Vice President for
Human Resources, Department supervisor, and members of the department’s
management team (if appropriate) to discuss the findings and the recommendations
presented in the final review.
Step V: Creating Transformation Activities
Following consultation with the Vice President for Student Services, Vice President,
Chief Business Officer, or the Vice President for Human Resources and the Department
supervisor, the department review coordinator will develop a final three year
transformation improvement plan that addresses the recommendation(s) outlined in the
Program Review Report for the Institutional Effectiveness Committee. The final
transformation improvement plan should specify proposed actions, implementation
strategies, and who is responsible for carrying out each action. If there are review
recommendations that the department is not in agreement with, the transformation action
plan should acknowledge these differences in thinking and where appropriate, present
alternative recommendations.
Step VI: Implementing the Department Transformation Plan
The Office of the Vice President will evaluate the progress on the department’s
transformation plan via the Program Review Yearly Update and the annual planning
processes.
67
Comprehensive Program Review Distribution of Activities Timeline and
Responsibilities
Date
Institutional Research
Ongoing
IR provides the Area/Program Coordinator with all necessary data to complete
the program review
Date
Vice President of the Division
June
SS&LS/AS creates assessment tool(s) and makes necessary adjustments to
the process for the following year
July
SS&LS/AS appoints the Area/Program Review Coordinator
Aug
SS&LS/AS provides orientation to the Area/Program Review Coordinator
Aug
SS&LS/AS provides Program Review Guide and the Program Review
template to the Area/Program Review Coordinator
Mar
SS&LS/AS monitors the submission of the Program Review to the appropriate
review bodies
Date
Area/Program Review Coordinator
July
Coordinator works with department to review and revise if necessary
mission statements, goals, learning outcomes, and assessment plans
Fall
Coordinator works with program members to prepare report
Sept
Coordinator recruits external reviewer and finalizes team membership
Sept
Coordinator arranges first meeting with review team members and schedules
future meetings as necessary
Sept
Coordinator distributes the Program Review Guide and the completed
Program Review Report to all review team members
Sept
Coordinator establishes review team member’s responsibilities/duties
Nov
Coordinator leads review team activities and assumes responsibility for
team’s preliminary report
Mar
Coordinator provides any supplementary information/documentation
needed by review team members
68
Apr
Coordinator makes any factual corrections to the final draft
Apr
Coordinator prepares an action plan addressing any program recommendations
identified including who is responsible for the action and the date of expected
completion
Apr
Coordinator sends one electronic and one hardcopy of the action plan to
the Office of Student Services and Learning Support or Administrative
Services by the appropriate deadline
Note: Exact dates will be established and provided for critical steps in the process.
Comprehensive Program Review Department Self-Study
Step 1: Department/Program Description
The self-study begins with a description of the department/program including information
about the area’s purpose, mission, background, and goals. Provide a short narrative on
how the department contributes to and advances the College’s goals and objectives.
Step 2: Past Quality Improvements
Provide a description (list or narrative) of the quality improvements that were
recommended from the department’s previous program review. Also discuss
improvements implemented by the department that were not specifically noted on the
previous program review.
Step 3: Outcomes Assessment
Address how successful the department is in determining intended program learning
outcomes and student learning outcomes, user satisfaction, and other noteworthy
performance results. Discuss the relevance of key performance measures the department
uses to track results. Included in this section is a discussion of the instruments used to
measure performance (surveys, benchmarking, internal or external review, focus groups,
etc.) and what assessment methods were used to determine whether the program
accomplished the program learning outcomes and/or students achieved the student
learning outcomes.
Departments should summarize the findings of the assessments and discuss the team’s
interpretation of the data as well as the strengths and challenges of the department as
indicated by the performance outcomes indicators.
The department program review team should provide concrete suggestions for a
transformation improvement plan including ways to build upon current strengths,
overcome challenges, and increase effectiveness and efficiency. Offer specific
transformation strategies, and timetable, for addressing the challenges affecting the
department.
69
This section of the workbook may be revised as appropriate for the department under
review.
Step 4: Cost Effectiveness
This is an opportunity to summarize the financial resources of the department.
Departments should comment on significant budget changes or variances from year to
year. How has the funding level changed over the past five years? What factors have
influenced these changes? What are the anticipated effects of this funding on
program/service delivery?
Step 5: Recommendations and Comments
Step five allows the department program review coordinator and outside readers to
comment on what they see as the strengths and challenges of the department. The
comments could encompass key program/service operational, resource and campus
environment challenges the department faces.
Step 6: Publication and Discussion of the Findings
Fully approved program reviews will be placed online and distributed to the District
Program Review Committee, the Division Vice President, the Institutional Effectiveness
Committee, and internal governing bodies.
Suggested List of Documents for Inclusion in the Comprehensive and Yearly
Summary Report
•
•
•
Organizational Chart
Department planning documents
Assessment results/reports including:
1. Surveys
2. Benchmark data
3. Data that measure progress of performance goals
4. Student learning outcomes data
5. Program learning outcomes data
70
APPENDIX A
Assessment and Transformation
Comprehensive Program Review Workbook
Review Coordinator
Review Team Members
Year Reviewed
71
SECTION I: Department Background, Purpose and Goals
A. Department Mission:
B. Background
C. Department Goals:
D. Describe how the goals of the program relate to the College’s Mission and
Strategic Plan.
SECTION II: Past Quality Improvements and Transformations
A. Obtain a copy of the program’s previous program review. Provide a description
(list or narrative) of the quality improvements that were recommended.
Recommended Quality Improvement
Accomplished?
(Y/N)
If Yes, Provide Details.
If No, Provide Explanation.
B. Provide a description (list or narrative) of other improvements accomplished
since the last review not specifically noted on the previous program review.
Other Improvements Accomplished
Provide Details
SECTION III: Outcomes Assessment
A. Provide the department’s program and student learning outcomes. Include or
attach a copy of the Outcomes Assessment Report.
Program Outcome(s)
Outcome
Process/Activities Contributing to
the Outcome (What does the
program do to create the
outcome?)
Measured/Documented By?
Student Learning Outcome(s)
Knowledge, value or skill
(What is being demonstrated
and assessed?)
Scope of Outcome
(In what context,
with whom, or in
what setting does
the outcome
occur?)
Process/Activities
Contributing to the
Outcome (What does
the program do to
create the outcome?)
72
Student Learning
Outcome
Statement
Measured/
Documented By?
Outcomes Assessment Report
Outcome
Assessment
Method
To Whom?
By Whom?
Administered
When?
Use of Results
Assessment
Status/
Findings
Transformations
B. What is the team’s interpretation of the data?
C. What are the strengths of your department as indicated by your performance
indicators?
D. What are the challenges for your department which are indicated by your
performance indicators?
E. What is your three year transformation plan to build upon current strengths,
overcome challenges, and increase effectiveness and efficiency? (Include a
timetable for implementing transformations)
STRATEGIES TO ADDRESS CHALLENGES/TRANSFORMATIONS
TIMETABLE FOR
IMPLEMENTATION
F. What resources are required (Personnel, funding, equipment, etc)?
73
Intended
Follow-up
SECTION IV: Cost Effectiveness
A. Indicate your budget figures and expenditures in recent years:
Categories
FY 2001
FY 2002
FY 2003
FY 2004
Operating Exp.
Budgeted
Expended
Other (if applicable) (
Budgeted
Expended
SUBTOTAL
Budgeted
Expended
Capital/Equipment
Budgeted
Expended
TOTAL
Budgeted
Expended
Surplus or (Deficit)
Generated Revenue
(if applicable)
Revenue Budget
Revenue Actual
B. Comment on significant budget changes or variances from year to year.
74
FY 2005
SECTION V: Conclusion
Department Comments:
Summary of discussion and findings relating to this review:
___________________________________________________ ____ / ____ / ____
Department Supervisor’s Signature
Date
External Reviewer Comments:
Summary of discussion and findings relating to the program review:
________________________________________________
____ / ____ / ____
External Reviewer Signature
Date
Program Review Committee Comments
Comments:
__________________________________________________
Co-Chair of the PRC
_____/_____/_____
Date
Academic Senate’s Comments
Comments:
__________________________________________________
Co-Presidents of the Senate
_____/_____/_____
Date
Institutional Effectiveness Committee’s Comments
Comments:
______________________________________________________ _____/_____/_____
Chair, Institutional Effectiveness Committee
Date
75
APPENDIX B
Assessment and Transformation
Program Review Yearly Update Report
Review Coordinator
Review Team Members
Year Reviewed
76
A. Outcomes Assessment Plan Report
Outcome
Assessment
Method
(Indicate the
assessment
methodology to
be used, or was
used, such as
focus group or
exit interview.)
To Whom?
By Whom?
Administere
d When?
Use of
Results
Assessment
Status/
Findings
(Indicate when
the data will be
available for
analysis or the
result of the
analysis. In the
next year,
indicate the
results of a
second
assessment
cycle
Transformations
(Indicate what
transformations
were
implemented as a
result of
assessments.)
Intended
Follow-up
B. Department Goals Report
Division Strategic
Direction
Division Goal
Department
Goal
FY2007-08
Action
Plan
Status of each
action item
Completed
In Progress
Collaboration
Partners
Resources Needed /
Reallocated
Staff
Staff Time
Completed
In Progress
New Money
Reallocated $
Space
Equipment
Other:
Completed
In Progress
Staff
Staff Time
Completed
In Progress
New Money
Reallocated $
Space
Equipment
Other:
77
C. Improvements or Rationale for Action: (Provide a brief summary of the major
accomplishments and achievements of your department during the past year.)
Vice President’s Recommendation
Comments:
______________________________________________________ _____/_____/_____
Vice President’s Signature
Date
Program Review Committee Comments
Comments:
__________________________________________________
Co-Chair of the PRC
_____/_____/_____
Date
Institutional Effectiveness Committee’s Recommendation Comments
Comments:
______________________________________________________ _____/_____/_____
Chair, Institutional Effectiveness Committee
Date
78
APPENDIX C
Program Review Committee
Student Services and Administrative Services Comprehensive Program
Review Evaluation Form
Program Title_________________________________________________________
Date and Time of the program review______________________________________
The PRC will provide the program with an external perspective concerning the
administrative and educational aspects of the program. This perspective includes a
judgment of program quality and an assessment of the effectiveness of the program.
Evaluation
1.
Does the program have mission and goals statements in place?  Yes
 No
2.
Are program goals consistent with the college’s mission?
 Yes
 No
3.
Has there been progress in implementing the last program review's
recommendations for program improvement?
 Yes
 No
4.
Has the program identified program and/or student learning outcomes that are
relevant to its mission?
 Yes No
5.
Are appropriate instruments used to assess learning outcomes?
 Yes
 No
6.
Are results of assessments used to improve the program?
 Yes
 No
7.
Are program strengths clearly articulated?
 Yes
 No
8.
Are program challenges clearly articulated?
 Yes
 No
Comments or suggestions to strengthen the program and/or the Program Review:
79
Program Review Committee
Student Services and Administrative Services Annual Program Review
Evaluation Form
Program Title_________________________________________________________
Date and Time of the program review______________________________________
The PRC will provide the program with an external perspective concerning the
administrative and educational aspects of the program. This perspective includes a
judgment of program quality and an assessment of the continuous improvement of the
program.
Evaluation
1. Has the program identified program and/or student learning outcomes that are
relevant to its mission?
 Yes
No
2. Are appropriate instruments used to assess learning outcomes?
 Yes
 No
Are results of assessments used to improve the program?
 Yes
 No
4. Are program goals consistent with the college’s mission?
 Yes
 No
5.
 Yes
 No
3.
Are improvements or rationale clearly articulated?
Comments or suggestions to strengthen the program and/or the Program Review:
80
Assessment Glossary
Program Mission: A description of the overall purpose of the program. Program
mission statements should align with the College and Division mission statements.
Program Goals: Goals are derived from the mission and reflect critical aims of the
program to be measured by achievable outcomes.
The mission and goal statements define what we do.
Outcomes: Outcomes are the expected product from the achievement of our goals.
Outcomes define why we do what we do.
Program/Administrative Outcome: Expected outcomes from program/administrative
processes to measure the efficiency and effectiveness of the program. Example: Students
will receive guidance and tools to chart their progress towards goals (measurements could
be # of Student Education Plans on file and/or satisfaction levels with specific
counseling/advising processes).
Student Learning Outcome: Something that we expect students to be able to do after
participating in a program or receiving services. Example: Transfer students will
demonstrate an understanding of general education transfer requirements (could be
documented by qualitative assessment through the advising process, surveys or # of
successful transfer students).
Data: Provides information for dialogue about the learning and necessary evidence to
determine if a transformation is needed.
Measurement: A quantitative data indicator or set of data indicators intended to reflect
the achievement of an outcome.
Documentation: Qualitative data or evidence intended to reflect the achievement of an
outcome.
Assessment Method: The data collection tools and methods that are utilized to produce
measurements and documentation.
Result: The level of performance for a specific outcome.
Baseline: The level of performance for a specific outcome in the first year of assessment.
Benchmark/Performance Target: The desired level of performance for a specific
outcome.
Use of Results:
• What does the data tell us about process? Goals; outcomes; means of
assessment; implementation results, data collection and analysis.
81
•
What does the data tell us about program improvement? What, if anything, do
we need to do to improve program and student learning outcomes? What
resources are necessary?
Transformations: What improvements will be implemented to improve the program
and/or student learning process?
Intended Follow-Up: Quantitative or qualitative data methodologies intended to clarify
initial results and/or measure transformations.
Data Element Definitions
Data element definitions will be aligned with California Community College
common definitions, when possible, in order to promote a shared understanding of
the data elements. The list of operational definitions below is taken from the
Research and Planning group of California. The list of data elements definitions
will continue to expand as measurements for program and student learning
outcomes are identified.
Operational Definitions from the Research and Planning Group of California
Common definitions have been established by the RP Group so that meaning is added to
data analyzed consistently over time, in relation to appropriate comparison groups used
by CA community college researchers. The definitions were adopted by the RP Group
(RP Group Executive Board Approval), after considerable consultation was provided by
researchers across the state. Definitions are related to Associate degree and nonAssociate applicable credit courses and do not include non-credit courses.
Course Enrollment:
Student is enrolled in course and receives a valid grade notation a the time the college
MIS data tape is submitted to the State Chancellor’s Office.
Valid grade notations
A, B, C, D, F, CR, NC, W, I (Incomplete)
Invalid grade notations
No grade of record (NGR), Report delayed (RD), In Progress (IP).
Course Retention I
(For comparing colleges to each other and for comparing departments) Student is
retained in the course to end of term. A, B, C, D, CR, I grade notations.
Retention Rate
Percent of students retained in courses out of total enrolled in courses. The retention rate
is calculated by dividing the numerator by the denominator and multiplying by 100.
82
Numerator
Number of Students (duplicated) with A, B, C, D, CR, I.
Denominator
Number of students (duplicated) with A, B, C, D, F, CR, NC, W, I.
Course Retention II
(For comparing colleges to each other and for comparing colleges to the CCC systemwide rate) Student is retained in the course to end of term. A, B, C, D, F, CR, NC, I
grade notations.
Retention Rate
Percent of students retained in courses out of total enrolled in courses. The retention rate
is calculated by dividing the numerator by the denominator and multiplying by 100.
Numerator
Number of students (duplicated) with A, B, C, D, F, CR, NC, I.
Denominator
Number of students (duplicated) with A, B, C, D, F, CR, NC, W, I.
Course Completion
None. Thought to be too similar to and less relevant than retention/success measures.
Course Success
Student succeeds in the course. A, B, C, CR grade notations.
Success Rate
Percent of students successful in courses out of total enrolled courses. The success rate is
calculated by dividing the numerator by the denominator and multiplying by 100.
Numerator
Number of students (duplicated) with A, B, C, CR.
Denominator
Number of students (duplicated) with A, B, C, D, F, CR, NC, W, I.
Term to Term Persistence
Student persists from one term to the next term.
First Term
Student is enrolled in at least one course. A, B, C, D, F, CR, NC, W, I grade notations.
Next Term
Student is enrolled in at least one course. A, B, C, D, F, CR, NC, W, I grade notations.
83
Persistence Rate
Percent of students enrolled in next term out of students enrolled in first term. The
persistence rate is calculated by dividing the numerator by the denominator and
multiplying by 100.
Numerator
Number of students with at least one course with A, B, C, D, F, CR, NC, W, I in the next
term (x100).
Denominator
Number of students with at least one course with A, B, C, D, F, CR, NC, W, I in the first
term.
Transfer Ready I
The rate by which a first-time entering freshmen cohort meets the basic transfer
requirements for admittance to California State University.
Transfer Directed
All freshmen enrolled in any transfer level English course and in any transfer level math
course within three years, and within six years of admission.
Transfer Ready II
The rate by which the transfer directed student cohort completes 60+ transfer units with a
2.0+ GPA at both three and six years, including the successful completion of any transfer
level English and any transfer level math course, earning grades A, B, C, or Credit.
84
SECTION VI
Calendar of Instructional Program Review Activities, Fall ‘07
Group I – Annual
Update
Annual Review
Due
PRC Review
Meeting
Revisions
PRC meets to
Due to PRC summarize
results
10/3/07
10/5/07
MSE Division
Reviews:
AG, Astro, Biol,
Chem, Eng, Envsc,
For, Geog, Geol,
Math, Meteo, MST,
NH, NR, Ocean,
Physc, Phys,
Sci-Prep
9/17/07
9/28/07
9/17/07
9/28/07
10/3/07
10/5/07
9/17/07
9/28/07
10/3/07
10/5/07
PE: Dance, HE, PE
AJ: AJ
Group A Comprehensive
BT, BUS, CIS, DM,
Addiction Studies
Comprehensive
Review Due
PRC Review
Meeting
10/1/07
10/12/07
Annual Review
Due
PRC Review
Meeting
Group II – Annual
Update
ALSS Division
10/15/07
Reviews:
Anth, Art, Drama,
Frnc, Germ, Ital, Jpn,
Music, NAS, Polsc,
Psych, Snlan, Soc,
Span
HO Division
10/15/07
Reviews:
HO, DA, MA, Lvn,
Nurs
85
Revisions
PRC meets to
Due to PRC summarize
results
10/17/07
10/19/07
10/26/07
Revisions
PRC meets to
Due to PRC summarize
results
10/31/07
11/2/07
10/26/07
10/31/07
11/2/07
Group III – Annual
Update
Hum/Com Division
Reviews:
Cine, Comm, ECE,
Engl, Guid, GS, GC,
Hist, IDS, Journ,
Phil, Spch
BTech Division
Reviews:
Econ, CET, HRC,
IT, RE, WT,
Group B Comprehensive
AT, DT, MT, DHET,
CT
Annual Review
Due
PRC Review
Meeting
Revisions
PRC meets to
Due to PRC summarize
results
10/29/07
11/9/07
11/14/07
11/16/07
10/29/07
11/9/07
11/14/07
11/16/07
Comprehensive
Review Due
PRC Review
Meeting
Revisions
PRC meets to
Due to PRC summarize
results
12/5/07
12/7/07
11/19/07
11/30/07
86
2008-2009
Draft Schedule For Instructional Annual Updates and Comprehensive Reviews
Calendar of Annual Review Update Activities, Fall ‘08
Group I – Annual
Update
MSE Division
Reviews:
Astro, Biol, Chem,
Eng, Envsc, Geog,
Geol, Math, Meteo,
Ocean, Physc, Phys,
Annual Review
Due
PRC Review
Meeting
Revisions
PRC meets to
Due to PRC summarize
results
10/8/08
10/10/08
9/22/08
10/3/08
9/22/08
10/3/08
10/8/08
10/10/08
9/22/08
10/3/08
10/8/08
10/10/08
PE: Dance, HE, PE
AJ Classes
Annual Review
Due
Group II – Annual
Update
ALSS Division
10/13/08
Reviews:
Anth, Art, Drama,
Frnc, Germ, Ital, Jpn,
Music, NAS, Polsc,
Psych, Snlan, Soc,
Span, Addiction
Studies
Group III – Annual
Update
Hum/Com Division
Reviews:
Cine, Comm, ECE,
Engl, Guid, GS, GC,
Hist, IDS, Journ,
Phil, Spch
PRC Review
Meeting
10/24/08
Revisions
PRC meets to
Due to PRC summarize
results
10/29/08
10/31/08
Annual Review
Due
PRC Review
Meeting
Revisions
PRC meets to
Due to PRC summarize
results
10/27/08
11/14/08
11/19/08
87
11/21/08
Group IV – Annual
Update
BTech Division
Reviews:
Econ, CET, HRC,
IT, RE, WT, BT,
BUS, CIS, DM, AT,
DT, MT, DHET, CT
Annual Review
Due
PRC Review
Meeting
PRC meets to
Revisions
Due to PRC summarize
results
11/17/08
12/5/08
12/10/08
12/12/08
Calendar of Comprehensive Review Activities, Spring ‘09
Group A Comprehensive
Comprehensive
Review Due
PRC Review
Meeting
2/2/09
2/20/09
Comprehensive
Review Due
PRC Review
Meeting
PRC meets to
Revisions
Due to PRC summarize
results
2/25/09
2/27/09
AG, FOR/NR, MST,
NH, Sci Trans Prep
Group B Comprehensive
HO Division
Reviews:
3/2/09
3/27/09
HO, DA, MA, LVN,
NURS
88
Revisions
PRC meets to
Due to PRC summarize
results
4/1/09
4/3/09
Draft Calendar: Fall, 2008 – Annual Updates
September, 2008
Mon
Tues
15
16
23
22
Group I Due
29
30
Wed
17
24
Thurs
18
25
Friday
19
26
Friday
3
Group I
Discussion
10
Group I
Summary
17
October, 2008
Mon
Tues
Wed
1
Thurs
2
6
7
9
13
Group II Due
20
14
8
Group I
Revisions
15
21
22
23
28
29
Group II
Revisions
30
27
Group III Due
16
24
Group II
Discussion
31
Group II
Summary
November, 2008
Mon
3
10
Tues
4
11
Wed
5
12
Thurs
6
13
17
Group IV Due
18
20
24
25
19
Group III
Revisions
26
Wed
3
Thurs
4
10
Group IV
Revisions
17
11
December, 2008
Mon
Tues
1
2
8
9
15
16
89
27
18
Friday
7
14
Group III
Discussion
21
Group III
Summary
28
Friday
5
Group IV
Discussion
12
Group IV
Summary
19
Draft Calendar: Winter/Spring 2009 – Comprehensive Reviews
January, 2009
Mon
19
26
Tues
20
27
Wed
21
28
Thurs
22
29
Friday
23
30
February, 2009
Mon
2
Group A Due
9
16
Tues
3
Wed
4
Thurs
5
Friday
6
10
17
11
18
12
19
23
24
25
Group A
Revisions
26
13
20
Group A
Discussion
27
Group A
Summary
March, 2009
Mon
2
Group B Due
9
16
Tues
3
Wed
4
Thurs
5
Friday
6
10
17
12
19
13
20
23
24
11
18 Spring
Break??
25
26
27
Group B
Discussion
30
31
Thurs
2
Friday
3
Group B
Summary
10
17
24
April, 2009
Mon
Tues
6
13
20
27
7
14
21
28
Wed
1
Group B
Revisions
8
15
22
29
90
9
16
23
30
Calendar of Student Services and Administrative Services Program Review
Activities, Fall ’07 and Spring ‘08
Comprehensive
• Athletics
• Student Health
Center
Annual Update
• Financial Aid
• Arcata Instructional
Site
• Bookstore
Executive Summary
of Student Services
mid-year annual to
PRC Subcommittee for
information only.
Annual Update
• Child Development
Center
• Facilities and
Grounds
• Information
Technology
Services
• Security
Comprehensive
•
•
•
Human Resources
Fiscal Services
Distance Education
Review Due to PRC Sub
PRC
Committee
Review
Meeting
10/1/07
10/09/07
9am-10am
HR Conf
Room
Revisions
Due to
PRC Sub
Committee
10/17/07
Summarized
results sent to PRC
10/1/07
NA
NA
NA
10/29/07
11/06/07
9am-10am
HR Conf
Room
11/09/07
11/14/07
Review Due
PRC Sub
Committee
Review
Meeting
03/18/08
9am-10am
HR Conf
Room
04/07/08
9am-10am
HR Conf
Room
04/23/08
9am-11am
03/10/08
Annual Update
• Food Service
• VP CBO
03/28/08
Annual Update
04/14/08
91
10/19/07
Revisions
Due to PRC
Sub
Committee
03/24/08
Summarized
results sent to
PRC
04/10/08
04/14/08
04/28/08
04/30/08
03/26/08
•
•
•
•
•
•
•
•
•
•
•
•
HR Conf
Room
Institutional
Research
Admissions &
Records
Residential Life
EOPS
Counseling/Advising
DSPS
Career Development
Library
ASC
Judicial Affairs
CalSOAP
Upward Bound
92
93
Download