PROGRAM REVIEW 2007-08 Self-Study Resource Guide College of the Redwoods Redwoods Community College District April 2007 i Table of Contents I. Introduction: What is Instructional Program Review? ............................................1 A. Purpose and Guiding Principles...................................................................1 B. How Program Review Integrates with Planning Assessment......................2 C. Constituents and Responsibilities ...............................................................5 D. List of Approved Programs..........................................................................9 E. Instructional Program Review Schedule for Approved Programs .............13 F. Instructional Program Review Schedule for Disciplines ...........................15 F. Student Services and Learning Support Program Review Schedule .........16 G. Program Review Distribution of Activities 2007-08 Calendar..................18 II. Comprehensive Instructional Program Review .....................................................19 A. Overview....................................................................................................20 B. The Comprehensive Program Review Self-Study Document Template ...22 C. Quality Improvement Plan (QIP) ..............................................................31 D. Program Review Committee Self-Study Review Guidelines ....................33 E. Program Review Participant Survey..........................................................40 III. Annual Instructional Program Review Update ......................................................42 IV. Appendices for Instructional Program Review......................................................47 A. Appendix A: Student Learning Outcomes, Course and Program Level ....48 B. Appendix B: Annual Update Forms ..........................................................54 C. Appendix C: Assessment and Strategic Planning......................................59 D. Appendix D: Glossary of Assessment Terms ............................................60 V. Student Services and Administrative Services Program Review ..........................61 A. Introduction................................................................................................62 B. Schedule of Reviews..................................................................................64 C. Comprehensive Program Review Process Timeline..................................65 D. Comprehensive Program Review Process Steps........................................69 E. Appendices for Student Services and Administrative Services Program Review ........................................................................................71 1. Appendix A: Comprehensive Program Review Workbook...........71 2. Appendix B: Program Review Annual Update Report..................76 3. Appendix C: Program Review Evaluation Forms .........................79 F. Assessment Glossary and Data Element Definitions.................................81 VI. Calendars for Program Review Activities .............................................................85 A. Instructional Programs Fall 2007.....................................................................85 B. Student Services and Administrative Services Fall 2007 & Spring 2008 .......91 Based on a compilation of program review models from Riverside Community College District (CA) and Bluegrass Community and Technical College (KY). Used with permission.. Section I Introduction--What is Program Review? Section I Introduction--What is Program Review? Purpose Program review is a means by which the college can systematically and regularly gather and analyze both qualitative and quantitative data in order to facilitate the continuous improvement of each program, to guide resource allocation, and to assist the faculty, administration and Board in making decisions about programs. As the primary process for program-level planning, program review provides mechanisms for: • • • • • • • improving program quality articulating the program mission and vision with those of the institution assessing program strengths, weaknesses, opportunities and threats determining program capabilities identifying and removing barriers between institutional areas of operation making use of college processes for fiscal, human, physical, equipment, and technology resource allocation assessing compliance with Board policy, WASC, and Ed Code The state Academic Senate has recognized that “the principal purposes of the review process are to recognize and acknowledge good performance, to enhance satisfactory performance and help programs which are performing satisfactorily to further their own growth, and to identify weak performance and assist programs in achieving needed improvement…” (“Program Review: Developing a Faculty Driven Process,” Spring 1996). The Program Review Committee also acknowledges potential inappropriate uses of the program review process. The District process will not be misused in the ways discussed below. It is not a process for evaluating individual performance. It is not a process for eliminating programs. Decisions about discontinuing programs should be made through an alternative process designed to address discontinuation issues directly. It should not be, or appear to be, an adversarial process between faculty and administration. The process should lend itself to developing more effective programs through dialogue and collaborative efforts. It should not be merely an exercise to fill out a document that sits on a shelf. It should have real impacts on program effectiveness and resource allocation decisions. • • • • Guiding Principles The following principles should guide the design and implementation of the Program Review self-study process: • Flexibility: the process needs to be open and flexible enough to accommodate 1 • • • • differences among instructional programs Collegiality: the process should be a collaborative process guided by a spirit of open and honest inquiry Relevance: the process should result in answers to important and relevant questions for programs Practicality: the process should: Minimize the reporting burden on programs Have measures that are few in number, simple, easy to implement and interpret, and useful for program improvement purposes Provide the necessary data resources to assist with the self-study process Effectiveness: the process should result in a clear sense of direction and accomplishment for participants. To facilitate the implementation of plans and the accomplishment of goals, resource needs identified through the self-study process will be linked with budgeting, facilities, and human resource planning. How Program Review Integrates with Planning and Assessment Planning – Self-study documents are intended to be key drivers informing the strategic planning process. Appendix C in Instructional Program Review (p. 59) provides a graphic representation of how assessment and the results of assessment feed into the planning process in support of the mission of the College. Student Learning Outcomes Assessment – In the self-study, programs are asked to develop a student learning outcomes assessment plan that enables them to examine the degree to which students have achieved the desired learning outcomes for a course, course sequence, or program of study. The Program Review Committee will help programs and courses develop instructional program and course student learning outcomes and assessment plans. Further details about how to implement student outcomes assessment follow later in this document. (See Instructional Program Review, Appendix A, p. 48.) 2 Model for Planning and Assessment Process 3 All instructional and student services programs identified annually by the Program Review Committee and approved by the Academic Senate will undergo self-study as part of a process that results in a comprehensive assessment of institutional effectiveness. The District Program Review Committee will forward approved self-studies and action plans to the Office of Academic Affairs, the Office of Student Services, and designated planning groups. These groups will review the self-studies and use information to inform resource allocation decisions. Linking program review and planning is essential; consequently, resource allocations and planning for your discipline will be directly impacted by this plan. The most important extrinsic purpose of this document is for you to receive the resources you need to serve your students. Your program’s priority for faculty or staff, equipment, space and supply allocations will be dependent upon your submission of this report. The intrinsic purpose of the self-study process is to help programs clarify and achieve their goals. In addition, the process aids programs in: • • • • Strengthening programs through self-improvement and self-determination; Strengthening bonds and fostering cooperation within the college community; Providing systematic feedback on student learning; Generating continuous and ongoing dialogue about how student learning can be enhanced through program and service improvements. There are two parts to the Program Review Process: • Comprehensive Program Review. Occurring every five years for instructional programs, an intensive, comprehensive program review will address each program’s courses, pedagogy and other developments related to the future of the program across the district. Beginning in 2007, instructional program reviews will follow a five-year comprehensive cycle with annual updates. As much as possible, comprehensive instructional reviews will be coordinated with the reviews required for external accreditation. Student services program reviews will also follow a five-year comprehensive cycle beginning in 2006. • Annual Program Review Update allows each program and each discipline to address its needs for resources (staff/faculty, space, equipment, etc.) and to report on outcomes assessment activities for the prior year. This annual update will also allow occupational and technical programs to report mandated criteria on program efficacy. All student services programs will also provide an annual update. The Program Self-Study Group (comprised of those individuals writing the comprehensive or the annual review) uses data provided by the program and by the Institutional Research office to analyze the program’s • Relevance to the college’s mission and its intended outcomes • Responsiveness to community needs and satisfaction of community demand • Accessibility to students and identification of unnecessary barriers • Student learning outcomes, including participation and successful completion rates • Measures of program quality • Adequacy of resources and efficiency of resource use 4 • • Performance, price, and enrollment compared to that of alternate local suppliers Direct and indirect revenues and costs to the college The Program Self-Study Group (PSSG) will submit its Self-Study Report (SSR) and a Quality Improvement Plan (QIP) to the Program Review Committee. The PRC will verify the information, complete a questionnaire based on this information, assess performance indicators, and provide feedback regarding the QIP and the Self-Study. The PRC will then forward all documentation (the Self-Study, the questionnaire and the QIP) to an established planning group. Constituents & Responsibilities Program Self-Study Group (PSSG) • Membership: All (when possible or feasible) of a program’s faculty or unit’s staff led by the program coordinator or unit administrator (and, whenever possible or feasible, the Advisory Board and students). For instructional programs with no full-time faculty member, the division chair, director, or campus vice president shall coordinate and help facilitate the process of forming the self-study group. For instructional programs with only one full-time faculty member, the program coordinator will facilitate the process of forming the self-study group. As much as possible, support staff, advisory groups and students should be involved in the self-study groups. • Responsibilities: Generates Self-Study Report; submits Self-Study Report and Quality Improvement Plan to PRC; reviews the PRC response questionnaire and recommendations and makes any necessary revisions; submits final Self-Study Report and QIP to PRC for final approval. IR • Responsibilities: On a schedule, collect and disseminate data to programs, review data with chairs, coordinators and/or unit administrators. District Program Review Committee • Membership: The Committee will be comprised of faculty, staff, and administrators from across the three campuses, as defined in AR122.01. • Responsibilities: Work with IR to review data; maintain the PR schedule; act as liaisons to PSSGs; inform programs of upcoming reviews; provide training; recommend either full or conditional approval of completed program reviews (conditional approval will require the program to make specified revisions to the report and/or the QIP to gain full approval); make fully approved program reviews available to the public; and submit these reviews along with QIPs to Academic Senate (for information purposes) and to designated planning groups (for planning purposes). Academic Senate • Responsibilities: Receives self-study and QIP for informational purposes. Planning Committee • Responsibilities: A designated planning committee will incorporate information resulting from reports into the planning and budgeting process. 5 Board of Trustees • Responsibilities: Review approved program reviews and recommendations for institutional improvement. Publication and Discussion of the Findings Fully approved self-studies will be placed online and distributed to the Office of Academic Affairs, the Office of Student Services, and the Board of Trustees. Presentations of self-studies may be made to the Board of Trustees, Cabinets, and the Academic Senate. 6 COMPREHENSIVE PROGRAM REVIEW PROCESS FLOW CHART Program Review Committee (PRC) ~Informs programs of upcoming reviews ~Trains program selfstudy group members Institutional Research ~Provides data to program Program Self-Study Group (PSSG) ~Generates Self-Study ~Submits Self-Study to Program Review Committee PRC ~Reviews preliminary SelfStudy ~Completes questionnaire and returns report, questionnaire, and recommendations to PSSG to do QIP 7 PSSG ~Reviews results of questionnaire ~Completes QIP ~Submits final report and QIP to PRC PRC ~Collects all final material QIP ~Approves report ~Submits fully approved reports to Senate, Academic Affairs, Student services for information and to planning group for action Planning Committee ~Submits to Board of Trustees for information ~Incorporates relevant information into planning, budgeting and goals development ANNUAL UPDATE PROCESS FLOW CHART Program Review Committee (PRC) ~Informs programs of upcoming reviews ~Trains program selfstudy group members Institutional Research ~Provides data to program PRC ~Reviews Annual Updates ~Submits fully approved reports to Senate, Academic Affairs, Student services for information and to planning group for action Discipline or Program Faculty ~Generates Annual Update and submits to Program Review Committee 8 Planning Committee ~Submits to Board of Trustees for information ~Incorporates relevant information into planning, budgeting and goals development List of Programs (Approved annually by the Academic Senate. Last approved 11/17/07.) Academic Instruction Programs Addiction Studies (CA) Administration of Justice Administration of Justice (AS) Administration of Justice (CA) Basic Law Enforcement (CA) Agriculture Agriculture (AS) Plant Science (AS) Plant Science (CA) Sustainable Agriculture (CA) Landscape Maintenance (CC) Nursery Practices (CC) Automotive Technology Automotive Technology (AS) Advanced Auto Tech (CA) Basic Auto Tech (CC) Business General Business (AS) General Business (CA) Medical Office Business Skills (CA) Bookkeeping (CC) Desktop Publishing (CC) Income Tax Preparation (CC) Payroll Clerk (CC) Word Processing (CC) Business Technology Computer Support Specialist (AS) Office Professional (AS) Computer Information Systems Computer Programming (AS) Computer Programming (CA) CIS Networking (AS) CIS Networking (CA) 9 Construction Technology Construction Technology (AS) Historic Preservation & Restoration (AS) Historic Preservation & Restoration (CC) Fine Woodworking I (CA) Fine Woodworking II (CA) Residential Construction I (CC) Residential Construction II (CA) Cabinetmaking and Millwork (CC) Residential Wiring (CC) Court Reporting (CA) Diesel Heavy Equipment Technology Diesel Heavy Equipment Technology (AS) Diesel Heavy Equipment Technology (CA) Diesel Engine Overhaul (CC) Diesel Fuel Injection & Electronics (CC) Diesel Truck Maintenance & Power Train (CC) Digital Media Digital Media (AS) Digital Media (CA) Game Art & Animation (CA) Game Programming (CA) Animation (CC) Internet Design & Electronic Commerce (CC) Drafting Technology Architectural Drafting (AS) Architectural Drafting (CA) Civil Design (AS) Civil Design (CA) Mechanical Drafting (AS) Mechanical Drafting (CA) Early Childhood Education ECE-Infant/Toddler (AS) ECE-Preschool (AS) ECE-School Age (AS) ECE-Special Needs (AS) ECE (CC) ECE (CA) Child Development Center 10 Education Paraprofessional Education (AS) Forestry and Natural Resources Technology Forestry & Natural Resources, Transfer (AS) Forestry Technology (AS) Forestry Technology (CA) Geographic Info Systems (AS) Natural Resources (AS) Graphic Communications Graphic Communications (AS) Graphic Communications (CA) Health Occupations Dental Assisting (AS) Dental Assisting (CA) Licensed Vocational Nursing (AS) Licensed Vocational Nursing (CA) Medical Assisting (AS) Medical Assisting (CA) Nursing Assistant (CA) Registered Nursing (AS) LVN to RN Upgrade (AS) LVN to RN Upgrade (CA) Hospitality, Restaurant, and Culinary Arts Culinary Management (AS) Restaurant Management (AS) Culinary Supervision (CA) Restaurant Supervision (CA) Culinary Operations (CC) Restaurant Operations(CC) Legal Studies Paralegal Studies (AS) Paralegal Studies (CA) Child Support Specialist (CC) Manufacturing Technology (AS, CA) CADD/CAM Manufacturing (AS, CA) Marine Science Technology Marine Science Technology (AS) Marine Science Technology (CA) 11 Natural History (CC) Real Estate (CA) Science Major University Prep Welding Technology Welding Technology (CC) Electric Arc & Oxyacetylene Welding (CC) MIG & TIG Welding (CC) AA—University Studies AA—General Studies AS—General Studies General Education (CR) Student Services Programs Academic Support Center Athletics CalSOAP Career Center/Student Employment Counseling and Advising DSPS Admissions and Records EOPS Library Residential Life and Campus Activities Student Health Center Upward Bound Judicial Affairs 12 Program Review Schedule for Instructional Programs Program Addiction Studies Administration of Justice Agriculture Automotive Technology Business Business Technology Computer Information Systems Construction Technology Court Reporting* Diesel Heavy Equipment Technology Digital Media Drafting Technology Early Childhood Education, incl. CDC Education* Forestry and Natural Resources Technology Graphic Communications All Health Occupations degrees and certificates Hospitality, Restaurant, and Culinary Arts Manufacturing Technology 2007-08 2012-13 2008-09 2013-14 2009-10 2014-15 2010-11 2015-16 2011-12 2016-17 C A A A A A A A A C A C C A A A A A A A C C A A A A A A A A C A A A A C A A A A C A A A A C C A A A A A A A A A A C A A A C A A A A A A C A A A C A A C A A A A A 13 Program 2007-08 2012-13 2008-09 2013-14 2009-10 2014-15 2010-11 2015-16 2011-12 2016-17 Marine Science Technology Natural History Real Estate Science Major University Prep Welding Technology AA—University Studies AA—General Studies AS—General Studies General Education A C A A A A A A C A C A C A A A A A A A A A C A A A A A A C A A A A C C=Comprehensive Program Review A=Annual Update *These programs were on the approved (11/17/06) list of programs but status of programs may change in 2007-08 academic year. 14 Program Review Schedule for Instructional Disciplines (not on the program list) Discipline 2007-08 2008-09 ACC ANTH ART ASTRO BIOL CHEM CINE CET COMM DANCE DRAMA ECON EMT ENGL ENGR ENVSC FS FRNC GS GEOG GEOL GERM GUID HE HIST HRC IT IDS ITAL JPN JOURN MA MATH METEO MUSIC NAS PHIL PE PHYSC PHYS POLSC SNLAN SOC SPAN SPCH A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A 15 Schedule of Reviews for Student Services and Learning Support C=Comprehensive Program Review Unit Admissions & Records Residential Life and Campus Activities Judicial Affairs EOPS Counseling and Advising DSPS CalSOAP Upward Bound Career Development/Student Employment Library Academic Support Center Athletics Student Health Center A=Annual Update Academic Year 2006- 200707 08 C A 200809 A 200910 A 201011 C 201112 A 201213 A 201314 A C A A A C A A A C C C A A A A A A A A A C C C A A A A A A A A A C C C C A A A A A A A A A A A A C C C C A A A A A A A A A A A A C C A A A A A A C C A A A A A A A A C C A A A A A A C C A A A A 16 Potential data to be provided by Institutional Research (Data required by each program review will vary by program) Data will be sorted by fall/spring for four years. • Curriculum history by semester and time-of-day • Current and historical course enrollment by headcount & FTES • Student cumulative GPAs by course • Retention by course (from census to end of term) • Retention vs. successful completion data for students who meet the advisory (if one exists) and for those who do not • Percentage of successful completion by course by student type • Attrition and retention of students continuing in a program by semester for two full years (e.g., fall 2005, spring 2006, fall 2006, spring 2007) • Current & historical count of graduates by student type • Current & historical count of transfers by student type • Average class sizes • Percentage of courses taught by full- and part-time faculty • Ratio of part-time faculty to full-time faculty and the percent of FTES taught per part-time faculty • Student transfer and transfer success tracking • Additional information available upon request (e.g., student demographics) • * Results of continuing & current student questionnaire • * Results of alumni and graduate surveys (student feedback on career and technical preparation) • * Employer satisfaction survey results (employer feedback on career and technical preparation) • * Alumni satisfaction survey results * Data currently unavailable; April 2007 Other data sources which might be helpful and which IR can help locate (although may not be responsible for): • Population Demographics • Labor Market Data • Student Program Intention • Feedback from Technical Advisory Committees • Student Placement/Wage Data • Other student tracking needs (tracking students who apply for state credentials, etc.) • Comparative Data • Mission statement and strategic plan 17 Comprehensive Program Review Distribution of Activities Calendar for 2007-08 Date Ongoing Institutional Research IR provides the Area/Program Coordinator or Student Services Administrator (when applicable) with all necessary data to complete the program review Program Review Committee Provides orientation and training to the Area/Program Coordinator or Student Services Administrator April 2007 Provides Program Review Guide to the Area/Program Coordinator or SS Administrator Nov. 1-15, 2007 Reviews the Self-Study Report and Quality Improvement Plan (QIP) Nov. 30 Makes recommendations to Self-Study Group (if necessary) Jan. 30, 2008 Receives and approves completed Self-Study and QIP; completes the Review Questionnaire Feb. 15 Monitors the submission of the Program Review Report and the Program’s Quality Improvement Plan to Academic Senate and Institutional Effectiveness by the appropriate deadlines Spring 2008 Posts a record for each program review that includes all key features of the review process Spring 2008 Compiles surveys and makes necessary adjustments to the process for the following year April 2007 Area/Program Coordinator/SS Administrator Apr-Nov.1, 2007 Works with self-study team/program members to prepare report (which should include input from students and, as applicable, advisory board members) Nov. 1 Submits report and QIP to the PRC Nov. 1-15 Provides any supplementary information/documentation needed by PRC Nov. 30 Begins meeting with program faculty to prepare an action plan addressing any program recommendations identified in the report, including who is responsible for the action and the date of expected completion Feb. 7, 2008 Sends one electronic and one hardcopy of the action plan to the PRC and the Academic Affairs Office by the appropriate deadline Program Review Annual Updates Distribution of Activities Calendar for 2007-08 Date Ongoing April 2007 April 2007 August 1 Jan 30, 2008 Feb. 15 Spring 2008 Spring 2008 Fall 2007 Oct. 15-Dec. 1 Jan. 2008 Feb. 7, 2008 Institutional Research IR provides the Area/Program Coordinator or Student Services Administrator (when applicable) with all necessary data to complete the program review (for fall and spring of previous year) Program Review Committee Provides orientation and training to the Area/Program Coordinator or Student Services Administrator Provides Program Review Guide to the Area/Program Coordinator or SS Administrator Determines IR has previous year’s data for programs and instructional disciplines Reviews the Annual Updates Monitors the submission of the Program Review Annual Updates to Academic Senate, General Education Committee and Institutional Effectiveness Posts a record for each program review that includes all key features of the review process Compiles surveys and makes necessary adjustments to the process for the following year Area/Program Coordinator/SS Administrator Works with self-study team/program members to prepare update Submits Annual Update to the PRC (due date set by PRC) Provides any supplementary information/documentation needed by PRC Sends one electronic and one hardcopy of the action plan to the PRC and the Academic Affairs Office by the appropriate deadline 18 Section II Comprehensive Instructional Program Review 19 Section II Comprehensive Instructional Program Review Overview of the Program Self-Study Group Process The normal time frame for the Comprehensive Program Review self study is one academic year. The process consists of two major steps: 1. 2. Initial Planning Self-Study Review Process The Program Self-Study Group will: • Develop detailed task list and timeline • Review data packet from Institutional Research • Discuss self-study guide and questions within the program • Gather any additional information the program deems necessary • Update course outlines and submit to Curriculum Committee if needed • Prepare draft documents resulting from dialogue within the program • Review and finalize the Self-Study Report • Submit Self-Study Report to the Program Review Committee (including updated course outlines) • Will review Program Review Committee’s completed questionnaire and list of recommendations for revising the Self-Study Report and for preparing a Quality Improvement Plan • Submit revised Self-Study and Quality Improvement Plan to Program Review Initial Planning & Training The initial orientation/training meeting will: • • • • Familiarize instructional programs with ACCJC/WASC accreditation standards, the program review process, student learning outcomes assessment, and the role of strategic planning. Allow programs to begin the self-study with hands-on training in data interpretation and assessment techniques. Assist programs in the development of a project plan with timelines and assignment of responsibilities. Generate discussion about the discipline/self-study process, including: 1. Look at the data provided by Institutional Research that have informed your document, discuss it with your colleagues, and identify major changes or trends you expect to be of particular relevance to your discipline in the next five years. 2. What will the implications of these trends be for your discipline and what changes will your discipline need to make to respond to these issues? 20 3. Comment on the enrollment trends and indicate plans as a result of that analysis. What trends, if any, exist in the data by gender, ethnicity, night vs. day, traditional vs. non-traditional age, etc.? Why might these trends be occurring? Are there additional data that would assist your discipline in analyzing these trends? As many members of the program as possible, including the discipline/area coordinator and the division chair should attend the training. Each program should select program representatives who will communicate with Institutional Research and the Program Review Committee and provide organizational leadership for completion of the self-study. 21 Comprehensive Instructional Program Review Document Programs are advised to review their prior Program Review submission (if one exists) and utilize those portions that remain relevant and appropriate to the current format. A. Mission and Relationship to the College(s) Describe how the activities and goals of your program relate to the mission statement and strategic plan of the college. B. Program Description, Curriculum, and Information 1. Provide official program description and list program student learning outcomes. Use the following table to align program student learning outcomes with courses in which each student learning outcome is addressed. Provide additional information relevant to your discipline. a. b. Official Program Description Develop a two-column matrix which presents: i. Program-Level Student Learning Outcomes (left-hand column) ii. Course(s) where these program-level SLOs are presented (right-hand column) Program Student Learning Outcome Statements SLO 1 SLO 2 Etc. Course (s) Course(s) addressing SLO 1 Course(s) addressing SLO 2 Etc. 2. Program-Specific Criteria and/or Admissions Guidelines (as applicable to program) 3. Outline the curriculum as it is being implemented for a full-time student completing a degree or certificate in this program. The outline should include course number, course title, units, lecture hours, and lab hours for each semester for the complete curriculum. Course Number Course Title Units 22 Lecture Hours Lab Hours 4. As part of your self-study, review and summarize the development of curriculum in the program. Include recent additions, deletions, or revisions of courses (attach current course outlines). Evaluate the timing, frequency and coordination of course offerings to determine the adequacy of course offerings relative to a transfer degree (articulation), vocational/occupational certificates, and other appropriate aspects of the district's/campus’/college’s mission. a) Please ensure that there is a thorough review of the course outlines of record and course content during this review period. Please indicate on the course outlines the date on which they were last revised. If the last major curriculum revision occurred more than five years ago, please indicate when the next major revision is planned. Please also review course prerequisites, co-requisites, and advisories as well as obtaining necessary approvals for distance education courses. b) Send updated course outlines to the Curriculum Committee. c) Please file the appropriate form with the Curriculum Committee to delete classes that have not been taught for three or more years unless you plan to teach them in the future. d) With respect to updating course outlines of record, list any relevant trends in your discipline with regard to: 1. Knowledge requirements 2. Skills/student learning outcome requirements 3. Instructional methods e) Describe the various educational delivery methods currently being utilized by the program. Examples include but are not limited to traditional in-person classroom delivery, in-person lab, field studies, online, interactive television, telecourses, clinical instruction, etc. f) Identify curricular revisions, program innovations, and new initiatives planned for the next five years. 5. If applicable, indicate the program external accreditation/approval status. Include a copy of the most recent notification of accreditation/approval status from the appropriate agency. If external accreditation is available but the program has chosen not to seek accreditation status, please explain. 23 C. Program History 1. History (update) since last review: What have been the major developments, activities, changes, and/or projects in your discipline over the past 2 or 4 years (longer if no recent review exists)? (This does not need to include curriculum updates, addressed above.) 2. What were the recommendations from your last program review (if any) and how has your discipline responded to those recommendations? 3. If you have goals from your previous program review, please list them along with the objectives related to your goals, the strategies being used to achieve objectives, and the documentation or evidence that demonstrates success. If no prior program review exists, skip #3. 4. If goals were not achieved, please explain. 5. Discuss any collaborative efforts you have undertaken with other programs (instructional or non-instructional) at College of the Redwoods District and offer an assessment of success and challenges, and potential changes in collaborative efforts. 6. Discuss any activities or projects you have undertaken with other educational institutions, the community, or business/industry. D. Measures of Effectiveness 1. Quality of Education a. Results of certifying, licensing, or registry examinations for each of the last five years (or list not applicable): 1st Attempt Year Number Attempting Pass Percent 20022003 20032004 20042005 24 2nd Attempt Pass Percent National % Goal State % Goal 20052006 20062007 b. Faculty Qualifications: (1) Is there one full-time faculty member whose primary assignment is responsibility for this program? Yes No If not, explain. (2) Are minimum faculty qualifications according to standards set by accrediting/approval bodies met? Yes No (a) State Chancellor’s Office? (b) External Accrediting/approval organizations Yes No If “No” for either response, explain. c. Student Outcomes Assessment: Attach course- and program-level outcomes and assessment reports for each of the last four years. (See Instructional Program Review Appendix A, pp. 48-49). Program-level outcomes should include information from graduate/alumni and employer surveys. d. 2. Vitality a. Attrition and Retention Number of students entering program in fall two academic years prior to this comprehensive program review year: e.g., Fall 2005 Number of students continuing in program (continuing students: all entering students except transfers or readmissions) Spring 2006 Fall 2006 Spring 2007 25 Number Graduating in 2 years (Fall 2005 Entering Class) Explain any attrition indicated by these data. b. Enrollment and Graduate Projections NOTE: Explain method used in making these projections. Use census day enrollments for this figure. Total Enrollment (unduplicated headcount) Graduates 20052006 N= 20062007 N= 20072008 N= 20082009 N= 20092010 N= N= N= N= N= N= c. Total number of 2006-2007 graduates by ethnic group and gender. d. Indicate the number and percent of graduates from the most recent graduating class who are employed in positions related to the program major or continuing in a higher degree program. NOTE: Indicate sources of information and provide any explanation as necessary. Total Number Of Graduates Employed in Related Field N % Continuing Education N % Year e. Indicate the beginning mean and median annual salary for graduates of the program employed full-time, as collected from alumni survey and program advisory meetings. 26 Mean and Median Beginning Annual Salary by Graduating Class Mean Beginning Salary Median Beginning Salary 2000-2000 2000-2000 2000-2000 2000-2000 2000-2000 f. Have there been any significant changes in enrollment, retention, success rates, or student demographics that impact your discipline? If so, please include data sheets (Excel or Word format) showing these changes. 3. Efficiency a. Describe current student population in terms of enrollment by ethnic group and gender. b. Composition of enrollment for selective admissions programs: Fall 2003 Selective Admission Programs Number of Applicants N= Number of Acceptances Number of Entries c. Fall 2004 Fall 2005 Fall 2006 Fall 2007 N= N= N= N= N= N= N= N= N= N= N= N= N= N= Faculty Staffing Pattern 2002-2003 a) Full-time faculty b) Part-time faculty 2003-2004 2004-2005 2005-2006 2006-2007 N= N= N= N= N= N= N= N= N= N= 27 d. Please rate the resources indicated in the table below with respect to how they support this program. Feel free to supplement the resource categories in the extra space provided. The four rating categories are defined as follows: Adequate: Minimally Adequate*: Inadequate*: Not Applicable: This one resource is adequate to permit the program to function effectively. Program functioning at minimally effective level. The limitations imposed by this one resource render the program marginal and will require immediate review for program improvement or continuation. Does not apply to this program. Resources Rating Adequate Minimally Inadequate Not Adequate Applicable a) Faculty Full-time Associate b) Support Staff Clerical Technical Instructional Support Other Personnel c) Current Expenses Office Supply Expenses Instructional/lab supply Expenses d) Equipment e) Library Resources f) Facilities g) Professional Development *If any component is rated Inadequate or Minimally Adequate, please explain. Include recommendations for reducing or eliminating the limitations. 28 e. Budgetary Categories--Please provide funding amount budgeted for each category during the most recent academic year and for each of the previous four years. (Please replicate this table for each year) Year: General Funds Amount Non-General Funds Amount Full-time Faculty Part-time Faculty Other, please specify 1. 2. 3 4 Sub-Total Travel Supplies Equipment Sub-Total TOTAL Explain any changes in funding and add any comments relevant to the adequacy of resources. 29 Summary and Recommendations 1. Summary ● Program Strengths (Please consider how these strengths can be marketed to the community and provide suggestions to PR department) ● Program Weaknesses 2. Recommendations for program improvement generated by self-study. (List by number and use these numbers in the Quality Improvement Plan.) 3. List program members who participated in completion of the self-study report. Vision and Goals 1. Based upon data from California 2025 and/or data from Institutional Research, briefly describe how you would like your discipline to evolve in the next five years. In what ways does your current state differ from your desired state? 2. What specific goals and objectives would you like to achieve to move you toward your vision? 3. What support from the college or district is needed to help you achieve your goals and objectives? 4. What documentation/evidence will demonstrate that you are making progress toward achieving your goals, objectives, and vision? 5. What changes will make the self-study process more helpful to you? 30 Completing the Quality Improvement Plan (QIP) The Quality Improvement Plan (QIP) is intended to assist the program in thinking and planning for a minimum of the next three years. Many factors that influence the implementation of the program’s plans can and do change over time. Each year, the coordinator, division chair or administrator, in consultation with other faculty, will be asked to update the QIP to make adjustments to the plans the program enters this year. The purpose of the QIP is to provide a format for translating both a) the self study group’s recommendations in the Instructional Program Review Self-Study and b) the review team’s recommendations in reviewing that report into actions for improving or maintaining the quality of the program; it is also a mechanism for administrators to determine input to the planning process. Program Self -Study teams complete this form after the Program Review Committee has met and reviewed the instructional program review self-study. Because this document will be used to inform planning processes, it is very important that all the requested information be provided. The form has been designed to elicit the information needed for this process. Each “block” on the form is for a single recommendation; thus, the program should complete all the fields for each of the recommendations cited in the self-study. To add blocks, please use the copy and paste functions in Word. 31 Quality Improvement Plan Program: Year: Field Recommendation # (from self-study report or from review-team report) Recommendation Planned Implementation Date Estimated Completion Date: Action/Tasks Measure of Success/Desired Outcome Estimated Cost(s) Who is responsible? Consequence if not funded External Accreditation Recommendations (if applicable) What to include Number should have been assigned to each recommendation in the Instructional Program Review Self-Study (IPRS). Use wording from IPRS. When does the program expect to begin to implement this recommendation? When does the program estimate this recommendation to be fully implemented? What steps must the program take to implement the recommendation? If the recommendation is implemented, what about the program will be improved? What difference will the implementation of this recommendation make in relation to students, the program’s purpose, and the college’s mission? How will this recommendation improve learning and help meet targeted objectives? This field is particularly important because the information the program enters here is the information that the Institutional Effectiveness/Planning committee will consider in its planning processes. Who will ensure the change is implemented and tracked? If this recommendation is not funded, how will students, the program, the college, or the community be negatively impacted? Does an external accrediting body require that action be taken? If so, what steps should be take to achieve results? What is the cost implication of this recommendation? 32 Program Review Committee Self-Study Review Guidelines I. Areas of Focus The PRC will provide the program with an external perspective concerning all administrative and educational aspects of the program. This perspective includes a judgment of program quality and an assessment of the effectiveness of the instructional program. Although certain information shall be incorporated in all program review reports, they are not intended to be totally comprehensive. Using the Questionnaire on the next page, the PRC will focus its review and recommendations on the following major areas: 1. Quality of degree programs and student learning 2. Quality of faculty/staff and program productivity 3. Quality of recruitment and retention of a diverse faculty, staff, and student body 4. Quality of management, adherence to policies, and collegial environment 5. Quality of planning, evaluation and resources maximizing program effectiveness in terms of allocation 33 II. Questionnaire In reviewing the Self-Study Report, the PRC should use this document to answer questions about each program. Please answer “yes” or “no” for each question and provide an explanation in the comments section when appropriate (all “no” responses must be accompanied by an explanation). Quality of Degree Program and Student Learning 1. Is the sequence of courses leading to the degree appropriate? Yes _______ 2. Comments: No _________ Comments: Are instructional techniques and policies appropriate for the curriculum and types of students enrolled? Yes _______ 5. No _________ Is there evidence that graduates are competent in reading, writing, oral communication, math and the basic use of computers? Yes _______ 4. Comments: Does the curriculum contain at least 18 semester hours of general education courses (for an Associate’s degree) including at least one course in each of the following areas identified by California Education Code: humanities, social science, natural science, writing, oral communications and analytical thinking? Yes _______ 3. No _________ No _________ Comments: Are students provided with written information (syllabus) describing course goals, requirements, content, evaluation methods, etc.? Yes _______ No _________ 34 Comments: 6. Are results obtained from evaluation of instruction used to improve the quality of instruction? Yes _______ 7. No _________ Comments: No _________ N/A _______ Comments: No _________ N/A _______ Comments: No _________ N/A _______ Comments: Are students’ files adequately maintained to determine students’ progress through the program? Yes _______ 14. Comments: Are results of evaluations of the orientation program used to improve the orientation program? Yes _______ 13. N/A _______ Is the orientation program evaluated regularly? Yes _______ 12. No _________ Is an effective orientation program made available to all full- and part-time students? Yes _______ 11. Comments: Are program length, clock hours or credit hours, tuition and fees appropriate for the degree awarded? Yes _______ 10. No _________ Are affiliations with outside agencies (e.g. clinicals in allied health programs) under the control and supervision of the educational institution? Yes _______ 9. Comments: Does instruction take place in an environment conducive to learning? Yes _______ 8. No _________ No _________ Comments: Are students satisfied with their educational experiences? 35 Yes _______ 15. No _________ Comments: Are board exam (e.g. certifying, licensing, and registry) results acceptable? Yes _______ No _________ N/A _______ Comments: Quality of Faculty/Staff and Program Productivity 16. Are the quality and quantity of full- and part-time faculty and staff adequate to support the program? Yes _______ 17. Comments: No _________ Comments: No _________ Comments: Are there adequate jobs for program graduates? Yes _______ 25. No _________ Is program enrollment appropriate? Yes _______ 24. Comments: Is graduation rate appropriate? Yes _______ 20. No _________ Are faculty workloads equitable and reasonable? Yes _______ 19. Comments: Is institutional support adequate for faculty/staff professional development? Yes _______ 18. No _________ No _________ Comments: Are salaries of graduates appropriate for their level of education? Yes _______ No _________ 36 Comments: Quality of Recruitment and Retention of a Diverse Faculty, Staff and Student Body. 26. Is there evidence that the program is involved in recruitment and retention of diverse faculty and staff? Yes _______ 27. Comments: No _________ Comments: Are admission policies appropriate for this program? Yes _______ 30. No _________ Does the program have clearly defined admission policies? Yes _______ 29. Comments: Is there evidence that the program is involved in recruitment and retention of a diverse student body? Yes _______ 28. No _________ No _________ Comments: Is there evidence that admission policies are followed as published? Yes _______ No _________ Comments: Quality of Management, Adherence to Policies and Collegial Environment 31. Is there an atmosphere of mutual respect among faculty, staff and students? Yes _______ 32. No _________ Comments: Is program coordination accomplished in an effective manner? Yes _______ No _________ 37 Comments: 33. Are management policies relating to program operation administered fairly and equitably? Yes _______ 34. No _________ Comments: Is there evidence of linkages/relationships with other post-secondary educational institutions? Yes _______ No _________ Comments: Quality of Planning, Evaluation, and Resources Maximizing Program Effectiveness in Terms of Allocation 35. Are planning and budgeting closely linked? Yes _______ 36. No _________ Comments: No _________ Comments: Are adequate instructional materials and supplies provided for the program? Yes _______ 40. Comments: Is adequate classroom and laboratory space provided for the program? Yes _______ 39. No _________ Are program goals and objectives consistent with college’s mission? Yes _______ 38. Comments: Is there broad-based faculty/staff input in the planning and budgeting process? Yes _______ 37. No _________ No _________ Comments: Are library resources adequate for faculty, staff and students? Yes _______ No _________ 38 Comments: 41. Is computer technology adequate for faculty, staff and students? Yes _______ 42. Comments: No _________ Comments: Are educational outcomes assessment results acceptable? Yes _______ 45. No _________ Are appropriate instruments used to assess educational outcomes? Yes _______ 44. Comments: Is there evidence the program advisory committee is utilized effectively? Yes _______ 43. No _________ No _________ Comments: Are results of assessment of educational outcomes used to improve the program? Yes _______ No _________ Comments: Using the five “Quality” statements as framework, compose a narrative describing the program’s strengths and weaknesses. Additionally, recommendations for program improvement must be provided for each weakness identified. Summary and Recommendations 1. Summary ● Program Strengths ● 2. Program Weaknesses Recommendations for program improvement generated by self study. (List by number and use these numbers in the Quality Improvement Plan.) 39 Program Review Survey This survey should be completed by all faculty working on the Comprehensive SelfStudy Report, the Annual Update, or involved in the Program Review process in any way. Thank you for assisting in the improvement of CR’s instructional programs. Please take a moment to answer the questions below. Your suggestions and comments benefit future program review processes and participants. Thank you for completing this survey. Circle One 1. I received proper orientation to understand my role in the program review process. Disagree Agree Strongly Agree N/A Suggestions for improvement: 2. The Program Review Self-Study Guide contained sufficient information to help me gain a full understanding of the program. Disagree Agree Strongly Agree N/A Suggestions for improvement: 3. I feel as though my contribution to the program review process was valued, appreciated, and respected. Disagree Agree Strongly Agree N/A Suggestions for improvement: 4. The program review process was open, honest, and healthy. Disagree Agree Strongly Agree N/A Suggestions for improvement: 40 5. For those programs which undergo external accreditation, CR’s program review process was helpful in completing your external report. Disagree Agree Strongly Agree N/A Suggestions for improvement: 6. Overall, the program review process is properly designed to achieve its intended goal of program improvement. Disagree Agree Strongly Agree N/A Suggestions for improvement: 7. Please provide any comments or suggestions on how the Program Review Committee can improve the process or facilitate your role. 41 Section III Annual Program Review Update 42 Section III Annual Program Review Update Instructions The Annual Self-Study is conducted district-wide by each program and each academic area and consists of a) analysis of general changes, staffing, resources, facilities, equipment and other needs, as well as b) reporting of curricular changes and outcomes assessment. The questions on the subsequent pages are intended to assist you in planning for your program or area. Input should be sought from all campuses. It should be submitted or renewed every year by the designated date in anticipation of budget planning for the next fiscal year. If there is no change from your prior report, you may simply resubmit that report (or any portion that remains constant) from the prior year with a new date. The majority of the information you will use for this annual update will be provided by IR. Please include pertinent documents such as student learning outcomes assessment reports (see Appendix A for assessment suggestions), and updated course outlines, as well as using data analysis to support any requests for new faculty, facilities, equipment, etc. You are encouraged to use the forms in Appendix B to clarify your requests for equipment, staffing, funding, etc. to make them easy for large committees to review quickly. Please retain this information for your discipline’s use, and submit an electronic copy to Program Review. 43 Annual Program Review Update *Be sure to include information from all three campuses. Program/Discipline: ____________________________ Date: ____________________________ Trends and Relevant Data 1. Has there been any change in the status of your program or area? (Have you shifted departments? Have new degrees or certificates been created by your program? Have you added or deleted courses? Have activities in other programs impacted your area or program? For example, a new nursing program could cause greater demand for life-science courses.) If not, skip to #2. Note: curricular changes should be addressed under 12-14. 2. Have there been any significant changes in enrollment, retention, success rates, or student demographics that impact your discipline? If so, please include data sheets (Excel or Word format) showing these changes. 3. Occupational programs must review the update of their labor-market data, some of it provided by Institutional Research, to illustrate that their program: a. Meets a documented labor market demand, b. Does not represent duplication of other training programs (in the region), and c. Is of demonstrated effectiveness as measured by the employment and completion success of its students. Other Resources 4. Do you have needs (professional development, library resources, and so forth) not previously required by the discipline or not previously addressed in budget or equipment considerations? Please describe. 5. Does your discipline need additional support from Student Services beyond that previously provided? Human Resource Needs 6. Complete the Faculty Employment Grids below (please list full- and part-time faculty numbers in separate rows): Faculty Load Distribution in the Program Discipline Name (e.g., Math, English, Accounting) Total Teaching Load for fall 2006 term % of Total Teaching Load by FullTime Faculty % of Total Teaching Load Taught by PartTime Faculty 44 Changes from fall 2005 Explanations and Additional Information (e.g., retirement, reassignment, etc.) Faculty Load Distribution in the Program Discipline Name (e.g., Math, English, Accounting) Total Teaching Load for spring 2007 term % of Total Teaching Load by FullTime Faculty % of Total Teaching Load Taught by PartTime Faculty Changes from spring 2006 Explanations and Additional Information (e.g., retirement, reassignment, etc.) Do you need more full-time faculty? Associate faculty? If yes, explain why and be sure to include data sheets justifying the need. 7. Complete the Staff Employment Grid below (please list full- and part-time staff numbers in separate rows: Staff Employed in the Program Assignment Full-time Part-time staff (e.g., Math, (classified) staff (give number) English) (give number) Gains over Prior Year Losses over Prior Year (give reason: retirement, reassignment, health, etc.) Do you need more full-time staff? Part-time staff? If yes, explain why and be sure to include data sheets justifying the need. 8. If necessary, to clarify your needs, please comment on current available staff and distribution of FTE's for contract and part-time faculty. Describe strengths and weaknesses of faculty/staff as appropriate to program's current status or future development. Facilities 9. Comment on facilities the program uses, their current adequacy, and any immediate needs. Have your discipline’s facilities needs changed? If so, how? Please provide a data-based justification for any request that requires new or additional facilities construction, renovation, remodeling or repairs. Equipment 10. Have your discipline’s equipment needs changed? If so, how? Is equipment in need of repair outside of your current budget? Please provide a data-based justification for any request that requires a new or additional budget allotment. 45 Learning Outcomes Assessment Update 11. How has your area or program been engaged in student learning outcomes assessment? a. Summarize your results. b. What did your program learn from these results that enabled you to improve teaching and learning in the discipline? c. How have part-time faculty been made aware of the need to assess SLOs? Curriculum Update (Reminder: Send updated course outlines to the Curriculum Committee.) 12. Identify curricular revisions, program innovations, and new initiatives undertaken in the last year. 13. Identify curricular revisions, program innovations, and new initiatives planned for the next year. 14. Complete the grid below Course Year Course Outline Year Next Update Last Updated Expected Goals and Plans 15. If you have recently undergone a comprehensive review, attach your Quality Improvement Plan if applicable. 16. If you do not have a QIP, what goals and plans does your area have for the coming year? 46 Instructional Program Review APPENDIX A: Student Learning Outcomes APPENDIX B: Forms for Annual Update Resource/Staffing Needs APPENDIX C: Assessment’s Link with Strategic Planning APPENDIX D: Glossary of Terms 47 Instructional Program Review—APPENDIX A: Course and Program Outcome Assessment Initiatives Student Outcomes Assessment at the course level 1. The Course Outline of Record template requires that each outline contain a list of student learning outcomes (SLOs) for the course, the core competencies that instructors expect students to leave the course with. Since all SLOs for every course will eventually have to be assessed, the Program Review Committee recommends that disciplines focus on comprehensive student learning outcomes, so that the typical course will have no more than six SLOs, with fewer than five entirely possible. ● ● For courses meeting general education requirements, disciplines are encouraged to link course SLOs with the institutional general education (GE) SLOs as much as possible, since this will be a way for us to assess GE outcomes at the same time we’re assessing the outcomes of individual courses. You will be asked by the template to link methods of evaluation and methods of instruction with specific SLOs, which helps make the SLOs “operational”—more easily measured. Evaluate all courses meeting (or potentially meeting) general education requirements against the institutional general education SLOs, using the Student Learning Outcomes for General Education. This process should be done collaboratively. ● CR has agreed on ____basic core student learning outcomes for its general education program, and all courses meeting GE requirements will link to one (or more) of those student learning outcomes. But we need evidence that they do. You’ll be asked to indicate where a particular GE course SLO matches (even in general terms) to an institution-wide GE SLO. ● As you revise course outlines, this is a good opportunity to adjust some course SLOs so that they more explicitly match the overall GE outcomes. (For example, Ed Code and IGETC requirements mandate that GE courses have critical thinking and writing SLOs, and now is the time to make them explicit.) This will permit PRC to assess overall GE outcomes as your program assesses specific course SLOs. 2. Do an assessment inventory and write a brief report about assessment efforts already undertaken in your discipline. Start with what you are already doing, even if you’re not calling it assessment. 48 Questions to consider: ● ● ● ● ● 3. What are the expected career/occupational outcomes for students in your discipline (e.g., transfer, employment)? Where do faculty members already evaluate student work collaboratively? Are there written criteria that guide this evaluation? Would such criteria be useful if they don’t presently exist? Do faculty members ever report, formally or informally, about strengths and weaknesses in student work? Does information about student learning ever become a topic for discussion at discipline or department meetings? Could it be? ○ Sometimes detailed minutes of division meetings contain extremely useful assessment information, particularly when faculty are encouraged to talk about what they see as areas where students are having trouble. What changes has your department or discipline undertaken recently to improve student learning? What beliefs or information led you to make these changes? What information regarding student success in your department or discipline might already exist that might lead you to reconsider teaching practices? ○ IR may be able to give you invaluable data for strengthening student learning. Begin to develop and implement a comprehensive plan to assess course-level SLOs in either of the ways described below, depending on the particular nature of your discipline: a. Option 1 If at least one course in the discipline is taught by a number of different instructors, choose such a course and work collaboratively to assess learning outcomes in the course. You may want to focus on a single SLO in the initial effort or perhaps several related SLOs. Involve part-time instructors in the process as much as possible. Focus on an assignment or examination toward the end of the semester in which students can b e expected to demonstrate what they’ve gained from the course. Depending on the discipline, you might develop a rubric by which to assess learning in sample essays (or other kinds of student work) taken from many different sections of the course, or—if your discipline routinely employs objective tests as a way of evaluating students—you may simply want to embed common questions in such tests and aggregate results to see where students are having success or difficulty. 49 b. ● Keep careful records of these course-based assessment projects, which should include sections on methodology, results, and analysis of results. ● Sometimes minutes of department meetings are useful appendices to show how results are used to improve teaching and learning. Option 2 If you teach in a discipline with few or no courses taught by multiple faculty members, you may not find collaboration feasible or desirable. In that case, ask every instructor (full- and part-time) to choose a course and develop an individual assessment project for it. Ask instructors to identify a significant SLO for a course they teach, then choose a major graded assignment in that class that they believe measures that SLO. Next, have them develop a rubric (or, for objective examinations, some other method of analysis) by which achievement of that SLO can be assessed. Once the course, SLO, and assignment are chosen and the method of assessment created, have instructors do the actual assessment. Finally, ask instructors to report on and analyze the results of the assessment, in writing. It’s useful to plan some discipline meeting time for discussion of these results, but even more vital that they be captured in written form. Make sure that the reports describe how the instructor would change or improve the teaching of this assignment (and the assignment itself). Discipline discussion should identify common problems instructors are having and potential solutions to those problems. 4. Provide a time line for implementing a plan that would systematically assess learning in other courses in the discipline. NOTE: Assessing student learning at the course level is a daunting process. It will not happen overnight, and no one expects it to. PRC believes it has developed guidelines that are not overly burdensome, while still enabling disciplines to generate the kind of information that will be of maximum benefit, to your discipline and to the college as a whole. Please bear in mind that this must be an ongoing process. For that reason, it’s vital not to conceive of assessment as something we do only when our discipline is up for program review. PRC is available to help you with all parts of the assessment process. 50 Program Outcomes Summary Program Name: Effectiveness Indicator Performance Standard Measurement/Data Source Responsible for collecting data I. Effectiveness A. Quality of Education Student Satisfaction At least 85% of AS program graduates Exit and Alumni will indicate they are satisfied with the Surveys quality of their education. Professional Licensure Program graduates will successfully Board and/or pass professional licensure exams at or Certification Exam above available norms (national or Scores state). Employer Satisfaction At least 85% of employers will indicate program graduates are academically prepared for employment. Employer Survey IR At least 85% of employers will indicate program graduates are competent in general education. Employer Survey IR 100% of faculty must meet ed code or WASC criteria. Faculty personnel file Academic Affairs 100% of faculty must meet external accreditation standards Faculty personnel file Program Coordinator Faculty Credentials 51 IR Program Coordinator (year) (year) (year) (year) Program Name: Effectiveness Indicator Performance Standard Measurement/Data Source Responsible (year) (year) (year) (year) for collecting data B. Vitality Job Placement Annual Unduplicated Headcount of Program Majors Advisory Committee Participation At least 90% of program graduates will Graduate/alumni survey secure employment in their field of study within six months of graduation. At least 12 students enrolled in program in most recent fall term. At least 75% of external members will Advisory Committee attend one meeting within the Members academic year. IR IR Program Coordinator Advisory Committee At least 85% of Advisory Committee survey respondents will indicate they agree or strongly agree with survey items. Program Advisory Committee Survey IR Graduates At least ? graduates per year. Data IR Student Retention Adequate to ensure program efficiency. Fall to fall retention data excluding transfers and readmits IR 52 Program Name: Effectiveness Indicator Performance Standard Measurement/Data Source Responsible (year) (year) (year) (year) for collecting data II. Efficiency Program Coordinator Annual Program Costs Annual Program Report Space Utilization Annual Program Report Average Section Size/Program Enrollment Annual Program Report Date 53 IR Instructional Program Review—Appendix B: Annual Update Forms Campus/Program Needs Worksheet Annual Program Review Update Resources 1. 2. 3. 4. 54 Low Moderate High List Resources Needed for Academic Year___________________ Not Approved Very High Please list/summarize the needs of your program on your campus below This section to be filled out by Subcommittee Degree of Justification (as Approval substantiated by Status the program review) Approved This section to be filled out by the program at each campus Campus/Program Needs Worksheet Annual Program Review Update Faculty 1. 2. 3. 4. 5. 55 Low Moderate High List Faculty Positions Needed for Academic Year___________________ Not Approved Very High Please list/summarize the needs of your program on your campus below This section to be filled out by Subcommittee Degree of Justification (as Approval substantiated by Status the program review) Approved This section to be filled out by the program at each campus Campus/Program Needs Worksheet Annual Program Review Update This section to be filled out by the program at each campus This section to be filled out by Subcommittee Please list/summarize the needs of your program on your campus below Approval Status List Staff Positions Needed for Academic Year___________________ Approved Staff 1. 2. 3. 4. 5. 56 Low Moderate High Not Approved Very High Degree of Justification (as substantiated by the program review) Campus/Program Needs Worksheet Annual Program Review Update Equipment 1. 2. 3. 4. 5. 6. 57 Low Moderate Approximate Cost High List Equipment or Equipment Repair Needed for Academic Year___________________ Not Approved Very High Please list/summarize the needs of your program on your campus below Approved This section to be filled out by Subcommittee Approval Degree of Status Justification (as substantiated by the program review) This section to be filled out by the program at each campus Campus/Program Needs Worksheet Annual Program Review Update Facilities 1. 2. 3. 4. 5. 6. 58 Low Moderate Approximate Cost High List Facility Needs for Academic Year___________________ (Remodels, Renovations or added new facilities) Not Approved Very High Please list/summarize the needs of your program on your campus below Approved This section to be filled out by Subcommittee Approval Degree of Status Justification (as substantiated by the program review) This section to be filled out by the program at each campus Instructional Program Review—Appendix C: Assessment’s Link with Strategic Planning The model below designed by Nichols (Nichols, 1997) shows the relationship of assessment to strategic planning. Positioned at the apex of the triangular model is the mission of the college. Both the development of strategic planning on the left side of the model and assessment practices on the right side flow from the core elements articulated in the College of the Redwoods’ mission statement. The right side of the model represents assessment practices and activities from which the College learns of ways to improve academic programming through the use of assessment results. The intended results, or expected educational results referred to in the WASC criteria, appear in the assessment side of the model along with the means of assessment, assessment results, and uses of results. The left side of the model represents the strategic planning process. Planning on this side of the model is guided by the underlying premise that strategic activities in academic, administrative, budget, and facilities planning are necessary to support the mission of the College. In concert, the information gleaned from assessment practices and use of those results will drive the strategic planning of the college to improve the overall mission of the college. The two sides of the triangle must be linked by budgetary support. When the model is viewed in its entirety it represents institutional effectiveness. 59 Instructional Program Review—Appendix D: Glossary of Terms Assessment: the process of defining, selecting, collecting, and using information to analyze and interpret student learning. Benchmark: a standard by which performance can be measured or judged. Criteria, assessment: a standard of judgment; describes the cognitive and affective outcomes of the educational program; specifies the content knowledge to be mastered. Evaluation: the judgments applied to assessment efforts. General Education Program: A core curriculum organized to promote well-educated individuals who have attained essential, agreed-upon competencies. Impact Statement: a measurable expression of the outcome of a learning experience. Indicator, core: a gauge used to monitor essential condition or performance; usually described in numerical values or indices which together illuminate the condition or direction of educational outcomes being studied. Institutional Effectiveness: the process of articulating the mission of the college, setting goals, defining how the college and community will know when the goals are being met, and using the data from assessment in an ongoing cycle of goal-setting and planning. Intended student outcomes; formal statements of the intended effects or consequences of a course of instruction. Measure, measurement: a set of questions or an inquiry designed to capture evidence on student performance. Stakeholders: persons or groups who have vested interests in the outcomes of the institution (e.g., faculty, staff, employers, licensing agencies, students). Student learning: elements of student learning include knowledge, aptitude, skills, and perceptions. 60 SECTION V Student Services and Learning Support and Administrative Services Program Reviews Introduction The Program Review Resource Guide developed by the Program Review Committee serves as the basis for the Student Services and Administrative Services program review process. Student Services and Administrative Services has representation on the Committee to ensure that the review timeline, procedures, and evaluative criteria used in the Student Services and Administrative Services reviews are relevant to the college’s broad program review effort. This section of the Program Review Resource Guide provides guidelines for use in Student Services and Administrative Services program review activities. These guidelines are indicative of Student Service’s and Administrative Service’s commitment to provide the information and coordination needed to support the district-wide program review process. Purpose and Objectives Program Review is a time consuming effort; however, it is important to note the positive impact program review will have on planning, decision-making, transformation, and stature in the college for student services and administrative services departments. The program review process gives Student Services and Administrative Services an opportunity to: 1) assess a department’s current service, organizational effectiveness and efficiencies, and resources; 2) uncover and validate their strengths; 3) develop program and/or student learning outcomes; 4) identify strategies and potential transformations to address the department’s challenges; and 5) provide an important communication tool within a department and between the department and the administration. The objectives of Student Services and Administrative Services program reviews are to: • • • • • Assess how well the department is performing in relation to its goals, the college strategic plans, and the division’s goals. Encourage strategic thinking about the department’s future plans. Provide evidence of the department’s effectiveness. Evaluate the extent to which a department has successfully addressed program and/or student learning outcomes. Uncover the challenges that may prevent a department from achieving its goals. Principles of Program Review in Student Services and Administrative Services Principle #1: The primary purpose for program review is to make sure that the programs 61 and offerings are connected to the mission, goals, and priorities of the division and the college. Principle #2: All student services and administrative services units contribute to the work of the divisions and therefore will be a part of the program review process. Principle #3: Broad faculty, staff and student participation in the reviews is essential in order to gain important insights and ideas for improvement as well as a broad recognition and understanding of department strengths. Principle #4: Program reviews are independent of any other type of review but, to ease workload and reduce unnecessary duplicative efforts, should be scheduled to compliment other department specific accrediting processes. Coordination of the Program Review Process Responsibilities for carrying out reviews will be shared between the Office of the Vice President for Student Services, the Office of the Vice President, Chief Business Officer, the Vice President of Human Resources, and the departments under review. Program review responsibilities handled by the Vice Presidents include: • • • • • • • Establishing and maintaining the overall Student Services and Administrative Services Program Review Schedule of Reviews and coordinate this schedule with the college’s Program Review Committee. Managing individual department program review calendars. Notifying departments scheduled for review. Planning and conducting department program review orientation meetings with participation from Institutional Research. Reviewing department self-studies for completeness of information. Monitoring improvement plan follow-up reporting, and Evaluating the review process itself and implement recommendations for improvements/revisions to the program review process as deemed necessary. The Program Review Area Coordinator will serve a team leader for the department under review. The Coordinator’s responsibilities include: • • • • • • • Working with program members to prepare report. Recruiting external and student review team members. Finalizing team membership. Arranging initial meetings with review team members and scheduling future meetings. Distributing the Program Review Guidelines and the completed Program Review Report to review team members. Establishing review team member’s responsibilities/duties. Leading review team activities and assuming responsibility for team’s preliminary report. 62 • • Providing any supplementary information/documentation needed by review team members. Making any factual corrections to the final draft. Program Review Cycle Student Services and Administrative Services programs will undergo review on a yearly basis. The comprehensive self-study program review process will be completed by each program once every five years. The annual program review update is to be completed by each program in the years the comprehensive review is not required. The program review schedule may be revised to align with state and/or federal program review reporting requirements. 63 Schedule of Reviews for Student Services and Learning Support C=Comprehensive Program Review Unit Admissions & Records Residential Life and Campus Activities Judicial Affairs EOPS Counseling and Advising DSPS CalSOAP Upward Bound Career Development/Student Employment Library Academic Support Center Athletics Student Health Center A=Annual Update Academic Year 2006- 2007- 200807 08 09 C A A 200910 A 201011 C 201112 A 201213 A 201314 A C A A A C A A A C C C A A A A A A A A A C C C A A A A A A A A A C C C C A A A A A A A A A A A A C C C C A A A A A A A A A A A A C C A A A A A A C C A A A A A A NA NA C C A A A A A A C C A A A A 64 Schedule of Reviews for Administrative Services C=Comprehensive Program Review Unit A=Annual Update Academic Year 20072008- 200908 09 10 A C A 201011 A 201112 A 201213 C 2013-14 Bookstore A A A C A A A Facilities and Grounds A A C A A A C Financial Aid A C A A A C A Fiscal Services C A A A C A A Food Service A A C A A A C Information Technology Services Security VP, CBO Human Resources Communications and Marketing Distance Education A C A A C A A A A C A C A A A A C A A A A C A C A A A A NA A A A C A A C A A A C A A Arcata Instructional Site A Program Review Process and Timeline The comprehensive program review process consists of six steps. These steps are: 1) prereview planning, 2) unit self-study, 3) external review, 4) writing of the final program review report, 5) creating the transformation improvement activities, and 6) implementing the department transformation improvement plan. The normal timeline required for a department program review includes a pre-review preparation period that establishes the groundwork for the review and 10-12 months for the review itself. The 10-12 months schedule creates a timeline that serves to structure and standardize the review process. 65 Step 1: Pre-Review Planning Each program should select program representatives who will communicate with Institutional Research and the Program Review Committee and provide organizational leadership for completion of the self-study. The pre-review planning meeting will: • • • • Familiarize programs with ACCJC/WASC accreditation standards, the program review process, and program and student learning outcomes assessment. Allow programs to begin the self-study with hands-on training in data interpretation and assessment techniques. Assist programs in the development of a project plan with timelines and assignment of responsibilities. Generate discussion about the area self-study process, including: 1. Looking at the data provided by Institutional Research that have informed your document, discuss it with your colleagues, and identify major changes or trends you expect to be of particular relevance to your area in the next five years. 2. Dialoguing about the implications of these trends for your area and what changes your area will need to make. 3. Commenting on the trends and indicate plans as a result of that analysis. What trends, if any, exist in the data? Why might these trends be occurring? Are there additional data that would assist your area in analyzing these trends? Step II: Department Self-Study The self-study represents a valuable opportunity for the department to make a candid assessment of itself and to consider future directions and options that would strengthen it. Each department will prepare a self-study report using as its framework the Comprehensive Workbook. The workbook is intended to provide a structure for the review and should be augmented by whatever information is deemed necessary. The selfstudy should be shared with all staff of the department and should represent consensus. The timeframe for a typical Comprehensive Program Review self-study review process is: Summer • Prepare draft documents resulting from dialogue within the program • Develop detailed task list and timeline • Review data packet from Institutional Research • Discuss self-study guide • Gather any additional information the program deems necessary • Update course outlines and submit to Curriculum Committee if needed • Discuss, and revise if necessary, department mission, goals, learning outcomes, and assessment plan November/December/January/February/March • Review, revise, and finalize the Program Review document 66 April/May • Submit to the appropriate review bodies Step III: External Review The external peer review procedure should be flexible and may be influenced by the results of the self-study. Generally, the external reviewer will conduct the review based on the department’s self-study report. The external reviewer will prepare a summary of recommendations with emphasis given to strengths and challenges of the department and how the department can continue to improve the quality of its programs and services. The summary will be included in the Final Program Review report. Step IV: Writing of the Department Review Report The department’s Program Review Coordinator is responsible for writing the final program review document. The final program review will be forwarded to the Vice President for Student Services and Learning Support, Vice President, Chief Business Officer, or the Vice President for Human Resources for review and comment. The Vice President will forward the report to the college’s Program Review Committee. Following receipt of the report, a meeting is scheduled with the Program Review Committee, Vice President of Student Services, Vice President, Chief Business Officer, Vice President for Human Resources, Department supervisor, and members of the department’s management team (if appropriate) to discuss the findings and the recommendations presented in the final review. Step V: Creating Transformation Activities Following consultation with the Vice President for Student Services, Vice President, Chief Business Officer, or the Vice President for Human Resources and the Department supervisor, the department review coordinator will develop a final three year transformation improvement plan that addresses the recommendation(s) outlined in the Program Review Report for the Institutional Effectiveness Committee. The final transformation improvement plan should specify proposed actions, implementation strategies, and who is responsible for carrying out each action. If there are review recommendations that the department is not in agreement with, the transformation action plan should acknowledge these differences in thinking and where appropriate, present alternative recommendations. Step VI: Implementing the Department Transformation Plan The Office of the Vice President will evaluate the progress on the department’s transformation plan via the Program Review Yearly Update and the annual planning processes. 67 Comprehensive Program Review Distribution of Activities Timeline and Responsibilities Date Institutional Research Ongoing IR provides the Area/Program Coordinator with all necessary data to complete the program review Date Vice President of the Division June SS&LS/AS creates assessment tool(s) and makes necessary adjustments to the process for the following year July SS&LS/AS appoints the Area/Program Review Coordinator Aug SS&LS/AS provides orientation to the Area/Program Review Coordinator Aug SS&LS/AS provides Program Review Guide and the Program Review template to the Area/Program Review Coordinator Mar SS&LS/AS monitors the submission of the Program Review to the appropriate review bodies Date Area/Program Review Coordinator July Coordinator works with department to review and revise if necessary mission statements, goals, learning outcomes, and assessment plans Fall Coordinator works with program members to prepare report Sept Coordinator recruits external reviewer and finalizes team membership Sept Coordinator arranges first meeting with review team members and schedules future meetings as necessary Sept Coordinator distributes the Program Review Guide and the completed Program Review Report to all review team members Sept Coordinator establishes review team member’s responsibilities/duties Nov Coordinator leads review team activities and assumes responsibility for team’s preliminary report Mar Coordinator provides any supplementary information/documentation needed by review team members 68 Apr Coordinator makes any factual corrections to the final draft Apr Coordinator prepares an action plan addressing any program recommendations identified including who is responsible for the action and the date of expected completion Apr Coordinator sends one electronic and one hardcopy of the action plan to the Office of Student Services and Learning Support or Administrative Services by the appropriate deadline Note: Exact dates will be established and provided for critical steps in the process. Comprehensive Program Review Department Self-Study Step 1: Department/Program Description The self-study begins with a description of the department/program including information about the area’s purpose, mission, background, and goals. Provide a short narrative on how the department contributes to and advances the College’s goals and objectives. Step 2: Past Quality Improvements Provide a description (list or narrative) of the quality improvements that were recommended from the department’s previous program review. Also discuss improvements implemented by the department that were not specifically noted on the previous program review. Step 3: Outcomes Assessment Address how successful the department is in determining intended program learning outcomes and student learning outcomes, user satisfaction, and other noteworthy performance results. Discuss the relevance of key performance measures the department uses to track results. Included in this section is a discussion of the instruments used to measure performance (surveys, benchmarking, internal or external review, focus groups, etc.) and what assessment methods were used to determine whether the program accomplished the program learning outcomes and/or students achieved the student learning outcomes. Departments should summarize the findings of the assessments and discuss the team’s interpretation of the data as well as the strengths and challenges of the department as indicated by the performance outcomes indicators. The department program review team should provide concrete suggestions for a transformation improvement plan including ways to build upon current strengths, overcome challenges, and increase effectiveness and efficiency. Offer specific transformation strategies, and timetable, for addressing the challenges affecting the department. 69 This section of the workbook may be revised as appropriate for the department under review. Step 4: Cost Effectiveness This is an opportunity to summarize the financial resources of the department. Departments should comment on significant budget changes or variances from year to year. How has the funding level changed over the past five years? What factors have influenced these changes? What are the anticipated effects of this funding on program/service delivery? Step 5: Recommendations and Comments Step five allows the department program review coordinator and outside readers to comment on what they see as the strengths and challenges of the department. The comments could encompass key program/service operational, resource and campus environment challenges the department faces. Step 6: Publication and Discussion of the Findings Fully approved program reviews will be placed online and distributed to the District Program Review Committee, the Division Vice President, the Institutional Effectiveness Committee, and internal governing bodies. Suggested List of Documents for Inclusion in the Comprehensive and Yearly Summary Report • • • Organizational Chart Department planning documents Assessment results/reports including: 1. Surveys 2. Benchmark data 3. Data that measure progress of performance goals 4. Student learning outcomes data 5. Program learning outcomes data 70 APPENDIX A Assessment and Transformation Comprehensive Program Review Workbook Review Coordinator Review Team Members Year Reviewed 71 SECTION I: Department Background, Purpose and Goals A. Department Mission: B. Background C. Department Goals: D. Describe how the goals of the program relate to the College’s Mission and Strategic Plan. SECTION II: Past Quality Improvements and Transformations A. Obtain a copy of the program’s previous program review. Provide a description (list or narrative) of the quality improvements that were recommended. Recommended Quality Improvement Accomplished? (Y/N) If Yes, Provide Details. If No, Provide Explanation. B. Provide a description (list or narrative) of other improvements accomplished since the last review not specifically noted on the previous program review. Other Improvements Accomplished Provide Details SECTION III: Outcomes Assessment A. Provide the department’s program and student learning outcomes. Include or attach a copy of the Outcomes Assessment Report. Program Outcome(s) Outcome Process/Activities Contributing to the Outcome (What does the program do to create the outcome?) Measured/Documented By? Student Learning Outcome(s) Knowledge, value or skill (What is being demonstrated and assessed?) Scope of Outcome (In what context, with whom, or in what setting does the outcome occur?) Process/Activities Contributing to the Outcome (What does the program do to create the outcome?) 72 Student Learning Outcome Statement Measured/ Documented By? Outcomes Assessment Report Outcome Assessment Method To Whom? By Whom? Administered When? Use of Results Assessment Status/ Findings Transformations B. What is the team’s interpretation of the data? C. What are the strengths of your department as indicated by your performance indicators? D. What are the challenges for your department which are indicated by your performance indicators? E. What is your three year transformation plan to build upon current strengths, overcome challenges, and increase effectiveness and efficiency? (Include a timetable for implementing transformations) STRATEGIES TO ADDRESS CHALLENGES/TRANSFORMATIONS TIMETABLE FOR IMPLEMENTATION F. What resources are required (Personnel, funding, equipment, etc)? 73 Intended Follow-up SECTION IV: Cost Effectiveness A. Indicate your budget figures and expenditures in recent years: Categories FY 2001 FY 2002 FY 2003 FY 2004 Operating Exp. Budgeted Expended Other (if applicable) ( Budgeted Expended SUBTOTAL Budgeted Expended Capital/Equipment Budgeted Expended TOTAL Budgeted Expended Surplus or (Deficit) Generated Revenue (if applicable) Revenue Budget Revenue Actual B. Comment on significant budget changes or variances from year to year. 74 FY 2005 SECTION V: Conclusion Department Comments: Summary of discussion and findings relating to this review: ___________________________________________________ ____ / ____ / ____ Department Supervisor’s Signature Date External Reviewer Comments: Summary of discussion and findings relating to the program review: ________________________________________________ ____ / ____ / ____ External Reviewer Signature Date Program Review Committee Comments Comments: __________________________________________________ Co-Chair of the PRC _____/_____/_____ Date Academic Senate’s Comments Comments: __________________________________________________ Co-Presidents of the Senate _____/_____/_____ Date Institutional Effectiveness Committee’s Comments Comments: ______________________________________________________ _____/_____/_____ Chair, Institutional Effectiveness Committee Date 75 APPENDIX B Assessment and Transformation Program Review Yearly Update Report Review Coordinator Review Team Members Year Reviewed 76 A. Outcomes Assessment Plan Report Outcome Assessment Method (Indicate the assessment methodology to be used, or was used, such as focus group or exit interview.) To Whom? By Whom? Administere d When? Use of Results Assessment Status/ Findings (Indicate when the data will be available for analysis or the result of the analysis. In the next year, indicate the results of a second assessment cycle Transformations (Indicate what transformations were implemented as a result of assessments.) Intended Follow-up B. Department Goals Report Division Strategic Direction Division Goal Department Goal FY2007-08 Action Plan Status of each action item Completed In Progress Collaboration Partners Resources Needed / Reallocated Staff Staff Time Completed In Progress New Money Reallocated $ Space Equipment Other: Completed In Progress Staff Staff Time Completed In Progress New Money Reallocated $ Space Equipment Other: 77 C. Improvements or Rationale for Action: (Provide a brief summary of the major accomplishments and achievements of your department during the past year.) Vice President’s Recommendation Comments: ______________________________________________________ _____/_____/_____ Vice President’s Signature Date Program Review Committee Comments Comments: __________________________________________________ Co-Chair of the PRC _____/_____/_____ Date Institutional Effectiveness Committee’s Recommendation Comments Comments: ______________________________________________________ _____/_____/_____ Chair, Institutional Effectiveness Committee Date 78 APPENDIX C Program Review Committee Student Services and Administrative Services Comprehensive Program Review Evaluation Form Program Title_________________________________________________________ Date and Time of the program review______________________________________ The PRC will provide the program with an external perspective concerning the administrative and educational aspects of the program. This perspective includes a judgment of program quality and an assessment of the effectiveness of the program. Evaluation 1. Does the program have mission and goals statements in place? Yes No 2. Are program goals consistent with the college’s mission? Yes No 3. Has there been progress in implementing the last program review's recommendations for program improvement? Yes No 4. Has the program identified program and/or student learning outcomes that are relevant to its mission? Yes No 5. Are appropriate instruments used to assess learning outcomes? Yes No 6. Are results of assessments used to improve the program? Yes No 7. Are program strengths clearly articulated? Yes No 8. Are program challenges clearly articulated? Yes No Comments or suggestions to strengthen the program and/or the Program Review: 79 Program Review Committee Student Services and Administrative Services Annual Program Review Evaluation Form Program Title_________________________________________________________ Date and Time of the program review______________________________________ The PRC will provide the program with an external perspective concerning the administrative and educational aspects of the program. This perspective includes a judgment of program quality and an assessment of the continuous improvement of the program. Evaluation 1. Has the program identified program and/or student learning outcomes that are relevant to its mission? Yes No 2. Are appropriate instruments used to assess learning outcomes? Yes No Are results of assessments used to improve the program? Yes No 4. Are program goals consistent with the college’s mission? Yes No 5. Yes No 3. Are improvements or rationale clearly articulated? Comments or suggestions to strengthen the program and/or the Program Review: 80 Assessment Glossary Program Mission: A description of the overall purpose of the program. Program mission statements should align with the College and Division mission statements. Program Goals: Goals are derived from the mission and reflect critical aims of the program to be measured by achievable outcomes. The mission and goal statements define what we do. Outcomes: Outcomes are the expected product from the achievement of our goals. Outcomes define why we do what we do. Program/Administrative Outcome: Expected outcomes from program/administrative processes to measure the efficiency and effectiveness of the program. Example: Students will receive guidance and tools to chart their progress towards goals (measurements could be # of Student Education Plans on file and/or satisfaction levels with specific counseling/advising processes). Student Learning Outcome: Something that we expect students to be able to do after participating in a program or receiving services. Example: Transfer students will demonstrate an understanding of general education transfer requirements (could be documented by qualitative assessment through the advising process, surveys or # of successful transfer students). Data: Provides information for dialogue about the learning and necessary evidence to determine if a transformation is needed. Measurement: A quantitative data indicator or set of data indicators intended to reflect the achievement of an outcome. Documentation: Qualitative data or evidence intended to reflect the achievement of an outcome. Assessment Method: The data collection tools and methods that are utilized to produce measurements and documentation. Result: The level of performance for a specific outcome. Baseline: The level of performance for a specific outcome in the first year of assessment. Benchmark/Performance Target: The desired level of performance for a specific outcome. Use of Results: • What does the data tell us about process? Goals; outcomes; means of assessment; implementation results, data collection and analysis. 81 • What does the data tell us about program improvement? What, if anything, do we need to do to improve program and student learning outcomes? What resources are necessary? Transformations: What improvements will be implemented to improve the program and/or student learning process? Intended Follow-Up: Quantitative or qualitative data methodologies intended to clarify initial results and/or measure transformations. Data Element Definitions Data element definitions will be aligned with California Community College common definitions, when possible, in order to promote a shared understanding of the data elements. The list of operational definitions below is taken from the Research and Planning group of California. The list of data elements definitions will continue to expand as measurements for program and student learning outcomes are identified. Operational Definitions from the Research and Planning Group of California Common definitions have been established by the RP Group so that meaning is added to data analyzed consistently over time, in relation to appropriate comparison groups used by CA community college researchers. The definitions were adopted by the RP Group (RP Group Executive Board Approval), after considerable consultation was provided by researchers across the state. Definitions are related to Associate degree and nonAssociate applicable credit courses and do not include non-credit courses. Course Enrollment: Student is enrolled in course and receives a valid grade notation a the time the college MIS data tape is submitted to the State Chancellor’s Office. Valid grade notations A, B, C, D, F, CR, NC, W, I (Incomplete) Invalid grade notations No grade of record (NGR), Report delayed (RD), In Progress (IP). Course Retention I (For comparing colleges to each other and for comparing departments) Student is retained in the course to end of term. A, B, C, D, CR, I grade notations. Retention Rate Percent of students retained in courses out of total enrolled in courses. The retention rate is calculated by dividing the numerator by the denominator and multiplying by 100. 82 Numerator Number of Students (duplicated) with A, B, C, D, CR, I. Denominator Number of students (duplicated) with A, B, C, D, F, CR, NC, W, I. Course Retention II (For comparing colleges to each other and for comparing colleges to the CCC systemwide rate) Student is retained in the course to end of term. A, B, C, D, F, CR, NC, I grade notations. Retention Rate Percent of students retained in courses out of total enrolled in courses. The retention rate is calculated by dividing the numerator by the denominator and multiplying by 100. Numerator Number of students (duplicated) with A, B, C, D, F, CR, NC, I. Denominator Number of students (duplicated) with A, B, C, D, F, CR, NC, W, I. Course Completion None. Thought to be too similar to and less relevant than retention/success measures. Course Success Student succeeds in the course. A, B, C, CR grade notations. Success Rate Percent of students successful in courses out of total enrolled courses. The success rate is calculated by dividing the numerator by the denominator and multiplying by 100. Numerator Number of students (duplicated) with A, B, C, CR. Denominator Number of students (duplicated) with A, B, C, D, F, CR, NC, W, I. Term to Term Persistence Student persists from one term to the next term. First Term Student is enrolled in at least one course. A, B, C, D, F, CR, NC, W, I grade notations. Next Term Student is enrolled in at least one course. A, B, C, D, F, CR, NC, W, I grade notations. 83 Persistence Rate Percent of students enrolled in next term out of students enrolled in first term. The persistence rate is calculated by dividing the numerator by the denominator and multiplying by 100. Numerator Number of students with at least one course with A, B, C, D, F, CR, NC, W, I in the next term (x100). Denominator Number of students with at least one course with A, B, C, D, F, CR, NC, W, I in the first term. Transfer Ready I The rate by which a first-time entering freshmen cohort meets the basic transfer requirements for admittance to California State University. Transfer Directed All freshmen enrolled in any transfer level English course and in any transfer level math course within three years, and within six years of admission. Transfer Ready II The rate by which the transfer directed student cohort completes 60+ transfer units with a 2.0+ GPA at both three and six years, including the successful completion of any transfer level English and any transfer level math course, earning grades A, B, C, or Credit. 84 SECTION VI Calendar of Instructional Program Review Activities, Fall ‘07 Group I – Annual Update Annual Review Due PRC Review Meeting Revisions PRC meets to Due to PRC summarize results 10/3/07 10/5/07 MSE Division Reviews: AG, Astro, Biol, Chem, Eng, Envsc, For, Geog, Geol, Math, Meteo, MST, NH, NR, Ocean, Physc, Phys, Sci-Prep 9/17/07 9/28/07 9/17/07 9/28/07 10/3/07 10/5/07 9/17/07 9/28/07 10/3/07 10/5/07 PE: Dance, HE, PE AJ: AJ Group A Comprehensive BT, BUS, CIS, DM, Addiction Studies Comprehensive Review Due PRC Review Meeting 10/1/07 10/12/07 Annual Review Due PRC Review Meeting Group II – Annual Update ALSS Division 10/15/07 Reviews: Anth, Art, Drama, Frnc, Germ, Ital, Jpn, Music, NAS, Polsc, Psych, Snlan, Soc, Span HO Division 10/15/07 Reviews: HO, DA, MA, Lvn, Nurs 85 Revisions PRC meets to Due to PRC summarize results 10/17/07 10/19/07 10/26/07 Revisions PRC meets to Due to PRC summarize results 10/31/07 11/2/07 10/26/07 10/31/07 11/2/07 Group III – Annual Update Hum/Com Division Reviews: Cine, Comm, ECE, Engl, Guid, GS, GC, Hist, IDS, Journ, Phil, Spch BTech Division Reviews: Econ, CET, HRC, IT, RE, WT, Group B Comprehensive AT, DT, MT, DHET, CT Annual Review Due PRC Review Meeting Revisions PRC meets to Due to PRC summarize results 10/29/07 11/9/07 11/14/07 11/16/07 10/29/07 11/9/07 11/14/07 11/16/07 Comprehensive Review Due PRC Review Meeting Revisions PRC meets to Due to PRC summarize results 12/5/07 12/7/07 11/19/07 11/30/07 86 2008-2009 Draft Schedule For Instructional Annual Updates and Comprehensive Reviews Calendar of Annual Review Update Activities, Fall ‘08 Group I – Annual Update MSE Division Reviews: Astro, Biol, Chem, Eng, Envsc, Geog, Geol, Math, Meteo, Ocean, Physc, Phys, Annual Review Due PRC Review Meeting Revisions PRC meets to Due to PRC summarize results 10/8/08 10/10/08 9/22/08 10/3/08 9/22/08 10/3/08 10/8/08 10/10/08 9/22/08 10/3/08 10/8/08 10/10/08 PE: Dance, HE, PE AJ Classes Annual Review Due Group II – Annual Update ALSS Division 10/13/08 Reviews: Anth, Art, Drama, Frnc, Germ, Ital, Jpn, Music, NAS, Polsc, Psych, Snlan, Soc, Span, Addiction Studies Group III – Annual Update Hum/Com Division Reviews: Cine, Comm, ECE, Engl, Guid, GS, GC, Hist, IDS, Journ, Phil, Spch PRC Review Meeting 10/24/08 Revisions PRC meets to Due to PRC summarize results 10/29/08 10/31/08 Annual Review Due PRC Review Meeting Revisions PRC meets to Due to PRC summarize results 10/27/08 11/14/08 11/19/08 87 11/21/08 Group IV – Annual Update BTech Division Reviews: Econ, CET, HRC, IT, RE, WT, BT, BUS, CIS, DM, AT, DT, MT, DHET, CT Annual Review Due PRC Review Meeting PRC meets to Revisions Due to PRC summarize results 11/17/08 12/5/08 12/10/08 12/12/08 Calendar of Comprehensive Review Activities, Spring ‘09 Group A Comprehensive Comprehensive Review Due PRC Review Meeting 2/2/09 2/20/09 Comprehensive Review Due PRC Review Meeting PRC meets to Revisions Due to PRC summarize results 2/25/09 2/27/09 AG, FOR/NR, MST, NH, Sci Trans Prep Group B Comprehensive HO Division Reviews: 3/2/09 3/27/09 HO, DA, MA, LVN, NURS 88 Revisions PRC meets to Due to PRC summarize results 4/1/09 4/3/09 Draft Calendar: Fall, 2008 – Annual Updates September, 2008 Mon Tues 15 16 23 22 Group I Due 29 30 Wed 17 24 Thurs 18 25 Friday 19 26 Friday 3 Group I Discussion 10 Group I Summary 17 October, 2008 Mon Tues Wed 1 Thurs 2 6 7 9 13 Group II Due 20 14 8 Group I Revisions 15 21 22 23 28 29 Group II Revisions 30 27 Group III Due 16 24 Group II Discussion 31 Group II Summary November, 2008 Mon 3 10 Tues 4 11 Wed 5 12 Thurs 6 13 17 Group IV Due 18 20 24 25 19 Group III Revisions 26 Wed 3 Thurs 4 10 Group IV Revisions 17 11 December, 2008 Mon Tues 1 2 8 9 15 16 89 27 18 Friday 7 14 Group III Discussion 21 Group III Summary 28 Friday 5 Group IV Discussion 12 Group IV Summary 19 Draft Calendar: Winter/Spring 2009 – Comprehensive Reviews January, 2009 Mon 19 26 Tues 20 27 Wed 21 28 Thurs 22 29 Friday 23 30 February, 2009 Mon 2 Group A Due 9 16 Tues 3 Wed 4 Thurs 5 Friday 6 10 17 11 18 12 19 23 24 25 Group A Revisions 26 13 20 Group A Discussion 27 Group A Summary March, 2009 Mon 2 Group B Due 9 16 Tues 3 Wed 4 Thurs 5 Friday 6 10 17 12 19 13 20 23 24 11 18 Spring Break?? 25 26 27 Group B Discussion 30 31 Thurs 2 Friday 3 Group B Summary 10 17 24 April, 2009 Mon Tues 6 13 20 27 7 14 21 28 Wed 1 Group B Revisions 8 15 22 29 90 9 16 23 30 Calendar of Student Services and Administrative Services Program Review Activities, Fall ’07 and Spring ‘08 Comprehensive • Athletics • Student Health Center Annual Update • Financial Aid • Arcata Instructional Site • Bookstore Executive Summary of Student Services mid-year annual to PRC Subcommittee for information only. Annual Update • Child Development Center • Facilities and Grounds • Information Technology Services • Security Comprehensive • • • Human Resources Fiscal Services Distance Education Review Due to PRC Sub PRC Committee Review Meeting 10/1/07 10/09/07 9am-10am HR Conf Room Revisions Due to PRC Sub Committee 10/17/07 Summarized results sent to PRC 10/1/07 NA NA NA 10/29/07 11/06/07 9am-10am HR Conf Room 11/09/07 11/14/07 Review Due PRC Sub Committee Review Meeting 03/18/08 9am-10am HR Conf Room 04/07/08 9am-10am HR Conf Room 04/23/08 9am-11am 03/10/08 Annual Update • Food Service • VP CBO 03/28/08 Annual Update 04/14/08 91 10/19/07 Revisions Due to PRC Sub Committee 03/24/08 Summarized results sent to PRC 04/10/08 04/14/08 04/28/08 04/30/08 03/26/08 • • • • • • • • • • • • HR Conf Room Institutional Research Admissions & Records Residential Life EOPS Counseling/Advising DSPS Career Development Library ASC Judicial Affairs CalSOAP Upward Bound 92 93