Using Program/Unit Review in Instructional, Business/Administrative and Student Services Areas to Facilitate Change The League for Innovation Innovations Conference 2005, NYC March 8, 2005 Terri M. Manning, EdD Denise Wells, MS Kathy Drumm, DBA, CPA Central Piedmont Community College Charlotte, North Carolina Some history…. CPCC is the largest community college in both the Carolinas with approximately 60,000 enrolled students. We have over 100 instructional program areas including for-credit, continuing education and literacy We have almost 50 administrative and student affairs units Some history…. When we started this process (1998), the College had done nothing truly “IE related” since the last SACS visit in 1992 when they received a recommendation for most must statements in section III. The IR office staff had rolled over and all institutional memory was gone. We had less than 5 years to put every program and unit through a review process. Why this Process? The evaluation of teaching and educational effectiveness is a top priority for most institutions However, the evaluation of business, administrative and student services units is often undervalued and overlooked Evaluating the effectiveness of services (not just customer satisfaction) makes good business sense Institutional Effectiveness and the Southern Association of College and Schools At the heart of SACS’ philosophy is the concept of institutional effectiveness (IE) IE involves a process of planning, evaluation, assessment and use of results Under the new core requirements and comprehensive standards, there are two “mandates” that specifically address program or unit reviews Under the New Core Requirements and Comprehensive Standards Core Requirement 2.5 states “The institution engages in ongoing, integrated, and institution-wide research-based planning and evaluation processes that incorporate a systematic review of programs and services that (a) results in continuing improvement, and (b) demonstrates that the institution is effectively accomplishing its mission.” Comprehensive Standard 3.3.1 States “The institution identifies expected outcomes for its educational programs and its administrative and educational support services; assesses whether it achieves these outcomes; and provides evidence of improvement based on analysis of those results.” IE at Central Piedmont CPCC established an IE committee in 1998 and an IE website in 2000 (http://inside.cpcc.edu/IE) (this website gets an average of 61 hits a day – most outside the college) Committee members were from various instructional departments, student services and administrative/business offices The committee established an IE plan for the college Visual of the IE Plan Four Major Overlapping Pieces of the IE Plan General Education Assessment Program/ Unit Review Annual Goal Setting Cycle College Assessment Process Annual goal setting and follow-up required by all units (Strategic Plan) Program/unit review on a 3-5 year cycle College assessment plan involving surveys, assessment and data analysis The evaluation of general education Developing the Process We wanted to create a meaningful program/unit review process We wanted programs to complete the process having learned something valuable (not a document to set on a shelf) We wanted the process to be “outcome-based” to stand the test of time Developing the Process The first review process was created for instructional programs During the 1998-99 year, 18 programs were reviewed (approximately 100 programs over five years) The perception at the beginning was that “this is just another academic exercise” But the results were very different than any review process previously completed Organizational Stages of Outcome Evaluation Stage 4 Stage 5 Acceptance & adaptation Challenge & competition Catalyst - Proactive Depression - exhaustion Compliance - Passive reactive Stage 3 Bargaining - no time/no money Seek outside sources Stage 2 Anger and antagonism Resistant & Reactive Stage 1 Disbelief & Denial Paralysis - Passive resistance Instructional Program Review The review process contained five sections: I. II. III. IV. V. The Program Profile Program Content Student Learning Outcomes Need for Change Future Issues Instructional Program Review Tasks accomplished through review: Defined their program Detailed faculty and staff (credentials, prof. dev, accomplishments) Unit mission or purpose Link unit mission to the college mission Defined the content of their program Defined the population served Set outcome objectives Performed some means of assessment Analyzed results Determined strengths and weaknesses Created strategies for improvement Determined future needs (including curricular change, needed resources, staff and space) Instructional Program Review Rules for the process included: Involvement on the part of all faculty in the department under review (not just one person). It is recommended that the program begin with a brief faculty retreat to discuss and divide tasks. All programs must use at least one external committee (advisory groups are fine) to provide feedback to programs. All programs must utilize feedback from students. What We Learned…. The first group was dragged kicking and screaming through the process Faculty had very little time for the details Faculty had trouble identifying “student learning outcomes” The results produced great marketing materials Once they saw the results, faculty embraced the process Identifying Outcomes We focused on three types of outcomes: Outcomes Program Outcomes Student Learning Outcomes Administrative Outcomes Outcomes and the Classroom "Outcomes" are benefits for people: changes in knowledge, values, position, skills, behavior or status. More simply stated, outcomes are typically what service providers hope recipients achieve once they complete a program or receive services. This is not the “what” but the “why” of education. Student learning outcomes are outcomes related to the learning that takes place in the classroom. We measure improvements in writing, speaking, understanding the scientific method, etc. Program outcomes are the benefits to a student who receives an associate degree in Nursing or completes a certificate in Network Administration? Typical outcomes might be licensure exam scores, job placement, etc. Program Outcome Model INPUTS ACTIVITIES OUTPUTS Resources Services Products or Results of Activities Staff Buildings Facilities State funds FTE Education (classes) Services Counseling Student activities Numbers served FTE (input next year) # Classes taught # Students recruited # Who participated Constraints Laws State regulations Program Outcomes Model INPUTS > ACTIVITIES > OUTPUTS > OUTCOMES Benefits for People *New knowledge *Increased skills *Changes in values *Modified behavior *Improved condition *Altered status *New opportunities Administrative Outcomes Many units do not directly serve students or they want results within their units that are not truly student learning outcomes. They want to improve services or approach an old problem in a new way. They want to become more efficient and effective. They will set administrative outcomes. My Administrative Outcomes 1. 80% of faculty/staff responding to the faculty/staff survey will perceive that Planning and Research responds quickly to their requests for data. 2. 80% of faculty/staff responding to the survey will perceive that Planning and Research makes a significant contribution to the College. 3. 80% of faculty/staff responding to the survey will perceive that Planning and Research contributes to the effectiveness of CPCC. 4. 80% of faculty/staff responding to the survey will indicate that Planning and Research produces enough reports to meet the planning and information needs of faculty and staff. We Had to do Training on How to Set Outcome Objectives There’s no magic number e.g. 80% or 90% What is reasonable? What can you afford? What realistically can your staff accomplish? May need to benchmark (e.g. enrollment growth) What percent shows you’re not committed and what percent shows you’re naïve? How to Set Outcome Objectives Examples: Fifty percent of students will be able to communicate effectively in writing (complete the writing exam with a grade of 60 [D] or better) By the end of the spring term, 95% of faculty and staff will have completed 20 contact hours of professional development (workshops, college courses, conferences, onsite trainings, etc.) More Realistic Seventy percent of students will be able to communicate effectively in writing (complete the writing exam with a grade of 75 [C+] or better) By the end of the spring term, the professional development office will increase their offerings for faculty and staff by 10% over what was offered last year (workshops, college courses, conferences, onsite trainings, etc.) What Happened Deans used the results to make a case for resources The administration became interested in what was learned through the review process Faculty knew their programs were working Unit Review – Lessons Learned To do this well you need several critical pieces: 1. Support from the top (President or Chancellor, Vice Presidents or Vice Chancellors) 2. Buy in from the grassroots level (participation in the development of the process) 3. Across-the-college participation (no one is exempt) 4. Technology to make it easy (web page and review templates) 5. Technical support from institutional research (liaison to mentor) IE Committee Decision The instructional program review process was successful The review of instructional units was important – but administrative/ESS units helped created an environment conducive to learning at the college and supported the learning process Administrative units should go through a similar process Create a committee to draft a similar process for the administrative and students services areas of the college Administrative units were spread across three VP areas Two representative from each VP area participated in drafting the new review process We spent approximately six months creating a workable process Administrative Units Reviewed Planning and Research Health and Safety Office Human Resources Professional Development Resource Development Equal Opportunity Employment Office Facilities Services (Distribution Services, Facilities Design and Construction, Facilities Management) Financial Services (Payroll, Budgeting, Financial Reporting, Purchasing, Accounts Payable, Cashiering) Administrative Units Reviewed, cont. Audit and Compliance Basic Skills Reporting Inventory Control College Services (Bookstore, Campus Printing and Vending) Information Technology Administrative Computing Services Security Executive Assistant to the President Marketing and Community Relations The College’s Foundation The Library Student Services Units Reviewed Counseling and Advising Career Services Graduation Office Financial Aid/Veteran’s Affairs Student Activities Admissions Registration Retention Services Recruitment Campus Student Services Information Services Welcome Center/Student Success Center Testing Center The Administrative Unit Review Design I. The Unit/Program Profile A. The Mission/Purpose 1. Role unit plays in the college mission 2. Unit/program goals as they relate to the college’s mission B. The Staff 1. Professional and administrative staff (since the last review) a. Position description/duties b. Credentials (full and part-time, if any) c. Accomplishments (if applicable) d. Service to college, community and nation The Administrative Unit Review (continued) B. The Staff (continued) e. Professional development activities 2. Classified Staff a. List of names and positions b. List of required credentials (if any) C. The Customer/Client Served 1. Breakdown of students/faculty or staff by type or demographic information (thorough explanation of who is served) The Administrative Unit Review (continued) II. Definition of Services or Program A. Definition of day-to-day duties of the unit B. Innovations, new projects, new initiatives, local, state-wide or national efforts C. Required functions of unit (description and status of compliance) 1. SACS requirements 2. State mandates 3. Federal mandates 4. Other The Administrative Unit Review (continued) III. Administrative Outcome Objectives and Student Outcomes (where appropriate) A. Administrative Outcome Objectives (2-3 objectives) B. Outcomes (or status if incomplete) of innovations, new projects, new initiatives, local, state or national efforts C. Assessment explanation (what was assessed, who, when, how many) D. Results of Administrative Outcome Objectives Based on Assessment Assessment of Administrative Units During the year of program review, the “Annual Faculty/Staff Survey” contained questions from those units being reviewed. Results were given to each unit and broken out by campus and job type http://inside.cpcc.edu/planning There’s a link to survey results. The Administrative Unit Review (continued) IV.Need for Change A. Strengths identified by external sources, faculty, staff and students B. Weaknesses identified by external sources, faculty, staff and students C. Recommendations by faculty, staff, external sources and students to improve the unit's services and programs D. Strategies for change (based on input from Sections A, B & C) - closing the loop E. A one-year follow-up brief report to the Unit VP reporting on the progress of D above (due at the end of the year following review) The Administrative Unit Review (continued) V. Future Issues - Resources needed for future efforts A. Market trends within the broad service unit or program area (based on bestpractices, the literature or training received) B. Anticipated future changes and needs (based on market trends) C. Resources, equipment, space, staffing and work load changes needs for future growth or continuation D. Future plans of unit Assistance from the Website The IE website: http://inside.cpcc.edu/IE Explanation of IE process Templates and forms for review “Perfect” examples for clarification Schedule for review Overall Benefits #1 No recommendations in the area of Institutional Effectiveness from SACS during the October 2002 re-accreditation visit The college became change-oriented Units had to define strategies for change Didn’t have to be perfect but rather making continuous progress Strategies for change helped identify needed resources for units Units had to “close the loop” with the one-year followup (couldn’t promise and not deliver) Overall Benefits Units became empowered to perform their functions in an optimal manner and to ask for what they needed (no one noticed them before, now they do) Created data to support needs Accomplishments were reported to major administrative groups across the college (President’s Cabinet, Planning Council, etc.) Overall Benefits The college community understood the purpose and function of every unit Senior administration realized the benefits and became strong supporters Data are reported annually from program/unit review VPs became stronger advocates for their units making changes to improve services Overall Lessons Learned Through Program/Unit Review The response rate to on-line surveys was double that of pencil-paper surveys Faculty and staff did not know what many units did Faculty and staff were unsure of many policies (how to obtain funds from the Foundation, etc.) Email was the preferred method of communication for faculty and staff Overall, faculty and staff were pleased with services Wednesday afternoons were the best time to hold trainings Overall Lessons Learned Through Program/Unit Review (cont.) Faculty and staff had good ideas on how to improve services – if we just asked them Use of P-Cards was saving the college time and money and faculty/staff preferred using them Faculty and staff wanted to continue to receive paycheck receipts in the mail (all employees are direct deposit) More than 80% of faculty and staff were using remote access to email Faculty and staff wanted changes in the budgeting process – too complicated Administrative Response Units spend a lot of time working on the program review. If we want them to take it seriously, we have to take it seriously. Once reviews are reported at the end of the year, their Dean or Vice President/Chancellor need to: 1. Read them 2. Respond to them From Personal Experience Sit down with the unit (group meeting) Give them my attention Discuss what was learned Discuss their major issues Discuss what they need to make improvements Actually attempt to direct resources to them to make those improvements Then they do not see it as an academic exercise How It Works for Us Administrative Units Reviewed in 2003 Administrative Technology Services Compliance & Audit, Inventory Control, Basic Skills Compliance & Reporting Facilities Services Information Technology Services Institutional Advancement/Foundation Planning and Research Resource Development Facilities Services Consists of the following departments: Facilities Design and Construction – Oversee planning, design, and construction of new facilities and major renovations Facilities Management – Operates and maintains existing facilities of the College Security – Provides physical security services at all campuses Distribution Services – Manages collection and delivery of U.S. and intercampus mail, package deliveries, and centralized shipping and receiving services Inventory Control – Manages the College’s inventory tracking, reporting, and disposal functions Facilities Services Survey respondents gave high rankings (in the 80-90% range) for: Newly constructed or newly renovated facilities being of high quality and meeting needs Knowing what was planned for the future facilities of their respective campuses Classrooms in buildings built since 1990 being adequate in size, furnishings and other amenities Facilities Services Classrooms and common areas being generally clean Lighting levels being adequate A safe and secure environment being provided for all members of the College community Knowing procedures to follow if a medical emergency on campus Providing mail, shipping and receiving, and inventory control services Areas Needing Improvement And Actions To Be Taken Only 73% of respondents were pleased with new office spaces. To address this concern, the Facilities Standards Subcommittee of Facilities Partners will review the adopted standards for offices and furnishings and make recommendations for revisions. Only 50% faculty and staff were pleased with the temperatures in classrooms. The Central Energy Plant began operating in the summer of 2003; when connected to the main body of Central Campus, the back-up capabilities of this facility should minimize the disruption of cooling. The College is also developing a plan for the future pairing of boilers to provide back-up heating for Central Campus buildings. At the suburban campuses, central plant approaches are being used to link buildings, thereby providing back-up capability. Areas Needing Improvement And Actions To Be Taken Only 57% of the respondents expressed satisfaction with the cleanliness of bathrooms. The College entered into a new housekeeping contract as of July 1, 2003 which employs a new vendor and has enforceable cleanliness standards that must be met. Additionally, Facilities Management is addressing the appearance of some of the older restroom facilities on Central Campus as part of a cyclical rehabilitation program. Future Considerations The College will be adapting its long-term capital planning to new guidelines from the County and seeking the use of alternative funding sources. Facilities Services will be working with other units of the College to ensure coordinated planning relating to the staffing and specialty services for new facilities. Facilities Services will be developing new strategies for meeting operational demands in the face of decreasing dollars per square feet in the County budgets. One Thing We Learned People have a lot to say when you hit them where they live (e.g. facilities services and information technology) The open-ended questions produced pages and pages of comments An Example from Instruction Workplace Basic Skills This program is a literacy initiative that goes directly into the worksite and teaches ESL classes, GED prep and GED classes. During their review, they surveyed both employers and students. This was the first time they had ever done this. What They Learned Employers said: 43.8% of employers reported increases in employee performance as a result of participation in the program. 31.3% reported a reduction in absenteeism by participants. 87.5% said classes improved the morale of their employees 37.5% said participants received raises 50% said communication had improved. What Students Said 70.2% reported being able to fill out job forms better 35.5% said they could now help their children with their homework 91.1% said they felt better about themselves 44.4% said they had received a raise, promotion or opportunity as a result of the courses 86.3% said their ability to communicate in the workplace had improved What Has Happened Since Their assessment data has shown up in their marketing brochures to employers. Their enrollment has grown dramatically. They have received funding and marketing support from Charlotte Reads (considered a model adult literacy program). “Best” Results of “Best” Practice Better use of data across the college We have become more client/customer focused Review gives direction for goals and needed changes Departments are empowered to do their jobs (can’t slip through the cracks and be unnoticed) Problems must be resolved (there is no hiding and no excuses) Surveys provide needs assessment data as well as evaluation which gives departments direction For a copy of this presentation Contact Terri Manning or Denise Wells terri.manning@cpcc.edu denise.wells@cpcc.edu Download or print presentation: http://inside.cpcc.edu/planning Click on “studies and reports” Listed as League Presentation