Fall 2006 Assessment Plans Student Affairs Compiled by Lori E. Varlotta, Vice President for Student Affairs Last Updated on: 12/11/07 -1- Updated by: Cyndi Rose In October 2005, the Vice President of Student Affairs launched a new assessment initiative within the Division. From Fall 2005 forward, she expected all departments to augment their current program improvement plans with student learning outcomes. Presently, each Student Affairs Director is charged with identifying and measuring at least one outcome related to student learning. The following pages detail each department’s emerging assessment plan. If you have general comments about the document, please send them to studentaffairs@csus.edu. If you have specific questions about the programs associated with a certain area, please contact the Director identified as point person for that particular department. Last Updated on: 12/11/07 -2- Updated by: Cyndi Rose Academic Advising Center As of July 2007 Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: The Academic Advising Center offers new student orientation, mandatory freshman advising, and advising on general education and graduation requirements for all students. The Center engages students in a developmental process that helps clarify and implement individual educational plans consistent with their skills, interests, and values. Through individual appointments, group advising sessions, and presentations help students understand the university’s academic requirements as well as its policies and procedures. As a result, students are better prepared to take responsibility for their education and persist towards a timely graduation. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Help students take responsibility for their education and persist towards a timely graduation. Goal 2: Contribute to increasing freshman to sophomore retention rates of Sacramento State students. Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Student Learning Outcome: After participating in a comprehensive freshman or transfer orientation program, eighty-five percent of incoming students will understand General Education requirements and University resources. Rationale: The newly implemented mandatory orientation policy will enable more students to gain a sense of belonging to the University and derive benefits from the information provided. A variety of modalities will be implemented to ensure that students have access to the University orientation program. Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. Develop an Orientation Survey of new students (pre/post test or post test only) to immediately assess the impact of orientation (through both online and traditional sessions) on students’ knowledge and understanding of academic requirements and resources. Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. The data were collected at summer orientation (11 freshman sessions in June and July) to measure learning outcomes. Transfer data were collected in June via scantrons to measure satisfaction with orientation. Data were collected via an online survey at the end of orientation before the students registered for classes. A random sample pre-test was conducted via pencil/paper. Transfer satisfaction data were collected via scantron forms. Last Updated on: 12/11/07 -3- Updated by: Cyndi Rose Results Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. Data collected right after the point-of-contact showed that students learned, for the most part, the information we had presented. See Attachment D1 for the pre-test and post-test survey results. Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. The data were analyzed last fall and minor changes were made to the survey. The post-test will be given online via StudentVoice. The data showed that some areas need to be emphasized more during the orientation sessions. For instance, less than sixty percent of orientation attendees could identify correctly the three academic areas that constitute the bachelorette degree. Because of this low correct response rate, future Orientation Leaders will participate in a two-day freshman program instead of the one day training currently in place. After the data was collected in Summer 2006, we noticed a very high correct response rate on pre-test questions. Though we saw an increase in knowledge on the post-test, it wasn’t as dramatic as we thought it would be. Later the Orientation Coordinator discovered that some of the Orientation Leaders had coached orientation attendees on the pre-test, because they thought low response rates would reflect poorly on their performance as a leader. Understanding the leaders’ misguided fear helped us better inform them about pre and post test methodology we are duplicating the data collection this year (2007) to see if we get different pre-test results. Program Objective: All freshmen will participate in a three-phase comprehensive, proactive advising and major/career exploration program by spring 2007. Rationale: This program will increase student confidence in their academic plan and enable them to make informed course selections and major/career decisions. Possible Measures All 2006-2007 advising appointments and sessions will be tracked through a database. Baseline data will be collected to measure the level of participation. Actual Measures The data were collected via the Academic Advising sign-in database. Also the Freshman Advisor tracked appointments in Microsoft Outlook and double-checked it with the sign-in system. Results Of the 2006 freshmen who were to participate in the phase 2 Freshman Advising Program, during Fall ’06 992 students were seen by peer mentors. The remaining 1014 students were scheduled for appointments in Academic Advising. Of the 1014 students scheduled, 856 were seen in the office. Of the 158 not seen, 33 withdrew from the university, 13 were disenrolled, and 41 had no holds placed because they received major advising through their academic departments. At the end of fall semester (minus the withdrawals, disenrolled, and major advising) the Freshman Advising Program saw 96.3% of participating freshmen. For Spring 2007 (Phase III), of the 1888 freshmen scheduled for appointments, the Freshman Advising Program saw 1715. Of the 173 not seen, 30 withdrew, 2 were disenrolled, and 36 had no holds placed because they Last Updated on: 12/11/07 -4- Updated by: Cyndi Rose received major advising through their academic departments. At the end of spring semester (minus the withdrawals, disenrolled, and major advising) the Freshman Advising Program saw 94.2% of participating freshmen. See Attachment D2 for complete details. Conclusions To keep the newly implemented freshmen advising program afloat we need at least two permanent academic advisors. Throughout the 2006-07 academic year we were able to secure- on a temporary basis- two EOP advisors who worked with the academic advising center on a one year appointment. To meet the basic expectations of the program these positions must be retained on a permanent basis. Program Objective: By December 2006, develop an instrument that measures the impact of proactive advising on freshmen continuation rates. Rationale: This freshmen advising program should increase first to second year continuation rates. Ultimately, it should also increase graduation rates, specifically impacting students’ time to degree. Possible Measures Have 70 percent of students participate in a survey that measures the impact of the advising program and how it affected their confidence in selecting a major and making informed decisions. Actual Measures Data was collected via StudentVoice and through a comprehensive online survey sent to all students via SacNotes. Results Two forms of data were collected for the Freshman Advising Program. Immediate (point-of-contact) responses were collected via PDAs and a web-based follow-up survey was emailed to the freshmen. We received 77% response rate from the point-of-contact surveys. We had only 82 responses to the web-based survey. The advisors each had a PDA with the Freshman Advising Program survey loaded on it. After each advising session the freshmen completed the survey. There were 1327 responses collected throughout the Phase III advising sessions. The survey included questions to evaluate Phase II as well. The data indicate that the freshmen overwhelmingly found the advising sessions helpful and useful, particularly with major advising and course selection (98% very to somewhat helpful). Although most of the advisors were not career counselors, students still felt that the advising they received on possible careers was helpful (88% very to somewhat helpful). A complete summary of the survey results are in Attachment D3 and D4. Conclusions We discussed preliminary survey results at length in staff meetings (including faculty/staff meetings, professional staff meetings, and joint meetings with the Career Center) starting the first part of May 2007. From the data, the program was a huge success. Everyone is anxious to see retention numbers to see if this may have contributed to students returning for a second year. Basically, the data collected was to justify how useful and helpful the program was to the students. We are already developing new surveys for next fall to measure areas that were not examined this year. Also, starting in Phase II, students will be surveyed at the point-ofcontact after each advising session. There are going to be some changes made to the program in Phases II and III not necessarily pointed out from the data. We are in the process of creating a home-grown career/advising assessment instrument targeted for Sac State freshmen. We found the Choices instrument not that useful for this venue. We are also going to switch the career component to Phase III. Students were not ready to discuss career options six weeks into their first semester. Basic career exploration will be discussed with students who are ready, and will be included Last Updated on: 12/11/07 -5- Updated by: Cyndi Rose in the talking points for the advisors to cover if relevant in Phase II. A more complete year-end report will be written in detail by the end of summer evaluating all aspects of the Freshman Advising Program. Program Objective: Increase the retention of students who have been reinstated as Undeclared majors by five percent. Eighty-five percent of these students indicate via survey that they received comprehensive advising and have achieved good academic standing by the following semester after reinstatement, and that the advising they received assisted them in staying at the University. Rationale: Students reinstated as Undeclared majors to the University will receive more mandatory proactive advising starting Fall 2006. Possible Measures Measure: Track through database and advising sessions. Grades will be analyzed by the end of the following semester after reinstatement. Measure: Have students participate in focus groups to discuss the impact of the advising sessions. Measure: Administer exit survey to students who do not return. Actual Measures Grade lists will be run to see academic progress of students reinstated as Undeclared students at the end of fall and spring semesters. Results The Academic Advising Center compared Fall 2005 reinstatements to Fall 2006 reinstatements to measure an increase in retention. The only data collected was retention rates. In Fall 2005 Academic Advising reinstated 68 Undeclared students. As of Summer 2006, 29 students were still enrolled and in good standing, academic probation, continued probation, or subject to academic review. In total, 42 percent of the Fall 2005 reinstatements were enrolled after one year. In Fall 2006 Academic Advising reinstated 94 Undeclared students. As of Summer 2007, 50 students were still enrolled and in good standing, academic probation, continued probation, or subject to academic review. In total, 53 percent of the Fall 2006 reinstated students are still registered and in better academic standing than before. This is an 11 percent increase in retention. Conclusions This objective received the least amount of attention because of the implementation of the Freshman Advising Program. The academic advisor responsible for this population met with reinstated students regularly but there was not an effective tracking system in place. Academic Advising is implementing a new scheduling/tracking system for Fall 2007 that will address this dilemma. No surveys were given and no focus groups were planned for 2006-07. This objective is being carried over to the 2007-08 academic year. Questions regarding the programs and outcomes delineated in this section should be addressed to Beth Merritt Miller, Academic Advising. (916) 278-6231. merrittmillerb@csus.edu. Last Updated on: 12/11/07 -6- Updated by: Cyndi Rose Attachment D1 1. What are the three areas of classes that make up an undergraduate degree at Sac State? A. B. C. D. Total Major, Minor, GE Elective, GE & Major Upper Division, Lower Division, Elective None of the above Pre-test Count Percent 131 39.34% 121 36.34% 55 16.52% 26 7.81% 333 100.00% Post-test Count Percent 445 22.72% 1091 55.69% 353 18.02% 70 3.57% 1959 100.00% % Change -16.62% 19.36% 1.50% -4.23% 0.00% 2. What is the minimum number of required units to graduate? A. B. C. D. Total 135 120 60 51 Pre-test Count Percent 42 12.54% 252 75.22% 33 9.85% 8 2.39% 335 100.00% Post-test Count Percent 19 0.97% 1917 97.86% 7 0.36% 16 0.82% 1959 100.00% % Change -11.57% 22.63% -9.49% -1.57% 0.00% 3. How many classes would you have to take each semester to graduate in 4 years, if each course is 3 units? % Pre-test Post-test Change Count Percent Count Percent A. 3 14 4.18% 28 1.43% -2.75% B. 4 120 35.82% 348 17.82% -18.00% C. 5 144 42.99% 1405 71.94% 28.96% D. 6 57 17.01% 172 8.81% -8.21% Total 335 100.00% 1953 100.00% 0.00% 4. Where do students go on campus to get information about student clubs and organizations? A. B. C. D. Total Career Center Student Activities Library WPE Office Last Updated on: 12/11/07 Pre-test Count Percent 17 5.06% 298 88.69% 3 0.89% 18 5.36% 336 100.00% -7- Post-test Count Percent 78 4.00% 1823 93.39% 11 0.56% 40 2.05% 1952 100.00% % Change -1.06% 4.70% -0.33% -3.31% 0.00% Updated by: Cyndi Rose 5. Which of the following examples fulfill the foreign language requirement to graduate at Sac State? A. B. C. D. Total Fluent in a language Passed the AP exam with a score of 3 Took three years of one language in high school All of the above Pre-test Count Percent 12 3.57% 42 12.50% 81 201 336 24.11% 59.82% 100.00% % Change Post-test Count Percent 9 0.46% 33 1.69% -3.11% -10.81% 370 1544 1956 -5.19% 19.12% 0.00% 18.92% 78.94% 100.00% 6. When can a student begin taking Upper Division courses, which are numbered 100-199? A. B. C. D. Total Freshman year After they pass a test Senior year Sophomore year Pre-test Count Percent 28 8.36% 100 29.85% 33 9.85% 174 51.94% 335 100.00% Post-test Count Percent 36 1.84% 280 14.34% 135 6.91% 1502 76.91% 1953 100.00% % Change -6.51% -15.51% -2.94% 24.97% 0.00% 7. Aside from special programs (athletes, EOP, CAMP, MEP), where can students get academic advising help? % Pre-test Post-test Change Count Percent Count Percent A. Academic Advising, Major Department, Minor Department 208 63.03% 1519 77.98% 14.95% B. Career Center 39 11.82% 190 9.75% -2.06% C. Associated Students 14 4.24% 34 1.75% -2.50% D. Tutoring office, major department 69 20.91% 205 10.52% -10.39% Total 330 100.00% 1948 100.00% 0.00% 8. What is the minimum GPA a student needs to be considered in good academic standing and graduate? % Pre-test Post-test Change Count Percent Count Percent A. 1.75 (D+) 1 0.30% 5 0.26% -0.05% B. 4.0 (A) 123 36.94% 176 8.98% -27.95% C. 2.0 (C) 195 58.56% 1752 89.43% 30.87% D. 3.0 (B) 13 3.90% 26 1.33% -2.58% Total 333 100.00% 1959 100.00% 0.00% Last Updated on: 12/11/07 -8- Updated by: Cyndi Rose 9. Which of the following factors should you take into consideration when making your class schedule? % Pre-test Post-test Change Count Percent Count Percent A. Work Load 6 1.81% 41 2.10% 0.29% B. Number of Classes 5 1.51% 22 1.13% -0.38% C. Family Commitments 3 0.90% 5 0.26% -0.65% D. Commute Time 2 0.60% 16 0.82% 0.22% E. All of the above 316 95.18% 1870 95.70% 0.52% Total 332 100.00% 1954 100.00% 0.00% 10. What is the registration system called at Sac State? A. B. C. D. Total CASPER Web SAILOR Web REGISTER Web HORNET Web Pre-test Count Percent 301 90.39% 1 0.30% 10 3.00% 21 6.31% 333 100.00% Post-test Count Percent 1944 99.69% 2 0.10% 3 0.15% 1 0.05% 1950 100.00% % Change 9.30% -0.20% -2.85% -6.26% 0.00% 11. What is the minimum number of units an undergraduate student has to take to be considered as a fulltime student? % Pre-test Post-test Change Count Percent Count Percent A. 15 units 43 12.95% 257 13.17% 0.22% B. 12 units 237 71.39% 1557 79.81% 8.42% C. 10 units 14 4.22% 39 2.00% -2.22% D. 9 units 38 11.45% 98 5.02% -6.42% Total 332 100.00% 1951 100.00% 0.00% 12. During the first week of classes you notice a class on your schedule that you don't remember registering for and don't want to take, you should: % Pre-test Post-test Change Count Percent Count Percent A. Not worry about it because it wasn't your fault 5 1.52% 9 0.46% -1.05% B. Not attend the class because you think the instructor will just drop you 3 0.91% 7 0.36% -0.55% C. Go to a computer lab and drop it online 215 65.15% 1749 89.46% 24.31% D. All of the above 107 32.42% 190 9.72% -22.71% 100.00 100.00 Total 330 % 1955 % 0.00% Last Updated on: 12/11/07 -9- Updated by: Cyndi Rose Attachment D2 Students-Actual Numbers Phase II - Fall 2006 Beginning of Semester Phase III - Spring 2007 2006 1888 992 NA 1014 1715 Withdrew 33 30 Disenrolled 13 2 No Hold - Major Advising 41 36 Not Seen - Hold Remains 71 105 Seen by Peer Mentor Seen in Academic Advising Last Updated on: 12/11/07 - 10 - Updated by: Cyndi Rose Attachment D3 Q1. Are you a first-time freshman? Count Percent 1372 97.72% Yes 32 2.28% No 1404 Respondents Q2. In which of the following classes are you currently/have you been enrolled? (Check all that apply; please scroll down) Respondent Response Count % % 625 45.99% 31.41% ENGL 1 129 9.49% 6.48% ENGL 1X 162 11.92% 8.14% LS 7A 197 14.50% 9.90% LS 7B 262 19.28% 13.17% LS 10A 0 0.00% 0.00% LS 10I 24 1.77% 1.21% LS 10X 65 4.78% 3.27% LS 15 32 2.35% 1.61% LS 86 80 5.89% 4.02% LS 87 44 3.24% 2.21% MATH 9 15 1.10% 0.75% MATH 11 355 26.12% 17.84% None of the above 1359 Respondents 1990 Responses Q3. How did you learn about your Fall 2006 semester advising session? (Check all that apply) Respondent Response Count % % 961 70.77% 56.83% Email & SacNote invitation 321 23.64% 18.98% Peer mentor 156 11.49% 9.23% Friend told me 253 18.63% 14.96% Registration hold 1358 Respondents 1691 Responses Q4. During Fall 2006 was your advising session at the Academic Advising Center? Count Percent 827 60.90% Yes 531 39.10% No 1358 Respondents Q5. Which type of session did you have? Count Percent 745 90.08% I had an individual advising session. 82 9.92% I attended a group advising session. 827 Respondents Q6. Were you advised by your peer mentor as part of your freshman seminar or learning community? Count Percent 467 87.95% Yes 64 12.05% No 531 Respondents Last Updated on: 12/11/07 - 11 - Updated by: Cyndi Rose Q7. This semester (spring 2007), how did you learn about your advising session? (Check all that apply) Respondent Response Count % % 1068 78.65% 65.56% Email & SacNote invitation 89 6.55% 5.46% Peer mentor 203 14.95% 12.46% Friend told me 269 19.81% 16.51% Registration hold 1358 Respondents 1629 Responses Q8. This semester (spring 2007), what type of advising session did you have at the Academic Advising Center? Count Percent 1310 96.47% An individual advising session 48 3.53% A group advising session 1358 Respondents Q9. How helpful was the advising you received about majors available at Sac State? Count Percent 1060 78.11% Very 267 19.68% Somewhat 24 1.77% Not very 6 0.44% Not at all 1357 Respondents Q10. How helpful was the advising you received about possible careers? Count Percent 590 43.48% Very 608 44.80% Somewhat 104 7.66% Not very 55 4.05% Not at all 1357 Respondents Q11. How helpful was the advising you received about course selection? Count Percent 1076 79.29% Very 257 18.94% Somewhat 18 1.33% Not very 6 0.44% Not at all 1357 Respondents Q12. How helpful was the advising you received about GE requirements? Count Percent 1137 83.79% Very 209 15.40% Somewhat 8 0.59% Not very 3 0.22% Not at all 1357 Respondents Last Updated on: 12/11/07 - 12 - Updated by: Cyndi Rose Q13. How helpful was the advising you received about how best to complete my degree (long-term planning)? Count Percent 653 48.12% Very 594 43.77% Somewhat 87 6.41% Not very 23 1.69% Not at all 1357 Respondents Q14. Please complete the sentence: As a result of my advising sessions, I know more about... (Check all that apply) Respondent Response Count % % 379 27.93% 10.63% Career options 555 40.90% 15.56% Major choices 1135 83.64% 31.82% General education requirements 990 72.96% 27.75% Choosing courses 508 37.44% 14.24% Graduation requirements 1357 Respondents 3567 Responses Q15. Do you have a major? Count Percent 1015 74.80% Yes 342 25.20% No 1357 Respondents Q16. When did you choose your major? (Check all that apply) Respondent Response Count % % 865 85.22% 74.70% Before enrolling in Sac State 131 12.91% 11.31% After taking classes that interest me 23 2.27% 1.99% After talking with professors in the department 54 5.32% 4.66% As a result of freshman advising sessions 71 7.00% 6.13% After talking with friends 14 1.38% 1.21% After going to the Career Center 1015 Respondents 1158 Responses Q17. How confident are you regarding your choice of major? Count Percent 617 60.79% Very 309 30.44% Somewhat 63 6.21% Not very 26 2.56% Not at all 1015 Respondents Q18. Are you close to choosing a major? Count Percent 127 37.13% Yes - I'm seriously considering a couple majors 134 39.18% Somewhat - I'm taking some classes that will help me choose a major Last Updated on: 12/11/07 - 13 - Updated by: Cyndi Rose Q18. Are you close to choosing a major? Count Percent 72 21.05% Not really - I'm still uncertain about choosing a major 9 2.63% No - I'm completely lost 342 Respondents Q19. Did freshman advising introduce you to new major possibilities? Count Percent 202 59.06% Yes 140 40.94% No 342 Respondents Q20. Are you going to take classes at Sac State next semester? Count Percent 1331 98.08% Yes 26 1.92% No 1357 Respondents Q21. Why are you not continuing at Sac State next semester? (Check all that apply; please scroll down) Respondent Response Count % % 1 3.85% 2.70% Academic difficulty 11 42.31% 29.73% Family or personal issues 1 3.85% 2.70% Work responsibilities 5 19.23% 13.51% Money concerns 1 3.85% 2.70% Military service 11 42.31% 29.73% Moving 5 19.23% 13.51% Campus environment 2 7.69% 5.41% Major I want is not offered here 0 0.00% 0.00% Could not enroll in the classes I need 26 Respondents 37 Responses Q22. What are your plans for next semester? (Check all that apply) Respondent Response Count % % 9 34.62% 27.27% Classes at community college 15 57.69% 45.45% Transferring to another 4-year college 1 3.85% 3.03% Trade or technical school 5 19.23% 15.15% Taking time off from school 1 3.85% 3.03% Enlisting in the military 2 7.69% 6.06% Working full-time 26 Respondents 33 Responses Q23. When do you plan on returning to Sac State? Count Percent Last Updated on: 12/11/07 - 14 - Updated by: Cyndi Rose Q23. When do you plan on returning to Sac State? Count Percent 8 30.77% After 1-2 semesters 5 19.23% At some point in the future 13 50.00% Never 26 Respondents Last Updated on: 12/11/07 - 15 - Updated by: Cyndi Rose Attachment D4 Q1. Are you a first-time freshman? Q2. In which of the following classes are you currently/have you been enrolled? (Check all that apply; please scroll down) Last Updated on: 12/11/07 - 16 - Updated by: Cyndi Rose Q3. How did you learn about your Fall 2006 semester advising session? (Check all that apply) Q4. During Fall 2006 was your advising session at the Academic Advising Center? Last Updated on: 12/11/07 - 17 - Updated by: Cyndi Rose Q5. Which type of session did you have? Q6. Were you advised by your peer mentor as part of your freshman seminar or learning community? Last Updated on: 12/11/07 - 18 - Updated by: Cyndi Rose Q7. This semester (spring 2007), how did you learn about your advising session? (Check all that apply) Last Updated on: 12/11/07 - 19 - Updated by: Cyndi Rose Q8. This semester (spring 2007), what type of advising session did you have at the Academic Advising Center? Last Updated on: 12/11/07 - 20 - Updated by: Cyndi Rose Q9. How helpful was the advising you received about majors available at Sac State? Q10. How helpful was the advising you received about possible careers? Last Updated on: 12/11/07 - 21 - Updated by: Cyndi Rose Q11. How helpful was the advising you received about course selection? Q12. How helpful was the advising you received about GE requirements? Last Updated on: 12/11/07 - 22 - Updated by: Cyndi Rose Q13. How helpful was the advising you received about how best to complete my degree (long-term planning)? Last Updated on: 12/11/07 - 23 - Updated by: Cyndi Rose Q14. Please complete the sentence: As a result of my advising sessions, I know more about... (Check all that apply) Q15. Do you have a major? Last Updated on: 12/11/07 - 24 - Updated by: Cyndi Rose Q16. When did you choose your major? (Check all that apply) Last Updated on: 12/11/07 - 25 - Updated by: Cyndi Rose Q17. How confident are you regarding your choice of major? Last Updated on: 12/11/07 - 26 - Updated by: Cyndi Rose Q18. Are you close to choosing a major? Last Updated on: 12/11/07 - 27 - Updated by: Cyndi Rose Q19. Did freshman advising introduce you to new major possibilities? Q20. Are you going to take classes at Sac State next semester? Last Updated on: 12/11/07 - 28 - Updated by: Cyndi Rose Q21. Why are you not continuing at Sac State next semester? (Check all that apply; please scroll down) Last Updated on: 12/11/07 - 29 - Updated by: Cyndi Rose Q22. What are your plans for next semester? (Check all that apply) Last Updated on: 12/11/07 - 30 - Updated by: Cyndi Rose Q23. When do you plan on returning to Sac State? Last Updated on: 12/11/07 - 31 - Updated by: Cyndi Rose Alcohol Education Program (AEP) As of July 2007 Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: The Alcohol Education Program strives to reduce the harm associated with high risk drinking through education, collaboration, student leadership development, and allocation of resources to support and encourage healthy choices among Sacramento State students. Rationale: Research has demonstrated that alcohol and other drug misuse can negatively impact student academic achievement, personal relationships, career aspirations, and emotional and physical well-being (Hingson, et al., 2005). The Alcohol Education Program utilizes a multifaceted research-based prevention model to address the use and abuse of alcohol, illicit, and prescription drugs. Program activities include: assessment, policy development, enforcement, outreach, peer education, media production and distribution, correction of behavioral misperceptions, and one-on-one interventions. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Decrease high risk drinking behaviors among students and the potential harm associated with these behaviors. Goal 2: Educate the campus community about alcohol, tobacco and other drugs. Goal 3: Collaborate with Student Affairs to further facilitate and enhance AEP programs. Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Student Learning Outcome: All students who participate in Alcohol Education Program activities, such as: “Don’t Cancel that Class” and CHOICES/E-Chug will demonstrate knowledge of: Laws and policies associated with alcohol and drug use Protective behaviors Potential consequences of high risk and underage drinking Resources available to address alcohol and other drug issues Rationale: The program provides many educational activities in support of this goal, including; 21 st birthday cards, the Social Norms campaign, peer health educators, educational presentations, outreach events, “Don’t Cancel that Class,” Choices program, and other collaborative projects. Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. Participating students will be surveyed with a pre-and post test to assess increased understanding of laws and policies associated with alcohol and drug use, protective healthy behaviors, the potential Last Updated on: 12/11/07 - 32 - Updated by: Cyndi Rose consequences of high risk and underage drinking, and resources available to address alcohol and other drug issues. Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. Collection Date(s): Method: Spring Semester 2007 Pre/Post Test web based as part of educational class. Please refer to Appendix A for surveys Students referred to the Level I education class Populations: Results Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. The learning outcomes assessment of our Choices/E-CHUG education program indicates students’ confusion regarding alcohol equivalencies by discussing both 80 and 100 proof hard liquors. Choices/E-Chug Pre & Post Test Results Fall 2006 Percent Correct 100.00% 80.00% 60.00% Pre Test 40.00% Post Test 20.00% 0.00% Question #1 Question #2 Question #3 Question #4 Question #5 Pre Test 56.40% 74.35% 74.35% 69.23% 61.53% Post Test 57.89% 50.00% 86.84% 92.10% 76.31% Question Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. The learning outcomes assessment of our Choices/E-CHUG education program indicates students’ confusion regarding alcohol equivalencies by discussing both 80 and 100 proof hard liquors. We will be modifying the educational curriculum to remedy this problem. (December 2006) Spring 2007assessment will indicate if revisions to the curriculum were successful in improving the intended learning outcome. As a result of the changes made to the curriculum, the percentage of students responding correctly to question #2 did not increase. Further modifications to the curriculum will be made and as part of the student facilitator training important content will be reviewed. Last Updated on: 12/11/07 - 33 - Updated by: Cyndi Rose Questions regarding the programs and outcomes delineated in this section should be addressed to Heather Dunn Carlton, Alcohol Education Program. (916) 278-6060. heather@csus.edu. Associated Students, Inc. As of July 2007 Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: Associated Students, Inc. serves as the official governing body of Sacramento State students and through operation and sponsorship of programs and services meets the varied needs of students. ASI provides experiential education, leadership opportunities, student representation, various business and recreational services, campus life programs and activities that support the campus and greater Sacramento community. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Serve as the representative entity for Sac State Students. Goal 2: Provide students with experiential education. Goal 3: Provide students with leadership opportunities. Goal 4: Provide business and recreational services. Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Program Objective: Increase the number of faculty who use ASI programs as a mechanism for prompting students to apply theory to practice. A. Increase by 10% the number of individual faculty and students who complete applied assignments in ASI program as of January 2008. B. Increase by 10% the number of academic departments who complete “applied” assignments in ASI programs as of January 2008. C. Achieve and maintain an 80% or greater frequency of “satisfied” or “very satisfied” responses by faculty and student participants by January 2008. D. With faculty, formulate, by June 2007, specific student learning outcomes that connect the theory and principles presented in CD32 and CD132 with the hands on experience gained in a ASI Children’s Center internship. E. Apply the student learning outcome model in (D) model to at least 10 more courses by June 2008. Rationale: ASI sees itself as a “learning lab” for Sacramento State students, where students apply, in a deliberate and structured way, the theory and principles covered in specific courses to particular ASI programs and services. ASI’s goal is for faculty from diverse disciplines to encourage their students to use Last Updated on: 12/11/07 - 34 - Updated by: Cyndi Rose ASI programs as a venue for completing “applied” assignments such as observations, interviews, projects or exercises. Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. Distribute and collect “ASI Survey of Current Student and Faculty Participants” during the Fall 2006 session. Use compiled data to establish “baselines” for number of faculty and students participants by February 2007. Use compiled data to establish “baselines” for number of different departments participating by February 2007. Use compiled data to determine level of faculty and student satisfaction with ASI participation, stated in percentages, by February 2007. Distribute and collect “Survey of Prospective Faculty” by August 2007. Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. Collection Date(s): Method: Fall Semester 2006 ASI Survey of Current Student and Faculty Participants available in hard copy and web based survey formats Available for faculty and students who utilized ASI Programs to fulfill “applied” assignments directly connected to an academic course Population: Results Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. ASI Survey of Current Student and Faculty Participants Fall Semester 2006 Results Associated Students (ASI) at Sacramento State strongly believes that students increase their understanding of theory and principles covered in the classroom setting when given the opportunity to apply and experience them outside the classroom. ASI sees itself as a “learning lab” for Sac State Students. We thank you for choosing ASI to fulfill your classroom assignment. We would like to know how we are doing in providing applied learning experiences for Sac State students and where we may improve our services and activities. Please complete the following survey to provide Associated Students with information that will help us to improve how we “make the connection” between the classroom and students’ experiences in ASI. 1. (Check one) [ ] Faculty = 14 [ ] Student = 257 2. Course: ___________________________ Section: ______________________________________________ Freshman Seminar AFROTC Communications Studies: 5, 21, 158 Education History 100 Journalism: 30, 130 RLS: 21, 100, 122, 136 MIS: 160, 161 MBA: 260 BIO 279 Last Updated on: 12/11/07 - 35 - Updated by: Cyndi Rose KINS: 09, 12, (Sailing – beg. & intermediate) 3. Semester: FALL____, Year: _2006_____ ASI Program: Peak Adventures – 149 Aquatic Center – 110 Children’s Center – 4 Leadership – 3 Green Sting – 1 4. Please rate your satisfaction with the following aspects of your ASI experience (e.g., program, workshop, presentation, etc.) Importance Satisfaction 1 – Very Unimportant 2 – Unimportant 3 – Neutral 4 – Important 5 – Very Important % 3.4 2.7 3.4 3.8 2.7 3.8 2.3 2.3 8.4 8.0 6.8 5.3 23.6 30.0 20.2 24.7 58.9 53.2 64.3 60.1 3.0 2.3 3.0 3.8 4.9 4.6 14.8 24.7 47.9 3.0 2.7 3.4 2.7 2.7 4.2 3.4 3.0 3.0 3.8 3.0 3.0 2.7 4.2 2.3 15.2 16.7 9.5 6.8 15.2 13.3 8.4 29.3 29.3 25.9 26.2 24.3 27.8 19.8 46.4 43.4 55.9 58.6 51.0 47.5 64.3 3.4 3.4 3.0 2.7 2.7 3.8 2.3 % % Please answer both sides 1– Very dissat isfied X– N/A % % % x – N/A 5 – Very satisfied 4 – Satisfied 3 – Neutral 2– Dissatisfied Quality of instruction Availability of staff Staff knowledge of the subject matter Staff preparation Relevance of experience to classroom subject matter Convenience of scheduling Program size Program facilities Program equipment Knowledge acquired Skills acquired Quality of your overall experience % % % % .7 .8 1.5 1.1 2.7 3.0 2.3 2.3 9.1 9.5 8.0 8.0 25.4 23.1 19.7 24.6 58.7 61.4 65.9 60.2 3.4 2.3 2.7 3.4 1.5 3.4 17.8 23.9 49.6 3.4 2.3 1.1 .8 .8 .8 .8 .8 6.1 4.2 2.7 1.9 1.5 1.1 .8 18.9 14.4 9.9 11.7 8.0 7.6 3.8 25.0 23.5 22.7 22.7 14.8 16.3 10.6 43.9 53.4 61.7 60.6 29.9 27.7 39.0 3.8 3.4 2.3 2.3 45.1 46.6 45.1 Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. The data allowed us to establish “baselines” for: The number of faculty and students participating in the survey The number of different departments represented in the survey participation The level of faculty and student satisfaction with aspects of their participation in ASI program(s) The percentage of respondents who selected “NA – not applicable” as their choice when rating their “satisfaction” with “knowledge acquired,” “skills acquired,” and “quality of your overall experience” was approximately 45%. This suggests that almost half of the respondents did not make explicit connections between their classroom leaning and the hands on experience they gained in an ASI program. For students to see this important connection ASI most work with faculty to help students translate theory to practice in ASI settings. Since the goal was to establish “baseline” data; no significant changes will be made to the survey. Staff will work to: Increase the number of faculty and student participants who complete “applied” assignments in ASI programs by 10% as of January 2008 Last Updated on: 12/11/07 - 36 - % Updated by: Cyndi Rose Increase the number of departments with faculty and students who complete “applied” assignments in ASI programs by 10% as of January 2008 Achieve and maintain an 80% or greater frequency of “satisfied” or “very satisfied” responses by faculty and student participants by January 2008 Student Learning Outcome: Survey, by July 2007, the alumni, employees, board members, and volunteers to determine how ASI prepared them for a leadership position beyond Sacramento State. Rationale: Involvement in ASI allows students to develop meaningful relationships with community members and organizations. The emerging network may provide ASI leaders, board members, and volunteers with a competitive edge when applying for community leadership positions, jobs, etc. Possible Measures Survey ASI alumni to: a) identify particular ASI experiences that gave them a competitive edge or opportunity in their post-graduate employment or community leadership pursuits, and b) identify leadership skills they developed specific to ASI experiences. Student Learning Outcome: Implement an ASI student employee mentoring program that includes one-onone meeting and structured programs geared at honing leadership skills. A. Identify and define specific “leadership competencies” by July 2007. B. Create a structured training program that will develop those competencies by June 2008. C. Develop assessment tools to evaluate pre and post-training skill levels by June 2008. Rationale: Each ASI staff supervisor should mentor the students under his or her direction so that they can develop their potential leadership skills. ASI must formalize this mentoring role so that all of its student employees have a more uniform and rewarding experience in being a mentee. Possible Measures The assessment tool developed by June 2008 will assess achievement of the objective. Program Objective: Improve the customer service skills of ASI student and professional staff. A. Identify and define specific “customer service competencies” by July 2007. B. Create a training program and develop those competencies by June 2008. C. Review “customer satisfaction” survey to discern level of achievement the customers’ expectations. Rationale: ASI prides itself on exceptional customer service. Therefore the organization must accurately measure the satisfaction end-users have with our programs and services. excellent delivery of services and programming and provides a means to target particular skills for our students to develop, giving them exceptional job experience and enhancing their employment and leadership opportunities beyond graduation. Possible Measures Establish assessment tools to evaluate pre and post-training skill levels by June 2008. Develop a “customer satisfaction” survey by January 2008. Last Updated on: 12/11/07 - 37 - Updated by: Cyndi Rose Conclusions Appendix A ASI Survey of Current Student and Faculty Participants Associated Students (ASI) at Sacramento State strongly believes that students increase their understanding of theory and principles covered in the classroom setting when given the opportunity to apply and experience them outside the classroom. ASI sees itself as a “learning lab” for Sac State Students. We thank you for choosing ASI to fulfill your classroom assignment. We would like to know how we are doing in providing applied learning experiences for Sac State students and where we may improve our services and activities. Please complete the following survey to provide Associated Students with information that will help us to improve how we “make the connection” between the classroom and students’ experiences in ASI. 1. (Check one) [ ] Faculty [ ] Student 2. Course: ___________________________ Section: ______________________________________________ 3. Semester: _________, Year: __________ ASI Program: _________________________________________ 4. Please rate your satisfaction with the following aspects of your ASI experience (e.g., program, workshop, presentation, etc.) Importance 1 – Very Unimportant 2 – Unimportant 3 – Neutral 4 – Important 5 – Very Important Satisfaction Please answer both sides X – N/A 1 1 1 1 2 2 2 2 3 3 3 3 4 4 4 4 5 5 5 5 x x x x 1 2 3 4 5 x 1 1 1 1 1 1 1 2 2 2 2 2 2 2 3 3 3 3 3 3 3 4 4 4 4 4 4 4 5 5 5 5 5 5 5 x x x x x x x 1 2 3 4 5 x Quality of instruction Availability of staff Staff knowledge of the subject matter Staff preparation Relevance of experience to classroom subject matter Convenience of scheduling Program size Program facilities Program equipment Knowledge acquired Skills acquired Quality of your overall experience Other: specified: ____________________ __________________________________ __________________________________ __________________________________ __________________________________ x – N/A 5 – Very satisfied 4 – Satisfied 3 – Neutral 2 – Dissatisfied 1 – Very dissatisfied 1 2 3 4 5 x 1 2 3 4 5 x 1 2 3 4 5 x 1 2 3 4 5 x 1 2 3 4 5 x 1 1 1 1 2 2 2 2 3 3 3 3 4 4 4 4 5 5 5 5 x x x x 1 2 3 4 5 x Please continue on the back. Last Updated on: 12/11/07 - 38 - Updated by: Cyndi Rose 5. If you responded “Dissatisfied” or “Very dissatisfied” to any item, please explain. ___________________________________________________________________________ ___________________________________________________________________________ ___________________________________________________________________________ ___________________________________________________________________________ 6. If you responded “Unimportant” or “Very Unimportant” to any item, please explain. ___________________________________________________________________________ ___________________________________________________________________________ ___________________________________________________________________________ ___________________________________________________________________________ 7. What was the most positive aspect of your ASI experience? ___________________________________________________________________________ ___________________________________________________________________________ ___________________________________________________________________________ ___________________________________________________________________________ 8. What additional experiential learning programs and services could ASI provide to help meet your needs? ___________________________________________________________________________ ___________________________________________________________________________ ___________________________________________________________________________ ___________________________________________________________________________ 9. Additional comments. ___________________________________________________________________________ ___________________________________________________________________________ ___________________________________________________________________________ ___________________________________________________________________________ 10. Contact information (optional). Name: ________________________________ Email: ____________________ Thank you for your participation! Questions regarding the programs and outcomes delineated in this section should be addressed to Pat Worley, Associated Students, Inc. (916) 278-6784. pcworley@csus.edu. Last Updated on: 12/11/07 - 39 - Updated by: Cyndi Rose Career Center As of July 2007 Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: The Career Center provides proactive and comprehensive career services to students through: career development, experiential learning, on-campus recruitment and employer relations. The professional staff creates a supportive and educational environment where students can actively participate in their own career development within the global work world. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Provide proactive and comprehensive career services. Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Student Learning Outcome: All first-time freshmen participating in the New Student Interactive Tours will understand and be able to utilize career and occupational services. Rationale: Career Center tours are arranged by the instructors of the Freshman Seminar, EOP (Educational Opportunity Program) and Learning Communities. Students from these courses come to the Career Center for an interactive, four-part tour. Each part of the tour teaches the student specific skills and information necessary to understand the services available through the Career Center. Each Interactive Tour includes the following: Introduction to the services the Career Center offers students during each of their four years at the university Exposure to the interactive stations that explain: o How to schedule a career counseling appointment o How to look up a part-time, internship, co-op job o How to look up salary information for jobs in California Information on choosing a major and/or what jobs are related to specific majors Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. Students will take a pre-test measuring knowledge of career resources and how to choose a major. Students will take an immediate post-test using the StudentVoice PDAs at each of the stations to measure knowledge gained in the tours. Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. The paper/pencil survey was given at the end of each interactive tour. See the attached assessment for question details. Results Last Updated on: 12/11/07 - 40 - Updated by: Cyndi Rose Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. Of the 660 first time freshmen who participated in the interactive tours, 495 of them completed the survey. Almost 79% of the respondents were 18 years old. Overall, the students acquired the information and demonstrated the skills indentified as the onset of the program. Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. A review of the data results has prompted us to refocus attention in a few areas. For instance only, 78% of participants could identify what was on the Career Connection main page after their tour. This should be higher than this, so this fall we will spend more time reviewing this with the participants. Also, only 78% of the students could identify all the services that are available to them for walk-in counseling. We want the identification rate to be higher (90%), so we will spend more time reviewing the career connection main page during the tour. Program Objective: Increase by 10% student contact in the Career Center and determine what majors the CC serves most frequently. Rationale: Determine which majors are underserved by the Career Center. Possible Measures Enter all students who utilize the Career Center into a database to track traffic flow. Revise and update current database to process more accurate information. Establish baseline data Fall 2006. Actual Measures Data have been entered into a database to collect baseline information. Results The Career Center increased walk-ins and individual appointments by 11.6% from Fall 2006 to Spring 2007. However, the database does not show the majors of the students counseled. In addition, the attendance at workshops was down by 4.6% in the same time period . Conclusions The Career Center does not know if it actually had an impact on underserved majors. Although our walk-ins and appointments met the target (largely due to the Freshman Advising Program), we realized we need to do more creative advertising for workshops and events. At the end of May 2007 we began working on a marketing plan for fall that involved a greater presence in the Library Quad and the North Quad. We will be doing more tabling, and have also piggy-backed on Welcome Week activities to begin advertising for the September career fair and other Career Center events. Questions regarding the programs and outcomes delineated in this section should be addressed to Beth Merritt Miller, Director, Career Center. 916-278-6231. merrittmillerb@csus.edu. Financial Aid As of July 2007 Last Updated on: 12/11/07 - 41 - Updated by: Cyndi Rose Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: The Financial Aid Office assists students who need financial support to meet the costs of obtaining a college education. Financial aid programs are funded and regulated by the federal government, the state, the University, and private donors. A financial aid package may include one or more of the following: scholarship(s), grant(s), loan(s), and/or work-study. Rationale: Although the primary responsibility for financing an education rests with the student and the student’s immediate family, the Financial Aid Office helps students who require financial assistance in meeting the cost of attending the University. Students who think they may need help in financing their educational costs are encouraged to apply for financial aid. Financial aid may be comprised of grants that are awarded on the basis of financial need (grants need not be repaid); loans, awarded on the basis of financial need and/or cost of attendance, made available from both the University and outside lending institutions (loans must be repaid); and/or federal work-study, awarded on the basis of financial need and earned through employment. Scholarships from the University and private donors (which need not be repaid) are available to eligible students. Financial Aid staff assist students with understanding eligibility criteria and the required procedures and completing necessary forms. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Help students meet the cost of obtaining a college education. Goal 2: Help students learn how to apply for financial aid and complete the process in a timely manner so that funds can be disbursed in a timely manner. Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Student Learning Outcome: Students will be able to 1) demonstrate understanding of the new Satisfactory Academic Progress (SAP) policy, 2) monitor their own academic progress as it relates to financial aid and 3) use their acquired knowledge to reduce the likelihood of financial aid disqualification. Rationale: Most financial aid programs require students to make Satisfactory Academic Progress (SAP) towards their educational objective as a condition for receiving financial aid. Students who fail to meet SAP will be disqualified from receiving aid; however, they may appeal to be reinstated if they have extenuating circumstances approved by the Appeals Board. Many comments regarding SAP policy indicate that neither faculty, staff nor students at Sacramento State fully understood the SAP policy in effect during that time. Therefore, the financial aid staff recommended simplifying the policy. The proposed changes were accepted by the university administration in Fall 2006. The Financial Aid Office rewrote the SAP policy to reflect the updates and the revised policy took effect in Fall 2007. To discern if students understand the revised SAP policy, Financial Aid will administer a “SAP Quiz” to financial aid applicants. When 80% of respondents answer 80% of the quiz questions correctly, the Financial Aid Office will be satisfied that the policy is fairly well understood. Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. Last Updated on: 12/11/07 - 42 - Updated by: Cyndi Rose Randomly select 5,000 potential financial aid candidates and send them the quiz by the end of August 2007 to establish a baseline marker. The Financial Aid Office will send a second quiz along with a copy of the SAP contract to the same 5,000 students by the end of September 2007 to measure their understanding of the policy. Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. Same as possible measures. Results Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. We have not finished collecting the data. We expect to have the results at the end of the Fall 2007 semester. Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. We expect to reach our conclusions at the end of Fall 2007. Student Learning Outcome: Financial aid loan applicants will be able to demonstrate a clear understanding of how to choose their Stafford Loan lender by the end of the fall 2007 semester. Rationale: From the academic year of 1997-1998 through 2005-2006 Sacramento State participated in Direct Loan program and students borrowed money directly from the US Dept. of Education. Though the Direct Loan program was fairly easy to understand, financial aid staff believed that it might be more beneficial to students if we switched to the Federal Family Educational Loan program (FFEL). Both programs are regulated by the federal government and operated under the same rules, however, there are more borrower benefits such as principal balance reduction, interest rate reduction, origination reduction and better customer service associated with FFEL program. In Fall 2006, Sacramento State switched to the FFEL program. This switch confused many students as they were required to choose a lender by themselves. To help students select the most suitable lender for their need, the Financial Aid Office is developing “How to select lender” flyer. The Financial Aid Office will know students understand the borrower benefits when 80% of students who took test answer 80% of the lender associated post test correctly. Possible Measures During the Fall 2007 semester, Financial Aid Office will administer a pre-test to students who see a Financial Aid counselor and conduct a post test after counseling is completed. Actual Measures Same as above. Results We have not finished collecting the data. We expect to have the results at the end of the Fall 2007 semester. Conclusions We expect to reach our conclusions at the end of Fall 2007. Last Updated on: 12/11/07 - 43 - Updated by: Cyndi Rose Questions regarding the programs and outcomes delineated in this section should be addressed to Craig Yamamoto, Financial Aid. (916) 278-6980. craigy@csus.edu. Gender Relations and Sexual Identity Center As of July 2007 Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience Mission: The mission of the Gender Relations and Sexual Identity Center is to confront sexism and homophobia and advocate for the respect and safety of all members of the campus community. The center offers educational programs and advocacy services that increase the awareness of women’s and LGBTI issues, sexual assault, and intimate partner dynamics. It also offers a men’s program that encourages male students to be strong allies in creating and maintaining a healthy and violence-free environment. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Develop students’ understanding of the contributions, opportunities and barriers facing women in contemporary society through participation in the WRC activities. Goal 2: Increase students’ awareness of issues and dynamics involved in violence towards women through participation in the WRC activities. Goal 3: Increase respect for and acceptance of the LGBTIAQQ community on campus as a result of participation in the PRIDE Programs. Goal 4: Build a welcoming campus by developing and implementing a Safe Zone Program that will foster respect for human differences and inclusiveness in a richly diverse community. Program Objective or Student Learning Outcome Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Student Learning Outcome: 70% of students who participate in WRC activities will indicate an increased understanding of the contributions, opportunities and barriers facing women in contemporary society. Rationale: In support of this goal, the Center plans the following: In order to increase the understanding of the contributions, opportunities, and barriers facing women in contemporary society the Women’s Resource Center (WRC) will provide the campus community with programs and advocacy that will focus on issues important to women. The WRC will host and schedule lectures from preeminent women who are experts in issues facing women today. In addition, the WRC will sponsor a monthly Last Updated on: 12/11/07 - 44 - Updated by: Cyndi Rose series of films by and/or about women and contemporary women’s issues. The WRC will also engage in social activism providing advocacy on women’s issues on campus and in the community. Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. General survey of participants (students, faculty, and staff) to measure satisfaction with the activity Focused survey of participants in one or two major activities to measure the impact of the activity on understanding the contributions, opportunities and barriers facing women in contemporary society Survey of participating faculty to measure their perception of the impact of the activity on students that attended one or two major activities Analysis of reflective essays prepared as extra credit to measure the impact of the “noon movies” program Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. Data Collection: Data was collected during Fall 2006 and Spring 07. Methods: The “one minute paper” technique was employed. This technique asks the participants attending the program to write everything they know about the program topic on one side of a blank sheet of paper before the program begins. They are given 60 seconds (1 minute) to complete the task. After the program, they are given 1 minute to write everything they know about the topic on the other side of the same paper (Fall 2006 & Spring 2007). A rubric was developed for each movie identifying the learning outcomes specific to each. The papers were rated according to the rubric, and the difference in scores between the pre and post viewing papers was recorded. A brief survey focused on learning outcomes for an event, Images of African Americans in Film, was employed as a post test only (Spring 07). Population: Students attending the special events or noontime movies. Results Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. In Fall 2006 the percentage of students reporting new learning as a result of viewing the movies was 63%. In the Spring 2007, the percentage increased to 73%. For the event, Images of African Americans in Film, we used the brief survey to collect data. Results indicated that 80% of the respondents identified the main points of learning outcomes. The scores on the individual points ranged from 87.9% to 75%. Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. There are several issues that could have impacted results. We used different raters for the different movies, who also were different from the person who developed the rubrics. It is a very subjective evaluation as to whether or not the students actually “hit” the points in their post papers. Some attention to “rater training” might be helpful in the future. Last Updated on: 12/11/07 - 45 - Updated by: Cyndi Rose We did not complete the data analysis from Fall 2006 in time to make any changes for the Spring 2007 semester. We are discussing different strategies for next year. In terms of the special event, Images of African Americans in Film, the survey data was deemed reliable. These are not events that are repeated so changes are not really feasible. I think that in general, moving to using Student Voice to collect data at special events may impact our response rate and outcomes. Student Learning Outcome: 70% of students who participate in WRC activities will demonstrate an increased understanding of the issues and dynamics involved in violence towards women. Rationale: In support of this goal the Center plans the following:In order to increase the understanding of the issues and dynamics involved in violence towards women The WRC will provide training to the campus community and will continually strive to remain in the forefront on the ongoing research surrounding the issues of violence against women. The WRC will offer multimedia classroom presentations on sexual assault and intimate partner violence throughout the academic year. Additionally, the WRC will examine issues involved in sexual assault cycle of violence and the impact on the campus community by surveying the campus community and keeping up to date with current research. Possible Measures Survey of participating students to measure student awareness of issues and dynamics involved in violence toward women (either pre/post or only post test). Percentage of correct responses to test items (from classroom exams) that relate to information provided in the presentation. Survey of participating faculty to measure their perception of the impact of the presentation on student awareness of the issues and dynamics involved in violence toward women. Actual Measures Data Collection: Data was collected during Summer and Fall 2006, and Spring 2007. Methods: A short survey was used to collect the data. The survey has 3 versions, one focused on presentations made to new students attending summer orientation, another version focused on classroom presentations, and the third was developed for presentations to student clubs and organizations. All surveys included 3 forced choice questions and 2 open ended questions. Data collected in Spring 07 was from classroom presentations only. There are no new student orientation presentations in Spring and data from new Greek member orientations is not available. Population: New students attending summer 2006 orientation session; students in class where presentations were made; and members of student clubs and organizations. Results The data indicated the students found the presentations useful (80.9%); there was no significant difference in the ratings of the peer health educators or staff, both were seen as knowledgeable (94- 92); and 90% indicated that they understood the material that was presented. The most noteworthy finding was that 80% of the respondents thought that the victim was never at fault for her/his assault. This means that 20% of the students believed that the victim is responsible for their assault! Conclusions Reviewing the data during semester break, we looked at redesigning the presentation to focus more specifically on the culpability of the perpetrator. (1/07) Last Updated on: 12/11/07 - 46 - Updated by: Cyndi Rose After the presentation was modified, the percentage of students believing that the victim is never at fault for their assault increased to 90%, according to data collected during Spring 2007. Those presentations were only classroom presentations, and the “real test” will be after the orientation sessions during Summer 2007. Along the other dimensions: usefulness of the information; presenter knowledge; and staff vs. peer health educators, the ratings remained the same as recorded in Summer/Fall 2006. Student Learning Outcome: 70% of students who participate in PRIDE Programs will demonstrate an increased understanding of the issues and dynamics that affect the LGBTIAQQ community on campus. Rationale: In support of this goal, the program plans the following: The Pride Center will increase the understanding of the issues and dynamics that affect eh LGBTIAQQ community by providing training to the campus community and continually examining and evaluating the issues faced on campus and the impact these issue have on campus. The Pride Center will provide to the campus community classroom presentations by panels of students from the Queer Community. Additionally, the Pride Center will examine and explore issues affecting the LGBTIAQQ community and culture intra-group diversity in order to provide appropriate training and education for the campus community. Possible Measures Survey of participating students to measure student awareness of issues and dynamics that affect the LGTIAQQ community (either pre/post or only post test). Percentage of correct responses to test items (from classroom exams) that relate to information provided in the presentation. Survey of participating faculty to measure their perception of the impact of the presentation on student awareness of the issues and dynamics that affect the LGTIAQQ community. Actual Measures Data Collection: Data was collected during Spring / Fall 2006 and Spring 07. Methods: A short survey was used to collect the data. The survey includes 5 forced choice questions and one open ended question (post-test only). Population: Students attending classroom panel presentations. Results The data from both time periods was very consistent. 88 to 89% of the respondents indicated that they had a better understanding of the social issues affecting the Queer community. 94% of the students felt it was a valuable classroom experience to have the panel discussions. Conclusions The panel presentations seem to be effective in meeting the objective. The panels can vary in membership from presentation to presentation, based on the volunteer’s availability. Training/preparation of potential volunteer panelists is being discussed as a way to increase ratings scores. Also, doing a pre test is being considered, as a way to assess impact of the presentations. Another issue under consideration is increasing the variety of the departments/classes that request presentations. Typically the same requests come in each semester. Program Objective: Offer “Safe Zone” training sessions for 25 people on campus by Spring 2006. Last Updated on: 12/11/07 - 47 - Updated by: Cyndi Rose Rationale: In support of this goal, the Center plans the following: The Pride Center will offer Safe Zone trainings for 25 people in order to create twenty five safe zones throughout the campus community. The training will develop a paradigm for competence of issues related to coming out, bisexual people, trans people, and inter-sex people. These trainings will promote understanding of heterosexism and phobias within the campus community, will provide the knowledge of available resources, will enable students and staff the understanding of individual responsibility in everyday actions. Possible Measures Tally of the number of sessions offered and attendance at each session. Tally of the number of Safe Zones created with the objective to develop 25 Safe Zones by Spring 2006. Tally of the demographics of allies to measure diversity of the Safe Zone program. Survey of participants to measure the allies’ understanding/competence regarding issues covered in the program (either pre/post or only post test). Post training survey to measure the allies’ satisfaction with the training program. Follow-up survey (1 semester after training) to measure the allies’ perception of the effectiveness of the program and problems/issues encountered. Actual Measures Data Collection: Data was collected during Spring, 2006 Methods: Tally of the number of participants completing the Safe Zone Training. Population: Students, faculty and staff attending the training sessions. Results A total of 26 people completed the training by the end of the 2005-06 academic year, and the target was met. Conclusions No specific target was set for academic year 2006-07, due to staff changes. However, four Safe Zone trainings were held during Spring 07, and 25 people completed the training. Questions regarding the programs and outcomes delineated in this section should be addressed to Pat Grady. General Relations and Sexual Identity Center. (916) 278-7388. pkgrady@csus.edu. Housing and Residential Life Submitted July 2007 Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: The Sacramento State Housing and Residential Life department provides on-campus residents a safe, well-maintained, attractive, and affordable living learning environment. In addition to the facilities, the department offers a myriad of academic support programs, recreational and social activities, and responsible decision making workshops that promote student success and retention. Rationale: Mission statement should embrace not only the institution’s mission and goals but also standards and guidelines contained within the ACUHO-I (Association of College and University Housing Officers Last Updated on: 12/11/07 - 48 - Updated by: Cyndi Rose International) Standards for College and University Student Housing as well as CAS Standards(Council for the Advancement of Standards for Student Services Development Programs) and professional organizations such as American College Personnel Association (ACPA) and National Association of Student Affairs Professionals (NASPA). Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Continue to offer safe, affordable residence facilities that foster a sense of community and promote academic success. Goal 2: Develop implement, and evaluate a multi-faceted residential education program that promotes personal development, academic achievement, and community responsibility. Goal 3: Use the rich diversity of the residential population as a spring board for planning and implementing programs that foster cross cultural competences. Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Student Learning Outcome: Resident Advisors will be able to describe 1) the types of knowledge they acquire through both formal and on the job training programs and 2) the values and or attributes they develop while serving in this role. Rationale: The Resident Advisor role should help students develop their intellectual and leadership potential, communications skills, and self-esteem. It should also help students engage in deliberate decision making in areas related to values clarification, career development, and health and wellness. Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. A pre-test in November and a post-test in April will be used to assess achievement of this goal. Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. A focus group of thirty Resident Advisors was convened in October 2006 during a regularly scheduled RA class. Feedback was gathered regarding both their formal training and their on the job experience (Appendix A). Initial responses were then used to formulate questions for an exit questionnaire. The exit questionnaire (Appendix B) was administered electronically in April 2007. Here, the RAs were asked to compare their current “comfort” levels with various aspects of the job with the comfort level they had when they were first hired. Results Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. Last Updated on: 12/11/07 - 49 - Updated by: Cyndi Rose RAs reported the most growth in areas related to “esteem” and “leadership.” Comparing responses from when they first started the job to when they ended the job, there was generally an increase from “not so comfortable” to “comfortable.” Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. Areas yielding a “first hired” score of “not so comfortable” or less will be an area of focus for 2007/08 RA training. We will provide more targeted training in each of the aforementioned areas. Program Objective: Ensure, by Spring 2007, that the plans for the new student housing facility (scheduled to open in Fall 2009) will: Explicitly target those student groups that have been identified as likely residents for the new facilities. Meet student needs Project good value both in cost and program Rationale: One of the primary purposes of the Housing and Residential Life program is student learning and development. It is important that the program is purposeful in the organization and management of resources and programs Possible Measures A Spring 2007 student demographic report will initially assess our progress in meeting this goal. Both an occupancy report and a student satisfaction survey will be used (Fall 2009) to assess actual achievement of this goal. A market comparison between on and off campus rental properties will be used (Spring 2007) to analyze cost/value issues. Actual Measures 2006-07 enrollment data will provide base- line on enrollment trends. This data will then be compared to a similar set of data for 2007-08, 2008-09, and opening year 2009-10. . The Customer Satisfaction Survey from the CSU Quality Improvement Program was administered in April 2007, and results were received in June 2007. The online survey was administered to 1,033 Sacramento State students. The response rate was 322 or 31.2%. Specific responses are included in the instrument. Results Campus results are considered preliminary since the CSU system still needs to tabulate the mean values for an inter-University comparison. After our initial review, however, scores are higher than we anticipated: for example, very few students expressed dissatisfaction with the programs and facilities offered in Residential Life. Conclusions When the mean values are published, the Sacramento State Housing staff will identify our strength and weakness’s in relation to other CSUs. We will then identify the areas where we hope to improve in 2008. Primary areas of focus may include substance use abuse and health and wellness education. Improved customer service practices are likely to be targeted as well. Last Updated on: 12/11/07 - 50 - Updated by: Cyndi Rose Program Objective: Develop, by Fall 2007, and implement, no later than Fall 2008, an online licensing program for students who wish to live on campus. Rationale: As stated in CAS standards, Housing and Residential Life must “maintain well-structured management functions, purchasing, contract administration, financial control and information systems.” The new housing facility which opens in Fall 2009 will double our current operations. To prepare for this expansion we are adding electronic technologies to help us process contracts in a timely fashion. Possible Measures Development and implementation of on-line licensing. Actual Measures “Development and implementation of on-line licensing” is a departmental goal and part of on-going operations. Results Specifications were developed for an RFP during the 2006/07 academic year. Several software products were reviewed, and a decision is expected to be made in August 2007. Conclusions Implementation of on-line licensing is an operational goal in progress. Collection of data concerning customer satisfaction will be incorporated elsewhere in other Housing assessments. Questions regarding the programs and outcomes delineated in this section should be addressed to Cynthia Cockrill. Housing and Residential Life. (916) 278-6655. cockrill@csus.edu. Multicultural Center As of October 20, 2006 Mission Statement Note: Mission represents an expanded statement of purpose. Your unit’s mission statement must be consistent with the University's mission statement. You should be able to see how the mission statement of your unit complements the identified mission of the University. Mission: The Mission of the Multicultural Center is to create a campus free of all form of oppression including, but not limited to: racism, sexism, classism, heterosexism, ableism, religious/spiritual oppression. Through education and advocacy, the MCC promotes an environment of cultural learning and exchange that fosters understanding and mutual respect for the uniqueness of each member of the campus community. By sponsoring various speakers and events, the MCC brings diverse perspectives and experiences to the students, faculty and staff as well as the Sacramento area community. Planning Goals Note: Goals are broad statements that describe the overarching long-range intended outcomes of an administrative unit. These goals are usually not measurable and need to be further developed as separate distinguishable outcomes, that when measured appropriately, provide evidence of how well you are accomplishing your goals. They are primarily used for general planning and are used as the starting point to the development and refinement of outcomes. (“Administrative Assessment Handbook” University of Central Florida) Last Updated on: 12/11/07 - 51 - Updated by: Cyndi Rose Create a campus free of racism, sexism, classism, heterosexism, ableism, religious/spiritual oppression and any other forms of oppression. Promote an environment of cultural learning and exchange Foster understanding and mutual respect for the uniqueness of each member of the campus community Planning Objectives Note: Objectives are specific statements that describe desired outcomes identified by an administrative unit. They are derived from the goal statements of the unit. Objectives may relate to the operations and processes of the unit and may also relate to intended behaviors that a student having used the services provided by the unit should demonstrate. (“Administrative Assessment Handbook” University of Central Florida) Objective 1: 70% of students who participate in MCC activities will indicate an increased understanding of the value and enrichment that diversity brings to the community. Rationale: In support of this goal the MCC plans the following: Schedule lectures and programs dealing with various issues of diversity. Engage in social activism around diversity issues. Possible Measures: General survey of participants (students, faculty, and staff) to measure satisfaction with the activity Focused survey of participants in one or two major activities to measure the impact of the activity on understanding the value of diversity. Survey of participating faculty to measure their perception of the impact of the activity on students that attended one or two major activities. Note: To Promote collaboration between Academic Affairs and Student Affairs. Objective 2: Collaborate with at least four faculty members to develop a co curricular activity for one of their classes that explores multiculturalism. Specific student learning outcomes will be developed for each of the classes. Rational: In support of this goal the MCC will: Recruit faculty members for various departments to participate in this initiative. Assist the class in developing and implementing their ideas for a new co curricular multicultural activity, Involve the class members in activities, projects or programs that have already been planned by the MCC. Possible measures: General survey of participants (students, faculty, and staff) to measure satisfaction with the co curricular activity. Focused survey of participants re the identified learning outcome. Survey of participating faculty to measure their perception of the impact of the activity on students that attended one or two major activities. Count the number of faculty who agree to participate in this initiative. Questions regarding the programs and outcomes delineated in this section should be addressed to Pat Grady, Multi-Cultural Center. (916) 278-7388. pkgrady@csus.edu. Last Updated on: 12/11/07 - 52 - Updated by: Cyndi Rose Outreach, Admissions and Records As of July, 2007 Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: Our mission is to attain optimal enrollment through the management of student admissions and registration, to maintain and safeguard the student record, and to ensure the integrity of progress toward degree. From inquiry to admission and enrollment to graduation, we are committed to providing students excellent customer service. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Develop and implement timely and efficient admission and registration processes. Goal 2: Maintain and safeguard student information. Goal 3: Ensure the integrity of the degree. Goal 4: Provide excellent customer services to students. Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Student Learning Outcome: All prospective and continuing students will demonstrate a thorough understanding of the enrollment requirements of Sacramento State. Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. Enrollment reports will be utilized to analyze progress in achieving this goal. Continuing transfer students will be surveyed in Spring 2007 to assess their understanding of registration processes, transfer credits, and graduation requirements. Freshmen will be surveyed in Fall 2007 to assess their understanding of registration processes, EPT and ELM requirements and student orientation (i.e., when/where/how). Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. Data was collected from November 2006 - May 2007. Last Updated on: 12/11/07 - 53 - Updated by: Cyndi Rose Participants in the “Graduation Workshop” completed either a written or electronic survey at the conclusion of the workshop. The surveys measured both direct student learning outcomes (e.g. the extent to which they understood graduation requirements) and student satisfaction levels (willingness to recommend session to a friend). Results Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. Three pilot workshops were held monthly October - December, 2006. Attendance at the first two workshops averaged 50+ students and 15 students attended the December workshop. From the paper surveys collected during those months, 100% agreed that the workshops were beneficial to them and that they would recommend the workshops to other students who were planning to graduate within the year. The paper evaluations also asked “What did you learn from this workshop that was most valuable to you? The written responses varied, but for the most part, students left with a better understanding of the Graduation application process: knowing where to access graduation information on the web and what to expect once they submit an application to graduate. A total of eight workshops were held Spring 2007. Over 95% of the students surveyed indicated that the workshops were beneficial and that they would recommend the workshop to students who were planning to graduate within the year. The survey measured four “take aways” (results are attached). The Application to Graduate initiates the Graduation Evaluation. The minimum number of units required to graduate is 120. Degrees are posted 8 to 10 weeks from the last day of the semester. If the student enrolls in a course not listed on their application to graduate he/she should first check the catalog to see if the course he enrolled in is an approved alternate (if not he should submit a Substitution Petition). Questions 1, 3 and 4 indicated that 88-100% of the students surveyed left with the desired take away. Question 2 indicated that 61% of the students left with the desired “take away.” (The remaining 39% left with an acceptable alternative take away.) Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. The pilot workshops held Fall 2006 were extremely well attended. Beginning Spring 2007 the workshops were held bimonthly. Workshops were available Mondays and Thursdays with a morning or afternoon option. There was also one evening workshop offered. The use of PDAs in assessing workshop effectiveness was also implemented Spring 2007. PDA assessment enabled us to obtain data with respect to specific “take aways” we wanted students to obtain concerning graduation requirements. The outcomes were overwhelmingly positive, an indication that the Graduation Workshops provided students with essential information with regard to graduation from Sac State. Objective #3 - Graduation Workshop Evaluation Results Q1. The Application to Graduate generates my Graduation Evaluation. Count Percent 84 92.31% True 7 7.69% False 91 Respondents Q2. What is the first thing I should do if I enroll in a course that I did not list on my graduation application? Count Percent Last Updated on: 12/11/07 - 54 - Updated by: Cyndi Rose Q2. What is the first thing I should do if I enroll in a course that I did not list on my graduation application? Count Percent 54 60.67% Check the catalog for major course requirements 35 39.33% Submit a Substitution Petition 0 0.00% Change my graduation date 89 Respondents Q3. What is the minimum number of units required to graduate? Count Percent 0 0.00% 51 89 100.00% 120 0 0.00% 160 89 Respondents Q4. My degree will be posted: Count Percent 4 4.49% Prior to commencement ceremony 79 88.76% 8 to 10 weeks from the last day of the semester 6 6.74% 6 months after the commencement ceremony 89 Respondents Q5. Please rate your level of agreement with the following: This graduation workshop was beneficial to me. Count Percent 43 48.31% Strongly agree 42 47.19% Agree 4 4.49% Neutral 0 0.00% Disagree 0 0.00% Strongly disagree 89 Respondents Q6. I would recommend this workshop to other students. Count Percent 55 61.80% Strongly agree 28 31.46% Agree 6 6.74% Neutral 0 0.00% Disagree 0 0.00% Strongly disagree 89 Respondents Q7. All students who plan to graduate within a year should attend this workshop. Count Percent 58 65.17% Strongly agree 28 31.46% Agree 3 3.37% Neutral 0 0.00% Disagree 0 0.00% Strongly disagree 89 Respondents Last Updated on: 12/11/07 - 55 - Updated by: Cyndi Rose Program Objective: Increase our recruitment at high schools and community colleges in the Southern California region for Fall 2007 as compared to Fall 2006. Possible Measures Review of the Southern California Transfer Day College Night schedule. The number of confirmed 2007 attendees will be compared to the number in prior years Actual Measures The data was collected in December 2006 via the Southern California “Transfer Day College” night calendar. Data on school visits is collected through the intersegmental coordinating council’s calendar of Transfer Day College night programs. OA&R confirms attendance and sends a representative to make the visit. Results Confirmation of Fall 2007 attendance at Southern California transfer day college night programs showed a 100% increase over Fall 2006 activities. Southern CA Transfer Day College Night Participation Fall 2007 vs Fall 2006 Date Time Host Fall 07 Fall 06 Yes No 18-Sep Tuesday 10 a.m. to 1 p.m. LOS ANGELES CITY X 19-Sep Wednesday 10 a.m. to 1 p.m. EAST LOS ANGELES X 20-Sep Thursday 10 a.m. to 1 p.m. WEST LOST ANGELES X 20-Sep Thursday 7 p.m. to 9 p.m. Fullerton USD - North @ @ Sonora High School X 24-Sep Monday 7 p.m. to 9 p.m. Fullerton USD - South @ Troy High School X 1-Oct Monday 5 p.m. to 7:00 p.m. COASTLINE (Le-Jao Ctr.) X 2-Oct Tuesday 10 a.m. to 1 p.m. SANTA MONICA X 3-Oct Wednesday 10 a.m. to 1 p.m. CERRITOS X 4-Oct Thursday 10 a.m. to 1 p.m. ORANGE COAST X 4-Oct Thursday 7 p.m. to 9 p.m. Santa Margarita @ Santa Margarita High School X 8-Oct Monday 7 p.m. to 9 p.m. Laguna Beach USD @ Laguna Beach High School X 9-Oct Tuesday 10 a.m. to 1 p.m. EL CAMINO X 10-Oct Wednesday 10 a.m. to 1 p.m. LOS ANGELES HARBOR X 11-Oct Thursday 10 a.m. to 1 p.m. RIO HONDO X 15-Oct Monday 6:30 p.m. to 8:30 p.m. Garden Grove USD @ Bolsa Grande High School X 16-Oct Tuesday 10 a.m. to 1 p.m. MOUNT SAN ANTONIO X 18-Oct Thursday 10 a.m. to 1 p.m. GOLDEN WEST X 18-Oct Thursday 6 p.m. to 8 p.m. Whittier UHSD @ Santa Fe High School X 22-Oct Monday 10 a.m. to 1 p.m. COMPTON X 22-Oct Monday 4:30 p.m. to 7 p.m. El Monte UHSD @ Arroyo High School X 23-Oct Tuesday 10 a.m. to 1 p.m. FULLERTON X 24-Oct Wednesday 10 a.m. to 1 p.m. SADDLEBACK X 25-Oct Thursday 10 a.m. to 1 p.m. CYPRESS X 25-Oct Thursday 7 p.m. to 9 p.m. Huntington Beach UHSD @ Marina High School X 29-Oct Monday 9:30 a.m. to 12:30 p.m. LOS ANGLES TRADE TECH X Last Updated on: 12/11/07 - 56 - Updated by: Cyndi Rose 29-Oct Monday 7 p.m. to 9 p.m. Bishop Amat High School X 30-Oct Tuesday 10 a.m. to 1 p.m. LOS ANGELES SOUTHWEST X 1-Nov Thursday 6:30 p.m. to 8:30 p.m. Hacienda/LaPuente USD @ Los Altos High School X 5-Nov Monday 10 a.m. to 1 p.m. IRVINE VALLEY X 6-Nov Tuesday 10 a.m. to 1 p.m. SANTIAGO CANYON X 7-Nov Wednesday 10 a.m. to 1 p.m. LONG BEACH CITY X 8-Nov Thursday 10 a.m. to 1 p.m. SANTA ANA X Coordinator's Name Barbara Sloan Conclusions Our visible presence in Southern California gives Sac State greater opportunity to interact and influence a larger pool of prospective students. High School and Community College counselors are given a designated contact in the Southern California area to call and ask questions or have a follow-up presentation. In Spring 2008, OA&R will implement an orientation program in the L.A/Orange county region for Southern California applicants. These changes will allow Sac State to gain greater visibility throughout the state. Program Objective: Increase New/Returning undergraduate student enrollment for Fall 2007, as compared to Fall 2006 New/Returning student enrollment. Possible Measures Census date enrollment focus reports will be utilized to assess the success of this objective. Daily application/admissions reports coupled with bi-monthly enrollment funnel reports will be utilized to monitor progress. Actual Measures Collection of data began in November of 2006. Daily application and admissions reports have been utilized and compared with previous year enrollment funnel reports. Early in the Fall 2007 term, data was not available because of the CMS conversion. Currently (June 2007), reports are available, and data can be compared to previous terms. Results The data shows an 8.8% increase in first-time freshman admits to the university and a 13% increase in transfer admits compared to the previous fall 2006 term. Fall 2007 Admissions Update 27-Jun-07 date to date admit numbers comparisons Last Updated on: 12/11/07 Admits Fall 06 Change from Fall 06 % Change from Fall 06 11,217 10,308 909 8.8% 513 531 -18 -3.4% Admits Fall 07 FTF Lower Division - 57 - Updated by: Cyndi Rose Transfer Upper Division Transfer 5,225 4,529 696 15.4% Total Transfer 5,738 5,060 678 13.4% Undergraduate totals 16,955 15,368 1,587 10.3% Conclusions In April 2007 OA&R implemented an admissions blitz with the goal to admit as many eligible applicants as possible before summer orientation began. Over 3,000 students were admitted in a three week time frame. Efforts to streamline and accelerate admission processing clearly helped increase the Fall 2007 admit numbers (even during the CMS implementation year). This approach will be utilized in the Fall 2007 to accelerate admissions decisions for Fall 2008 first time freshman applicants. It is too early to make a full assessment regarding the admission blitz as many factors may have had an influence on the intended outcome of this objective. The final measure of this objective will take place in Fall 2007, on the last day of instruction during the fourth week of classes. Program Objective: All first-time transfers (who have confirmed their intent to enroll by May 1st) will have a completed Transfer Credit Evaluation (TCE) prior to the end of their first semester at Sacramento State. Possible Measures The Transfer Credit Production Report will be used to demonstrate that this objective has been met. Actual Measures CMS query reports were used to obtain data that was then put into an excel sheet for reporting of transfer credit production. Results This objective was achieved in Spring of 2006. The data for Fall 2006 and Spring 2007 shows that 6,228 students entered during these terms though 5,132 transfer credit evaluations were completed by the end of the Spring 2007 term,1,096 transfer credit evaluations still need to be processed as of May 2007. TCE/ Workload as of 5/24/07 Transfer Credit Evaluation # To be Processed # Completed to date # Remaining Time Frame to completion **Tentative TCE files Fall 2006 Spring 2007 Fall 2006 /Spr 2007 Data Entry #Student Contact On-Going 4200 2028 2850* On-Going 570 4,104 1028 2,800 1034 0 96 1000 117 On-Going August 31, 2007^ August 31, 2007^ On-Going On-Going Last Updated on: 12/11/07 - 58 - Updated by: Cyndi Rose Conclusions The new CMS system required evaluators to review several more data screens per student than the former program. Additional screens meant additional steps, and this increase slowed the evaluation process. Additionally, time on the new system took them “off task” for significant periods of time. The system itself also needed to be tweaked and adjusted during this initial start up phase. A Transfer Credit team has been identified and will focus 100% of their time to address the backlog and achieve this objective for the spring term of 2008. Program Objective: All applicants (with complete information) applying for admission within the priority filing period (Oct/Nov) will receive a decision on their status of admission to Sacramento State by the national standard of April 1st. Possible Measures The Priority Filing Period Admissions Report will be assessed for trends to demonstrate that this objective has been met. Actual Measures Focus reports from the OA&R office were utilized to establish applicant numbers. Query reports were created to confirm the actual number of students admitted by the established deadline of April 1 st. Results This objective was not completely achieved. The data shows that 15,000 first time freshman applied to Sac State during the months of October and November. Of these applicants 5,000 where admitted by April 1 st. Conclusions Owing to the transition to CMS this objective was significantly influenced and affected by the transition to the new system. The data reveals that only 33% of this goal was achieved. This objective will be restated for the 07-08 planning objectives. Questions regarding the programs and outcomes delineated in this section should be addressed to Emiliano Diaz, Outreach, Admissions & Records. (916) 278-2674. ediaz@csus.edu. Psychological Counseling Services Submitted July 2007 Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: The mission of Psychological Counseling Services (PCS) is to facilitate Sacramento State students’ learning by helping reduce psychological symptoms and developmental stressors, cope with difficult life events, and balance academic and social life. PCS accomplishes these functions by providing high-quality, timeeffective, and culturally-sensitive counseling, consultation, outreach, testing, training, and psychiatric services. Last Updated on: 12/11/07 - 59 - Updated by: Cyndi Rose Rationale: Psychological counseling has been shown to improve retention (Illovsky, 1997; Wilson, et al., 1997; Turner & Berry, 2000.) In addition, counseling can lead to improved student satisfaction, adjustment, and wellbeing. Overall, counseling removes the impediments to student success. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Help remove students’ psychological/emotional barriers that threaten academic excellence and personal well-being Goal 2: Teach students how to manage strong feelings, resolve conflicts and develop healthy relationships Goal 3: Achieve the above by providing the highest quality, time-effective and culturally sensitive: Individual, couples, and group therapy Crisis intervention and emergency services Outreach Consultation Referral resources Training and Supervision Program evaluation Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Program Objective: Conduct a Needs Assessment with all members of the PCS staff by Fall 2006. Rationale: After several years of interim leadership, it was helpful to determine the direction of the agency as expressed in the interests of each staff member. Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. The new director completed initial one-on-one meetings with staff members in August 2006 and September 2006. The meetings assessed, through individual interview questions, staff beliefs about strengths and areas for growth for PCS. Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. Collection Date(s): August and September, 2006 Method: Individual Interview by Director Populations: Staff members of PCS Results Last Updated on: 12/11/07 - 60 - Updated by: Cyndi Rose Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. 1. What are PCS’ Strengths? Strong clinical skills/talent/experience Dedicated to training/The training program Multidisciplinary and multicultural staff & not in a hierarchy Work well together, like a family, with good staff relationships 2. PCS’ Areas for Growth: More staff Increased service utilization/efficiency (i.e. session limits, other scheduling improvements, reduced wait list) Expanded Group Program/Outreach Program/Workshops Larger and better physical facility Expanded Training Program/APA Accreditation 3. What do you really like about PCS in general? What do you personally like about working here? Colleagues/The staff/Interaction/Relationships/Vibrancy/The environment Seeing client progress/The clientele The cycle of the academic year 4. What would you not want to change? Having us all be in the same location Keep session limits (maybe at 8) 5. What do you not like about PCS? What would you like changed? Too few staff for number of students The physical facility Better community referral options For psychiatrists, not having reliable speech recognition software 6. General goals for PCS Improve training program/APA accreditation 7. What do you think our goals should be in the following areas [These should be measurable, attainable in the coming year, and significantly challenging]: A. Program/Service Improvement… Revamp intake/disposition process o Lower session limits & more groups to decrease wait list Implement Outcome Measurement system B. Student Learning Outcomes (Changes in students’ attitudes/behaviors in response to using our services) Measure therapy outcome with revised client survey Reduce stigma with more workshops and mental health awareness promotion C. Collaborative goal: Student Health Center More training of RAs and Res Life staff Women’s Resource Center Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. Last Updated on: 12/11/07 - 61 - Updated by: Cyndi Rose The director compiled this information and presented it in aggregate form to the entire staff at the end of October 2006. The information was used as a springboard to begin a series of discussions regarding areas for improvement of services. The director worked with administration on staffing issues, and staff worked together to consider internal efficiency issues. We expanded our group program to include three consistent groups, a 33% increase in number of groups over the previous year. We worked with the health center and administration to design a part of the new Recreation & Wellness Center for PCS. We created the position of Case Manager (who will research and assist with community referral options) and hired a person who will start in October 2007. We purchased voice recognition software for our psychiatrists. Student Learning Outcome: To adopt and utilize a comprehensive therapy outcome/client questionnaire to track changes in overall client functioning, symptom reduction and impact on academic performance. Such that: At least 60% of respondents who have received services on at least three occasions will indicate counseling services have been effective in the above-mentioned areas. More specifically, at least 60% of respondents who have received services on at least three occasions will indicate counseling services have led to a positive (behavioral) change in their life. Our revised questionnaire will be used by Fall 2006, and a new national, more comprehensive questionnaire, the CCAPS, will be used in 2007 or 2008. Rationale: Assessing our client’s report of improvements provides us the opportunity to determine areas for improvement as well as demonstrate areas of strength. Possible Measures PCS vastly redesigned their Outcome & Satisfaction Survey in Fall 2006 in order to measure client functioning, symptom reduction and impact on academic performance. This instrument provides a self-report from students, as measured several times during the academic year. PCS is also now administering this survey on PDAs using Student Voice technology. Actual Measures Collection Date(s): Four two week periods, at the beginning and end of Fall and Spring terms 2006-2007. Method: Students are administered the 40 question survey a PDA prior to their session. Population: Students who have been seen at PCS as clients at least twice. Results As to effectiveness, 76% of students indicated either “good” or “excellent” effectiveness of services. As to positive changes, 94% of students indicated that what they learned in therapy will have a positive change in their life. We also obtained data on a number of other specific questions. Please see Appendix A for complete survey results. Conclusions The final administration of the survey occurred at the end of the Spring term. Full staff will not return until August, at which time we will discuss these results. However, from the very positive results, it appears initially that no significant changes are needed in terms of students’ perceived effectiveness of services and amount of positive change that resulted. Last Updated on: 12/11/07 - 62 - Updated by: Cyndi Rose In regard to the instrument used, PCS plans to participate in a national counseling center project that uses a standardized instrument, the CCAPS, along with the Titanium scheduling system, for data collection (which ties into Goal 3 below.) This national outcome measure will be useful for PCS because we can compare our aggregate results against those of other institutions, and we can choose to participate in national research projects. Currently, the national project group is working to refine the standardized instrument. Program Objective: To implement a new scheduling system (and possibly, an electronic record-keeping system, as well) to track utilization and improve communication with providers. Scheduled system will be used by Fall 2006; electronic record-keeping will be used by Fall 2007 or Fall 2008, depending on the technological safeguards that are required. Rationale: Keeping more detailed data on our service provision will help us identify areas in which to improve. The scheduling system also allows for enhanced communication amongst providers and our front desk personnel, allowing for more efficient scheduling. Keeping records in saved electronic fashion allows for easier access to client records and a more streamlined system. It is a step towards moving to an all-electronic medical records (EMR) system. Possible Measures Reports as provided by the Titanium Scheduling Software. Actual Measures Collection Date: Titanium was installed in summer 2006. After a semester of use, we reviewed its generating capacity. Method: The software program keeps a variety of statistics on appointments. It has a “reports” function that is accessed to measure data over discrete periods of time. Population: Students/Counselors Results We determined that clients are in counseling for an average of 5 sessions, precisely the national average. We also determined that our no-show rate is 10% and our cancellation rate is 10%. Our cancellation rate is 5% higher than the national average. Effective July 1, 2007 all notes will be done electronically. For 2007-2008 we will continue to keep backup paper records. We then will consider eliminating paper records by Fall 2008, after consultation with the IT Department in regards to security and backup capabilities. Conclusions Given that our cancellation rate is high we proposed to the Student Fee Committee that students who no-show or cancel with less than 24 hours notice be charged $10. The Fee Committee approved our proposition, and we expect to implement this fee in the fall. Program Objective: Examine and modify existing policies and procedures to improve time utilization of providers and improve student access to services. This is a continual process; however, we hope to make significant inroads by Fall 2007. Last Updated on: 12/11/07 - 63 - Updated by: Cyndi Rose Rationale: Increased efficiency our service provision will allow us to see more students with the same funding. Possible Measures Staff discussion regarding alternate forms of scheduling and different methods of calling clients who are on wait list. Actual Measures Collection Date(s): Throughout 2006-2007 academic year at regular and permanent staff meetings. Method: Discussion of various proposals. Population: Clinical and Reception PCS staff. Results Staff determined that our current methods of scheduling may not be the most efficient, and several staff members agreed to pilot alternate methods. In addition, we determined that calls to clients on our wait list could be made more quickly. Lastly, in the Spring term we determined a need for additional permanent staff meetings as well as additional discussion amongst the director and a few permanent staff members. Conclusions Staff decided we would experiment with alternate forms of scheduling beginning in the summer in order to maximize efficiency. We are piloting the reduction of time periods for students to return calls regarding scheduling. We also implemented a new permanent staff member meeting at least once a month, as well as a Resource Team composed of several key staff members meeting with the director weekly. Program Objective: Improve data analysis of student demographic and presenting issues. Rationale: Additional information in this area will allow for greater allocation of resources as well as programmatic decisions. Possible Measures In the past we collected a minimal amount of data from students when they presented to our center. We decided in Spring 2007 to expand our intake form and reviewed similar forms used by other counseling centers. In addition, as part of our participation in the Center for Study of Counseling Services in Mental Health’s national outcome project (referenced above) we incorporated the core elements of their new national intake form, the Standardized Data Set (SDS.) Actual Measures Collection Date(s): Intake form modified Spring 2007; Will be used in Fall 2007 Method: Reviewed forms of other counseling centers and used core items from SDS to develop new intake paperwork. Will administer to student beginning Fall 2007. Population: PCS clients. Results Last Updated on: 12/11/07 - 64 - Updated by: Cyndi Rose The result of our research into intake paperwork is our revised PCS Intake Questionnaire. See copy in Appendix B. (Note due to margin formatting, the actual document is more compressed than the one reproduced here.) Conclusions Staff will meet at the beginning of the Fall 2007 semester to finalize the Intake Questionnaire format. We will then begin using it with students. We will enter the data into our Titanium Program and begin analysis in the 0708 academic year. Appendix A PCS Outcome & Satisfaction Survey 2006-2007 Q1. Today's provider: (scroll down to see all answer choices) Count Percent 17 7.94% Adam 10 4.67% Ann 7 3.27% Bert 10 4.67% Brian 13 6.07% Cheri 18 8.41% David 10 4.67% Esther 18 8.41% Janice 12 5.61% Lori 11 5.14% Mary 4 1.87% Ona 11 5.14% Paul 3 1.40% Sandra 9 4.21% Starla 15 7.01% Tani 25 11.68% Dr. Berg 21 9.81% Dr. Boyle 214 Respondents Q2. Age: Count Percent 2 0.95% 41 19.43% 18-21 113 53.55% 22-29 36 17.06% 30-39 19 9.00% 211 Under 18 40+ Respondents Q3. Ethnicity: Count Respondent % Response % 1 0.48% 0.47% American Indian or Alaska Native 20 9.52% 9.30% Asian American or Pacific Islander 14 6.67% 6.51% African American/Black 27 12.86% 12.56% Latino(a)/Chicano(a)/Hispanic 122 58.10% 56.74% European-American/ Caucasian/White 16 7.62% 7.44% Biracial/Multiracial 2 0.95% 0.93% International 13 6.19% 6.05% Other (please specify) Last Updated on: 12/11/07 - 65 - Updated by: Cyndi Rose Q3. Ethnicity: Count Respondent % 210 Respondents 215 Responses Response % Q4. Gender: Count Percent 55 26.19% Male 154 73.33% Female 1 0.48% 210 Transgender Respondents Q5. Your School/College: Respondent % Response % 46 22.22% 21.90% Arts and Letters 12 5.80% 5.71% Business Administration 2 0.97% 0.95% Continuing Education 26 12.56% 12.38% Education 12 5.80% 5.71% Engineering and Computer Science 38 18.36% 18.10% Health and Human Services 19 9.18% 9.05% Natural Sciences and Mathematics 40 19.32% 19.05% Social Sciences and Interdisciplinary Studies 15 7.25% 7.14% Undecided Count 207 Respondents 210 Responses Q6. Class Standing: Count Percent 8 3.86% First-Year (Freshman) 13 6.28% Sophomore 43 20.77% Junior 83 40.10% Senior 56 27.05% Graduate student 4 1.93% 207 Other: Respondents Q7. Type of services you received this semester: Count 167 Respondent % Response % 81.07% 62.08% On-going individual counseling 9 4.37% 3.35% Group therapy 66 32.04% 24.54% Psychiatric (medication management) services 22 10.68% 8.18% Testing/LD assessment 5 2.43% 1.86% Biofeedback 206 Respondents 269 Responses Q8. Approximate number of on-going individual sessions: Count Percent 12 9.23% 47 36.15% 2-3 19 14.62% 4-5 12 9.23% 6-7 18 13.85% 8-9 22 16.92% 10 + 130 1 Respondents Q9. The courtesy and helpfulness of the front desk staff. Count Percent 155 76.73% Excellent 40 19.80% Good 5 2.48% Fair 1 0.50% Poor 1 0.50% N/A 202 Respondents Last Updated on: 12/11/07 - 66 - Updated by: Cyndi Rose Q10. Your comfort waiting in the reception area. Count Percent 101 50.25% Excellent 78 38.81% Good 18 8.96% Fair 2 1.00% Poor 2 1.00% N/A 201 Respondents Q11. The availability of same day drop-in appointments. Count Percent 55 27.36% Excellent 45 22.39% Good 22 10.95% Fair 10 4.98% Poor 69 34.33% N/A 201 Respondents Q12. The ability to get an appointment after the initial drop-in session in an acceptable amount of time. Count Percent 53 26.37% Excellent 64 31.84% Good 31 15.42% Fair 23 11.44% Poor 30 14.93% N/A 201 Respondents Q13. Comfort with staff handling of confidentiality. Count Percent 142 70.65% Excellent 48 23.88% Good 4 1.99% Fair 2 1.00% Poor 5 2.49% N/A 201 Respondents Q14. Satisfaction with the clinician you saw today. Count Percent 127 63.18% Excellent 48 23.88% Good 5 2.49% Fair 1 0.50% Poor 20 9.95% N/A 201 Respondents Q15. Satisfaction with the PCS brief therapy model. Count Percent 57 28.36% Excellent 48 23.88% Good 21 10.45% Fair 3 1.49% Poor 72 35.82% N/A 201 Respondents Q16. Effectiveness of your counseling/psychiatric/testing services. Count Percent 86 42.79% Excellent 66 32.84% Good 11 5.47% Fair 3 1.49% Poor 35 17.41% N/A 201 Respondents Q17. Clinician’s sensitivity to cultural and individual differences. Count Percent 99 49.50% Excellent Last Updated on: 12/11/07 - 67 - Updated by: Cyndi Rose Q17. Clinician’s sensitivity to cultural and individual differences. Count Percent 61 30.50% 5 2.50% Fair 1 0.50% Poor 34 17.00% N/A 200 Good Respondents Q18. The overall service you are receiving. Count Percent 114 57.00% Excellent 78 39.00% Good 5 2.50% Fair 1 0.50% Poor 2 1.00% N/A 200 Respondents Q19. I would recommend PCS services to a friend. Count Percent 139 69.50% Strongly Agree 54 27.00% Agree 0 0.00% Disagree 1 0.50% Strongly Disagree 6 3.00% N/A 200 Respondents Q20. I would return to PCS if I needed help again. Count Percent 145 72.50% Strongly Agree 50 25.00% Agree 1 0.50% Disagree 1 0.50% Strongly Disagree 3 1.50% N/A 200 Respondents Q21. I regard PCS as a necessary part of the University. Count Percent 171 85.50% Strongly Agree 27 13.50% Agree 0 0.00% Disagree 1 0.50% Strongly Disagree 1 0.50% N/A 200 Respondents Q22. Increase concentration Count Percent 40 20.00% Strongly Agree 95 47.50% Agree 11 5.50% Disagree 2 1.00% Strongly Disagree 52 26.00% 200 N/A Respondents Q23. Attend classes regularly Count Percent 68 34.00% Strongly Agree 68 34.00% Agree 13 6.50% Disagree 2 1.00% Strongly Disagree 49 24.50% 200 N/A Respondents Q24. Reduce procrastination Count Percent 30 15.00% Strongly Agree 80 40.00% Agree Last Updated on: 12/11/07 - 68 - Updated by: Cyndi Rose Q24. Reduce procrastination Count Percent 33 16.50% 3 1.50% 54 27.00% 200 Disagree Strongly Disagree N/A Respondents Q25. Manage test anxiety Count Percent 34 17.09% Strongly Agree 68 34.17% Agree 16 8.04% Disagree 2 1.01% Strongly Disagree 79 39.70% 199 N/A Respondents Q26. Improve my study skills Count Percent 23 11.56% Strongly Agree 75 37.69% Agree 21 10.55% Disagree 3 1.51% 77 38.69% 199 Strongly Disagree N/A Respondents Q27. Improve time management Count Percent 24 12.12% Strongly Agree 73 36.87% Agree 27 13.64% Disagree 1 0.51% 73 36.87% 198 Strongly Disagree N/A Respondents Q28. Find useful campus resources Count Percent 37 18.69% Strongly Agree 67 33.84% Agree 9 4.55% Disagree 0 0.00% Strongly Disagree 85 42.93% 198 N/A Respondents Q29. Prior to counseling, I was thinking of leaving the University before completing my degree. Count Percent 14 7.07% 26 13.13% Agree 37 18.69% Disagree 68 34.34% Strongly Disagree 53 26.77% N/A 198 Strongly Agree Respondents Q30. Counseling helped me to stay at Sacramento State. Count Percent 13 32.50% Strongly Agree 17 42.50% Agree 2 5.00% Disagree 2 5.00% Strongly Disagree 6 15.00% 40 N/A Respondents Q31. PCS services helped me improve my relationships. Count Percent 50 25.25% Strongly Agree 92 46.46% Agree 12 6.06% Disagree Last Updated on: 12/11/07 - 69 - Updated by: Cyndi Rose Q31. PCS services helped me improve my relationships. Count Percent 1 0.51% 43 21.72% 198 Strongly Disagree N/A Respondents Q32. PCS services helped me feel less stressed/anxious. Count Percent 53 26.77% Strongly Agree 106 53.54% Agree 10 5.05% Disagree 1 0.51% Strongly Disagree 28 14.14% 198 N/A Respondents Q33. PCS services helped me feel less sad/depressed. Count Percent 53 26.77% Strongly Agree 100 50.51% Agree 11 5.56% Disagree 4 2.02% Strongly Disagree 30 15.15% 198 N/A Respondents Q34. PCS services helped me make important decisions. Count Percent 43 21.72% Strongly Agree 89 44.95% Agree 11 5.56% Disagree 1 0.51% Strongly Disagree 54 27.27% 198 N/A Respondents Q35. PCS services helped me feel more connected at Sacramento State. Count Percent 26 13.13% Strongly Agree 76 38.38% Agree 32 16.16% Disagree 6 3.03% 58 29.29% 198 Strongly Disagree N/A Respondents Q36. PCS services helped me make healthier lifestyle choices. Count Percent 45 22.73% Strongly Agree 82 41.41% Agree 15 7.58% Disagree 4 2.02% Strongly Disagree 52 26.26% 198 N/A Respondents Q37. PCS services helped me cope better with the challenges of life. Count Percent 67 34.01% Strongly Agree 88 44.67% Agree 8 4.06% Disagree 3 1.52% Strongly Disagree 31 15.74% 197 N/A Respondents Q38. PCS services helped me be more assertive. Count Percent 38 19.29% Strongly Agree 80 40.61% Agree 16 8.12% Disagree 1 0.51% Strongly Disagree Last Updated on: 12/11/07 - 70 - Updated by: Cyndi Rose Q38. PCS services helped me be more assertive. Count Percent 62 31.47% 197 N/A Respondents Q39. PCS services helped me develop skills that will be useful in a variety of other settings. Count Percent 53 26.90% Strongly Agree 88 44.67% Agree 6 3.05% Disagree 1 0.51% Strongly Disagree 49 24.87% 197 N/A Respondents Q40. What I have learned from coming to PCS has led to positive changes in my life. Count Percent 186 94.42% Yes 11 5.58% No 197 Respondents Appendix B Psychological Counseling Services (PCS) Drop-In Counseling, Intake Questionnaire Date Name Social Security Number __ Name you would like us to call you (if different from above) ________________________________________________ Who referred you to us? __________________ D.O.B. Age __ Marital Status _____________ Phone (home) ____ ____ (cell) ___________ Where may we call and leave a message?: home______ Address (work) ______ cell ______ work ______ City Zip ________________ NOTE: EMAIL IS NOT CONSIDERED A SECURE FORM OF COMMUNICATION Email ___________________________________ May we contact you by email for scheduling? Y N In case of emergency contact ______________________________________________________________________ Relationship _____________________________ Phone _____________ Last Updated on: 12/11/07 - 71 - Updated by: Cyndi Rose Racial/Ethnic Group: ___ African-American/Black/African ___ American Indian/Alaskan Native ___ Arab American/Arab/Persian ___ Asian American/Asian ___ East Indian ___ European American/White/Caucasian ___ Native Hawaiian/Pacific Islander ___ Hispanic/Latino/Latina ___ Multi-racial ___ Prefer Not to Answer ___ Other (please specify) ________________________ If not U.S., what is your country of origin?___________________________ Is English your first language? Y N If no, what is your native language? _____________________________ Do you have medical insurance? Y NName of Insurance Company __________________________________ Have you had previous counseling here at PCS? Y N When? ____________________________________ Somewhere else? Y N Where? ________________________ When? ______________________________ What kind of housing do you currently have? ___ On-campus ___ Fraternity/sorority house ___ Live w/in 5 miles of campus ___ Live over 5 miles from campus Do you work? __ Y __ N If so, where: ____________________ Average number of hours worked per week: ______ ------------------------------------------- Please complete information on reverse side ----------------------------------------------------To be completed by Drop-In Therapist DROP-IN PROVIDER _________________________ PRIORITY? 1 2 3 4 1:1/Cpls/Bio _________ Testing __________ Med Eval 1 2 3 4 5 Group _____________ CLINICIAN PREFERENCES: Anyone _______ Other (specify): Have you ever been enlisted in any branch of the US Military (active duty, veteran, guard, reserves)? Y* N Last Updated on: 12/11/07 - 72 - Updated by: Cyndi Rose 5 * Did your military experiences include highly traumatic/stressful experiences that continue to bother you? Y N PLEASE CHECK ALL THAT APPLY: __ Freshman/First-Year __Full Time Student (7 units or more) Major __ Sophomore __Part Time Student (6 units or less) ______________ __ Junior __International Student __ Senior __Veteran GPA __ Graduate Student __ Disabled/Challenged ________ __ Male __ Female __ Transgender __ Other __ Prefer Not 2 Answer Do you have a diagnosed and documented disability? __ Attention Deficit/Hyperactivity Disorders __ Deaf or Hard of Hearing __ Learning Disorders __ Mobility Impairments __ Neurological Disorders __ Physical/health related __ Visual Impairments __ Other Disorders __ Psychological Disorder/Condition If you checked “Other” above, please describe: _________________________________________________________ Do you have any other significant medical conditions? If yes, please describe: Yes No _____________________________________________________________________________ What medications (if any) are you taking specifically for a mental health condition? __________________________________________________________________________________________ ______ Are you currently taking any other medications? Yes No Name of medication(s): ______________________________________________________ Please indicate if and when you have had the following experiences: 1. 2. 3. 4. 5. 6. 7. Attended counseling for mental health concerns _____ Never _____ Prior to college _____ After starting college _____ Both Taken a prescribed medication for mental health concerns _____ Never _____ Prior to college _____ After starting college _____ Both Been hospitalized for mental health concerns _____ Never _____ Prior to college _____ After starting college _____ Both Received treatment for alcohol or drug use _____ Never _____ Prior to college _____ After starting college _____ Both Purposely injured yourself without suicidal intent (e.g., cutting, hitting, burning, hair pulling, etc.) _____ Never _____ Prior to college _____ After starting college _____ Both Seriously considered attempting suicide _____ Never _____ Prior to college _____ After starting college _____ Both Made a suicide attempt Last Updated on: 12/11/07 - 73 - Updated by: Cyndi Rose _____ Never _____ Prior to college _____ After starting college _____ Both Seriously considered injuring another person _____ Never _____ Prior to college _____ After starting college _____ Both Intentionally injured another person _____ Never _____ Prior to college _____ After starting college _____ Both Had unwanted sexual contact(s) or experience(s) _____ Never _____ Prior to college _____ After starting college _____ Both Experienced harassing, controlling, and/or abusive behavior from another person? (e.g., friend, family member, partner, or authority figure) _____ Never _____ Prior to college _____ After starting college _____ Both 8. 9. 10. 11. 12. Major reasons for seeking help? 13. Since when have these things bothered you? 14. Below is a list of concerns commonly experienced by college students. Using the following scale, please circle the number indicating the degree to which each item is presently a concern for you. Not at All 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 2 A little bit 3 Moderately 4 Quite a bit 5 Very Much 5 1. Dealing with stress or pressure 5 2. Absent from classes too often 5 3. Thinking of dropping out of school 5 4. Adjusting to the university, campus or living environment 5 5. Feeling depressed, sad, or down 5 6. Choosing a major or establishing a career direction 5 7. Death or illness of a significant person 5 8. Academic progress, courses, test or performance anxiety, time management 5 9. Difficulties related to sexual identity or sexual orientation 5 10. Relationships with family members (parents, siblings, children, relatives) 5 11. Feeling anxious, fearful, worried or panicky 5 12. Feeling unmotivated, procrastination, or difficulty concentrating 5 13. Feeling irritable, tense, angry, or hostile 5 14. Money, finances 5 15. Feeling lonely, isolated, or uncomfortable with others 5 16. Values, beliefs, religion, or spirituality 5 17. Sexual trauma (sexual abuse/assault, incest, rape) 5 18. Low self-esteem or self-confidence 5 19. Legal matters 5 20. Someone else’s habits or behaviors 5 21. Unwanted/out-of-control behaviors, habits, or thoughts 5 22. Problems with assertiveness or shyness 5 23. Sleep problems 5 24. Pregnancy 5 25. Eating problems (bingeing, restricting, low appetite, vomiting, laxative use, etc.) 5 26. Relationships with romantic partner/spouse 5 27. Physical health problems (headache pain, fainting, injury, fatigue, etc.) 5 28. Sexual matters (sexually transmitted disease, sexual functioning, etc.) 5 29. Relationships with instructors or other university personnel 5 30. Discrimination 5 31. Feelings of guilt or self-criticism 5 32. Weight or body image problems 5 33. Difficulties trusting others Last Updated on: 12/11/07 - 74 - Updated by: Cyndi Rose 1 2 3 4 5 1 2 3 4 5 1 2 3 4 5 34. Addiction or Substance Use 35. Suicidal/Homicidal Thoughts or Intentions 36. Other: ______________________________________________________________ Questions regarding the programs and outcomes delineated in this section should be addressed to Bert Epstein, Psychological Counseling Services. (916) 278-3905. epsteinb@csus.edu Student Activities As of July 2007 Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: Student Activities promotes co-curricular learning by providing students with opportunities to join organizations, participate in recreational activities, and engage in training programs. The aforementioned programs and services hone students’ leadership, lifelong learning, and healthy/ethical decision-making skills. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Increase student involvement in co-curricular activities Goal 2: Enhance experiential learning, healthy decision-making, and life skill development Goal 3: Maximize leadership opportunities, both on and off campus, through interaction and collaboration Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Program Objective: By December 2006 design a multi-faceted Leadership Education program that Enhances the leadership competencies of current student leaders. Rationale: Student Activities will invite various student affairs offices, academic departments, and students to critique proposed program elements and recommend changes. The comprehensive program will include opportunities to develop leadership competencies in areas such as communication, decision making, and group dynamics. Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. During the Fall 2006 semester, the director will summarize and analyze current literature and existing leadership education programs at other higher education institutions. This report will include a compilation of leadership Last Updated on: 12/11/07 - 75 - Updated by: Cyndi Rose courses listed in the California State University, Sacramento 2006-2008 Catalog. Based upon findings, he will purpose the type of multi-faceted program he feels is best suited for Sacramento State. Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. “Premise for a Comprehensive Leadership Education Program” was presented to the Vice President of Student Affairs for her consideration, review, and support. [Report attached: Appendix #1] Results Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. From the information and ideas gleaned while researching the report, a number of initiatives and activities have been undertaken: The Social Change Model of leadership was identified as the program most closely aligned with Sacramento State’s mission statement. This model will frame the emerging leadership education program being developed by the Student Activities office. Two Program Advisors attended a Social Change summer workshop to hone their understanding of this particular model. Dr. Mike Lee spoke about infusing the curriculum with social justice issues, which serves as the basis of the Social Change Model. Additional efforts will be made to provide resources to the Center for Teaching and Learning. Copies of “A Social Change Model of Leadership Development Guidebook” were purchased for departments that are expected to partner with Student Activities to operationalize the model. Departments that received copies of the social change model include: ASI, Residential Life, Gender Equity and Sexual Identity Programs, the Office of Community Engagement, and UNIQUE Programs. To avoid over programming and duplication of efforts, and to maximize promotional opportunities, a 4month planning calendar is being developed within Student Activities. Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. In addition to the aforementioned initiatives the following activities are planned for the upcoming semester: This years annual Student Activities leadership retreat will include an abbreviated Challenge Course activity as a means of building community among student leaders. Working with Alumni Affairs, the Student Activities office will feature - on a poster series- six to ten successful alumni who used their co-curricular leadership experience as a stepping stone to professional positions. Additionally, Alumni Affairs will sponsor a U.S. mailing to all Greek Alumni prior to Homecoming. In co-operation with Communication Studies, Student Activities will assign a student intern to develop a public relations campaign to promote our activities. The need for such an intern was identified through the focus groups conducted as part of Goal #2. Student Activities is supporting a current student leader who is bringing together campus constituents via an Online Student Leadership Forum. The student has sought and received support from ASI leaders, Alumni Affairs, and the President. Last Updated on: 12/11/07 - 76 - Updated by: Cyndi Rose Program Objective: Conduct focus groups and/or Student Voice surveys to determine the needs of emerging and established leaders. Rationale: While the Student Activities Office provides a variety of services and programs, unmet student needs have not been identified. The Student Activities Office will explore what students know about Student Activities, what students say they need from Student Activities, and what barriers prevent them from participating more fully in Student Activities programs and services. Possible Measures Identify and invite students selected at ‘random’ from three different populations: Students who were believed to have had more than one contact with Student Activities; Students who were believed to have had one contact with Student Activities; and Students chosen at random from a list provided by the Registrar’s Office. Actual Measures Focus groups of both active participants and the broader student population were conducted to determine what types of student needs are unmet and how Student Activities can better communicate with students. Focus Groups were conducted by Dr. Elaine Gale, Communication Studies, in February and March 2007. The Focus Groups were designed specifically for students who were believed to have had more than one contact with Student Activities; students who were believed to have had one contact with Student Activities; and a random list from the Registrar’s Office of students who may or may not had any contact with Student Activities. Participants were told the session would last no more than an hour, that they would be paid $20 for participating, and that food/beverages would be provided. Dr. Gale reported participants in each group were very active. It should be noted that no activities staff member participated in any of the focus groups. Important perceptions and misperceptions were identified during the course of these focus groups. Students were also asked to complete a paper survey to determine some basic demographic information and levels of involvement, on and off campus. Results See Appendix #2: Focus Group Responses; #2B: Paper Questionnaire Results from Focus Groups; #2C: What we want to know Conclusions Students are confused about and often interchange ASI, UNIQUE, and Student Activities. On the one hand, the Student Affairs Office is blamed for things they do not like about ASI, on the other hand, we get credit for some of the excellent programming produced by UNIQUE. Students are very active on campus and in the community. Over 95% of the students expect to hold leadership roles in the future. Word of mouth is the most effective way to get information out to students. Most students spend over two hours a day chatting online with friends. Program Objective: By December 2006, begin to implement some of the issues and activities identified in Goal 1 and 2; increase overall student participation in Student Activities programs by 30%. Last Updated on: 12/11/07 - 77 - Updated by: Cyndi Rose Rationale: To take full advantage of Student Activities, students must be cognizant of what we offer. A comprehensive marketing program will be developed and implemented to ensure that students are fully aware of the services and support provided by Student Activities. Possible Measures Track students who participate in programs during the Spring 2007 semester. Use this data as a baseline for future comparisons. Actual Measures Students were able to register for workshops and learning opportunities online through our website. This process enabled us to track student registration. In addition, we continue to request program evaluations from those who completed the workshops. Results In an effort to avoid conflicts with final exams, the dates for the Focus Groups were moved from November/December 2006 to February 2007. We gathered information from focus group participation regarding their preferred communication. The bottom line is students use electronic media more than other forms. For example, 65% of the students surveyed have a myspace account. Over 50% of the respondents use internet chat rooms 1 – 2 hours per day. Given this, Student Activities has started a myspace page at www.myspace.com/csusstudentactivities to provide them with information. In addition, the theme for involvement this year will be “SacStateTube.com” based upon the very popular “YouTube.com” website. This will be interactive and we have ‘purchased’ the domain name. Starting in the Fall 2007, we will allow student organizations to indirectly upload video content. Attendance at the Nuts and Bolts workshop increased during 2006 – 2007 over 2005 – 2006 by only 6%. Student Activities will supervision and direction for a student intern through Communication Studies specializing in public relations. It is my hope to continue this program beyond the initial semester. Conclusions A review of the leadership research and comparable programs at other institutions has indicated that Sacramento State is at the beginning of this comprehensive program development. To move forward the Student Activities staff must continue to participate in professional activities, familiarize themselves with additional research, and continue to collect baseline data. Progress on each of those activities will be reported at the end of the Fall 2007 semester. Questions regarding the programs and outcomes delineated in this section should be addressed to Lou Camera, Student Activities. (916) 278-6595. cameral@csus.edu Student-Athlete Resource Center As of July 2007 Mission Statement Last Updated on: 12/11/07 - 78 - Updated by: Cyndi Rose Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: The Student-Athlete Resource Center (SARC) provides Division I student-athletes with the support to develop as independent and successful young adults and the skills to achieve their academic goals, to persist towards graduation, and to meet all NCAA eligibility requirements. SARC offers academic advising, tutoring and mentoring, NCAA rules compliance and eligibility education, financial services support, and life skills programming to these athletes. SARC’s services and programming are provided in an assortment of ways, including: individual advising, orientations, group and team meetings, coaches’ compliance and rules education sessions, freshmen seminar classes, and various student development and academic workshops. Rationale: SARC staff helps student-athletes navigate through many of the campuses’ academic procedures and policies, understand NCAA rules compliance and eligibility requirements, and provide services that place student welfare and the development of the whole student as a top priority. In addition, the SARC assists athletic coaches, staff, faculty, and administration in achieving overall integrity and success within our Division I athletics program. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Provide student-athletes with the support to develop as independent and successful young adults Goal 2: Plan, implement, and assess student athlete retention programs Goal 3: Assist student-athletes in developing the skills necessary for achieving their academic goals Goal 4: Continuously inform student athletes about NCAA eligibility requirements Goal 5: Provide rules education and compliance monitoring on NCAA and affiliated conference rules and requirements Goal 6: Augment the internal and external communities’ understanding of NCAA-related academic data and its impact on student-athletes, the Athletic Department and the overall campus. Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Program Objective: Increase student-athletes’ comprehension of GE requirements, campus academic policies, NCAA academic requirements and their awareness of other campus support services, such that 75% of studentathletes surveyed/tested will demonstrate a “good” understanding of these services, support systems, and policies. (80% score or higher on post-programming tests/surveys). Rationale: A primary role of SARC is to provide students with information required for their continued academic success. The Center will measure students’ understanding in order to gauge its effectiveness in information dissemination. Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. . Last Updated on: 12/11/07 - 79 - Updated by: Cyndi Rose Test a large cross-section of the Sacramento State student-athlete population during the Spring 2007 semester to determine the extent to which they understand campus academic requirements and NCAA eligibility rules. Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. After their Spring 2007 mandatory advising meeting with SARC staff, student athletes were required to take a Student Voice “quiz” via the web or PDA. The survey included four questions that identified their status as an athlete: sport, year in school, scholarship or non-scholarship and whether they started at Sacramento State as a freshman or transfer. The final six questions assessed a student-athlete’s knowledge of basic athletic eligibility and campus academic policies. 139 student-athletes (75 women and 64 men) were surveyed from 14 sports. Results Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. On the six direct-survey questions the correct response was selected accurately by the following percentage of student-athletes surveyed: (see survey in appendix A for specific questions) Question 5 – 81.34% of respondents answered correctly Question 6 – 77.61% of respondents answered correctly Question 7 – 81.34% of respondents answered correctly Question 8 – 97.01% of respondents answered correctly Question 9 – 78.36% of respondents answered correctly Question 10 – 87.31% of respondents answered correctly Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. The survey results confirmed that student-athletes are fairly knowledgeable regarding the eligibility and academic requirements. Our goal that 75% of our student-athletes are knowledgeable of NCAA and academic requirements was reached, as correct responses ranged from 77.6% to 97.0% on the six “knowledge” questions. The SARC staff believes that the mandatory advising sessions help student-athletes gain and retain this knowledge, therefore, we will continue this service. We also will continue to monitor knowledge gained via this service. Further comparisons of individual sports will determine whether additional rules education is necessary for specific sports. Questions regarding the programs and outcomes delineated in this section should be addressed to Paul Edwards, Student-Athlete Resource Center. (916) 278-7796. edwardsp@csus.edu. Student Health Center July, 2007 Mission Statement Note: Departmental mission must be directly aligned with those of the University and the Division. This statement should include approximately 3-5 sentences that identifies the name of the department, its primary functions, modes of delivery and target audience. Mission: The Student Health Center’s mission is to promote the health and wellness of Sacramento State Last Updated on: 12/11/07 - 80 - Updated by: Cyndi Rose students by providing quality medical services, prospective health programs, and educational activities in a highly competent, cost-effective, and compassionate manner. Rationale: The Sacramento State Student Health Center (SHC) develops and implements programs designed to create a healthy campus environment and to improve student health and wellness. Students' health has a great impact on their academic performance; therefore the SHC will strive to enhance the academic performance of students through accessible and high quality medical care, health prevention, and educational programs. Planning Goals Note: Planning Goals are broad statements that describe the overarching, long-range intentions of an administrative unit. Goals are used primarily for general planning, as the starting point for the development and refinement of program objectives or student learning outcomes. (UCF Administrative Handbook, University of Central Florida). Goal 1: Address some of the reproductive health needs of Sacramento State students. Program Objective or Student Learning Outcomes Note: The Objectives or Outcomes can be one of two types: program objectives or student learning outcomes. The former are related to program improvement around issues like timeliness, efficiency and participant satisfaction. The latter addresses what a student learns or how a student changes by participating in the program or utilizing the service. Both program objectives and student learning outcomes are measurable statements that provide evidence as to how well you are reaching your goals. Student Learning Outcome: Provide low income students at Sacramento State access to health education, counseling, and family planning services in an effort to: increase students’ knowledge regarding reproductive health; and possibly change their behavior regarding unprotected sexual activities (e.g. reduce STIs.) Possible Measures Note: Measures describe the methodology and timeframe for data collection. Measures also should identify the population being surveyed and/or tested. During the 2006-2007 academic year, the SHC will review its current data collection methods and revise them as needed relative to unintended pregnancies. Complete Quality Assurance study in Spring 2007 to determine the number of unintended pregnancies has been reduced from Fall 2006 to Spring 2007. ***This was not undertaken – due to our existing systems – we would not be able to definitely ascertain the number of unintended pregnancies due to SHC services or intervention. . Actual Measures Note: (If different from possible measures): Specify the data collection timeframe, population and methodology. Provide materials such as survey instruments, check lists, focus group protocols, etc. in an appendix. Quantitative – Gathered from daily patient encounter forms Qualitative – Data is collected from the following patient surveys: Health Education Sexual Health Program – fall 2006. Students voluntarily complete ten-item survey at the end of their Family PACT education visit. Point of Service – February 2007. Student voluntarily complete survey at the end of their health center visit during the first week of February. Results Note: Results include a brief narrative of findings, and/or essential tables or graphs. The results should indicate the extent to which the program objective or student learning outcome was met. Last Updated on: 12/11/07 - 81 - Updated by: Cyndi Rose Four surveys were conducted: (1) February 2006, (2) April 2006, (3) November 2006, and (4) February 2007. The four surveys yielded 352 responses to two questions regarding Family PACT which addressed learning behavior: A. Based upon the educational information you received during Family PACT, have you considered making any changes to improve your reproductive health? And B. Based upon the educational information you received during Family PACT, have you actually made any changes to improve your reproductive health? Results of these two questions indicated that: A. 30% of students responded with a “yes” to considering making changes, and B. 21% of students responded with a “yes” to having made changes to improving their reproductive health. As a follow up, students were asked to identify their intended changes or actual changes the most common answers were “having more health exams” and “using safer sex methods, e.g. condoms.” See APPENDIX A for additional information Conclusions Note: The conclusion should summarize briefly the collection and analyses of data. It also should “close the loop” by identifying what decisions and/or program modifications were made on the basis of these analyses. The outcome measures that were initially identified for the evaluation of the Family PACT program require more extensive research in order to yield the desired information (reduction of unintended pregnancies). Therefore, we will re-examine our choice of outcome measures to evaluate the program’s impact on student behavior and learning outcomes. Program Objective: By 2010 develop a multi-dimensional wellness program that assists students in making healthy lifestyle choices. This multi-dimensional wellness program requires planning and development with contributions from various campus constituents. To begin this work the SHC will develop, by May 2008, a formal document that delineates the program’s structure, implementation, processes and evaluation. Possible Measures Interim progress reports will be completed. The first interim progress report was submitted in June 2006. Complete the multi-dimensional wellness model by June 2008. Wellness Program implemented by Fall 2010. Actual Measures Create and implement an assessment plan for Fit HELP (formerly the Eating and Healthy Weight Program) as part of the first phase of the Wellness Program development. Results Please see Appendix B. Conclusions Last Updated on: 12/11/07 - 82 - Updated by: Cyndi Rose The initial phase of program development has been very successful, but the objective may need to be more focused - it’s far too broad in its current form. The Student Health Center will develop a more concise objective and corollary measures for the 2007-08 academic year. Program Objective: At the conclusion of the spring 2007 semester, 90% of Student Health Center medical providers will follow the Centers for Disease Control guidelines for evaluating and treating acute respiratory disease. Viruses are responsible for most cases of acute pharyngitis, and in 2001 the CDC developed evidence-based guidelines for treating adults with acute respiratory disease. According to SHC utilization data, upper respiratory infection is the number one reason students seek campus medical care services, and as such is a condition that may significantly impact a student’s health status. Possible Measures Conduct a chart review of 100 student medical records coded for pharyngitis/tonsillitis. Each record will be evaluated to determine if the Centers for Disease Control guidelines are being followed relative to the use of appropriate diagnostic criteria and antibiotic therapy at the SHC. A written quality assurance study will be completed in spring 2007 by one of the SHC staff physicians documenting the chart review and its subsequent findings. The study will include comparison of findings from the 2006-2007 chart review with previous review of 2005-2006. Actual Measures Review of 100 student medical records coded for pharyngitis/tonsillitis Results Forty-seven of the 100 patients evaluated for acute pharyngitis were prescribed antibiotics. Of the 31 patients seen by the providers hired in the last 2 years, 29 (93.5%) were treated, whereas, only 18 of 69 patients (26%) were treated by providers who were hired more than 3 years ago. During the chart review, another trend was discovered for one of the new providers – the vast majority (94%) of all pharyngitis patients evaluated by this provider was noted to have three Center criteria, and as such were treated with antibiotics. Though most of these patients had normal temperatures in the clinic, each of them stated they had a history of fever. This self-reported undocumented fever apparently was used as one of the criteria for treatment. This follow-up study had interesting and important findings. The medical providers who have worked at the Student Health Center for more than 3 years have shown a major improvement in adherence to pharyngitis treatment guidelines. This has resulted in a large decline in inappropriate use of antibiotic therapy. The recently hired providers, who have not had as much exposure to the previous pharyngitis studies and subsequent education, show an extremely high rate of unnecessary antibiotic prescribing. Perhaps one reason is that new providers utilized streptococcal testing far less frequently to confirm the presence of bacterial infection, leading to inappropriate antimicrobial prescriptions for the majority of patients they evaluated for pharyngitis. Another reason is that the criteria associated with risk of streptococcal infection were not used consistently for clinical screening and decision making. It also appeared that an adequate history may not have been obtained for some patients or may have been too loosely interpreted. It is not likely that a fever was present in more than 94 per cent of those seen for sore throat. If a patient did not measure their temperature with a thermometer, a self-reported “fever” may not be correct and thus not a true criterion. Last Updated on: 12/11/07 - 83 - Updated by: Cyndi Rose Conclusions In this era of antibiotic resistance, with subsequent emphasis on appropriate use of antimicrobials, it is important that all medical providers are educated about this practice soon after hiring. Continued monitoring and education regarding adherence to clinical standards is important to ensure quality of care. Also, in the future, more effort should be made to encourage testing for ambiguous cases. It appears that utilizing the second option discussed in CDC guidelines would be helpful, i.e. testing patients with two or three Center criteria and limiting treatment to those with a positive test or patients with four criteria. It is anticipated that these measures will increase and maintain a rational approach to therapy of this common medical condition. The results of the study were shared at a medical provider meeting in Fall 2006. The newly hired medical providers were individually counseled regarding the CDC guidelines and the study’s findings, and they felt that the study findings were due to their coding interpretation, i.e. they only coded treated patients as pharyngitis/tonsillitis. These medical providers plan to modify their coding process to match that of the rest of the SHC medical provider group in the future. Questions regarding the programs and outcomes delineated in this section should be addressed to Laurie Bisset Grady, Student Health Center, (916) 278-5422, lbgrady@csus.edu. Last Updated on: 12/11/07 - 84 - Updated by: Cyndi Rose APPENDIX A. Last Updated on: 12/11/07 - 85 - Updated by: Cyndi Rose APPENDIX B. Fit HELP (Fitness, Healthy Eating and Lifestyle Program) As of May 2007 Mission Statement Note: Mission represents an expanded statement of purpose. Your unit’s mission statement must be consistent with the University's mission statement. You should be able to see how the mission statement of your unit complements the identified mission of the University. Fit HELP advances student wellness through promotion of healthy lifestyles and the removal of health impediments to academic success. Planning Goals Note: Goals should represent a desired accomplishment that deals with intended outcomes, or objectives. If achieved, the goal will advance the unit toward an outcome. The outcome should be challenging, attainable, measurable numerically and accomplished within a given period of time. Outcomes may state what students will accomplish, and/or what the unit will accomplish. Goal 1: Increase students’ knowledge of wellness and preventative health concepts. Goal 2: Increase students’ proactive health behaviors and reduce risk factors for disease. Goal 3: Provide appropriate treatment related to nutrition, fitness, and associated mental health issues. Objectives Note: Objectives are statements that describe the desired quality of key services within the administrative unit. It defines what the service should promote. Objectives can also be stated in terms of student learning outcomes. This is most appropriate in services that aim to increase students’ knowledge or understanding of specific concepts. (UCF Administrative Handbook, University of Central Florida). Objective 1: By Fall 2007, initiate analysis of client profiles and utilization patterns to assist with program development, marketing and resource allocation. Rationale: As a new initiative under the auspices of the Student Health Center, Fit HELP must assess client profiles and utilization patterns in order to continuously improve service to current users, as well as identify underserved populations. This first year of assessment will help to establish baseline data to guide the program. Possible Measures: Assessment of client intake data form User/Non-user focus groups Feedback survey of past users Objective 2: By Fall 2007, develop a comprehensive marketing plan for Fit HELP to increase campus community awareness; referrals to the program by staff; and utilization by students. Rationale: During the inaugural year of the Eating and Healthy Weight Program there were positive efforts made to raise awareness of the program and its services through the use of both paid and earned media. With the establishment of a new, more descriptive name, the program must again pursue cost-effective strategies to raise awareness and market itself to faculty, staff and a new student cohort. Possible Measures: Number of students receiving information as a part of New Student Orientation Intake reporting in response to “How did you hear about the program?” Number of faculty teaching Freshman Seminar and Learning Community courses requesting classroom presentations Number of referrals made by Student Health Center staff Objective 3: By Spring 2008, develop a tool to assess students’ utilization of Fit HELP tool (“the book”) and level of follow-through with lifestyle plan. Last Updated on: 12/11/07 - 86 - Updated by: Cyndi Rose Rationale: Fit HELP has developed a new, comprehensive tool (“the book”) for use with students seeking positive lifestyle change. During the next year, staff will develop a tool to gather user feedback regarding the usefulness of “the book” and their level of utilization. This information will then be used to revise the document to ensure maximum value and use. Possible Measures: Successful development of a tool to measure utilization and usefulness of “the book” Initial implementation of the assessment tool Objective 4: At the end of spring semester 2008, when asked for follow-up feedback, 70% of students will report positive satisfaction related to treatment services received. Rationale: Satisfaction is how well an experience meets expectations. Student satisfaction with their experience with Fit HELP will be based on both qualitative (did they find the information or treatment received helpful or appropriate) and quantitative measures (did they achieve their weight management or fitness goals). Possible Measures: Development of a patient satisfaction assessment tool Initial implementation of the patient satisfaction assessment tool Questions regarding the programs and outcomes delineated in this section should be addressed to Joyce Harris, Student Health Center. (916) 278-6461. harrisj@csus.edu. Testing Center As of October 12, 2006 Mission Statement Mission: The Testing Center provides direction and support for meaningful student assessment and promotes awareness of ethics in testing through the administration of CSU system-wide tests and national standardized tests, as well as providing testing accommodations for students with special needs. Assessment is the primary measurement for student learning, thereby making the Testing Center an integral part of the University and an essential element to the academic success of our students. Planning Goals To provide direction and support for meaningful student assessment. To promote awareness of ethics in testing. Planning Objectives Objective 1: All students who take the EPT/ELM Test Taker Survey will rate “test dates offered” and “access to the on-line EPT/ELM Information Bulletin” “Good” or “Excellent”. Rationale: Many factors related to testing are managed by testing companies and are beyond the control of the Center. However, the testing environment, convenience of testing, and knowledgeable support to students are elements that are within the scope of the Center’s influence. It is important to the Center to Last Updated on: 12/11/07 - 87 - Updated by: Cyndi Rose ensure that we serve students in the most effective and efficient manner possible and develop and maintain services that meet the needs of our students and community. Possible Measures: The EPT/ELM Test Taker Survey will be used to measure accomplishment of this objective during the 2007 spring semester. Objective 2: Through improved signage there will be a reduction of students arriving late for the Saturday Testing Programs. Rationale: Many students arrive late to the Saturday Testing Program because of problems such as signage, direction, and parking. Improved procedures will enable more students to arrive at the appropriate time. Possible Measures: The test administrator will record the number of late arrivals, and conduct a point of service interview to determine their reasons so that pre and post-signage comparisons can be made between the Fall 2006 semester without signs and the Spring 2007semester with signs. Objective 3: Develop computer-based training program for proctors, readers, and scribes. Rationale: A computer-based training program would substantially enhance the quality of test administration and streamline the training process. Currently it is impossible to train proctors as a group because of their conflicting class schedules and the short time between hiring and beginning employment. It is not only time consuming to train each of the 10 to 20 proctors individually, the quality of individual training is not standardized. A computer-based training program would solve this problem and improve the quality of test administration by fall 2007. Possible Measures: Developing test questions about testing protocol with proctors achieving 80% or higher. Monitoring proctors throughout the year to determine if the quality of work improved compared to those past proctors who did not go through the computer training. Documenting time saved by the new process will also measure the accomplishment of this objective. Questions regarding the programs and outcomes delineated in this section should be addressed to Sheila Deaner, Testing Center. (916) 278-6296. deaner@csus.edu. University Union As of August 29, 2006 Mission Statement Note: Mission represents an expanded statement of purpose. Your unit’s mission statement must be consistent with the University's mission statement. You should be able to see how the mission statement of your unit complements the identified mission of the University. Mission: The University Union exists for the benefit of students, offering a welcoming environment where students, faculty, staff, alumni and the greater community participate in campus life. The programs, services and facilities of the Union foster personal growth, encourage social interaction, and develop leadership skills. This involvement leads to memorable experiences and builds a community that cultivates enduring commitment, pride and loyalty to the University. Planning Goals Note: Goals are broad statements that describe the overarching long-range intended outcomes of an administrative unit. These goals are usually not measurable and need to be further developed as separate distinguishable outcomes, that when measured appropriately, provide evidence of how well you are accomplishing your goals. They are primarily used for general planning and are used as the starting point to the development and refinement of outcomes. Last Updated on: 12/11/07 - 88 - Updated by: Cyndi Rose Train, develop and encourage a highly committed and conscientious student staff to provide exceptional, professional, customer-oriented service which is sensitive and responsive to the needs of our multicultural community. Enhance students’ (volunteers, interns, and student assistants) interpersonal, leadership and critical thinking skills; develop nurturing and supportive networks; help prepare them for active citizenship beyond the collegiate experience. Increase the use of the University Union’s programs, services and facilities to enhance continual growth through superior programs, activities and creative marketing in order to advance the mission of the Union, Student Affairs, and the University. Planning Objectives Note: Objectives are specific statements that describe desired outcomes identified by an administrative unit. They are derived from the goal statements of the unit. Objectives may relate to the operations and processes of the unit and may also relate to intended behaviors that a student having used the services provided by the unit should demonstrate. Objective 1: Offer a welcoming environment to students Note: Specific assessment tools and targets for each area are given in the Measures section below Rationale: The University Union is the centerpiece and the campus home for students, faculty, staff and alumni of California State University, Sacramento. Additionally, the University Union is often the only contact someone from off campus has with the University, through attendance at an event, meeting or conference. We believe the performance of the people of the Union are at the core of service to the campus, and we MUST provide customer focused service and be a leader in responding to the ever changing needs of our diverse campus. Possible Measures: Continue the use of customer satisfaction surveys to indicate the level of satisfaction with the Union and our efforts to exceed expectations in the following areas: A. Food Services B. Event Services Continue the use of customer satisfaction surveys to indicate the level of satisfaction with the Union and our efforts to exceed expectations in the following areas: A. 80% of respondents to Food Services Feedback Cards rate the University Union and University enterprises Inc as good quality, Value, good customer service, clean facility. i. Tools: a. Food Service Feedback Cards b. EBI Benchmarking Survey Questions 28 through 36. (Variety, food prices, quality) B. 85% of the respondents to Event Services Feedback cards will rate the University Union as i. Clean: a. Event Services Feedback Cards b. EBI Benchmarking Survey :Questions #46-#48 C. On the EBI benchmarking study, the Union is rated as 5.0 or better on a seven point scale on the following areas: i. Factor 3: the Union is Student oriented: Questions #11-14 ii. Factor 2: college union has a positive environment: Questions #7 through #10 D. EBI Benchmarking survey: Overall analysis Objective 2: Foster student personal growth and develop student leadership skills. Note: Specific assessment tools and targets for each area are given in the Measures section below. Rationale: Students’ utilization of the Union as a community member, an event attendee, a customer, a program volunteer or as an employee are critical opportunities for students to learn about themselves, their Last Updated on: 12/11/07 - 89 - Updated by: Cyndi Rose campus community, the surrounding community, and the world they will be entering when they leave Sac State. We want to identify how employment and volunteering at the Union impacts a student personal development and prepares them for life beyond the University. Possible Measures: Establish and implement a program of measuring student learning at the University Union: Use pre and post tests, and/or exit interviews, to measure customer service, leadership, and professionalism skills development in the following areas: A. Student Assistants at the University Union : 1. Measures: a. Student assistants will score 100% on all skill check sheets prior to being assigned a work shift. b. 85% of student assistants will demonstrate personal growth during their experiences at the university Union (as measured by pre and post tests administered at fall staff orientation and either at Exit interview or end of spring Semester.) 2. Student Areas: a. Event Services Set Up: 1. Skill Check sheet 2. Pre Test and Post Test b. Maintenance Services: 1. Skill Check sheet 2. Pre Test and Post Test c. Program Services: 1. Skill Check sheet 2. Pre Test and Post Test d. Graphics interns: 1. Skill Check sheet 2. Pre Test and Post Test B. Student Volunteers: 1. 85% of Student Volunteers will demonstrate personal growth during their experiences at the University Union. 2. Measure: Pre and Post Test 3. Student Areas: a. Student volunteers: Board Members b. Student volunteers: UNIQUE Objective 3: Increase opportunities for social interaction among students through increased utilization of the University Union facilities and services. Note: Specific assessment tools and targets for each area are given in the Measures section below. Rationale: The University Union is supported primarily through student fee dollars; consequently we are obligated to reach out to the campus community and encourage use of the facilities, programs and services. The Union intends to validate our effectiveness in reaching the community by measuring the success of our programs and services through attendance, growth and observation of trends. Possible Measures: Create new measures for customer satisfaction regarding programs offered by the Union. A. B. C. Last Updated on: 12/11/07 Track attendance at Union sponsored events. Track door counts. Track Web site utilization: OPUS Union - 90 - Updated by: Cyndi Rose D. Campus Events Calendar Satisfaction with Union services will be assessed Utilize the data from the Spring 2006 benchmarking study to address areas that could be improved. Questions regarding the programs and outcomes delineated in this section should be addressed to Leslie Davis, University Union, (916) 278-6744, leslied@csus.edu. Last Updated on: 12/11/07 - 91 - Updated by: Cyndi Rose