STUDENT LEARNING OUTCOMES (SLOS) ASSESSMENT PLAN WORKSHOP Conducted by Peggy Liggit and Lisa Klopfer 109 Bruce T. Halle Library Building Tuesday, September 14, 1-2:30 PM Friday, September 17, 10:30 AM – 12:00 PM Monday, September 20, 9 – 10:30 AM Tuesday, September 21, 2:30 – 4:00 PM Monday, September 27, 1 – 2:30 PM Friday, October 1, 1 – 2:30 PM Wednesday, October 6, 2:30 – 4:00 PM Peggy Liggit Peggy.Liggit@emich.edu 734-487-0199 Director of Academic Assessment (I) Office of Institutional Effectiveness, 234 McKenny EMU: Office of Institutional Effectiveness and Accountability Thursday, October 7, 9 – 10:30 AM Monday, October 11, 10:30 AM – 12:00 PM Thursday, October 14, 1 – 2:30 PM Tuesday, October 19, 9 – 10:30 AM Wednesday, October 20, 3 – 4:30 PM Wednesday, October 27, 10:30 AM – 12:00 PM Thursday, October 28, 3 – 4:30 PM Lisa Klopfer Lisa.Klopfer@emich.edu 734-487-0020 ext. 2114 Director, Bruce K. Nelson Faculty Development Center (I) 109 Bruce T. Halle Library Building Page 1 Table of Contents Workshop Goals ............................................................................................................................................................................................................ 3 Project Timeline ............................................................................................................................................................................................................ 4 Notes and Encouragement from Peggy ........................................................................................................................................................................ 5 Overview of Embedded Assessment ............................................................................................................................................................................ 6 Plan for Program-Level Assessment of Student Learning ............................................................................................................................................ 8 Methodology Plan Templates ....................................................................................................................................................................................... 9 Self-Evaluation Questions -- Program-Level Assessment Plans for Student Learning .............................................................................................. 13 Submitting Your Program-Level Assessment ............................................................................................................................................................. 17 Electronic Resources ................................................................................................................................................................................................... 22 Appendix: Supplementary Material ........................................................................................................................................................................... 23 EMU: Office of Institutional Effectiveness and Accountability Page 2 WORKSHOP GOALS Welcome to a new academic year. Let’s remember that the purpose of this project is to share with others your genuine interest and intention in evaluating student performance as a means to improve the learning experience for the students enrolled in the programs associated with your department or school. The workshop facilitators and their office teams are here to support your efforts, particularly: forming and implementing plans for assessing student learning, and successfully reporting assessment plans and activities within the Academic Review and Continuous Improvement process for 2010-2011. During this workshop you will practice: comparing and contrasting multiple methods for writing effective program-level assessment plans, practice writing student learning outcomes (or revising outcomes that are already written), curriculum mapping: that is, mapping program outcomes to learning experiences conducted throughout the curriculum, and writing a methodology plan that describes how the academic program is going to improve student learning for at least one outcome in your program. Additional activities that will be covered: Describe how embedded assessment can be used to identify student misconceptions. Explain that when the embedded assessment process is written as a narrative (the Teach, Assess, Analyze, and Adjust steps), this narrative can serve as one example of an accountability record for documenting how you are improving student learning in the program area. IMPORTANT NOTE: EMU: Office of Institutional Effectiveness and Accountability This methodology plan is due by October 30, 2010. Page 3 PROJECT TIMELINE March 1 - May 2, 2011 February 11 and February 18, 2011 January 14, 2011 November 12 and December 3rd, 2010 October 30, 2010 Assessment plans due for Departments/ Internal Review Retreats Assessment plas for Departments/Schools completing Full Review due for submission to the College Deans Roundtable Discussions (Scheduled on Fridays, due to limited space) Internal Review Retreats Deans' comments submitted 1/31/11 Campus Comment Opened 2/1/11 Schools completing Annual Planning EMU: Office of Institutional Effectiveness and Accountability Page 4 NOTES AND ENCOURAGEMENT FROM PEGGY “The connections made by good teachers are held not in their methods but in their hearts – meaning heart in its ancient sense, as a place where intellect and emotion and spirit and will converge in the human self.” --Parker Palmer, Courage to Teach. After analyzing over 100 hours of faculty interviews in the last year, it is clear to me that faculty at our institution teach with heart. “You should never worry about your good ideas being stolen in educational reform, because even when people are sincerely motivated to learn from you, they have a devil of a time doing so.” --Michael Fullan, Change Forces: The Sequel Why There are Few ‘Good’ Assessment models: o Michael Fullan (1999, 2006) - Innovative ideas are “difficult to disseminate and replicate.” Not easily transferable - difficult to capture “subtleties of the reform practice.” The inability to replicate another’s model, “replicating the wrong thing” “The reform itself, instead of the conditions which spawned the success.” Problems with scale: small -scale may not work well on a wider-scale. After returning from the Higher Learning Commission Conference and reading many articles in the current literature about assessment, there are no assessment models/templates that one can simple take and fill-in to meet the needs of an individual program. The best plan is to look at several examples of assessment models/templates and modify/revise what can work in your own discipline. I suggest beginning this work with what you already know and practice … EMU: Office of Institutional Effectiveness and Accountability (continue to the next page) Page 5 EMBEDDED ASSESSMENT Take a Moment Here: Reflect and write about a misconception/problem in student learning that you did something about. What/how were you teaching at the time? What activity were you doing to determine student’s thinking/reasoning? What did students understand/not understand about what you were teaching? What did you do to help students better understand? You have just documented a, presumably, unconscious competent activity. Teach, Assess, Analyze, and Adjust = Embedded Assessment EMU: Office of Institutional Effectiveness and Accountability Page 6 EMU: Office of Institutional Effectiveness and Accountability Page 7 PLAN FOR PROGRAM-LEVEL ASSESSMENT OF STUDENT LEARNING* 1. Generate Program-Level Student Learning Outcomes: At the end of the program, students will be able to <<action verb>> <<something>>. 2. Identify Learning Activities Associated with Program Outcomes: Describe what student performance of an outcome looks like. Identify learning activities, or create them if none exist, to provide opportunities for students to move closer toward mastery of the program’s outcomes. How well a student performs on a particular outcome can be demonstrated through a learning object (or artifact). For example, exam questions, projects, research papers, presentations, portfolios, exhibitions directly offer evidence regarding students’ performance in relation to a particular outcome. Some type of “assessment instrument” to evaluate how well students master the learning activity accompanies the learning object, such as, a scoring guideline, rubric, or answer key. 3. Curriculum Mapping: Identify courses spanning the curriculum that provide learning opportunities for students to achieve program outcomes. Students do not master an outcome with one or two experiences; they need multiple opportunities from the beginning to the end of their program, with increasing difficulty and challenge as they developmentally advance. 4. Capture Student Performance of Outcomes: Collect information about student work from learning objects (artifacts) associated with particular program outcomes, so you can learn about students’ progress with those outcomes. How you learn about student performance depends on the kinds of artifacts you collect. For instance, if you test students on important concepts, you might look at their test scores. Alternatively, if you ask students to create a performance, write a report, etc. , you might use those. Whatever you choose, make sure that: a) It represents learning outcomes that are important for the your discipline; b) You have a way to assess the artifacts that you collect. If you’re looking at test scores, you have the scores themselves. However, if you’re looking at something like a paper, portfolio, or performance, you will need to create a method to analyze these artifacts. EMU: Office of Institutional Effectiveness and Accountability It is not necessary to collect data in every course every semester. For each program outcome, it is reasonable to collect 3-5 scores over the number of years a student is in the program – documenting that students have multiple opportunities that bring them closer to “mastering” each outcome. 5. Interpret Data: Analyze trends and patterns in student performance to determine to what extent students are achieving program outcomes. What do you see that could be improved upon with regard to students learning these outcomes? 6. Take Action to Improve Your Program: If students are not meeting your program outcomes to the extent or manner in which you have intended, what changes in the program need to be made? Implement an improvement plan, such actions might include, but are not limited to: improving advising, implementing prerequisites, changing the curriculum, mentoring faculty for improving instructional delivery or revising learning activities and objects, acquiring different or updated program resources and technologies, creating community partnerships. Repeat steps 4 - 6 as necessary. 7. Share Your Progress with Program Stakeholders: Our internal and external communities want to know to what extent students are learning and mastering the content, skills, and attitudes of the disciplines that constitute your programs. 8. Revisit, Revise, Repeat steps of this process as often as necessary to improve student learning. *Although this looks like a clean and regimented process written here as text, it is really a much more fluid and, at times, even “messy” process. Within this framework, every program must create an assessment system that works for them, and it is understood this system will look a little different from program to program. To create such a system requires meaningful dialog and collaboration, identification and prioritization of our educational values, and the understanding and patience that this is a dynamic and human process. Page 8 TEMPLATES TO HELP YOU GET STARTED WITH YOUR METHODOLOGY PLAN (Remember, these are only graphic organizers to help you with your thinking/planning. As we know, models rarely transfer from one program to the next, from one department to the next, from one institution to the next. Take-in what information helps, leave behind what doesn’t and create your own model and plan). 1. List of Program-Level Student Learning Outcomes: When writing these outcomes consider the following: – Write outcomes that are measurable and specify definite, observable behaviors. (Remember, data will be collected on these outcomes). • – Students will <<action verb>>< <something>>. Program level outcomes should also be written to communicate mastery and the highest levels learning (see Bloom’s Taxonomy). What skills, knowledge, or behaviors do you want students to master at the end of the program? Outcome 1 – Outcome 2 – Outcome 3 – 2. List of assessments, also known as learning objects, from course or program assignments or exams used to evaluate student learning outcomes. Outcome 1 Assessment/learning object Outcome 2 Assessment/learning object EMU: Office of Institutional Effectiveness and Accountability Page 9 3. Alignment of course and program assessments with student learning outcomes – Curriculum Map List Of Student Learning Outcomes Students can: 1. 2. 3. 4. 5. 6. 7. Student Learning Outcomes 1 Courses/ Assignments 2 3 4 5 6 7 EMU: Office of Institutional Effectiveness and Accountability Page 10 4. Design a methodology plan to show evidence that students are meeting the outcomes (What data will be collected, how often will it be collected, how will it be summarized and analyzed?) Suggestion: Design a small pilot plan you can implement and manage throughout 2010/2011. Start with one or two outcomes, assess these in a beginning, mid-program, and capstone classes, ask questions about student learning – What do students struggle with and what are you testing to see what is helping this? Collect data to answer your questions. What worked, what will you do next? 5. Use of assessment results to improve the program: In this section, present the evidence that assessment results have been analyzed and been (or will be used) to improve student performance and strengthen the program. This description should not link improvements to individual assessments but, rather it should summarize principal findings from the evidence, the faculty’s interpretations of those findings, and changes made in (or planned for) the program as a result. Describe the steps program faculty has taken to use information from assessment for improvement of both student performance and the program. EMU: Office of Institutional Effectiveness and Accountability Page 11 EMBEDDED ASSESSMENT METHODOLOGY FOR PROGRAM-LEVEL ASSESSMENT (ANNUAL PLANNING) Teach/Implement - (What activity/concept/skill are you going to be teaching/implementing?) Assess – (What activity/exam/project are you going to have students do to determine their thinking/reasoning/performance?) Analyze – (What are you going to be looking for to understand what student know and don’t know about the concept/skill/task/idea/reasoning they are supposed to be learning?) Adjust – (What did you do to improve student learning based on your findings? EMU: Office of Institutional Effectiveness and Accountability Page 12 SELF-EVALUATION QUESTIONS – PROGRAM-LEVEL ASSESSMENT PLANS FOR STUDENT LEARNING ORGANIZATION AND FORMAT EFFECTIVE ASSESSMENT PLANS SOMEWHAT EFFECTIVE ASSESSMENT PLANS INEFFECTIVE ASSESSMENT PLANS Who is responsible for the assessment plan in our program? Assessment is collective effort involving both tenured and non-tenured faculty. Clear direction and leadership by a “point person” or small group of persons. Large portion of faculty participate in assessment process, although some are not yet included. Direction and leadership for the process may be divided among too many individuals or groups to be effective in directing assessment. Little to no participation from faculty. No directing “point person.” Rules and responsibilities are not clearly defined. Do we use a common language in our assessment plan? Program faculty have agreed on common terminology. Program’s terminology coincides with that of the University as much as possible. For the most part, there is agreement on terminology among faculty, but inconsistencies remain. Little agreement on terminology with the University as a whole. How many Student Learning Outcomes are included in our assessment program? Short list, of appropriate length for your program, of clearly defined program-level student learning outcomes. List is of manageable size, but not all SLOs are necessary (e.g. some similar SLOs can be rephrased and combined into a single outcome, thus increasing the efficiency of the assessment process). For similar programs that are grouped together for convenience of reporting: Did we include at least one unique outcome to distinguish on program from another? At least one and, ideally, more than one unique outcome has been included for each program Some programs have unique learning outcomes, but not all in group Essentially no agreement on terminology at any institutional level; numerous different terminologies and standards for measure employed. Excessive complexity of outcomes. “Laundry list” of numerous program outcomes. Numerous SLOs overlap with each other, and are often redundant. No unique outcomes to distinguish programs; all programs covered by general set of outcomes EMU: Office of Institutional Effectiveness and Accountability Page 13 ORGANIZATION AND FORMAT EFFECTIVE ASSESSMENT PLANS SOMEWHAT EFFECTIVE ASSESSMENT PLANS INEFFECTIVE ASSESSMENT PLANS Are our Student Learning Outcomes written in an appropriate format? SLOs address issues that are concise and measurable through the format: “Students will be able to <<action verb>> <<something>>.” For example: “Students will be able to design solutions based upon customer requirements.” SLOs assess advanced skills of students graduating from the program. Program faculty understand and agree upon the definition of the written outcome. Complexity and measurability of outcomes is somewhat inconsistent: e.g. some are concise and clearly measurable, but others are excessively length and/or lack empirical standards of measurement. Some SLOs do not address the most advanced skills of students graduating from the program Although there is consensus on the majority of outcomes, there is some degree of disagreement amongst faculty over the definitions of some written outcomes. SLOs are excessively complex. Issues are not measurable, difficult or impossible to assess. For example: “Students will be able to identify, define, and analyze the major causes, effects, and implications of $150-a-barrell oil prices on the transportation, food, and housing industries.” SLOs assess basic understandings that develop early in the curriculum. Little to no agreement upon definition of written outcomes amongst program faculty. Are our outcomes supported by core courses? Core curriculum courses designed to ensure students are given opportunity to develop competences in program-level SLOs. Faculty have systematically examined curriculum, and created a “curriculum map” to identify inadequately supported outcomes, areas of overlap, or overlooked outcomes. Some outcomes are not adequately supported by the program’s core courses, but faculty are aware of these blind spots thanks to curriculum mapping. Some or all of the SLOs are unsupported by the program’s core curriculum courses. No “curriculum map” has been created; little awareness of relation between core curriculum and SLOs. EMU: Office of Institutional Effectiveness and Accountability Page 14 DATA COLLECTION, METHODOLOGY, AND ANALYSIS Does our assessment plan rely on direct measures of student learning? How do we assess, collect, and organize data? What do we do with our data after it has been collected? EFFECTIVE ASSESSMENT PLANS SOMEWHAT EFFECTIVE ASSESSMENT PLANS INEFFECTIVE ASSESSMENT PLANS Assessment based upon direct examination of work completed by students, either formatively or summatively. Possible vehicles for the direct examination of student work include essays, problem sets, exams, etc. Work itself directly tied to one or more SLO. Assessment methods are chosen to address the specific characteristics of the program . Assessment plans implemented in systematic, ongoing manner. Faculty regularly meet and exchange student work for assessment. The majority of assessment is based upon direct examination of student work, although conjecture and opinion surveys are sometimes used to shore up the gaps. Assessment depends upon indirect measures of student outcomes, such surveys asking students how much they have learned, or extrapolation from student satisfaction data. Data collection, assessment, and analysis utilizes a hybrid of specific and nonspecific methods. At least some degree of planning prior to undertaking assessment. Faculty meetings for the exchange of student work do take place, although perhaps they do not include a large enough portion of the faculty, or do not occur with enough regularity. Data is retained and studied over time to identify different patterns of evidence. These patterns include “patterns of consistency” (developed by studying data from the same outcome over a period of time), “patterns of consensus” (developed by disaggregating data to compare the achievement levels. Either data is retained over time but not enough is done to analyze and act on the information, or data is analyzed upon collection but not retained for comparison with future data for the purpose of identifying patterns. Data collected and assessed primarily from nonspecific sources, such as standardized tests. Assessment plans implemented “at the last minute,” shortly before accreditation or reaccreditation visit. Little to no inter-faculty exchange of student work. Little to no action taken after data collection; data stored and forgotten. Little to no analysis of data to identify trends over time. EMU: Office of Institutional Effectiveness and Accountability Page 15 DATA COLLECTION, METHODOLOGY, AND ANALYSIS How have we used our collected data to improve learning and teaching? EFFECTIVE ASSESSMENT PLANS Faculty, staff, and student development activities – such as workshops, presentations, and discussions – have been organized to insure that everyone in the program fully understands the program assessment plan. When necessary and appropriate, program policies and procedures may be revised to take into account the findings of assessment (e.g. revising the criteria for admission to the program, etc.). Program curriculum is revised to take into account any gaps or deficiencies that may be have become apparent during assessment. EMU: Office of Institutional Effectiveness and Accountability SOMEWHAT EFFECTIVE ASSESSMENT PLANS Some development activities have been prepared, but they may not include a large enough portion of the program’s faculty, staff, and student populations. Some program policies and procedures have been revised. Program curriculum has been revised, but some gaps may remain unaddressed. INEFFECTIVE ASSESSMENT PLANS There has been little to no preparation of development activities for the program’s faculty, students, and staff. Little to no revision of program policies and procedures to take into account findings of assessment. Numerous gaps in the program’s curriculum remain largely unaddressed. Page 16 SUBMITTING YOUR PROGRAM-LEVEL ASSESSMENT PLAN my.emich REPORTS In my.emich, go to the “Reports” tab. Under “IRIM Reports,” select “Academic Review and Continuous Improvement.” After entering your my.emich ID and password, you will be taken to the Modules page. MODULES PAGE Select the appropriate Department or School from the drop-down menu at the top of the page, then click the “Switch Dept/Schl” button. If your department/school is in Full Review, click on “Review Module.” If your department/school is in Annual Planning, click on “Planning Module.” EMU: Office of Institutional Effectiveness and Accountability Page 17 SUBMITTING PLANS IN FULL REVIEW DEPT./SCHOOL INFORMATION Once you have clicked on the Review Module, you will be taken to the Department/School Information page. From there, you can use the link on the left-hand sidebar to navigate to the Program Review Page. Be sure that the box at the bottom of this first page – the one that acknowledges that the list of programs is correct – is checked; otherwise you will not be able move forward. Make sure to click “Save” after checking the box. PROGRAM REVIEW Use the “Save,” “Save and Continue,” and “Continue” buttons at the bottom of each page to navigate your way through the various pages. Navigating away from a page without clicking “Save” or “Save and Continue” will erase any information you have entered. EMU: Office of Institutional Effectiveness and Accountability Page 18 Selecting “Program Review” will first take you to the “Program Overview” section. From there, you can use the left-hand sidebar to navigate (make sure to save each time you navigate to a new page!) through the different sections of Program Review. “Summary of Past Activities/Improvement Status” takes you to the page where you can enter information on the status of past program improvement efforts. “Assurance of Quality (Assessment of Student Learning Outcomes)” takes you to the page where you can submit information on your program-level Student Learning Outcomes. Always remember to Save! EMU: Office of Institutional Effectiveness and Accountability Page 19 SUBMITTING PLANS IN ANNUAL PLANNING After selecting the “Planning Module” you will be prompted to select the year for planning. Select the current year in the annual planning cycle and click “Continue.” This will take you to the “Department/School Information” Page. Click “Program Planning” on the left-hand sidebar and select the appropriate program from the list. EMU: Office of Institutional Effectiveness and Accountability Page 20 This will take you to the page where you can submit your program’s Annual Planning information on Student Learning Outcomes. Before navigating away from this page, or any others, be sure to click either “Save” or “Save and Continue” at the bottom of the page. EMU: Office of Institutional Effectiveness and Accountability Page 21 ELECTRONIC RESOURCES Electronic copies of the resources used in this workbook (listed in the bibliography), along with further information on the assessment process, including articles, sample assessment plans, and links to helpful outside resources can be found on the Student Learning Outcomes Assessment site on eFolio. To access the site, follow these steps: 1. Go to http://tinyurl.com/emuassess This will take you to the main page for Student Learning Outcomes Assessment, where you can find guides and resources to help you with assessment, as well as a schedule of Assessment Plan Workshops. 2. On the left-hand sidebar, you will see several links underneath “Student Learning Outcomes Assessment.” “Resource Library” – this includes links to electronic copies to articles/resources that were used in the development of this workbook, and that may be helpful to you during your assessment process. “Samples” – here you can find sample assessment plans from programs that have previously undergone assessment. These illustrate a variety of different approaches to the assessment process. “Finding and Fixing Misconceptions” – this section houses resources and a bibliography from a different workshop, but feel free to look! There are some interesting articles. 3. In order to access these pages, you may be prompted to login. If so, use the following information: Username: SLO Password: SLO Be aware, both the username and password are case sensitive. EMU: Office of Institutional Effectiveness and Accountability Page 22 APPENDIX: SUPPLEMENTARY MATERIAL Bloom’s Taxonomy “Assessing Your Program-Level Assessment Plan,” Susan Hatfield EMU: Office of Institutional Effectiveness and Accountability Page 23 COMPREHENSION APPLICATION KNOWLEDGE Cite Count Define Draw Identify List Name Point Quote Read Recite Record Repeat Select State Tabulate Tell Trace Underline Associate Classify Compare Compute Contrast Differentiate Discuss Distinguish Estimate Explain Express Extrapolate Interpolate Locate Predict Report Restate Review Tell Translate EMU: Office of Institutional Effectiveness and Accountability Apply Calculate Classify Demonstrate Determine Dramatize Employ Examine Illustrate Interpret Locate Operate Order Practice Report Restructure Schedule Sketch Solve Translate Use Write EVALUATION SYNTHESIS ANALYSIS Analyze Appraise Calculate Categorize Classify Compare Debate Diagram Differentiate Distinguish Examine Experiment Inspect Inventory Question Separate Summarize Test Arrange Assemble Collect Compose Construct Create Design Formulate Integrate Manage Organize Plan Prepare Prescribe Produce Propose Specify Synthesize Write Appraise Assess Choose Compare Criticize Determine Estimate Evaluate Grade Judge Measure Rank Rate Recommend Revise Score Select Standardize Test Validate Page 24 Consider using the NEW Bloom’s Taxonomy EMU: Office of Institutional Effectiveness and Accountability Page 25