Chapter 1. Introduction: The World of Educational Assessment 1 The word ‘assess’ . . . comes from the Latin verb ‘assidere’ meaning ‘to sit with’. Thus, in assessment one is supposed to sit with the learner. This implies it is something we do with and for students and not to students. (Green, J. M. 1998, February. Constructing the way forward for all students. A speech delivered at “Innovations for Effective Schools” OECD/New Zealand joint follow-up conference, Christchurch, New Zealand.) 2 consider The McNamara Fallacy attributed to economist Charles Handy ... The first step is to measure whatever can be easily measured. This is okay as far as it goes. The second step is to disregard that which can’t be easily measured, or to give it an arbitrary value. This is artificial and misleading. The third step is to presume that what can’t be measured easily isn’t important. This is blindness. The fourth step is to say that what can’t be easily measured doesn’t exist. This is suicide. Bottom Line of the Robert McNamara fallacy: “What does not get counted does not count.” 3 Edward L. Thorndike (1874-1949) . . . his pioneer investigations in the fields of human and animal learning are among the most influential in the history of Psychology. “Whatever exists, exists in some quantity. If it exists in quantity, it can be measured.” – E. L. Thorndike (quote paraphrased) 4 So, as a teacher, How do you make sure you are counting the most important things (aka . . . the most important aspects of learning) in your assessment system? It is the overall intent of this class to help you develop useful attitudes and skills in this area. 5 Four Sections of Study I. Basic Principles of Assessments II. Teacher- Made Assessments III. External Assessments IV. Special Applications 6 I. Basic Principles of Assessment Statistics (Ch 2) Reliability (Ch 3) Validity (Ch 4) Interpretation (Ch 5) 7 II. Teacher Made Assessments Planning for Assessments (Ch 6) Selected-response (Ch 7) Constructed-response (Ch 8) Interests, attitudes, creativity (Ch 9) Administering and analyzing (Ch 10) 8 III. External Assessments Achievement (Ch 11) Ability (Ch 12) Behavioral (Ch 12) 9 IV. Special Applications Grading and Reporting (Ch 13) Laws related to assessment (Ch 14) Evaluating your own teaching (Ch 15) 10 What Professional Groups Want You To Know About Assessment Examples of Professional Statements BCOE Conceptual Framework – Reflection in Action Ohio Standards for the Teaching Profession Ohio Integrated Systems Model (OISM) Praxis II PLT Category II – Assessment NCATE SPAs – Specialized Professional Associations Main Points These Groups Make 1. Have clear learning objectives. 2. Develop skill with variety of assessment methods. 3. Be familiar with standardized tests. 4. Know technical stuff (e.g., reliability, validity). 5. Know how to grade, give feedback. 6. Be familiar with legal and ethical issues. 11 Purposes of Assessment 1. Certify Student Learning 2. Motivate Student Learning 3. Instructional Planning or Diagnosis 4. Feedback to Important Users of Assessment as to how we are doing. To the parents To the school administrators To the public community 5. Research 12 Current Assessment Trends Assessment is Built on Classical Test Theory The drive to make educational assessment as “scientific” as possible Standards-based Education Its characteristics Content Standards (Learning Goals/Objectives) Performance Standards Emphasis on HIGH Performance Standards These HIGH Standards Apply to All Learners Public Reporting Today’s example - No Child Left Behind (NCLB) 13 Current Trends (cont.) Accountability schools must demonstrate their outcomes teachers must demonstrate competence Legal Activism laws and courts are used to pursue social goals concern for fairness and equity Alternative Performance Assessment concepts of authentic assessment portfolios 14 Trends (cont.) State, National, International bodies are increasingly interested in group effects (often related to “return on investment” thinking). State (explosive growth right now) National - National Assessment of Educational Progress (NAEP) International – (increasing attention) 15 Terms Measurement Quantification of observation or a believed underlying reality Testing Using a selected measurement process or measuring device Assessment Drawing a conclusion or making a judgment about information from testing. 16 Terms (cont.) Standardized Testing may involve: Uniform, clearly specified methods and procedures for administering the test. The test has scoring norms based on many (perhaps thousands) previous cases Group administered, machine-scored, multiple choice tests. “I don’t do well on standardized tests!” Means which of the above, exactly? Your response? 17 Terms (cont.) Distinctions Among Types of Tests Group vs. Individual Speed vs. Power Maximum vs. Typical Performance Criterion vs. Norm Referenced Paper-and-Pencil vs. Performance Summative vs. Formative Selected-Response vs. Constructed Response 18 High-Stakes Testing A situation in which the outcome of a test has exceptionally important consequences for an individual. You fail . . . You don’t get in or get out. You fail . . . You are nobody to somebody. You fail . . . You wasted your time. This can create test anxiety in the individual and could lead to non-performance or under performance. 19 Terms/Concepts to Review and Study on Your Own (1) accountability assessment authentic assessment classical test theory constructed-response criterion-referenced formative evaluation 20 Terms/Concepts to Review and Study on Your Own (2) high-stakes testing measurement NAEP No Child Left Behind (NCLB) Act norm-referenced performance assessment portfolio assessment 21 Terms/Concepts to Review and Study on Your Own (3) reliability selected-response standardized standards-based education summative evaluation test TIMSS validity 22