Review: Alternative Assessments • Alternative/Authentic assessment • Real-life setting • Performance based • Techniques: • • • • • Observation Individual or Group Projects Portfolios Exhibitions Student Logs or Journals • Developing alternative assessments • • • • • Determine purpose Define the target Select the appropriate assessment task Set performance criteria Determine assessment quality Review: Grading • Grading process: Objectives of instruction Test selection and administration Results compared to standards • Making grading fair, reliable, and valid • • • • • • • Determine defensible objectives Ability group students Construct tests which reflect objectivity No test is perfectly reliable Grades should reflect status, not improvement Do not use grades to reward good effort Consider grades as measurements, not evaluations Final grades Cognitive Assessments Physical Fitness Knowledge Physical Fitness Knowledge HPER 3150 Dr. Ayers Test Planning • Types Mastery (driver’s license) Achievement (mid-term) Table of Specifications (content-related validity) • Content Objectives history, values, equipment, etiquette, safety, rules, strategy, techniques of play • Educational Objectives (Blooms’ taxonomy, 1956) knowledge, comprehension, application, analysis, synthesis, evaluation Table of Specifications for a 33 Item Exercise Physiology Concepts Test (Ask-PE, Ayers, 2003) T of SPECS-E.doc Test Characteristics • When to test • Often enough for reliability but not too often to be useless • How many questions (p. 185-6 guidelines) • More items yield greater reliability • Format to use (p. 186 guidelines) • Oral (NO), group (NO), written (YES) • Open book/note, take-home: (dis)advantages of both • Question types • Semi-objective (short-answer, completion, mathematical) • Objective (t/f, matching, multiple-choice, classification) • Essay Semi-objective Questions • Short-answer, completion, mathematical • When to use (factual & recall material) • Weaknesses • Construction Recommendations (p. 190) • Scoring Recommendations Objective Questions • True/False, matching, multiple-choice • When to use (M-C: MOST IDEAL) • FORM7 (B,E).doc • Pg. 196-203: M-C guidelines • Construction Recommendations (p. 191-200) • Scoring Recommendations Cognitive Assessments I • Explain one thing that you learned today to a classmate Review: Cognitive Assessments I • Test types • Mastery Achievement • Table of Specifications (value, use, purpose) • Questions Types • Semi-objective: short-answer, completion, mathematical • Objective: t/f, match, multiple-choice • Essay (we did not get this far) Figure 10.1 The difference between extrinsic and intrinsic ambiguity (A is correct) B B B A A A D D C C D C Too easy Extrinsic ambiguity Intrinsic Ambiguity (weak Ss miss) (all foils = appealing) Essay Questions • When to use (definitions, interpretations, comparisons) • Weaknesses • Scoring • Objectivity • Construction & Scoring recommendations (p. 205-7) Administering the Written Test • Before the Test • During the Test • After the Test Characteristics of Good Test Items • • • • • • Leave little to "chance" Reliable Relevant Valid Average difficulty Discriminate Gotten correct by more knowledgeable students Missed by less knowledgeable students • Time consuming to write Quality of the Test • Reliability and Validity • Overall Test Quality Individual Item Quality Item Analysis • Used to determine quality of individual test items • Item Difficulty Percent answering correctly • Item Discrimination How well the item "functions“ Also how “valid” the item is based on the total test score criterion Item Difficulty 0 (nobody got right) – 100 (everybody got right) Goal=50% Uc Lc Difficulty * 100 Un Ln Item Discrimination <20% & negative (poor) 20-40% (acceptable) Goal > 40% Uc Lc Discri min ation * 100 Un Figure 10.4 The relationship between item discrimination and difficulty Sources of Written Tests • Professionally Constructed Tests (FitSmart, Ask-PE) • Textbooks (McGee & Farrow, 1987) • Periodicals, Theses, and Dissertations Questionnaires • • • • • • • • Determine the objectives Delimit the sample Construct the questionnaire Conduct a pilot study Write a cover letter Send the questionnaire Follow-up with non-respondents Analyze the results and prepare the report Constructing Open-Ended Questions • Advantages Allow for creative answers Allow for respondent to detail answers Can be used when possible categories are large Probably better when complex questions are involved • Disadvantages Analysis is difficult because of non-standard responses Require more respondent time to complete Can be ambiguous Can result in irrelevant data Constructing Closed-Ended Questions • Advantages Easy to code Result in standard responses Usually less ambiguous Ease of response • Disadvantages Frustration if correct category is not present Respondent may chose inappropriate category May require many categories to get ALL responses Subject to possible recording errors Factors Affecting the Questionnaire Response • Cover Letter Be brief and informative • Ease of Return You DO want it back! • Neatness and Length Be professional and brief • Inducements Money and flattery • Timing and Deadlines Time of year and sufficient time to complete • Follow-up At least once (2 about the best response rate you will get) The BIG Issues in Questionnaire Development • Reliability Consistency of measurement • Validity Truthfulness of response • Representativeness of the sample To whom can you generalize? Cognitive Assessments II • Ask for clarity on something that challenged you today