Questioning Your Questions: Critically Examining Science Assessment Items Phyllis Kirkpatrick

advertisement
Questioning Your Questions:
Critically Examining Science
Assessment Items
Presented by
Phyllis Kirkpatrick, senior science program coordinator
Barbara Taylor, senior science program coordinator
at NSTA’s 56th National Conference on Science Education, Boston, MA
1
Reflection
Do I have the knowledge and skills to generate
assessment items for:
Classroom assessments?
District (including benchmark) assessments?
State assessments?
2
Session Objectives
• Increase participants awareness of the
purposes and types of assessments.
• Introduce participants to assessment item
guidelines and test construction guidelines.
• Practice using assessment item guidelines to
edit selected items.
3
Looking at Assessments
What do YOUR tests TEST?
4
Purposes of Assessment
1.
Diagnose student strengths and weaknesses
2.
Assign grades
3.
Determine instructional effectiveness
4.
Influence the public’s perception of educational effectiveness
5.
Guide instructional decisions
6.
Hold schools accountable
7.
Sort, select, rank, and compare students and schools
8.
Evaluate instructional programs
9.
Allocate resources
10.
Match students to a special program
5
“The purpose of assessment
MUST BE to improve student
learning. Period.”
Cathy Seeley,
Past President of NCTM
9/94
6
Use a Variety of Assessments
Essays
Performance Tasks
Hands-on Tasks
Logbooks or Journals
Portfolios
Selected Response
7
Know where you are going
“To begin with the end in mind means to start
with a clear understanding of your
destination. It means to know where you are
going so that you better understand where
you are now so that the steps you take are
always in the right direction.”
Stephen R. Covey
The Seven Habits of Highly Effective People
8
Selected Response Questions
Remember—no test question is perfect; it can always
be made better.
9
Question: Does the item match the depth and
complexity of the identified standard?
Which of the following is a sex-linked characteristic?
A.
Blood type
C. Color blindness
B.
Eye color
D. Tongue rolling
What is the question asking the student to do?
The question is asking the student to recall.
10
Question: Does the item match the depth and
complexity of the identified standards?
Color blindness is a sex-linked characteristic.
What are the chances of a normal male and a
Color-blind female having a color blind son?
A. 0%
B. 25%
C. 75%
D. 100%
This question requires the
student to make a prediction
based on an understanding
of the heredity of sex-linked
characteristics.
11
Question: Is there only ONE arguably correct
response, regardless of the teaching?
Which of these will result if all colors are mixed?
A.
Black
C. Green
B.
White
D. Yellow
The question is ambiguous:
Are the colors ‘light’ or
‘pigment’?
12
Question: Do questions avoid negatives/prompts?
Which of these is NOT a characteristic of animals?
A.
B.
C.
D.
They
They
They
They
eat other organisms.
can move.
do not make their own food.
have cell walls.
13
Question: Do questions avoid negatives/prompts?
Which of these is a characteristic that all animals
have in common?
A.
B.
C.
D.
Unicellular
Photosynthetic
Multicellular
Carnivorous
This question requires the
student to identify the
characteristics of animals
and other organisms.
14
Question: Does the item ask a question? Does
it avoid incomplete thoughts?
________describes characteristics you can detect
with your senses.
A.Physical properties
B.Exterior properties
C.Appearance
D.Chemical properties
This item has multiple
correct answers.
15
Question: Are any of the answer options clued by the
words in the stem?
The Millennium Seed Bank is a project designed to
protect the future of __________.
A. Sperm
B. Eggs
C. Plant species
D. Animal species
The correct answer is
clued— most students
could relate seeds
to plants. Or, it is possible
that it can be answered
only if this is specifically
taught.
16
Reviewing and Editing Test Items
17
Test Items
1. The item stem should ask a question.
2. Items should be written in active voice using simple
sentences.
3. Artwork should provide useful information and be
necessary for answering the question.
4. Items should avoid using superfluous language.
5. Make sure that the information in the item stem and
options is factual.
18
6. The items should focus on important content, not on trivial
details.
7. The reading level of the item should be from a half year
below grade level to a half year above grade level. Keep in
mind that scientific terms tend to increase the reading level.
8. Items should avoid obvious biases.
9. Avoid using negatives in the stem such as not and least.
Edit these items to become except items, with except at the
end of the stem.
10. The items should be written clearly enough so that
there is only one arguably correct answer. The wording of
the stem should not allow for wiggle room.
19
Test Item Options
1. Item options need to be parallel. This means that the
length and wording of each option should be checked so
that students are not drawn to one option over the others.
2. Item options should not be clued by words in the stem.
3. If the options are numerical, they should be written in
ascending or descending order, not in random order.
4. All incorrect options should be logical. If they are
mathematical, they should be based on miscalculations
students are likely to make. The incorrect options should
give a teacher an indication of what the students were
thinking or the miscalculation they made that caused them
to choose the incorrect option.
20
5. In items that require calculations, check to make sure
that a correct answer cannot be obtained by doing the
mathematics the wrong way. In other words, a student
should not be able to randomly add numbers found in
the problem and get the correct answer.
6. Avoid using repetitious information in options.
7. Avoid having opposites as options.
8. Avoid overlap in options, especially in numerical answer
sets.
21
Practice on Your Samples
You have some sample items in your packet. Work
with a partner and be prepared to report out:
Which of the items need to be improved?
How did you and your partner make the item
better?
22
When Constructing a Test
1. Test items should cover the full range of Bloom’s taxonomy.
2. No more then 10% of test items should be at the knowledge
level.
3. Try to have at least 5 items for each main idea being tested.
4. Place easier items at the beginning and end of the test. The
harder items should be in the middle of the test.
5. Try to place items on the test so that there is adequate white
space. Do not place artwork/scenarios on a page different
(adjacent or back side) from the page the related item(s) is
on.
23
6. Distribute the answer choices randomly; try not to
have the same answer more then three times in row.
7. Make sure that the length of the test is appropriate.
Most students should be able to complete the test in
the class period. Rule of thumb—multiply the time it
takes for you to take the test and make the answer key
by 4 to roughly determine the average time your
students will need to complete the test.
8. Make sure items don’t key the answers to other items.
The stem of one item should not contain information
that answers another item.
24
Teachers should NOT have to
put aside GOOD INSTRUCTION
to prepare for a TEST…
25
Reflection
1.
2.
3.
What did I learn about my practices and beliefs
regarding assessment item design and test
construction?
What action steps would I like to take in assessment
item design and test construction this coming school
year?
What was the most important thing I took away from
this experience?
26
Resources
Marzano, Robert J., Pickering, Debra J., and Pollock, Jane E. (2001). Classroom
instruction that works: Research-based strategies for increasing student
achievement. Alexandria, VA: Association for Supervision and Curriculum
Development.
National Research Council. (2001). Classroom assessment and the National
Science Education Standards. Committee on Classroom Assessment and the
National Science Education Standards. J. Myron Atkin, Paul Black, and Janet
Coffey (Eds.). Center for Education, Division of Behavioral and Social Sciences
and Education. Washington, DC: National Academy Press.
Sevenair, John P., and Burkett, Allan R. (2001, Fall). “Materials to accompany a
presentation on multiple choice questions, examinations, and test banks.”
Unpublished paper, Xavier University of Louisiana. Available at
http://webusers.xula.edu/jsevenai/objective/ObjectiveTests.pdf
Stiggins, Richard J. (2002, June). “Assessment crisis: The absence of assessment
FOR learning.” Phi Delta Kappan. Retrieved November 16, 2005, from
http://www.pdkintl.org/kappan/k0206sti.htm
27
Contact Information
Phyllis Kirkpatrick
pkirkpat@mail.utexas.edu
Barbara Taylor
Barbara_taylor@mail.utexas.edu
28
Download