Eastern Michigan University Peggy Liggit – Director of Academic Assessment

advertisement
Eastern Michigan University
Peggy Liggit – Director of Academic Assessment
Steven Pernecky – Professor of Chemistry
Micheal Tew-Professor of Communications, Media, Theatre Arts
Peggy Liggit
Shared Assumption:
Conscious Competent Learning Model
Unconscious
Conscious
Incompetent
Incompetent
Conscious Competent
Unconscious Competent
Most faculty who achieve tenure are
operating at this level in their discipline
y “The connections made by good teachers are held not in their methods but yg
in their hearts – meaning heart in its ancient sense, as a place where intellect and emotion and spirit and will converge in the human self.” ‐‐Parker Palmer, Courage to Teach.
y “You should never worry about your good ideas being stolen in educational reform, because even when people are sincerely motivated to y ,
y
g
learn from you, they have a devil of a time doing so.” ‐‐Michael Fullan, Change Forces: The Sequel.
y “One important feature of embedded assessment is that it “blurs the lines” between teaching and assessment.” ‐‐James J. Gallagher, Improving Science Teaching and Student Achievement through Embedded Assessment.
Why There are Few ‘Good’ Assessment models
Michael Fullan (1999, 2006)
y Innovative ideas are “difficult to disseminate and replicate.”
y Not easily transferable ‐ difficult to capture “subtleties of the reform practice.” y The inability to replicate another’s model, “replicating the Th i bili
li
h ’
d l “ li i
h
wrong thing” y
“The
The reform itself, instead of the conditions which spawned reform itself instead of the conditions which spawned
the success.” y
Problems with scale: small ‐scale may not work well on a y
wider‐scale. Reflect about a misconception/problem in g
y
g
student learning that you did something about.
y What/how were you teaching at the time?
y What activity were you doing to determine student’s y
y
g
thinking/reasoning?
y What did students understand/not understand about what you were teaching?
y What did you do to help students better understand?
Care to share?
Care to share?
You have just documented a, presumably,
d
t d
bl
unconscious competent activity. Teach, Assess, Analyze, and Adj t
Adjust = Embedded Assessment
E b dd d A
t
Include an activity that intentionally
Gather information about learning through
focuses on a learning outcome and
work samples that expose
reveals what students’ are
students’ ideas/skills students
“thinking” – not just
e.g. constructing
Assess
focusing on a
representations
final or
Assess
of what they
correct
know.
Assess
answer.
Teach
Teach
Teach
Start
Analyze
Analyze
Analyze
y
Adjust
Adjust
Decide and
i l
implement
t nextt
instructional moves.
Adjust
Analyze what
students struggle
with as learning
g
happens in
real time?
Embedded Assessment Cycle
Embedded Assessment Cycle “Blurring the lines” between instruction and assessment. • Read how other’s have found misconceptions
R d h th ’ h f
d i
ti
• Try it yourself!
E b dd d A
t
Embedded Assessment:
Identifying Misconceptions Exercise
Identifying Misconceptions Exercise
Located ~10 diverse articles about misconceptions in varied disciplines. Looked for these trends:
1) Wh t
1) What was the misconception?
th
i
ti ?
2) How was the misconception discovered? 3) What was the suggestion on how to improve student understanding about the misconception. E
l A i l
Example Articles:
Mapping the structures of resonance
(M ld 2009.
(Malde.
2009 N
National
ti
lA
Association
i ti off T
Teachers
h
off Si
Singing)
i )
“Body Mapping helps people discover and correct
misconceptions about the way their bodies are built…”
Misconceptions Articles: Results
y Misconceptions found through various activities:
y performance or application activities,
y disciplinary writing,
y problem solving “show your work” exercises
y Most remedies included instruction with an activity
(hands-on or some kind of problem set.)
Embedded assessment is a useful model for Embedded assessment is a useful model for
finding and fixing misconceptions and supports the conscious competent learning model
conscious competent
learning model
AND
serves as documentation for reporting on p
g
accountability, …thus humanizing the process.
Writing Better Exam Essays in Human Physiology
With permission from D H
Dr. Howard Booth (EMU‐
d B th (EMU Biology)
Bi l
)
ASSESS
HUMAN PHYSIOLOGY
TEACHING IMPROVEMENT PROJECT
ANALYZE
TEACH
ADJUST
y and evolution into a SOTL p
j
History
project
My conviction: science students need to communicate in their discipline
Revealing misconceptions: Student writing is a prime place to asses their
understanding of content but poor writing skills distorts this
assessment and lowers exam scores disproportionately
disproportionately.
(Good students were losing many essay points due to poor writing)
The challenge: good writing is hard work for teachers and students
Traditional resolutions: ignore it, require it, or teach it
Hypothesis: If students understood and applied good organizational and
proofreading techniques to their writing, essay scores would move up to be
comparable to the rest of the exam and provide a truer assessment of their learning
learning.
Goal of the project: to improve the Big Essay scores (worth 25% of the exam; 4 exams per semester)
-Through a focused effort to teach students better essay writing skills
-Through
g a secondary
y focus on improving
p
g organization
g
and coherence
Students Learning Outcome 2: Students will communicate scientific knowledge and concepts in written form.
Students Learning Outcome 3: Students will synthesize scientific knowledge in higher level thinking problems.
17
ASSESS
TEACH Phase 1: Getting Organized
The SOTL Experiment
ANALYZE
TEACH
ADJUST
Control group: Student exam and essay scores from 2008 (before essay focus)
Experimental Group: Fall 2009 and Winter 2010 classes
-All students: essay score and exam scores
Subgroup: student scores from Rewrite Workshop
-Subgroup:
Time Line --- Methods and materials
Pre exam 1: All lecture instructions “how to take the exam” (5 min)
P
Post
exam1:
1
-All lecture illustration of an excellent essay answer (10 min)
-Essay rewrite workshop: For those “performing poorly” or “want to do better”
Instruction-discussion
Instruction
discussion on writing & organization skills (40 min)
Re-write exam1 essay: take home, self timed (30 min)
Re-Grade: scores, written comments & personal discussion
Postt exam 2:
P
2 lecture
l t
di
discussion/illustration
i /ill t ti off a conceptt map off essay (15 min)
i )
Pre exam 4: Opinion survey on rubric & essay scoring
Inclusion of an additional rubric statement.
18
ASSESS:
Essay 1 Example
“Bill”,
A good
Student
89%=“A-” on
Non-essay
Part of exam
Earning
A
Poor
essay
Score
16 of 25
is
64%=“D”
19
ASSESS
TEACH Phase 2: The Workshop
ANALYZE
TEACH
Doing the “Experiment”
Workshop Highlights
•
•
•
•
•
•
ADJUST
Blue print - how to write a good paragraph: Theme introduced in
topic sentence, supporting sentences in logical order, a conclusion.
Science
Sc
e ce writing
g iss concise,
co c se, de
detailed,
a ed, accu
accurate,
a e, b
breadth
ead o
of the
e theme
e e
Building blocks- analyze the question (see printout)- finding out what to include
Main theme, mentioned areas of interest, required 9 terms/phrases
Organizing the construction site- organization of logical sequence
Outline areas of interest and terms-terms flow diagram & concept map
Building it-- apply above -- also, some hints on good construction
Set aside enough time, avoid “wordy” superficial statements,
use the space
p
p
provided- not more , not less. Write neatly,
y 14-18point
p
too small hard to read, too large wastes space. Spell correctly.
Building inspection-- proof read, all “areas of interest” included?’
& terms used correctly, explained, circled and numbered?
match your essay with rubric - avoid losing unnecessary points
Avoid building disasters-- study ! If your blocks crumble----if you don’t know the content, good writing won’t help much
20
ASSESS:
“Bill’ ”
“Bill’s”
Rewrite
Scored
92%
On
The
Rewrite
28%
Better
From a
“D”
p to
Up
an “A-”
21
ASSESS
ANALYZE: What we were looking for
and what we found
The Results
ANALYZE
TEACH
• The quality of the re-writes improved:
F’09, 13/28 (46%) to 23/28 (82%) a gain of 36 %
ADJUST
p
y
• Improvement
carried into exam 2 and beyond
13/28 to 19.8/28 (up 70.1%) on exam 2.
• Overall exam scores moved up over 5% from 74.4% to 79.7% .
• Numerous positive student comments about the experience and their increased
comfort with essay writing.
writing “ I loved the essay format
format- very helpful”
helpful F
F’09
09.
•Participation increased from 15/60(25%) F’09 to 38/61 W’10(62%) (up 37%)
•Rubric Opinion Survey to assess student misconceptions or concern on scoring
•Change was greatest in the poorest writers, about 65% improved dramatically
Data Chart of Essay Score Averages before (W’08) and now (W’10)
Exam 1
Exam 2
Exam 3
Exam 4
17.8
18.3
18.5
19.8
63.5%
65.2%
66.1%
70.8%
16.8
21.0
22.0
22.0
% chang e numbe r/28 possibl e
59.9%
75.5%
78.6%
77.8%
W'08 C ontrol all stude nts' essays
61.7%
63.3%
63.0%
66.1%
All students Averag e
% chang e numbe r/28 possible
Workshop
p stude nt averages
g
And see
the Graph
next slide
lid
22
ASSESS
ANALYZE
TEACH
ANALYZE : Results Graph and Summary
Graph of Essay Score Averages
Before and Now
ADJUST
90.0%
W'08 Control
80.0%
W'10 All
Students
W'10 Workshop
students
70.0%
60.0%
50.0%
40.0%
30.0%
20.0%
Series1
control
RESULTS SUMMARY
1. W'10 Workshop participants impoved most from a D+ to a B
2. W'10 all students improved modestly from a C- to a C+
3. W'08 control students improved the least from C- to a C
4. Workshop participants improved most from exam I to exam 2
and continued to improve in exams 3&4
5. All W'10 students benefited compared to W'08 control
6. Essay scores no longer "Disproportionately low"
Workshop 5% above, AllW'10 at all exam averages.
All
W'10
Series2
Workshop
Series3
10.0%
0.0%
1
2
Exams
3
1, 2, 3
&
4
4
I c o mpleted th e s u rvey f o r a B onu s pt. Name: _ ______ _ _____ _ _____ _ _
-- ----- ----- ------- ------- ------- ----- ---tear o ff o r c u t h ere -- --- ------- ----- ------- ---
ANALYZE:
Rubric
Rubric Survey Results
Opi n ion Su rvey on Big Essay P o in ts Dis trib ut ion (sc o ri n g r u bric)
P lease circ le th e lett e r b e fo re
a= s tro n g ly agre e
b = agree
c= n eith er agree n or d isagree
d= d isagree
e= s tro n g ly d isagree
th e statem e n t to in dica te y o ur le v e l of agree m en t w ith it.
5p t
4p t
3p t
2p t
1p t
(An y co mm en ts e x p lain in g y ou r th o ug h ts w o u ld a lso b e v er y h e lp fu l,. T h an ks)
40 s tude n ts resp
p o n ded ( o u t o f 53 still atte n d in g n ear th e en d of t h e se m es ter )
Conclusions:
1 U
1.
Understanding
d
t di
the essay
scoring is not
much of a
problem
2. It’s easy to fix it
for the few who
were concerned
((add a statement to
the exam)
1. ( a b c d e ) A s I read th e essa y ques tio n itse lf , I thin k I h a v e a c lear idea o f h ow I ca n ear n fu ll p o in ts.
Resu lts: 4 @ 2p t, 3@ 3p t, 24 @ 4p t 9 @ 5p t = a v erage 4 .0 o u t o f 5.0
An aly sis: M o st d id n' t see ru b ric sc o rin g as a p ro bl em , b u t 7 (2 0% ) d id h a v e co n cer n s h er e.
C o mm en ts: " v er y c lear"
" th e essa y ques tio n s are usua lly s tra ig h t fo rwar d an d y o u k n o w y ow th ey ar e supp o sed to b e an swer in g." ,
" o n ce I we n t to th e w orks h o p I was ab le to b ett e r u n ders ta n d h o w th e essa y was to b e w ritt en "
2. ( a b c d e ) A separa te ch eck li st a n d e x p lana tio n , n o t par t of th e ques tio n itse lf e x p lain in g p o in t
d istrib u tio n (sc o rin g ru b ric), w o u ld m ake th is p roc ess m o re u n ders ta n da b le.
Resu lts: 1 @ 1, 5 @ 2p t, 9 @ 3 pt ,18 @ 4, 7 @ 5p ts = average o f 3. 6 o u t of 5.0
An aly sis: Qu ite a stro n g fee li n g t h is w o u ld h elp. Bu t read in g t h e co mm en ts left m e th in k in g th is was n' t
v er y im p o rtan t to th em .
("ca n' t h ur t")
C o mm en ts:
"g
g oo d idea, T h is w ill a ll ow stude n ts to h av e a n e x ac t acc o u n t for a ll p o in ts
ts"
"y o ur sc o rin g m eth o d is a lread y c lear"
"h elp fu l for fi rs t e x a m , I f e lt v er y u n co mf o rtab le w ritin g it b ecause I h ad n o c lue w h a t to e x pec t. A g rad in g
ru b ric w o u ld let m e k n o w w h at y ou were e x pec ting"
3. ( a b c d e ) C o n fus io n a b o u t h ow I ear n th e essa y p o in ts c o n cer n s m e as I w rite th e essa y (" n ot
k n o w in g" w h ere y o u mi g h t ga in o r l o se p o in ts in ter feres w ith y ou r th o ug h t pr og ress io n as y o u are p lann in g
a n d wr itin g th e essa y ).
Resu lts: 2 @ 1p t, 13 @ 2 pt, 9 @ 3p t, 8 @ 4p t, 5@5p t, = av erag e o f 2.9 out o f 5
An aly sis: Near n eu tra l n ot stro n g fee li n g o r c o m m e n ts. N o t m uc h o f a n issue for m o st
C o mm en ts:
"Wh il e I'm wr itin g t h e essa y I 'm jus t tryi n g t o d o th e b es t I ca n . I'm n ot w orr ied ab o u t th e p o in ts ." " N o ,
dep th o f a n swer an d lack o f tim e ar e my c o n cer n s", "T h is ca n b e t ro u b li n g b ecause t h e p o in t sy stem isn' t
cu t a n d dr y "
4. (a b c d e ) I th in k m o re in fo rm atio n o n th e sc o rin g ru b ric w o u ld im pro v e m y essa y sc o res.
4
res
Resu lts: 3 @ 1p t, 7 @ 2 pt, 1 4 @ 3p t, 14 @ 4, 1@ 5p t = a v erage o f 3.1 o u t o f 5
An aly sis: A ga in alm o st n eu tra l o v er all b u t a b o u t 30% at leas t fa v o red th e idea. C o mm e n ts:
"m o re in fo rm atio n n e v er h urts", "a sc o rin g ru b ric d o es n' t e nh a n ce o r inh ib it my w ritin g."
5. An y ot h e r sugges tio n s?
24 y ,
"A grad in g ru b ric a t th e b eg inn in g of th e se m es ter w o u ld b e h e lp fu l, I f e lt v er y n er v o us ab o u t th e fi rs t essa
a fter th at it ge ts eas ier"
ASSESS
ADJUST: Continuous fine tuning
ANALYZE
TEACH
ADJUST
1. Already have incorporated a series of changes
•Three “editions” of workshop notes
half page outline of my talking points
full page on writing
two page handout on writing,
organizing, proofreading and rubric
•Adding concept map to lecture presentation W’10
•Expanded “ workshop
workshop” to a take
take-home
home option W’10
W 10
•Last week upgraded the handout and sent e-copies to all
of my spring students as a Pre Exam 1 study sheet.
•Plan to subdivide workshop participants into working
groups this spring.
25
ASSESS
Conclusions
Yes, teach
it again
!
g
TEACH
ANALYZE
Experiment was a success:
ADJUST
-hypothesis supported
-embedded
data
b dd d d
t collected
ll t d
-students engaged & learning outcome goals were met
-essay scores improved and now more accurately
reflect learning (no longer disproportionately low)
-improved essays: better assessment of misconceptions
Cost/ Benefit Analysis
-costs: clocked about 10 hours per semester
-benefits:
students
t d t write
it better--realize&
b tt
li & appreciate
i t itit-- “you
“
care””
course & the department get assessment quantified
I followed my convictions --teaching “good” writing in science
improvement keeps teaching interesting and enjoyable
Bottom line: It was, and is a success at many levels
26
y
Steve Pernecky EMU Biochemistry Program Assessment
EMU Biochemistry Program Assessment
What did we need to get started?
y
• Passion: Committed core of faculty
• Motivation: AQIP/Program Review
• Goal: Continuous Improvement
• Framework: Progression of Student Knowledge
ASSESS
ANALYZE
TEACH
ADJUST
ASSESS
™
Develop four Student Learning Outcomes, each containing
four to six components.
components
Conceptual/quantitative problem solving in chemistry
Lab techniques in chemistry
P f
Professional
i
l skills/standards
kill / t d d
Practice of chemistry as scientific endeavor
™
Develop curriculum map Æ links Student Learning
Outcomes to courses.
™
Brought
g Student Learning
g Outcomes and Curriculum
Mapping to Chemistry faculty to refine/approve .
ASSESS
p
Curriculum Map:
Biochemistry Program
R
R
R
R
R
123/ 124/
281/
121 122 125 126 270 271 282
R
361
R
R
371 372
R
R
R R
373/
Res
376 381 461 463 465 451 452 453 481 475 Coop 415 411 412 413
Students will evaluate
problems in chemistry using
conceptual and quantitative
approaches
h
Stoichiometry
*
*
*
*
*
Phys.-Chem. properties
Chemical bonding and
reactivity
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
Thermodynamics
*
*
*
*
*
*
*
*
*
*
*
Kinetics
*
*
*
*
*
*
Equilibria
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
*
R = Required course in General Biochemistry Curriculum
AND Professional Biochemistry program
*
*
*
Why Start with Equilibrium?
Why Start with Equilibrium?
y Teacher’s Misconceptions about the Effects of Addition of More Reactants or Products on
y
y
y
y
y
y
Chemical Equilibrium (2009) Cheung, D, Ma, H., Yang, J., Int. J. Sci. Math. Ed. 7, 1111-1133.
pp
g Chemical
Effectiveness of Instruction Based on the Constructivist Approach
on Understanding
Equilibrium Concepts (2003) Akkus, H., Kadayifci, H., Atasov, B., Geban, O., J. Res. Sci. Tech.
Ed. 21, 209-227.
Dynamic Processes of Conceptual Change: Analysis of Constructing Mental Models of Chemical
Equilibrium
q
((2002)) Chiu,, M.,, Chou,, C.,, Liu,, C.,, J. Res. Sci. Teach. 39,, 688-712.
Investigating the Relationship between Students’ Multiple Conceptions of Action and Reaction in
Cases of Static Equilibrium (2001) Palmer, D., Res. Sci. Tech. Ed. 19, 193-204.
An In-Depth Study of Misconceptions in Stoichiometry and Chemical Equilibrium at a South
African University (1996) Huddle, P., Pillay, A., J. Res. Sci. Teach. 33, 65
65-77.
77.
Students’ and Teachers’ Misapplication of Le Chatelier’s Principle: Implications for the Teaching
of Chemical Equilibrium (1995) Quilez-Pardo, J., Solaz-Portoles, J.J., J. Res. Sci. Teach. 32, 939957.
Teaching Chemical Equilibrium and Thermodynamics in Undergraduate General Chemistry
Classes (1995), Banerjee, A.C., J. Chem. Ed. 72, 879-881.
We had a good sense from teaching our own curriculum
that there were misconceptions about chemical equilibrium.
ASSESS
™ Development of an assessment project: Develop an use
instrument for one component of SLO #1: Equilibrium (F
‘08).
™ Faculty
y surveyy of equilibrium
q
exam q
questions used in
chemistry courses (F ‘08).
™ Develop set of questions from survey that can be used to
assess “knowledge
“k
l d progression”
i ” (W ‘09)
‘09).
• Variables tested: (1) Reversibility of chemical reactions; (2) Le
Chatelier’s
Chatelier
s Principle; (3) Quantitative aspects of equilibrium
equilibrium.
• Bloom’s level for each question.
K=Knowledge level, A=Applied level, S=Synthesis level
ASSESS
Distribution of Equilibrium Questions in Four Chemistry Courses
Variable
1: Reactions are reversible
2: Le Chatelier's
Principle
3: Quantitative aspects
Bloom's Level (K, A, S)
(K A S)a
Course
CHEM123
CHEM281
CHEM372 CHEM451
K
*
*
*
*
A
S
*
‐‐‐
‐‐‐
*
*
‐‐‐
*
K
‐‐‐
‐‐‐
*
A
S
‐‐
*
*
*
*
‐‐‐
*
*
K
*
*
‐‐‐
*
A
S
*
‐‐‐
*
‐‐‐
‐‐‐
‐‐‐
*
ASSESS
y instrument in courses ((Su/F ‘09))
™Verify
• Interview students: Determine if equilibrium questions
were clearly stated.
• Instructor
I t t review
i
off questions
ti
and
d rubrics
b i (F ‘09).
‘09) Thi
This
was very important for buy-in.
™Modify instrument (F’09)
(F 09)
• CHEM123/General Chemistry and CHEM451/Biochemistry I:
Did not change in question.
• CHEM372/Organic Chemistry II: Scrapped instrument,
developed new instrument.
• CHEM281/Quantitative Analysis: Make questions reflect
language used in course.
ASSESS
Original
question in
CHEM281
1. Answer the following questions about the bicarbonate buffer system. The pKa for H2CO3 is 6.1. A patient is found to have a [HCO
p
[
/
[ 2CO3]] of 2.4 3­]] of 29.5 mmol/L and a [H
mmol/L. The pH of normal blood is 7.4. Is the patient suffering from acidosis (a blood pH more acidic than 7.4) or from alkalosis (a blood pH more basic than 7.4)? SHOW your answer assuming that the bicarbonate buffer system is the only buffer system controlling blood pH. Variable: 3, Level: K, A Refined
question in
CHEM281
1. HCO3‐ (bicarbonate) is an important species in biological systems. It comes from CO2 dissolving in the blood (pH=7.42). a) Write the equilibrium reactions for CO2 dissolving in blood (Hint: since blood is mostly water, you can consider it to be water in your reactions). b) Some medicines cause people to eliminate HCO3‐ in excess. How would loss of HCO3‐ affect blood pH? c) By adding more of which of the existing species in the system would the pH return to normal (pH 7.4)? ASSESS
1. Bicarbonate buffer system question
a) CO2 + H2O ⇄ H2CO3; H2CO3 ⇄ H+ + HCO3Mark the answer as correct if there are no errors or only minor errors.
Mark the answer as incorrect if there are any significant errors, such as one or
Rubric and
scoring for
question in
CHEM281
more incorrect equations.
b) Elimination of HCO3- would cause the second equilibrium to shift to the right,
increasing H+ concentration and therefore decreasing pH.
Mark the answer as correct if the explanation is appropriate and the conclusion
(decrease in pH) is correct.
Instructor ______________________________
Number
int or
this
M k the
Mark
th answer as incorrect
i
t if th
the
conclusion
l of
i students
is
i incorrect
i
the
th section
explanation
l ________
ti iis
incorrect or incomplete.
Bicarbonate buffer system question
c) Adding
shift the second
decreasing
H+
a) equilibrium
________ to the left,students
answered
the question correctly
students answered the question incorrectly
concentration and therefore increasing
g ________
p
pH.
________
students left the question blank
Mark the answer as correct if the explanation is appropriate.
b) ________
students answered the question correctly
Mark the answer as incorrect if the explanation
is incorrect or incomplete.
________
students answered the question incorrectly
________
students left the question blank
c) ________
students answered the question correctly
________
students answered the question incorrectly
________
students left the question blank
HCO3- would
1.
ASSESS
™Administration of Instrument (W, Su, F ‘10)
Course and Number of Sections/Students in which
Equilibrium was Assessed
Course
CHEM123/Gen. Chem. II
Number of Sections Number of Students
10
339
CHEM281/Quant. Anal.
4
119
CHEM372/Org Chem.
CHEM372/Org.
Chem II
3
39
CHEM451/Biochem. I
2
66
ANALYZE
ADJUST
™ Instructors of each disciplinary group were invited to meet
with to discuss and analyze the data (W ’11)
™ Discussion lead to a number of questions.
y Evaluating the instrument
o
o
Should we change
g the q
questions for each course?
Was Bloom’s level appropriate for each question?
y Core values
o
o
How important is it that students can apply (good “A”
A level) concepts
without really understanding them (poor “K” level)?
How important is it that we address a particular concept for each course
or in the program?
y Adjustments
o
o
Should pedagogy change to incorporate more conceptual teaching of
chemistry concepts?
How willing are we to change a course, the curriculum, what we teach,
and the way we teach it?
General Observations about Assessment Across the C rric l m
Across the Curriculum
™ Personnel: A core group of motivated workers who believe in
assessing for improvement can lay the groundwork quickly!!
• Student Learning Outcomes and Curriculum Mapping
• Develop/verify assessment instruments, level of testing (K, A,S), and
variables tested.
tested
• Chemistry Education researcher – Dr. Amy Johnson
™ Time: Expect
p
to spend
p
lots of time on refinement of the
assessment instruments
• Interview students to validate questions
• Get feedback on questions by instructors to get buy-in
buy in
• Take time enough to get good data
™ Process: Develop a comfortable mechanism (assume no
release
l
titime)) th
thatt will
ill b
benefit
fit your curriculum,
i l
nott jjustt an
institutional requirement.
Michael Tew
Professor
C
Communication, Media and Theatre Arts
i i M di d Th
A
B i t
t l
t f t
Barriers to natural assessment of outcome based learning across disciplines:
y Perceived differences
y Disciplinary habits
y Language and interpretation of outcomes
y Instructional strategy
gy
y Misconceptions
y different work = incompatible data
y Common outcome = common work
y Coordination = compromised integrity
Multicourse Learning in EMU
Multicourse Learning in EMU General Education
y Quantitative Reasoning
y 12 Outcomes
y One of Eight courses in four
disciplines
y Perspectives On a Diverse World
y Global Awareness
5 Outcomes
y One of 16 courses in 12
di i li
disciplines
y US Diversity
y 5 Outcomes
y One
O off 20 courses iin 14
disciplines
y
y Arts
y 6 Outcomes
y Two of 20 courses in 5
disciplines
p
y Humanities
y 7 Outcomes
y Two of 40 courses in 12
disciplines
y Social Sciences
y 7 Outcomes
y Two of 20 courses in 11
disciplines
y Natural Sciences
y 13 Outcomes
y Two of 23 Courses in 7
di i li
disciplines
Multicourse Assessment
y Our Challenge: Multiple courses across different
disciplines + common outcomes
y Our Commitments: to disciplinary habits; to
embedded assessment; to direct assessment
y Our Question:
y Can we understand student learning in common ways
when learning is manifested in different ways/in
different performances?
An Extension of the Model
• Outcome in the discipline
• Master themes
• Evidence in the discipline
Teach
Assess
Adjust
Analyze
• Shared values and challenges
GEMAP
Faculty Focus Groups
Interpret Outcomes
Offer Exemplars
Common Language/Values/Challenges/Themes
Build Outcome Based Analytic Tool
Compile Outcome Portfolio of student work
Data Generation and Analysis
Application of Tools
Contribution to Understanding of Student C
ib i U d
di f S d Learning
d
U.S. Diversity Outcome: examine causes and consequences of social intolerance in the U.S.
y Native American
Literature Course:
y A critical
iti l essay on
selected literature
y History Course in Race,
Class, and Gender
y An essay exam
y African American
Theatre Course:
y Performance
P f
off a
scene with
interpretation
annotations
y Inter-racial/ethnic
C
Communication
i ti C
Course:
y A persuasive speech
Contributions to Process
y Common indicators of
outcome based
learning:
y Students use evidence
to defend
arguments/perspectives
y Students articulate
p p
perspectives
in
meaningful ways
y Students use language
in a way
a that
demonstrates an
understanding of
intersectionality
y Students present
moral/ethical concerns
with an awareness of
historical/cultural
context
Process Guided by Questions:
TEACH:
y
y
1. What does a
specific outcome
mean in your
discipline?
Based on naturally
occurring outcome
reflective
assignments, what
do you value and
what is
challenging?
ASSESS:
y 2. What
common
themes/language/expec
tations and
understandings do we
share?
y How does that inform
our understanding of
student learning as
reflected in student
work?
Process Guided by Questions:
ANALYZE:
y
y
Independent of
disciplinary differences,
what indicators of
student learning
emerge from our
shared understanding?
g
How can we
characterize levels of
student
t d t performance
f
against those indicators
(for a particular
outcome)?
ADJUST:
y Applying “this rubric” to a
portfolio of
multidiscipline/multi genre
multidiscipline/multi-genre
assignments,
y What contribution can we
make to our understanding
of student learning in
General Education?
Community Committed to
Community Committed to Humanized Assessment
Common values l grounding outcomes
Tolerance for divergent practices
Enlightened U d t di Understanding of Student Learning
The Added Values
y Prizing Humanized Processes
y Pedagogy and Teaching at the Center
y Mirror of natural instructional processes
y Disciplinary Habits + Common Values
y Record of ongoing dialogue provides documentation
for accountability:
y
y
y
y
What should they know?
How do we see that?
What have we learned?
Wh t are we doing
What
d i about
b t th
that?
t?
T k H
E b dd d A
t
Take Home; Embedded Assessment: ASSESS
ANALYZE
TEACH
ADJUST
Fosters professional development:
y shifts unconscious competent to
conscious reflection while
purposefully “looking for”
misconceptions;
i
ti
y As a narrative , it “captures”
improvements for
accountability;
… thus, humanizing assessment.
Unconscious Competent
Conscious Reflection
Awareness of continuous improvement of student learning
References
y
Gallagher, J. J. 1999. Improving science teaching and student achievement through embedded
assessment. MSTA Journal. http://www.msta-mich.org/index.php/publications/journalArticle/20
y
Gallagher, J. J. 2007. Teaching for Understanding: A practical guide for middle and high school
teaching. Columbus, Ohio. Pearson, Merrill Prentice Hall.
y
Liggit, P. Humanizing Academic Accountability: Embedded Assessment Gets at the Heart of
Teaching and Learning. Spiraling Upwards: The Scholarship of Teaching and Learning. Ypsilanti,
MI Bruce
MI:
B
K.
K Nelson
N l
Faculty
F
lt Development
D
l
t Center,
C t Eastern
E t
Michigan
Mi hi
U
University.
i
it IIn press.
y
Lin Ding et. al. 2009. Are we asking the right questions? Validating clicker question sequences by
student interviews American Association of Physics Teachers.
y
Kuhle, Barry X. and Jessica M. Barber, and Adam S. Bristol. "Predicting Students'
Performance in Introductory Psychology from their Psychology Misconceptions." Journal of
Instructional Psychology. 36.2 (2009): 119-24.
y
Malde, Melissa. "Mapping the Structures of Resonance." Journal of Singing. 65.5
(2009): 521-9.
C
I f
Contact Info
Peggy Liggit, Ph.D.
Director of Academic Assessment
Office of Institutional Effectiveness
234 McKenny
Eastern Michigan University
Ypsilanti, MI 48197
Peggy.Liggit@emich.edu
734-487-0199 (direct)
734-487-8288 (Office)
734-487-8290 (FAX)
Visit our Assessment Website
Steven Pernecky, Ph.D.
Professor
Undergraduate
Undergraduate Coordinator/Associate Chair
Dept. of Chemistry
Eastern Michigan University
Ypsilanti, MI 48197
Phone: 734‐487‐0383
Fax: 734‐487‐1496
www.tinyurl.com/emuassess
Download