T 2005–2009 CLA INSTITUTIONAL REPOR Winston-Salem State University

advertisement
2005–2009 CLA INSTITUTIONAL REPORT
Winston-Salem State
University
Acknowledgements and Notes
We would like to thank the
Lumina Foundation for Education
for funding the fall 05 – spring 09
CLA longitudinal study.
We would also like to thank your
institution for participating.
This report presents the results of
your institution across the three
phases of this longitudinal study.
Additional summary findings from
of this study will be published in
the coming months.
Institutional Report
1 Purpose of the CLA (1)
2 Your Results (2-7)
3 Longitudinal Cohort (8-12)
4 Diagnostic Guidance (13)
5 Moving Forward (14)
Appendices
A Task Overview (16-19)
B Task Development (20)
C Scoring Criteria (21-23)
D Scoring Process (24-25)
E Scaling Procedures (26-27)
F Student Data File (28)
G CAE Board of Trustees and Officers (29)
Institutional Report
1: Purpose of the CLA
The Collegiate Learning Assessment
The continuous improvement model
The CLA uses detailed scoring
(CLA) offers an authentic approach
also requires multiple assessment
guides to precisely and reliably
to assessment and improvement
indicators beyond the CLA because
evaluate student responses. It
of teaching and learning in higher
no single test to benchmark student
also encourages institutions to
education. Over 400 institutions and
learning in higher education is
compare their student learning
180,000 students have participated
feasible or desirable.
results on the CLA with learning
to date. Growing commitment on
at other institutions and on other
the part of higher education to assess
This, however, does not mean certain
student learning makes this a good
skills judged to be important by most
time to review the distinguishing
faculty and administrators across
The signaling quality of the CLA
features of the CLA and how it
virtually all institutions cannot be
is important because institutions
connects to improving teaching and
measured; indeed, the higher order
need to benchmark (have a frame
learning on your campus.
skills the CLA focuses on fall into
of reference for) where they stand
this measurable category.
and how much progress their
The CLA is intended primarily to
assessments.
students have made relative to the
assist faculty, department chairs,
The CLA presents realistic problems
progress of students at other colleges.
school administrators and others
that require students to analyze
Otherwise, how do they know how
interested in programmatic change
complex materials. Several different
well they are doing?
to improve teaching and learning,
types of materials are used that vary
particularly with respect to
in relevance to the task, credibility,
Yet, the CLA is not about ranking
strengthening higher order skills.
and other characteristics. Students’
institutions. Rather, it is about
written responses to the task are
highlighting differences between
The CLA helps campuses follow a
graded to assess their abilities to
them that can lead to improvements
continuous improvement model that
think critically, reason analytically,
in teaching and learning.
positions faculty as central actors.
solve problems, and communicate
clearly and cogently.
CLA Education (described on
While the CLA is indeed an
assessment instrument, it is
page 14) does just that by focusing
The institution—not the student—is
deliberately designed to contribute
on curriculum and pedagogy and
the initial primary unit of analysis.
directly to the improvement of
the link between assessment and
The CLA is designed to measure
teaching and learning. In this respect
teaching and learning.
an institution’s contribution, or
it is in a league of its own.
value added, to the development of
these competencies, including the
effects of changes to curriculum and
pedagogy.
1
2: Your Results
Table 1 below provides the number of your students
with both Entering Academic Ability (EAA) scores
and CLA scores in different phases of the longitudinal
assessment.
EAA scores represent SAT Math + Verbal, ACT
Composite, or Scholastic Level Exam (SLE) scores on
the SAT scale.
Per sampling restrictions of Phase 1 of the assessment,
the sample of students throughout the report also have
EAA scores.
1
Number of your students with CLA scores in different longitudinal phases
Phase 1
and
Phase 2
Phase 2
and
Phase 3
Phase 1
and
Phase 3
Performance Task
15
8
41
Analytic Writing Task
4
2
32
Make-an-Argument
5
2
33
Critique-an-Argument
5
3
33
4
2
31
F 2005
F 2005
F 2005
S 2007
S 2007
S 2007
S 2009
S 2009
S 2009
Total Score
Phases referenced
in columns above
2
2: Your Results
For each longitudinal phase, Table 2 below presents
summary statistics for your students including counts,
means, 25th and 75th percentiles, and standard
deviations.
We also calculated these statistics across all students
and all schools. See Tables 7 and 8 on pages 10 and 11.
2
Summary statistics for your students participating in the longitudinal assessment
Number of
Students
25th
Percentile
Mean
Score
75th
Percentile
Standard
Deviation
Performance Task
186
875
983
1099
159
Analytic Writing Task
131
906
983
1055
124
Make-an-Argument
138
801
985
1084
154
Critique-an-Agument
138
869
976
1018
147
Total Score
129
923
994
1061
117
EAA Score
183
830
911
970
109
Number of
Students
25th
Percentile
Mean
Score
75th
Percentile
Standard
Deviation
Performance Task
15
910
1029
1155
151
Analytic Writing Task
5
906
951
980
83
Make-an-Argument
5
942
942
942
100
Critique-an-Agument
Phase 1
Phase 2
6
869
968
1018
122
Total Score
5
908
1014
1066
124
EAA Score
15
910
970
1050
125
Number of
Students
25th
Percentile
Mean
Score
75th
Percentile
Standard
Deviation
Performance Task
42
946
1058
1192
158
Analytic Writing Task
42
952
1045
1127
136
Make-an-Argument
42
911
1042
1163
186
Critique-an-Agument
42
905
1047
1198
158
Total Score
42
953
1052
1139
112
EAA Score
42
830
951
1050
137
Phase 3
3
2: Your Results
Table 3 provides summary statistical data
The larger the positive effect size, the
The last column shows the correlation
on your students across all three phases.
greater the improvement. Effect sizes
between students’ scores in different
We present these data in three groups to
greater than 0.50 are generally considered
phases. A high positive correlation
examine change across different phases.
“large.” Negative effect sizes would indicate
indicates that the students who scored
that the students scored higher during an
relatively highly in one phase (relative
The “effect size” column of Table
earlier phase. To place your performance
to their classmates) also tended to score
3 is particularly important. This
in context, Figure A on page 5 plots these
relatively highly in another phase. In other
column indicates how much change
effect sizes versus all other participating
words, the improvement in scores was fairly
occurred between two different phases.
institutions.
consistent across students.
3
Comparison of your students’ scores in Phase 1 (Fall 2005), Phase 2 (Spring 2007), and Phase 3 (Spring 2009)
Phase 1
Phase 2
Difference
Summary Statistics
Number of
Students
Mean
Score
Standard
Deviation
Mean
Score
Standard
Deviation
Mean
Score
Standard
Deviation
Effect
Size*
Mean Score
Correlation
Performance Task
15
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
Analytic Writing Task
4
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
Make-an-Argument
5
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
Critique-an-Agument
5
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
4
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
Number of
Students
Mean
Score
Standard
Deviation
Mean
Score
Standard
Deviation
Mean
Score
Standard
Deviation
Effect
Size
Mean Score
Correlation
Performance Task
8
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
Analytic Writing Task
2
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
Make-an-Argument
2
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
Critique-an-Agument
3
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
2
N/A
N/A
N/A
N/A
N/A
N/A
N/A
N/A
Number of
Students
Mean
Score
Standard
Deviation
Mean
Score
Standard
Deviation
Mean
Score
Performance Task
41
1005
163
1055
159
Analytic Writing Task
32
1000
116
1041
140
Make-an-Argument
33
1002
150
1039
Critique-an-Agument
33
986
152
31
1021
113
Phases 1 and 2
Total Score
Phase 2
Phases 2 and 3
Total Score
Phase 3
Phase 1
Phases 1 and 3
Total Score
Difference
Phase 3
Summary Statistics
Difference
Summary Statistics
Standard
Deviation
Effect
Size
Mean Score
Correlation
50
-4
0.31
0.27
41
24
0.35
0.22
194
37
44
0.25
0.13
1053
159
67
7
0.44
0.32
1045
117
24
4
0.21
0.42
* A different method was used previously to calculate effect sizes, so these results differ slightly from those in the 2005–2007 Institutional Report.
Previously, we divided mean differences between phases by the CLA score standard deviation of all participating students in the earlier phase,
which would tend to underestimate effect sizes. Now, we divide by the standard deviation of your participating students in the earlier phase.
4
2: Your Results
Mean Score Changes between Phase 1, Phase 2, and Phase 3 by CLA Measure
1
A
Effect Size Box and Whisker Plots
CLA Measure
Between Phases
Effect Size
-.75
-.50
-.25
0
.25
.50
.75
1.00
1.25
1.50
1.75
-.75
-.50
-.25
0
.25
.50
.75
1.00
1.25
1.50
1.75
Performance Task
1
Analytic Writing Task
2
Make-an-Argument
Critique-an-Argument
Total Score
Performance Task
Analytic Writing Task
2
3
Make-an-Argument
Critique-an-Argument
Total Score
Performance Task
Analytic Writing Task
1
3
Make-an-Argument
Critique-an-Argument
Total Score
Key for Phases
When
Phase
The Figure A
1 “box and whisker” plots show effect size distributions across
CLA measures for longitudinal schools. The top cluster displays effect sizes
between Phase 1 (fall 2005 first-year students) and Phase 2 (spring 2007
rising juniors); the middle cluster repeats this for Phases 2 and 3 (spring 2009
seniors); and the bottom cluster does the same for Phases 1 and 3.
Who
1
Fall 2005
First-year
2
Spring 2007
Rising Junior
3
Spring 2009
Senior
In each plot, the extreme left hand vertical bar shows the 5th percentile. The
“box” itself shows the 25th (left face), 50th (internal vertical line), and 75th
(right face) percentile points. The extreme right hand vertical line shows the
95th percentile. The horizontal x-axis shows the effect size (in standard
deviation units).
Key for Box and Whisker Plots
5th
50th
Your School
Effect sizes are one way to measure change between two time points. Effect
sizes were calculated at a school by taking the difference in mean (or average)
CLA scores of the same students from two different phases (e.g., Phase 1 and
Phase 3). This difference is then divided by the spread of scores (standard
deviation) from the first phase (e.g., Phase 1) to produce an effect size.
95th
Circles
25th
identify effect sizes for your school (see Table 53 for values).
75th
5
2: Your Results
Table 4 below indicates whether students
Colleges with scores between -1.00 and
Figure B on the next page displays your
scored well above, above, at, below, or well
+1.00 standard errors from their expected
deviation scores in a scatterplot and
below what would be expected given their
scores are categorized as At Expected.
documents the regression equations from
scores as first-year students. Results are
Institutions with scores greater than one
which they were derived.
expressed in the form of value-added scores
standard error (but less than two standard
that correspond to standard errors.
errors) from their expected scores are in
The two regression equations (Phase 1 to
the Above Expected or Below Expected
Phase 2 in blue; Phase 1 to Phase 3 in red)
categories (depending on the direction
are each based on data from all institutions
of the deviation). The schools with scores
where at least 25 students received CLA
greater than two standard errors from
total scores in both phases in the model.
their expected scores are in the Well
Therefore, the sample of institutions is not
Above Expected or Well Below Expected
identical across the testing phases reported
categories.
here, but the use of inclusive samples
maximizes the strength of each equation.
4
Value added results for your school
Deviation
Performance
Deviation
Performance
Score
Level
Score
Level
Performance Task
N/A
N/A
-0.2
At
Analytic Writing Task
N/A
N/A
-0.8
At
Make-an-Argument
N/A
N/A
-0.6
At
Critique-an-Argument
N/A
N/A
-0.8
At
N/A
N/A
-0.9
At
F 2005
F 2005
F 2005
F 2005
S 2007
S 2007
S 2007
S 2007
S 2009
S 2009
S 2009
S 2009
Total Score
Phases referenced
in columns above
6
2: Your Results
B
Relationship between Spring 2009, Spring 2007, and Fall 2005 CLA Performance
1500
First-year (Phase 1) to Rising Junior (Phase 2)
First-year (Phase 1) to Senior (Phase 3)
CLA Total Score (Phase 2 or Phase 3)
1400
1300
1200
1100
1000
900
Regression
Intercept
Slope
R-square
800
800
900
1000
1100
1200
1300
55 -85
0.97 1.15
0.73 0.80
1400
1500
CLA Total Score (Phase 1)
The vertical y-axis in Figure B above
The horizontal x-axis shows a school’s
A school’s data point is above its
shows a school’s mean CLA Total score
mean CLA Total score for Phase 1.
corresponding line if its students in
for both Phase 2 (in blue) and Phase 3
Phase 2 (or Phase 3) performed better
(in red).
on the CLA than what would be
expected relative to their Phase 1 scores.
7
3: Longitudinal Cohort
During the spring 2009 testing cycle, 31 institutions
tested a sufficient number of students to provide the
school-level analyses and results presented in this
report.
Table 5 shows CLA longitudinal schools grouped
by Basic Carnegie Classification. The spread of
schools differs slightly from that of the 1,713 fouryear institutions across the nation, with Doctorategranting Universities constituting a higher percentage
among CLA schools than nationally. Accordingly,
representation among both Master’s Colleges and
Universities as well as Baccalaureate Colleges is lower
among CLA longitudinal schools than it is nationally.
5
Nation
Carnegie Classification
CLA
Number
Percentage
Number
Percentage
Doctorate-granting Universities
282
16%
12
39%
Master’s Colleges and Universities
664
39%
9
29%
Baccalaureate Colleges
767
45%
10
32%
1,713
31
Source: Carnegie Foundation for the Advancement of Teaching, Carnegie Classifications
Data File, June 11, 2008.
8
3: Longitudinal Cohort
Table 6 provides comparative statistics for colleges and
universities across the nation and CLA longitudinal
schools in terms of some important characteristics.
These statistics suggest that CLA longitudinal
schools are fairly representative of institutions
nationally. Percentage public, percentage HBCU, and
undergraduate student body size are exceptions.
6
School Characteristic
Nation
CLA
Percentage public
37%
52%
Percentage Historically Black College or University (HBCU)
5%
13%
Mean percentage of undergraduates receiving Pell grants
34%
31%
Mean four-year graduation rate
36%
40%
Mean six-year graduation rate
52%
61%
Mean first-year retention rate
73%
82%
Mean Barron’s selectivity rating
3.4
3.8
1067
1106
Mean estimated median SAT score
Mean number of FTE undergraduate students (rounded)
Mean student-related expenditures per FTE student (rounded)
4,320
9,040
$12,365
$14,180
Source: College Results Online dataset, managed by and obtained with permission from the Education Trust, covers
most 4-year Title IV-eligible higher-education institutions in the United States. Data were constructed from IPEDS
and other sources. Because all schools did not report on every measure in the table, the averages and percentages
may be based on slightly different denominators.
9
3: Longitudinal Cohort
For each longitudinal phase, Table 7 below presents
summary statistics for all students including counts,
means, 25th and 75th percentiles, and standard
deviations.
7
Student
Count
25th
Percentile
Mean
Score
75th
Percentile
Standard
Deviation
Performance Task
11437
971
1087
1209
189
Analytic Writing Task
9221
977
1085
1196
165
Make-an-Argument
9879
942
1080
1225
188
Critique-an-Agument
9627
869
1079
1167
186
Total Score
9168
990
1096
1198
149
EAA Score
11360
950
1093
1230
195
Student
Count
25th
Percentile
Mean
Score
75th
Percentile
Standard
Deviation
Performance Task
3327
1012
1157
1277
207
Analytic Writing Task
3161
980
1119
1267
161
Make-an-Argument
3227
942
1114
1225
181
Critique-an-Agument
3202
1018
1120
1316
180
Total Score
3141
1025
1141
1249
160
EAA Score
3329
1010
1134
1270
188
Student
Count
25th
Percentile
Mean
Score
75th
Percentile
Standard
Deviation
Performance Task
2374
1093
1216
1337
194
Analytic Writing Task
2308
1091
1221
1363
186
Make-an-Argument
2318
1051
1199
1354
214
Critique-an-Agument
Phase 1 (student level)
Phase 2 (student level)
Phase 3 (student level)
2317
1083
1240
1393
210
Total Score
2290
1115
1222
1339
163
EAA Score
2380
1030
1153
1300
184
10
3: Longitudinal Cohort
For each longitudinal phase, Table 8 below presents
summary statistics for all schools including counts,
means, 25th and 75th percentiles, and standard
deviations.
8
School
Count
25th
Percentile
Mean
Score
75th
Percentile
Standard
Deviation
Performance Task
48
1027
1094
1168
104
Analytic Writing Task
47
1021
1089
1164
98
Make-an-Argument
48
1018
1081
1159
103
Critique-an-Agument
48
1006
1084
1153
103
Total Score
47
1038
1100
1170
93
EAA Score
48
1011
1104
1209
152
School
Count
25th
Percentile
Mean
Score
75th
Percentile
Standard
Deviation
Performance Task
32
1068
1147
1194
103
Analytic Writing Task
32
1057
1114
1170
93
Make-an-Argument
32
1062
1111
1166
94
Critique-an-Agument
32
1051
1114
1160
94
Total Score
32
1073
1132
1185
96
EAA Score
32
1044
1116
1193
129
School
Count
25th
Percentile
Mean
Score
75th
Percentile
Standard
Deviation
Performance Task
31
1126
1193
1256
107
Analytic Writing Task
30
1120
1201
1267
109
Make-an-Argument
30
1095
1181
1251
109
Critique-an-Agument
Phase 1 (school level)
Phase 2 (school level)
Phase 3 (school level)
30
1152
1219
1281
111
Total Score
30
1138
1202
1267
104
EAA Score
32
1049
1126
1217
137
11
3: Longitudinal Cohort
The institutions listed here in alphabetical order
agreed to be identified as participating schools. All
participated in Phase 3 of the longitudinal assessment,
but some may not have tested enough students to be
included in all comparative analyses.
Auburn University
Belmont University
Bowling Green State University
California State Polytechnic University-Pomona
California State University-Northridge
Carleton College
Central Michigan University
City University of New York City College
City University of New York Herbert H. Lehman College
Cleveland State University
Colorado College
Earlham College
Fayetteville State University
Grand Valley State University
Kalamazoo College
Loyola University of Chicago
Macalester College
Morgan State University
North Carolina Agricultural and Technical State University
Northern Arizona University
Pace University
Saint Xavier University
Spelman College
St. Olaf College
Syracuse University
The George Washington University
The Ohio State University
University of California, Riverside
University of Charleston
University of North Carolina at Charlotte
University of Saint Thomas
University of the Pacific
University of Wyoming
Wagner College
Winston-Salem State University
Wofford College
12
4: Diagnostic Guidance
CLA results operate as a signaling tool of overall institutional performance on tasks that measure
higher order skills holistically. However, the three types of CLA tasks—Performance, Make-anArgument, and Critique-an-Argument—differ slightly in the combination of skills necessary to
perform well.
Indeed, some schools score significantly lower on one type than on another. Examining
performance across CLA task types can serve as an initial diagnostic exercise. Specifically, cases of
performance Well Below or Below Expected on a particular task type indicate that students are
not demonstrating the expected level of skill at analyzing complex, realistic scenarios; writing a
persuasive, analytic essay to support a position on an issue; and/or critiquing written arguments.
Performance Task
Make-an-Argument
Critique-an-Argument
Analyzing
complex, realistic scenarios
Writing
a persuasive, analytic essay
Critiquing
written arguments
Synthesizing information from multiple
Establishing a thesis or a position on an
Identifying a variety of logical flaws or
sources; recognizing conflicting
issue; maintaining the thesis throughout
fallacies in a specific argument; explaining
evidence, weighing the credibility of
the essay; supporting the thesis with
how or why the logical flaws affect
different sources of evidence; identifying
relevant and persuasive examples (e.g.,
the conclusions in that argument; and
logical fallacies, interpreting data,
from personal experience, history,
presenting their critique in a written
tables, and figures correctly; drawing
art, literature, pop culture, or current
response that is a grammatically correct,
reasonable and logical inferences from
events); anticipating and countering
organized, well-developed, logically
the available information; developing
opposing arguments to the position,
sound, and neutral in tone.
sound conclusions based on all available
fully developing ideas, examples, and
evidence; and utilizing the most relevant
arguments; crafting an overall response
and credible evidence available to justify
that generates interest, provokes thought,
their conclusion.
and persuades the reader; organizing the
structure of the essay (e.g., paragraphing,
ordering of ideas and sentences within
paragraphs); employing transitions and
varied sentence structure to maintain
the flow of the argument; and utilizing
sophisticated grammar and vocabulary.
13
5: Moving Forward
We encourage institutions to examine
CLA Education focuses on curriculum
Through the steps noted above we
performance across CLA tasks and
and pedagogy and embraces the crucial
encourage institutions to move toward
communicate results across campus,
role that faculty play in the process of
a continuous system of improvement in
link student-level CLA results with
assessment.
teaching and learning stimulated by the
other data sources, pursue in-depth
CLA. Without your contributions, the
sampling, stay informed through the
The flagship program of CLA
CLA would not be on the exciting path
CLA Spotlight, and participate in CLA
Education is the Performance Task
that it is today. We look forward to your
Education offerings.
Academy, which shifts the focus from
continued involvement!
general assessment to the course-level
Student-level CLA results are provided
work of faculty. The Performance Task
for you to link with other data sources
Academy provides an opportunity for
(e.g., course-taking patterns, grades,
faculty members to learn to diagnose
portfolios, student satisfaction and
their individual students’ work and to
engagement, major-specific tests, etc.).
receive guidance in creating their own
performance tasks, which are designed
These internal analyses can help you
to supplement the educational reform
generate hypotheses for additional
movement toward a case and problem
research, which you can pursue through
approach in learning and teaching.
CLA in-depth sampling in experimental
areas (e.g., programs or colleges within
A CLA Education web site also has
your campus) in subsequent years or
been formed as a clearing house for
simultaneously.
performance tasks developed by
faculty. For more information, visit
We welcome and encourage your
www.claintheclassroom.org, or contact
participation in the CLA Spotlight—a
Director of CLA Education, Dr. Marc
series of free informational web
Chun at mchun@cae.org.
conferences. Each CLA Spotlight
features campuses doing promising work
using the CLA, guest-speakers from the
larger world of assessment, and/or CLA
staff members who provide updates or
insights to CLA-related programs and
projects.
14
Appendices
A: Task Overview
Introduction
The CLA is comprised of three types of prompts
within two types of task: the Performance Task
and the Analytic Writing Task. Your students
were asked to take all three prompts. The Analytic
Writing Task includes a pair of prompts called
Make-an-Argument and Critique-an-Argument.
The CLA uses direct measures of skills in which
students perform cognitively demanding tasks
from which quality of response is scored. All CLA
measures are administered online and contain
open-ended prompts that require constructed
responses. There are no multiple-choice questions.
The CLA tasks require students to integrate critical
thinking, analytic reasoning, problem solving,
and written communication skills. The holistic
integration of these skills on the CLA tasks mirrors
the requirements of serious thinking and writing
tasks faced in life outside of the classroom.
16
A: Task Overview
Performance Task
Each Performance Task requires
No two Performance Tasks assess
Performance Tasks often require
students to use an integrated set of
the exact same combination of skills.
students to marshal evidence from
critical thinking, analytic reasoning,
Some ask students to identify and then
different sources; distinguish rational
problem solving, and written
compare and contrast the strengths and
from emotional arguments and fact
communication skills to answer
limitations of alternative hypotheses,
from opinion; understand data in tables
several open-ended questions about a
points of view, courses of action, etc. To
and figures; deal with inadequate,
hypothetical but realistic situation. In
perform these and other tasks, students
ambiguous, and/or conflicting
addition to directions and questions,
may have to weigh different types of
information; spot deception and holes
each Performance Task also has its
evidence, evaluate the credibility of
in the arguments made by others;
own document library that includes a
various documents, spot possible bias,
recognize information that is and is not
range of information sources, such as
and identify questionable or critical
relevant to the task at hand; identify
letters, memos, summaries of research
assumptions.
additional information that would help
to resolve issues; and weigh, organize,
reports, newspaper articles, maps,
photographs, diagrams, tables, charts,
Performance Tasks also may ask
and synthesize information from several
and interview notes or transcripts.
students to suggest or select a course
sources.
Students are instructed to use these
of action to resolve conflicting or
materials in preparing their answers to
competing strategies and then provide
the Performance Task’s questions within
a rationale for that decision, including
the allotted 90 minutes.
why it is likely to be better than one or
more other approaches. For example,
The first portion of each Performance
students may be asked to anticipate
Task contains general instructions and
potential difficulties or hazards that are
introductory material. The student is
associated with different ways of dealing
then presented with a split screen. On
with a problem, including the likely
the right side of the screen is a list of the
short- and long-term consequences and
materials in the Document Library. The
implications of these strategies. Students
student selects a particular document
may then be asked to suggest and
to view by using a pull-down menu. On
defend one or more of these approaches.
the left side of the screen are a question
Alternatively, students may be asked
and a response box. There is no limit
to review a collection of materials or
on how much a student can type. Upon
a set of options, analyze and organize
completing a question, students then
them on multiple dimensions, and then
select the next question in the queue.
defend that organization.
17
A: Task Overview
Analytic Writing Task
Make-an-Argument
Critique-an-Argument
Students write answers to two types of
A “Make-an-Argument” prompt
A “Critique-an-Argument” prompt
essay prompts, namely: a “Make-an-
typically presents an opinion on some
asks students, in 30 minutes, to critique
Argument” question that asks them to
issue and asks students to write, in 45
an argument by discussing how well
support or reject a position on some
minutes, a persuasive, analytic essay to
reasoned they find it to be (rather than
issue; and a “Critique-an-Argument”
support a position on the issue. Key
simply agreeing or disagreeing with the
question that asks them to evaluate
elements include: establishing a thesis
position presented). Key elements of
the validity of an argument made
or a position on an issue; maintaining
the essay include: identifying a variety
by someone else. Both of these tasks
the thesis throughout the essay;
of logical flaws or fallacies in a specific
measure a student’s skill in articulating
supporting the thesis with relevant and
argument; explaining how or why the
complex ideas, examining claims and
persuasive examples (e.g., from personal
logical flaws affect the conclusions
evidence, supporting ideas with relevant
experience, history, art, literature, pop
in that argument; and presenting a
reasons and examples, sustaining a
culture, or current events); anticipating
critique in a written response that is a
coherent discussion, and using standard
and countering opposing arguments
grammatically correct, organized, well-
written English.
to the position, fully developing ideas,
developed, logically sound, and neutral
examples, and arguments; crafting an
in tone.
overall response that generates interest,
provokes thought, and persuades the
reader; organizing the structure of the
essay (e.g., paragraphing, ordering of
ideas and sentences within paragraphs);
employing transitions and varied
sentence structure to maintain the
flow of the argument; and utilizing
sophisticated grammar and vocabulary.
18
A: Task Overview
Example Performance Task
Example Document Library
You advise Pat Williams, the president
‚‚ Newspaper article about the accident
of DynaTech, a company that makes
‚‚ Federal Accident Report on in-flight
breakups in single-engine planes
precision electronic instruments and
navigational equipment. Sally Evans,
a member of DynaTech’s sales force,
recommended that DynaTech buy a
small private plane (a SwiftAir 235)
that she and other members of the
sales force could use to visit customers.
Pat was about to approve the purchase
when there was an accident involving
a SwiftAir 235. Your document library
Example Questions
‚‚ Do the available data tend to support
or refute the claim that the type of
wing on the SwiftAir 235 leads to
more in-flight breakups?
‚‚ Internal Correspondence (Pat’s e-mail
to you and Sally’s e-mail to Pat)
‚‚ What is the basis for your conclusion?
‚‚ Charts relating to SwiftAir’s
performance characteristics
‚‚ What other factors might have
contributed to the accident and
should be taken into account?
‚‚ Excerpt from magazine article
comparing SwiftAir 235 to similar
planes
‚‚ Pictures and descriptions of SwiftAir
Models 180 and 235
‚‚ What is your preliminary
recommendation about whether
or not DynaTech should buy the
plane and what is the basis for this
recommendation?
contains the following materials:
Example Make-an-Argument
Example Critique-an-Argument
There is no such thing as “truth” in the
A well-­respected professional journal
sample group were more likely to be
media. The one true thing about the
with a readership that includes
overweight––relative to the national
information media is that it exists only
elementary school principals recently
average. Based on this study, the
to entertain.
published the results of a t­ wo-­year
principal of Jones Elementary School
study on childhood obesity. (Obese
decided to confront her school’s obesity
individuals are usually considered to
problem by opposing any fast food
be those who are 20 percent above
restaurant openings near her school.
their recommended weight for
height and age.) This study sampled
50 schoolchildren, ages 5-11, from
Smith Elementary School. A fast food
restaurant opened near the school just
before the study began. After two years,
students who remained in the
19
B: Task Development
Iterative Development Process
A team of researchers and writers
While developing a Performance Task,
After several rounds of revision, the
generate ideas for Make-an-Argument
a list of the intended content from each
most promising of the Performance
and Critique-an-Argument prompts,
document is established and revised.
Tasks and the Make-an-Argument
and Performance Task storylines, and
This list is used to ensure that each piece
and Critique-an-Argument prompts
then contribute to the development
of information is clearly reflected in the
are selected for pre-piloting. Student
and revision of the prompts and
document and/or across documents,
responses from the pilot test are
Performance Task documents.
and to ensure that no additional pieces
examined to identify what pieces
of information are embedded in the
of information are unintentionally
For Analytic Writing Tasks, multiple
document that were not intended. This
ambiguous, what pieces of information
prompts are generated, revised and
list serves as a draft starting point for
in the documents should be removed,
pre-piloted, and those prompts that
the analytic scoring items used in the
etc. After revision and additional pre-
elicit good critical thinking and writing
Performance Task scoring rubrics.
piloting, the best functioning tasks (i.e.,
those that elicit the intended types and
responses during pre-piloting are further
revised and submitted to more extensive
During revision, information is either
ranges of student responses) are selected
piloting.
added to documents or removed from
for full piloting.
documents to ensure that students could
During the development of
arrive at approximately three or four
During piloting, students complete
Performance Tasks, care is taken to
different conclusions based on a variety
both an operational task and one of the
ensure that sufficient information is
of evidence to back up each conclusion.
new tasks. At this point, draft scoring
provided to permit multiple reasonable
Typically, some conclusions are designed
rubrics are revised and tested in grading
solutions to the issues present in
to be supported better than others.
the pilot responses, and final revisions
are made to the tasks to ensure that the
the Performance Task. Documents
are crafted such that information is
Questions for the Performance Task
task is eliciting the types of responses
presented in multiple formats (e.g.,
are also drafted and revised during the
intended.
tables, figures, news articles, editorials,
development of the documents. The
letters, etc.).
questions are designed such that the
initial questions prompt the student
to read and attend to multiple sources
of information in the documents, and
later questions require the student to
evaluate the documents and then use
their analysis to draw conclusions and
justify those conclusions.
20
C: Scoring Criteria
Introduction
Assessing Critical Thinking,
Analytic Reasoning and
Problem Solving
Assessing Writing
This section summarizes the
Applied in combination, critical
Analytic writing skills invariably
types of questions addressed
thinking, analytic reasoning
depend on clarity of thought.
by CLA scoring of all task
and problem solving skills are
Therefore, analytic writing
types. Because each CLA
required to perform well on
and critical thinking, analytic
task and their scoring rubrics
CLA tasks. We define these
reasoning, and problem
differ, not every item listed is
skills as how well students can
solving are related skills sets.
applicable to every task. The
evaluate and analyze source
The CLA measures critical
tasks cover different aspects
information, and subsequently
thinking performance by asking
of critical thinking, analytic
to draw conclusions and
students to explain in writing
reasoning, problem solving, and
present an argument based
their rationale for various
writing and in doing so can, in
upon that analysis. In scoring,
conclusions. In doing so, their
combination, better assess the
we specifically consider the
performance is dependent
entire domain of performance.
following items to be important
on both writing and critical
aspects of these skills.
thinking as integrated rather
See page 22 for detail.
than separate skills. We evaluate
writing performance using
holistic scores that consider
several aspects of writing
depending on the task. The
following are illustrations of the
types of questions we address in
scoring writing on the various
tasks.
See page 23 for detail.
21
C: Scoring Criteria
Assessing Critical Thinking,
Analytic Reasoning and
Problem Solving
Evaluation of evidence
Analysis and synthesis of evidence
How well does the student assess the quality and relevance
How well does the student analyze and synthesize data and
of evidence, including:
information, including:
‚‚ Determining what information is or is not pertinent to
the task at hand
‚‚ Presenting his/her own analysis of the data or
information (rather than “as is”)
‚‚ Distinguishing between rational claims and emotional
ones, fact from opinion
‚‚ Committing or failing to recognize logical flaws (e.g.,
distinguishing correlation from causation)
‚‚ Recognizing the ways in which the evidence might be
limited or compromised
‚‚ Breaking down the evidence into its component parts;
‚‚ Spotting deception and holes in the arguments of others
‚‚ Considering all sources of evidence
‚‚ Drawing connections between discrete sources of data
and information
‚‚ Attending to contradictory, inadequate or ambiguous
information
Drawing conclusions
Acknowledging alternative explanations/viewpoints
How well does the student form a conclusion from their
How well does the student acknowledge additional
analysis, including:
perspectives and consider other options, including:
‚‚ Constructing cogent arguments rooted in data/
information rather than speculation/opinion
‚‚ Recognizing that the problem is complex with no clear
answer
‚‚ Selecting the strongest set of supporting data
‚‚ Proposing other options and weighing them in the
decision
‚‚ Prioritizing components of the argument
‚‚ Avoiding overstated or understated conclusions
‚‚ Identifying holes in the evidence and subsequently
suggesting additional information that might resolve the
issue
‚‚ Considering all stakeholders or affected parties in
suggesting a course of action
‚‚ Qualifying responses and acknowledging the need
for additional information in making an absolute
determination
22
C: Scoring Criteria
Interest
How well does the student maintain the reader’s interest?
Does the...
‚‚ Student use creative and engaging examples or
descriptions
Assessing Writing
‚‚ Structure, syntax and organization add to the interest of
their writing
‚‚ Student use colorful but relevant metaphors, similes, etc.
‚‚ Writing engage the reader
‚‚ Writing leave the reader thinking
Presentation
Development
How clear and concise is the argument? Does the student…
How effective is the structure? Does the student…
‚‚ Clearly articulate the argument and the context for that
argument
‚‚ Correctly and precisely use evidence to defend the
argument
‚‚ Comprehensibly and coherently present evidence
‚‚ Logically and cohesively organize the argument
‚‚ Avoid extraneous elements in the argument’s
development
‚‚ Present evidence in an order that contributes to a
persuasive and coherent argument
Persuasiveness
Mechanics
How well does the student defend the argument? Does the
What is the quality of the student’s writing?
student…
‚‚ Effectively present evidence in support of the argument
‚‚ Draw thoroughly and extensively from the available
range of evidence
‚‚ Analyze the evidence in addition to simply presenting it
‚‚ Consider counterarguments and address weaknesses in
his/her own argument
‚‚ Is vocabulary and punctuation used correctly
‚‚ Is the student’s understanding of grammar strong
‚‚ Is the sentence structure basic, or more complex and
creative
‚‚ Does the student use proper transitions
‚‚ Are the paragraphs structured logically and effectively
23
D: Scoring Process
Score Sheet
There are two types of items that appear
Performance Task holistic items are
Make-an-Argument score sheets contain
on a CLA score sheet: analytic and
scored on four or seven-point scales
only holistic items scored on four or
holistic. Analytic scoring items are
(i.e., 1-4 or 1-7). There are multiple
seven-point scales (i.e., 1-4 or 1-7). The
particular to each prompt and holistic
holistic items per Performance Task that
holistic items include ratings for various
items refer to general dimensions, such
require graders to provide an evaluation
aspects of writing (e.g., organization,
as evaluation of evidence, drawing
of different aspects of critical thinking
mechanics, etc.) and critical thinking
conclusions, acknowledging alternative
and reasoning in the student responses.
(e.g., reasoning and logic, sophistication
explanations and viewpoints, and overall
These holistic items include areas
and depth of treatment of the issues
writing. We compute raw scores for each
such as the student’s use of the most
raised in the prompt) as well as two
task by adding up all points on all items
relevant information in the Performance
overall assessments of writing and
(i.e., calculating a unit-weighted sum).
Task, their recognition of strengths
critical thinking.
and weaknesses of various pieces of
Performance Task scoring is tailored
information, overall critical thinking,
For all task types, blank responses or
to each specific prompt and includes
and overall writing.
responses that are entirely unrelated to
a combination of both holistic and
the task (e.g., writing about what they
analytic scoring items. Though there
Critique-an-Argument score sheets also
had for breakfast) are assigned a 0 and
are many types of analytic items on the
include a combination of analytic and
are flagged for removal from the school-
Performance Task score sheets, the most
holistic scores. Critique-an-Argument
level results.
common represent a list of the possible
analytic items are a list of possible
pieces of information a student could
critiques of the argument presented in
or should raise in their response. These
the prompt. In addition, a few holistic
cover the information presented in the
items are used to rate the overall quality,
Performance Task documents as well
critical thinking and writing over the
as information that can be deduced
entire response.
from comparing information across
documents. The analytic items are
generally given a score of 0 if the student
did not use the information in their
response, or 1 if they did. The number
of analytic items varies by prompt.
24
D: Scoring Process
Scoring Procedure
All scorer candidates undergo rigorous training in
order to become certified CLA scorers. Training
includes an orientation to the prompt and score sheet,
instruction on how to evaluate the scoring items,
repeated practice grading a wide range of student
responses, and extensive feedback and discussion after
scoring each response.
After participating in training, scorers complete a
reliability check where they score the same set of
student responses. Scorers with low agreement or
reliability (determined by comparisons of raw score
means, standard deviations and correlations among the
scorers) are either further coached or removed from
scoring.
In fall 2008 and spring 2009, a combination of
machine and human scoring was used for the Analytic
Writing Task.
The CLA utilizes Pearson Knowledge Technology’s
Intelligent Essay Assessor program for evaluating
responses to the Make-an-Argument and Critique-anArgument prompts.
The machine scoring engine was developed and tested
using scores from a broad range of responses that were
previously scored by humans (often double scored).
In some cases the automated scoring engine is unable
to score off-topic or abnormally short/long responses.
These student responses are scored by humans.
25
E: Scaling Procedures
To facilitate reporting results across
schools, ACT scores were converted
Standard ACT to SAT
Conversion Table
(using the ACT-SAT crosswalk to the
right) to the scale of measurement used
ACT to SAT
to report SAT scores.
36
1600
35
1580
34
1520
33
1470
32
1420
31
1380
open admission schools), we make
30
1340
available the Scholastic Level Exam
29
1300
28
1260
27
1220
26
1180
25
1140
and Recentered SAT I Sum Scores” by
were converted to SAT scores using data
24
1110
N.J. Dorans, C.F. Lyu, M. Pommerich,
from 1,148 students participating in
23
1070
and W.M. Houston (1997), College and
spring 2006 that had both SAT and SLE
22
1030
21
990
University, 73, 24-31; “Concordance
20
950
19
910
18
870
N.J. Dorans, Research Notes (RN-07),
17
830
College Entrance Examination Board:
16
780
1999; “Correspondences between ACT
15
740
14
680
and SAT I Scores” by N.J. Dorans, College
13
620
12
560
Entrance Examination Board: 1999; ETS
11
500
Research Report 99-2, Educational Testing
For institutions where a majority of
students did not have ACT or SAT
scores (e.g., two-year institutions and
(SLE), a short-form cognitive ability
measure, as part of the CLA. The SLE is
produced by Wonderlic, Inc. SLE scores
scores. These converted scores (both
ACT to SAT and SLE to SAT) are
referred to simply as entering academic
ability (EAA) scores.
Students receive a single score on a
CLA task because each task assesses
an integrated set of critical thinking,
analytic reasoning, problem solving, and
written communication skills.
Sources:
“Concordance Between ACT Assessment
between SAT I and ACT Scores for
Individual Students” by D. Schneider and
Board Research Report 99-1, College
Service: 1999.
26
E: Scaling Procedures
Each Performance Task and Analytic
A linear scale transformation is used
the same as the lowest EAA score of any
Writing Task has a unique scoring
to convert reader-assigned raw scores
freshman who took that task. On very
rubric, and the maximum number of
to scale scores. This process results
rare occasions, a student may achieve an
reader assigned raw score points differs
in a scale score distribution with the
exceptionally high or low raw score (i.e.,
across tasks. Consequently, a given
same mean and standard deviation as
well above or below the other students
reader-assigned raw score, such as 15
the Entering Academic Ability (EAA)
taking that task). When this occurs,
points, may be a relatively high score on
scores of the freshmen who took that
it results in assigning a student a scale
one task but a low score on another task.
measure. This type of scaling preserves
score that is outside of the normal EAA
the shape of the raw score distribution
range. Prior to the spring of 2007, scores
To adjust for such differences, reader-
and maintains the relative standing of
were capped at 1600. Capping was
assigned raw scores on the different
students. For example, the student with
discontinued starting in fall 2007.
tasks are converted to a common scale
the highest raw score on a task will also
of measurement. This process results
have the highest scale score on that task,
In the past, CAE revised its scaling
in scale scores that reflect comparable
the student with the next highest raw
equations each fall. However, many
levels of proficiency across tasks. For
score will be assigned the next highest
institutions would like to make year-
example, a given CLA scale score
scale score, and so on.
to-year comparisons (i.e., as opposed
indicates approximately the same
to just fall to spring). To facilitate this
percentile rank regardless of the task
This type of scaling generally results in
activity, in fall 2007 CAE began using
on which it was earned. This feature of
the highest raw score earned on a task
the same scaling equations it developed
the CLA scale scores allows combining
receiving a scale score of approximately
for the fall 2006 administration and
scores from different tasks to compute
the same value as the maximum EAA
has done so for new tasks introduced
a school’s mean scale score for each task
score of any freshman who took that
since then. As a result of this policy, a
type as well as a total average scale score
task. Similarly, the lowest raw score
given raw score on a task will receive the
across types.
earned on a task would be assigned a
same scale score regardless of when the
scale score value that is approximately
student took the task.
27
F: Student Data File
We provide a CLA Student Data
We provide student-level information
Student-level scores are not designed to
File, which includes variables across
to link with other data you collect (e.g.,
be diagnostic at the individual level and
three categories: student self-reported
from NSSE, CIRP, portfolios, local
should be considered as only one piece of
information from their CLA on-line
assessments, course-taking patterns,
evidence about a student’s skills.
profile; CLA scores and identifiers; and
participation in specialized programs,
information provided/verified by the
etc.) to help you hypothesize about
registrar.
campus-specific factors related to overall
institutional performance.
Self-Reported Data
‚‚ Student ID, E-mail address,
and Name (first, middle
initial, last)
‚‚ Age
‚‚ Gender
‚‚ Race/Ethnicity
‚‚ Primary and Secondary
Academic Major (34
categories)
CLA Scores and Identifiers
‚‚ CLA scores for Performance Task,
Analytic Writing Task, Make-anArgument, Critique-an-Argument,
and Total CLA Score (depending
on the number of tasks taken and
completeness of responses):
‚‚ Transfer Student Status
‚‚ Program ID and Name (for
classification of students into
different schools, majors,
studies, programs, etc.)
CLA scale scores;
‚‚ Entering Academic Ability
(EAA) Score
--
Student Performance Level categories (i.e., well below expected,
below expected, at expected,
above expected, well above
expected) if CLA scale score and
entering academic ability (EAA)
scores are available;
‚‚ SAT I - Math
Percentile Rank in the CLA
(among students in the same class
year; based on scale score); and
‚‚ SAT I - Writing (Essay subscore)
‚‚ English as primary language
--
‚‚ Attended school as Freshman,
Sophomore, Junior, Senior
‚‚ CLA Local Survey Responses
‚‚ Class Standing
--
‚‚ Field of Study (6 categories;
based on primary academic
major)
‚‚ Total years at school
Registrar Data
--
Percentile Rank at School
(among students in the same class
year; based on scale score).
‚‚ SAT I - Verbal / Critical
Reading
‚‚ SAT Total (Math + Verbal)
‚‚ SAT I - Writing
‚‚ SAT I - Writing (Multiplechoice subscore)
‚‚ ACT - Composite
‚‚ Unique CLA numeric identifiers
‚‚ ACT - English
‚‚ Year, Administration (Fall or Spring),
Type of Test (90 or 180-minute), Date
of test
‚‚ ACT - Reading
‚‚ ACT - Mathematics
‚‚ ACT - Science
‚‚ ACT - Writing
‚‚ GPA
28
G: CAE Board of Trustees and Officers
Roger Benjamin
President & CEO
James Hundley
Executive Vice President & COO
Benno Schmidt
Chairman, CAE
Richard Atkinson
President Emeritus, University of California System
Doug Bennett
President, Earlham College
Michael Crow
President, Arizona State University
Russell C. Deyo
Vice President & General Counsel, Johnson & Johnson
Richard Foster
Managing Partner, Millbrook Management Group, LLC
Ronald Gidwitz
Chairman, GCG Partners
Lewis B. Kaden
Vice Chairman, Citigroup Inc.
Michael Lomax
President, United Negro College Fund
Katharine Lyall
President Emeritus, University of Wisconsin System
Eduardo Marti
President, Queensborough Community College, CUNY
Ronald Mason
President, Jackson State University
Diana Natalicio
President, University of Texas at El Paso
Charles Reed
Chancellor, California State University
Michael D. Rich
Executive Vice President, RAND Corporation
Farris W. Womack
Executive Vice President and Chief Financial Officer, Emeritus
Professor, Emeritus
The University of Michigan
29
pb
3030
Download