2B 1100 Harding_Sandra.ppt

advertisement
Extreme measures
Professor Sandra Harding
James Cook University
Acknowledgements: Dr Nick Szorenyi-Reischl, Floris
van der Leest and Jasper Taylor, JCU
People in labour force with University
qualifications 2006
Good reasons to measure
• Public accountability
• Providing information to marketplace
• Driving particular outcomes:
–
–
–
–
–
–
–
–
Employability
Student satisfaction
Learning outcomes
Retention
Engagement
Equity and Indigenous education
Generic skills
Research output: quality, quantity, impact, translation
• Cross national comparison and competitiveness
• Improvement: AUQA, group, sector level…
Renewed interest in measures..
• Australia
– Reviews – Bradley, Cutler, CRC, RQF->ERA
– Compacts
– AUQA
– International standing queries
• Abroad
– Bologna process – comparable, transportable
– OECD: measuring teaching cross-nationally
– US: their competitiveness
– UK: improving measures, including in
upskilling
– League tables and their impacts
Getting the measure…..
Institutional level
• DEEWR reporting – IAF institutional level
– Organisational sustainability
– Quality of outcomes
– Compliance with legislation
– Monitoring equity
– Indigenous educational outcomes
• Student evaluation of teaching
• Student evaluation of subjects
• Sundry other surveys
Sector level – data gathered at level
of the institution
Teaching and learning:
• Graduate Destination Survey
–
–
–
–
Course comment
Labour market status
Employment details
Current further study
• Course Experience Questionnaire
–
–
–
–
–
Good teaching
Clear goals and standards
Appropriate workload
Appropriate assessment
Generic skills
Sector level…..
• AUSSE: Australian Universities Survey of
Student Engagement (25 universities in 2007),
ACER sponsored
• First year experience questionnaire (since 1994)
– Demographic data
– FY student experience
• Universities Australia
– Statistical data collection from members;
Sector level…
• Various other benchmarking exercises;
and across groups of institutions
• Research measures…. Higher Education
Research Data Collection (HERDC)
– Research income – Australian Competitive
Grants (and others)
– Publications (quality outlets, referreed)
– PhD completions
– Student load
Excellence for Research in Australia
(ERA)
• 141 Fields of Research groupings (4-digit FOR
codes ANZSRC; 8 clusters)
• Three broad categories of indicators:
– Measures of activity and intensity (total research
income and total number of outputs)
– Indicators of research quality (publications over 6
years; competitive research income over 3 years;
other discipline specific measures)
– Indicators of excellent applied research and
translation of research outcomes (no detail as yet)
Internationally…
• OECD data: spend on higher education
• World ranking of Universities - league
tables:
– Shanghai Jiao Tong Academic Ranking of
World Universities
– Times Higher Education Supplement– QS
World University Rankings
Fragility
• Data used for purposes other than originally
intended
– CEQ and GDS being used as a proxy for quality and
to distribute funding:
• Teaching and Learning Performance Fund
• Good Universities Guide
• Research data uses: distribute block funding
(IGS, RIDC, RTS, APAs, CTSs)
• Low response rates
• Reliability and validity problems
• Data lagged
• Disconnect between institutionally gathered data
on satisfaction and CEQ outcomes
Hot topics
• LTPF (Scanlon, James):
–
–
–
–
–
–
–
Allegations of institutional bad behaviour
Reporting on whole of institution, not field of study
Can students report?
Large impact of small differences between institutions
Dubious measure of institutional performance
Self reported generic skills a poor indicator
CEQ tells little about learning in absolute or value add
terms
Hot topics …
• SJTU and THES League tables
(Marginson)
– Both have problems,
• The THES ranking particularly – volatile; opinion
based; not reproducible
• Shanghai Jiao Tong: underplays humanities and
social science and publishing outside of English
language journals, history is important
Hot topics …
• Measurement driving loss of diversity?
• Risk of losing interdisciplinary research: a
challenge for ERA
• International developments – Bologna impacts
• Measuring outcomes for student learning and
value add:
– Assessment tasks: uncalibrated, unknown reliability
and validity
– Standards elusive – need for focus on value add
outcomes to deal with sector diversity, measurement
based on outcomes
• Measuring international students’ English
literacy – IELTS predictive validity unknown
What do we need?
• Better, targeted measures – fewer? Wonder…
• Measures being used appropriately
• Measures that drive the right behaviour
What don’t we have? Good measures of:
• Impact of research;
• Standard of English
• Standard of learning
• Diversity
• Interdisciplinarity
• The value added through education
• Industry outcomes/alumni outcomes
Download