Thinking it through: Australian students` skills in creative problem

advertisement
Australian Council for Educational Research
ACEReSearch
OECD Programme for International Student
Assessment (PISA Australia)
National and International Surveys
9-2014
Thinking it through: Australian students’ skills in
creative problem solving
Lisa De Bortoli
ACER, Lisa.DeBortoli@acer.edu.au
Greg Macaskill
ACER
Follow this and additional works at: http://research.acer.edu.au/ozpisa
Part of the Educational Assessment, Evaluation, and Research Commons, International and
Comparative Education Commons, Science and Mathematics Education Commons, and the
Secondary Education and Teaching Commons
Recommended Citation
De Bortoli, Lisa and Macaskill, Greg, "Thinking it through: Australian students’ skills in creative
problem solving" (2014).
http://research.acer.edu.au/ozpisa/18
This Book is brought to you by the National and International Surveys at ACEReSearch. It has been accepted for inclusion in OECD Programme for
International Student Assessment (PISA Australia) by an authorized administrator of ACEReSearch. For more information, please contact
repository@acer.edu.au.
Programme for International Student Assessment (PISA)
Thinking it through
Australian students’ skills
in creative problem solving
Lisa De Bortoli
Greg Macaskill
Australian Council for Educational Research
First published 2014
by Australian Council for Educational Research Ltd
19 Prospect Hill Road, Camberwell, Victoria, 3124, Australia
www.acer.edu.au
www.acer.edu.au/ozpisa/reports/
Text © Australian Council for Educational Research Ltd 2014
Design and typography © ACER Creative Services 2014
This book is copyright. All rights reserved. Except under the conditions described in the Copyright Act 1968 of
Australia and subsequent amendments, and any exceptions permitted under the current statutory licence scheme
administered by Copyright Agency (www.copyright.com.au), no part of this publication may be reproduced, stored in
a retrieval system, transmitted, broadcast or communicated in any form or by any means, optical, digital, electronic,
mechanical, photocopying, recording or otherwise, without the written permission of the publisher.
Cover design, text design and typesetting by ACER Creative Services
Edited by Amanda Coleiro
National Library of Australia Cataloguing-in-Publication entry
Author:
DeBortoli, Lisa, author.
Title:Thinking it through : Australian students’ skills in creative problem solving / Lisa DeBortoli, Greg
Macaskill.
ISBN:
9781742862576 (paperback)
Notes:
Includes bibliographical references.
Subjects:
Programme for International Student Assessment.
Problem-based learning--Australia
Critical thinking in children--Australia
Problem solving--Ability testing--Australia.
Educational evaluation.
Other Authors/Contributors:
Macaskill, Greg, author.
Australian Council for Educational Research, issuing body.
Dewey Number: 371.39
The views expressed in this report are those of the authors and not necessarily those of the Commonwealth, State
and Territory governments.
Contents
Executive summary
v
List of figures
xii
List of tables
xiv
Reader’s guide
xv
CHAPTER 1 Introduction
1
The main goals of PISA
1
The importance of assessing problem solving
2
What participants did
2
Participants in PISA 2012
3
How results are reported
6
Organisation of the report
6
Further information
6
CHAPTER 2 The assessment of problem solving
7
How is problem solving defined in PISA?
7
How is problem solving assessed in PISA?
7
The PISA 2012 problem-solving assessment structure
CHAPTER 3 Australian students’ performance in problem solving
10
23
Australia’s problem-solving performance from an international perspective
23
Australia’s problem-solving performance in a national context
31
Variations in problem-solving performance between and within schools
41
Comparing students’ performance in problem solving with mathematics, science and reading
44
Relative performance in problem solving in Australia
46
CHAPTER 4 Students’ strengths and weaknesses in problem solving
48
Students’ strengths and weaknesses in problem-solving processes
50
Students’ strengths and weaknesses in the nature of the problem situation
53
Students’ strengths and weaknesses on the response formats
55
Grouping countries by their strengths and weaknesses in problem solving
59
iii
CHAPTER 5 Australian students’ motivation towards problem solving
61
Perseverance61
Students’ openness to experience in problem solving
66
References70
iv
Contents
Executive summary
In PISA 2003, an assessment of cross-disciplinary problem solving was undertaken as a paper-based
assessment. In PISA 2012, problem solving was once again assessed with 44 of the 65 participating
countries and economies completing an optional computer-based assessment of problem solving.
The problem-solving assessment focuses on students’ general-reasoning skills, their ability to regulate
problem-solving processes and their willingness to do so, by presenting students with problems that do
not require specific curricular knowledge to solve.
PISA 2012 defines problem solving as:
an individual’s capacity to engage in cognitive processing to understand and resolve
problem situations where a method of solution is not immediately obvious. It includes
the willingness to engage with such situations in order to achieve one’s potential as a
constructive and reflective citizen. (OECD, 2014, p. 30)
There are three main aspects in the problem-solving framework that guided the development of
assessment items: 1) problem-solving processes—the cognitive process involved in problem solving:
exploring and understanding, representing and formulating, planning and executing, and monitoring
and reflecting; 2) the nature of the problem situation: interactive or static; and 3) the problem context:
technological or not, personal or social.
This report presents the results of the PISA 2012 problem-solving assessment that measured how well
prepared today’s 15-year-old students are in solving complex, unfamiliar problems that they may
encounter outside curricular contexts.
v
Australian students’ performance in problem solving
Reporting student performance
Similar to the reporting of results for other assessed domains in PISA, statistics such as
mean scores and measures of distribution of performance and proficiency levels are used to
examine student performance.
Mean scores
Mean scores provide a summary of student performance and allow comparisons of the
relative standing between different countries and different subgroups.
Proficiency levels
There are six levels in the PISA problem-solving proficiency scale, ranging from Level 6 (the
highest proficiency level) to Level 1 (the lowest proficiency level).
»» Students achieving a proficiency of Level 5 or 6 are considered top performers.
»» Level 2 has been defined as a baseline proficiency level and is the level of
achievement on the PISA scale at which students begin to demonstrate the
problem-solving competencies that will enable them to actively participate in reallife situations.
»» Students failing to reach Level 2 (those students placed at Level 1 or below) are
considered low performers.
Results across participating countries
» Overall, Australian students performed very well in the PISA 2012 problem-solving assessment,
and are well equipped to apply their skills and knowledge to solve challenging problems.
» Australia achieved a mean score of 523 points on the problem-solving assessment, which was
significantly above the OECD average of 500 score points.
» Australia was one of the high-performing countries, outperformed by only seven of the 44
participating countries and economies.
» Three countries and four economic regions, all from the Asian continent, performed significantly
higher than Australia. These were Singapore, Korea, Japan, Macao–China, Hong Kong–China,
Shanghai–China and Chinese Taipei.
» Australia’s performance was not significantly different from three countries: Canada, Finland
and England.
» Australia’s performance was significantly higher than 33 countries, including the United States
and Ireland.
» Sixteen per cent of Australian students were top performers compared to 30% of students in
Singapore and 12% of students across the OECD.
» Sixteen per cent of Australian students were low performers compared to 8% of students in
Singapore and 21% of students across the OECD.
vi
Executive summary
Results across the Australian jurisdictions
» All jurisdictions achieved statistically similar scores, except for Tasmania which performed
significantly lower than all other jurisdictions.
» Six jurisdictions (Western Australia, the Australian Capital Territory, New South Wales, Victoria,
Queensland and South Australia) performed at a significantly higher level than the OECD
average. The Northern Territory performed at a level not significantly different to the OECD
average and Tasmania performed significantly lower than the OECD average.
» The proportion of top performers in problem solving ranged from 11% in Tasmania to 19% in
the Australian Capital Territory.
» The proportion of low performers in problem solving ranged from 13% in Western Australia to
27% in Tasmania.
Results for females and males
» Across OECD countries, males performed significantly higher than females (by 6 score points
on average). Approximately half the countries had significant sex differences in favour of males,
while 11% of the countries had significant sex differences in favour of females.
» Australian females and males performed at a level that was not significantly different in
problem solving.
» In Australia, 16% of females and 18% of males were top performers compared to 10% of females
and 13% of males across OECD countries.
» In Australia, 15% of females and 16% of males were low performers compared to 22% of females
and 22% of males across OECD countries.
» Significant sex differences were found in only one jurisdiction, Western Australia, with males
achieving 18 score points on average higher than females.
» All jurisdictions, except Tasmania, achieved a higher proportion of top-performing males
compared to the OECD average (13%), while all jurisdictions achieved a higher proportion of
top-performing females compared to the OECD average (10%).
» The proportion of low-performing males in Tasmania and the Northern Territory was higher
than the OECD average (22%), while the proportion of low-performing females was higher in
Tasmania than across the OECD (22%).
Results for geographic location of schools
The geographic location of schools was classified using the broad categories metropolitan, provincial and
remote, as defined in the MCEECDYA Schools Geographic Location Classification1.
» Students attending metropolitan schools performed at a significantly higher level (528 score
points on average) than students in schools from provincial areas (510 score points on average)
and remote areas (475 score points on average). Students attending provincial schools significantly
outperformed students attending remote schools.
» Eighteen per cent of students from metropolitan schools and 12% of students from provincial
schools were top performers compared to 9% of students from remote schools.
» Fifteen per cent of students from metropolitan schools and 18% of students from provincial
schools were low performers compared to 30% of students from remote schools.
Refer to the Reader’s Guide for details about the MCEECDYA Schools Geographic Location Classification.
1
Executive summary
vii
Results for Indigenous students
Students’ Indigenous background was derived from information provided by the school.2
» Indigenous students achieved on average 454 score points in problem solving, which was
significantly lower than for non-Indigenous students (526 score points on average) and for
students across the OECD.
» Four per cent of Indigenous students were top performers compared to 18% of
non-Indigenous students.
» Thirty-seven per cent of Indigenous students were low performers compared to 15% of
non-Indigenous students.
» Indigenous females and males performed at a level that was not significantly different in
problem solving.
» A small, yet similar, proportion of Indigenous females (3%) and males (4%) were top
performers in problem solving, while 35% of Indigenous females and 39% of Indigenous males
were low performers.
Results for socioeconomic background
Socioeconomic background in PISA is measured by an index of Economic, Social and Cultural Status
(ESCS), which captures the wider aspects of a student’s family and home background.3
» Students in the highest socioeconomic quartile achieved an average score of 560 points, which
was 73 score points higher than those students in the lowest socioeconomic quartile.
» Twenty-seven per cent of students in the highest socioeconomic quartile were top performers
compared to 9% of students in the lowest socioeconomic quartile.
» Eight per cent of students in the highest socioeconomic quartile were low performers compared
to 25% of students in the lowest socioeconomic quartile.
Results for immigrant background
Immigrant background was measured on students’ self-report of where they and their parents were born.4
» Australian-born students achieved an average score of 523 points, which was not significantly
different from the performance of foreign-born students (517 points), but significantly lower than
the mean score achieved for first-generation students (531 points).
» Sixteen per cent of Australian-born students, 19% of first-generation students and 16% of
foreign-born students were top performers.
» Fifteen per cent of Australian-born students, 14% of first-generation students and 18% of
foreign-born students were low performers.
Results for language background
Language background was based on students’ responses regarding the main language spoken at home—
English or another language.5
The Reader’s Guide provides more information about the definition of Indigenous background.
Refer to the Reader’s Guide for details about the Economic, Social and Cultural Status index.
4
Refer to the Reader’s Guide for details about the definitions of immigrant background.
5
Refer to the Reader’s Guide for details about the definitions of language background.
2
3
viii
Executive summary
»
»
»
Students who spoke English at home performed significantly higher (average score of 526 points)
than those students who spoke a language other than English at home (average score of 509 points).
Eighteen per cent of students who spoke English at home and 16% of students who spoke a
language other than English at home were top performers.
Fifteen per cent of students who spoke English at home and 21% of students who spoke a
language other than English at home were low performers.
Variations in problem-solving performance between and within schools
The variation in performance within countries can be divided into a measure of performance difference
between students from the same school and a measure of performance difference between groups of
students from different schools.
» In Australia, the amount of variation in performance within schools was 75% and was higher
than the OECD average (61%), while the amount of variation in performance between
Australian schools was 28% and lower than the OECD average (38%).
» On average, the variation in problem-solving performance that was observed between schools
ranged from 19% in South Australia to 39% in Tasmania, while the variation in problem-solving
performance that was observed within schools ranged from 72% in Victoria to 94% in the
Northern Territory.
Variation in problem-solving performance associated with performance in mathematics,
science and reading
An analysis examined the variation in problem-solving performance that was associated with skills
measured in the problem-solving assessment and the variation in problem-solving performance that was
also measured in one of the three regular literacy domain assessments.
» Across the OECD, 68% of the problem-solving variance ref lected skills that were also measured
in one of the three literacy domains regularly assessed in PISA. The remaining 32% ref lected
skills that were uniquely measured in the problem-solving assessment.
» In Australia, 71% of the problem-solving variance ref lected skills that were also measured in
one of the three literacy domains regularly assessed in PISA. The remaining 29% of the score
ref lected skills that were uniquely measured in the problem-solving assessment.
Relative performance in problem solving in Australia
»
Australian students performed better than expected in problem solving, based on their
performance in mathematics. The difference between observed and expected performance is
particularly large among students with strong performance in mathematics.
Students’ strengths and weaknesses in problem solving
Focusing on the different aspects of the problem-solving framework, analyses were undertaken to identify
comparative strengths and weaknesses within countries and within different social groups.
Executive summary
ix
Strengths and weakness in the problem-solving processes
» Generally, the higher performing countries in problem solving performed relatively stronger on
the exploring and understanding process and on the representing and formulating process, and
relatively weaker on the planning and executing process and on the monitoring and ref lecting
process. (These comparisons take into account the countries’ overall performance.)
» Australian students are comparatively stronger on the exploring and understanding process and
on the representing and formulating process, and are relatively weaker on the planning and
executing process. (These comparisons take into account the countries’ overall performance.)
» Students from Western Australia performed relatively stronger on the exploring and
understanding process, while in New South Wales students performed relatively weaker on
this process. Students from Queensland performed relatively stronger on the representing and
formulating process.
» Females’ problem-solving process skills were relatively stronger on the monitoring and ref lecting
process, and relatively weaker on the representing and formulating process. The opposite was
found for males, where their relative strength was found on the representing and formulating
process, and their relative weakness was found on the monitoring and ref lecting process.
» Indigenous students were relatively weaker on the exploring and understanding process, while
non-Indigenous students were found to be relatively stronger on this process.
» Students in the lowest socioeconomic quartile performed relatively stronger on the planning and
executing process, and relatively weaker on the exploring and understanding process. The reverse
was found for students in the highest socioeconomic quartile.
Strengths and weakness in the nature of the problem situation
» No clear pattern emerged of relative strength or weakness in static or interactive items by
countries’ overall performance in problem solving.
» In Tasmania, students performed relatively stronger on the static tasks, while in Queensland
students performed relatively stronger on the interactive tasks.
» No relative strengths or weaknesses in static or interactive tasks were found across the
different social groups.
Strengths and weakness on the response formats
» Generally, the higher performing countries and the lower performing countries in problem
solving performed relatively stronger on the selected-response format items and weaker on the
constructed-response format items.
» In Australia, students performed relatively stronger on the constructed-response format items
and weaker on the selected-response format items.
» Tasmanian students performed relatively stronger on the constructed-response format items and
Queensland students performed relatively stronger on the selected-response format items.
» In Australia, females performed stronger than males on the constructed-response format items.
» Australian-born students performed relatively stronger on the constructed-response format items,
where the effect was consistent, but weaker for students who spoke English at home. Foreignborn students were relatively stronger on the selected-response format items. This was also the
case, to a lesser extent, for students who spoke a language other than English at home.
x
Executive summary
Australian students’ perseverance and openness in problem solving
The PISA definition of problem solving acknowledges that solving a problem relies on motivational and
affective factors. In PISA 2012, students completed a questionnaire that collected information about their
engagement with and at school, their drive and the beliefs they hold about themselves as learners. This
included measures of perseverance and openness in problem solving.
Perseverance in problem solving
In PISA, perseverance relates to a student’s willingness to work on problems.
» Australian students reported a significantly higher level of perseverance than the OECD average.
» Australian males reported significantly higher levels of perseverance than Australian females.
» All jurisdictions reported higher mean scores on the perseverance index compared to the OECD
average, with students from the Australian Capital Territory reporting the highest levels and
students from the Northern Territory reporting the lowest levels of perseverance.
» Non-Indigenous students, students from metropolitan schools and students in the highest
socioeconomic quartile reported higher levels of perseverance than their counterparts.
Students’ openness to experience in problem solving
Openness relates to a student’s willingness to engage with problems and to be open to new challenges in
order to be able to solve complex problems and situations.
» Australian students reported a lower level of openness to problem solving than the
OECD average.
» Australian males reported significantly higher levels of openness to problem solving than
Australian females.
» The Australian Capital Territory was the only jurisdiction to have an average score that was
higher than the OECD average. The Northern Territory had the same index score as the OECD
average, while all other jurisdictions had a lower index score than the OECD average. The
Australian Capital Territory had the highest mean score on the openness to problem-solving
index, while South Australia and Queensland had the lowest mean scores on this index.
» Similar to the findings on perseverance, non-Indigenous students, students from metropolitan
schools and students in the highest socioeconomic quartile reported higher levels of openness to
problem solving than their counterparts.
Executive summary
xi
Figures
Figure 1.1
Countries participating in PISA 2012
4
Figure 2.1
Main features of the PISA 2012 problem-solving assessment framework
8
Figure 2.2
The relationship between items and students on the PISA problem-solving scale
12
Figure 2.3
Summary descriptions of the six levels on the problem-solving proficiency scale
14
Figure 2.4Map of selected problem-solving items, illustrating the proficiency level and assigned
problem-solving process and nature of the problem
15
Figure 3.1
Mean scores and distribution of students’ performance on the problem-solving scale
26
Figure 3.2
Percentage of students across the problem-solving proficiency scale, by country
28
Figure 3.3Mean scores and differences between sexes in students’ performance on the problemsolving scale, by country
30
Figure 3.4Percentage of students across the problem-solving proficiency scale by sex, for Australia and
the OECD average
31
Figure 3.5Mean scores and distribution of students’ performance on the problem-solving scale, by
jurisdiction32
Figure 3.6
Percentage of students across the problem-solving proficiency scale, by jurisdiction
Figure 3.7Mean scores and differences in students’ performance on the problem-solving scale, by
jurisdiction and sex
Figure 3.8
Percentage of students across the problem-solving proficiency scale, by jurisdiction and sex
Figure 3.9Mean scores and distribution of students’ performance on the problem-solving scale, by
geographic location
Figure 3.10
Percentage of students across the problem-solving proficiency scale, by geographical location
Figure 3.11Mean scores and distribution of students’ performance on the problem-solving scale, by
Indigenous background
Figure 3.12
Percentage of students across the problem-solving proficiency scale, by Indigenous background
34
34
35
36
36
37
Figure 3.13Mean scores and distribution of students’ performance on the problem-solving scale, by
Indigenous background and sex
37
Figure 3.14Percentage of students across the problem-solving proficiency scale, by Indigenous
background and sex
38
Figure 3.15Mean scores and distribution of students’ performance on the problem-solving scale, by
socioeconomic background
38
Figure 3.16
Percentage of students across the problem-solving proficiency scale, by socioeconomic background 39
Figure 3.17Mean scores and distribution of students’ performance on the problem-solving scale, by
immigrant background
xii
33
39
Figure 3.18
Percentage of students across the problem-solving proficiency scale, by immigrant background
Figure 3.19Mean scores and distribution of students’ performance on the problem-solving scale, by
language background
40
40
Figure 3.20
Percentage of students across the problem-solving proficiency scale, by language background
41
Figure 3.21
Variation in problem-solving performance between and within schools, by country
42
Figure 3.22
Variation in problem-solving performance between and within schools, by jurisdiction
43
Figure 3.23Variation in problem-solving performance associated with performance in mathematics,
science and reading, by country
45
Figure 3.24Variation in problem-solving performance associated with performance in mathematics,
science and reading, by jurisdiction
46
Figure 3.25Relative performance in problem solving at different levels on the mathematics scale for
Australia, England and the United States
47
Figure 4.1
Relative strengths and weaknesses in problem-solving processes, by countries
51
Figure 4.2
Relative strengths and weaknesses in problem-solving processes, by jurisdictions
52
Figure 4.3Relative strengths and weaknesses in problem-solving processes, by different social groups
for Australia
53
Figure 4.4Relative strengths and weaknesses on problem-solving tasks by the nature of the problem
situation, across countries
54
Figure 4.5Relative strengths and weaknesses on problem-solving tasks by the nature of the problem
situation, across jurisdictions
55
Figure 4.6
Relative strengths and weaknesses on problem-solving tasks by response format, across countries 56
Figure 4.7
Relative strengths and weaknesses on problem-solving tasks by response format, across
jurisdictions
57
Figure 4.8Relative strengths and weaknesses on problem-solving tasks by response format, across
different social groups
58
Figure 4.9
Joint analysis of strengths and weaknesses, by nature of the problem and by process, for countries 59
Figure 4.10Joint analysis of strengths and weaknesses, by nature of the problem and by process, for
Australian jurisdictions and social groups Figure 5.1
Relationship between Australian students’ perseverance and problem-solving performance
Figure 5.2Relationship between Australian students’ openness to problem solving and
problem-solving performance
60
64
68
Figures
xiii
Tables
Table 1.1
Number of Australian PISA 2012 schools, by jurisdiction and school sector
5
Table 1.2
Number of Australian PISA 2012 students, by jurisdiction and school sector
5
Table 2.1Classification of problem-solving items, by cognitive process and the nature of the problem
situation11
Table 3.1
Multiple comparisons of mean problem-solving performance, by jurisdiction
Table 3.2Relationship between performance in problem solving, mathematics, science and reading
across the OECD
44
Table 3.3Relationship between performance in problem solving, mathematics, science and reading
for Australia
44
Table 5.1Students’ perseverance in problem solving for Australia and comparison countries
62
Table 5.2Index of perseverance for Australia and comparison countries
63
Table 5.3Index of perseverance for Australia and comparison countries, by sex
64
Table 5.4Students’ perseverance in problem solving, by sex, jurisdiction, geographic location,
Indigenous background and socioeconomic background
65
Table 5.5
Students’ openness to problem solving for Australia and comparison countries
66
Table 5.6
Index of openness to problem solving for Australia and comparison countries
67
Table 5.7
Index of openness to problem solving for Australia and comparison countries, by sex
67
Table 5.8Students’ openness to problem solving, by sex, jurisdiction, geographic location,
Indigenous background and socioeconomic background
xiv
32
69
Reader’s guide
Target population for PISA
This report uses ‘15-year-olds’ as shorthand for the PISA target population. In practice, the target
population was students who were aged between 15 years and 3 (complete) months and 16 years and 2
(complete) months at the beginning of the assessment period, and who were enrolled in an educational
institution that they were attending full-time or part-time. Since the largest part (but not all) of the PISA
target population is made up of 15-year-olds, the target population is often referred to as 15-year-olds.
OECD average
An OECD average was calculated for most indicators in this report and is presented for comparative
purposes. The OECD average represents OECD countries as a single entity and each country contributes
to the average with equal weight. The OECD average is equivalent to the arithmetic mean of the
respective country statistics.
Rounding of figures
Because of rounding, some numbers in tables may not exactly add to the totals reported. Totals,
differences and averages are always calculated on the basis of exact numbers and are rounded only after
calculation. When standard errors have been rounded to one or two decimal places and the value 0.0
or 0.00 is shown, this does not imply that the standard error is zero, but that it is smaller than 0.05 or
0.005 respectively.
Confidence intervals and standard errors
In this and other publications, student achievement is often described by a mean score. For PISA, each
mean score is calculated from the sample of students who undertook the PISA assessment and is referred
to as the sample mean. These sample means are an approximation of the actual mean score (known
as the population mean) that would have been obtained had all students in a country actually sat the
PISA assessment.
Since the sample mean is just one point along the range of student achievement scores, more information
is needed to gauge whether the sample mean is an under estimation or an over estimation of the
population mean. The calculation of confidence intervals can assist assessment of a sample mean’s
precision as a population mean. Confidence intervals provide a range of scores within which we are
confident that the population mean actually lies.
In this report, sample means are presented with an associated standard error. The confidence interval—
which can be calculated using the standard error—indicates that there is a 95% chance that the actual
population mean lies within plus or minus 1.96 standard errors of the sample mean.
The term significantly is used through this report to describe a difference that meets the requirements of
statistical significance at the 0.05 level, indicating that the difference is real and would be found in at least
95 analyses out of 100 if the comparison was to be repeated.
xv
Mean performance
Mean scores provide a summary of students’ performance and allow comparisons of the relative standing
between different countries and different subgroups. In addition, the distribution of scores (reported at
the 5th, 10th, 25th, 75th, 90th and 95th percentiles) are reported in graphical format.
Proficiency levels
To summarise data from responses to the PISA assessment, performance scales were constructed for each
assessment domain. The scales are used to describe the performance of students in different countries,
including in terms of described performance levels. The described performance levels are known as
proficiency levels.
This publication uses top performers as shorthand for those students proficient at Level 5 or 6 of the
assessment and low performers for those students proficient below Level 2 of the assessment.
PISA indices
The measures that are presented as indices summarise students’ responses to a series of related items
constructed on the basis of previous research. In describing students in terms of each characteristic (for
example, instrumental motivation to learn mathematics, or disciplinary climate), scales were constructed
on which the average OECD student was given an index value of 0, and about two-thirds of the OECD
population were given values between –1 and +1 (that is, the index has a mean of 0 and a standard
deviation of 1). Negative values on an index do not necessarily imply that students responded negatively
to the underlying items. Rather, students with a negative score responded less positively than students on
average across OECD countries.
The indices are based on four categories for each item, whereas the reported percentages are collapsed
into two categories. Due to this and the weighting of responses, a ranking based on the value of the
indices will sometimes not exactly correspond to one based, say, on the average of the percentages.
Information about school characteristics was collected through the school questionnaire, which was
completed by the principal. In this report, responses from principals were weighted so that they are
proportionate to the number of 15-year-olds enrolled in the school.
Bonferroni correction
The Bonferroni correction states that if an experimenter is testing n independent hypotheses on a set of
data, then the statistical significance level that should be used for each hypothesis separately is 1/n times
what it would be if only one hypothesis was tested. The Bonferroni correction was used in the multiple
comparison tables in earlier PISA publications (for PISA 2000 and PISA 2003). It is widely acknowledged
that there are technical issues with using the Bonferroni correction for such a large group of countries and
its results are conservative. As such, the Bonferroni correction has not been used in PISA 2012.
Correlational analysis
An analysis of the correlation between two variables can be used to investigate the association between
them. If there is a significant positive correlation, it does not imply that one factor depends on the other
or that there is a cause–effect relationship between them; it simply means that they occur together.
Further analysis and investigation are needed to determine the nature of the association. The most
commonly used measure is the Pearson correlation coefficient, which is abbreviated as r.
xvi
Reader’s guide
The correlation coefficient measures the strength between two variables. Values of the correlation
coefficient can range from –1 (a negative correlation—as one value increases the other value decreases) to
a +1 (a positive correlation—as one value increases the other value increases). In this report, as a general
rule, the correlation coefficients have been interpreted as follows:
Correlation coefficient range
Strength of association
r < –0.50
strong/high negative association
–0.50 > r < –0.30
moderate/medium negative association
–0.30 > r < –0.10
small/low negative association
–0.10 > r < +0.10
very small or no association
+0.10 > r < +0.30
small/low positive association
+0.30 > r < +0.50
moderate/medium positive association
r > +0.50
strong/high positive association
Definition of problem solving
The term creative problem solving emphasises the difference from the PISA 2003 assessment of crosscurricular problem solving and the PISA 2012 assessment, which examined the necessary role creativity
plays in real problem solving, i.e., where problems are set in novel situations and the ways of achieving a
goal is not immediately obvious. For consistency with the corresponding OECD report, the title of this
report refers to creative problem solving, while the term problem solving is used throughout the report.
Definitions of background characteristics
There are a number of definitions used in this report that are particular to the Australian context, as well
as many that are relevant to the international context. This section provides an explanation for those that
are not self-evident.
Indigenous background
Indigenous background is derived from information provided by the school, which was taken from school
records. Students were identified as being of Australian Aboriginal or Torres Strait Islander descent.
For the purposes of this publication, data for the two groups are presented together under the term
Indigenous Australian students.
Socioeconomic background
Socioeconomic background is based on the answers of students to questions about their parents’
education, parents’ occupation and items in the home. Two measures are used by the OECD to represent
elements of socioeconomic background. One is the highest level of the father’s and mother’s occupation
(known as HISEI), which is coded in accordance with the International Labour Organization’s
International Standard Classification of Occupations. The other measure is the index of Economic,
Social and Cultural Status (ESCS), which was created to capture the wider aspects of a student’s family
and home background. The ESCS is based on three indices: the highest occupational status of parents
(HISEI); the highest educational level of parents in years of education (PARED); and home possessions
(HOMEPOS). The index of home possessions (HOMEPOS) comprises all items on the indices of family
wealth (WEALTH), cultural resources (CULTPOSS), access to home educational and cultural resources
(HEDRES), and books in the home.
Reader’s guide
xvii
Geographic location
In Australia, participating schools were coded with respect to the MCEECDYA Schools Geographic
Location Classification. For the analysis in this report, only the broadest categories are used:
» Metropolitan—including mainland capital cities or major urban districts with a population of
100,000 or more (for example, ACT-Queanbeyan, Cairns, Geelong, Hobart)
» Provincial—including provincial cities and other nonremote provincial areas (for example,
Darwin, Ballarat, Bundaberg, Geraldton, Tamworth)
» Remote—including remote areas and very remote areas. Remote: very restricted accessibility
of goods, services and opportunities for social interaction (for example, Coolabah, Mallacoota,
Capella, Mt Isa, Port Lincoln, Port Hedland, Swansea, Alice Springs). Very remote: very little
accessibility of goods, services and opportunities for social interaction (for example, Bourke,
Thursday Island, Yalata, Condingup, Nhulunbuy).
Immigrant background
For the analysis in this report, immigrant background has been defined by the following categories:
» Australian-born students—students born in Australia with both parents born in Australia
» First-generation students—students born in Australia with at least one parent born overseas
» Foreign-born students—students born overseas with both parents also born overseas.
Language background
The language spoken at home indicates whether a student has a language background other than English.
The question asked about the language spoken at home most of the time.
Sample surveys
PISA is a sample survey and, as such, a random sample of students was selected to represent the population
of 15-year-old students. The PISA sample was designed as a two-stage stratified sample. The first stage
involved the sampling of schools in which 15-year-old students could be enrolled. The second stage of the
selection process sampled students within the sampled schools.
The following variables were used in the stratification of the school sample: jurisdiction; school sector;
geographic location (based on the MCEECDYA’s Schools Geographic Location Classification); sex of
students at the school; a socioeconomic background variable (based on the Australian Bureau of Statistics’
Socio-Eeconomic Indexes for Areas—SEIFA; the SEIFA consists of four indexes that rank geographic
areas across Australia in terms of their relative socioeconomic advantage and disadvantage); and an
achievement variable (based on a Year 9 NAPLAN numeracy school-level score).
Online statistical tables
The data underlying the figures in this report are provided in Excel spreadsheets and are available from
the national PISA website: www.acer.edu.au/ozpisa/reports/.
Acknowledgements
Information included in this report about the problem-solving framework, proficiency scale, sample items
and discussion about the results—including some of the international tables—has been assembled from the
OECD problem-solving report (OECD, 2014).
xviii
Reader’s guide
CHAPTER 1
Introduction
In every PISA survey, students from every participating country are assessed in the core domains of
mathematics, science and reading literacy. In addition to assessing these literacy domains, the OECD
proposes additional assessments in other domains. In PISA 2003, a paper-based assessment of crossdisciplinary problem solving was first assessed, when it was included as a core domain. In PISA 2012,
problem solving was once again assessed, this time as an optional computer-based assessment.
The focus of the PISA 2012 assessment of problem solving was: Are today’s 15-year-old students
acquiring the problem-solving skills that will prepare them to meet the challenges of the future? This
report describes how PISA defines and measures problem solving, and presents the performance of
Australian students in problem solving.
The main goals of PISA
PISA seeks to measure how well young adults, at age 151 and near the end of compulsory schooling in
most participating education systems, are prepared to use knowledge and skills in particular areas to meet
real-life challenges. This is in contrast to assessments that seek to measure the extent to which students
have mastered a specific curriculum. PISA’s orientation reflects a change in the goals and objectives of
curricula, which increasingly address how well students are able to apply what they learn at school.
As part of the PISA process, students complete an assessment of mathematical literacy, scientific literacy and
reading literacy, as well as an extensive background student questionnaire. School principals complete a
school questionnaire describing the context of education at their school, including the level of resources in
the school and the qualifications of staff. From this, the reporting of PISA findings is able to focus on:
» How well are young adults prepared to meet the challenges of the future? Can they analyse,
reason and communicate their ideas effectively? What skills do they possess that will facilitate
their capacity to adapt to rapid societal change?
» Are some ways of organising schools or school learning more effective than others?
» What inf luence does the quality of school resources have on student outcomes?
» What educational structures and practices maximise the opportunities of students from
disadvantaged backgrounds? How equitable is the provision of education within a country
or across countries?
Refer to the Reader’s Guide for more information about the target population for PISA.
1 1
The importance of assessing problem solving
The ability to solve complex problems enables an individual to adapt to changes in society or the
environment and to learn from their mistakes. Individuals that are proficient problem solvers will be
personally fulfilled, have greater opportunities for employment and contribute to economic growth
(Autor, Levy & Murnane, 2003; Hanushek, Jamison, Jamison & Woessmann, 2008).
Every individual, regardless of who they are and what they do, will encounter problems that they will
need to solve. Some problems will be easy to solve, while others will be more complex, requiring a
number of steps and strategies to solve the problem. For some problems, the skills and knowledge that
were acquired during school will be used to solve a problem; for example, managing a credit card,
applying measurement conversions when working with a recipe, or making an informed decision about
climate change. However, there are other problems that are not related to the skills or knowledge
acquired in school and that individuals will be unfamiliar with. These types of problems require a
different set of skills: existing knowledge needs to be reorganised and combined with new knowledge
using a range of reasoning skills. It is these skills that the PISA 2012 computer-based assessment measured.
The PISA 2012 assessment of problem solving assessed students’ general reasoning skills, their ability to
regulate problem-solving processes and their willingness to do so by presenting students with problems
that can be solved without domain-specific knowledge.
What participants did
Australian students who participated in PISA 2012 completed a paper-based assessment booklet that
contained questions assessing mathematical literacy and questions assessing either reading literacy,
scientific literacy or both, and a student questionnaire. All students in the 44 countries that opted to
participate in problem solving completed a computer-based assessment that assessed one or more of
problem-solving, mathematical and reading literacy.
Cognitive assessment
In PISA 2012, the majority of the assessment was devoted to mathematical literacy, with scientific literacy
and reading literacy assessed to a lesser extent. Participating students each responded to a two-hour
cognitive assessment, which took place in the morning.
After a lunch break, all students2 completed a 40-minute computer-based cognitive assessment. Students
completed a practice test before responding to one of 24 forms. Each form consisted of two clusters (of
20 minutes each) allocated according to a rotated test design among four clusters of computer problemsolving items, four clusters of computer mathematical literacy items and two clusters of computer reading
literacy items.
In the cognitive assessments, students were presented with units that required them to construct responses
to a stimulus and a series of questions (or items). Context was represented in each unit by the stimulus
material, which was typically a brief written passage or text accompanied by a table, chart, graph,
photograph or diagram. Each unit then contained several items related to the stimulus material.
A range of item-response formats was employed to cover the full range of cognitive abilities and
knowledge identified in the assessment frameworks. There were five types of item format: multiplechoice and complex multiple-choice items, in which students selected from among several possible
answers; closed constructed-response items, in which students were required to provide an unambiguous
single word, a number or diagrammatic answer; and open constructed-response and short-response
items, in which students provided a written response, showing the methods and thought processes they
had used.
In other participating countries, the computer-based assessment was only administered to a subsample of the students who had been assessed in PISA.
2 2
Thinking it through: Australian students’ skills in problem solving
Context questionnaires
PISA 2012 collected contextual information from students and principals. The internationally
standardised student questionnaire sought information on students and their family background, aspects
of motivation, learning and instruction in mathematics, and context of instruction including instructional
time and class size. Students were randomly assigned one of three questionnaires. Each questionnaire
comprised questions about the student and their family background and a selection of questions from the
remaining pool of questions.
Students were allowed up to 40 minutes to complete the student questionnaire, which they responded
to after the completion of the paper-based assessment and before the completion of the computerbased assessment.
The principal (or the principal’s delegate) completed the school questionnaire. It collected descriptive
information about the school, including the quality of the school’s human and material resources,
decision-making processes, instructional practices, and school and classroom climate. In Australia, the
school questionnaire was administered online and took around 30 minutes to complete.
Time of testing
PISA standards stipulate that testing should take place in the second half of the academic year. In
Australia, the PISA assessment took place in a six-week period from late July to early September 2012.
For most countries in the Northern Hemisphere, the testing period took place between March and May
2012. Together with appropriate application of the student age definition, this resulted in the Australian
students being at both a comparable age and a comparable stage in the school year to those in the
Northern Hemisphere who had been tested earlier in 2012.
Participants in PISA 2012
Countries
Although PISA was originally an OECD assessment created by the governments of OECD countries, it
has become a major assessment in many regions and countries around the world. Since the first assessment
in 2000, when PISA was implemented in 32 OECD countries, it has expanded to include nonOECD countries, referred to as partner countries and economies3. Sixty-five countries and economies
participated in PISA 2012, including 34 OECD countries and 31 partner countries or economies.4
Forty-four countries participated in the optional computer-based assessment, including 28 OECD and
16 partner countries. Around 85,000 students were assessed in problem solving, representing around 19
million 15-year-olds in the schools of participating countries.
Figure 1.1 (p. 4) shows the countries that participated in PISA 2012 (in purple) and also identifies those
countries that participated in the PISA computer-based assessment (denoted with an asterisk in the table).
Economic regions are required to meet the same PISA technical standards as other participating countries. Results for an economic region are only
representative of the region assessed and are not representative of the country.
4
Although Chinese Taipei, Hong Kong–China, Macao–China and Shanghai–China are economic regions, for convenience they will be referred to
throughout this report as countries.
3
Introduction
3
OECD countries
Australia*
Partner countries/economies
Hungary*
Poland*
Albania
Kazakhstan
Shanghai–China*
Austria*
Iceland
Portugal*
Argentina
Latvia
Singapore*
Belgium*
Ireland*
Slovak Republic*
Brazil*
Liechtenstein
Thailand
Canada*
Israel*
Slovenia*
Bulgaria*
Lithuania
Tunisia
Chile*
Italy*
Spain*
Chinese Taipei*
Macao–China*
United Arab Emirates*
Czech Republic*
Japan*
Sweden*
Colombia*
Malaysia*
Uruguay*
Denmark*
Korea*
Switzerland
Costa Rica
Montenegro*
Vietnam
Estonia*
Luxembourg
Turkey*
Croatia*
Peru
Finland*
Mexico
United Kingdom*
Cyprus*
Qatar
France*
Netherlands*
United States*
Hong Kong–China*
Romania
Germany*
New Zealand
Indonesia
Russian Federation*
Greece
Norway*
Jordan
Serbia*
Note: Those countries that participated in the computer-based assessment of problem solving in PISA 2012 are denoted with an asterisk.
Figure 1.1 Countries participating in PISA 20125
In this report, the OECD average refers to the average scores for students in the 28 OECD countries who
participated in the computer-based assessment of problem solving.
Schools and students
The target population for PISA is students who are 15 years old and enrolled in an educational institution,
either full- or part-time, at the time of testing. In most countries, 150 schools and 35 students in each
school were randomly selected to participate in PISA. In some countries, including Australia, a larger
sample of schools and students participated.
Only England participated in the computer-based assessment.
5
4
Thinking it through: Australian students’ skills in problem solving
The Australian PISA 2012 school sample consisted of 775 schools (Table 1.1). The sample was designed so
that schools were selected with a probability proportional to the enrolment of 15-year-olds in each school.
Stratification of the sample ensured that the PISA sample was nationally representative of the 15-year-old
population. Several variables were used in the stratification of the school sample including jurisdiction6,
school sector, geographic location, sex of students at the school, a socioeconomic background variable7
and an achievement variable8.
Of the Australian PISA schools, 85% were coeducational. Eight per cent of schools catered only for female
students, while 7% catered only for male students. Of the PISA schools that were single-sex schools, 2%
(17 schools) were government schools, almost 8% (62 schools) were Catholic and 4% (34 schools) were
independent schools.
The Australian PISA 2012 sample of 14,481 students, whose results feature in the national and
international reports, was drawn from all jurisdictions and school sectors according to the distributions
shown in Table 1.2.
Table 1.1 Number of Australian PISA 2012 schools, by jurisdiction and school sector
Sector
Jurisdiction
Government
Catholic
Independent
Total
ACT
26
8
11
45
NSW
113
43
28
184
VIC
77
31
26
134
QLD
83
24
25
132
SA
56
18
18
92
WA
51
18
21
90
TAS
47
12
12
71
NT
17
5
5
27
470
159
146
775
Australia
Note: The numbers are based on unweighted data.
Table 1.2 Number of Australian PISA 2012 students, by jurisdiction and school sector
Jurisdiction
ACT
NSW
VIC
QLD
N students
501
2133
1362
1769
Weighted N
2386
47964
35446
30539
SA
WA
TAS
NT
Total
931
1020
869
256
8841
10268
15363
3842
1341
147149
Government
Catholic
N students
209
828
571
497
306
330
235
81
3057
Weighted N
1500
19389
15636
10200
3691
5742
1221
210
57589
N students
198
486
473
456
336
388
154
92
2583
Weighted N
827
12155
11312
10044
3668
6431
832
703
45972
Independent
Australia
N students
908
3447
2406
2722
1573
1738
1258
429
14481
Weighted N
4713
79508
62394
50783
17627
27536
5895
2254
250710
Notes: N students is based on the achieved (unweighted) sample.
Weighted N is based on the number of students in the target population represented by the sample.
Throughout this report, the Australian states and territories will be collectively referred to as jurisdictions.
Based on the Australian Bureau of Statistic’s Socio-Economic Indexes for Areas (SEIFA).
8
Based on a NAPLAN numeracy school-level score.
6
7
Introduction
5
How results are reported
International comparative studies have provided an arena to observe the similarities and differences
between educational policies and practices. They enable researchers and others to observe what is possible
for students to achieve and what environment is most likely to facilitate their learning. PISA provides
regular information on educational outcomes within and across countries by providing insight into the
range of skills and competencies, in different assessment domains, that are considered to be essential to an
individual’s ability to participate in and contribute to society.
Similar to other international studies, PISA results are reported as mean scores that indicate average
performance and various statistics that reflect the distribution of performance. School and student
variables further enhance the understanding of student performance. PISA also attaches meaning to
the performance scale by providing a profile of what skills and knowledge students have achieved.
The performance scale is divided into levels of difficulty, referred to as proficiency levels. Students at a
particular level not only typically demonstrate the knowledge and skills associated with that level, but
also the proficiencies required at lower levels. For the domain of problem-solving literacy, six proficiency
levels have been defined to describe the scale.
Organisation of the report
This report focuses on the results for Australian students in problem solving in PISA 2012. Chapter 2
provides a brief overview of the PISA problem-solving framework. Chapter 3 presents results on the
performance of Australian students in problem solving. Results are compared to other participating
countries and economies, across jurisdictions and for different social groups. Chapter 4 presents a
discussion of students’ strengths and weaknesses in performing certain types of tasks. The final chapter
examines students’ attitudes related to problem solving.
Further information
This report focuses on the computer-based assessment of problem solving, which was offered for the
first time in PISA 2012. Details about the PISA 2012 paper-based assessment of mathematical literacy,
scientific literacy and reading literacy, and the computer-based assessment of mathematical literacy and
reading literacy can be found in the national report, PISA 2012: How Australia measures up.
Further information about PISA in Australia is available from the national PISA website:
www.acer.edu.au/ozpisa/.
6
Thinking it through: Australian students’ skills in problem solving
CHAPTER 2
The assessment of problem solving
The aim of the PISA 2012 problem-solving assessment was to assess an individual’s problem-solving
competency.1 This chapter describes the framework underlying the PISA 2012 computer-based, problemsolving assessment. It commences with a definition of problem solving, followed by a discussion of the
main elements of the problem-solving assessment framework and an overview of the general structure
of the assessment. The last section presents examples of problem-solving items from the PISA 2012
assessment2.
How is problem solving defined in PISA?
In PISA 2012, problem solving has been defined as:
an individual’s capacity to engage in cognitive processing to understand and resolve
problem situations where a method of solution is not immediately obvious. It includes
the willingness to engage with such situations in order to achieve one’s potential as a
constructive and reflective citizen (OECD, 2014, p. 30).
How is problem solving assessed in PISA?
The main features of the PISA 2012 problem-solving assessment framework are shown in Figure 2.1
(p. 8). The PISA framework for problem solving is organised into three aspects: the nature of the problem
situation; the problem-solving process; and the problem context.
For ease of reading, problem-solving competence will be referred to as problem solving.
Details about the problem-solving framework and sample items have been taken from OECD (2014).
1
2
7
NATURE OF THE
PROBLEM SITUATION
Is all the information needed
to solve the problem disclosed
at the outset?
Static: all relevant information for solving the problem is
disclosed at the outset
Interactive: not all information is disclosed; some information
can be uncovered by exploring the problem situation
Exploring and understanding: building mental
representations of each of the pieces of information presented
in the problem
PROBLEM SOLVING
PROCESSES
What are the main cognitive
processes involved in the
particular task?
Representing and formulating: constructing graphical,
tabular, symbolic or verbal representations of the problem
situation and formulating hypotheses about the relevant
factors and the relationships between them
Planning and executing: devising a plan by setting goals
and subgoals, and executing the sequential steps identified in
the plan
Monitoring and reflecting: monitoring progress, reacting
to feedback, and reflecting on the solution, the information
provided with the problem or the strategy adopted
PROBLEM CONTEXT
In what everyday scenario is the
problem embedded?
Setting: does the scenario involve a technological device?
- Technology (involves a technological device)
- Nontechnology (doesn’t involve a technological device)
Focus: what environment does the problem relate to?
- Personal (the student, family or close peers)
- Social (the community or society in general)
Figure 2.1 Main features of the PISA 2012 problem-solving assessment framework
Nature of the problem situation
This aspect of the problem-solving framework relates to how a problem is presented. In PISA, problem
situations are defined as static or interactive.
When all the relevant information about the problem is disclosed to the problem solver at the outset, the
problem situation is considered static. Static problem situations have a single goal and are not dynamic in
nature; that is, its state will not change of its own accord during the course of solving a problem. Solving
a jigsaw puzzle is an example of a static problem situation.
On the other hand, when the problem solver is not presented with all the information at the outset and
the problem solver is required to explore the situation to uncover additional relevant information, the
problem situation is considered interactive. Interactive problem situations can be dynamic in nature.
Using a train ticket vending machine, or another technological device, for the first time is an example of
an interactive problem situation. In this situation, it may not be obvious what steps need to be taken until
the user starts to interact with the ticket vending machine.
Problem-solving processes
The second aspect in the PISA 2012 problem-solving framework relates to the cognitive processes
involved in solving a problem. In assessing problem solving in PISA 2012, these cognitive processes have
been grouped into four processes:
8
Thinking it through: Australian students’ skills in problem solving
»» Exploring and understanding involves building mental representations of each of the pieces
of information presented in the problem by exploring the situation (observing it, interacting
with it, searching for information and finding limitations or obstacles) and demonstrating an
understanding of the given information and the information discovered while interacting with
the problem situation.
» Representing and formulating involves building a coherent mental representation of the
problem situation by selecting, mentally organising and integrating the relevant information with
relevant prior knowledge. This is achieved by representing the problem (by using tables, graphs,
symbols or words to represent aspects of the problem situation) and formulating hypotheses (by
identifying the relevant factors in a problem and their interrelationships, organising and critically
evaluating information).
» Planning and executing involves planning (clarifying the overall goal and setting subgoals,
where necessary), devising a plan or strategy to reach the goal and executing the plan (carrying
out a plan).
» Monitoring and ref lecting involves monitoring progress towards the goal at each stage
(including checking intermediate and final results, detecting unexpected events and taking
remedial action when required) and ref lecting on the solution from different perspectives,
critically evaluating assumptions and alternative solutions, identifying the need for additional
information or clarification, and communicating progress in a suitable manner.
In solving a particular problem, the processes may not be sequential and not all processes will be involved
in solving a problem. In the problem-solving assessment, single items were intended to have one of these
processes as their main focus, although often several processes occurred simultaneously, or in succession,
while solving a particular item.
A major distinction among tasks is between acquisition and use of knowledge. In knowledge-acquisition
tasks, the goal is for students to develop or refine their mental representation of the problem space.
Students need to generate and manipulate the information in a mental representation. The movement
is from concrete to abstract, from information to knowledge. In the context of the PISA assessment of
problem solving, knowledge-acquisition tasks may be classified either as exploring and understanding tasks
or as representing and formulating tasks. The distinction within knowledge-acquisition tasks between the
two processes is sometimes small, and may relate to the amount of scaffolding provided for exploring and
representing the problem space.
In knowledge-utilisation tasks, the goal is for students to solve a concrete problem. The movement is
from abstract to concrete, from knowledge to action. Knowledge-utilisation tasks correspond to the
process of planning and executing. Within the PISA assessment of problem solving, tasks would only be
classified as planning and executing if the execution of a plan is the dominant cognitive demand of the item
(and likewise for other problem-solving processes).
Monitoring and reflecting tasks are intentionally left out of this distinction, because they often combine both
knowledge-acquisition and knowledge-utilisation aspects.
Problem context
The problem context consists of two dimensions: the setting (technology and non-technology) and the
focus (personal or social). These have been identified to ensure that the assessment items cover a range of
contexts, are authentic and are of interest to 15-year-olds.
Problems are set in a technology or a nontechnology context. Examples of technological devices
include digital clocks, mobile phones and remote controls for appliances. Students are led to explore and
understand the functionality of a device, as preparation for controlling the device or for troubleshooting
The assessment of problem solving
9
its malfunctioning. Examples of nontechnology contexts include task scheduling, route planning and
decision making.
Problems are also set in a personal or a social focus. Personal contexts include those relating primarily to
the self, family and peer groups, while social contexts relate to broader situations in the community or
society in general.
An item about the rules that govern the function of an MP3 player would be classified as having a
technological and personal context, whereas an item relating to the seating plan for a birthday party has a
nontechnological and social context.
The PISA 2012 problem-solving assessment structure
The PISA 2012 problem-solving assessment framework serves as the conceptual basis for assessing
students’ proficiency in problem solving. Items were developed to reflect the concepts in the framework.
Structure of the computer-based assessment
As with the PISA 2012 paper-based assessment, items in the computer-based assessment were grouped
into units and the units were grouped into clusters. The PISA 2012 problem-solving assessment consisted
of 42 items with a total of 16 units. The units were grouped into four clusters, providing a total
assessment time of 80 minutes in problem solving. There were also four mathematical literacy clusters and
two reading literacy clusters to assess these literacy domains in a computer-based environment.3
Students were randomly assigned to one of 24 computer forms, each consisting of two clusters. Each form
consisted of none, one or two of the problem-solving clusters according to a separate balanced rotation
design, i.e., students answered only some of the items in the total item pool.
Students were allocated 40 minutes to complete the assessment. Only a basic level of information and
communication technologies (ICT) competence was required. Prior to the assessment, students undertook
a 20-minute tutorial that covered information about how to navigate the test interface and the different
response formats. The basic ICT skills that were needed to participate in the computer-based assessment
were: using a keyboard, using a mouse or touchpad, clicking radio buttons, drag-and-drop, scrolling, and
use of pull-down menus and hyperlinks.
The reading load was minimised by using stimulus material and task statements that were clear, simple
and brief, as well as using animations and pictures of diagrams.
Delivery of the computer-based assessment
The paper-based PISA 2003 problem-solving assessment typically assessed static problem situations.
However, advancements in software-development tools have provided an opportunity to assess students’
problem-solving skills in a new dimension. Students can be presented with relatively complex problems
that require direct interaction to uncover and discover the relevant information to solve the problem.
These interactive problem situations can be simulated in a testing setting using a computer, allowing
for a wider range of authentic, real-life scenarios. The capability to administer dynamic and interactive
problems also engages students’ interest more fully. Another benefit of measuring problem solving
through a computer-based assessment is the opportunity to collect data that relate to the processes and
strategies, such as the frequency, length and sequences of actions performed by students as they respond
to items.
The results of the computer-based mathematical literacy assessment and digital reading assessment can be found in the national report, PISA 2012:
How Australia measures up.
3
10
Thinking it through: Australian students’ skills in problem solving
The appearance of the test interface was consistent across items.4 For each item, the stimulus material was
shown in the top part of the screen, while the item appeared in the lower part of the screen, with a border
separating the stimulus from the item. All of the stimulus material was shown so students did not have to
scroll up or down to see all the information. There was a timing bar located in the top right-hand corner
of the screen, which showed students how much time was remaining in the assessment. There was also
another indicator of progress along the top left-hand side of the screen, identifying the number of items in
the unit and the item they were currently completing.
Items within units and units within clusters were delivered in a fixed, lockstep order. This meant that
as students completed an item (by clicking the arrow button), they were presented with a dialog box
displaying a message that they were about to move forward in the assessment and that it would not be
possible to return to this item. Students were then able to confirm that they wanted to move to the next
item or return to the current item.
Distribution of items
The PISA 2012 computer-based problem-solving assessment was based on 42 items. Items were developed
to measure how well students performed when the various problem-solving processes were exercised with
the two different types of problem situations across a range of contexts.
Items were classified by two main aspects, the main cognitive process involved in the particular task
and the nature of the problem situation. The assessment was constructed so that there was no strong
association between the main cognitive process involved in the task and the nature of the problem
situation; i.e., strengths and weaknesses in particular cognitive processes were unlikely to influence
strengths and weaknesses that were found in interactive or static tasks.
The distribution of items by the main cognitive process involved in the task was: 38% planning and
executing, 24% exploring and understanding, 21% representing and formulating, and 17% monitoring
and reflecting. Two-thirds of the items were classified as interactive and one-third of the items were
static. Table 2.1 presents the number of items in the problem-solving assessment, by cognitive process and
the nature of the problem situation.
Table 2.1 Classification of problem-solving items, by cognitive process and the nature of the problem situation
Problem-solving process
Nature of the
problem situation
Exploring and
understanding
Representing and
formulating
Planning and
executing
Monitoring and
reflecting
Total items
Static
5
2
6
2
15
Interactive
5
7
10
5
27
Total items
10
9
16
7
42
Items in the assessment were also classified by the particular context in which the problem situation
occured and according to their response format. About one-third of the items were simple multiplechoice items with one correct answer, or complex multiple-choice items with two or three separate
multiple-choice selections. Students were required to select their response by clicking a radio button or by
selecting a drop-down menu. These items were automatically coded.
The remaining two-thirds of items were constructed-response items. The majority of these items (80%)
were closed constructed-response items that required students to construct their response by entering
a number, dragging shapes, drawing lines between points or highlighting part of a diagram. These
closed constructed-response items were automatically coded. The remaining constructed-response items
were open constructed-response items, where students wrote a short explanation to show the method
T he computer-based assessment was delivered on computers running Windows XP, Windows Vista or Windows 7. Monitors used to display the
computer-based assessment had to support at least 1024 by 768 image resolution. 4
The assessment of problem solving
11
and thought process they had used in constructing their response. Trained coders coded these open
constructed-response items.
Scaling the problem-solving tasks
The assessment design—similar to those used in the regular PISA assessments of mathematics, science and
reading—allowed a single scale of proficiency in problem solving to be constructed.
The scale for problem solving was constructed using item-response theory, with each item associated with
a particular point on the scale indicating its difficulty and each student’s performance associated with a
particular point on the same scale indicating their estimated problem-solving proficiency. On this scale,
the relative difficulty of items in an assessment can be estimated by considering the proportion of students
responding to each item correctly. It is possible to estimate the location of individual students and to
describe the degree of problem solving that they possess.
The relationship between items and students on the problem-solving scale (shown in Figure 2.2) is
probabilistic. The estimate of student proficiency reflects the kinds of tasks they would be expected to
successfully complete. A student whose ability places them at a certain point on the PISA problem-solving
scale would most likely be able to successfully complete tasks at or below that location. They would
increasingly be more likely to be able to complete tasks located at progressively lower points on the scale.
They would be less likely to be able to complete tasks above that point on the scale and they would be
increasingly less likely to be able to complete tasks located at progressively higher points on the scale.
Problem-solving
scale
Item VI
Items with relatively
high difficulty
Items with moderate
difficulty
Items with relatively
low difficulty
Student A, with
relatively high
proficiency
It is expected that student A will be able
to complete items I to V successfully,
and probably item VI as well.
Student B,
with moderate
proficiency
It is expected that student B will be able
to complete items I, II and III successfully,
will have a lower probability of completing
item IV and is unlikely to complete items
V and VI successfully.
Item V
Item IV
Item III
Item II
Item I
Student C, with
relatively low
proficiency
It is expected that student C will be unable
to complete items II to VI successfully,
and will also have a low probability of
completing item I successfully.
Figure 2.2 The relationship between items and students on the PISA problem-solving scale
Defining problem-solving proficiency levels in PISA 2012
The PISA 2012 problem-solving assessment provides an overall problem-solving proficiency scale, which
draws on all problem-solving items in the assessment. The problem-solving scale was constructed to have
a mean score of 500 points across OECD countries and a standard deviation of 100 score points. Twothirds of the students across OECD countries scored between 400 and 600 points.
12
Thinking it through: Australian students’ skills in problem solving
While mean scores provide a convenient summary of student performance, proficiency levels are
developed in PISA to provide a description of the knowledge and skills students could be expected to
have at particular levels. The problem-solving scale is divided into six proficiency levels. The proficiency
levels range in difficulty from the lowest described level, Level 1—which corresponds to an elementary
level of problem-solving skills—to the highest described level, Level 6—which corresponds to the
advanced level of problem-solving skills. There is also an unbounded level, below Level 1—which does
not have a description about these students, as there is an insufficient number of items on which to base a
description of these students’ problem-solving proficiency.
Students with a proficiency score within the range of Level 1 are expected to complete most Level 1 tasks
successfully, but are unlikely to be able to complete tasks at higher levels. Students with scores in the
Level 6 range are likely to be able to successfully complete all tasks included in the PISA assessment of
problem solving.
The descriptions of what students can typically do at each of the proficiency levels in problem solving are
shown in Figure 2.3 (p. 14). A difference of 65 score points represents one proficiency level on the PISA
2012 problem-solving scale.
Students who perform at Level 5 or 6 (scoring 618 points or higher) are considered top-performing
students. These students are highly proficient in problem solving. Students who are placed at Level 1
or below (scoring 422 points or lower) are considered low-performing students. These students have
not reached Level 2, which has been defined internationally as a baseline proficiency level and defines
the level of performance on the PISA scale at which students begin to demonstrate the problem-solving
competencies that will enable them to actively participate in the 21st century work force and contribute
as productive citizens.
Students who perform below Level 1 can not be reliably described because there are not enough problemsolving items in this lower region of the scale. These students show limited problem-solving skills.
Sample problem-solving items
A small number of items have been publicly released to illustrate the types of items that students
responded to in the computer-based assessment. Figure 2.4 (p. 15) presents how these items map onto
the described problem-solving proficiency scale. The most difficult items are located at the top of the
figure at the higher proficiency levels and the least difficult items are located at the bottom in the lower
levels. Cut-off score points between proficiency levels are also displayed. Each of the items is placed in the
relevant proficiency level according to the difficulty of the item (the number in brackets) and is identified
by its main problem-solving process and nature of the problem situation.
Items included in the same unit can have a range of difficulties. For example, the unit Tickets comprises
items at Levels 2 to 5, showing that a single unit may cover a broad section of the PISA problem-solving
scale. Some items, for example, the second item in the unit Climate control is a partial credit item. If
students correctly responded to the item, they were rewarded with full credit, but it was possible for some
students to be rewarded with partial credit.
Only a few tasks in the problem-solving assessment are associated with difficulty levels below Level 1.
Among the released items, one item, the first question in the unit Traffic, is located below the lowest level
of proficiency described. All of the following interactive sample items are set in technology contexts. The
assessment also included interactive problems in nontechnology contexts; for example, asking students to
orient themselves in a maze.
The assessment of problem solving
13
Proficiency level
6
What students can typically do at each level
Students can develop complete, coherent mental models of diverse problem scenarios, enabling them to solve complex
problems efficiently. They can explore a scenario in a highly strategic manner to understand all information pertaining to the
problem. The information may be presented in different formats, requiring interpretation and integration of related parts. When
confronted with very complex devices, such as home appliances that work in an unusual or unexpected manner, they quickly
learn how to control the devices to achieve a goal in an optimal way. Level 6 problem-solvers can set up general hypotheses
about a system and thoroughly test them. They can follow a premise through to a logical conclusion or recognise when there is
not enough information available to reach one. In order to reach a solution, these highly proficient problem-solvers can create
complex, flexible, multistep plans that they continually monitor during execution. Where necessary, they modify their strategies,
taking all constraints into account, both explicit and implicit.
683 score points
5
Students can systematically explore a complex problem scenario to gain an understanding of how relevant information is
structured. When faced with unfamiliar, moderately complex devices, such as vending machines or home appliances, they
respond quickly to feedback in order to control the device. In order to reach a solution, Level 5 problem-solvers think ahead to
find the best strategy that addresses all the given constraints. They can immediately adjust their plans or backtrack when they
detect unexpected difficulties or when they make mistakes that take them off course.
618 score points
4
Students can explore a moderately complex problem scenario in a focused way. They grasp the links among the components
of the scenario that are required to solve the problem. They can control moderately complex digital devices, such as unfamiliar
vending machines or home appliances, but they don’t always do so efficiently. These students can plan a few steps ahead and
monitor the progress of their plans. They are usually able to adjust these plans or reformulate a goal in light of feedback. They
can systematically try out different possibilities and check whether multiple conditions have been satisfied. They can form a
hypothesis about why a system is malfunctioning and describe how to test it.
553 score points
3
Students can handle information presented in several different formats. They can explore a problem scenario and infer simple
relationships among its components. They can control simple digital devices, but have trouble with more complex devices.
Problem-solvers at Level 3 can fully deal with one condition; for example, by generating several solutions and checking to see
whether these satisfy the condition. When there are multiple conditions or interrelated features, they can hold one variable
constant to see the effect of change on the other variables. They can devise and execute tests to confirm or refute a given
hypothesis. They understand the need to plan ahead and monitor progress, and are able to try a different option if necessary.
488 score points
2
Students can explore an unfamiliar problem scenario and understand a small part of it. They try, but only partially succeed,
to understand and control digital devices with unfamiliar controls, such as home appliances and vending machines. Level 2
problem-solvers can test a simple hypothesis that is given to them and can solve a problem that has a single, specific
constraint. They can plan and carry out one step at a time to achieve a subgoal and have some capacity to monitor overall
progress towards a solution.
423 score points
1
Students can explore a problem scenario only in a limited way, but tend to do so only when they have encountered very similar
situations before. Based on their observations of familiar scenarios, these students are able only to partially describe the
behaviour of a simple, everyday device. In general, students at Level 1 can solve straightforward problems provided there is a
simple condition to be satisfied and there are only one or two steps to be performed to reach the goal. Level 1 students tend not
to be able to plan ahead or set subgoals.
358 score points
Figure 2.3 Summary descriptions of the six levels on the problem-solving proficiency scale
14
Thinking it through: Australian students’ skills in problem solving
Proficiency level
6
Item
Task
score
Main problem-solving process
Nature of the problem
Robot cleaner
Item 3 full credit (CP002Q06)
701
Representing and formulating
Static
Climate control
Item 2 full credit (CP025Q02)
672
Planning and executing
Interactive
Tickets
Item 2 full credit (CP038Q01)
638
Exploring and understanding
Static
Climate control
Item 2 partial credit (CP0025Q02)
592
Planning and executing
Interactive
Tickets
Item 3 (CP038Q03)
579
Monitoring and reflecting
Interactive
Robot cleaner
Item 2 (CP002Q07)
559
Exploring and understanding
Static
Tickets
Item 1 (CP038Q02)
526
Planning and executing
Interactive
Climate control
Item 1 full credit (CP025Q01)
523
Climate control
Item 1 partial credit (CP025Q01)
Representing and formulating
Interactive
492
Robot cleaner
Item 1 (CP002Q08)
490
Exploring and understanding
Static
Tickets
Item 2 partial credit (CP083Q01)
453
Exploring and understanding
Static
Traffic
Item 2 (CP007Q02)
446
Planning and executing
Static
Robot cleaner
Item 3 partial credit (CP002Q06)
414
Representing and formulating
Static
Traffic
Item 3 (CP007Q03)
408
Monitoring and reflecting
Static
Traffic
Item 1 (CP007Q01)
340
Planning and executing
Static
683 score points
5
618 score points
4
553 score points
3
488 score points
2
423 score points
1
358 score points
Below 1
Figure 2.4 M
ap of selected problem-solving items, illustrating the proficiency level and assigned problem-solving process
and nature of the problem
The assessment of problem solving
15
The four units, Robot cleaner, Climate control, Tickets and Traffic are described below. For each unit, a
screenshot of the stimulus information is provided, together with a brief description of the context of the
unit. This is followed by a screenshot and description of each item from that unit.5
Robot cleaner
The animation shows the movement of a new robotic vacuum cleaner.
It is being tested.
Click the START button to see what the vacuum cleaner does when
it meets different types of objects.
You can use the RESET button to place the vacuum cleaner back in its
starting position at any time.
The unit Robot cleaner presents students with an animation showing the behaviour of a robot cleaner in
a room. The robotic vacuum cleaner moves forward until it meets an obstacle and then behaves according
to a few deterministic rules, depending on the kind of obstacle. Students can run the animation as many
times as they wish to observe this behaviour.
Despite the animated task prompt, the problem situations in this unit are static, because the student
cannot intervene to change the behaviour of the vacuum cleaner or aspects of the environment. The
context for the items in this unit is classified as social and nontechnological.
Item 1: Robot cleaner CP002Q08
What does the vacuum cleaner do when it meets a red block?
m
It immediately moves to another red block.
m
It turns and moves to the nearest yellow block.
m
It turns a quarter circle (90 degrees) and moves forward until it meets something else.
m
It turns a half circle (180 degrees) and moves forward until it meets something else.
In the first item, students must understand the behaviour of the vacuum cleaner when it meets a red
block. The item is classified as exploring and understanding. To show their understanding, students
need to select among a list of four options. Based on observation of the animation, the description that
corresponds to the behaviour of the robot cleaner in this situation is: “It turns a quarter circle (90 degrees)
and moves forward until it meets something else.”
These units are available for viewing at http://cbasq.acer.edu.au
5
16
Thinking it through: Australian students’ skills in problem solving
Item 2: Robot cleaner CP002Q07
At the beginning of the animation, the vacuum cleaner is facing the left wall. By the end of the
animation it has pushed two yellow blocks.
If, instead of facing the left wall at the beginning of the animation, the vacuum cleaner was facing
the right wall, how many yellow blocks would it have pushed by the end of the animation?
m
0m
1
m
2m
3
In the second item, students must predict the behaviour of the vacuum cleaner using spatial reasoning.
How many obstacles would the vacuum cleaner encounter if it started in a different position? This item
is also an exploring and understanding item because the correct prediction of the robot’s behaviour
requires at least a partial understanding of the rules and careful observation of the animation to grasp the
information needed. It is made easier if the student notes that the new starting position corresponds to an
intermediate state of the robot’s trajectory in the animation. Response options are provided.
Item 3: Robot cleaner CP002Q06
The vacuum cleaner’s behaviour follows a set of rules. Based on the animation, write a rule that
describes what the vacuum cleaner does when it meets a yellow block.
The final item in this unit is classified as representing and formulating. It asks students to describe the
behaviour of the robot cleaner when it meets a yellow block. In contrast to the first item, students must
formulate the answer themselves by entering it in a text box. This item requires expert scoring for credit.
Full credit answers are those that describe both of the rules that govern the robot’s behaviour (for
example, “it pushes the yellow block as far as it can and then turns around”). Partial credit was available
for answers that only partially described the behaviour (for example, by listing only one of the two rules).
Only a small percentage of students across participating countries obtained full credit for this item.
The assessment of problem solving
17
Climate control
You have no instructions for your new air conditioner. You need to work
out how to use it.
You can change the top, central and bottom controls on the left by using
the sliders ( ). The initial setting for each control is indicated by p.
By clicking APPLY, you will see any changes in the temperature and
humidity of the room in the temperature and humidity graphs. The box
to the left of each graph shows the current level of temperature or
humidity.
In the unit Climate control, students are told that they have a new air conditioner but no instructions for
it. Students can use three controls (sliders) to vary temperature and humidity levels, but first they need to
understand which control does what. A measure of temperature and humidity in the room appears in the
top right-hand part of the screen, both in numerical and graphical form. All items in this unit present an
interactive problem situation, with context classified as personal and technological.
The unit Climate control is an example of a system of causal relations involving only a few variables that
have to be explored and controlled in order to reach the assigned goal states. In the first knowledgegeneration phase, the student has to control up to three input variables. The increase in the level of an
input variable leads to an increase, a decrease, a mixed effect (increase and decrease for different variables)
or no effect in one or more output variables. Students typically have to demonstrate rule knowledge after
this first phase. Students are then asked to control the system to reach a certain target by choosing the
appropriate input levels.
Item 1: Climate control CP025Q01
Find out whether each control influences
temperature and humidity by changing the
sliders. You can start again by clicking RESET.
Draw lines in the diagram on the right to show
what each control influences.
To draw a line, click on a control and then click
on either Temperature or Humidity. You can
remove any line by clicking on it.
In the first item, students are invited to change the sliders to find out whether each control influences
the temperature or the humidity level. The problem-solving process for this item is representing and
formulating: the student must experiment to determine which controls have an impact on temperature
and which on humidity, and then represent the causal relations by drawing arrows between the three
controls and the two outputs (temperature and humidity). There is no restriction on the number of
rounds of exploration that the student is allowed. Full credit for this question requires that the causal
diagram is correctly completed. Partial credit for this question is given if the student explores the
relationships among variables efficiently, by varying only one input at a time, but fails to correctly
represent them in a diagram.
18
Thinking it through: Australian students’ skills in problem solving
Item 2: Climate control CP025Q02
The correct relationship between the three
controls, Temperature and Humidity is shown
on the right.
Use the controls to set the temperature and
humidity to the target levels. Do this in a
maximum of four steps. The target levels are
shown by the red bands across the Temperature
and Humidity graphs. The range of values for
each target level is 18–20 and is shown to the
left of each red band. You can only click
APPLY four times and there is no RESET
button.
The second item asks students to apply their new knowledge of how the air conditioner works to set
temperature and humidity at specified target levels (lower than the initial state). This is a planning
and executing item. To ensure that no further exploration is needed beyond the one conducted in the
previous item, a diagram shows how the controls are related to temperature and humidity levels (students
could not return to any previous item during the test). Because only four rounds of manipulation are
permitted, students need to plan a few steps ahead and use a systematic, if simple, strategy to succeed in
this task. The target levels of temperature and humidity provided can be reached in several ways within
four steps—the minimum number of steps needed is two—and a mistake can often be corrected, if
immediate remedial action is taken. A possible strategy, for instance, is to set separate subgoals and to
focus on temperature and humidity in successive steps. If the student is able to bring temperature and
humidity both closer to their target levels within the four rounds of manipulation permitted, but does not
reach the target for both, then partial credit is given.
The assessment of problem solving
19
Tickets
A train station has an automated ticketing machine. You use the touch
screen on the right to buy a ticket. You must make three choices.
• Choose the train network you want (subway or country).
• Choose the type of fare (full or concession).
• Choose a daily ticket or a ticket for a specified number of trips.
Daily tickets give you unlimited travel on the day of purchase. If you
buy a ticket with a specified number of trips, you can use the trips
on different days.
The BUY button appears when you have made these three choices.
There is a cancel button that can be used at any time BEFORE you
press the BUY button.
In the unit Tickets, students are invited to imagine that they have just arrived at a train station that
has an automated ticketing machine. The context for the items in this unit is classified as social and
technological. At the machine, students can buy subway or country train tickets, with full or concession
fares; and they can choose daily tickets or a ticket for a specified number of trips. All items in this unit
present an interactive problem situation: students are required to engage with the unfamiliar machine and
to use the machine to satisfy their needs.
Item 1: Tickets CP038Q02
Buy a full fare, country train ticket with two individual trips.
Once you have pressed BUY, you cannot return to the question.
In the first item, students are invited to buy a full fare, country train ticket with two individual trips.
This item measures the process of planning and executing. Students first have to select the network
(country trains), choose the fare type (full fare), select between a daily ticket and one for multiple
individual trips, and finally indicate the number of trips (two). The solution requires multiple steps and
the instructions are not given in the same order as they need to be applied. This is a relatively linear
problem compared to the following ones, but it is the first encounter with this machine, which increases
its level of difficulty relative to the following items.
Item 2: Tickets CP038Q01
You plan to take four trips around the city on the subway today. You are a student, so you can use
concession fares.
Use the ticketing machine to find the cheapest ticket and press BUY.
Once you have pressed BUY, you cannot return to the question.
20
Thinking it through: Australian students’ skills in problem solving
In the second item, students are asked to find and buy the cheapest ticket that allows them to take four
trips around the city on the subway, within a single day. As students, they can use concession fares. This
item is classified as exploring and understanding because this is the most crucial problem-solving process
involved. Indeed, to accomplish the task, students must use a targeted exploration strategy, first generating
at least the two most obvious possible alternatives (a daily subway ticket with concession fare, or an
individual concession fare ticket with four trips), and then verifying which of these is the cheapest ticket.
If students visit both screens before buying the cheapest ticket (which is the individual ticket with four
trips), they are given full credit. Students who buy one of the two tickets without comparing the prices
for the two only earn partial credit. Solving this problem involves multiple steps.
Item 3: Tickets CP038Q03
You want to buy a ticket with two individual trips for the city subway. You are a student, so you can
use concession fares.
Use the ticketing machine to purchase the best ticket available.
In the third item, students are asked to buy a ticket for two individual trips on the subway. They are
told that they are eligible for concession fares. This item is classified as monitoring and reflecting, since
it requires them to modify their initial plan (to buy concession-fare tickets for the subway). When
concession fares are selected, the machine says that “there are no tickets of this type available”. In this
item, students must realise that it is not possible to carry through their initial plan and so must adjust this
plan by buying a full-fare ticket for the subway instead.
The assessment of problem solving
21
Traffic
Here is a map of a system of roads that links the
suburbs within a city. The map shows the travel time
in minutes at 7.00 am on each section of the road.
You can add a road to your route by clicking on it.
Clicking on a road highlights the road and adds the
time to the Total Time box.
You can remove a road from your route by clicking on
it again. You can use the RESET button to remove all
roads from your route.
In the unit Traffic, students are given a map of a road network with travel times indicated. While this
is a unit with static items because all the information about travel times is provided at the outset, it still
exploits the advantages of computer delivery. Students can click on the map to highlight a route, with a
calculator in the bottom left-hand corner adding up travel times for the selected route. The context for
the items in this unit is classified as social and nontechnological.
Item 1: Traffic CP007Q01
Pepe is at Sakharov and wants to travel to Emerald. He wants to complete his trip as quickly as
possible. What is the shortest time for his trip?
m 20 minutes
m 21 minutes
m 24 minutes
m 28 minutes
In the first item—a planning and executing item—students are asked about the shortest time to travel
from Sakharov to Emerald, two relatively close points shown on the map. Four response options
are provided.
Item 2: Traffic CP007Q02
Maria wants to travel from Diamond to Einstein. The quickest route takes 31 minutes.
Highlight this route.
The second item is a similar planning and executing item. It asks students to find the quickest route
between Diamond and Einstein, two distant points on the map. This time, students must provide their
answer by highlighting this route. Students can use the indication that the quickest route takes 31 minutes
to avoid generating all possible alternatives systematically; instead, they can explore the network in a
targeted way to find the route that takes 31 minutes.
Item 3: Traffic CP007Q03
Julio lives in Silver, Maria lives in Lincoln and Don lives in Nobel. They want to meet in a suburb on
the map. No-one wants to travel for more than 15 minutes.
Where could they meet?
In the third item, students have to use a drop-down menu to select the meeting point that satisfies a
condition on travel times for all three participants to meet. The demand in this third item is classified
as a monitoring and reflecting item, because students have to evaluate possible solutions against a
given condition.
22
Thinking it through: Australian students’ skills in problem solving
CHAPTER 3
Australian students’ performance
in problem solving
This chapter presents Australian students’ performance in problem solving in PISA 2012. Results
are reported by means (average scores) and by proficiency levels across the problem-solving scale.
Comparisons of student performance in problem solving are provided at an international level, describing
Australia’s performance relative to other participating countries, and at a national level, where the
performance of different (social) groups is examined.
Australia’s problem-solving performance from an international
perspective
Problem-solving performance across countries
Overall, Australian students performed very well in problem solving, achieving a mean score of
523 points. Seven countries (two OECD and five partner countries and economies) performed
significantly higher than Australia. Singapore and Korea, the highest performing countries in problem
solving, achieved a mean score of 562 points and 561 points respectively, followed by Japan (552 score
points), Macao–China and Hong Kong–China (540 score points), Shanghai–China (536 score points)
and Chinese Taipei (534 score points). Three countries, Canada, Finland and England, performed not
significantly different to Australia, while all other countries, including the United States and Ireland,
performed at a significantly lower level than Australia. The lowest scoring countries were Uruguay,
Bulgaria and Colombia with scores of 403 points or lower.
Australia was one of 19 countries (14 OECD countries and five partner countries) that achieved a mean
score that was significantly higher than the OECD average. These countries were: Singapore, Korea,
Japan, Macao–China, Hong Kong–China, Shanghai–China, Chinese Taipei, Canada, Australia, Finland,
England, Estonia, France, the Netherlands, Italy, Czech Republic, Germany, the United States and
Belgium. Five countries (Austria, Norway, Ireland, Denmark and Portugal) achieved a mean score that
was not significantly different from the OECD average, and 20 countries (nine OECD countries and
11 partner countries) performed significantly lower than the OECD average. These countries were:
Sweden, the Russian Federation, the Slovak Republic, Poland, Spain, Slovenia, Serbia, Croatia, Hungary,
Turkey, Israel, Chile, Cyprus, Brazil, Malaysia, the United Arab Emirates, Montenegro, Uruguay,
Bulgaria and Colombia.
23
How problem solving is reported in PISA
Similar to the reporting of results for other assessed domains in PISA, statistics such as
mean scores and measures of distribution of performance and proficiency levels are used to
examine students’ performance.
Mean scores and distribution of scores
Mean scores provide a summary of students’ performance and allow comparisons of the
relative standing between different countries and different subgroups. As problem solving
was assessed for the first time as a computer-based assessment in PISA 2012, the mean
score across OECD countries was set at 500 score points, with a standard deviation of 100
score points. This establishes the benchmark against which each country’s problem-solving
performance in PISA 2012 is compared.
The distribution of scores along the problem-solving scale also provides further detail about
students’ performance. Results are reported at the 5th, 10th, 25th, 75th, 90th and 95th
percentiles in graphical format to observe the variation in students’ performance within a
country or subgroup.
Proficiency levels
Proficiency levels provide results in descriptive terms, where descriptions of the skills and
knowledge students can typically demonstrate are attached to achievement results. The
problem-solving proficiency scale spans from Level 1 (the lowest proficiency level) to Level 6
(the highest proficiency level). Students who are placed at Level 5 or 6 are considered
top-performing students, while students who fail to reach Level 2 are considered lowperforming students.
Interpreting differences in PISA scores: How big is ‘big’?
How do we go about understanding the difference in average problemsolving scores between two groups of students?
A difference of 65 score points represents one proficiency level on the PISA problemsolving scale. In substantive terms, this can be considered a comparatively large difference
in students’ performance. For example, compare the skill set for those students who
are proficient at Level 2 and those students who are proficient at Level 3. Students who
perform at Level 2 on the problem-solving scale are only starting to demonstrate problemsolving competence. They can explore an unfamiliar problem scenario and understand a
small part of it, and they can test a simple hypothesis and solve a problem that has a single,
specific constraint. In contrast, students who reach Level 3 are proficient with the tasks at
Level 2 and can also explore a problem scenario and infer simple relationships among its
components, and they can devise and execute tests to confirm or refute a given hypothesis.
The difference in average performance between the highest- and lowest-performing
countries is 163 score points. Across the OECD, the difference between the highest- and
lowest-performing countries is 113 score points. For comparison, the difference between
the highest- and lowest-performing jurisdictions is 36 score points.
Treating all OECD countries as a single unit, one standard deviation in the distribution of
students’ performance on the problem-solving scale corresponds to 100 score points, which
means that—on average within OECD countries—two-thirds of the student population have
scores within 100 score points of the OECD mean (500 score points).
24
Thinking it through: Australian students’ skills in problem solving
The difference in mean scores between students in the 5th and 95th percentiles varied considerably
within countries, showing no clear relationship between average achievement and the degree of spread
in students’ scores. In Australia, the difference in achievement between the most capable problem
solvers (those students scoring at the 95th percentile) and the least capable (those students scoring at
the 5th percentile) was 320 score points. The OECD average between the 5th and 95th percentiles was
314 score points.
Among the OECD countries, the widest differences between the lowest and highest achieving students
were found in Israel (405 score points), Belgium (348 score points), Spain (346 score points) and Hungary
(345 score points). For partner countries, the widest differences were found in Bulgaria (351 score points)
and the United Arab Emirates (347 score points).
The narrowest differences between the lowest and highest achieving students were found in the partner
country Macao–China, with 259 score points between the 5th and 95th percentiles, and in Turkey, with
262 score points between students in the 5th and 95th percentiles.
Figure 3.1 (p. 26) provides the average problem-solving scores, along with the standard errors, confidence
intervals around the average and the difference between the 5th and 95th percentiles. In addition, this
figure also shows the graphical distribution of student performance. Countries are shown in order from
the highest to the lowest average problem-solving score and the three colour bands indicate whether a
particular country has performed at a significantly higher or lower level, or whether they performed at a
level not significantly different to Australia.
Australian students’ performance in problem solving
25
Significantly higher
than Australia
Not significantly
different from
Australia
Significantly lower
than Australia
Country
Mean score
SE
Confidence
interval
Difference
between
5th and 95th
percentiles
Singapore
562
1.2
560–565
312
Korea
561
4.3
553–570
292
Japan
552
3.1
546–558
280
Macao–China
540
1.0
538–542
259
Hong Kong–China
540
3.9
532–547
304
Shanghai–China
536
3.3
530–543
295
Chinese Taipei
534
2.9
529–540
297
Canada
526
2.4
521–530
327
Australia
523
1.9
519–527
320
Finland
523
2.3
518–527
307
England
517
4.2
509–525
315
Estonia
515
2.5
510–520
287
France
511
3.4
504–518
313
Netherlands
511
4.4
502–519
326
Italy
510
4.0
502–518
293
Czech Republic
509
3.1
503–515
312
Germany
509
3.6
502–516
324
United States
508
3.9
500–516
306
Belgium
508
2.5
503–513
348
Austria
506
3.6
499–513
305
Norway
503
3.3
497–510
337
OECD average
500
0.7
499–501
314
Ireland
498
3.2
492–505
307
Denmark
497
2.9
491–503
302
Portugal
494
3.6
487–501
288
Sweden
491
2.9
485–496
316
Russian Federation
489
3.4
482–496
290
Slovak Republic
483
3.6
476–490
324
Poland
481
4.4
472–489
313
Spain
477
4.1
469–485
346
Slovenia
476
1.5
473–479
318
Serbia
473
3.1
467–480
294
Croatia
466
3.9
459–474
302
345
Hungary
459
4.0
451–467
Turkey
454
4.0
447–462
262
Israel
454
5.5
443–465
405
Chile
448
3.7
441–455
283
Cyprus
445
1.4
442–448
326
299
Brazil
428
4.7
419–438
Malaysia
422
3.5
416–429
274
United Arab Emirates
411
2.8
406–417
347
Montenegro
407
1.2
404–409
300
Uruguay
403
3.5
397–410
322
Bulgaria
402
5.1
392–412
351
Colombia
399
3.5
392–406
300
Distribution of scores
200
300
400
500
600
Mean problem-solving performance
Figure 3.1 Mean scores and distribution of students’ performance on the problem-solving scale
26
Thinking it through: Australian students’ skills in problem solving
700
800
Each country’s results are represented in horizontal bars with various colours. On the left
end of the bar is the 5th percentile—this is the score below which 5% of the students
have scored. The next two lines indicate the 10th percentile and the 25th percentile. The
next line at the left of the white band is the lower limit of the confidence interval for the
mean—i.e., there is 95% confidence that the mean will lie in this white band. The line in the
centre of the white band is the mean. The lines to the right of the white band indicate the
75th, 90th and 95th percentiles.
Confidence
interval
10th
percentile
5th
percentile
25th
percentile
Mean
90th
percentile
75th
percentile
95th
percentile
Students’ problem-solving competencies across countries
PISA uses proficiency levels to provide further meaning about students’ capabilities in problem solving.
Six levels of proficiency have been defined in problem solving, ranging from Level 1 (the lowest
proficiency level) to Level 6 (the highest proficiency level). A seventh proficiency level, below Level 1,
includes those students who are unable to successfully complete many of the items of Level 1 difficulty.
The average proportion of students at each problem-solving proficiency level by country is presented in
Figure 3.2 (p. 28). Countries have been ordered by the percentage of students classified as below Level 2,
the internationally assigned baseline benchmark. Countries with the lowest proportion of students below
Level 2 are placed at the top of the figure and countries with the highest proportion of students below Level 2
are placed at the bottom.
Students who achieved a score of 683 points or higher were placed at the highest proficiency level,
Level 6. These students are highly proficient problem solvers who can develop complete coherent, mental
models of diverse problem scenarios, enabling them to solve complex problems efficiently.
On average, 3% of students across OECD countries performed at this level. In Singapore, one in 10
students, and in Korea, around one in 12 students were highly skilled problem solvers. These two highest
performing countries had the greatest proportion of students who achieved a proficiency of Level 6. The
other countries ( Japan, Macao–China, Hong Kong–China, Shanghai–China and Chinese Taipei) which
performed significantly higher than Australia achieved between 3 and 5% of their students achieving
Level 6. Australia, along with Finland, was among the countries with the greatest proportion of students
placed at Level 6 with 4%. In nine countries (Brazil, Bulgaria, Chile, Colombia, Malaysia, Montenegro,
Turkey, the United Arab Emirates and Uruguay) less than 1% of students performed at Level 6.
Students proficient at Level 5 were able to systemically explore a complex problem scenario to gain an
understanding of how relevant information is structured and when faced with unfamiliar, moderately
complex devices, they could respond quickly to feedback in order to control the device.
Across OECD countries, 12% of students were proficient at Level 5 or higher. These students are
considered top performers. In Singapore, 30% of students were top performers, compared to 28% of
students in Korea and 22% of students in Japan. Almost one in five students were top performers in
Hong Kong–China and Chinese Taipei, while there were 18% of students in Shanghai–China, 17%
of students in Macao–China and Canada, and 16% of students in Australia who achieved these levels.
In the lower-performing countries of Colombia, Uruguay, Montenegro, Bulgaria and Malaysia, less
than 2% of students were top performers.
Australian students’ performance in problem solving
27
2 5
Korea
13
24
Japan
2 5
15
Singapore
2 6
14
Macao–China
2 6
18
3 7
16
Chinese Taipei
3 8
18
Shanghai–China
3 8
18
10
20
Estonia
4
11
22
5
10
Italy
5
11
Australia
5
11
19
England
6
11
20
France
7
10
21
11
20
26
26
15
4
27
26
14
24
7
12
21
Austria
7
12
22
Ireland
7
13
24
Germany
8
12
11
3
20
9
3
27
22
10 2
27
22
7
13
24
8
13
22
Portugal
7
14
Norway
8
13
28
10 3
19
26
9
28
18
25
18
25
10 3
22
11
Sweden
9
15
24
26
18
Slovak Republic
11
15
24
26
16
16
26
26
16
18
27
26
14
10
10
Serbia
27
28
16
7 2
6 2
6
4
Spain
13
15
24
24
16
6 2
11
17
25
24
16
6
Croatia
12
20
17
18
27
23
24
Turkey
11
25
31
Chile
15
23
29
22
Israel
17
20
Cyprus
22
Brazil
25
28
13
21
19
14
10 3
27
17
7 2
28
16
5
23
United Arab Emirates
30
25
22
14
6 2
Bulgaria
33
23
22
14
6
30
27
24
Uruguay
32
26
22
100
33
80
60
28
40
20
7 2
20
Malaysia
Colombia
9 2
26
Montenegro
14
5
13
22
11
0
20
4
5
9 2
22
20
21
13
22
3
6
Slovenia
Hungary
5
4
40
60
Percentage of students
Below Level 1
Level 1
Level 2
Level 3
Level 4
Level 5
Level 6
Note: In cases in which the proportion of students in a proficiency level is 1% or less, the level still appears in the figure
but the numeric label 1 does not. This convention has been used for all figures about proficiency levels in this chapter.
Figure 3.2 Percentage of students across the problem-solving proficiency scale, by country
28
Thinking it through: Australian students’ skills in problem solving
3
6
19
15
Poland
7 2
20
7
Russian Federation
7 2
22
26
22
9 2
19
26
Denmark
3
10 2
22
28
20
4
11
23
27
OECD average
12
23
26
5
9 2
23
28
Czech Republic
12
12
22
27
4
10 2
23
28
26
4
11
22
26
23
3
5
23
13
14
14
19
6
5
10
29
29
United States
9
17
27
27
Canada
8
20
30
5
Belgium
29
27
27
Finland
7
20
27
22
Hong Kong–China
Netherlands
29
80
100
Students proficient at Level 4 were able to explore a moderately complex problem scenario in a focused
way. They could understand the links among the components of the scenario that were required to solve
the problem and could control moderately complex digital devices, but not always so efficiently.
On average within the OECD, one in three students were proficient at Level 4 or higher. In Singapore,
Korea and Japan, over one in two students were able to complete Level 4 tasks and almost one in two
students in the other top-performing Asian countries performed at Level 4 or higher. In Australia
and Finland, 39% of students achieved Level 4 or higher, which was a similar proportion to Canada
(40%). Those countries that achieved the lowest mean score had less than one in 10 students placed at
these levels.
Students proficient at Level 3 were able to handle information presented in several different formats.
They could explore a problem scenario and infer simple relationships among its components. They
could control simple digital devices, but had trouble with more complex devices.
Across OECD countries, more than half (58%) of students were proficient at Level 3 or higher. In those
countries performing significantly higher than Australia, more than 70% of students performed at Level 3
or higher. In Australia, 65% of students achieved a proficiency of at least Level 3, which was similar to
Canada and Finland (66%), and England (64%).
Students proficient at Level 2 were able to explore an unfamiliar problem scenario and understand a small
part of it. They were only able to partially succeed in understanding and controlling digital devices with
unfamiliar controls. At this level of proficiency, students were able to engage with everyday problems,
make progress towards solving the problem and sometimes successfully solve the problem.
Across OECD countries, 80% of students were proficient at Level 2 or higher. Level 2 is considered a
baseline level of proficiency, where students begin to demonstrate the competencies in problem solving
that will enable them to effectively and productively participate in today’s society. The majority of
students (93% or higher) in Korea, Macao–China, Singapore and Japan, and 89% of students in
Hong Kong–China, Chinese Taipei and Shanghai–China, were placed at Level 2 or higher. In
Australia, England, France and Italy, 84% of students were proficient at Level 2 or higher. In the
United Arab Emirates, Bulgaria, Montenegro, Uruguay and Colombia, less than one in two students
achieved a proficiency of Level 2 or higher.
Students proficient at Level 1 were able to explore a problem scenario only in a limited way, but could
only do so when they had encountered similar situations before. Students at Level 1 were able to solve
straightforward problems provided there was a simple condition to be satisfied and the need for only one
or two steps to be performed to solve the problem.
Students who scored less than 358 points were placed below Level 1. Although there were insufficient
items to fully describe the proficiencies of students that were placed below Level 1, some of these students
were able to use an unsystematic strategy to solve a simple problem in a familiar context and they may be
able to find solutions when a limited number of well-defined possibilities are presented. Students who are
placed below Level 1 are limited in their ability to solve problems.
On average, one-fifth of students across the OECD were placed at Level 1 or below Level 1. Less than
one in 10 students from Singapore, Korea, Japan and Macao–China were placed at these low proficiency
levels. Students who were placed at Level 1 or below Level 1 are considered low performers. In Australia,
16% of students were low performers, which was a similar proportion to students in Canada and Finland
(15%), England and France (17%) and the United States (19%).
Problem-solving performance by sex across countries
On average across OECD countries, males scored significantly higher than females in problem solving,
on average by six score points. Differences between sexes were found to be in favour of males in
more countries than females. Figure 3.3 (p. 30) shows males significantly outperformed females in
Australian students’ performance in problem solving
29
Females
Country
Males
Mean score
SE
Mean score
SE
424
3.2
398
4.6
Bulgaria
410
5.3
394
5.8
Cyprus
449
2.0
440
1.8
United Arab Emirates
Finland
526
2.6
520
2.8
Montenegro
409
1.8
404
1.8
Slovenia
478
2.2
474
2.1
Sweden
493
3.1
489
3.7
Norway
505
3.8
502
3.6
Poland
481
4.6
481
4.9
Spain
476
4.1
478
4.8
Australia
522
2.2
524
2.4
United States
506
4.2
509
4.2
Hungary
457
4.3
461
5.0
France
509
3.5
513
4.0
Estonia
513
2.6
517
3.3
Netherlands
508
4.5
513
4.9
Ireland
496
3.2
501
4.8
Canada
523
2.5
528
2.8
England
514
4.6
520
5.4
Israel
451
4.1
457
8.9
Germany
505
3.7
512
4.1
OECD average
497
0.7
503
0.8
Czech Republic
505
3.5
513
3.9
Malaysia
419
4.0
427
3.9
Belgium
504
3.1
512
3.1
Russian Federation
485
3.7
493
3.9
Singapore
558
1.7
567
1.8
Denmark
492
2.9
502
3.7
Macao–China
535
1.3
546
1.5
Uruguay
398
3.8
409
4.0
Austria
500
4.1
512
4.4
Chinese Taipei
528
4.1
540
4.5
Korea
554
5.1
567
5.1
Chile
441
3.7
455
4.5
Hong Kong–China
532
4.8
546
4.6
Serbia
466
3.2
481
3.8
Turkey
447
4.6
462
4.3
Croatia
459
4.0
474
4.8
Portugal
486
3.6
502
4.0
Italy
500
4.5
518
5.2
Japan
542
3.0
561
4.1
Slovak Republic
472
4.1
494
4.2
Brazil
418
4.6
440
5.4
Shanghai–China
524
3.8
549
3.4
Colombia
385
3.9
415
4.1
Difference in mean score
Females
score
higher
40
30
20
Males
score
higher
10
Sex difference significant
0
10
20
Sex difference not significant
Figure 3.3 Mean scores and differences between sexes in students’ performance on the problem-solving scale, by country
30
Thinking it through: Australian students’ skills in problem solving
30
40
approximately half the countries. The largest differences in favour of males were found in Colombia,
Shanghai–China, Brazil and the Slovak Republic, with males scoring between 22 and 30 score points
higher than females. In five countries (11%), females performed significantly higher than males, with the
largest differences found in the United Arab Emirates (26 score points) and Bulgaria (16 score points).
Australian males achieved a mean score of 524 points, which was not significantly different to the mean
score of 522 points for females.
The average proportion of females and males at each level on the problem-solving scale for Australia and
the OECD average is shown in Figure 3.4. There were higher proportions of males than females who
achieved Level 5 or 6. In Australia, 18% of males compared to 16% of females, and across the OECD,
13% of males compared to 10% of females were top performers.
OECD average
Australia
The proportion of males and females who failed to reach Level 2 was similar. In Australia, 16% of males
compared to 15% of females failed to reach Level 2, while across the OECD the proportion of males and
females was the same at 22%.
Females
5
10
20
Males
5
11
19
Females
8
14
23
Males
9
13
21
100
80
Below Level 1
60
Level 1
40
Level 2
20
27
Level 4
Level 5
19
40
5
8 2
10 3
20
25
4
12
13
22
25
0
20
Percentage of students
Level 3
23
27
60
80
100
Level 6
Figure 3.4 Percentage of students across the problem-solving proficiency scale by sex, for Australia and the OECD average
Australia’s problem-solving performance in a national context
Problem-solving performance across the Australian jurisdictions
Figure 3.5 (p. 32) shows the mean scores and distribution of problem-solving scores for each jurisdiction.
The mean score and distribution for Australia and the highest performing country (Singapore) have also
been included for comparison. The mean scores ranged from 528 points in Western Australia to 490
points in Tasmania.
The Northern Territory, along with Tasmania, showed the widest distribution of scores, with a range
of 364 and 349 score points respectively between the lowest and highest performing students. South
Australia, with 305 score points, had the narrowest range between the 5th and 95th percentiles.
Australian students’ performance in problem solving
31
Difference
between 5th and
95th percentiles
Mean score
SE
Confidence
interval
ACT
526
3.7
518–533
338
NSW
525
3.5
518–532
328
VIC
523
4.1
515–531
311
QLD
522
3.4
515–529
317
SA
520
4.1
512–528
305
WA
528
4.0
520–536
314
TAS
490
4.0
482–498
349
NT
513
7.9
497–529
364
Australia
523
1.9
519–527
320
Singapore
562
1.2
560–565
312
OECD average
500
0.7
499–501
314
Jurisdiction
Distribution of scores
200
300
400
500
600
700
800
Mean problem-solving performance
Figure 3.5 Mean scores and distribution of students’ performance on the problem-solving scale, by jurisdiction
Table 3.1 is a multiple-comparison table that provides further details about the performance of each
jurisdiction compared to the other jurisdictions. Almost all the jurisdictions performed at a level not
significantly different to one another. Western Australia, the Australian Capital Territory, New South
Wales, Victoria, Queensland, South Australia and the Northern Territory were on a par with each other,
while Tasmania was the only jurisdiction that performed significantly lower than all other jurisdictions.
Six jurisdictions (Western Australia, the Australian Capital Territory, New South Wales, Victoria,
Queensland and South Australia) performed at a significantly higher level than the OECD average, the
Northern Territory performed at a level not significantly different to the OECD average, and Tasmania
performed significantly lower than the OECD average.
Table 3.1 Multiple comparisons of mean problem-solving performance, by jurisdiction
Jurisdiction
WA
ACT
NSW
VIC
QLD
SA
NT
TAS
OECD
average
l
l
l
l
l
l
p
p
l
l
l
l
l
p
p
l
l
l
l
p
p
l
l
l
p
p
l
l
p
p
l
p
p
p
l
Mean score
SE
WA
528
4.0
ACT
526
3.7
l
NSW
525
3.5
l
l
VIC
523
4.1
l
l
l
QLD
522
3.4
l
l
l
l
SA
520
4.1
l
l
l
l
l
NT
513
7.9
l
l
l
l
l
l
TAS
490
4.0
q
q
q
q
q
q
p
OECD average
500
0.7
q
q
q
q
q
q
l
Note: Read across the row to compare a jurisdiction’s performance with the performance of each jurisdiction listed in the column heading.
32
p
Average performance significantly higher than in comparison jurisdiction
l
No statistically significant difference from comparison jurisdiction
q
Average performance significantly lower than in comparison jurisdiction
Thinking it through: Australian students’ skills in problem solving
q
p
Figure 3.6 shows the average proportion of students at each problem-solving proficiency level by
jurisdiction. Five per cent of students from three jurisdictions, the Australian Capital Territory,
New South Wales and the Northern Territory, achieved Level 6. This was half the proportion of
students in Singapore at this level. In other jurisdictions, 3 or 4% of students achieved Level 6,
which was higher than the 2% of students across the OECD.
The proportion of top performers across the jurisdictions ranged from 11% in Tasmania to 19% in the
Australian Capital Territory. All jurisdictions, except Tasmania, achieved higher proportions of top
performers than the OECD average (11%).
Just over one-quarter (27%) of students in Tasmania and one-fifth (21%) of students in the Northern
Territory were low performers compared to 16% of students in the Australian Capital Territory and
Queensland, and 15% of students in New South Wales, Victoria and South Australia. The smallest
proportion of low performers was in Western Australia with 13%.
WA
4
9
19
26
NSW
5
10
19
26
VIC
5
10
20
26
SA
4
11
21
ACT
6
10
18
QLD
5
11
20
12
18
NT
9
TAS
10
17
Australia
5
Singapore
8
80
60
40
20
22
22
0
20
4
5
4
10
9 2
20
40
5
20
27
26
3
12
23
22
12
8 3
16
26
4
12
22
23
12
12
23
22
4
5
14
24
26
14
13
23
24
19
2 6
OECD average
100
11
13
22
27
23
13
25
60
80
100
Percentage of students
Below Level 1
Level 1
Level 2
Level 3
Level 4
Level 5
Level 6
Figure 3.6 Percentage of students across the problem-solving proficiency scale, by jurisdiction
Problem-solving performance by sex across the Australian jurisdictions
Overall, no significant differences between the sexes were found in Australia for problem solving.
Figure 3.7 (p. 34) shows that only one jurisdiction, Western Australia, was found to have significant
differences between the sexes, with males performing significantly higher than females. In Western
Australia, males achieved a mean score of 537 points, which was 18 score points on average higher than
females. This difference was three times larger than the OECD average (6 score points).
Figure 3.8 (p. 34) shows the proportion of males and females in each problem-solving proficiency level by
jurisdiction and across the OECD. Generally, there were higher proportions of males than females
who were top performers (achieving Level 5 or 6) and low performers (failed to reach Level 2).
The proportion of top-performing males ranged from 12% in Tasmania to 22% in the Northern
Territory, and the proportion of top-performing females ranged from 11% in Tasmania to 17% in the
Australian Capital Territory and New South Wales. All jurisdictions, except Tasmania, achieved a higher
proportion of top-performing males compared to the OECD average (13%), while all jurisdictions had a
higher proportion of top-performing females compared to the OECD average (10%). The difference
Australian students’ performance in problem solving
33
Females
Jurisdiction
Males
Mean score
SE
Mean score
SE
ACT
529
4.9
522
5.9
Difference in mean score
TAS
491
5.5
489
5.4
SA
521
4.8
519
4.7
NSW
525
4.1
525
5.1
QLD
521
4.2
523
4.1
VIC
522
4.5
524
4.9
NT
507
11.1
519
10.1
WA
519
5.6
537
5.5
Females
score
higher
Males
score
higher
30
40
20
0
10
Sex difference significant
10
20
30
40
Sex difference not significant
Figure 3.7 Mean scores and differences in students’ performance on the problem-solving scale, by jurisdiction and sex
9
18
24
10
17
24
10
19
11
19
5
Females
8
Males
4
Females
6
Males
5
14
22
5
10
20
26
Males
5
11
19
26
Females
5
10
20
Males
5
11
19
10
21
6
13
21
Females
4
13
23
27
24
4
13
26
4
12
23
QLD
VIC
NSW
ACT
between top-performing males and females was larger in two jurisdictions (Western Australia and the
Northern Territory), where there were between 5 and 10% more males than females reaching these high
proficiency levels. In other jurisdictions, the difference between top-performing males and females was
smaller, with 1 or 2%.
4
3
12
22
29
WA
SA
Females
5
12
23
25
4
12
22
27
4
13
22
Males
5
11
21
26
22
12
4
Females
5
10
20
27
23
11
4
8
12
Males
10
13
10 2
22
26
20
4
TAS
NT
Females
8
17
21
21
17
12
Males
8 3
15
25
25
16
8
Females
5
15
26
25
17
9
4
Males
10
20
Males
84
5
11
19
Australia
5
OECD
average
85
Females
Females
8
14
23
Males
9
13
21
100
80
60
40
20
23
27
27
20
19
40
4
12
5
8 2
10 3
20
25
8
13
22
25
0
14
21
17
16
60
80
Percentage of students
Below Level 1
Level 1
Level 2
Level 3
Level 4
Level 5
Level 6
Figure 3.8 Percentage of students across the problem-solving proficiency scale, by jurisdiction and sex
34
Thinking it through: Australian students’ skills in problem solving
100
The proportion of low-performing males ranged from 13% in Western Australia to 29% in Tasmania,
and the proportion of low-performing females ranged from 14% in the Australian Capital Territory,
New South Wales and South Australia to 24% in Tasmania. The proportion of low-performing males in
Tasmania and the Northern Territory was higher than the OECD average (22%), while the proportion
of low-performing females was higher in Tasmania than across the OECD (22%). The difference
between low-performing males and females was 5% in Tasmania and 4% in the Australian Capital
Territory, and smaller (3% or less) in other jurisdictions.
Problem-solving performance by geographic location of school
Australian schools in PISA were assigned to one of three geographic locations (metropolitan, provincial
and remote)1, so that students’ performance in each of these groups could be investigated. Figure 3.9
shows the mean problem-solving scores for the three geographic location categories, along with the
standard error, confidence intervals and the distribution of scores graphically.
Students in metropolitan schools achieved a mean score of 528 points, which was significantly higher
than for students in provincial schools (510 score points on average) and significantly higher than for
students in remote schools (475 score points on average). Compared to the OECD average, students in
metropolitan and provincial schools performed significantly higher, while students in remote schools
performed significantly lower, with a difference of 25 score points on average.
The mean score difference between students’ performance in metropolitan and remote schools was
53 score points or a difference of almost one proficiency level, which was larger than the mean score
difference between students’ performance in provincial and remote schools (35 score points) or between
students’ performance in metropolitan and provincial schools (18 score points).
The spread of scores between the lowest and highest performing students was widest for students in
remote schools, followed by students in metropolitan schools, with a spread of scores that was the same as
across the OECD. Students in provincial schools had the narrowest range of problem-solving scores.
At the 5th percentile, students in remote schools performed about one proficiency level lower than
students in metropolitan or provincial schools. At the 95th percentile, students in remote schools
performed about one-half of a proficiency level lower than students in provincial schools and about
three-quarters of a proficiency level lower than students in metropolitan schools.
SE
Confidence
interval
Difference between
5th and 95th
percentiles
528
2.4
524–533
320
Provincial
510
3.2
504–517
314
Remote
475
16.4
443–508
344
Geographic
location
Mean score
Metropolitan
Distribution of scores
200
300
400
500
600
700
800
Mean problem-solving performance
Figure 3.9 Mean scores and distribution of students’ performance on the problem-solving scale, by geographic location
Using the MCEECDYA Schools Geographic Location Classification. The Reader’s Guide provides more information about the MCEECDYA Schools
Geographic Location Classification.
1
Australian students’ performance in problem solving
35
Figure 3.10 shows that higher proportions of students in metropolitan schools (18%) and provincial
schools (12%) were top performers compared to students in remote schools (9%). At the lower end of
the proficiency scale, there were higher proportions of students in remote schools (30%) who were low
performers compared to students in metropolitan schools (15%) and provincial schools (18%). For those
students who were placed at below Level 1, there were more than twice as many students in remote
schools compared to students in metropolitan and provincial schools.
Metropolitan
5
Provincial
10
6
Remote
13
100
80
60
40
19
12
22
17
20
25
23
28
21
20
26
0
20
13
16
40
9
5
3
8
60
80
100
Percentage of students
Below Level 1
Level 1
Level 2
Level 3
Level 4
Level 5
Level 6
Figure 3.10 Percentage of students across the problem-solving proficiency scale, by geographical location
Problem-solving performance by Indigenous background
Details about students’ Indigenous background were derived from information provided by the school.2
Figure 3.11 shows that the performance of Indigenous students in problem solving was significantly lower
compared to non-Indigenous students. Indigenous students achieved a mean score of 454 points, which
was 72 score points less than the mean score achieved by non-Indigenous students with 526 score points.
The score difference between Indigenous and non-Indigenous students is equivalent to more than one
proficiency level. Indigenous students performed significantly lower than students across the OECD by
46 score points on average, while non-Indigenous students performed significantly higher.
Although the spread of scores between the highest and lowest performing students was similar for
Indigenous (314 score points) and non-Indigenous students (317 score points), the lowest performing
Indigenous students scored 294 points on average compared to the lowest performing non-Indigenous
students, with 362 score points on average. Indigenous students’ scores at each of the other percentiles
were lower compared to the corresponding scores for non-Indigenous students.
Indigenous
background
Mean score
SE
Confidence
interval
Difference between 5th
and 95th percentiles
Indigenous
454
4.2
446–462
314
Non-Indigenous
526
1.9
522–529
317
Distribution of scores
200
300
400
500
600
700
800
Mean problem-solving performance
Figure 3.11 Mean scores and distribution of students’ performance on the problem-solving scale, by Indigenous background
Figure 3.12 provides details about Indigenous and non-Indigenous students’ proficiencies in problem
solving. Four per cent of Indigenous students were top performers in problem solving compared to 18%
of non-Indigenous students. Less than 1% (0.6%) of Indigenous students achieved the highest proficiency
level (Level 6) considerably lower than the 5% of non-Indigenous students or the 3% of students across
the OECD who achieved at this level.
The Reader’s Guide provides more information about the definition of Indigenous background.
2
36
Thinking it through: Australian students’ skills in problem solving
Thirty-seven per cent of Indigenous students were low performers compared to 15% of non-Indigenous
students. These low-performing students will be limited in their capacity to solve problems. The
proportion of low-performing Indigenous students was almost twice that for students across the OECD
(21%).
There were almost half as many Indigenous students as non-Indigenous students who achieved Level 4,
while similar proportions of Indigenous and non-Indigenous students achieved Level 2 or 3 (47% and
45% respectively).
Indigenous
16
21
Non-Indigenous
5
100
80
60
40
20
25
10
22
19
12
26
0
20
3
23
40
13
60
5
80
100
Percentage of students
Below Level 1
Level 1
Level 2
Level 3
Level 4
Level 5
Level 6
Figure 3.12 Percentage of students across the problem-solving proficiency scale, by Indigenous background
Problem-solving performance by sex and Indigenous background
As shown in Figure 3.13, there were no significant differences between the performance of Indigenous
females and males in problem solving. This was also the case for non-Indigenous females and males.
Indigenous
background
Females
Males
Mean score
SE
Mean score
SE
Indigenous
456
4.9
452
5.8
Non-Indigenous
524
2.2
527
2.5
Figure 3.13 Mean scores and distribution of
performance on the problem-solving scale, by Indigenous
background and sex
Distribution of scores
Females
score
higher
40
30
Males
score
higher
20
10
Sex difference significant
0
10
20
30
40
students’
Sex difference not significant
There were similar proportions of Indigenous females (3%) and males (4%) who were top performers in
problem solving (Figure 3.14, p. 38). The proportions of top-performing non-Indigenous females and
males were also similar (16% and 18% respectively).
Thirty-five per cent of Indigenous females and 39% of Indigenous males were low performers, while
14% of non-Indigenous females and 15% of non-Indigenous males were low performers.
For Indigenous students, there were similar proportions of top-performing females and males, but there
was a higher proportion of males who were low performers compared to females. For non-Indigenous
students, the proportion of top-performing females and males was similar, while for the low-performing
students there were slightly more males than females.
Australian students’ performance in problem solving
37
26
21
25
Indigenous
20
Non-Indigenous
15
Females
Females
4
10
20
Males
5
10
19
18
Males
100
80
60
Below Level 1
40
Level 1
Level 2
20
Level 3
Level 4
13
23
25
40
4
12
23
27
Level 5
3
12
20
0
20
Percentage of students
3
12
23
60
5
80
100
Level 6
Figure 3.14 Percentage of students across the problem-solving proficiency scale, by Indigenous background and sex
Problem-solving performance by socioeconomic background
Socioeconomic background in PISA is measured by the index of Economic, Social and Cultural Status
(ESCS), which captures the wider aspects of a student’s family and home background.3 Figure 3.15 shows
the positive relationship between socioeconomic background and student performance, with students in
the higher socioeconomic quartiles achieving a score of 560 points on average, which was significantly
higher in problem solving than students in the lower socioeconomic quartiles.
Students in the highest socioeconomic quartile achieved an average of 73 score points higher than
students in the lowest socioeconomic quartile. This difference represents more than one proficiency level
on the problem-solving proficiency scale. The difference between each socioeconomic quartile and the
next was significant at around 25 score points on average or around one-third of a proficiency level.
Students in the two highest socioeconomic quartiles showed a similar spread of scores between the lowest
and highest performing students, which were slightly narrower than students in the two lowest
socioeconomic quartiles.
Socioeconomic
background
Difference between
5th and 95th
percentiles
Mean score
SE
Confidence
interval
Lowest quartile
487
2.5
482–492
313
Second quartile
512
2.4
507–516
307
Third quartile
538
2.9
532–543
301
Highest quartile
560
2.5
555–565
302
Distribution of scores
200
300
400
500
600
700
800
Mean problem-solving performance
Figure 3.15 Mean scores and distribution of students’ performance on the problem-solving scale, by socioeconomic background
The Reader’s Guide provides more information about socioeconomic background and the ESCS index.
3
38
Thinking it through: Australian students’ skills in problem solving
Figure 3.16 shows the proportion of top performers was higher with each increase in the
socioeconomic quartile, while the proportion of low performers was lower with each decrease in the
socioeconomic quartile.
At the higher end of the proficiency scale, 9% of students in the lowest socioeconomic quartile were
top performers compared to 12% in the second socioeconomic quartile, 19% of students in the third
socioeconomic quartile and 27% of students in the highest socioeconomic quartile.
At the lower end of the proficiency scale, one-quarter of the students in the lowest socioeconomic
quartile were low performers compared to 18% of students in the second socioeconomic quartile, 11% of
students in the third socioeconomic quartile and 8% of students in the highest socioeconomic quartile.
Lowest quartile
Second quartile
6
Third quartile
3 8
Highest quartile
2 6
100
80
60
40
20
0
20
40
60
5
8
19
27
24
13
14
26
27
17
3
9
21
28
22
12
7 2
17
25
25
16
9
80
100
Percentage of students
Below Level 1
Level 1
Level 2
Level 3
Level 4
Level 5
Level 6
Figure 3.16 Percentage of students across the problem-solving proficiency scale, by socioeconomic background
Problem-solving performance by immigrant background
Immigrant background was measured on students’ self-report of where they and their parents were born.4
Figure 3.17 shows first-generation students achieved a mean score of 531 points, which was significantly
higher than the mean score for Australian-born students (523 points) and foreign-born students (517 points).
Australian-born students’ performance in problem solving was not significantly different from foreignborn students.
The range of scores was 315 points for Australian-born students, which was similar to the range of scores
for first-generation students (318 points), while the range of scores for foreign-born students was wider at
328 points.
Immigrant
background
Difference between
5th and 95th
percentiles
Mean score
SE
Confidence
interval
Australian-born
523
2.1
519–527
315
First-generation
531
2.9
525–536
318
Foreign-born
517
3.6
510–524
328
Distribution of scores
200
300
400
500
600
700
800
Mean problem-solving performance
Figure 3.17 Mean scores and distribution of students’ performance on the problem-solving scale, by immigrant background
The Reader’s Guide provides more information about the definition of immigrant background.
4
Australian students’ performance in problem solving
39
Figure 3.18 shows the proportions of students at each proficiency level on the problem-solving scale by
immigrant background. For the top performers, the proportion of first-generation students (19%) was
slightly higher than the proportion of Australian-born and foreign-born students (16%), while for low
performers, there was almost one-fifth (18%) of foreign-born students and similar proportions of
Australian-born and first-generation students (15 and 14% respectively).
Australian-born
5
10
19
27
First-generation
4
10
19
25
23
12
20
25
21
6
Foreign-born
100
80
60
40
20
0
23
12
4
14
5
12
20
40
60
Level 4
Level 5
Level 6
4
80
100
Percentage of students
Below Level 1
Level 1
Level 2
Level 3
Figure 3.18 Percentage of students across the problem-solving proficiency scale, by immigrant background
Problem-solving performance by language background
Students who spoke English as their main language at home performed significantly higher in problem
solving (526 score points on average) than those students whose main language at home was a language
other than English (509 score points on average).
Figure 3.19 shows that the spread of scores between the lowest and highest performing students for
English speakers was narrower (316 score points), than the spread of scores for students who speak a
language other than English at home (341 score points).
Difference between
5th and 95th
percentiles
Language background
Mean score
SE
Confidence
interval
English spoken at home
526
1.9
522–529
316
Language other than
English spoken at home
509
4.4
500–518
341
Distribution of scores
200
300
400
500
600
700
800
Mean problem-solving performance
Figure 3.19 Mean scores and distribution of students’ performance on the problem-solving scale, by language background
Figure 3.20 shows 18% of students who spoke English at home were top performers in problem solving,
which was simliar to the 16% of students who spoke a language other than English at home. One-fifth
(21%) of students who spoke a language other than English at home were low performers compared to
15% of students who spoke English at home.
40
Thinking it through: Australian students’ skills in problem solving
English spoken at home
5
Language other than
English spoken at home
100
8
80
60
40
20
10
13
19
26
20
23
0
20
23
13
20
40
12
60
5
4
80
100
Percentage of students
Below Level 1
Level 1
Level 2
Level 3
Level 4
Level 5
Level 6
Figure 3.20 Percentage of students across the problem-solving proficiency scale, by language background
Variations in problem-solving performance between and
within schools
The variation in performance within countries can be divided into a measure of performance differences
between students from the same school and a measure of performance differences between groups of
students from different schools. Figure 3.21 (p. 42) shows the proportion of variance in achievement for
each country, divided into the amount of variation that occurs between schools (i.e., the performance
variation attributable to differences in students’ results in different schools) and the amount of variation
that occurs within schools (the performance variation attributable to the range of students’ results that
cannot be attributed to differences between schools).
Across OECD countries, the amount of variation in performance within schools was 61% and the
amount of variation in performance between schools was 38%. In Australia, the amount of variation
in performance within schools was 75% and was higher than the OECD average, while the amount of
variation in performance between schools in Australia was 28% and lower than the OECD average.
Australian students’ performance in problem solving
41
Variation within schools
(as proportion of average OECD total)
Variation between schools
(as proportion of average OECD total)
Israel
OECD
average
61%
Hungary
Bulgaria
OECD
average
38%
Netherlands
United Arab Emirates
Belgium
Germany
Slovenia
Slovak Republic
Czech Republic
Austria
Uruguay
Brazil
OECD average
Italy
Cyprus
Croatia
Poland
Shanghai–China
Turkey
Chinese Taipei
Montenegro
Chile
Spain
Colombia
Singapore
Hong Kong–China
Serbia
Russian Federation
England
Malaysia
Korea
Australia
United States
Japan
Denmark
Portugal
Canada
Norway
Ireland
Macao–China
Estonia
Sweden
Finland
100
80
60
40
20
0
20
40
60
80
100
Percentage of variation within and between schools
Figure 3.21 Variation in problem-solving performance between and within schools, by country 5,6,7
Expressed as a percentage of the average variation in student performance across OECD countries.
Data for France is not available.
7 The variation within and between schools will not add up to 100 because it is a percentage of the average OECD total variance.
5
6
42
Thinking it through: Australian students’ skills in problem solving
In Finland, there was little variation in performance between schools (10%). This was followed by
Sweden, Estonia and Macao–China, where around 20% of the variation was due to differences in
performance between schools. In these countries, students can expect to achieve similar results regardless
of the school they attend. In countries that report larger amounts of variation between schools, for
example in Hungary and Israel, the amount of variance in performance between schools was 70% and
84% respectively, meaning it would make a difference which schools students attend.
Figure 3.22 shows the proportion of between—and within—school variation in problem-solving
performance for jurisdictions. On average, the variation in problem-solving performance that was
observed between schools ranged from 19% in South Australia to 39% in Tasmania. This suggests that
the school students attend in Tasmania will influence problem-solving performance more than the school
students attend in other jurisdictions, where there were smaller proportions of between-school variations
in problem solving.
On average across jurisdictions, the variation in student performance that was observed within schools
ranged from 72% in Victoria to 94% in the Northern Territory. This means that for students in a
Northern Territory school, their score in problem solving would be more heterogeneous than
for students in a jurisdiction where the within-school variation in problem solving was smaller.
Variation within schools
(as proportion of average OECD total)
Variation between schools
(as proportion of average OECD total)
TAS
OECD
average
61%
NT
ACT
OECD
average
38%
NSW
QLD
VIC
WA
SA
Australia
OECD average
100
80
60
40
20
0
20
40
60
80
100
Percentage of variation within and between schools
Figure 3.22 Variation in problem-solving performance between and within schools, by jurisdiction
Australian students’ performance in problem solving
43
Comparing students’ performance in problem solving with
mathematics, science and reading
PISA assesses problem solving in two contexts. There are the regular assessments of mathematics,
science and reading that include problem-solving tasks that assess students’ abilities to apply the skills
and knowledge they have learned in school, and there is the assessment that assesses students’ general
reasoning skills, focusing on the cognitive process that is essential for successful problem solving.
It is expected that performance in problem solving is positively correlated with students’ performance
in mathematics, science and reading. That is, students who do well in problem solving are likely to do
well in other literacy domains and students who perform poorly in problem solving are likely to perform
poorly in other literacy domains. The strength of the relationship between performance in the regular
PISA assessments of mathematics, science and reading, and performance in problem solving is shown in
Table 3.2. The data in the table are latent correlations for the OECD average, where the closer to 0.00
implies no relationship and the closer to 1.00 implies the strongest positive relationship. Comparing the
strength of the relationship between problem solving and the three literacy domains, it can be seen that
the strongest correlation is between problem solving and mathematics (0.81) and the weakest correlation
is between problem solving and reading (0.75).
Table 3.2 Relationship between performance in problem solving, mathematics, science and reading across the OECD
Science
Mathematics
Reading
Problem solving
0.78
0.81
0.75
Reading
0.88
0.85
Mathematics
0.90
Table 3.3 shows the corresponding correlations for Australia. It can be seen that the largest correlation is
again between problem solving and mathematics and the smallest correlation is between problem
solving and reading. The strengths of the relationships for Australia are slightly stronger compared
to the OECD average.
Table 3.3 Relationship between performance in problem solving, mathematics, science and reading for Australia
Science
Mathematics
Reading
Problem solving
0.81
0.83
0.77
Reading
0.90
0.87
Mathematics
0.91
The correlations between problem solving and mathematics were 0.81 for Western Australia, 0.82
for the Australian Capital Territory, Queensland and South Australia, 0.83 for Tasmania and the
Northern Territory, 0.84 for Victoria and 0.86 for New South Wales.
An analysis that relates the variation in problem-solving performance jointly to the variation in the
performance in mathematics, science and reading shows that skills assessed in the problem-solving
assessment were also used in a wide range of contexts. Across the OECD, 68% of the problem-solving
variance reflected skills that were also measured in one of the three literacy domains regularly assessed
in PISA. The remaining 32% reflected skills that were uniquely measured in the problem-solving
assessment. Of the 68% of variation that problem-solving performance shared with other literacy
domains, the largest proportion was shared with all three regular literacy assessment domains (62% of
the total variation), about 5% was uniquely shared between problem solving and mathematics, and
about 1% was based on skills that were specifically measured in the science and reading assessments.
44
Thinking it through: Australian students’ skills in problem solving
Figure 3.23 shows the association of problem-solving skills with performance in mathematics, science
and reading was, in general, of similar strength across countries. In Colombia, the Russian Federation,
Spain, Japan, Italy and Hong Kong–China, less than 60% of the problem-solving variance reflects skills
Colombia
Russian Federation
Spain
Japan
Hong Kong–China
Denmark
Canada
Poland
Norway
Italy
Macao–China
Uruguay
Cyprus
Portugal
Ireland
Austria
Montenegro
Chile
Sweden
Korea
United Arab Emirates
Belgium
Bulgaria
OECD average
Slovenia
Brazil
Singapore
Serbia
France
Malaysia
Hungary
Turkey
Shanghai–China
Australia
Germany
Finland
Estonia
Croatia
Variation associated with more than one domain
Slovak Republic
Variation uniquely associated with mathematics
performance
England
Variation uniquely associated with reading
performance
United States
Netherlands
Variation uniquely associated with science
performance
Israel
Residual (unexplained) variation
Chinese Taipei
Czech Republic
0
20
40
60
Percentage of variance explained
80
100
120
Figure 3.23 Variation in problem-solving performance associated with performance in mathematics, science and reading, by country
Australian students’ performance in problem solving
45
that were also measured in the other literacy domains. This indicates comparatively weak associations
between the skills measured in the problem-solving assessment and performance in mathematics, science
and reading. On the other hand, the Czech Republic, Chinese Taipei and Israel had comparatively
stronger associations between the problem-solving skills measured in the assessment and performance
in mathematics, science and reading, with over 75% of the total explained variation. In Australia, 71%
of the problem-solving variance reflected skills that were also measured in the regular PISA assessments.
Approximately 30% of the variation in Australian problem-solving performance was uniquely measured
in the problem-solving assessment. These findings suggest that problem-solving ability, as defined in
PISA, is highly correlated with the core PISA literacy domains.
Figure 3.24 shows the variation in problem-solving performance that is associated with one or more of
the three literacy domains and the variation in problem-solving performance that is associated with only
problem-solving skills. The proportion of the problem-solving variance that reflected skills that were also
measured in one of the three regular assessment literacy domains ranged from 68% in Western Australia
to 75% in New South Wales, while the proportion of the problem-solving variance that reflected skills
that were uniquely captured by the problem-solving assessment ranged from 25% in New South Wales to
32% in Western Australia.
Western Australia had comparatively weak associations between the skills assessment in the problemsolving assessment and performance in mathematics, science and reading. New South Wales had
comparatively stronger associations between the skills assessment in the problem-solving assessment and
performance in mathematics, science and reading.
WA
QLD
SA
ACT
Variation associated with more than one domain
TAS
Variation uniquely associated with mathematics
performance
VIC
Variation uniquely associated with reading
performance
NT
Variation uniquely associated with science
performance
NSW
Australia
Residual (unexplained) variation
OECD average
0
20
40
60
80
100
Percentage of variance explained
Figure 3.24 Variation in problem-solving performance associated with performance in mathematics, science and reading, by jurisdiction
Relative performance in problem solving in Australia
It is possible to infer whether students perform the same as, above or below students with similar
proficiency in mathematics by comparing the performance of students from one country to the average
performance observed across participating countries at a given level of proficiency in mathematics.
In Australia, England and the United States, the best students in mathematics also had excellent
problem-solving skills. These countries’ good performance in problem solving was mainly due to strong
performers in mathematics. This may suggest that in these countries, top performers in mathematics have
access to—and take advantage of—the kinds of learning opportunities that are also useful for improving
their problem-solving skills (Figure 3.25).
46
Thinking it through: Australian students’ skills in problem solving
700
Problem-solving score
600
500
400
300
300
400
500
600
700
Mathematics score
Notes: The dotted line shows the average performance in problem solving, across students from all participating countries, at different levels of performance in mathematics.
The continuous line shows the pattern of relative performance in problem solving.
Figure 3.25 R elative performance in problem solving at different levels on the mathematics scale for Australia,
England and the United States
In Japan, Korea and Italy, the good performance in problem solving was, to a large extent, due to the fact
that lower performing students score beyond expectations in the problem-solving assessment; however, in
Singapore, for example, students’ average performance in problem solving across all levels were similar to
the average performance in mathematics.
There were similar differences among countries with overall weak performance in problem solving,
relative to their students’ performance in mathematics. In several of the countries (Bulgaria, Colombia,
Croatia, Denmark, Estonia, Germany, Hungary Ireland, Israel, the Netherlands, Slovenia, Spain and the
United Arab Emirates), specific difficulties in problem solving were most apparent among students with
poor mathematics skills. Students with strong mathematics skills often perform on or close to par with
students in other countries.
In Austria, Belgium, Malaysia, Montenegro, Poland, Shanghai–China, Singapore, the Slovak Republic
and Uruguay, weak performance in problem solving—relative to mathematics performance—was mainly
due to students’ performance being similar across all proficiency levels.
Australian students’ performance in problem solving
47
CHAPTER 4
Students’ strengths and weaknesses
in problem solving
Chapter 3 described student performance on the PISA 2012 problem-solving scale. This chapter examines
problem-solving performance by taking a closer look at the problem-solving items, analysing how
students interact with the test items to identify comparative strengths and weaknesses within countries
and within different social groups. In Chapter 3, results were reported on the overall problem-solving
scale. This chapter focuses on the problem-solving aspects, analysing how students perform on the
problem-solving processes, the nature of the problem situation and the response formats, and identifies the
skills that some students master better than other students. The final part of this chapter presents details of
two of the problem-solving framework aspects—the nature of the problem situation and problem-solving
processes—and groups the countries, the Australian jurisdictions and different social groups within
Australia by their strengths and weaknesses in problem solving.
48
Reporting students’ strengths and weaknesses in
problem solving
PISA reports the performance of all students on the problem-solving assessment
on an overall scale. While this approach has many advantages, it can potentially hide
interesting differences in patterns of performance at lower levels of aggregation, that
is, on single items or on subsets of items. To look at patterns of performance at a
more detailed level than that given by the overall problem-solving scale, the approach
used was to look at the unscaled item of students who were administered each item1.
Internationally, average percentages of correct responses are computed at the country
level, where a correct response is taken as a full-credit answer, and nonreached2
items are taken as incorrect. For the average of a group of items, the simple average
percentage of the relevant items is used. The Australian results considered in this
chapter are calculated over the relevant groups of students, such as jurisdictions or
Indigenous background.
Across countries, a measure of the difficulty of the items is their average percentage.
The average across the OECD was used as a basis for international comparison.
The relative difficulty of two distinct sets of items can be given by comparing their
average percentages. So, by comparing the percentage correct across two sets of
items and across countries, the relative strengths and weaknesses of each country
can be identified. For each subset of items and for each country, the result of this
comparison is reported as an odds ratio where a ratio of 1 indicates that the pattern of
performance across items is in line with the average OECD pattern of performance.
A ratio of more than 1 indicates that the items in this subset were relatively easier
for students of a particular country than for students across OECD countries, after
accounting for the overall differences in performance. Conversely, a ratio of less than 1
meant that students in a particular country found these items relatively harder.
For the within Australia comparisons in this report, it was decided to use the overall
results for Australia as a basis of comparison rather than the results across the OECD.
As such, an odds ratio of more than 1 would indicate, for example, that Indigenous
students found a set of items relatively easier than the students in Australia as a
whole, after accounting for their overall performance.
In the international results, significant differences from the OECD averages at the 5%
level are shown. For the Australian results, differences from the Australian averages
are shown at a moderate significance (the 10% level), as well as at the 5% level.
The purpose of this chapter is to provide a profile of students’ strengths and
weaknesses in problem solving. However, it is important to place a caveat around the
results. Performance has been compared to the OECD average, at the international
level, and to the Australian average, at the national level, to identify comparative
strengths and weaknesses. These averages were selected for pragmatic reasons;
however, it could be argued that the balance between the various aspects of problemsolving competence could be challenged to be different than what has been used in
these analyses. Also, the results were based on a small number of items. There were
42 items in the problem-solving assessment and, in some cases, analyses were based
on small sets of items.
Students only completed a subset of items from the whole problem-solving item pool.
These are items that students have not attempted.
1
2
Students’ strengths and weaknesses in problem solving
49
Students’ strengths and weaknesses in problem-solving processes
The main cognitive processes involved in solving a problem were outlined in Chapter 2. These were:
exploring and understanding; representing and formulating; planning and executing; and monitoring
and reflecting. Each of the items in the PISA 2012 problem-solving assessment was classified to reflect
the main demand of each item; however, other processes may have been involved while solving a
particular item.
Strengths and weaknesses on problem-solving tasks by problem-solving processes,
across countries3
Figure 4.1 summarises countries’ strengths and weaknesses in problem-solving processes and shows
there is a pattern of high-performing countries performing relatively stronger on the exploring and
understanding and the representing and formulating processes, and relatively weaker on the planning and
executing and the monitoring and reflecting processes. The lower performing countries generally show
the opposite of this pattern. Of course, there is some variation in this, but the effect seems quite clear.
The middle-ranked countries show no definite pattern, though there are some countries that show a
significant difference from the OECD average in only one or none of the problem-solving processes.
Australian students are comparatively stronger on the exploring and understanding and the representing
and formulating processes, and are relatively weaker on the planning and executing process.
In terms of problem-solving skills, Australian students are good at generating new knowledge and can
be characterised as quick learners—questioning their knowledge and challenging assumptions, and
generating and experimenting with alternatives—and good at abstract-information processing.
Students who are typically good at planning and executing processes use the knowledge they have, and
are characterised as goal-driven and persistent. This is an area where Australian students’ skills could be
improved; they need to be able to use their knowledge to devise a plan and execute the plan in order to
solve a problem.
These comparisons take into account the countries’ overall performance.
3 50
Thinking it through: Australian students’ skills in problem solving
Difference between observed and expected performance,
by problem-solving process
Country
Mean score in
problem solving
Singapore
562
Korea
561
Japan
552
Macao–China
540
Hong Kong–China
540
Shanghai–China
536
Chinese Taipei
534
Canada
526
Australia
523
Finland
523
England
517
Estonia
515
France
511
Netherlands
511
Italy
510
Czech Republic
509
Germany
509
United States
508
Belgium
508
Austria
506
Norway
503
Ireland
498
Denmark
497
Portugal
494
Sweden
491
Russian Federation
489
Slovak Republic
483
Poland
481
Spain
477
Slovenia
476
Serbia
473
Croatia
466
Hungary
459
Turkey
454
Israel
454
Chile
448
Cyprus
445
Brazil
428
Malaysia
422
United Arab Emirates
411
Montenegro
407
Uruguay
403
Bulgaria
402
Colombia
399
Exploring and
understanding
Representing
and formulating
Planning and
executing
Monitoring and
reflecting
Stronger than expected performance on the problem-solving process
Non significant strength or weakness
Weaker than expected performance on the problem-solving process
Figure 4.1 Relative strengths and weaknesses in problem-solving processes, by countries
Students’ strengths and weaknesses in problem solving
51
Strengths and weaknesses on problem-solving tasks by problem-solving processes,
within Australia4
The highest performing jurisdiction, Western Australia, performed relatively better on the exploring
and understanding process than Australia as whole. The opposite was true for New South Wales, with
students performing relatively weaker on this process compared to Australia as a whole. Queensland
performed stronger than expected on the representing and formulating process and moderately weaker on
the planning and executing process. New South Wales performed relatively stronger on the representing
and formulating process. Figure 4.2 shows there were no clear patterns in different performance between
the jurisdictions.
Difference between observed and expected performance,
by problem-solving process
Jurisdiction
Mean score in
problem solving
WA
528
ACT
526
NSW
525
VIC
523
QLD
522
SA
520
NT
513
TAS
490
Exploring and
understanding
Representing
and formulating
Planning and
executing
Monitoring and
reflecting
Stronger than expected on the problem-solving process
Moderately stronger than expected performance on the problem-solving process
Nonsignificant strength or weakness
Moderately weaker than expected performance on the problem-solving process
Weaker than expected performance on the problem-solving process
Figure 4.2 Relative strengths and weaknesses in problem-solving processes, by jurisdictions
Australia’s results across the problem-solving processes were disaggregated into national subgroups,
as presented in Figure 4.3. Females showed a strong positive bias on the monitoring and reflecting
process and a moderate positive bias on the planning and executing process, but were weaker on the
representing and formulating process. Indigenous students performed relatively weaker on the exploring
and understanding process, while non-Indigenous students performed relatively stronger on the same
process. Internationally, this was also what was observed, with lower performing countries also being
relatively weaker in this process. There were no clear patterns in performance across the problem-solving
processes for geographic location, despite large differences in overall performance. Interestingly, students
in the lowest quartile of socioeconomic background performed relatively stronger on the planning and
executing process, and performed relatively weaker on the exploring and understanding process. The
opposite findings were found for students in the highest quartile of socioeconomic background.
These comparisons take into account Australia’s overall performance.
4
52
Thinking it through: Australian students’ skills in problem solving
Difference between observed and expected performance,
by problem-solving process
Mean score in
problem solving
Exploring and
understanding
Representing
and formulating
Planning and
executing
Monitoring and
reflecting
Sex
Female
522
Male
524
Indigenous background
Indigenous
454
Non-Indigenous
526
Geographic location
Metropolitan
528
Provincial
510
Remote
475
Socioeconomic background
Lowest quartile
487
Second quartile
512
Third quartile
538
Highest quartile
560
Immigrant status
Australian-born
523
First-generation
531
Foreign-born
517
Language at home
English
526
Language other than English
509
Stronger than expected on the problem-solving process
Moderately stronger than expected performance on the problem-solving process
Nonsignificant strength or weakness
Moderately weaker than expected performance on the problem-solving process
Weaker than expected performance on the problem-solving process
Figure 4.3 Relative strengths and weaknesses in problem-solving processes, by different social groups for Australia
Students’ strengths and weaknesses in the nature of the
problem situation
In the problem-solving framework, items have been classified by how information about a problem has
been presented. Information about the problem that is disclosed at the outset is considered static, while
information that is discovered as students explore the problem is considered interactive.
Strengths and weaknesses on problem-solving tasks by the nature of the problem
situation, across countries
It is difficult to discern any pattern of relative strength or weakness in these results by overall performance
by nature of the problem situation, unlike those that appeared for problem-solving processes. About half
the countries show no significant bias and where an effect exists, the number of countries showing either
direction is about the same, with a slightly higher number of lower performing countries showing a
preference for static tasks (Figure 4.4, p. 54).
Students’ strengths and weaknesses in problem solving
53
Country
Mean score in
problem solving
Singapore
562
Korea
561
Japan
552
Macao–China
540
Hong Kong–China
540
Shanghai–China
536
Chinese Taipei
534
Canada
526
Australia
523
Finland
523
England
517
Estonia
515
France
511
Netherlands
511
Italy
510
Czech Republic
509
Germany
509
United States
508
Belgium
508
Austria
506
Norway
503
Ireland
498
Denmark
497
Portugal
494
Sweden
491
Russian Federation
489
Slovak Republic
483
Poland
481
Spain
477
Slovenia
476
Serbia
473
Croatia
466
Hungary
459
Turkey
454
Israel
454
Chile
448
Cyprus
445
Brazil
428
Malaysia
422
United Arab Emirates
411
Montenegro
407
Uruguay
403
Bulgaria
402
Colombia
399
Difference between observed and expected
performance, by nature of the problem
Static tasks
Interactive tasks
Stronger than expected performance on the nature of the problem
Nonsignificant strength or weakness
Weaker than expected performance on the nature of the problem
Figure 4.4 Relative strengths and weaknesses on problem-solving tasks by the nature of the problem situation, across countries
54
Thinking it through: Australian students’ skills in problem solving
Strengths and weaknesses on problem-solving tasks by the nature of the problem
situation, within Australia
As seen in the international results, no clear pattern emerged for the Australian jurisdictions. Although, in
the lowest performing jurisdiction—Tasmania—performance was stronger on static tasks, which occurred
somewhat more frequently in lower performing countries. Queensland was the only jurisdiction to
perform relatively stronger on interactive tasks (Figure 4.5).
Jurisdiction
Mean score
in problem solving
ACT
526
NSW
525
VIC
523
QLD
522
SA
520
WA
528
TAS
490
NT
513
Difference between observed
and expected performance,
by nature of the problem
Static tasks
Interactive tasks
Stronger than expected on the problem-solving process
Moderately stronger than expected performance on the problem-solving process
Nonsignificant strength or weakness
Moderately weaker than expected performance on the problem-solving process
Weaker than expected performance on the problem-solving process
Figure 4.5 Relative strengths and weaknesses on problem-solving tasks by the nature of the problem situation, across jurisdictions
No relative strengths or weaknesses in static or interactive tasks were found across the different Australian
social groups (therefore, no figure has been presented).
Students’ strengths and weaknesses on the response formats
In the PISA problem-solving assessment, one-third of the items were simple multiple-choice items. Twothirds of the items were constructed-response format items, requiring students to write an answer, draw
lines between two points or drag a shape.
Strengths and weaknesses on problem-solving tasks by response format,
across countries
Figure 4.6 (p. 56) shows the relative strengths and weaknesses by response formats across countries.
Unlike the comparison between the static and interactive items, there were some clear patterns evident
with the different types of response formats. Of the seven highest performing countries, except
Singapore, these Asian countries’ performances were stronger on the selected-response format items.
Twelve of the 13 lowest performing countries in problem solving also showed stronger performance on
the selected-response format items. Canada, Australia, England, Estonia, Belgium, Ireland and Denmark
showed a bias towards the constructed-response format items.
Students’ strengths and weaknesses in problem solving
55
Country
Mean score in
problem solving
Singapore
562
Korea
561
Japan
552
Macao–China
540
Hong Kong–China
540
Shanghai–China
536
Chinese Taipei
534
Canada
526
Australia
523
Finland
523
England
517
Estonia
515
France
511
Netherlands
511
Italy
510
Czech Republic
509
Germany
509
United States
508
Belgium
508
Austria
506
Norway
503
Ireland
498
Denmark
497
Portugal
494
Sweden
491
Russian Federation
489
Slovak Republic
483
Poland
481
Spain
477
Slovenia
476
Serbia
473
Croatia
466
Hungary
459
Turkey
454
Israel
454
Chile
448
Cyprus
445
Brazil
428
Malaysia
422
United Arab Emirates
411
Montenegro
407
Uruguay
403
Bulgaria
402
Colombia
399
Difference between observed and expected
performance, by response format
Selected responses
Constructed responses
Stronger than expected performance on the problem-solving process
Nonsignificant strength or weakness
Weaker than expected performance on the problem-solving process
Figure 4.6 Relative strengths and weaknesses on problem-solving tasks by response format, across countries
56
Thinking it through: Australian students’ skills in problem solving
Strengths and weaknesses on problem-solving tasks by response format, within Australia
The lowest performing state, Tasmania, was the only jurisdiction to show a significant bias towards the
constructed-response format items, while Queensland showed a significant bias towards the selectedresponse format items (Figure 4.7).
Difference between observed and expected
performance, by response format
Jurisdiction
Mean score in
problem solving
ACT
526
NSW
525
VIC
523
QLD
522
SA
520
WA
528
TAS
490
NT
513
Selected responses
Constructed responses
Stronger than expected on the problem-solving process
Moderately stronger than expected performance on the problem-solving process
Nonsignificant strength or weakness
Moderately weaker than expected performance on the problem-solving process
Weaker than expected performance on the problem-solving process
Figure 4.7 Relative strengths and weaknesses on problem-solving tasks by response format, across jurisdictions
Students’ strengths and weaknesses in problem solving
57
Figure 4.8 shows that females do relatively better than males on the constructed-response format items.
Australian-born students were relatively better on the constructed-response format items, where the
effect was consistent, but weaker with students who spoke English at home. Foreign-born students were
relatively stronger on the selected-response format items. This was also the case, to a lesser extent, for
students who spoke a language other than English at home. Since these items tend to include a higher
proportion of reading matter, these results could be expected.
Mean score in
problem solving
Difference between observed and expected
performance, by response format
Selected responses
Constructed responses
Sex
Female
522
Male
524
Indigenous
454
Non-Indigenous
526
Indigenous background
Geographic location
Metropolitan
528
Provincial
510
Remote
475
Socioeconomic background
Lowest quartile
487
Second quartile
512
Third quartile
538
Highest quartile
560
Immigrant status
Australian-born
523
First-generation
531
Foreign-born
517
English
526
Language other than English
509
Language at home
Stronger than expected on the problem-solving process
Moderately stronger than expected performance on the problem-solving process
Nonsignificant strength or weakness
Moderately weaker than expected performance on the problem-solving process
Weaker than expected performance on the problem-solving process
Figure 4.8 Relative strengths and weaknesses on problem-solving tasks by response format, across different social groups
58
Thinking it through: Australian students’ skills in problem solving
Grouping countries by their strengths and weaknesses in
problem solving
In this chapter, differences in performance patterns across the problem-solving processes, the nature of
the problem situation and response-format types have been identified. Figure 4.9 shows the performance
difference according to the nature of the problem situation (static and interactive tasks) and the main
problem-solving process—in this figure, defined as knowledge-acquisition tasks and knowledgeutilisation tasks.5
The Chinese speaking countries (all high-performing countries) differ from the other Asian high
performers (Singapore, Korea and Japan) in being weaker on interactive items, although not all the
differences from the OECD are significant. Korea, Singapore, Hong Kong–China, Macao–China,
Chinese Taipei and Shanghai–China were more successful on knowledge-acquisition tasks (the exploring
and understanding process and the representing and formulating process).
Among the lower performing countries, only Brazil is significantly better on interactive items, although
like all the lower performing countries, it is weaker on knowledge-acquisition tasks. With the exception
of the Czech Republic, all Eastern European countries—mainly lower performing countries—lie in the
quadrant with weaker-than-expected performance in the interactive and knowledge-acquisition tasks.
Most Western European countries were reasonably close to the knowledge-acquisition axis, but some
clearly favour static- or interactive-task items.
While the average problem-solving performance was different between Australia, Japan and Italy (where
Japan performed significantly higher than Australia, which in turn performed significantly higher than
Italy), these countries showed a similar balance of skills when compared to each other. These countries
all performed close to their expected level on interactive items (based on the OECD average pattern of
performance) and slightly above their expected level on knowledge-acquisition tasks.
Stronger-than-expected performance on interactive items
and on knowledge-acquisition tasks
OECD average
Better performance on interactive tasks, relative to static tasks
Stronger-than-expected performance on interactive items,
weaker-than-expected performance on knowledge-acquisition tasks
Ireland
Brazil
Germany
England
Portugal
United Arab Emirates
Spain
Colombia
Czech Republic
Chile
Estonia
Russian Federation
Malaysia
Turkey
Uruguay
Poland
Serbia
Croatia
Hungary
Netherlands
Slovenia
Finland
Slovak Republic
Denmark
Montenegro
United States
France
Korea
Canada
Belgium
Austria
Israel
Norway
Italy
Japan
Australia
Singapore
OECD average
Hong Kong–China
Macao–China
Sweden
Chinese Taipei
Shanghai–China
Bulgaria
Weaker-than-expected performance on interactive items
and on knowledge-acquisition tasks
Weaker-than-expected performance on interactive items,
stronger-than-expected perfomance on knowledge-acquistion tasks
Better performance on knowledge-acquisition tasks, relative to knowledge-utilisation tasks
Figure 4.9 Joint analysis of strengths and weaknesses, by nature of the problem and by process, for countries
In the PISA problem-solving assessment, the problem-solving processes of exploring and understanding and representing and formulating can be
classified as knowledge-acquisition tasks, while the planning and executing process can be classified as knowledge-utilisation tasks.
5
Students’ strengths and weaknesses in problem solving
59
Figure 4.10 shows that, in general, lower performing groups—such as Indigenous students and students
in the lowest socioeconomic quartile—appear in the lower-left quadrant, indicating that these students’
performance was weaker than expected on interactive tasks and on knowledge-acquisition tasks.
The performance across the jurisdictions was widely scattered across the figure. Tasmania was quite a
distance from the other jurisdictions, in the lower-left quadrant, and was a relatively lower performing
jurisdiction overall.
Females perform relatively poorly on knowledge-acquisition tasks and appear in the upper-left quadrant,
while males appear in the bottom-right quadrant, indicating stronger-than-expected performance on
knowledge-acquisition tasks.
Stronger-than-expected performance on interactive items
and on knowledge-acquisition tasks
QLD
AUS average
Better performance on interactive tasks, relative to static tasks
Stronger-than-expected performance on interactive items,
weaker-than-expected performance on knowledge-acquisition tasks
ACT
NT
Females
Lowest quartile SES
Third quartile
SES
English spoken
at home
Highest quartile SES
Schools in metropolitan areas
non-Indigenous
Australian-born
Foreign-born
NSW
Language other than
VIC
English spoken at home
SA
Males
Schools in remote areas
AUS average
Indigenous
First-generation
Second quartile SES
Schools in
provincial areas
WA
TAS
Weaker-than-expected performance on interactive items
and on knowledge-acquisition tasks
Weaker-than-expected performance on interactive items,
stronger-than-expected perfomance on knowledge-acquistion tasks
Better performance on knowledge-acquisition tasks, relative to knowledge-utilisation tasks
Figure 4.10 J oint analysis of strengths and weaknesses, by nature of the problem and by process, for Australian jurisdictions
and social groups
60
Thinking it through: Australian students’ skills in problem solving
CHAPTER 5
Australian students’ motivation
towards problem solving
In PISA’s definition of problem solving, there is recognition that the use of skills and knowledge to solve
a problem depends on motivational and affective factors. It includes: “the willingness to engage with such
situations” (OECD, 2014, p. 30). Engaging with a problem is an integral part of problem solving. In PISA
2012, students completed a questionnaire that collected information about their engagement with and at
school, their drive and the beliefs they hold about themselves as learners.
Results from the assessment of the PISA 2012 core domains showed that “students’ ability to perform at high
levels is not only a function of their aptitude and talent; if students do not cultivate their intelligence with
hard work and perseverance, they will not achieve mastery in any field” (OECD, 2014, p. 111). This chapter
takes a closer look at perseverance and openness to problem solving in association with problem-solving
performance. A number of countries have been reported alongside Australia to provide an international
context. The countries of comparison are the high-performing Asian countries (Chinese Taipei, Hong
Kong–China, Japan, Korea, Macao–China, Singapore and Shanghai–China), the countries that performed
on par with Australia (Canada, England and Finland) and the two remaining English-speaking countries
(Ireland and the United States). Australian students’ results are reported at jurisdictional levels, as well as by
subgroups according to sex, geographic location, Indigenous background and socioeconomic background.
Perseverance
Perseverance relates to students’ willingness to work on tasks that are difficult, even when they encounter
setbacks. In PISA 2012, perseverance1 was assessed by asking students how well each of the statements
described them, using a 5-point Likert scale (very much like me, mostly like me, somewhat like me, not
much like me, and not at all like me):
» When confronted with a problem, I give up easily.
» I put off difficult problems.
» I remain interested in the tasks that I start.
» I continue working on tasks until everything is perfect.
» When confronted with a problem, I do more than what is expected of me.
In PISA, perseverance is not an objective measure of how students engaged or persisted with the assessment itself, but of how students perceive
themselves in terms of perseverance.
1
61
Table 5.1 shows the average percentage of students who thought the statements were ‘very much like
me’ and ‘mostly like me’ or ‘not much like me’ and ‘not at all like me’2 for Australia, the OECD average
and comparison countries. There were fewer students in Japan than any of the comparison countries
who reported that it was not like them to give up easily when confronted with a problem, or to put
off difficult problems. Students in the United States reported the highest proportion of students than
any other comparison countries who indicated that it wasn’t like them to give up easily or to put off
difficult problems.
Fewer students in Japan than any of the comparison countries also reported that it was like them to
remain interested in tasks they start, to continue working on tasks until everything was perfect, and
when confronted with a problem, to do more than what is expected of them. Of the high-performing
countries, students from Shanghai–China, Singapore and Macao–China, as well as the United States, had
the highest proportion of students who indicated it was like them to remain interested in tasks that they
started, to continue working on tasks until everything was perfect and to do more than was expected
of them.
Approximately two-thirds of Australian students indicated it was not like them to give up easily when
confronted with a problem and almost half of the Australian students indicated it was not like them to
put off difficult problems. Half the Australian students reported that it was like them to remain interested
in the tasks they started and to continue working on tasks until everything was perfect, while one-third
of Australian students reported that it was like them to do more than what was expected of them when
confronted with a problem.
Table 5.1 Students’ perseverance in problem solving for Australia and comparison countries
Percentage of students who reported the following
statement descriptions are ‘not much like me’
or ‘not at all like me’
When confronted with a
problem, I give up easily.
I put off difficult problems.
Percentage of students who reported the following statement descriptions are
‘very much like me’ or ‘mostly like me’
I remain interested in the
tasks that I start.
I continue working
on tasks until everything
is perfect.
When confronted with a
problem, I do more than
what is expected of me.
Country
%
SE
%
SE
%
SE
%
SE
%
SE
Australia
62
0.6
44
0.7
50
0.7
46
0.6
31
0.4
Canada
67
0.7
44
0.7
52
0.8
51
0.7
39
0.6
Chinese Taipei
59
0.8
45
0.8
35
1.0
31
0.8
28
0.8
England
59
0.9
44
0.7
52
0.9
47
1.0
36
0.7
Finland
59
0.8
46
0.9
45
0.9
40
0.8
28
0.6
Hong Kong–China
61
0.9
37
0.9
52
0.9
50
0.8
35
0.9
Ireland
61
0.9
45
0.9
55
0.8
48
1.0
33
0.9
Japan
32
0.9
16
0.6
29
0.9
25
0.7
12
0.7
Korea
40
1.1
20
0.8
60
1.1
44
1.1
27
1.0
Macao–China
50
0.8
34
0.7
51
0.9
53
1.0
46
0.8
Shanghai–China
53
1.0
37
0.9
73
0.7
55
1.1
38
0.8
Singapore
62
0.7
44
0.7
58
0.8
61
0.9
45
0.9
United States
70
0.8
49
0.8
57
1.0
55
0.8
44
1.1
OECD average
56
0.2
37
0.1
49
0.2
44
0.2
34
0.1
The index of perseverance was created using these five items, and standardised to have a mean of 0
and a standard deviation of 1 across the OECD student population. Higher scores on the index were
representative of higher levels of perseverance. Table 5.2 presents the mean scores for students in Australia,
the OECD average and comparison countries. These mean scores are reported overall for countries
(Table 5.2) and for females and males separately (Table 5.3, p. 64).
For ease of reading ‘very much like me’ and ‘mostly like me’ will be referred to as ‘like them’ and ‘not much like me’ and ‘not at all like me’ will be
referred to as ‘not like them’ throughout this chapter.
2
62
Thinking it through: Australian students’ skills in problem solving
Australian students’ mean score on the index of perseverance—or how persistently students were willing
to work on problems—was higher than the OECD average, indicating that Australian students reported
slightly higher levels of perseverance compared to students across the OECD. The mean index scores for
the high-performing countries were not consistent. In Japan, Korea and Chinese Taipei, students reported
the lowest mean index scores, meaning that they had lower levels of perseverance, while Shanghai–China
and Singapore reported the highest mean index scores across the high-performing countries, indicating
higher levels of perseverance. Of the comparison countries, students from the United States reported the
highest mean scores on the index of perseverance.
Table 5.2 Index of perseverance for Australia and comparison countries
All students
Mean index
SE
Australia
0.10
0.01
Canada
0.22
0.01
Chinese Taipei
–0.08
0.02
England
0.11
0.02
Country
Finland
0.00
0.02
Hong Kong–China
0.12
0.02
Ireland
0.15
0.02
Japan
–0.59
0.02
Korea
–0.09
0.02
Macao–China
0.15
0.01
Shanghai–China
0.25
0.02
Singapore
0.29
0.02
United States
0.38
0.02
OECD average
0.00
0.00
Across the OECD, there was a significant difference between the sexes on the index of perseverance with
males indicating they were more persistent in willing to work on problems than females. In all of the
comparison countries, except the United States, males reported significantly higher levels of perseverance
than females, with the largest differences in sex found in England and Korea. The United States was the
only comparison country where no significant difference on the index of perseverance was found, with
males and females reporting similar mean index scores. Both the mean index scores for Australian males
and females were significantly higher than the OECD average.
Australian students’ motivation towards problem solving
63
Table 5.3 Index of perseverance for Australia and comparison countries, by sex
Males
Country
Females
Sex difference (M–F)
Mean index
SE
Mean index
SE
Dif.
SE
Australia
0.19
0.02
0.00
0.02
0.19
0.02
Canada
0.25
0.02
0.19
0.02
0.06
0.02
Chinese Taipei
–0.04
0.02
–0.12
0.02
0.09
0.03
England
0.23
0.03
–0.01
0.02
0.23
0.03
Finland
0.07
0.02
–0.07
0.02
0.13
0.03
Hong Kong–China
0.18
0.02
0.05
0.02
0.13
0.03
Ireland
0.22
0.03
0.07
0.03
0.15
0.04
Japan
–0.55
0.02
–0.64
0.02
0.10
0.03
Korea
0.03
0.02
–0.22
0.02
0.25
0.03
Macao–China
0.19
0.02
0.11
0.02
0.09
0.02
Shanghai–China
0.32
0.03
0.17
0.02
0.16
0.03
Singapore
0.36
0.02
0.23
0.02
0.13
0.03
United States
0.39
0.03
0.36
0.03
0.03
0.04
OECD average
0.05
0.00
–0.05
0.00
0.10
0.01
Note: Bolded values indicate a statistically significant difference.
The index of perseverance was divided into quartiles. Figure 5.1 shows the relationship between
perseverance and problem-solving performance for Australia. There was a small positive association (0.22)
between the perseverance index and problem-solving performance. The pattern is linear, with higher
levels of perseverance likely to be associated with higher performance in problem solving.
Mean problem-solving performance scale
700
600
500
400
Lowest quartile
Second quartile
Third quartile
Highest quartile
Index of perseverance
Figure 5.1 Relationship between Australian students’ perseverance and problem-solving performance
Table 5.4 shows the percentage of students who indicated that the statements were like themselves or
not like themselves and the mean score on the index of perseverance according to sex, jurisdiction,
geographic location, Indigenous background and socioeconomic background. Results for Australia and
the OECD average have been included in the table for comparison.
Earlier in this chapter, it was noted that males reported higher levels of perseverance than females.
Table 5.4 provides further detail about these differences by sex. Around 10% more males than females
indicated that they were less likely to give up easily when confronted with a problem and to put off
difficult problems, and more likely to remain interested in the tasks they started.
64
Thinking it through: Australian students’ skills in problem solving
All jurisdictions had index scores that were higher than the OECD average, with students from the
Australian Capital Territory reporting the highest levels and students from the Northern Territory
reporting the lowest levels of perseverance.
Students from schools in remote areas had lower levels of perseverance than students in metropolitan or
provincial schools. Students in metropolitan schools had the highest levels of perseverance compared to
students in provincial or remote schools. Interestingly, almost half the students in remote schools indicated
it was not like them to give up easily when confronted with a problem, compared to approximately twothirds of students in metropolitan or provincial schools. There were also 10% less students in the rural
schools who indicated it was not like them to put off difficult problems compared to their peers.
The level of perseverance for Indigenous students was lower than that of non-Indigenous students and
across the OECD average. This was also the case for students in the lowest socioeconomic quartile, while
for students in the other socioeconomic quartiles, the level of perseverance increased with each increase in
socioeconomic quartile.
Table 5.4 S tudents’ perseverance in problem solving, by sex, jurisdiction, geographic location, Indigenous background
and socioeconomic background
Percentage of students who reported the
following statement descriptions are ‘not
much like me’ or ‘not at all like me’
When confronted
with a problem,
I give up easily.
I put off difficult
problems.
Percentage of students who reported the following statements
descriptions are ‘very much like me’ or ‘mostly like me’
I remain interested
in the tasks that I
start.
I continue working
on tasks until
everything is perfect.
When confronted
with a problem, I do
more than what is
expected of me.
Index of perseverance
%
SE
%
SE
%
SE
%
SE
%
SE
Mean score
SE
Female
55
0.7
37
0.9
45
0.9
47
0.9
29
0.7
0.00
0.02
Male
68
0.8
51
0.9
54
0.9
45
0.7
33
0.7
0.19
0.02
ACT
68
2.0
46
2.4
47
2.7
49
2.2
32
2.5
0.16
0.05
Sex
Jurisdiction
NSW
63
1.2
45
1.1
51
1.2
48
1.1
32
1.0
0.13
0.02
VIC
63
1.3
45
1.5
51
1.5
46
1.4
30
1.1
0.11
0.03
QLD
59
1.3
43
1.4
49
1.4
44
1.4
32
1.3
0.07
0.03
SA
61
1.7
43
1.4
48
1.7
45
1.8
29
1.4
0.06
0.03
WA
63
1.7
44
1.9
48
1.9
45
1.6
30
1.3
0.10
0.03
TAS
61
1.8
45
2.0
45
1.9
42
1.8
33
2.0
0.07
0.04
NT
56
4.2
43
3.9
49
5.0
38
4.1
31
4.7
0.02
0.10
Geographic location
Metropolitan
63
0.7
44
0.8
51
0.9
48
0.7
33
0.5
0.14
0.02
Provincial
61
1.3
45
1.3
45
1.2
40
1.2
27
1.0
0.00
0.02
Remote
46
8.8
36
3.5
46
3.2
40
4.8
31
2.2
–0.05
0.08
Indigenous background
Indigenous
48
2.0
35
1.5
41
1.7
35
1.7
74
1.3
–0.20
0.04
Non-Indigenous
63
0.6
45
0.7
50
0.7
46
0.6
31
0.5
0.11
0.01
Socioeconomic background
Lowest quartile
52
1.1
38
1.2
43
1.3
39
1.0
27
0.8
–0.10
0.02
Second quartile
59
1.2
43
1.3
46
1.3
43
1.4
29
1.1
0.04
0.03
Third quartile
67
1.2
47
1.2
53
1.4
48
1.3
31
1.2
0.17
0.02
Highest quartile
71
1.0
49
1.3
58
1.3
53
1.2
37
1.3
0.31
0.03
Country
Australia
62
0.6
44
0.7
50
0.7
46
0.6
31
0.4
0.10
0.01
OECD average
56
0.2
37
0.1
49
0.2
44
0.2
34
0.1
0.00
0.00
Australian students’ motivation towards problem solving
65
Students’ openness to experience in problem solving
Students need to be willing to engage with problems and to be open to new challenges in order to be
able to solve complex problems and situations. In PISA 2012, students’ openness to problem solving was
measured by asking students how well each of the statements described them using a 5-point Likert scale
(very much like me, mostly like me, somewhat like me, not much like me, and not at all like me):
» I can handle a lot of information.
» I am quick to understand things.
» I seek explanations of things.
» I can easily link facts together.
» I like to solve complex problems.
Table 5.5 shows the average percentage of students who thought the statements were ‘very much like me’
or ‘mostly like me’ for Australia, the OECD average and comparison countries.
Table 5.5 Students’ openness to problem solving for Australia and comparison countries
Percentage of students who reported the following statement descriptions are ‘very much like me’ or ‘mostly like me’
Country
I can handle a lot
of information.
I am quick to
understand things.
%
%
SE
SE
I seek explanations for
things.
%
SE
I can easily link
facts together.
%
SE
I like to solve
complex problems.
%
SE
Australia
49
0.7
52
0.7
63
0.6
53
0.7
31
0.6
Canada
57
0.6
61
0.6
65
0.7
60
0.7
37
0.7
Chinese Taipei
30
0.9
42
0.8
54
0.9
39
0.9
26
0.7
England
52
1.1
52
0.8
60
0.8
57
0.9
37
0.8
Finland
41
1.0
52
0.9
53
0.9
57
1.0
34
0.8
Hong Kong–China
35
0.8
48
0.9
48
0.8
42
0.8
31
1.0
Ireland
52
1.0
55
0.9
66
0.8
57
0.9
30
0.7
Japan
26
0.7
35
0.8
32
0.7
26
0.8
19
0.7
Korea
30
1.1
37
1.2
52
1.2
47
1.2
23
1.0
Macao–China
31
0.8
38
0.8
49
0.8
38
0.8
25
0.8
Shanghai–China
47
1.0
55
0.9
66
0.9
62
1.1
36
0.8
Singapore
44
0.9
50
0.9
69
0.7
52
1.0
39
0.9
United States
58
0.8
58
1.0
66
0.9
60
1.0
39
0.9
OECD average
53
0.2
57
0.2
61
0.2
57
0.2
33
0.1
The lowest percentages in agreement across all countries and for the OECD average were for the
statement ‘I like to solve complex problems’. Australian students’ percentages were slightly lower than
the OECD average for all statements except ‘I seek explanations for things’ (63% compared to 61%
respectively).
Students from Japan had the lowest percentage agreement across all five statements compared to
comparison countries and the OECD average. Students from the United States were more positive in
their engagement with problem solving, scoring higher than most other countries on all five statements.
In Australia, two-thirds of students indicated they sought explanations for things, approximately half the
students indicated they could handle a lot of information, were quick to understand things and could
easily link facts together, and one-third of the students reported to like solving complex problems.
66
Thinking it through: Australian students’ skills in problem solving
Students’ responses were standardised to calculate the index for openness to problem solving, with higher
index scores illustrative of higher levels of openness to problem solving. Table 5.6 shows that students
from Japan had the lowest levels of openness to problem solving, whereas students from the United States
reported the highest levels of openness to problem solving. Students from Australia scored below the
OECD average.
Table 5.6 Index of openness to problem solving for Australia and comparison countries
All students
Country
Mean index
SE
–0.07
0.02
Canada
0.14
0.01
Chinese Taipei
–0.33
0.02
Australia
England
–0.02
0.02
Finland
–0.11
0.02
Hong Kong–China
–0.25
0.02
Ireland
–0.02
0.02
Japan
–0.73
0.02
Korea
–0.37
0.02
Macao–China
–0.34
0.01
Shanghai–China
0.07
0.02
Singapore
0.01
0.02
United States
0.18
0.02
OECD average
0.00
0.00
Males scored significantly higher than females on the index for openness to problem solving in all the
countries listed in Table 5.7. The largest gaps between sexes were found in Japan and Hong Kong–China.
The mean index scores for males and females in Canada, Shanghai–China and the United States were
higher than the OECD average, while the mean index scores for males and females in Australia, Chinese
Taipei, Finland, Hong Kong–China, Japan, Korea and Macao–China were lower than across the OECD.
Table 5.7 Index of openness to problem solving for Australia and comparison countries, by sex
Males
Country
Females
Sex difference (M–F)
Mean index
SE
Mean index
SE
Dif.
SE
Australia
0.05
0.02
–0.19
0.02
0.24
0.02
Canada
0.26
0.02
0.02
0.02
0.23
0.03
Chinese Taipei
–0.19
0.03
–0.48
0.03
0.29
0.04
England
0.09
0.02
–0.12
0.02
0.22
0.03
Finland
0.00
0.03
–0.21
0.02
0.21
0.03
Hong Kong–China
–0.08
0.02
–0.44
0.02
0.36
0.03
Ireland
0.06
0.03
–0.10
0.02
0.16
0.04
Japan
–0.54
0.03
–0.94
0.03
0.40
0.04
Korea
–0.26
0.03
–0.50
0.03
0.24
0.04
Macao–China
–0.26
0.02
–0.42
0.02
0.17
0.03
Shanghai–China
0.21
0.03
–0.08
0.02
0.29
0.03
Singapore
0.13
0.02
–0.12
0.02
0.26
0.03
United States
0.29
0.03
0.08
0.03
0.21
0.05
OECD average
0.12
0.00
–0.12
0.00
0.23
0.01
Note: Bolded values indicate a statistically significant difference.
Australian students’ motivation towards problem solving
67
The relationship between the quartiles of the openness to problem-solving index and mean problemsolving performance is presented in Figure 5.2. It shows that students in the higher quartiles of openness
to problem solving achieved higher mean scores on problem solving. The relationship between openness
to problem solving and performance was even stronger than the relationship found between perseverance
and performance, with a correlation of 0.32, which is considered a moderate association according to
Cohen’s (1988) criteria.
Mean problem-solving performance score
700
650
600
550
500
450
400
Lowest quartile
Second quartile
Third quartile
Highest quartile
Index of openess to problem solving
Figure 5.2 Relationship between Australian students’ openness to problem solving and problem-solving performance
The percentages of students who indicated the openness to problem-solving statements were ‘very much
like me’ or ‘mostly like me’, and the index scores for different Australian subgroups, Australia and the
OECD average are shown in Table 5.8.
There were approximately 10% more males than females that reported they liked solving complex
problems, were quick to understand things and could easily link facts together. The index scores across
the jurisdictions ranged from –0.12 to 0.12. The Australian Capital Territory had the highest mean score
on the openness to problem-solving index, meaning that students from this jurisdiction reported more
willingness to engage with problems and be open to new challenges in order to be able to solve complex
problems and situations, compared to students with lower mean index scores. South Australia and
Queensland had the lowest mean index scores, while the mean index score for the Northern Territory
was the same as the OECD average.
The mean score on the index of openness to problem solving was around the OECD average for students
in metropolitan schools, whereas the mean index scores for students in provincial and—even more
so—remote schools were lower than that for students in metropolitan schools. Indigenous students had
lower mean scores on the index compared to non-Indigenous students, indicating they were less likely
to engage with problems and were less open to new challenges. This was also the case for students in the
lowest socioeconomic quartile, who had lower mean index scores compared to students in the higher
socioeconomic quartiles.
68
Thinking it through: Australian students’ skills in problem solving
Table 5.8 S tudents’ openness to problem solving, by sex, jurisdiction, geographic location, Indigenous background
and socioeconomic background
Percentage of students who reported the following statement descriptions are ‘very much like me’ or ‘mostly like me’
I can handle a lot of
information.
I am quick to
understand things.
I seek explanations
for things.
I can easily link
facts together.
I like to solve
complex problems.
Index of openness to
problem solving
%
SE
%
SE
%
SE
%
SE
%
SE
Mean
score
SE
Female
46
0.9
47
0.9
64
0.9
49
0.9
24
0.8
–0.19
0.02
Male
53
0.9
58
0.9
62
0.7
58
1.0
37
0.9
0.05
0.02
Sex
Jurisdiction
ACT
55
2.4
58
2.2
71
2.3
61
2.5
36
2.1
0.12
0.05
NSW
51
1.1
55
1.1
63
1.0
54
1.1
33
1.1
–0.03
0.02
VIC
51
1.2
53
1.3
64
1.2
54
1.5
30
1.2
–0.08
0.03
QLD
46
1.4
49
1.4
60
1.4
52
1.3
29
1.2
–0.12
0.03
SA
46
1.8
51
1.9
62
1.6
53
1.8
27
1.6
–0.12
0.04
WA
48
1.9
50
1.8
62
1.7
54
1.5
31
1.4
–0.08
0.03
TAS
49
1.9
53
1.8
65
1.7
54
2.0
32
2.0
–0.05
0.04
NT
47
3.6
52
3.3
67
3.2
53
3.8
33
3.6
0.00
0.08
Geographic location
Metropolitan
51
0.8
54
0.8
65
0.7
55
0.8
32
0.7
–0.03
0.02
Provincial
44
1.1
48
1.0
58
1.1
48
1.2
28
1.1
–0.19
0.02
Remote
42
2.8
45
4.2
56
4.3
44
4.7
28
3.2
–0.29
0.09
Indigenous background
Indigenous
37
1.7
42
2.1
52
1.6
39
1.8
23
1.5
–0.38
0.04
Non-Indigenous
50
0.7
53
0.7
63
0.6
54
0.7
31
0.6
–0.06
0.02
Socioeconomic background
Lowest quartile
39
1.1
44
1.0
54
1.1
40
1.0
24
1.0
–0.33
0.02
Second quartile
46
1.2
48
1.3
60
1.1
49
1.3
28
1.0
–0.18
0.03
Third quartile
52
1.3
55
1.3
67
1.2
58
1.4
32
1.1
0.01
0.03
Highest quartile
61
1.2
63
1.1
71
1.0
68
1.2
40
1.2
0.24
0.02
Country
Australia
49
0.7
52
0.7
63
0.6
53
0.7
31
0.6
–0.07
0.02
OECD average
53
0.2
57
0.2
61
0.2
57
0.2
33
0.1
0.00
0.00
Australian students’ motivation towards problem solving
69
References
Autor, D. H., Levy, F., & Murnane, R. J. (2003). The skill content of recent technological change: An
empirical exploration. Quarterly Journal of Economics, 118(4), 1279–1333.
Cohen, J. (1988). Statistical power analysis for the behavioural sciences (2nd ed.). New Jersey: Lawrence
Erlbaum Associates.
Hanushek, E. A., Jamison, D. T., Jamison, E. A., & Woessmann, L. (2008). Education and economic
growth: It’s not just going to school but learning something while there that matters. Education
Next, 8(2), 62–70.
OECD (2014). PISA 2012 results: Creative problem solving: Students’ skills in tackling real-life problems (Vol. V).
Paris: OECD Publishing.
Thomson, S., De Bortoli, L., & Buckley, S. (2013). PISA 2012: How Australia measures up. Melbourne:
Australian Council for Educational Research.
70
www.oecd.org
www.ozpisa.acer.edu.au
ISBN 978-1-74286-257-6
9 781742 862576
Australian Council for Educational Research
Download