Teachers’ Knowledge of Error Analysis School of Education, University of the Witwatersrand Funded by the Gauteng Department of Education 9 January 2012 Updated June 2013 Teachers’ knowledge of Error Analysis School of Education, University of the Witwatersrand Funded by the Gauteng Department of Education School of Education University of the Witwatersrand 27 St Andrews’ Road Parktown Johannesburg 2193 http://www.wits.ac.za/academic/humanities/education/l © School of Education, University of the Witwatersrand This work is licensed under a Creative Commons Attribution-NonCommercial 2.5 South Africa License. Permission is granted under a Creative Commons Attribution-NonCommercial license to replicate, copy, distribute, transmit, or adapt this work for non-commercial purposes only provided that attribution is provided as illustrated in the citation below. To view a copy of this licence visit http://creativecommons.org/licenses/by-nc/2.5/za/deed.en . Citation: Shalem, Y., & Sapire, I. 2012. Teachers’ Knowledge of Error Analysis. Johannesburg: Saide. Contents page Contents page ........................................................................... 5 Tables ......................................................................................... 7 Acknowledgements ................................................................. 9 Introduction ............................................................................ 11 Report Plan .............................................................................. 15 Section One: Teacher knowledge of mathematical errors .................................................................................................... 19 Section two: Activity and Process ....................................... 29 2.1. The Activity .....................................................................................30 Section three: Evaluation analysis methodology ............ 37 3.1 Aim of the evaluation.....................................................................38 3.2 Items evaluated ...............................................................................38 3.3 Evaluation Instrument ...................................................................39 3.3.1 The coding template (see Appendix 4): ........................................................ 40 3.4 Training of coders and coding process .......................................44 3.5 Validity and reliability check .......................................................44 3.6 Data analysis ....................................................................................44 Section Four: Analysis of data ............................................. 47 4.1 An overview of the comparison of Round 1 and Round 2 .....48 4.2. Analysis of individual criteria ....................................................49 4.2.1 Procedural explanations .................................................................................. 49 4.2.2 Conceptual explanations ................................................................................. 55 4.2.3 Awareness of Mathematical Error ................................................................. 61 4.2.4 Diagnostic reasoning ....................................................................................... 68 4.2.5 Multiple explanations ..................................................................................... 75 4.2.6 Use of the everyday in explanations of the error ....................................... 84 4.2.7 Comparative strengths and weaknesses between the two sets of grouped grades .......................................................................................................... 90 4.2.8 Comparative strengths and weaknesses according to content area ........ 93 Section Five: General findings ............................................ 95 5.1 What are the findings about teachers reasoning about learners’ errors? .....................................................................................96 5.2 Teachers’ experiences of the error analysis activity.................97 5.3 Findings from the quantitative analysis ..................................100 5.4 Summary .........................................................................................109 Section Six: Implications for professional development: developing diagnostic judgement .................................... 113 Recommendations for professional development and for further research..................................................................... 123 Recommendations for professional development ............................................ 123 Recommendations for further research............................................................... 123 References .............................................................................. 125 Appendices: ........................................................................... 129 Appendix 1................................................................................................................ 130 Appendix 2................................................................................................................ 134 Appendix 3................................................................................................................ 137 Appendix 4................................................................................................................ 140 Appendix 5................................................................................................................ 142 Appendix 6................................................................................................................ 145 Appendix 7................................................................................................................ 150 Appendix 8................................................................................................................ 155 Appendix 9................................................................................................................ 159 Appendix 10.............................................................................................................. 164 Appendix 11.............................................................................................................. 168 6 Tables Table 1: Domains of teacher knowledge and related error analysis categories ................. 21 Table 2 “Error-focused activities” and “error-related activities” ........................................ 30 Table 3: Round 1 Error analysis ................................................................................................ 32 Table 4: Round 2 Curriculum Mapping and Error Analysis ................................................ 33 Table 5: Number of items analysed by groups (Rounds 1 and 2)........................................ 35 Table 6: Sample summary ......................................................................................................... 38 Table 7: Grades 3-6 procedural explanations demonstrated in teacher text explanations50 Table 8: Grades 7-9 procedural explanations demonstrated in teacher text explanations51 Table 9: Procedural explanations demonstrated in teacher test explanations ................... 52 Table 10: Grades 3-6 conceptual explanations demonstrated in teacher text explanations ....................................................................................................................................................... 55 Table 11: Grades 7-9 conceptual explanations demonstrated in teacher text explanations ....................................................................................................................................................... 56 Table 12: Conceptual explanations demonstrated in teacher test explanations ................ 58 Table 13: Grades 3-6 awareness of mathematical error demonstrated in teacher text explanations ................................................................................................................................ 62 Table 14: Grades 7-9 awareness of mathematical error demonstrated in teacher text explanations ................................................................................................................................ 63 Table 15: Awareness of the mathematical error demonstrated in teacher test explanations ....................................................................................................................................................... 65 Table 16: Grades 3-6 diagnostic reasoning demonstrated in teacher text explanations ... 68 Table 17: Grades 7-9 diagnostic reasoning demonstrated in teacher text explanations ... 69 Table 18: Diagnostic reasoning in relation to the error demonstrated in teacher test explanations ................................................................................................................................ 71 Table 19: Grades 3-6 multiple explanations demonstrated in teacher text explanations . 77 Table 20: Grades 7-9 multiple explanations demonstrated in teacher text explanations . 78 Table 21: Multiple explanations of the error demonstrated in teacher test explanations 81 Table 22: Grades 3-6 use of the everyday demonstrated in teacher text explanations of the error........................................................................................................................................ 85 Table 23: Grades 7-9 use of the everyday demonstrated in teacher text explanations of the error........................................................................................................................................ 86 Table 24: Use of the everyday in explanations of the error demonstrated in teacher test explanations ................................................................................................................................ 88 Table 25: Content areas according to strength and weakness in explanations .................. 93 Table 26: An example of an explanation that is accurate but incomplete ........................ 110 Table 27: Grade 5 test item explanations illustrating poor diagnostic judgement .......... 111 Table A1: Category descriptors for “procedural explanations” ........................................ 145 Table A2: Criterion 1 – Procedural explanation of the choice of the correct solution in relation to one item................................................................................................................... 146 7 Table A3: Criterion 1 – Procedural explanation of the choice of the correct solution in relation to a range of items ...................................................................................................... 147 Table A4: Category descriptors for “conceptual explanations”......................................... 150 Table A5: Criterion 2 – Conceptual explanation of the choice of the correct solution in relation to one item................................................................................................................... 151 Table A6: Criterion 2 – Conceptual explanation of the choice of the correct solution in relation to a range of items ...................................................................................................... 152 Table A7: Category descriptors for “awareness of mathematical error”.......................... 155 Table A8: Criterion 3 – Awareness of the error embedded in the incorrect solution in relation to one item................................................................................................................... 155 Table A9: Criterion 3 – Awareness of the error embedded in the incorrect solution in relation to a range of items ...................................................................................................... 157 Table A10: Category descriptors for “diagnostic reasoning”............................................. 159 Table A11: Criterion 4 – Diagnostic reasoning of learner when selecting the incorrect solution in relation to one item ............................................................................................... 160 Table A12: Criterion 4 – Diagnostic reasoning of learner when selecting the incorrect solution in relation to a range of items .................................................................................. 161 Table A13: Category descriptors for “use of everyday” ..................................................... 164 Table A14: Use of the everyday exemplars ........................................................................... 165 Table A15: Category descriptors for “multiple explanations” ........................................... 168 Table A16: Multiple explanations exemplars ....................................................................... 168 8 Acknowledgements This report was prepared by: Professor Yael Shalem (Wits School of Education) Ingrid Sapire (Wits School of Education) Tessa Welch (Saide) Maryla Bialobrzeska (Saide) Liora Hellmann (Saide) We would like to thank the following people for their role in the preparation of the report: M Alejandra Sorto (Maths Education Consultant) David Merand (Maths Education Consultant) Gareth Roberts (Statistical computation and analysis) Carola Steinberg (Editing) The project team for DIPIP Phases 1 and 2 Project Director: Prof Yael Shalem Project Leader: Prof Karin Brodie Project coordinators: Lynne Manson, Nico Molefe, Ingrid Sapire The project team and evaluators would like to thank the Gauteng Department of Education and in particular Reena Rampersad and Prem Govender for their support of the project. We would also like to thank all of the teachers, departmental subject facilitators, Wits academic staff and students and the project administrators (Karen Clohessy, Shalati Mabunda and team) who participated over the three years of the project. 9 10 Introduction New imperatives held in high regard in South Africa are the growing emphasis on learners’ performance as an indicator of accountability and the use of assessment to inform learners’ learning and to monitor teachers’ work. Current policy is very clear about the accountability imperative. The Department of Education has stipulated that teachers and district officials monitor learners’ performance continuously and write reports on their progress. It has also conducted and participated in a range of regional and international evaluations, amongst them: the Trends in International Mathematics and Science Study (TIMSS: 1995, 1999, 2003); Systemic Evaluation for grades 3 and 6 in Numeracy and Literacy; participation in Southern and Eastern Africa Consortium for Monitoring Educational Quality (SACMEQ: 2003,2007), the Progress in International Reading Literacy Study (PIRLS), and the most recent one being the Annual National Assessment (ANA). As part of the regional evaluation exercise, the Gauteng Department of Education used the ICAS test (International Competitions and Assessments for Schools). The ICAS test is an international standardized, primarily multiple-choice test, which was administered in Gauteng to a population of about 55 000 learners in public and private schools in Grades 3 to 11 in 2006, 2007 and 2008. The tests and data of learners’ errors on the 2006 and 2007 tests of Grade 3-9 provided the starting point for teacher engagement with learners’ errors, the main subject of this report. The data from standardised evaluations has been used by mathematical experts, economists and statisticians at a systemic level and solely for benchmarking.1 Results have been used to monitor the process of educational reform.2 In this process the common messages heard by the public in general and by teachers in particular are that “high participation” has come at the expense of “poor quality”; that the system demonstrates low value and poor accountability; and that South Africa performs poorly in comparison to countries in Africa poorer than her. Teachers were blamed and shamed for their low subject matter content knowledge, and poor professionalism. The general view that has emerged is that good and committed teachers are pivotal for the desired change. As important as they are, these kind of messages have done very little to change the results, and certainly contributed to teachers’ low morale and loss of public confidence in the capacity of education to deliver a meaningful change (Taylor, 2008; Shalem & Hoadley, 2009; Van den Berg, 2011). Important for our task here, is that although the Department of Education (and the public) expect teachers to use the results In South Africa, matric exams, Grade 3 and 6 systemic evaluation and international and regional tests have been used to benchmark the performance of the system or as a tool which, by controlling social and economic factors that affect learners’ performance, reports to the public on learners’ standards of proficiently (van Den Berg & Louw 2006; van den Berg & Shepherd; 2008; Taylor 2007, 2008). 1 2 In South Africa since 1994 there have been three curriculum changes to which teachers and learners have had to adjust. 11 to improve their practice accordingly, to date teachers have not been invited to interpret the results nor been shown how to integrate them into their practice. Teachers have not been given the opportunity to develop skills to interpret results from systemic evaluations. This is despite the current rhetoric of alignment between standardised assessment and classroom teaching. With the advent of the Annual National Assessments (ANA), there have been new kinds of pronouncements. For example, the Department of Basic Education is reported to have stated the following: ANA is intended to provide regular, well-timed, valid and credible data on pupil achievement in the education system. Assessment of pupils’ performance in the GET Band (Grades 1- 9) has previously been done at school level. Unlike examinations that are designed to inform decisions on learner promotion and progression, ANA data is meant to be used for both diagnostic purposes at individual learner level and decision-making purposes at systemic level. At the individual learner level, the ANA results will provide teachers with empirical evidence on what the learner can and/or cannot do at a particular stage or grade and do so at the beginning of the school year. Schools will inform parents of their child’s ANA performance in March 2011 (our emphasis). 3 The above statement includes two important claims, which together suggest an approach that attempts to go beyond the current use of standardized assessment for benchmarking purposes. The first claim is that assessment should now be dealt with on two levels and in two forms: at a school level, through the use of specific and localised forms of assessment and at a national level, through the use of general and standardised forms of assessment. This is not a new claim and is in line with the Department of Education policy of formative assessment. It suggests that worthwhile assessment data should consist of evidence that is valid for specific learning contexts (or schools) together with reliable and generalizable evidence that represents different learning contexts (or schools). In assessment argot, worthwhile assessment data consist of external assessment and classroom assessment (Gipps, 1989; Brookhart et al, 2009; Looney, 2011)4. The second claim, which is particularly relevant for this report, is new and suggests a new direction in the Department of Education policy. This claim suggests that data from external assessment is intended to be used diagnostically both at the level of the individual learner and at the systemic level of a grade cohort. The idea of informing local knowledge using a systemic set of evidence, diagnostically, is not sufficiently theorized. Teachers have always needed to recognise learners’ errors, a skill without which they would not have been able to assess learners’ work. The difference now is that teachers are required “to interpret their own learners’ performance in national (and other) assessments” (DBE and DHET, 2011, p2) and develop better lessons on the basis of these interpretations. According to the above statements, the tests will Media statement issued by the Department of Basic Education on the Annual National Assessments (ANA): 04 February 2011. http://www.education.gov.za/Newsroom/MediaReleases/tabid/347/ctl/Details/mid/1389/ItemID/3148/Default.aspx 3 Gipps argues that externally administered standardised assessments (systematic assessment) are considered more reliable, in terms of their process (design and marking). Classroom assessment, if well designed and on-going, provides a variety of contextually-relevant data. 4 12 provide teachers with information that will tell them, in broad strokes, how close or far their learners are from national and/or international standards. In other words, teachers will be given a set of “evaluation data” in the form of pass rates and proficiency scores. This type of information is commonly known as “criterion-referenced proficiency classification” (Looney, 2011, p27). What is not clear and has not been researched is what is involved on the part of teachers in interpreting this data: what do teachers do and what should teachers be doing so that the interpretation process is productive? How does teachers’ tacit knowledge about learners’ errors (which they have acquired from years of marking homework and tests, encountering learners’ errors in teaching, or through hearing about errors from their colleagues) inform or misinform their reasoning about evaluation data? In what ways should teachers work with proficiency information when they plan lessons, when they teach or when they design assessment task? Important to remember is that curriculum statements about assessment standards, recurring standardised assessments, or reports on pass rates and proficiency levels do not, in themselves, make standards clear (let alone bring about a change in practice). Reported data of standardized assessment provides information about what learners can or can’t do but does not analyse what the learners do not understand, how that may affect their poor performance, and how preferred instructional practice could afford or constrain addressing these difficulties (Shepard, 2009, p37). These are ideas that teachers need to interpret from the evaluation data. Hence Katz et al are correct when they say: Data don’t “tell” us anything; they are benign... The meaning that comes from data comes from interpretation, and interpretation is a human endeavour that involves a mix of insights from evidence and the tacit knowledge that the group brings to the discussion. (2009, p.28) 5 These kinds of questions, stated above, inform our approach to teachers’ learning about their practice through engagement with error analysis. They motivated our professional development work in the Data-Informed Practice Improvement Project. The DataInformed Practice Improvement Project (henceforth, the DIPIP project) has been the first attempt in South Africa to include teachers in a systematic way in a process of interpretation of 55, 000 learners’ performance on a standardized test. The three-year research and development programme6 included 62 mathematics teachers from Grades 3 to 9 from a variety of Johannesburg schools. The project involved teachers in analysing learners’ errors on multiple choice items of the ICAS mathematics test. In the project For Hargreaves (2010), focussing teachers’ learning from data is important for building collegiality. He argues that the future of collegiality may best be addressed by (inter alia) taking professional discussion and dialogue out of the privacy of the classroom and basing it on visible public evidence and data of teachers’ performance and practices, such as shared samples of student work or public presentations of student performance data (p.524). 6 There is currently a third phase of DIPIP which is located in certain schools following a very similar process with teacher groups in these schools. 5 13 teachers mapped the ICAS test’s items onto the curriculum, analysed learners’ errors, designed lessons, taught and reflected on their instructional practices, constructed test items, all in a format of “professional learning communities”.7 With regard to its focus on working with learners’ errors in evaluation data, the work done with teachers in the DIPIP project is an exception. Its main challenge has been to design a set of meaningful opportunities for teachers to reason about the evaluation data collated from the results of 55,000 learners’ who wrote the test, in each of the years that it ran (2006-2008). The central conceptual questions we addressed in researching the results of this process are: What does the idea of teachers’ interpreting learner performance diagnostically, mean in a context of a standardized assessment test? What do teachers, in fact, do when they interpret learners’ errors? In what ways can what they do be mapped on the domains of teacher knowledge? In the field of mathematics, Prediger (2010) uses the notion of “diagnostic competence” to distinguish reasoning about learners’ errors from merely grading their answers: The notion of diagnostic competence (that, in English, might have some medical connotations) is used for conceptualizing a teacher’s competence to analyse and understand student thinking and learning processes without immediately grading them. (p76) Prediger (2010) argues that teachers seem to be better interpreters of learners’ errors when they have an interest in learners’ rationality, are aware of approaches to learning, interrogate meanings of concepts (in contrast to just knowing their definitions), and have studied domain specific mathematical knowledge. His view is consistent with Shepard’s work on formative assessment, in particular with the idea of using insights from student work, formatively, to adjust instruction (2009, p34). Research in mathematics education has shown that a focus on errors, as evidence of reasonable and interesting mathematical thinking on the part of learners, helps teachers to understand learner thinking, to adjust the ways they engage with learners in the classroom situation, as well as to revise their teaching approach (Borasi, 1994; Nesher, 1987; Smith, DiSessa & Roschelle, 1993). The field of Maths education has developed various classifications of learners’ errors (Radatz, 1979; Brodie & Berger, 2010) but there is hardly any work on what kinds of criteria can be used to assess teacher knowledge of error analysis. There is no literature that examines the ways teachers’ reason about evaluation data. There is hardly any literature that shows what would be involved in assessing the quality of teachers’ reasoning about evaluation data gathered from a systematic assessment.8 See Shalem, Sapire, Welch, Bialobrzeska, & Hellman, 2011. See also Brodie & Shalem, 2010. Peng and Luo (2009) and Peng (2010) present one attempt to classify the tasks teachers engage with when they analyse learners’ errors (identifying, addressing, diagnosing, and correcting errors). 7 8 14 In this report we hope to show a way of assessing teachers’ knowledge of error analysis in relation to specific criteria and mathematical content. Our central claim is that without teachers’ being able to develop this kind of knowledge the “empirical evidence on what the learner can and/or cannot do at a particular stage or grade”,9 will remain non-integrated with teacher practice. Report Plan Section one This is the conceptual section of the report. The conceptual aim of this report is to locate the idea of error analysis in teacher knowledge research, in order to develop criteria for assessing what it is that teachers do when they interpret evaluation data. We develop the idea of interpreting evaluation data diagnostically and show a way of mapping its six constitutive aspects against the domains of teacher knowledge, as put forward by Ball et al (2005, 2008) and Hill et al (2008a). By locating these aspects in the broader discussion of teacher knowledge, we are able to demonstrate the specialisation involved in the process of teachers’ reasoning about learners errors. We use the word ‘mapping’ consciously, because we believe that the task of error analysis requires both subject matter knowledge and pedagogical content knowledge. Section two In this section we describe the error analysis activity, in the context of the DIPIP project. We draw a distinction between different error analysis activities, between “errorfocused activities” and “error-related activities”. This distinction structures the sequence of activities in the DIPIP project. We describe the time-line of the project, as some of the activities were repeated twice and even three times throughout its threeyear duration. In this report we only examine the error analysis activity, which forms the central focus of the “error-focused activities” and which was done twice. We refer to this time-line as Round 1 and Round 2 and use it to compare teachers’ reasoning about learners’ errors across these two rounds. The main difference between the two rounds is that in Round 1 the teachers worked in small groups led by a group leader, while in Round 2 the small groups worked without a group leader. In comparing the two rounds, we are able to suggest some ideas about the role of group leaders in this kind of professional development project. In this section we also describe the ICAS test we used for the error analysis activity and we provide information about the number of items analysed by the groups in Round 1 and 2 as well as the training provided for the group leaders. Media statement issued by the Department of Basic Education on the Annual National Assessments (ANA): 04 February 2011. http://www.education.gov.za/Newsroom/MediaReleases/tabid/347/ctl/Details/mid/1389/ItemID/3148/Default.aspx 9 15 Section three In this section we describe the methodology we followed for the evaluation of the findings of the error analysis activity. We state the three central aims of the evaluation of the error analysis activity. We then describe how we selected the items used as data for the evaluation, the evaluation instrument, the coding template used to code the data, the training of the coders and the coding process, including its validity and reliability. We conclude with the description of the three-level approach we followed to analyse the data. Section four This section presents what we call the level 1 analysis of findings. In this section we detail the findings for each of the six criteria we selected for the analysis of teachers’ reasoning about the correct and the erroneous answers. This section of the report brings us to the empirical aim of the report, which is to measure and analyze the quality of teachers’ ability to do diagnostic error analysis, across the constitutive aspects of interpreting evaluation data. The empirical analysis intends to answer the following three empirical questions: On what criteria is the groups’ interpretation weaker and on what criteria is the groups’ interpretation stronger? In which mathematical content areas do teachers produce better judgement on the errors? Is there a difference in the above between two sets of grouped grades (Grade 3-6 and Grade 7-9)? Section five This section presents what we call the level 2 analysis of findings. In this section we begin to draw together broader findings across criteria and groups. We present several findings that help us construct an argument about the relationship between subject matter knowledge and pedagogical content knowledge as it pertains to error analysis. We also present findings that compare the two grouped grades (Grade 3-6 and Grade 79) and describe the role of the group leaders. We conclude with some implications for the idea of interpreting evaluation data diagnostically. Section six This section presents what we call the level 3 analysis of findings. In terms of the professional development aspect of the report, the discussion in this section begins with the ways in which teachers could be involved in analysing evaluation data. Then, based on the findings of the error analysis activity and the mapping of error analysis against 16 the domains of teacher knowledge, we conclude with a short description of “diagnostic judgment”. Diagnostic judgement is the construct we propose for describing what teachers do when they interpret evaluation data. We argue that understanding teachers’ professionalism in this way may help the discourse to shift away from the dominance of “accounting” to “accountability”. We conclude with lessons to be learned from the project and recommendations both for professional development and research. 17 18 Section One: Teacher knowledge of mathematical errors 19 No matter how the South African curriculum debate of the last 10 years will be resolved or the change away from OBE to a content-based curriculum is enacted, no matter what educational philosophy informs our debate on good teaching, the research on teacher knowledge insists that teachers today are expected (or should be able) to make sound judgments on sequence, pacing and evaluative criteria so as to understand learners’ reasoning and to inform learners’ learning progression (Muller, 2006; Rusznyak, 2011; Shalem & Slonimsky, 2011). The central question that frames studies on teacher knowledge, in the field of mathematics education, goes as follows: “is there a professional knowledge of mathematics for teaching which is tailored to the work teachers do with curriculum materials, instruction and students?” (Ball, Hill & Bass, 2005, p16). This question draws on Shulman’s work, specifically, on his attempt to define “knowledge-in-use in teaching”. Shulman’s main project was to situate subject matter knowledge within the broader typology of professional knowledge. Continuing within this tradition Ball, Thames and Phelps (2008) elaborate Shulman’s notion of pedagogical content knowledge and its relation to subject matter knowledge for mathematics teaching. Ball, Thames and Phelps define mathematical knowledge for teaching as “the mathematical knowledge needed to carry out the work of teaching mathematics” (p395). In order to investigate what this in fact means, they collected extensive records of specific episodes, analyses of curriculum materials, and examples of student work. They also draw on Ball’s personal experience of teaching and researching in a Grade 3 mathematics class for a year. On the basis of these resources, they then developed over 250 multiple-choice items “designed to measure teachers’ common and specialized mathematical knowledge for teaching” (Ball, Hill & Bass 2005, p43). Some of the items emphasise mathematical reasoning alone and others include the more specialized knowledge for teaching (Hill, Ball & Schilling, 2008, p376; Ball, Hill & Bass 2005, pp 22 and 43). They then conducted large-scale surveys of thousands of practicing teachers as well as interviews with smaller number of teachers. They wanted to know “what ‘average’ teachers know about students’ mathematical thinking” (Hill, Ball & Schilling, 2008, p376), whether this specialized knowledge is different from mathematical reasoning and whether it can be shown to affect the quality of instruction (Ball et al, 2008b). Their work is summarised in several papers, showing that teachers’ mathematical knowledge consists of four key domains (Ball, Hill & Bass, 2005; Ball, Thames & Phelps 2008; Hill, Ball & Schilling, 2008).10 The first two domains elaborate the specialisation of 10 They also include two other domains, Knowledge of Curriculum and Knowledge at the mathematical horizon. 20 subject-matter knowledge (“common content knowledge” and “specialized content knowledge”). The second two domains elaborate the specialisation involved in teaching mathematics from the perceptive of students, curriculum and pedagogy. These domains (“knowledge of content and students” and “knowledge of content and teaching”) elaborate Shulman’s notion of “pedagogical content knowledge”. 11 Although there is some debate on the specificity of the domains of knowledge in the realm of teachers’ knowledge in mathematics and other authors have expanded on them in a number of academic papers (e.g. Adler, 2005), for the purpose of the DIPIP error analysis evaluation we use Ball’s classification of the domains and in what follows we present each of the domains, by foregrounding key aspects of ‘error analysis’ relevant to each of the domains. We also formulate the criteria that are aligned with the domain and explain why. The following table presents the way we mapped the evaluation criteria in relation to teacher knowledge. Table 1: Domains of teacher knowledge and related error analysis categories Subject Matter Knowledge Pedagogical Content Knowledge Knowledge domain Knowledge domain Knowledge domain Knowledge domain Common content knowledge (CCK) Specialized content knowledge (SCK) Knowledge of content and students (KCS) Knowledge of content and teaching (KCT) DIPIP category DIPIP category DIPIP category DIPIP category Procedural understanding of correct answers Awareness of Errors Diagnostic reasoning of learners’ thinking in relation to errors N/A (only in lesson design and teaching) Conceptual understanding of correct answers Use of everyday links in explanations of errors Multiple explanations The principle point behind this typology of the 4 domains is that mathematics teaching is a specialized practice, which combines mathematical and pedagogical perspectives. The argument behind this work is that a specialized mathematical perspective includes things like “determining the validity of mathematical arguments or selecting a mathematically appropriate representation”, (Ball, Thames and Bass, 2008 p398) but also “skills, habits of mind, and insight” (p399) or mathematical reasoning. The specialized pedagogical perspective on the other hand requires knowledge of curriculum, of learners and of teaching - it focuses on learners’ reasoning and requires specialised pedagogical tasks. 11 21 of errors The first domain is common content knowledge, which is general subject matter knowledge (CCK). Ball, Thames and Bass (2008) define this domain of teacher knowledge as follows: Teachers need to know the material they teach; they must recognize when their students give wrong answers or when the textbook gives an inaccurate definition. When teachers write on the board, they need to use terms and notation correctly. In short, they must be able to do the work that they assign their students. But some of this requires mathematical knowledge and skill that others have as well—thus, it is not special to the work of teaching. By “common,” however, we do not mean to suggest that everyone has this knowledge. Rather, we mean to indicate that this is knowledge of a kind used in a wide variety of settings—in other words, not unique to teaching. (pp. 398-399) The argument here is that like any other mathematician, a mathematics teacher with good subject matter knowledge uses mathematical terms correctly, is able to follow a procedure fully, and can evaluate that a textbook defines a mathematical term correctly or incorrectly. In terms of error analysis, this knowledge is about recognizing that a learner’s answer is correct or not. Recognition of errors is a necessary component of teachers’ content knowledge. It is necessary, as it shows the “boundaries of the practice” of doing mathematics, of “what is acceptable and what is not” (Brodie, 2011, p66). The underlying condition, i.e. the pre-condition that enables teachers to recognize error mathematically, is for teachers to be able to explain the mathematical solutions of the problem, both procedurally and conceptually. The emphasis in this domain is on teachers’ ability to explain the correct answer. Recognizing “when their students give wrong answers or when the textbook gives an inaccurate definition” relies on knowing the explanation of a solution in full. Without a full knowledge of the explanation, teachers may recognize the error only partially. “Only partial”, because they may not know what the crucial steps that make up the solution are or what their sequence needs to be (procedural explanations). “Only partial”, because they may not know what the underlying conceptual links are that the student needs to acquire in order not to err. Although Ball et al argue that there are some aspects of what teachers know which is not unique to teaching, and although, one would argue, it is not expected that mathematicians will distinguish procedural from conceptual explanations when they address a mathematical solution, these two kinds of explanations are essential aspects of teachers’ content knowledge and are the enablers of recognition of learners’ errors. Star notes that “it is generally agreed that knowledge of concepts and knowledge of procedures are positively correlated and that the two are learned in tandem rather than independently” (2000, p80). The first aspect of mathematical knowledge that teachers need in order to “carry out the work of teaching mathematics”, specifically of error 22 analysis, is content knowledge. This involves both knowledge of the crucial sequenced steps needed to get to the correct answer (procedural knowledge) and their conceptual links (conceptual knowledge) 12. Because this knowledge underlies recognition of error, we call it “content knowledge” (irrespective of whether it is common or not to mathematical experts in general). This means we included two criteria under common content knowledge. Criteria for recognizing Common Content Knowledge: Procedural understanding The emphasis of the criterion is on the quality of the teachers’ procedural explanations when discussing the solution to a mathematical problem. Teaching mathematics involves a great deal of procedural explanation which should be done fully and accurately for the learners to grasp and become competent in working with the procedures themselves. Conceptual understanding The emphasis of the criterion is on the quality of the teachers’ conceptual links made in their explanations when discussing the solution to a mathematical problem. Teaching mathematics involves conceptual explanations which should be made with as many links as possible and in such a way that concepts can be generalised by learners and applied correctly in a variety of contexts. The second domain is Specialised Content Knowledge (SCK), which is mathematical knowledge specific to teaching and which, according to Ball et al, general mathematicians do not need. A different aspect of error analysis is located in this domain, which involves activities such as,“looking for patterns in student errors or … sizing up whether a nonstandard approach would work in general” (Ball, Thames & Bass, 2008, p400). Whereas teacher knowledge of the full explanation of the correct answer enables a teacher to spot the error, teacher knowledge of mathematical knowledge for teaching enables a teacher to interpret a learner’s solution and evaluate its plausibility, by recognizing the missing steps and /or conceptual links and taking into account, we would add, other factors such as the age of the learner, or the time the learner was given to complete the task, the complexity in the design of the question etc. Notwithstanding our addition, the main idea about “sizing up whether a nonstandard approach would work in general”, assumes that teachers recognize errors relationally. Some mathematical problems lend themselves more to procedural explanations while in others the procedural and the conceptual are more closely linked. There is a progression in mathematical concepts – so that what may be conceptual for a grade 3 learner (for example, basic addition of single digit numbers) is procedural for a grade 9 learner who will have progressed to operations at a higher level. 12 23 They evaluate the “nonstandard approach” and/or “the error” in relation to what is generally considered the correct approach, taking into account the context of the “nonstandard approach” and/or the error. In Ball et al’s words, knowledge of this domain enables teachers to “size up the source of a mathematical error” (Ball, Thames & Bass, 2008, p397) and identify what mathematical step/s would produce a particular error. Error analysis is a common practice among mathematicians in the course of their own work; the task in teaching differs only in that it focuses on the errors produced by learners… Teachers confront all kinds of student solutions. They have to figure out what students have done, whether the thinking is mathematically correct for the problem, and whether the approach would work in general. (ibid)13 It is important to emphasise that although, as Ball et al define above, the knowledge of errors in this domain focuses on what “students have done”, teachers’ reasoning about the error is mathematical (i.e. not pedagogical) in the main. In their analysis it forms the second domain of teachers’ content knowledge: deciding whether a method or procedure would work in general requires mathematical knowledge and skill, not knowledge of students or teaching. It is a form of mathematical problem solving used in the work of teaching. Likewise, determining the validity of a mathematical argument, or selecting a mathematically appropriate representation, requires mathematical knowledge and skill important for teaching yet not entailing knowledge of students or teaching. (p398) Ball, Thames and Bass characterize this type of knowledge as “decompressed mathematical knowledge” (2008b, p400) which a teacher uses when s/he unpacks a topic for the learner or make “features of particular content visible to and learned by students” (ibid, see also Prediger, 2000, p79). Teaching about place value, for example, requires understanding the place-value system in a self-conscious way that goes beyond the kind of tacit understanding of place value needed by most people. Teachers, however, must be able to talk explicitly about how mathematical language is used (e.g., how the mathematical meaning of edge is different from the everyday reference to the edge of a table); how to choose, make, and use mathematical representations effectively (e.g., recognizing advantages and disadvantages of using rectangles or circles to compare fractions); and how to explain and justify one’s mathematical ideas (e.g., why you invert and multiply to divide fractions). All of these are examples of ways in which teachers work with mathematics in its decompressed or unpacked form. (ibid) So, for example (Ball, 2011, Presentation), when marking their learners’ work, teachers need to judge and be able to explain to the learners which definition of a concept (‘rectangle’, in the following example) is more accurate: 13 a rectangle is a figure with four straight sides, two long and two shorter a rectangle is a shape with exactly four connected straight line segments meeting at right angles a rectangle is flat, and has four straight line segments, four square corners, and it is closed all the way around. 24 In relation to Shulman’s distinction between subject matter knowledge and pedagogical content knowledge, Ball, Thames and Bass, (2008) argue that the above two domains are framed, primarily, by subject matter knowledge. This is very important from the perspective of examining teachers’ interpretation of evaluation data. As Peng and Luo (2009) argue, if teachers identify learner’s errors but interpret them with wrong mathematical knowledge, their evaluation of student performance or their plan for a teaching intervention are both meaningless. In other words, the tasks that teachers engage with, in error analysis, such as sizing up the error or interpreting the source of its production, are possible because of the mathematical reasoning that these domains of teacher knowledge equip them with. The idea behind this is that strong mathematics teachers recruit content knowledge into an analysis of a teaching situation and do so by recruiting their “mathematical reasoning” more than their knowledge of students, teaching or curriculum. This mathematical reasoning enables strong mathematics teachers to size up the error from the perspective of socializing others into the general field of mathematics. We included one criterion under Specialized Content Knowledge. Criteria for recognizing Specialized Content Knowledge Awareness of error This criterion focuses on teachers’ explanations of the actual mathematical error and not on learners’ reasoning. The emphasis in the criterion is on the mathematical quality of teachers’ explanations of the actual mathematical error. The third domain of Knowledge of Content and Students orients teachers’ mathematical perspective to the kind of knowing typical of learners of different ages and social contexts in specific mathematical topics. Teachers develop this orientation from their teaching experience and from specialized educational knowledge of typical misconceptions that learners develop when they learn specific topics. Ball, Thames and Bass (2008) state as examples of this domain of teacher knowledge, “the kinds of shapes young students are likely to identify as triangles, the likelihood that they may write 405 for 45, and problems where confusion between area and perimeter lead to erroneous answers” (p401). This knowledge also includes common misinterpretation of specific topics or levels of the development in representing a mathematical construct (e.g. van Hiele levels of development of geometric thinking). Teacher knowledge in this domain contains “knowledge that observant teachers might glean from working with students, but that have not been codified in the literature” (Hill, Ball & Schilling, 2008, p378). So for example, after years of teaching, teachers gain an understanding of how to define a mathematical concept that is both accurate and appropriate to Grade 2 learners, they come to know what aspect of the definition of a mathematical concept is more difficult for these learners, and how the learners’ everyday knowledge (of the particular age 25 group) can come in the way of acquiring the specialised mathematical knowledge (Ball, 2011). From the point of error analysis, this knowledge domain involves knowing specific mathematical content from the perspective of how learners typically learn the topic or “the mistakes or misconceptions that commonly arise during the process” of learning the topic (Hill, Ball & Schilling, 2008, p375). Ball, Thames and Bass (2008) emphasise that this type of knowledge builds on the above two specialized domains but yet it is distinct: … Recognizing a wrong answer is common content knowledge (CCK), whereas sizing up the nature of an error, especially an unfamiliar error, typically requires nimbleness in thinking about numbers, attention to patterns, and flexible thinking about meaning in ways that are distinctive of specialized content knowledge (SCK). In contrast, familiarity with common errors and deciding which of several errors students are most likely to make are examples of knowledge of content and students (KCS). (p401) The knowledge of this domain enables teachers to explain and provide a rationale for the way the learners were reasoning when they produced the error. Since it is focused on learners’ reasoning, it includes, we argue, the ability to provide multiple explanations of the error. Because contexts of learning (such as age and social background) affect understanding and because in some topics the learning develops through initial misconceptions, teachers will need to develop a repertoire of explanations, with a view to addressing differences in the classroom. How is this knowledge about error analysis different from the first two domains? The first two domains, being subject-matter based, combine knowledge of the correct solution, (which includes both a knowledge of the procedure to be taken as well as the underlying concept, “Common Content Knowledge”) with knowledge about errors (“Specialized Content Knowledge” which is focused on the relation between the error / the non-standard solution and the correct answer / the answer that is generally accepted). The social context of making the error is secondary to the analysis of the mathematical content knowledge involved in explanting the solution or the error and therefore is back-grounded. “Knowledge of content and students”, the third domain, is predominantly concerned with context-specific experiential knowledge about error that teachers develop, drawing on their knowledge of students (Hill et al, 2008b, p 385).14 Hill et al discuss the difficulties they found in measuring KCS. They believe that “logically, teachers must be able to examine and interpret the mathematics behind student errors prior to invoking knowledge of how students went astray” (2008b, p390). They found that teachers’ mathematical knowledge and reasoning (first and second domains) compensate when their knowledge of content and student is weak. This has implications for their attempt to measure the distinctiveness of KCS, which is beyond the scope of this report. 14 26 The diagnostic aspect of this domain is focused on learner reasoning and the general mathematical content knowledge of the correct answer is secondary and therefore is back-grounded. We included three criteria under Knowledge of Content and Students. Criteria for recognizing Knowledge of Content and Students Diagnostic reasoning The idea of error analysis goes beyond identifying the actual mathematical error (“awareness of error”). The idea is to understand how teachers go beyond the mathematical error and follow the way learners were reasoning when they produced the error. The emphasis in this criterion is on the quality of the teachers’ attempt to provide a rationale for how learners were reasoning mathematically when they chose a distracter. Use of everyday knowledge Teachers sometimes explain why learners make mathematical errors by appealing to everyday experiences that learners draw on and/or confuse with the mathematical context of the question. The emphasis in this criterion is on the quality of the use of everyday knowledge in the explanation of the error, judged by the links made to the mathematical understanding that the teachers attempt to advance. Multiple explanations of error One of the challenges in the teaching of mathematics is that learners need to hear more than one explanation of the error. This is because some explanations are more accurate or more accessible than others and errors may need to be explained in different ways for different learners. This criterion examines the teachers’ ability to offer alternative explanations of the error when they are engaging with learners’ errors. Knowledge of Content and Teaching (KCT) is the fourth domain of teacher knowledge. This type of knowledge links between subject matter knowledge content (CCK+SCK) and knowledge about instruction, which takes into account knowledge about students (KCS). Based on their knowledge of these three domains, teachers use their knowledge of teaching to decide on the sequence and pacing of lesson content or on things such as which learner’s contribution to take up and which to ignore. This domain of knowledge includes “knowledge of teaching moves”, such as “how best to build on student mathematical thinking or how to remedy student errors” (Hill et al, 2008b, p378). As this domain is concerned with teachers’ actively teaching it falls outside the scope of this report. 27 It is worth noting that the four domains follow a sequence. The idea of a sequence between the first two subject knowledge-based domains and the second two pedagogical content knowledge-based domains is important for its implication for teacher development of error analysis. It suggests that the first three domains, when contextualized in approaches to teaching and instruction, equip teachers to design their teaching environment. It is only when teachers have learnt to understand patterns of error, to evaluate non-standard solutions, or to unpack a procedure, that they will be able to anticipate errors in their teaching or in their assessment, and prepare for these in advance. Research (Heritage et al, 2009) suggests that the move from error analysis to lesson planning is very difficult for teachers. The sequence of the domains in this model suggests the importance of knowledge about error analysis for developing mathematical knowledge for teaching. 28 Section two: Activity and Process 29 2.1. The Activity DIPIP engaged the teachers in six activities. These can be divided into two types: errorfocused activities and error-related activities. Error-focused activities engaged the teachers directly in error analysis. Error-related activities engaged teachers in activities that built on the error analysis but were focused on learning and teaching more broadly. Table 2 “Error-focused activities” and “error-related activities” Error-focused activities Error-related activities Analysis of learner results on the ICAS mathematics tests (with a focus on the multiple choice questions and the possible reasons behind the errors that led to learner choices of the distractors)15 Analysis of learners’ errors on tests that were designed by the teachers An interview with one learner to probe his/her mathematical reasoning in relation to errors made in the test. Mapping of ICAS test items in relation to the South African mathematics curriculum; Development of lesson plans which engaged with learners’ errors in relation to two mathematical concepts (equal sign; visualisation and problem solving) Teaching the lesson/s and reflecting on one’s teaching in "small grade-level groups” and presenting it to “large groups16 In the course of Phases 1 and 2 of the DIPIP project, the error analysis activity followed after the curriculum mapping activity17 and structured a context for further professional conversations among teachers about the ICAS and other test data. Teachers used the tests to analyse the correct answers and also the errors embedded in the distractors presented in the multiple choice options and the incorrect solutions given by learners on the open tests. “Distractors” are the three or four incorrect answers in multiple choice test items. They are designed to be close enough to the correct answer to ‘distract’ the person answering the question. 15 The small groups consisted of a group leader (a mathematics specialist – Wits School of Education staff member or post graduate student who could contribute knowledge from outside the workplace), a Gauteng Department of Education (GDE) mathematics subject facilitator/advisor and two or three mathematics teachers (from the same grade but from different schools). This meant that the groups were structured to include different authorities and different kinds of knowledge bases. These were called small grade-level groups (or groups). As professional learning communities, the groups worked together for a long period of time (weekly meetings during term time at the Wits Education Campus for up to three years), sharing ideas and learning from each other and exposing their practice to each other. In these close knit communities, teachers worked collaboratively on curriculum mapping, error analysis, lesson and interview planning, test setting and reflection. For certain tasks (such as presenting lesson plans, video clips of lessons taught or video clips of learner interviews) the groups were asked to present to large groups. A large group consisted of the grade-level groups coming together into larger combined groups, each consisting of four to six small groups (henceforth the large groups). This further expanded the opportunities for learning across traditional boundaries. (See Shalem, Sapire, Welch, Bialobrzeska, & Hellman, 2011, pp.5-6) 16 17 See the full Curriculum Mapping report, Shalem, Y., & Sapire, I. (2011) 30 2.2 The ICAS test To provide a basis for systematic analysis of learners’ errors, the results of Gauteng learners on an international, standardized, multiple-choice test, the ICAS test, were used. The International Competitions and Assessments for Schools (ICAS) test is conducted by Educational Assessment Australia (EAA), University of New South Wales (UNSW) Global Pty Limited. Students from over 20 countries in Asia, Africa, Europe, the Pacific and the USA participate in ICAS each year. EAA produces ICAS papers that test students in a range of subject areas including Mathematics. Certain schools in Gauteng province, both public and private schools, used the ICAS tests in 2006, 2007 and 2008, and it was the results of these learners on the 2006 and 2007 tests that provided the starting point for teacher engagement with learners’ errors. The ICAS test includes multiple choice and some open items. The Grade 3-6 tests consist of 40 multiple choice questions and the Grade 6-11 tests consist of 35 multiple choice items and 5 open questions. 2.3 Time line The first round of the error analysis activity ran for ten weeks, from July 2008 to October 2008. In this round, which we refer to as Round 1, the teachers analysed data from the ICAS 2006 tests. As in the curriculum mapping activity, the groups worked on either the even or odd numbered items (the same items for which they had completed the curriculum mapping activity). The second round of error analysis ran for five weeks in September and October 2010. In this round, which we refer to as Round 2, the teachers in their groups analysed data from the ICAS 2007 tests as well as tests which they had set themselves. The number of items analysed in Round 2 had to be cut down because of time limits. 2.4 The Process Round 1 error analysis gave the teachers the opportunity in their groups to discuss the mathematical reasoning that is required to select the correct option in the ICAS 2006 multiple choice test items, as well as to provide explanations for learners’ choices of each of the distractors (incorrect answers). In order to deepen the teachers’ conceptual understanding and appreciation of learners’ errors, the groups had to provide several explanations for the choices learners made. This was intended to develop a more differentiated understanding of reasons underlying learners’ errors amongst the teachers. The focus of the error analysis in this Round was on the multiple choice items because the test designers provided a statistical analysis of learner responses for these, which served as the starting point for teachers’ analysis. Some groups also discussed 31 the open ended items. Groups had to analyse either odd or even numbered items, as they had done in the curriculum mapping activity. Some groups analysed more items than others, but all completed the analysis of the 20 odd/even numbered items they were assigned. Table 3 below18, lists the material given to the groups, the error analysis task and the number and types of groups (small or large or both) in which the teachers worked in Round 1 error analysis. Table 3: Round 1 Error analysis Material Tasks Group type Exemplar templates (See Appendix 1) Group leader training: Group leaders with project team Guidelines for the completion of the error analysis template Discussion of error analysis of selected items to generate completed template for the groups to work with. ICAS achievement results for Gauteng Grading items by difficulty level. Our method for grading item difficulty was as follows: The achievement stats for learners, who wrote the ICAS tests in the Gauteng sample were entered onto excel sheets and items were graded from the least to the most difficult, based on this data. The items were thus all assigned a “difficulty level” from 1 to 40. 19 Project team Test items and answer key with learner performance data (per item) from the 2006 ICAS test. Error analysis of learner performance using template: 14 small grade-specific groups Statistical analysis of learner performance for correct answer. Identification of the mathematical reasoning behind the correct response Provision of possible explanations for learners’ choice of each of the three distracters. 3 distractors Error analysis template. (see 18 For more detail see Shalem, Sapire, Welch, Bialobrzeska, & Hellman, 2011 Out initial plan was to only work on a selected number of tests items (ICAS 2006). We were planning to choose the items that proved to be most difficult for the learners who wrote the test at the time (55,000 learners in Gauteng from both public and independent schools). At the time of the start of the project (last quarter of 2007), the EAA’s Rasch analysis of each of the tests items was not available. 19 32 Material Tasks Group type Appendix 2) In Round 2 Error analysis eleven20 of the groups repeated the error analysis and curriculum mapping activities on selected ICAS 2007 items or on their own tests21, but without the facilitation of the group leader. This was intended to establish how effectively the experiences of Round 1 error analysis and other project activities had enabled the teachers, in their groups, to work independently. Table 4 below22, lists the material given to the groups, the error analysis task and the number and types of group in which the teachers worked in Round 2 error analysis. Table 4: Round 2 Curriculum Mapping and Error Analysis Material Tasks Group type ICAS 2007 test items. Error analysis of ICAS 2007 tests. Error analysis template with curriculum alignment built into the template The groups were assigned 12 items for analysis. Not all the small groups completed the analysis of these items. The small groups conducted the same analysis as they did for the ICAS 2006 test items. 6 groups (One of each of Grades 3, 4, 5, 6, 7, and 9) Error analysis template including curriculum mapping, which was modified in order to accommodate analysis of “own tests” (see Appendix 3)23 Error analysis of “own tests”. As with the ICAS, the groups were asked 5 groups on own test items to pool together the results of all the learners in their class and to work out the achievement statistics on the test (One of each of Grades 3, 4, 5, 6 and 8) to place the results on a scale and on that basis judge the difficulty of each item to analyse the ways in which the learners got the correct answer to each of the questions as well as the ways the learners got each of the items wrong. Round 2 error analysis was completed by 11 of the 14 teacher groups since three small groups did a third round of teaching in the last phase of the project. 20 21 See Shalem, Sapire, Welch, Bialobrzeska, & Hellman, 2011 on the process and the design of “own tests” 22 For more detail see Shalem, Sapire, Welch, Bialobrzeska, & Hellman, 2011. Due to time constraints, as well as to the fact that a further round of teaching was not going to follow this exercise, the revised template did not include a question on whether and how teachers taught the concept underlying the test item. 23 33 Material Tasks Group type Verbal instructions to groups Error analysis presentations: Three Large groups All groups were requested to present the error analysis of one item to the larger groups. Group leaders were invited to attend the presentation. Grades 3 and 4, Grades 5 and 6 and Grades 7 and 9. 2.5 Group leader training A plenary meeting with group leaders for the error analysis mapping activity was held prior to the commencement of the small group sessions. The session was led by the project leader, Prof Karin Brodie, with input taken from all group leaders. When the error analysis activity was presented to the groups, there was a briefing session where all groups sat together, before they spilt into their smaller groups and started on the activity. Project coordinators moved around and answered questions while groups worked. The groups were encouraged to complete templates in full and settled queries where possible. When the first templates were returned, groups were given feedback and asked to add/amend their templates as necessary – to try for even better completion of templates. Some of the group discussions were recorded24. After the activity was completed, we conducted two large group de-briefing discussions (one with the Grade 3, 4, 5 and 6 groups and one with the Grades 7, 8 and 9 groups). These debriefing sessions were recorded. This report is based on the written templates, not on the recordings. 2.6 Item analysis The error analysis component of the DIPIP project resulted in a large number of completed templates in which groups had recorded their analysis of the items. This involved analysis of both correct and incorrect solutions. Table 5 summarises the total number of items analysed by the groups during Round 1 and 2. Groups are called “odd” and “even” according to the test item numbers assigned to them for initial curriculum mapping in Round 1 (see Column Two). Column Three details the number of ICAS 2006 items analysed by the groups in Round 1. Column Four details the number of ICAS 2007 / Own tests analysed by the groups in Round 2. 24 Three groups were recorded in Round 1 and eleven groups were recorded in Round 2. 34 Table 5: Number of items analysed by groups (Rounds 1 and 2) Grade Numbers Round 1 Round 2 (ICAS 2006 test) (ICAS 2007 test/Own test) 3 Odd 17 Own test (6) 3 Even 16 ICAS (10) 4 Odd 20 Own test (5) 4 Even 20 ICAS (11) 5 Odd 20 Own test (6) 5 Even 20 ICAS (4) 6 Odd 20 Own test (4) 6 Even 20 ICAS (11) 7 Odd 20 Not done 7 Even 20 ICAS (8) 8 Odd 11 Own test (6) 8 Even 13 Not done 9 Odd 16 ICAS (11) 9 Even 17 Not done 250 items 82 items Total Round 1 error analysis produced a much bigger set of data (250 items) than Round 2 (82 items). In Round 1 certain groups analysed more items than others. The expectation for error analysis of items in Round 1 was a minimum of 20 items per group. No group completed more than 20 ICAS 2006 items, although some groups assisted other groups with editing and improving their analysis when they had completed the 20 items allocated to their group. Grade 6 – 9 ICAS tests include open ended questions. Some of the groups chose not to analyse these items. This can be seen in the Grade 8 and 9 groups that mapped less than 20 items. Neither of the Grade 3 groups completed the mapping of all 20 even/odd items. The expectation for error analysis of items in Round 2 was a minimum of 12 ICAS 2007 items and 6 “own test” items per group. No group completed all the 12 ICAS 2007 items assigned to it. Three of the six groups completed 11 items. Three of the six groups that analysed their own test completed all the 6 questions. 35 In summary, in Round 2 only 11 of the 14 groups participated and they analysed fewer items due to time constraints. Some groups in Round 2 analysed the ICAS 2007 test while others analysed tests that they had set themselves. The membership of most of the groups that continued across two rounds of error analysis remained. The format of the error analysis was the same for both rounds. Groups received feedback in both rounds. The main difference between rounds was that group leaders were not present in Round 1. In between the two rounds all of the groups were involved in lesson design, teaching and reflection where learners’ errors played a significant role in planning, teaching and reflecting on teaching. It was intended that in coming back to error analysis in Round 2 the groups would be more familiar with the idea of error analysis both in teaching and related activities. In view of the differences between the two rounds comparisons between rounds are possible, although they should be taken tentatively. The comparison does imply any inferences about the impact of Round 1 activities on those of Round 2. 36 Section three: Evaluation analysis methodology 37 3.1 Aim of the evaluation We aim to evaluate the quality of error analysis done by the groups. In this we aim to evaluate the quality of teachers reasoning about error, evident in the group texts (both in the “correct analysis texts” and in the “error analysis texts”) and the value of the group leaders in leading the process. The empirical analysis intends to answer the following four empirical questions: On what criteria is the groups’ error analysis weaker and on what criteria is the groups’ error analysis stronger? In which mathematical content areas do the groups produce better judgements on the error? Is there a difference in the above between primary (Grade 3-6 groups) and high school (Grade 7-9 groups) teachers? What does the change in performance between Round 1 and 2 suggest about the role of group leaders? 3.2 Items evaluated For the purpose of analysis, we selected a sample of ten ICAS 2006 items per grade per group for Round 1. These ten items (per group) were made up of five items from the first half of the test and five items from the second half of the test. Since fewer items were mapped in Round 2 mapping activity, all the items mapped by groups that participated in Round 2 were analysed. Only one distractor (or incorrect answer) per item was selected for the evaluation analysis. The selected distractors were the ones that were most highly selected according to the test data, in other words, the “most popular” incorrect learner choice. Table 6 below summarises the sample of data used in the evaluation analysis of Rounds 1 and 2 error analyses: Table 6: Sample summary Full set of data to be considered Sample Round 1: ICAS 2006 items: 10 items selected per group, hence 20 items selected per grade. The selected items were chosen to represent the curriculum. Completed error analysis templates based on ICAS 2006 test data (Grades 3-9, 14 groups) Groups: All groups Grades 3-9 (14), 140 items in sample. All texts related to the correct answer and the chosen distractor of the selected items formed the data which was analysed – from Round 1 there were 572 texts coded (316 answer texts and 252 error texts). 38 Full set of data to be considered Sample Round 2: ICAS 2007: Some groups Grades 3-9 (6), mapped items selected per group, various item numbers coded per grade. Completed error analysis templates based on ICAS 2007 test as well as data from tests which groups had developed “own test” data “Own tests”: Some groups Grades 3-9 (5), mapped items from tests that they had set as groups for the learner interview activity, various item numbers coded per grade. 82 items in sample. (Grades 3-9, 11 groups) All texts related to the correct answer and the chosen distractor of the selected items formed the data which was analysed – from Round 2 there were 284 texts coded (173 answer texts and 111 error texts). The sample size in Round 2 was smaller than the sample size in Round 1. In both rounds the number of answer texts was higher than the number of error texts. 3.3 Evaluation Instrument The initial evaluation criteria and their corresponding categories were drawn up into a template by three members of the project team. The template was then used to code the full data sample, piecemeal, while on-going discussions were held to refine the criteria over a four month period. Agreement was reached after the coding of four Round 1 grades was completed and the final format of the template and wording of the criteria and categories was constructed. See more on the template below. Two external coders, one expert in the field of maths teacher education and one an experienced mathematician with experience in teacher education, coded the groups’ texts. The coders were given group’s texts; to code. A group’s text is divided to two texts types: “Correct answer texts” or the explanation provided by the group for the correct answer. “Error answer texts” or the explanations provided by the group for the distractor selected for the evaluation. 39 Groups provided more than one explanation for the correct answer or for the selected distractor for some items. This meant that some items had one correct answer text and one error text per the distractor while other items may have had several correct answer texts and/or several error for the distractor. The coders were given all of the texts written by the groups in explanation of the correct answer and of the selected distractor (error). The texts were arranged for the coders for each item in a large excel coding sheet so that all coders worked according to the same identification of texts. Coders were asked to give a code to each text. In this way a code was assigned to each of the texts for each of the items selected for the evaluation. 3.3.1 The coding template (see Appendix 4): A coding template was prepared for the coders so that all coding related to the same texts, and all the texts are arranged in the same format. The coding template included general information about the group text, six columns of criteria and their corresponding categories, and a final comment column. One template per grade was prepared, with all the texts inserted into the template next to the relevant item. The coding template for the ICAS test items consists of all the items and all the texts (correct answer and error answer texts) associated with the item. The coding template for “own test” was the same as that for the ICAS tests. The only difference is that for the “own tests” template the actual incorrect answer to be discussed was pasted into the “Distractor” column since it could not be represented by a key. Instrument overview General information on group texts: Grade – indicates the grade of the item texts. Item – indicates the item number for which the groups produced the correct answer and the error answer texts. Maths Content area - indicates the curriculum area (number, pattern/algebra, space, measurement or data handling) of the item. Distractor – indicates which distractor had been selected for analysis. Text number – indicates the number of the text to be coded. Numbering started from one for each correct answer text and then from one again for each error text. Text – refers to the actual text as recorded by the groups in their completed templates. The texts were pasted into the coding template, verbatim. Texts that could not be pasted into the coding template were sent to coders in a separate document attachment and referred to in the templates by grade and number. 40 Criterion code columns: In addition to the above information, the key aspect of the template is the criteria and their relevant categories. The templates consist of six columns, which are related to the six evaluation criteria selected. Each criterion is further divided into four categories (“not present”, “inaccurate”, “partial” and “full”) that capture the quality of the correct and the error answer text, in terms of teachers’ reasoning about the correct answer and the distractor (or error in “own texts”). For purpose of simplicity each category was coded with a number (4 assigned to “full” and 1 to “not present”). Coders had to insert their coding decisions for each of the four categories into each column, for each identified text. Procedural explanations – a code of 1 to 4 assigned to each text, which correspond to the four categories. This criterion measures the fullness and the correctness of the groups’ description of the procedure to be followed in order to answer the question. Conceptual explanations – a code of 1 to 4 assigned to each text, which correspond to the four categories. This criterion measures the fullness and the correctness of the groups’ description of the conceptual links that should be made in order to answer the question. Awareness of error – a code of 1 to 4 assigned to each text, which correspond to the four categories. This criterion measures general knowledge of the mathematical error. Diagnostic reasoning – a code of 1 to 4 assigned to each text, which correspond to the four categories. This criterion measures groups’ explanation of learner’ reasoning behind the error. Use of everyday – a code of 1 to 4 assigned to each text, which correspond to the four categories. This criterion measure groups use of everyday knowledge about learner’ reasoning. Multiple explanations – a code of “n” or “f” assigned25 to each text from which codes of 1 to 4 could be assigned to “multiple explanations” for each text. This criterion measure the feasibility of the alternative explanations the groups offered to explain leaners reasoning behind the error. Comment – The final comment allowed coders to write any particular comments they wished to note about the answer or error text in that row. These comments could be used in the alignment discussion and evaluation. The coders inserted their codes according to the given column heading for criterion coding. A wide range of exemplars are provided in the appendices in order to demonstrate the operationalization of the criteria (see Appendices 6 – 11.). Exemplar items were chosen from Round 1 and Round 2 and in such a way that a spread of grades “n” was assigned to a particular text when it was considered not mathematically feasible in relation to the error under discussion and “f “was assigned when the explanation was considered mathematically feasible. The final code for “multiple explanations” was decided on the number of feasible explanations given of the chosen distractor. 25 41 is represented with both strong and weak explanations. Items were also selected so that there are exemplars across all five of the mathematical content areas represented in the curriculum. The criteria Whilst Section One offers explanations for the choice and the meaning of the criteria, we offer a brief explanation in this section too. A wide range of exemplars are provided in the appendices in order to demonstrate the operationalization of the criteria. Exemplar items were chosen from Round 1 and Round 2 and in such a way that a spread of grades is represented with both strong and weak explanations. Items were also selected so that there are exemplars across all five of the mathematical content areas represented in the curriculum. In what follows we provide a brief description of each criterion and the Appendix number which includes the full wording of each criterion and its corresponding four categories. For the full set of the six criteria and their categories see Appendix 5 Answer Text Criteria The first two criteria were used to code the groups’ explanations of the correct answers. Every “correct answer text” was thus coded according to two criteria: procedural and conceptual. Procedural explanations The literature emphasises that the quality of teachers’ explanations depends on the balance they achieve between explaining the procedure required for addressing a mathematical question and the mathematical concepts underlying the procedure. This criterion aims to grade the quality of the teachers’ procedural explanations of the correct answer. The emphasis in the criterion is on the quality of the teachers’ procedural explanations when discussing the solution to a mathematical problem through engaging with learner test data. Teaching mathematics involves a great deal of procedural explanation which should be done fully and accurately for the learners to grasp and become competent in working with the procedures themselves (see Appendix 6). Conceptual explanations The emphasis in this criterion is on the conceptual links made by the teachers in their explanations of the learners’ mathematical reasoning in relation to the correct answer. Mathematical procedures need to be unpacked and linked to the concepts to which they relate in order for learners to understand the mathematics embedded in the procedure. The emphasis of the criterion is on the quality of the teachers’ conceptual links made in 42 their explanations when discussing the solution to a mathematical problem through engaging with learner test data (see Appendix 7). Error Text Criteria The next four criteria were used to code the groups’ explanations of the distractor. These explanations are about teachers’ engagement with the errors. Four criteria were selected: “awareness of error”, “diagnostic reasoning”, “use of everyday knowledge” and “multiple explanations”. Every “error answer text” was coded according to these four criteria. Awareness of error The emphasis in this criterion is on the teachers’ explanations of the actual mathematical error (and not on the learners’ reasoning). The emphasis in the criterion is on the mathematical quality of the teachers’ explanations of the actual mathematical error when discussing the solution to a mathematical problem (see Appendix 8). Diagnostic Reasoning The idea of error analysis goes beyond identifying the actual mathematical error. The idea is to understand how teachers go beyond the mathematical error and follow the way learners were reasoning when they made the error. The emphasis in the criterion is on the quality of the teachers’ attempt to provide a rationale for how learners were reasoning mathematically when they chose a distractor (see Appendix 9). Use of the everyday Teachers often explain why learners make mathematical errors by appealing to everyday experiences that learners draw on and may confuse with the mathematical context of the question. The emphasis in this criterion is on the quality of the use of everyday knowledge, judged by the links made to the mathematical understanding that the teachers attempt to advance (see Appendix 10). Multiple explanations One of the challenges in the teaching of mathematics is that learners need to hear more than one explanation of the error. This is because some explanations are more accurate or more accessible than others and errors need to be explained in different ways for different learners. This criterion examines the teachers’ ability to offer alternative explanations of the error when they are engaging with learners’ errors through analysis of learner test data (see Appendix 11). 43 3.4 Training of coders and coding process Coders started by coding one full set of Grade 3 texts, after an initial discussion of the coding criteria and template. The evaluation analysis team discussed the assigned Grade 3 codes with the coders. After this and more general discussions about the format and the criterion, the final format of the coding template was agreed on. This format was to include all texts pasted into the template, so that there could be no confusion as to which part of the teachers’ explanations the coders were referring when they allocated codes. Discussion meetings with the evaluation analysis team were held after each set of coding (per grade) was completed. The discussions allowed for refinement of the wording of the criteria – and when refinements were made, coders reviewed their completed grades in order to be sure that all coding was in line with the criteria. Coding continued per grade, with discussion after the completion of each grade for Grades 4, 5 and 6. After each full set of codes for a grade was completed by both coders, the percentages of alignment between coders was calculated and sent to coders. The alignment differences were used to guide the above discussions and the evaluation coordinator gave comments as to which areas the coders should focus on and think more carefully about in order to improve on alignment. After the fourth set of codes was completed and discussed it was decided that there was sufficient agreement between the two coders for them to continue with the remaining sets of codes and the process set was in motion. All remaining templates were prepared and emailed to coders. 3.5 Validity and reliability check The full set of codes was completed and then reviewed by the evaluation analysis team. Consensus discussions between the coders were held on certain items in order to hone agreement between them. The final set of codes used in the analysis was agreed on in discussion with and through arbitration by a third expert (a member of the evaluation analysis team). The alignment between coders was 57% before the review, 71% after the review and 100% after the arbitration. 3.6 Data analysis The coded data were analysed quantitatively, finding observable trends and relationships evident in the sample. The content of the templates was entered into excel spread sheets to facilitate the overall analysis of the findings and comparison of the various criteria. Data were summarised for descriptive analysis and t-tests were done to 44 establish whether there were any significant changes between the error analysis in Rounds 1 and 2. An unpaired t-test assuming an unequal variance (since the sample sizes were different) was used to test for significance. This was considered to be the most conservative measure and appropriate for use for our data set. Correlations between codes for the answer texts (procedural and conceptual explanations) and error texts (awareness of error and diagnostic reasoning) were calculated using Pearson’s r coefficient. The analysis follows three levels: Level one: For each of the six criteria, we grouped the data into two sets of groupedgrades (Grade 3-6 and Grade 7-9). This enabled us to discern patterns in performance across the 14 small groups, between primary and secondary teachers (see Section four). For each of the six criteria we provide the following: Graphs and tables comparing the findings for two sets of grouped grades (Grade 3-6 and Grade 7-9 groups) between Rounds 1 and 2. Graphical representations of the overall findings across all groups. Graphs representing the findings for mathematical content areas. Level two: We generate broad findings across two or more criteria and about the total 14 groups analysed (see Section five). These findings enabled us to see patterns in teacher knowledge of error analysis, divided along subject matter and pedagogical content knowledge. This type of analysis also enabled us to compare the overall performance across the criteria between the two sets of grouped grades (Grade 3-6 and Grade 7-9). Lastly the analysis enables us to show the importance of guided instruction by group leaders, for a successful process of accountability conversations, albeit with some differentiation across the criteria. With these findings we are able to then construct a broad description of what is involved in teachers “diagnostic judgment”. Level three: We draw on the finding to generate a broad description (working towards a conceptual definition) of what we mean by “diagnostic judgment” in teachers reasoning about error (see Section five). 45 46 Section Four: Analysis of data 47 4.1 An overview of the comparison of Round 1 and Round 2 An overview of the comparison of Round 1 and Round 2 shows immediately that the groups worked to a higher level in Round 1 when they had leaders. This overall impression of quality differences between Rounds 1 and 2 shows a drop in the percentage of texts demonstrating “full” explanations on every criterion between Round 1 and 2 and an increase in the percentage of texts demonstrating “not present” explanations on every criterion between Round 1 and 2. Figure 1: Round 1 – overall codes assigned on all six criteria Round 1 Number Spread of Codes 100 90 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Figure 2: Round 2 – overall codes assigned on all six criteria Round 2 Overall Spread of codes 100 90 80 70 60 50 40 30 20 10 0 Not present Inaccurate Subject Matter Knowledge Multiple Everyday Diagnostic Awareness Conceptual Procedural Partial Full Pedagogical Content Knowledge 48 The correlation between procedural explanations and conceptual explanations was high in both rounds, although it decreased in Round 2 (r = 0,74 in Round 1 and r = 0,66 in Round 2). The correlation between awareness of the mathematical error and diagnostic reasoning was also high and increased in Round 2 (r = 0,651 in Round 1 and r = 0,71 in Round 2)26. The downwards trend and the increases and decreases in texts demonstrating different levels of the criteria will now be analysed individually for each of the criteria. 4.2. Analysis of individual criteria We now discuss the criteria separately, focusing on grouped grades (3-6 and 7-9) and mathematical content for each of the six criteria. We first note our observation separately for each of the grouped grades, and then for the all of the 14 groups together. 4.2.1 Procedural explanations Figure 3: Procedural explanations of answers – Round 1 and 2 by grouped grades Procedural explanations Round 1 and Round 2 by grade 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% Grade 3-6 Grade 7-9 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 Procedural explanation not given Procedural explanation inaccurate or incomplete Procedural Procedural explanation correct explanation but missing some accurate and full steps Since these correlations were high, exemplars of explanations which received the same level coding are given in Appendices 6, 7, 8 and 9, in addition to other exemplar texts. 26 49 Analysis of procedural explanations in Grade 3-6 small groups’ texts: Table 7: Grades 3-6 procedural explanations demonstrated in teacher text explanations Change Round Round Strength of Procedural Explanations between Round 1 Round 2 1 2 explanation rounds Procedural explanation not given 3% 17% 14%*** Procedural explanation inaccurate or incomplete 8% 2% -6% Procedural explanation correct but missing some steps 46% 61% 15%* Procedural explanation accurate and full 43% 20% Weaker 11% 19% Stronger 89% 81% -23%*** *** Difference significant at a 99% level of confidence * Difference significant at a 90% level of confidence Observations about groups’ procedural explanations in relation to correct answers in Rounds 1 and 2 for the grade 3-6 groups: 1. The number of texts of the Grade 3-6 group that demonstrate weak procedural explanations increased slightly between Rounds 1 and 2 (from 11% in Round 1 to 19% in Round 2). 2. The number of texts where no attempt was made to give procedural explanations increased by 14% from 3% in Round 1 to 17% of texts in Round 2. 3. The number of texts where procedural explanations were inaccurate or incomplete decreased by 6% (from 8% in Round 1 to 2% in Round 2). 4. The number of texts of the Grade 3-6 group that demonstrate strong procedural explanations slightly decreased by 8% in Round 2 (from 89% in Round 1 to 81% in Round 2). 5. The gap between incomplete and systematic explanations was not very high in Round 1 but it grew much wider in Round 2 (from 3% in Round 1 to 41% in Round 2). 50 Analysis of procedural explanations in Grade 7-9 small groups’ texts: Table 8: Grades 7-9 procedural explanations demonstrated in teacher text explanations Change Round Round Strength of Procedural Explanations between Round 1 Round 2 1 2 explanation rounds Procedural explanation not given 3% 9% 6% Procedural explanation inaccurate or incomplete 8% 10% 2% Procedural explanation correct but missing some steps 43% 56% 13%** Procedural explanation accurate and full 46% 26% Weaker 11% 19% Stronger 89% 81%27 20%*** *** Difference significant at a 99% level of confidence ** Difference significant at a 95% level of confidence Observations about groups’ procedural explanations in relation to correct answers in Rounds 1 and 2 for the grade 7-9 groups: 1. The number of texts of the grade 7-9 group that demonstrate weak procedural explanations slightly increased (from 11% in Round 1 to 19% in Round 2). 2. The number of texts where no attempt was made to give procedural explanations increased by 6% (from 3% in Round 1 to 9% in Round 2). 3. The number of texts where procedural explanations were inaccurate or incomplete increased by 2% (from 8% in Round 1 to 10% in Round 2). 4. The number of texts of the Grade 7-9 group that demonstrate strong procedural explanations decreased by 7% in Round 2 (from 89% in Round 1 to 82% in Round 2). 5. The gap between incomplete and systematic explanations was not very high in Round 1 (3%) but grew wider in Round 2 (31%). Rounding of Round 1 and 2 percentages is done individually but Round totals are added and then rounded hence there may be small discrepancies such as the one highlighted in this table. This occurs in certain table in this report but is only noted here, the first time such a discrepancy arises. 27 51 Overall findings on the procedural explanations of the correct answer – Rounds 1 and 2 Figure 4: Procedural explanations of answers – Round 1 and 2 Procedural explanations - Overall - Round 1 and 2 100% 90% 80% 70% 56% 60% 50% 46% 44% 40% 30% 23% 16% 20% 10% 3% 8% 5% Round 1 Round 2 0% Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 Procedural explanation Procedural explanation Procedural explanation Procedural explanation not given inaccurate or correct but missing accurate and full incomplete some steps Table 9: Procedural explanations demonstrated in teacher test explanations Change Round Round Strength of Procedural Explanations between Round 1 1 2 explanation rounds Procedural explanation not given 3%* 16%* 13% Procedural explanation inaccurate or incomplete 8% 5% -3% Procedural explanation correct but missing some steps 44% 57% 13%*** Procedural explanation accurate and full 46% 23% Round 2 Weaker 11% 20% Stronger 89% 80% -16%*** *** Difference significant at a 99% level of confidence Observations about groups’ procedural explanations in relation to correct answers in Rounds 1 and 2: 52 1. The overall number of texts that demonstrate weak procedural explanations increased slightly between Rounds 1 and 2 (from 11% in Round 1 to 20% in Round 2). 2. The number of texts where no attempt was made to give procedural explanations increased by 13% from 3% in Round 1 to 16% of texts in Round 2. This increase was significant at 99%. 3. The number of texts where procedural explanations were inaccurate or incomplete decreased by 3% (from 8% in Round 1 to 5% in Round 2). This decrease was not significant. 4. The overall number of texts that demonstrate strong procedural explanations slightly decreased by 9% in Round 2 (from 89% in Round 1 to 80% in Round 2). 5. The gap between incomplete and systematic explanations was not very high in Round 1 but it grew much wider in Round 2 (from 2% in Round 1 to 34% in Round 2). 6. The decrease between Round 1 and Round 2 in the number of texts with full or partially complete explanations was significant at 99%. Procedural explanations, by mathematical content Number was the content area in which the groups’ procedural explanations of the correct answer were strongest. Data was the content area in which the groups’ procedural explanations of the correct answer were the weakest. The graphs below represent the percentages of procedural explanations for these two content areas across the two rounds. Figure 5: Round 1 and 2 procedural explanations of answers – content area number Number Procedural explanations Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 2.272727273 7.954545455 29.54545455 60.22727273 Round 2 10.9375 1.5625 51.5625 35.9375 53 Change in the strong content area: The number of texts that demonstrate full and accurate procedural explanations decreased from 60.23% in Round 1 to 35.94% in Round 2. The number of texts that demonstrate partial procedural explanations (correct but, missing some steps) increased from 29.55% in Round 1 to 51.56% in Round 2. The number of texts in which procedural explanations were inaccurate or incomplete decreased from 7.95% to 1.56% in Round 2. The number of texts where procedural explanations were not given increased from 2.27% in Round 1 to 10.94% in Round 2. Figure 6: Round 1 and 2 procedural explanations of answers – content area data Data Procedural explanations Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 5.882352941 1.960784314 47.05882353 45.09803922 Round 2 25 12.5 56.25 6.25 Change in the weak content area: The number of texts that demonstrate full and accurate procedural explanations decreased from 45.10% in Round 1 to 6.25% in Round 2. The number of texts that demonstrate partial procedural explanations (correct but, missing some steps) increased from 47.06% in Round 1 to 56.25% in Round 2. The number of texts that procedural explanation inaccurate or incomplete increased from 1.96% to 12.5% in Round 2. The number of texts that did not include procedural explanations increased from 5.88% in Round 1 to 25% in Round 2. 54 In comparison, the trends between the two rounds in these two content areas are similar, bar the decrease in Round 2 in the number of texts with inaccurate explanation in the number content area. 4.2.2 Conceptual explanations Figure 7: Conceptual explanations in explanations – Round 1 and 2 by grouped grades Conceptual Explanations - Round 1 and Round 2 by grade 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% Grade 3-6 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 No conceptual links Explanation in explanation includes poorly conceived conceptual links Explanation includes some but not all conceptual links Grade 7-9 Explanation includes conceptual links that explain process and background Analysis of conceptual explanations in Grade 3-6 small groups’ texts: Table 10: Grades 3-6 conceptual explanations demonstrated in teacher text explanations Change Round Round Strength of Round Conceptual explanations between Round 2 1 2 explanation 1 rounds No conceptual links in explanation 4% 30% 26%* Explanation includes poorly conceived conceptual links 19% 10% -9% Explanation includes some but not all conceptual links 39% 49% 10% Explanation includes conceptual links that explain process and background 38% 11% Weaker 23% 40% Stronger 77% 60% -27%** 55 *** Difference significant at a 99% level of confidence * Difference significant at a 90% level of confidence Observations about groups’ conceptual explanations in relation to correct answers in Rounds 1 and 2 for the grade 3-6 groups: 1. The number of texts of the Grade 3-6 group that demonstrate weak conceptual explanations increased by 17% between Rounds 1 and 2 (from 23% in Round 1 to 40% in Round 2). 2. The number of texts where no conceptual links were evident increased by 26% from 4% in Round 1 to 30% of texts in Round 2. 3. The number of texts where poorly conceived conceptual links were made decreased by 9% (from 19% in Round 1 to 10% in Round 2). 4. The number of texts of the Grade 3-6 group that demonstrate strong conceptual explanations decreased by 17% in Round 2 (from 77% in Round 1 to 60% in Round 2). 5. The gap between explanations that include conceptual links that explain the background and process of the answer and explanations that include some but not all of the conceptual links grew wider in Round 2 (from 1% in Round 1 to 39% in round 2). Analysis of conceptual explanations in Grade 7-9 small groups’ texts: Table 11: Grades 7-9 conceptual explanations demonstrated in teacher text explanations Change Round Round Strength of Conceptual explanations between Round 1 Round 2 1 2 explanation rounds No conceptual links in explanation 6% 33% 27%*** Explanation includes poorly conceived conceptual links 15% 4% -11%** Explanation includes some but not all conceptual links 40% 34% -6% Explanation includes conceptual links that explain process and background 38% 29% Weaker 21% 37% Stronger 79% 63% -9% *** Difference significant at a 99% level of confidence ** Difference significant at a 95% level of confidence 56 Observations about groups’ conceptual explanations in relation to correct answers in Rounds 1 and 2 for the grade 7-9 groups: 1. The number of texts of the Grade 7-9 group that demonstrate weak conceptual explanations increased (from 21% in Round 1 to 37% in Round 2). 2. The number of texts where no conceptual links were evident increased by 27% from 6% to 33% of texts. 3. The number of texts where poorly conceived conceptual links were made decreased by 9% (from 15% in Round 1 to 4% in Round 2). 4. The number of texts of the Grade 7-9 group that demonstrate strong conceptual explanations decreased by 15% in Round 2 (from 78% in Round 1 to 63% in Round 2). 5. The gap between explanations that include conceptual links that explain the background and process of the answer and explanations that include some but not all of the conceptual links was not at all high in Round 1 (2%) and 5% in Round 2. Overall findings on the conceptual explanations of the correct answer – Rounds 1 and 2 Figure 8: Conceptual links in explanations of answers -Round 1 and 2 Conceptual explanations - Overall - Round 1 and 2 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% 39% 45% 40% 29% 17% 5% Round 1 Round 2 Round 1 16% 10% Round 2 No conceptual links in Explanation includes explanation poorly conceived conceptual links Round 1 Round 2 Explanation includes some but not all conceptual links Round 1 Round 2 Explanation includes conceptual links that explain process and background 57 Table 12: Conceptual explanations demonstrated in teacher test explanations Change Round Round Strength of Conceptual explanations between Round 1 1 2 explanation rounds No conceptual links in explanation 5% 29% 24%*** Explanation includes poorly conceived conceptual links 17% 10% 7%** Explanation includes some but not all conceptual links 39% 45% 6% Explanation includes conceptual links that explain process and background 40% 16% Round 2 Weaker 22% 39% Stronger 78% 61% -14%*** *** Difference significant at a 99% level of confidence ** Difference significant at a 95% level of confidence Observations about groups’ conceptual explanations in relation to correct answers in Rounds 1 and 2: 1. The overall number of texts that demonstrate weak conceptual explanations increased between Rounds 1 and 2 (from 22% in Round 1 to 39% in Round 2). 2. The number of texts where no attempt was made to give procedural explanations increased by 14% from 5% in Round 1 to 29% of texts in Round 2. This increase was significant at 99%. 3. The number of texts where conceptual explanations included poor conceptual links decreased by 7% (from 17% in Round 1 to 10% in Round 2). This decrease was significant at 95%. 4. The overall number of texts that demonstrate strong conceptual explanations decreased by 17% in Round 2 (from 78% in Round 1 to 61% in Round 2). 5. The gap between explanations that include some but not all conceptual links and conceptual explanations with links that explain the process and background was 9% in Round 1 but it grew much wider to 29% in Round 2. 6. The decrease between Round 1 and Round 2 in the number of texts with conceptual explanations with links that explain the process and background was significant at 99%. 58 Conceptual explanations, by mathematical content Number was the content area in which the groups’ conceptual explanations of the correct answer are the strongest. Algebra is the content area in which the groups’ conceptual explanations of the correct answer were the weakest. The graphs below represent the percentages of conceptual explanations for these two content areas across the two rounds. Figure 9: Round 1 and 2 conceptual explanations of answers – content area number Number Conceptual explanations Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 3.409090909 15.90909091 32.95454545 47.72727273 Round 2 25 9.375 40.625 25 Change in the strong content area: The number of texts that include conceptual links that explain process and background decreased from 47.73% in Round 1 to 25% in Round 2. The number of texts that include some but not all conceptual links increased from 32.95% in Round 1 to 40% in Round 2. The number of texts that include poorly conceived conceptual links decreased from 15.91% in Round 1 to 9.38% in Round 2. The number of texts that include no conceptual links in explanation increased from 3.41 in Round 1 to 25% in Round 2. 59 Figure 10: Round 1 and 2 conceptual explanations of answers – content area algebra Alegbra Conceptual explanations Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 11.36363636 27.27272727 36.36363636 25 Round 2 41.66666667 4.166666667 45.83333333 8.333333333 Change in the weak area: The number of texts that include conceptual links that explain process and background decreased from 25%% in Round 1 to 8.33% in Round 2. The number of texts that include some but not all conceptual links increased from 36.36% in Round 1 to 45.83% in Round 2. The number of texts that include poorly conceived conceptual links decreased from 27.27% in round 1 to 4.17% in round 2. The number of texts that include no conceptual links in explanation increased from 11.36 in Round 1 to 41.67% in round 2. In comparison, the trends between the two rounds in these two content areas are similar. Noted is the far higher increase of texts that include no conceptual links in explanation in the weaker area. 60 4.2.3 Awareness of Mathematical Error Figure 11: Awareness of error in explanations – Round 1 and 2 by grouped grades Awareness of Error - Round 1 and Round 2 by grade 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% Grade 3-6 Grade 7-9 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 Mathematical Mathematical explanation of explanation of error not present error is flawed or incomplete Mathematical explanation emphasises the procedural Mathematical explanation emphasises the conceptual Analysis of awareness of error in Grade 3-6 small groups’ texts: The information in this graph is presented in the table below, classifying groups’ explanation texts according to the above four criteria: 61 Table 13: Grades 3-6 awareness of mathematical error demonstrated in teacher text explanations Change Level of mathematical Round Round Strength of between Round 1 awareness 1 2 explanation rounds In these texts a mathematical explanation of the particular error is not included. 18% 21% 3% In these texts the explanation of the particular error is mathematically inaccurate or incomplete and hence potentially confusing. 12% 4% -8%* In these texts the explanation of the particular error is mathematically sound but does not link to common misconceptions or errors. The explanation is predominantly procedural 39%* 53%* 14%** In these texts the explanation of the particular error is mathematically sound and suggests links to common misconceptions or errors. 30% 21% Round 2 Weaker 30% 25% Stronger 69% 75% 9% ** Difference significant at a 95% level of confidence * Difference significant at a 90% level of confidence Observations about groups’ awareness of the mathematical error in Rounds 1 and 2 for the grade 3-6 groups: 1. The number of texts of the Grade 3-6 group that demonstrate weak mathematical awareness decreased in Round 2 (From 30% in Round 1 to 25% in Round 2). 2. The number of texts with mathematical flaws and or that are potentially confusing was reduced by 8%, from 12% to 4%. 3. Group 3-6 demonstrates a slight increase in the number of texts that demonstrate strong awareness (from 69% to 75%). 4. For the texts in Round 2 that demonstrate strong mathematical awareness, the gap between texts where explanation of the error is predominantly procedural and texts 62 where explanation of error links to common misconceptions of errors grew wider, from 9% in Round 1 to 32% in Round 2. Analysis of Grade 7-9 small groups’ texts: Table 14: Grades 7-9 awareness of mathematical error demonstrated in teacher text explanations Change Level of mathematical Round Round Strength of between Round 1 awareness 1 2 explanation rounds In these texts a mathematical explanation of the particular error is not included. 15% 22% 7% In these texts the explanation of the particular error is mathematically inaccurate or incomplete and hence potentially confusing. 24% 15% -9% In these texts the explanation of the particular error is mathematically sound but does not link to common misconceptions or errors. The explanation is predominantly procedural. 35% 38% 3% In these texts the explanation of the particular error is mathematically sound and suggests links to common misconceptions or errors. 26% 25% Round 2 Weaker 39% 37% Stronger 61% 63% -1% Observations about groups’ awareness of the mathematical error in Rounds 1 and 2 for the grade 7-9 groups: 1. The number of texts of the Grade 7-9 group that demonstrate weak mathematical awareness remained the same in Round 2 (39% in Round 1 and 37% in Round 2). 2. The number of texts with mathematical flaws or that are potentially confusing was reduced in Round 2 by 9% (from 24% in Round 1 to 15%, in Round 2). 63 3. The number of general texts increased (from 15% in Round 1 to 22% in Round 2). 4. The number of texts of the Grade 7-9 group that demonstrate strong mathematical awareness increased very slightly in Round 2 (from 61% in Round 1 to 63% in Round 2). 5. The gap between texts where explanation of the error is predominantly procedural and texts where explanation of the error links to common misconceptions or errors also only slightly grew wider in Round 2 (from 9% in Round 1 to 13% in Round 2). Overall findings on the awareness of the mathematical error – Rounds 1 and 2 Figure 12: Awareness of error in explanations – Round 1 and 2 Awareness of error - Overall - Round 1 and 2 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% 50% 36% 18% 22% 29% 17% 22% 6% Round 1 Round 2 Mathematical explanation of error not present Round 1 Round 2 Mathematical explanation of error is flawed or incomplete Round 1 Round 2 Mathematical explanation emphasises the procedural Round 1 Round 2 Mathematical explanation emphasises the conceptual 64 Table 15: Awareness of the mathematical error demonstrated in teacher test explanations Change Level of mathematical Round Round Strength of Round Round between awareness 1 2 explanation 1 2 rounds In these texts a mathematical explanation of the particular error is not included. 18% 22% 4% In these texts the explanation of the particular error is mathematically inaccurate or incomplete and hence potentially confusing. 17% 6% -11%*** In these texts the explanation of the particular error is mathematically sound but does not link to common misconceptions or errors. The explanation is predominantly procedural. 36% 50% 14%*** In these texts the explanation of the particular error is mathematically sound and suggests links to common misconceptions or errors. 29% 22% Weaker 36% 28% Stronger 64% 72% -7% *** Difference significant at a 99% level of confidence Observations about groups’ awareness of the mathematical error in Rounds 1 and 2: 1. The overall number of texts that demonstrate weak awareness of error decreased between Rounds 1 and 2 (from 36% in Round 1 to 28% in Round 2). 2. The number of texts where no mathematical awareness of the error was evident increased by 4% from 18% in Round 1 to 22% of texts in Round 2. This increase was not significant. 3. The number of texts where explanations were mathematically inaccurate decreased by 11% (from 17% in Round 1 to 6% in Round 2). This decrease was significant at 99%. 4. The overall number of texts that demonstrate strong awareness of the mathematical error increased by 8% in Round 2 (from 64% in Round 1 to 72% in Round 2). 65 5. The gap between explanations that show more procedural than conceptual awareness of the error was 9% in Round 1 but it grew much wider to 28% in Round 2. 6. The increase between Round 1 and Round 2 in the number of texts show more procedural awareness of the error was significant at 99%. Awareness of error, by mathematical content Number was the content area in which the groups’ awareness of error in their explanations of the choice of incorrect answer was the strongest. Shape was the content area in which the groups’ awareness of error in their explanations of the choice of incorrect answer was the weakest. The graphs below represent the percentages of explanations demonstrating awareness of error for these two content areas across the two rounds. Figure 13: Round 1 and 2 awareness of error – content area number Number Awarenss of error Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 20.89552239 11.94029851 28.35820896 38.80597015 Round 2 18.18181818 3.03030303 54.54545455 24.24242424 Change in the strong content area: The number of texts that demonstrate awareness of concept decreased from 38.81% in Round 1 to 24.24% in Round 2. The number of texts that demonstrate awareness of procedure increased from 28.36% in Round 1 to 54.55% in Round 2. The number of texts that have mathematical flaws decreased from 11.94% in Round 1 to 3.03% in Round 2. 66 The number of texts that includes no mathematical awareness decreased from 20.90 in Round 1 to 18.18% in Round 2. Figure 14: Round 1 and 2 awareness of error – content area shape Shape Awarenss of error Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 17.85714286 26.78571429 41.07142857 14.28571429 Round 2 22.22222222 5.555555556 61.11111111 11.11111111 Change in the weak content area: The number of texts that demonstrate awareness of concept decreased from 14.29% in Round 1 to 11.11% in Round 2. The number of texts that demonstrate awareness of procedure increased from 41.07% in Round 1 to 61.11% in Round 2. The number of texts that have mathematical flaws decreased from 26.79% in Round 1 to 5.56% in Round 2. The number of texts that includes no mathematical awareness increased from 17.86% in Round 1 to 22.22% in round 2. 67 4.2.4 Diagnostic reasoning Figure 15: Diagnostic reasoning in explanations – Round 1 and 2 by grouped grades Diagnostic Reasoning - Round 1 and Round 2 by grade 100% 90% 80% 70% 60% 50% 40% Grade 3-6 30% Grade 7-9 20% 10% 0% Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 No attempt to explain learner reasoning Description of Description of Description of learner reasoning learner reasoning learner reasoning does not hone in hones in on error is systematic and on error but is incomplete hones in on error Analysis of diagnostic reasoning in Grade 3-6 small groups’ texts: Table 16: Grades 3-6 diagnostic reasoning demonstrated in teacher text explanations Change Round Round Strength of Diagnostic awareness between Round 1 1 2 explanation rounds In these texts, no attempt is seen to describe learners’ mathematical reasoning behind the particular error 24% 29% Round 2 5% In these texts the description of the learners’ mathematical reasoning does not hone in on the particular error. 31% 29% -2% In these texts the description of the learners’ mathematical reasoning is 35% 30% -5% Weaker 55% 58% Stronger 45% 42% 68 incomplete although it does hone in on the particular error. In these texts the description of the steps of learners’ mathematical reasoning is systematic and hones in on the particular error. 10% 12% 2% Observations about groups’ diagnostic reasoning in relation to the mathematical error in Rounds 1 and 2 for the grade 3-6 groups: 1. The number of texts of the Grade 3-6 group that demonstrate weak diagnostic reasoning increased slightly between Rounds 1 and 2 (from 55% in Round 1 to 58% in Round 2). 2. The number of texts where no attempt was made to explain learner reasoning behind the error increased by 4% from 24% in Round 1 to 29% of texts in Round 2. 3. The number of texts where learner reasoning was described but did not hone in on the error slightly decreased by 3% (from 31% in Round 1 to 29% in Round 2). 4. The number of texts of the Grade 3-6 group that demonstrate strong diagnostic reasoning slightly decreased by 3% in Round 2 (from 45% in Round 1 to 42% in Round 2). 5. The gap between incomplete and systematic explanations was high but decreased Round 2 (from 25% in Round 1 to 18% in Round 2). Analysis of diagnostic reasoning in Grade 7-9 small groups’ texts: Table 17: Grades 7-9 diagnostic reasoning demonstrated in teacher text explanations Change Round Round Strength of Round Diagnostic awareness between 1 2 explanation 1 rounds In these texts, no attempt is seen to describe learners’ mathematical reasoning behind the particular error 23% In these texts the description of the learners’ mathematical reasoning does not hone in on the particular error 36% 25% 2% Weaker 12% Round 2 59% 37% -24%*** 69 In these texts the description of the learners’ mathematical reasoning is incomplete although it does hone in on the particular error 25% In these texts the description of the steps of learners’ mathematical reasoning is systematic and hones in on the particular error. 16% 53% 28%*** Stronger 11% 41% 64% -5% *** Difference significant at a 99% level of confidence Observations about groups’ diagnostic reasoning in relation to the mathematical error in Rounds 1 and 2 for the grade 7-9 groups: 1. There was a big decrease in Round 2 in the number of texts of the Grade 7-9 group that demonstrate weak diagnostic reasoning (from 59% in Round 1 to 37% in Round 2). 2. The number of texts where no attempt was made to explain learner reasoning increased by 2% from 23% to 25% of texts. 3. The number of texts where teachers attempt to describe learner reasoning but the description does not hone in on the error decreased by 24% (from 36% in Round 1 to 12% in Round 2). 4. The number of texts of the Grade 7-9 group that demonstrate strong diagnostic reasoning increased by 23% in Round 2 (from 41% in Round 1 to 64% in Round 2). 5. The gap between incomplete and systematic explanations was not very high in Round 1 (9%) but grew wider in Round 2 (41%). 70 Overall findings on the diagnostic reasoning in relation to the mathematical error – Rounds 1 and 2 Figure 16: Diagnostic reasoning in relation to the error in explanations – Round 1 and 2 Diagnostic reasoning - Overall - Round 1 and 2 100% 90% 80% 70% 60% 50% 40% 30% 26% 28% 32% 24% 30% 36% 20% 12% 12% Round 1 Round 2 10% 0% Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 No attempt to explain Description of learner Description of learner Description of learner learner reasoning reasoning does not reasoning hones in on reasoning is systematic hone in on error error but is incomplete and hones in on error Table 18: Diagnostic reasoning in relation to the error demonstrated in teacher test explanations Change Round Round Strength of Diagnostic awareness between Round 1 1 2 explanation rounds No attempt to explain learner reasoning 26% 28% 2% Description of learner reasoning does not hone in on error 32% 24% -12% Description of learner reasoning hones in on error but is incomplete 30% 36% 6% Description of learner reasoning is systematic and hones in on error 12% 12% Round 2 Weaker 58% 52% Stronger 42% 48% 0% 71 Observations about groups’ diagnostic reasoning in relation to the mathematical error in Rounds 1 and 2: 1. The overall number of texts that demonstrate weak diagnostic reasoning in relation to the error decreased slightly between Rounds 1 and 2 (from 58% to 52%). 2. The number of texts where there was no attempt to explain the learners’ errors was evident increased by 2% from 26% in Round 1 to 28% of texts in Round 2. This increase was not significant. 3. The number of texts where explanations did not hone in on the error decreased by 6% (from 32% in Round 1 to 24% in Round 2). This decrease was not significant. 4. The overall number of texts that demonstrate strong diagnostic reasoning in relation to the error increased by 6% in Round 2 (from 42% in Round 1 to 48% in Round 2). 5. The gap between explanations that hone in on the error but are either incomplete or complete was 18% in Round 1 but it grew to 24% in Round 2. 6. The number of texts with incomplete explanations but that do hone in on the error increased from 30% in Round 1 to 36% in Round 2. This increase was not significant. 7. The number of texts with complete explanations that hone in on the error was 12% of all texts and did not change between Rounds 1 and 2. Diagnostic reasoning, by mathematical content Measurement was the content area in which the groups’ diagnostic reasoning their explanations of the choice of incorrect answer was the strongest. Shape was the content area in which the groups’ diagnostic reasoning their explanations of the choice of incorrect answer was the weakest. The graphs below represent the percentages of explanations demonstrating diagnostic reasoning in the explanation of errors for these two content areas across the two rounds. 72 Figure 17: Round 1 and 2 diagnostic reasoning – content area measurement Percentage of texts Measurement Diagnostic reasoning Rounds 1 and 2 100 90 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 25.92592593 24.07407407 35.18518519 14.81481481 Round 2 21.42857143 25 28.57142857 25 Change in the strong content area: The number of texts in which the description of learner reasoning is systematic and hones in on error increased from 14.81% in Round 1 to 25% in Round 2. The number of texts in which the description of learner reasoning hones in on error but is incomplete decreased from 35.19% in Round 1 to 28.57% in Round 2. The number of texts in which the description of learner reasoning does not hone in on error increased from 15.93% in Round 1 to 21.43% in Round 2. The number of texts that include no attempt to explain learner reasoning decreased from 25.93 in Round 1 to 21.43% in Round 2. 73 Figure 18: Round 1 and 2 diagnostic reasoning – content area shape Shape Diagnostic reasoning Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 28.57142857 41.07142857 21.42857143 8.928571429 Round 2 33.33333333 27.77777778 38.88888889 0 Change in the weak content area: The number of texts in which the description of learner reasoning is systematic and hones in on error decreased from 8.93% in Round 1 to 0% in Round 2. The number of texts in which the description of learner reasoning hones in on error but is incomplete increased from 21.43% in Round 1 to 38.89% in Round 2. The number of texts in which the description of learner reasoning does not hone in on error decreased from 41.07% in Round 1 to 27.78% in Round 2. The number of texts that include no attempt to explain learner reasoning increased from 28.57% in Round 1 to 33.33% in Round 2. 74 4.2.5 Multiple explanations In order to code teachers’ use of multiple explanations, all the items were first coded as mathematically feasible or not. The coding on this criterion was per item since the code (per item) was assigned based on the overall coding of all of the texts relating to each item. Non-feasible explanations include explanations that are not mathematically focused on the error under discussion (for example, ‘Simply a lack in conceptual understanding of an odd number’) or are what we call “general texts” (for example, ‘We could not ascertain the reasoning behind this distractor as there was no clear or obvious reason why a learner would choose it’).The final code for multiple explanations was assigned according to the number of feasible/non-feasible explanations given for texts relating to each item. Figure 19: Feasible and non-feasible explanations – Round 1 and 2 Feasible and not feasible explanations - Round 1 and Round 2 by grade 100% 90% 85% 86% 80% 78% 80% 70% 60% 50% Grade 3-6 40% Grade 7-9 30% 20% 22% 15% 14% 20% 10% 0% Round 1 Round 2 Feasible Round 1 Round 2 Not feasible Observation The differences between Rounds 1 and 2 show a small decrease in the number of feasible explanations offered by the two groups (Grade 3-6 and Grade 7-9) in explanation of error and a corresponding increase in non-feasible explanations. This finding points to the value of the presence of an expert leader in the groups. 75 The findings reported below show that although the decrease in Round 2 in the number of feasible explanations is small (5%) for the Grade 3-6, and slightly higher (8%) for the Grade 7-9, teachers across the two groups provide mainly one mathematically feasible/convincing explanation (with/without general explanation). Figure 20: Multiple explanations in explanations – Round 1 and 2 by grouped grades Multiple Explanations - Round 1 and Round 2 by grade 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% Grade 3-6 Grade 7-9 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 No feasible mathematical explanation One feasible mathematical explanation with/without general explanations Two feasible mathematical explanations combined with general explanations Two (or more) feasible mathematical explanations 76 Analysis of multiple explanations in Grade 3-6 small groups’ texts: Table 19: Grades 3-6 multiple explanations demonstrated in teacher text explanations Change Round Round Strength of Multiple explanations between Round 1 1 2 explanation rounds In these items no mathematically feasible/convincing explanation is provided 4% 9% 5% In these items one mathematically feasible/convincing explanation is provided (with/without general explanations) 62% 74% 8% In these items, at least two of the mathematical explanations are feasible/convincing (combined with general explanations) 13% 7% -6% In these items, all of the explanations (two or more) are mathematically feasible/convincing 21% 10% Round 2 Weaker 66% 83% Stronger 34% 17% -11% Observations about groups’ multiple explanations of mathematical errors in Rounds 1 and 2 for the grade 3-6 groups: 1. Grade 3-6 use of multiple explanations was low in both Rounds 1 and 2. 66% of the items in Round 1 and 83% of the items in Round 2 demonstrate little use of multiple explanations. 2. The number of items with no feasible mathematical explanation at all increased very slightly (from 4% in Round 1 to 9% in Round 2). 3. The number of items with one feasible mathematical explanation was high and increased by 11% in Round 2, (from 62% in Round 1 to 73%, in Round 2). 4. Multiple explanations (two or more) were not highly evident in Round 1 and even less so in Round 2 in the grade 3-6 groups. 5. The number of items in the Grade 3-6 group that demonstrate use of multiple explanations decreased by 17% in Round 2 (from 34% in Round 1 to 17% in Round 2). The decrease is evident in both types of stronger explanations (6% and 11% in the respective types). 77 6. The Grade 3-6 groups provide mainly one mathematically feasible/convincing explanation (with/without general explanations). Analysis of multiple explanations in Grade 7-9 small groups’ texts: Table 20: Grades 7-9 multiple explanations demonstrated in teacher text explanations Change Round Round Strength of Multiple explanations between Round 1 1 2 explanation rounds In these items no mathematically feasible/convincing explanation is provided 5% 17% 12% In these items one mathematically feasible/convincing explanation is provided (with/without general explanations) 48% 70% 22% In these items, at least two of the mathematical explanations are feasible/convincing (combined with general explanations) 8% 0% -8% In these items, all of the explanations (two or more) are mathematically feasible/convincing 38% 13% Round 2 Weaker 53% 87% Stronger 46% 13% -25% Observations about groups’ multiple explanations of mathematical errors in Rounds 1 and 2 for the grade 7-9 groups: 1. Grade 7-9 use of multiple explanations was low in both Rounds 1 and 2. 53% of the items in Round 1 and 87% of the items in Round 2 demonstrate little use of multiple explanations. 2. The number of items with no feasible mathematical explanation at all increased by 12% (from 5% in Round 1 to 17% in Round 2). 3. The number of items with one feasible mathematical explanation radically increased in Round 2 by 22% (from 48% in Round 1 to 70%, in Round 2). 4. Multiple explanations (two or more) were not highly evident in Round 1 and even less so in Round 2 in the Grade 7-9 groups. 78 5. The number of items of the Grade 7-9 groups that demonstrate use of multiple explanations radically decreased, by 30% in Round 2 (from 46% in Round 1 to 13% in Round 2). 6. There was a difference of 30% between the percentage of items where two or more feasible mathematical explanations were offered with and without general explanations in Round 1 and a 13% gap between these multiple explanation categories in Round 2. This means that there was a decrease in both types of multiple explanations from Round 1 to Round 2. 7. There were 8% of items with two or more mathematical and general explanations and 25% of items where two or more mathematical explanations were put forward. In other words, Grades 7-9 did have 13% of items for which they offered two or more feasible mathematical explanations without general explanations but all other explanations they gave fell into the “weaker” categories. 8. The Grade 7-9 groups provide mainly one mathematically feasible/convincing explanation (with/without general explanations). 79 Overall findings on the multiple explanations of the mathematical error – Rounds 1 and 2 Figure 21: Multiple explanations of the error in explanations – Round 1 and 2 Multiple Explanations- Overall - Round 1 and 2 100% 90% 80% 74% 70% 56% 60% 50% 40% 30% 20% 10% 11% 11% 4% 12% 11% Round 1 Round 2 4% 0% Round 1 Round 2 No feasible mathematical explanation Round 1 Round 2 Round 1 Round 2 One feasible Two feasible Two (or more) feasible mathematical mathematical mathematical explanation explanations combined explanations with/without general with general explanations explanations 80 Table 21: Multiple explanations of the error demonstrated in teacher test explanations Change Round Round Strength of Round Multiple explanations between 1 2 explanation 1 rounds No feasible mathematical explanation 4% 11% 7% One feasible mathematical explanation with/without general explanations 56% 74% 18% Two feasible mathematical explanations combined with general explanations 11% 4% -7% Two (or more) feasible mathematical explanations 12% 11% Round 2 Weaker 60% 85% Stronger 23% 15% -1% Observations about groups’ multiple explanations of the mathematical error in Rounds 1 and 2: 1. The use of multiple explanations was low in both Rounds 1 and 2. 2. 60% of the items in Round 1 and 85% of the items in Round 2 demonstrate little use of multiple explanations. 3. The number of items with no feasible mathematical explanation at all increased by 7% (from 4% in Round 1 to 11% in Round 2). 4. The number of items with one feasible mathematical explanation increased in Round 2 by 18% (from 56% in Round 1 to 74% in Round 2). 5. There were very few multiple explanations (two or more) in Round 1 and even less so in Round 2. 6. The number of items where any kind of multiple explanations of the error were given decreased, by 8% in Round 2 (from 23% in Round 1 to 15% in Round 2). 7. However the decrease was not in the strongest category of multiple explanations (which decreased from 12% in Round 1 to 11% in Round 2), it was in the category of two feasible mathematical explanations combined with general explanations (which decreased from 11% in Round 1 to 4% in Round 2). 8. All groups provide mainly one mathematically feasible/convincing explanation (with/without general explanations). Multiple explanations, by mathematical content Shape was the content area in which the groups’ multiple explanations of the choice of the incorrect answer were the strongest. Algebra was the content area in which the 81 groups’ multiple explanations of the choice of incorrect answer were the weakest. The graphs below represent the percentages of explanations demonstrating multiple explanations of the error for these two content areas across the two rounds. Figure 22: Round 1 and 2 Multiple explanations – content area shape Shape Multiple Explanations Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 3.448275862 58.62068966 10.34482759 27.5862069 Round 2 15.38461538 61.53846154 0 23.07692308 Change in the strong content area: The percentage of items for which all explanations (two or more) were mathematically feasible/convincing decreased from 27.59% in Round 1 to 23% in Round 2. The percentage of items for which at least two of the mathematical explanations were feasible/convincing but which were combined with general explanations decreased from 10.34% in Round 1 to 0% in Round 2. The percentage of items in which one mathematically feasible/convincing explanation was provided (with/without general explanations) increased from 58.62% in Round 1 to 61.54% in Round 2. The percentage of items for which no mathematically feasible/convincing explanation was provided increased from 3.45% in Round 1 to 15.38% in Round 2. 82 Figure 23: Round 1 and 2 Multiple explanations – content area algebra Algebra Multiple Explanations Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 5.263157895 68.42105263 10.52631579 15.78947368 Round 2 7.142857143 92.85714286 0 0 Change in the weak content area: The percentage of items for which all explanations (two or more) were mathematically feasible/convincing decreased from 15.79% in Round 1 to 0% in Round 2. The percentage of items for which at least two of the mathematical explanations were feasible/convincing but which were combined with general explanations decreased from 10.53% in Round 1 to 0% in Round 2. The percentage of items in which one mathematically feasible/convincing explanation was provided (with/without general explanations) increased from 68.42% in Round 1 to 92.86% in Round 2. The percentage of items for which no mathematically feasible/convincing explanation was provided increased from 5.26% in Round 1 to 7.14% in Round 2. 83 4.2.6 Use of the everyday in explanations of the error “Use of the everyday” was considered in the analysis since the NCS Curriculum emphasises linking the teaching of mathematics to everyday contexts or using everyday contexts to enlighten learners about mathematical concepts. The ICAS tests included predominantly contextualized questions (ranging from 98% in the Grade 3 test to 85% in the Grade 9 test) and yet teachers made very few references to these contexts in their explanations. In certain questions there may not have been contextualized explanations that would have assisted in the explanation of the solution. The code “no discussion of the everyday” includes all explanations which did not make reference to the everyday whether it was appropriate or not. Further more detailed analysis of the kinds of contextualized explanations which were offered and whether they were missing when they should have been present could be carried out since it was not within the scope of this report. Figure 24: Use of the everyday in explanations of the error – Round 1 and 2 by grouped grades Everyday Explanations - Round 1 and Round 2 by grade 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% Grade 3-6 Grade 7-9 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 Round 1 Round 2 No discussion of everyday Everyday dominates and obscures the explanation Everyday link is relevant but not mathematically grounded Everyday explanation links clearly to the mathematcial concept illustrated Can see that Grades 3-6 make more reference to the everyday in their explanations, but this is minimal. 84 Analysis of use of the everyday in Grade 3-6 small groups’ texts: Table 22: Grades 3-6 use of the everyday demonstrated in teacher text explanations of the error Change Use of everyday in Round Round Strength of Round Round between explanations 1 2 explanation 1 2 rounds In these texts no discussion of everyday is evident 88% 97% 9%*** In these texts teachers’ use of the ‘everyday’ dominates and obscures the mathematical understanding, no link to mathematical understanding is made 5% 1% -4%** In these texts teachers’ use of the ‘everyday’ is relevant but does not properly explain the link to mathematical understanding 5% 2% -3%* In these texts teachers’ use of the ‘everyday’ enables mathematical understanding by making the link between the everyday and the mathematical clear 2% 0% low 93% 98% high 7% 2% -2%* *** Difference significant at a 99% level of confidence ** Difference significant at a 95% level of confidence * Difference significant at a 90% level of confidence Observations about the quality of the use of error analysis between Round 1 and Round 2 for the grade 3-6 groups: 1. Grade 3-6 use of everyday explanations was extremely low in both Rounds 1 and 2. 2. In 88% of the texts in Round 1 and in 97% of the texts in Round 2, Grade 3-6 group did not discuss every day contexts when they dealt with the errors. 3. Together with the texts where the use of everyday context was obscuring (5% in Round 1 and 1% in round 2) the number of texts for Grade 3-6 groups in which discussion of everyday contexts in the explanation of error is low increased by 5% (from 93% in Round 1 to 98% in Round 2) 85 4. Grade 3-6 use of everyday explanations which was relevant but incomplete and or enabled mathematical understating of the error was very low in both Rounds 1 and 2 (7% in Round 1 and 2% in Round 2). 5. The difference in use of everyday (relevant but incomplete and enabled mathematical understating) was small and almost the same in the two rounds ( 3% in Round 1 and 2 % in Round 2) Analysis of use of the everyday in explanations in Grade 7-9 small groups’ texts: Table 23: Grades 7-9 use of the everyday demonstrated in teacher text explanations of the error Change Use of everyday in Round Round Strength of Round Round between explanations 1 2 explanation 1 2 rounds In these texts no discussion of everyday is evident 96%* 100%* 4%** In these texts teachers’ use of the ‘everyday’ dominates and obscures the mathematical understanding, no link to mathematical understanding is made 1% 0% -1% In these texts teachers’ use of the ‘everyday’ is relevant but does not properly explain the link to mathematical understanding 3% 0% -3%* In these texts teachers’ use of the ‘everyday’ enables mathematical understanding by making the link between the everyday and the mathematical clear 1% 0% Weaker 97% 100% Stronger 3% 0% -1%* ** Difference significant at a 95% level of confidence * Difference significant at a 90% level of confidence 86 Observations about the quality of the use of error analysis between Round 1 and Round 2 for the grade 7-9 groups: 1. Grade 7-9 use of everyday explanations was extremely low in both Rounds 1 and 2. 2. In 96% of the texts in Round 1 and in 100% of the texts in Round 2, Grade 7-9 group did not discuss every day contexts when they dealt with the errors. 3. In Round 2 texts with no discussion of everyday contexts increased by 9%. 4. Together with the texts where the use of everyday context was obscuring (1% in Round 1 and 0% in round 2) the number of texts for Grade 7-9 groups in which discussion of everyday contexts in the explanation of error is negligible. 5. Grade 7-9 use of everyday explanations which was relevant but incomplete and or enabled mathematical understating of the error was very low in both Rounds 1 and 2 (3% in Round 1 and 0% in Round 2). 6. Only in Round there was a difference in use of everyday (relevant but incomplete and enabled mathematical understating), which was negligible. Overall findings on the use of the everyday in explanations of the mathematical error – Rounds 1 and 2 Figure 25: Use of the everyday demonstrated in teacher explanations of the error – Round 1 and 2 Use of everyday in explanations- Overall - Round 1 and 2 100% 98% 91% 90% 80% 70% 60% 50% 40% 30% 20% 12% 10% 4% 1% 4% 1% 0% 0% Round 1 Round 2 No discussion of everyday Round 1 Round 2 Everyday dominates and obscures the explanation Round 1 Round 2 Everyday link is relevant but not mathematically grounded Round 1 Round 2 Everyday explanation links clearly to the mathematcial concept illustrated 87 Table 24: Use of the everyday in explanations of the error demonstrated in teacher test explanations Change Use of everyday in Round Round Strength of Round between explanations 1 2 explanation 1 rounds In these texts no discussion of everyday is evident 91% 98% 7% In these texts teachers’ use of the ‘everyday’ dominates and obscures the mathematical understanding, no link to mathematical understanding is made 4% 1% -3%** In these texts teachers’ use of the ‘everyday’ is relevant but does not properly explain the link to mathematical understanding 4% 1% -3%* In these texts teachers’ use of the ‘everyday’ enables mathematical understanding by making the link between the everyday and the mathematical clear 12% 0% Round 2 Weaker 94% 99% Stronger 16% 1% -12%*** *** Difference significant at a 99% level of confidence ** Difference significant at a 95% level of confidence * Difference significant at a 90% level of confidence Observations about groups’ use of the everyday in explanations of the mathematical error in Rounds 1 and 2: 1. The overall use of everyday in explanations of the error was extremely low in both rounds 1 and 2. 2. In 91% of the texts in Round 1 and 98% of the texts, the groups did not discuss every day related explanations of the error. 3. Only in 12% of the texts in Round 1 the groups used ‘everyday’ in ways that enables mathematical understanding, by making the link between the everyday and the mathematical clear 88 Use of the everyday, by mathematical content There was very little reference to the everyday in mathematical explanations across all content areas, but since such minimal reference was made in this section we highlight the two areas in which at least some reference was made to the everyday in Rounds 1 and 2. The graphs below represent the percentages of explanations where some reference to the everyday was demonstrated in the explanations of errors for these two content areas across the two rounds. Figure 26: Round 1 Use of everyday in explanations – content area number Measurement use of the everyday Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 85.18518519 1.851851852 5.555555556 7.407407407 Round 2 96.42857143 3.571428571 0 0 Change in the area of measurement The number of texts that demonstrate groups’ use of the ‘everyday’ that enables mathematical understanding by making the link between the everyday and the mathematical clear, decreased from 7.41% in Round 1 to 0% in Round 2. The number of texts in which teachers’ use of the ‘everyday’ is relevant but does not properly explain the link to mathematical understanding, decreased from 5.56% in Round 1 to 0% in Round 2. The number of texts in which teachers’ use of the ‘everyday’ dominates and obscures the mathematical understanding, and where no link to mathematical understanding is made, is negligible, decreasing from 1.85% in Round 1 to 0% in Round 2. The number of texts where no discussion of everyday is evident increased from 85.19% in Round 1 to 96.43% in Round 2. 89 Figure 27: Round 1 Use of everyday in explanations – content area measurement Number use of the everyday Rounds 1 and 2 100 90 Percentage of texts 80 70 60 50 40 30 20 10 0 Not present Inaccurate Partial Full Round 1 92.53731343 4.47761194 2.985074627 0 Round 2 96.96969697 0 3.03030303 0 Change in the area of number: In both rounds no text was found that demonstrates groups’ use of the ‘everyday’ that enables mathematical understanding by making the link between the everyday and the mathematical clear. The number of texts in which teachers’ use of the ‘everyday’ is relevant but does not properly explain the link to mathematical understanding, was negligible in both rounds, increasing from 2.09% in Round 1 to 3% in Round 2. The number of texts in which teachers’ use of the ‘everyday’ dominates and obscures the mathematical understanding, and where no link to mathematical understanding is made, decreased from 4.48% in Round 1 to 0% in Round 2. The number of texts where no discussion of everyday is evident increased from 92.54% in Round 1 to 96.97% in Round 2. 4.2.7 Comparative strengths and weaknesses between the two sets of grouped grades The graphs below show the changes between rounds in the number of texts demonstrating texts which do not have mathematical content “not present” and texts which are mathematcially inaccurate (“inaccurate”) in the procedural, conceptual, awareness and diganostic criteria. The first bar (blue) in each pair represents the Grade 3-6 group and the second bar (red) represents the Grade 7-9 group. Positive changes 90 represnet an increase between Round 1 and Round 2 while negative changes represent a decrease between Round 1 and Round 2. The actual changes in percentages betweeen rounds are given in the table below the graph. Figure 28: Grade 3-6 and Grade 7-9 changes in “not present” and “inaccurate” explanations Changes between rounds on criterion levels "not present" and "inaccurate" Percentage change between rounds 40 30 20 10 0 -10 -20 -30 Not Inaccurat Not Inaccurat Not Inaccurat Not Inaccurat present e present e present e present e Procedural Conceptual Subject Matter Knowledge Awareness Diagnostic PCK Grade 3-6 14.220681 -5.828103 25.956199 -8.311805 3.0235019 -8.373027 4.6592256 -2.041653 Grade 7-9 5.760582 1.436078927.698413 -11.32615 7.554726 -9.288372 1.4304409 -23.89219 These two criterion levels are the “weaker” levels and thus an increase here represents a weakening while an increase represents strengthening between rounds. The absence of mathematical content in an explanation increased across all criteria for both grade groups but the presence of inaccurate texts decreased more markedly for the Grade 7-9 group, which suggests that the Grade 7-9 group held better quality when they worked on their own without group leaders. The graphs below show the changes between rounds in the number of texts demonstrating texts which demonstrate incomplete mathematical explanations “partial” and texts which demonstrate full mathematical explanations (“full”) in the procedural, conceptual, awareness and diganostic criteria. The first bar (blue) in each pair represents the Grade 3-6 group and the second bar (red) represents the Grade 7-9 group. Positive 91 changes represnet an increase between Round 1 and Round 2 while negative changes represent a decrease between Round 1 and Round 2. The actual changes in percentages betweeen rounds are given in the table below the graph. Figure 29: Grade 3-6 and Grade 7-9 changes in “partial” and “full” explanations Changes between rounds on criterion levels "partial" and "full" Percentage change between rounds 40 30 20 10 0 -10 -20 -30 Partial Full Procedural Partial Full Partial Conceptual Subject Matter Knowledge Full Awareness Partial Full Diagnostic PCK Grade 3-6 14.617097 -23.00967 9.2269335 -26.87133 14.17001 -8.820485 -5.533302 2.9157297 Grade 7-9 12.502145 -19.69881 -6.657372 -9.714893 3.2271822 -1.493536 28.27757 -5.815822 These two criterion levels are the “stronger” levels and thus an increase here represents a strengthening while a decrease represents weakening between rounds. The lack of completeness in mathematical explanations increased across all criteria for both grade groups but less so for the Grade 7-9 group, while the fullness of correct explanations decreased but again less so for the Grade 7-9 group, which again suggests that the Grade 7-9 group held better quality when they worked on their own without group leaders. 92 4.2.8 Comparative strengths and weaknesses according to content area The findings in relation to content areas in which groups were strongest and weakest are summarised in Table 25 below. Strength was measured by taking the content area where higher level explanations (partial or full) were the strongest in the two rounds. An analysis of the changes between Round 1 and 2 for the various categories and types of explanations offered shows that the groups explanations were most often stronger in the area of number than in any other of the mathematical content areas. Weakness was measured by taking into account where lower level explanations (not present or inaccurate) were the most common in both rounds, or were the highest in Round 2. There was not one single content area where weakness was demonstrated consistently more than in other areas. Weakness was evident in the areas of data, algebra and shape. Table 25: Content areas according to strength and weakness in explanations Text Criteria Strong content area Weak content area Procedural Number Data Number Algebra Awareness of error Number Shape Diagnostic Measurement Shape Shape Algebra Number, n/a explanation of Correct Answer text correct answer Conceptual explanation of correct answer reasoning Error text Multiple Explanations Use of everyday Measurement 93 94 Section Five: General findings 95 5.1 What are the findings about teachers reasoning about learners’ errors? This section begins with a short qualitative analysis of teachers’ experience of the error analysis activity. In this we report on three common perceptions of teachers when interviewed on their experiences of the error analysis activity and its influence on their practice. In view of the division between the six criteria we used to evaluate teachers’ knowledge of the items (correct solution and errors) we summarize the findings in two parts. We first look at what the data shows about the three criteria that mapped against Ball’s et al two Subject Matter Knowledge (SMK’s) domains. We then look at the three criteria that mapped against one of the domains Ball et al’s specify for Pedagogical Content Knowledge (PCK). We do this in order to look at the quality of teachers’ reasoning both in relation to the correct solution and in relation to the error. We answer the question: On what criteria is the groups’ error analysis weaker and on what criteria is the groups’ error analysis stronger? We proceed by looking at the difference between grouped grades’ (3-6 and 7-9) performance on the different criteria between rounds. We answer the question: Is there a difference between primary (Grade 3-6 groups) and high school (Grade 7-9 groups) teachers in sustaining the performance on the Round 1 error analysis activity? Although this was not a key focus of the project, in view of the analysis of teacher’s knowledge we examine the differences in performance of the groups across different mathematical content areas. We answer the question: In which mathematical content areas do the groups produce better judgement on the error? Given the model of teacher development in which groups were working with group leaders with academic we tentatively suggest the areas in which they contributed to the groups in Round 1. We answer the question: What does the change in performance between Round 1 and 2 suggest about the role of group leaders? 96 5.2 Teachers’ experiences of the error analysis activity In the various interviews conducted throughout the project, teachers, subject facilitators and group leaders confirmed that the error analysis activity was interesting and worthwhile. The different participants emphasized repeatedly that working with questions actually made a difference to them- it made them aware that questions can be asked in different ways, and they started thinking about asking their questions in a different way. Before, they said, they knew what answer they wanted to get and created a question to get to this answer, but they didn’t really think that some formulations of the same question are better than others. Teachers found it interesting to examine the ICAS questions, to see what multiple choice questions are like, how they work, what kind of information they make available and how. Many of them took ICAS questions and tried them in their classes and came back and reflected on their learners’ performance. They also enjoyed very much creating their own questions, testing their learners and analysing the errors, as can be seen in the following three excerpts from the focus group discussion after the activity. We didn’t know how they came up with the answer. We had no clue. And the only way (laughter) we thought we would have a clue … clue is to let the kids do it … and see what they came up with and then ask them why they chose that answer. That is the only way, because we had no clue. It was this funny picture with stars all over the show and it said, which of the following is rotational symmetry and none of them actually had rotational symmetry if you removed one. So, to us it didn’t make sense and we thought we’d throw it back to the kids. I did the same with the grade 9s. There was one question we couldn’t get to the answer, we got the correct answer but why did, because there was I think 32% chose this specific wrong answer. So I took it to my grade 9 classes, three of them, and then just for the fun of it, I gave it to my grade 10 classes, two of the grade 10 classes. And the responses was…and then just to be really now getting to the bottom of this, I gave it to my grade 8 son at home. And it was very interesting coming from the students sitting in front of you. I also did it with the grade 3s for quite a few questions and I got them to write at the back how did they get to the answer. I didn’t ask them to explain it to me, I told them to write it down. And if they had to draw it, then they drew it, if they had to colour in, they coloured in. And it was very interesting. And many, many of them were able to explain to us how they got to it, and some of them actually said there I guessed because I don’t know. They wrote: I guessed because I don’t know. In what follows we record other teachers’ comments about what they feel they learned from doing the error analysis activity. First, teachers reported that they understand that they need to gain a far deeper understanding of the questions they set for their learners. Although this sounds obvious, 97 teachers reported that their approach to asking questions was simple or even unthoughtful before the error analysis. Well, I don’t think I’ve ever thought about the way we ask questions before. You know, you just do it because it’s been done like that for so many years. So now for the first time you start thinking: but, maybe I can ask it in a different way? Maybe there is still something wrong with the question, how it’s been put itself. Teachers suggested that before this activity they did not have a way to think about test or activity questions: “I think we all just set a question: now work out the perimeter of this, and, you know, that’s it.” They wanted an answer to the question but did not know how to think about the question in depth: “I don’t think as a maths teacher when we originally set any test that you really analyse and probe in depth into the question… .” The teachers felt that together, the mapping activity (see Process Document) and the error analysis activity, showed them a way to examine the conceptual depth of questions. In the following two quotations, the respective teachers point to two related matters of conceptual depth. The first quotation refers to the idea that questions draw on more than one concept. Even though a question may have a specific conceptual focus, it could draw on a number of related concepts and so it is essential that teachers unpack questions when testing learners. Precisely because questions draw on webs of concepts, learners may misunderstand these questions, particularly when the focus of the question is not clear. And maybe to add onto that, I think one other thing that came out of this is that you look at a question and try to go…or maybe try to go back to school and say, what skills do they need? What concepts do they need in order for them to, like she say, there are multiple skills in one question there, what is it that they need to know in order to answer this question? What skills? For instance, you might discover that is a story sum but we need Pythagoras there. Some skills that they need to identify that they need to answer the question. Whilst the first quotation raises the conceptual background of questions which has implications for learners’ preparedness, the next quotation raises the importance of knowing what the question is in fact asking. Does it obscure its focus in ways that learners, in particular, young and inexperienced ones, may struggle to understand what’s expected of them. The teacher in the following quote posits direct and useful questions that she now asks when formulating a question: I think one other thing is that when we look at the questions now and probably now when I look at a question that I set as well, I think about what is it I want from this, what do I want them to do, and is my question going to allow it? And is it set in such a way that they can see the steps that they need to do, or they understand the various operations that are required? Especially at grade 3 level… . And now when I look at a 98 question and especially when it comes to problem solving and things like that I look at that specifically to see, is it going to give us the answer we want, and are the kids going to understand what’s expected of them? Ok, that’s one thing for me to set the question and I understand what I want them to do, but do they understand? Second, In addition to thinking about questions more deeply, teachers reported that their orientation to learners’ approach to questions has changed. Doing the error analysis activity, trying to figure out what the learners were thinking when they chose the correct answer or the distractor, has given them a lens to think about learners’ reasoning: I also thought that I should encourage my children to explain what they say. I should ask them WHY? if they give me an answer… Not just leave it at that. This is very different to what teachers were used to do: We only give learners corrections where they needed corrections. I must say, we never tell them there’s a misconception here, this is how you were supposed to do it, we just give the corrections. To most teachers the notion of “misconception” was new and many struggled to develop productive relationship with the idea. So whilst the literature distinguishes misconceptions from errors and posits that misconceptions are part of learning, for many of the teachers, a misconception remained something that they need to foresee and avoid. In almost every interview, the interviewer needed to remind the teachers that a misconception is a mathematical construct that develops through learning the development of which is not necessarily the teacher’s fault. More than that, misconceptions can be a productive tool for teachers to deepen their understanding of learners’ thinking and inform their teaching. Even towards the end of the project, the teachers tended to treat misconceptions as errors that needed to be avoided: Interviewee: What I can say, when we plan, so we try by all means to avoid the misconception side. Interviewer: To avoid it? Interviewee: To avoid it. I think we try to avoid it. Interviewer: Why would you avoid it? Interviewee: Because it will distract the learners. It will confuse them. So I think that will confuse them, I don’t think we can rely on it. You know, part of good teaching is that you make people aware of the pitfalls. So no, I don’t think you’d avoid it entirely, but being made aware of the pitfalls is different to putting up a booby trap. And I think that’s what she’s saying is don’t put booby traps in the way. Does that make sense? When asked to reflect on how the error analysis activity helped the teachers in planning their lessons and in teaching the planned lesson, teachers acknowledged its influence in 99 two ways. In the following quotation the teacher speaks about awareness. Overcoming misconceptions, she says “was right up front…the whole time”: I would say awareness. If one is aware of the distractors that we identified, that brought it to mind, and I think that we were going through it and when we were drawing up the worksheets and drawing up the things, we did have that at the back of our minds and saying, well let’s not fall into the same trap of having those distractors. And of course as we were designing the learning program, the whole focus was on overcoming the misconception. So it was right up front, it was there the whole time that was our motivation the whole time, was to overcome that misconception of the equal sign. This is an important point in that it suggests that in addition to following a syllabus or a plan made by an HOD, the teacher feels that there is another lens that informs her practice- the challenge to try and overcome misconceptions. I mean, you’re given your frames and your time works, and your milestones and whatever type of thing, but when you actually sit down and plan the lesson, maybe foremost in our mind it should actually be, how am I going to teach this that the learners are going to not make these errors and understand it properly, and if you’re aware of what those errors are, then you can address them. But that also comes with experience, I suppose. So, if I plan a lesson and I see where they’ve made the errors, the next time I plan that lesson I might bring in different things to that lesson, bearing in mind what those errors were. So adapting and changing your lessons. You might even re-teach that lesson again. I also think that in the error analysis you found where their misconceptions were and then you had to find some way of addressing it, so the point of the lesson plan is to now go and see how you can practically correct these misconceptions…. The reported experiences cited above are interesting and certainly encouraging. Notwithstanding, they must be taken with caution and in relation to the quantitative analysis that follows. 5.3 Findings from the quantitative analysis On what criteria is the groups’ error analysis weaker and on what criteria is the groups’ error analysis stronger? Finding 1: Groups drew primarily on mathematical knowledge and less so on other possible explanations to explain the correct answer or the errors. When analysing the correct answer and the error, in about 70%-80% of the texts, groups drew primarily on mathematical knowledge and much less so, on other discourses. In both rounds the groups tried hard not to resort to common teacher talk on error such as test-related explanations (the learners did not read the question well, or the learners 100 guessed) or learner-related (the question is not within the learners’ field of experience) or curriculum-related (they have not learned this work). Interestingly, this finding is aligned with groups’ use of everyday knowledge in explaining errors: Despite the almost mandatory instruction by the NCS to link to everyday experiences when explaining mathematical concepts, the findings on this criterion point to a very different reality, with groups hardly referring to everyday contexts in their explanations of mathematical errors. The use of everyday as part of error explanations was minimal across all grades and all content areas. It was more evident in Round 1, when the groups worked with expert leaders, albeit in 12% of the texts only. Only this small number of texts in the sample demonstrates explanations of errors that make clear the link between an everyday phenomenon and the mathematical content of the item, and thus shows a use of ‘everyday’ that enables mathematical understanding. In Round 2 it was barely evident at all.28 Finding 2: Most of groups’ explanations of the correct solutions and of the error are incomplete, missing crucial steps in the analysis of what mathematics is needed to answer the question. Most of the explanations texts fall within the “incomplete” category, more so in Round 2. 57% of the procedural explanations of the correct answer, 45% of the conceptual explanations of the correct answer, and 50% of awareness of error explanations are incomplete. This is an important finding about the teachers’ knowledge. The evidence of teachers using predominantly incomplete procedural explanations is worrying. Incomplete procedural explanations of the mathematics involved in solving a particular mathematic problem, may impede on teachers’ capacity to identify mathematical errors, let alone to diagnose learners’ reasoning behind the errors. Finding 3: There is a correlation between groups’ procedural and conceptual explanations The correlation between procedural explanations and conceptual explanations was high in both rounds, although it decreased in Round 2 (r = 0,74 in Round 1 and r = 0,66 in Round 2). This suggests that when teachers are able to provide a full explanation of the steps to be taken to arrive at a solution they are also more aware about the conceptual underpinnings of the solution and vice versa. The weaker the procedural explanations are, the weaker the conceptual explanations, and vice versa. The higher correlation in Round 1 than in Round 2, between procedural and conceptual explanations suggests It is possible that this finding is specific to error analysis of learners’ response on (predominantly) multiple choice questions, where three feasible mathematical responses are given. More work is being done now to interrogate this finding, specifically, with regard to the teachers’ analysis of learner’s errors “own test”, which did not include any multiple choice questions, and during teaching. Further examination of this findings is done in relation to teachers’ ways of engaging with error, during teaching (referring to the 4th domain of teacher knowledge, Knowledge content and teaching (KCT) see table 1. 28 101 that when working with the group leaders teachers’ unpacked the concepts underlying the procedures more consistently. Much research in South Africa suggests that teachers use procedural and not enough conceptual explanations (Carnoy et al, 2011) and that this may explain learners’ poor performance. A new and important lesson can be learned from the correlation between procedural and conceptual explanations of the correct answer. The correlation suggests that there is conceptual interdependence between the procedural and conceptual aspects in teachers’ explanation of the mathematics that underlie a mathematical problem. When the quality of teachers’ procedural explanations goes higher, the quality of the conceptual explanations is also improved (and vice versa). This suggests that mathematical explanations of procedure and of concept rely on good subject matter knowledge of both, and a good procedural knowledge can help teachers in building their understanding of the underling concept. Instead of foregrounding the lack of conceptual understanding in teachers’ mathematical knowledge, more effort is needed to improve the quality of teachers’ procedural explanations, making sure that teachers are aware of which steps are crucial for addressing a mathematical problem and what counts as a full procedural explanation. (Examples of the range of procedural and conceptual explanations can be found in Appendix 6 and Appendix 7.) More research is needed to differentiate between strong and weak procedural explanations, in general and in different mathematical content areas. More research is needed to understand the quality of the relationship between full(er) procedural explanations and conceptual understanding. This is important for building data-base for teachers on crucial steps in explanations of leverage topics. We believe that this will help the effort of building mathematics knowledge for teaching. Finding 4: With more practice groups’ explanations demonstrated decreased inaccuracy, but at the expense of quality in other categories of explanations within the three criteria of Subject Matter Knowledge In all the Subject Matter Knowledge (SMK) explanations there was decreased inaccuracy (inaccurate explanations) in Round 2. In fact this seems to be the category of explanation in which they improved the most. This improvement needs to be understood relationally, in the context of the three remaining categories within each of the three SMK criteria. In what follows we look at these three criteria. Procedural explanation of the correct answer: There are very few inaccuracies when groups describe the steps learners need to take to arrive at the correct solution (8% of the texts in Round 1 and 5% of the text in Round 2). In Round 2 when the groups worked without leaders they kept a similar level of accuracy but when they were not sure about the steps needed to be followed to arrive at the correct solution 102 explanation, they did not mention procedure in their explanation (in Round 2 these explanations increased from 3% to 16%, a significant increase). The majority of correct answer procedural explanations in both rounds were stronger. The ratio between weaker and stronger procedural explanations of the correct answer in Round 2 was maintained in favour of stronger explanations, in particular of correct but incomplete explanations (57% of the procedural explanations of the correct answer in Round 2 were correct but incomplete). These findings point to groups engaging meaningfully with procedural explanations of the correct answer. That said, in Round 2 the number of full procedural explanations decreased (from 46% in round 1 to 23% in round 2). Taking these findings together, the following observation can be made: The significant decrease of full procedural explanations and the significant increase of explanations without any and/or with less information on procedure suggest that when groups worked without their group leaders their performance on procedural explanations of the correct answer was weaker. Conceptual explanations of the correct answer: Maintaining accuracy in conceptual explanations of the correct answer was more difficult for the groups. There was a significant decrease in inaccurate conceptual explanations of the correct answer between rounds (17% in Round 1 to 10% in Round 2). This change was at the expense of a significant increase in explanations which had no conceptual links (5% in Round 1 increased to 29% in Round 2). About 30% of the explanations in Round 2 offered no conceptual links (almost double of the equivalent procedural explanations category). Notwithstanding the quantitative difference, the trend between these two explanations is similar: In Round 2, when the groups worked without leaders, when they were not sure about the correct mathematical explanation, they tended to provide explanations without conceptual links. Similarly to procedural explanations of the correct answer, the majority of the conceptual explanations in both rounds were stronger. The ratio between weaker and stronger conceptual explanations of the correct answer in Round 2 was maintained in favour of stronger explanations. These findings point to groups engaging meaningfully with conceptual explanations of the correct answer. That said, in Round 2 the number of full conceptual explanations decreased (from 40% in Round 1 to 16% in Round 2). Taking these findings together, the following observation can be made: The significant decrease in full conceptual explanations, the insignificant improvement of partial conceptual explanations, and the overall increase of weak explanations in Round 2, suggest that when groups work without their group leaders their performance on conceptual explanations was weaker. It was also weaker than their performance on procedural explanations. In both rounds the groups started with 40% texts with full explanations (procedural and conceptual). The decrease in Round 2 in both of these explanations was significant, but higher in conceptual explanations. Whilst close to 60% of the procedural 103 explanations of the correct answer were accurate but incomplete, their parallel in conceptual was less than 50%. Awareness of error: In Round 2, when the groups worked without a group leader, they significantly reduced the number of inaccurate texts (or incomplete and hence potentially confusing explanations). Only 6% of Round 2 awareness of error explanations was inaccurate. This figure is consistent with the very low number of inaccurate procedural explanations of the correct answer in Round 2 (5%). As in the above subject matter knowledge-related explanations of the correct answer, groups decreased inaccuracies in their explanations of the error, which is an important improvement across these three criteria. On two categories of awareness explanations, the groups’ performance did not change: In both rounds, 20% of the texts offered no mathematical awareness of the error (this finding is consistent with the percentage of non-feasible explanations found in Round 1 and 2 texts, which was about 20%, see below). Second, in both rounds the number texts that demonstrated a full awareness of errors (in which the explanation of the error is mathematically sound and suggests links to common errors) remained at about 25% of the total error-related texts. Other than reduction of inaccuracy, a notable change in Round 2 was found in the 3rd category of this criterion: In 50% of the error texts in Round 2, the explanation of the error was mathematically sound but did not link to common errors. Taken together, the following observation can be made: when the groups worked without leaders, the ratio between stronger and weaker explanations remained in favor of stronger awareness of error in their explanations. The percentage of stronger explanations improved mainly due to the increase to 50% of the 3rd category explanation (see above). Together with the significant decrease of mathematically inaccurate texts, the groups maintained their performance in Round 2. Finding 5: There is a correlation between groups’ awareness of error and their diagnostic of learners’ reasoning behind the error The correlation between awareness of the mathematical error and diagnostic reasoning was also high and increased in Round 2 (r = 0,651 in Round 1 and r = 0,71 in Round 2).When groups demonstrate high awareness of the mathematical error (SMK) they are more likely to give the appropriate diagnosis of the learner thinking behind that error (PCK). The correlation between awareness and diagnostic reasoning merits reflection. The lesson it provides is that when teachers can describe the error, mathematically well (SMK), they are more likely to be able to delve into the cognitive process taken by the learners and describe the reasoning that led to the production of the error (PCK). 104 Although this is only a correlational finding, tentatively we suggest that improving teachers’ mathematical awareness of errors could help improve teachers’ diagnostic reasoning. Furthermore in view of the finding that teachers’ procedural explanation of the mathematics underlying the correct answer is itself weak (missing steps, incomplete), we suggest that the finding that the teachers struggled to describe the mathematical way in which the learners produced the error, is expected. All of this has implications for the relationship between SMK and PCK. The correlation between awareness and diagnostic reasoning, and between procedural and conceptual knowledge, bring the importance of subject matter knowledge in teaching. Finding 6: Groups struggled to describe learners’ reasoning behind the error Close to 30% of the texts in both rounds did not attempt to explain learners’ error and another 27% of the texts in both rounds described learners’ reasoning without honing in on the error. Altogether more than 50% of the texts in both rounds demonstrated weak diagnostic reasoning. About 33% of the texts in both rounds honed on the error but the description of learner reasoning was incomplete. In both rounds then, in about 90% of the texts, groups offered no or incomplete explanation of learners’ reasoning. Only 12% of texts in each of the rounds were systematic and honed in on the learners’ reasoning about the error. This is the criterion in which groups’ performance was the weakest, more so (albeit, insignificantly) in Round 1, and proportionally more in the Grade 3-6 group. As in awareness of error explanations, the groups performed better on texts that do hone in on the error but are incomplete. The weakness in explaining learners’ reasoning is also evident and is consistent with the difficulty of the groups to produce more than one explanation to explain the error. All groups provided mainly one mathematically feasible/convincing explanation (with/without general explanations). 60% of the items in Round 1 and 85% of the items in Round 2 demonstrate little use of multiple explanations. These two findings suggest that groups struggle to think about the mathematical content covered by the item from the perspective of how learners typically learn that content. According to Ball et al (Ball, Thames and Bass and Ball, 2011), this type of thinking implies “nimbleness in thinking” and “flexible thinking about meaning”. The groups’ difficulty to think meaningfully about rationales for the ways in which the learners were reasoning and their inability, even in a group situation, to think of alternative explanations is evident to them lacking these two qualities in the way they approach errors. Finding 7: In Round 2 across all criteria the number of texts without a mathematical explanation increased while the number of inaccurate texts decreased The number of texts with mathematically inaccurate explanations generally decreased in Round 2, which suggests that just by doing more items, even without leaders, the likelihood of producing explanations which were not mathematically flawed improved. 105 Bearing this in mind, it was interesting to note that while inaccurate texts decreased in number between Rounds 1 and 2, texts that did not include a mathematical explanation increased. This could be an indication that the group leaders seem to be more useful, albeit not significantly so, in enhancing awareness of the mathematical concepts and in focusing the explanations on the mathematical content of the question since in Round 2 when they were absent, a higher number of explanations that did not have mathematical substance were given. Is there a difference between primary (Grade 3-6 groups) and high school (Grade 7-9 groups) teachers in sustaining the performance on the Round 1 error analysis activity? Finding 8: The Grade 7-9 group held better quality when working without group leaders than the Grade 3-6 group An examination of the different performance by the two sets of grouped grades in the weaker and stronger level criteria both point to the better quality held by the Grade 7-9 group when they worked without a group leader in Round 2. One particularly notable change was that in Round 2 on the diagnostic reasoning criteria there was a high decrease in inaccurate mathematical explanations and a corresponding high increase in correct though partially complete mathematical explanations for the Grade 7-9 group. This is very different from the general pattern where the decrease in inaccurate explanations was seen in conjunction with an increase in general texts. This suggests that the learning in this group was strongest in relation to the diagnosis of error. Finding 9: Teachers reasoning in relation to mathematical concepts and errors seems to be strongest in the content area of number while weakness is spread across the areas of data, shape, algebra and measurement Number is the content area which is most often taught in schools and so this finding corresponds with the knowledge expected of teachers. It is interesting to note that diagnostic reasoning was strongest in the area of measurement and that most multiple explanations were offered in the area of shape. Use of the everyday in explanations was done very little, but the two content areas in which most (albeit very little) reference to the everyday in explanations were number and measurement. These are findings that could be further investigated. Data is a relatively “new” content area in the South African mathematics curriculum which could explain the weakness here while algebra and shape are recognized internationally as more difficult mathematical content areas. 106 What does the change in performance between Round 1 and 2 suggest about the role of group leaders? Finding 10: Group leaders are important When worked with leaders, groups: completed more items, provided more full explanations and less partial explanations, provided more conceptual explanations and less procedural explanations, provided more mathematically focused explanations and less “general texts” types of explanations, unpacked the concepts underlying the procedures more consistently, and gave more multiple explanations of errors. 107 108 5.4 Summary In what follows we summarise the main findings of the quantitative analysis and draw conceptual implications for the construct we propose to frame the idea of teacher’s thinking about error- “diagnostic judgment”. In both Rounds (more so in Round 2), groups tended to give mathematically accurate but incomplete procedural and/or conceptual explanations of the correct answer and/or of the error (SMK-related criteria). This means that groups were able to describe, albeit, incompletely some of the steps that learners should follow and some of the conceptual links they need to understand to arrive at the solution, and when the groups identified the error, their level of mathematical accuracy was high. Only in very few texts were the groups’ procedural and conceptual explanations of the correct answer found to be inaccurate. This needs to be taken together with groups being found to give more “stronger” than “weaker” explanations and maintaining this strength in Round 2. This suggests that generally teachers’ did not mis-recognise the procedure or the concept the correct answer consists off, and they could recognise an error. Nevertheless, in all these three acts of recognition, many of their explanations is incomplete and some were inaccurate. The implication we draw from this is that in order for teachers to be able to improve their error recognition and their explanations of the correct answer, they need to develop their content knowledge so that they will be able to produce fuller explanations. To repeat, the distinctive feature of teachers’ knowledge, which distinguishes them from other professionals who need mathematics for their work is that “teachers work with mathematics in its decompressed or unpacked form” (Ball, Thames and Bass, 2008b, p400). When teachers’ subject matter knowledge is strong they will have acquired an ability to judge when and how to move from compressed to unpacked mathematical knowledge, and how to provide explanations that is both accurate and full. What this study of error analysis suggests is that teachers’ when explanations are basically correct their incomplete form may cause confusion. Incomplete explanations, we suggest signal weak content knowledge. In the following example we give two explanations given for a grade 8 question. The first explanation is of the correct answer and it is an example of a basically correct but incomplete explanation. The second explanation is of the error. It is both inaccurate and confusing. 109 Table 26: An example of an explanation that is accurate but incomplete ICAS 2006, Grade 8, Question 10 Correct answer: B Selected distractor: D Content Area: Geometry – Rotational symmetry Teacher explanation of the correct answer By removing the star in the centre all the other stars would still be symmetrical. Limitations of explanation of the correct answer This explanation is a typical compressed explanation where the concept is assumed not elaborated. Furthermore what is included by way of explanation is that the removal of the centre star is the correct solution but is not an explanation. Teachers should explain that learners need to visualise rotation to determine the centre is not affected by rotation and therefore it can be removed. This could be based on a broader explanation of the meaning of rotational symmetry so that the explanation is generalisable. Teacher explanation of the error The learners could have imagined an axis of symmetry between A and B. they could have then counted out the number of stars on either side which totals five stars, therefore D is out. Limitations of explanation of the error This explanation further demonstrates the lack of knowledge of the concept - rotational symmetry. The counting of the stars on either side of a line relates to line symmetry and not rotational symmetry. The position of the imagined line of symmetry (which would not even be used to find the rotational symmetry) is not well described Teachers particularly struggled to try and imagine learners’ reasoning, to offer variety of explanations, and to meaningfully reflect on the everyday context of questions and link this to the mathematical context of questions. Groups’ judgment about learners’ thinking (PCK) was very weak and their ability to approach leaners’ thinking in diversified ways was also weak. So despite demonstrating a very small percentage of inaccurate explanations, the aspect of subject matter knowledge that came to the fore in this study is the incompleteness in groups’ explanations of the error (awareness of error), which 110 seems to correlate with explanations that hone in on learners’ thinking (diagnostic reasoning). This suggests a relationship of dependence between: what teachers can do creatively for learners (explain ideas in a differentiated way, connect between every day and mathematical contexts of questions, PCK) and imagine the ways learners think (PCK) and the quality of the subject matter knowledge they acquire and develop. In the following example we give two explanations given for a grade 5 question. In this example both explanations, of the correct answer and of the error, demonstrate insufficient content knowledge. In the case of the error text this leads to an inability to make a full diagnostic judgement. Table 27: Grade 5 test item explanations illustrating poor diagnostic judgement ICAS 2007, Grade 5, Question 16 Correct answer: B Selected distractor: D Content Area: Numbers and Operations – completing an equation Teacher explanation of the correct answer The learner could have subtracted 6 from 21 and divided by 3. Limitations of explanation of the correct answer This explanation offers no conceptual links. Conceptually an understanding of the equality of the two sides of an equation underlies the solution. Teacher explanation of the error The learner added 6 and 21 and got 27 and divided 27 by 3 to get 9 Limitations of explanation of the error Procedurally this could be exactly what a learner may have done to reach this incorrect solution but the explanation does not hone in on the misconception that may have led to such procedural activity. This misconception relates to a lack of understanding of the equality of the two sides of an equation the learners reposition the equal sign without adjusting the operations accordingly. 111 Notwithstanding, three strengths of the project can be recognised in the comparison between Rounds 1 and 2: The main one being that in Round 2, groups gave fewer explanations which were mathematically inaccurate. This is an important achievement that needs to be recognised. In Round 2, when groups worked without a group leader they managed to decrease their inaccuracy across the different criteria. As discussed before, this improvement within the weaker category of explanations often came at the expense of an increase in texts that did not include mathematical explanations relevant to the criteria. We propose that having participated other DIPIP project activities and listened to feedback, groups were more cautious, and as a result, when they were not sure (and having no group leader to lean on), they did not give mathematical explanations. Secondly, important to acknowledge as an achievement, that when analysing the correct answer and the error, in about 70%-80% of the texts, groups drew primarily on mathematical knowledge and much less so, on other discourses. Thirdly, in Round 2, in the three SMK-related criteria, most of the explanations remained within the two stronger categories. 112 Section Six: Implications for professional development: developing diagnostic judgement 113 The central predicament of audit culture is that it produces massive amount of data, which is good for “external accountability” but often remains remote from “internal accountability”. Elmore who made this distinction notes the following: Internal accountability precedes external accountability. That is, school personnel must share a coherent, explicit set of norms and expectations about what a good school looks like before they can use signals from the outside to improve student learning. Giving test results to an incoherent, atomized, badly run school doesn't automatically make it a better school. The ability of a school to make improvements has to do with the beliefs, norms, expectations, and practices that people in the organization share, not with the kind of information they receive about their performance. Low-performing schools aren't coherent enough to respond to external demands for accountability … Low-performing schools, and the people who work in them, don't know what to do. If they did, they would be doing it already. You can't improve a school's performance, or the performance of any teacher or student in it, without increasing the investment in teachers' knowledge, pedagogical skills, and understanding of students. This work can be influenced by an external accountability system, but it cannot be done by that system. (2002, 5-6) In other words, the important and very complicated condition for achieving the link between teachers’ tacit knowledge, their motivation to change and performance is teachers' belief about their competence. This belief is linked directly to their learning experiences, to meaningful opportunities to learn what specific curriculum standards actually mean. The argument about internal accountability, first, is that with sufficient meaningful professional development, teachers can be helped to see that their efforts can bear fruits (Shalem, 2003). Teachers need to be able to see the reasons for change, understand its core principles and be convinced that it is feasible and will benefit their learners. This gives rise to the question of what constitutes a meaningful learning opportunity for teachers. “Alignment” between formative and summative assessments is the notion which the international literature uses to describe the reciprocity that is needed for teacher learning. International research shows that just having another set of data (in the form of benchmarking, targets and progress reports) that ‘name and shame’ schools leads to resentment and compliance but not really to improvement of learning and teaching (McNeil, 2000; Earl & Fullan, 2003; Fuhrman & Elmore, 2004). In South Africa, Kanjee (2007) sums up the challenge: For national assessment studies to be effectively and efficiently applied to improve the performance of all learners, the active participation of teacher and schools is essential. … Teachers need relevant and timeous information from national (as well as international) assessment studies, as well as support on how to use this information to improve learning and teaching practice. Thus a critical challenge would be to introduce appropriate polices and systems to disseminate information to teachers. For example, teacher-support materials could be developed using test items administered in national assessments. (p. 493) 114 Katz et al (2005) draw an important distinction between “accounting” and “accountability” (which follows on from Elmore’s distinction between external and internal accountability). They define the former as the practice of gathering and organising of data for benchmarking, which is the mechanism that the Department of Education has put in place in order to ensure the public that it gets a good service for its investment. They define the latter as “teachers-led educational conversations” about what specific data means and how it can inform teaching and learning (Earl & Fullan, 2003; Earl & Katz, 2005, Katz et al, 2009). In this view, the international and local benchmarking tests mentioned above form an “accounting” feature of the South African educational landscape. Katz et al want to re-cast “accountability”. They want to recast it from its reporting emphasis to making informed professional judgement, where judgment is constructed through data-based conversations on evidence that is gathered from systematic research and from assessments. They argue: The meaning that comes from data comes from interpretation, and interpretation is a human endeavour that involves a mix of insights from evidence and the tacit knowledge that the group brings to the discussion. ... This is at the heart of what’s involving in determining a needs-based focus. The process begins by tapping into tacit knowledge, by engaging in a hypothesis generation exercise and recasting “what we know” into “what we think we know”... Instead of suspending beliefs in the service of data or adamantly defending unsubstantiated beliefs, the conversation is a forum for examining both and making the interrelationship explicit. (Katz et al 2009, p28) “Accountability” as phrased here emphasizes two ideas. First, in order for teachers to learn from evaluation data what their learners “can and/or cannot do at a particular stage or grade”29, they need to be engaging their tacit knowledge in some or other of a structured learning opportunity (“recasting ‘what we know’ into ‘what we think we know’). Secondly, accountability (in the above sense) constitutes teachers reflection as a shared process, as a structured conversation between professionals. And so, since these conversations are structured conversations, “accountability” enables teachers to hold themselves and each other “to account” by explaining their ideas and actions, to each other, in terms of their experiences as well as their professional knowledge (Brodie & Shalem, 2010). In this way, Katz et al argue, accountability conversations can give participants imagination for possibilities that they do not yet see and help them making tacit knowledge more explicit and shared. 30 In our project, we used the evaluation data of proficiency results on the ICAS test as an artefact around which to structure a learning opportunity to experience a process of diagnostic assessment on learners’ error in mathematic evaluation data. Media statement issued by the Department of Basic Education on the Annual National Assessments (ANA): 04 February 2011. http://www.education.gov.za/Newsroom/MediaReleases/tabid/347/ctl/Details/mid/1389/ItemID/3148/Default.aspx 29 30 For the idea of “Accountability conversations” see Brodie & Shalem, 2011. 115 Research has only recently begun to engage with the question of how to use the data beyond that of an indicator of quality i.e. beyond benchmarking (Katz, Sutherland, & Earl, 2005; Boudett, City & Muuname, 2005; Katz, Earl, & Ben Jaafar, 2009). Some conceptual attempts to examine a more balanced way between accounting and accountability include Shavelson et al (2002)31, Nichols et al (2009)32, and Black and Wiliam (2006). In the South African context, Dempster (2006) and Dempster and Zuma , (2010), Reddy (2005) and Long (2007) have each conducted small case studies on testitem profiling, arguing that this can provide a useful data that can be used for formative and diagnostic purposes. Arguably, awareness about learners’ errors is a useful step in the process of significantly improving practice. Hill and Ball (2009) see “analyzing student errors” as one of the four mathematical tasks of teaching “that recur across different curriculum materials or approaches to instruction (p70)33. Studies on teaching dealing with errors show that teachers’ interpretive stance is essential for the process of remediation of error, without which, teacher simply re-teach without engaging with the mathematical source of the error, or with it metacognitive structure (Peng, 2009, 2010; Prediger, 2010; Gagatsis & Kyriakides, 2000). Shavelson et al devised what they call “a logical analysis” that test-designers or teachers can use in order to “psychologize about the nature of the problem space that a student constructs when confronted with an assessment task” (2002, p15). They argue that effective error analysis should use a “logical analysis” of task demands together with “empirical analysis” of the kinds of cognitive processes a task, has in fact, evoked by students. In more common terms the two steps include tasks analysis followed by an error analysis. In constructing the process of “logical analysis”, Shavelson et al used a coding scheme consisting of numerous categories that together make up 4 types of knowledge- “declarative” (knowing that) “procedural” (knowing how) , “schematic” (knowing why) and “strategic” knowledge (knowing when knowledge applies). These were used to align the type of knowledge intended for Science tasks or their “construct definitions” with the observations made about “the student’s cognitive activities that were evoked by the task as well as the student’s level of performance” (p6). Shavelson et al nest this kind of analysis within what they call as “the assessment square”, a framework that they develop to increase the alignment between the type of knowledge structure to be measured (the 4 types of knowledge stated above), the kind of task designed for measuring learners’ performance of the knowledge structure (including the design features of different tasks), the variety of responses elicited by the learners, and the inferences made on the basis of these analyses. This type of analysis is useful for task analysis; it helps to unpack the cognitive demands embedded in tasks, their degree of openness and level of complexity. 32 In their work, Nichols et al attempt to create a framework that will link between information from performance on a particular assessment and student learning. Their model suggests several interpretive acts build in a sequence that connects between three main data “structures”: “student data” or sets of data derived from systemic assessment, “knowledge domain model” or the knowledge and skills associated with a learning construct, “a student model” or the representation of current understanding of a learner’s or specific class of learners’ knowledge and “a teaching model” or a sets of methods that have been shown to be effective in relation to the other domains (the knowledge and the student domain). What holds these structures together and gives them meaning are the recurring interpretive acts required by the teacher: Data interpretation involves reasoning from a handful of particular things students say, do, or make in particular circumstances, to their status on more broadly construed knowledge, skills, and abilities that constitute the student model. (p17) 33 The others being: “encountering unconventional solutions, choosing examples, or assessing the mathematical integrity of a representation in a textbook”. (ibid) 31 116 In South Africa, Adler sees teachers’ knowledge of error analysis as a component of what she calls mathematics for teaching (2005, p3). She asks: What do teachers need to know and know how to do (mathematical problem solving) in order to deal with ranging learner responses (and so some error analysis), and in ways that produce what is usefully referred to as “mathematical proficiency”, a blend of conceptual understanding, procedural fluency and mathematical reasoning and problem solving skills? (ibid) From this point of view, teachers’ diagnostic judgment is essential; it is used to make decisions which affirm learners’ performance, but also to make decisions which classify learners and select them - or not - for particular futures. What does DIPIP phases 1&2 research suggest diagnostic judgment entails? Diagnostic Judgement Prediger’s argument that teachers who have an interest in learners’ rationality are aware of approaches to learning implies that diagnostic judgment encompasses far more than reasoning about the way in which learners think mathematically when they answer mathematics test questions. This implies that diagnostic judgement should include understanding about learners, such as their contexts, culture, learning styles and possible barriers to learning. It also implies that diagnostic judgement will look different in different teaching practices, for example during teaching, when planning a lesson planning, and when making assessments. The data of this project suggests that error analysis of mathematical evaluation data relies on a firm SMK- specifically correctness and fullness of knowledge of the steps to be followed to arrive at the solution and the most relevant concept underling the question. The distinction between accuracy and fullness is important. Whilst a general mathematician can figure out an accurate solution with little attention to the number, kind and sequence of steps to be followed, for a teacher this is different. Teachers need to know both- they need to be able to distinguish between explanations that are correct but are not completely full from those that are correct and full. Teachers need to unpack what is compressed behind the correct answer or behind an error. This has to do with their role in socialising learners into doing mathematics. Unlike the general mathematician or other professionals, teachers need to show learners which explanation is better and why, what steps are there to be followed, in what sequence and why. As the analysis of teacher knowledge suggests, without a full knowledge of the explanation, teachers may recognize the error only partially, as they may not know what the crucial steps that make up the solution are or their sequence and the conceptual relation between them. Without thorough knowledge of the mathematical content and procedures relating to that content (Common Content Knowledge) it is unlikely that teachers will be able to systematically look for “a pattern in students error” or “size up 117 whether a non-standard approach would work in general” (Specialized Content Knowledge , see Ball, Thames &Bass, 2008, p400). The study shows that there is a conceptual interdependence between the procedural and conceptual aspects in teachers’ explanation of the mathematics that underlie a mathematical problem. It also suggests that there is interdependence between awareness of the mathematical error and the quality of teachers’ diagnostic reasoning. Together this means that diagnostic reasoning is dependent on prior content knowledge, that following the learner’s reasoning is dependent on teachers having a full picture of what the question is in fact asking, how is it asking it, what is the underlying concept of the question, what kind of procedure it relies on and what are the crucial steps that have to be completed. Only with this knowledge, contextual factors about learners’ misrecognition can add value to how teachers are analysing learners’ errors. So we propose a vertical form construct to describe teachers’ knowledge of errors: When teachers can describe the steps and the underlying concepts of a question in full (or mathematically well), they are more likely to be able to size up the error or interpret its source (SMK). When teachers can describe the error, mathematically well (SMK) they are more likely to be able to delve into the cognitive process taken by different kinds of learners and describe the reasoning that led to the production of an error (PCK). The idea of developing teachers’ diagnostic Judgement is implied by other research in South Africa and internationally The teacher today is required to demonstrate different types of professional knowledge which range from knowledge of subject matter to recognizing that an answer is wrong, or that a particular mathematical solution cannot be accepted because it cannot work in general, to much more subtle kinds of knowledge such as 'knowledge of self'. This kind of teacher is regarded a specialist for being able to teach subject matter knowledge with attunement to the curriculum, to the knowledge field and to the learner. This kind of attunement requires a distinct mode of awareness that is often informed by dispositions and perspectives that are not readily accessible to teachers in terms of their pre-service training, more so in South Africa (Adler & Hulliet, 2008, p19). This is consistent with the works of Nesher (1987), Borasi (1994), Gagatsis and Kyriakides (2000), Kramarski and Zoldan (2008), Predriger (2010) which show that a focus on learner’s mathematical thinking when producing errors provides a mechanism for teachers to learn to think more carefully about their teaching and for learners to change their attitudes towards errors (Borasi, 1994, p189). Borasi argues that experiencing learning through errors shifted learners’ conception about the use of errors which until that point were seen solely as mistakes that need to be remediated. In terms of teaching, Borasi argues, errors are “a valuable source of information about the learning process, a clue that researchers and teachers should take advantage of in order to uncover what students really know and how they have constructed such knowledge” (p170). 118 The development of “formative assessment” opens up new possibilities for teachers with regard to this. The notion of formative assessment has changed the pedagogical role of teachers in assessment. The following definition proposed by Black and Wiliam, explains the idea of “formative”, grounding it in awareness of learners’ errors. Practice in the classroom is formative to the extent that evidence about student achievement is elicited, interpreted, and used by teachers, learners, or their peers, to make decisions about the next steps in instruction that are likely to be better, or better founded, than the decisions they would have taken in the absence of the evidence that was elicited. (2009, p9) At the core of this view is a pedagogical practice that consists of a sequence of activities structured by the teacher in response to evidence s/he gathers about students’ learning from classroom and/or systemic assessment. This also includes classroom discussions amongst learners and their peers and amongst teachers on the meaning of evaluative criteria, on errors that appear to be common, and on ways of addressing them (See also Brookhart, 2011). Having these kinds of engaging discussions, Black and Wiliam argue, will promote self-regulation and autonomy in students and “insight into the mental life that lies behind the student’s utterances” in teachers (2009, p13). Nichols et al (2009) offer the idea of “a reasoned argument” which they believe is the key to a successful process of for formative assessment. For them formative assessment is valid, when it is used to inform instruction and its use to inform instruction is based on a valid interpretation: The claim for formative assessment is that the information derived from students’ assessment performance can be used to improve student achievement. It is how that information is used, not what the assessment tells us about current achievement, that impacts future achievement. Therefore, use, based on a valid interpretation, is the primary focus of the validity argument for formative assessments. (our emphasis, p15) Nichols et al look for a model of formative assessment that can be trusted as valid. They look for a model that will ensure that when teachers make judgments about learners’ reasoning behind a mathematical error (about why the learners think in that way, and how their thinking can be addressed in teaching), they can account for the way they reasoned about the evidence, and can show coherently how the evidence informed their instructional plan. Their model was criticized strongly by Shepard 2009, p34) and for good reasons, what is relevant for us, however, is the idea of “reasoned argument”, of teachers needing to ask and interpret “what does the evidence from student data say about students’ knowledge, skills and abilities” (p.16)? This idea is consistent with Prediger’s idea of “diagnostic competence” (Teachers analyze and interpret learner reasoning), Kats’ et al notion of “accountability” (ability to account for), and with Black’s and Wiliam’s idea of formative assessment (evidence about student achievement is elicited and interpreted). Together, they support the notion of we would like to call as “diagnostic judgment”. To move away from the instrumentalist use of the current audit 119 culture, of a culture that uses evaluation data to mainly benchmark learner performance (and to blame and shame teachers), professional development of teachers need to work out ways to develop in teachers diagnostic judgment. In this we mean: the ability to find evidence of learning and to know how to work with that evidence. As professionals and as specialists of a particular subject matter, teachers need to know how to elicit and reason about evidence of learners’ errors. Lessons about error analysis of mathematical evaluation data that can be taken from DIPIP Phases 1 and 2: The last question we need to address, is given these results, what could DIPIP have done differently? Here we refer to the process document which describes the linear thinking that informed the design of its main activities. The groups followed the error analysis with two different applications- lesson plan and teaching. All the groups worked on their lesson plans for several weeks and repeated it. Then all groups designed a test and interviewed some learners on errors identified in their tests. Finally eleven groups continued into a second round of curriculum mapping and error analysis while some groups went on to a third round.. The idea behind this design sequence was that once the groups mapped the curriculum and conducted error analysis, they will recontextualised their learning of the error analysis into two core teaching activities. In other words, they will start with “error-focused” activities and move to “error-related” activities and then go back to error-focused activities. This will enable them to be more reflective in their lesson design and in their teaching as in both of these activities the groups will work with what the curriculum requires (assessment standards) but also with errors to be anticipated in teaching a particular mathematical topic. This design sequence was also informed by ideas from the field of assessment, in particular the emphasis on integration between teaching and assessment - the key construct in the idea of alignment between standardised and formative assessment. The above analysis of teachers’ thinking about learners’ errors in evaluation tests suggests that the idea of sequential development between these activities was too ambitious. To ask teachers to apply their error knowledge which at most was only partial in terms of its SMK aspect and very weak in terms of its diagnostic of learners reasoning was certainly too ambitious, at least for teachers in SA. Teachers needed repeated practice in “error-focused activities” before they can be expected to transfer this knowledge into Knowledge of content and teaching (KCT), Ball et al’s 4th domain of teacher knowledge (or second domain within PCK). The idea was ambitious in two ways. First teachers needed more repetitive experience in doing error analysis. Error analysis of evaluation data is new for teachers and they needed to develop their competence in doing more of the same. They also needed a variation of questions. There are pros and cons in working on multiple content questions and groups needed to work with a variety of questions. Secondly and even more importantly, our analysis shows 120 that teachers need to work on errors in relation to specific mathematical content. This is consistent with Cohen and Ball (1999) who argue that change of teaching practice comes as support for a certain content area in the curriculum and not vice versa. Teachers cannot change their teaching practices in all the content areas they are teaching (Cohen and Ball, 1999, p.9).34 If accepted, the conception of diagnostic judgement that we propose shows that teachers need to know the content well and think about it carefully when they are working on errors. 34 Cohen, D. K. and Ball, D. L. (1999) Instruction, Capacity and Improvement CPRE Research Report Series RR-43. 121 122 Recommendations for professional development and for further research DIPIP Phases 1 and 2 set out to include teachers in different professional activities all linked to evaluation data and focused on errors. It had a development and a research component. In this report we evaluate the error analysis activity and not the whole process of development this is done in the process report. Based on the experience of this project activity we make the following recommendations with regard to teacher development and further research. Recommendations for professional development Error analysis activities should be content-focused. Different error-focused activities should be used and coherently sequenced. These should include error analysis in relation to different forms of mathematical questions (open and multiple choice); the use of systemic tests and tests set by the teachers; design of different kinds of mathematical questions, interviewing learners and design of rubrics for evaluation of learners. Error analysis should be done in conjunction with curriculum mapping as this helps teachers to understand the content and purpose of questions. We suggest that this should be done with a specific content focus. Error analysis activities should be situated within content development programmes. Structured content input is necessary for teachers in South Africa since this is a prerequisite for diagnostic judgement. In order to avoid compromising on quality modeling of what counts as full explanations (including all procedural steps and conceptual links) is necessary. At all times and through all activities the emphasis on full explanations must be made explicitly which we believe will strengthen teachers’ content knowledge. Learning materials for content-focused professional development of error analysis should be developed. This was one of the aims of DIPIP 1 and 2. Expert leadership of groups working on error analysis activities is important to enable quality instruction of mathematical content, promote depth of analysis and fullness and variety of explanations. Never-the-less judgement should be made as to when is the appropriate time to remove this expert leadership. Recommendations for further research More research is needed to identify different qualities of mathematical content specific procedural explanations. This could inform the teaching of mathematical content. Further research could investigate other discourses teachers draw on when they analyse error on a standardized test and on their own tests. 123 A more in-depth study could be done on the use of everyday contexts in the explanation of mathematical concepts and learners’ errors. The role of expert leaders in supporting teacher development through error analysis needs further investigation. More research is needed to inform ways of aligning evaluation data results and teachers’ practice. 124 References Adler, J (2005). Mathematics for teaching: what is it and why is it important that we talk about it? Pythagoras, 62 2-11. Adler, J. & Huillet, D. (2008) The social production of mathematics for teaching. In T. Wood S. & P. Sullivan (Vol. Eds) International handbook of mathematics teacher education: Vol.1. Sullivan, P., (Ed.), Knowledge and beliefs in mathematics teaching and teaching development. Rotterdam, the Netherlands: Sense Publishers. (pp.195222). Rotterdam: Sense. Ball, D.L. (2011, January) Knowing Mathematics well enough to teach it. Mathematical Knowledge for teaching (MKT). Presentation for the Initiative for Applied Research in Education expert committee at the Israel Academy of Sciences and Humanities, Jerusalem, Israel, January 30, 2011. Ball, D.L., Hill, H.C., & Bass, H. (2005) “Knowing mathematics for teaching: who knows mathematics well enough to teach third grade, and how can we decide?” American Educator, 22: 14-22; 43-46. Ball, D.L., Thames, M.H., & Phelps, G. (2008) “Content knowledge for teaching: what makes it special?” Journal of Teacher Education 9 (5): 389-407. Borasi, R. (1994) “Capitalizing on errors as “springboards for inquiry”: A teaching experiment in Journal for Research in Mathematical Education, 25(2): 166-208. Boudett, K.P., City, E. & Murnane, R. (2005) (Eds) Data Wise A Step-by-Step Guide to Using Assessment Results to Improve Teaching and Learning. Cambridge: Harvard Education Press Black, P. & Wiliam, D. (2006) Developing a theory of formative assessment in Gardner, John (ed.) Assessment and Learning (London: Sage). Brookhart S M (Ed) (2009) Special Issue on the Validity of Formative and Interim assessment, Educational Measurement: Issues and Practice, Vol. 28 (3) Brodie, K. (2011) Learning to engage with learners’ mathematical errors: an uneven trajectory. In Maimala, T and Kwayisi, F. (Eds) Proceedings of the seventeenth annual meeting of the South African Association for Research in Mathematics, Science and Technology Education (SAARMSTE), (Vol 1, pp. 65-76). North-West University, Mafikeng: SAARMSTE. Brodie, K. & Berger, M (2010) Toward a discursive framework for learner errors in Mathematics Proceedings of the eighteenth annual meeting of the Southern African Association for Research in Mathematics, Science and Technology Education (SAARMSTE) (pp. 169 - 181) Durban: University of KwaZulu-Natal Brodie, K. & Shalem, Y. (2011) Accountability conversations: Mathematics teachers’ learning through challenge and solidarity, Journal of Mathematics teacher Education 14: 420-438 125 Dempster, E. ( 2006). Strategies for answering multiple choice questions among south African learners: what can we learn from TIMSS 2003, in Proceedings for the 4th Sub-Regional Conference on Assessment in Education, UMALUSI. Dempster, E. & Zuma, S. (2010) Reasoning used by isiZulu-speaking children when answering science questions in English. Journal of Education 50 pp. 35-59. Departments of Basic Education and Higher Education and Training. (2011). Integrated Strategic Planning Framework for Teacher Education and Development in South Africa 2011-2025: Frequently Asked Questions. Pretoria: DoE. Earl, L & Fullan M. (2003) Using data in Leadership for Learning (Taylor &Francis Group: Canada) Earl, L., & Katz, S. (2005). Learning from networked learning communities: Phase 2 key features and inevitable tensions. http://networkedlearning.ncsl.org.uk/collections/network-researchseries/reports/nlg-external-evaluation-phase-2-report.pdf. Elmore, RE. 2002. Unwarranted intrusion. Education next. [Online]. Available url: http://www.educationnext.org/20021/30.html Fuhrman SH & Elmore, RE (2004) Redesigning Accountability Systems for Education. (Teacher College Press: New York). Gagatsis, A. and Kyriakides, L (2000) “Teachers’ attitudes towards their pupils’ mathematical errors”, Educational Research and Evaluation 6 (1): 24 – 58. Gipps, C. (1999) “Socio-cultural aspects of assessment”. Review of Research in Education 24. pp 355-392. Heritage, M., Kim, J., Vendlinski, T., & Herman, J. (2009), From evidence to action: A seamless process in formative assessment? Op cit pp. 24-31. In S. Brookhart (Ed) (2009) Special Issue on the Validity of Formative and Interim assessment, Educational Measurement: Issues and Practice, Vol. 28 (3) Hill, H.C., Ball, D.L., & Schilling, G.S. (2008a) “Unpacking pedagogical content knowledge: conceptualizing and measuring teachers’ topic-specific knowledge of students” Journal for Research in Mathematics Education 39 (4): 372-400. Hill, C. H., Blunk, M.L., Charalambos, C.Y., Lewis, J.M., Phelps, G.C., Sleep, L &Ball, D.L (2008b) Mathematical knowledge for teaching and the mathematical quality instruction: an exploratory study. Cognition and Instruction 26 (4) 430-511. Hill, H.C. & Ball, D.L. (2009) “ The curious- and crucial- case of mathematical knowledge for teaching Kappan 9 (2): 68-71. Katz, S., Sutherland, S., & Earl, L. (2005). “Towards an evaluation habit of mind: Mapping the journey”, Teachers College Record, 107(10): 2326–2350. Katz, S., Earl, L., & Ben Jaafar, S. (2009). Building and connecting learning communities: The power of networks for school improvement. Thousand Oaks, CA: Corwin. 126 Kramarski, B. & Zoldan, S. (2008). Using errors as springboards for enhancing mathematical reasoning with three metacognitive approaches. The Journal of Educational Research, 102 (2), 137-151. Louw, M. & van der Berg, S & Yu, D. (2006) "Educational attainment and intergenerational social mobility in South Africa Working Papers 09/2006, Stellenbosch University, Department of Economics Looney, J. W. (2011). Integrating Formative And Summative Assessment: Progress Toward A Seamless System? OECD Education Working Paper No. 58 EDU/WKP 4 http://essantoandre.yolasite.com/resources.pdf Long, C. (2007) What can we learn from TIMSS 2003? http://www.amesa.org.za/AMESA2007/Volumes/Vol11.pdf McNeil, LM. (2000) Contradictions of school reform: Educational costs of standardized testing. New York: Routledge. Muller, J. (2006) Differentiation an Progression in the curriculum in M. Young and J. Gamble (Eds). Knowledge, Curriculum and Qualifications for South African Further Education. Cape Town: HSRC Press. Nesher, P. (1987) Towards an instructional theory: The role of learners’ misconception for the learning of Mathematics. For the Learning of Mathematics, 7(3): 33-39. Nichols, P D Meyers, J L & Burling, K S (2009) A Framework for evaluating and planning assessments intended to improve student achievement in S. Brookhart (Ed), (2009), op cit pp. 14–23. Peng, A. (2010). Teacher knowledge of students’ mathematical errors. www.tktk.ee/bw_client_files/tktk_pealeht/.../MAVI16_Peng.pdf Peng, A., & Luo, Z. (2009). A framework for examining mathematics teacher knowledge as used in error analysis. For the Learning of Mathematics, 29(3), 22-25. Prediger, S (2010) “How to develop mathematics-for-teaching and for understanding: the case of meanings of the equal sign” , Journal of Math Teacher Education 13:73– 93. Radatz, H. (1979) Error Analysis in Mathematics Education in Journal for Research in Mathematics Education, 10 (3): 163- 172 http://www.jstor.org/stable/748804 Accessed: 08/06/2011 05:53. Reddy, V. (2006) Mathematics and Science Achievement at South African Schools in TIMSS 2003. Cape Town: Human Sciences Research Council. Rusznyak, L. (2011) Seeking substance in student teaching in Y. Shalem & S Pendlebury (Eds) Retrieving Teaching: Critical issues in Curriculum, Pedagogy and learning. Claremont: Juta. pp 117-129 Shalem, Y. (2003) Do we have a theory of change? Calling change models to account. Perspectives in Education 21 (1): 29 –49. Shalem, Y., & Sapire, I. (2011). Curriculum Mapping: DIPIP 1 and 2. Johannesburg: Saide 127 Shalem, Y., Sapire, I., Welch, T., Bialobrzeska, M., & Hellman, L. (2011). Professional learning communities for teacher development: The collaborative enquiry process in the Data Informed Practice Improvement Project. Johannesburg: Saide. Available from http://www.oerafrica.org/teachered/TeacherEducationOERResources/SearchResults/tabi d/934/mctl/Details/id/38939/Default.aspx Shalem, Y. & Hoadley, U. (2009) The dual economy of schooling and teacher morale in South Africa. International Studies in Sociology of Education. Vol. 19 (2) 119–134. Shalem,Y. & Slonimsky,L. (2010) The concept of teaching in Y. Shalem & S Pendlebury (Eds) Op cit. pp 12-23 Shavelson, R. Li, M. Ruiz-Primo, MA & Ayala, CC (2002) Evaluating new approaches to assessing learning. Centre for the Study of Evaluation Report 604 http://www.cse.ucla.edu/products/Reports/R604.pdf Shepard, L A (2009) Commentary: evaluating the validity of formative and Interim assessment: Op cit pp. 32-37. Smith, J.P., diSessa A. A., & Roschelle, J. (1993) Misconceptions Reconceived: A Constructivist analysis of Knowledge in Transition. Journal of Learning Sciences 3(2): 115-163. Star, J.R. (2000). On the relationship between knowing and doing in procedural learning. In B. Fishman & S. O’Connor-Divelbiss (Eds), Fourth International Conference of the Learning Sciences. 80-86. Mahwah, NJ: Erlbaum. Taylor, N. (2007) “Equity, efficiency and the development of South African schools”. In International handbook of school effectiveness and improvement, ed. T. Townsend, 523. Dordrecht: Springer. Taylor, N. (2008) What’s wrong with South African Schools? Presentation to the Conference What’s Working in School Development. JET Education Services 28-29. Van den Berg, S. (2011) “Low Quality Education as a poverty trap” University of Stellenbosch. Van den Berg, S. & Shepherd, D. (2008) Signalling performance: An analysis of continuous assessment and matriculation examination marks in South African schools. Pretoria: Umalusi 128 Appendices: 129 Appendix 1 Exemplar template- Error analysis – used in group leader training prior to the activity Based on ICAS 2006, Grade 8 Question 22 Part A: Details Grade 8 Question 22 Content Solve a simple linear equation that has one unknown pronumeral Area Algebra Difficulty 27 Part B: Learner responses (%) A 20 B 17 C 28 D 33 Correct answer A Analysis More learners chose incorrect answers C and D, together constituting three times the correct choice A. 130 Part C: Analysis of correct answer Correct Answer A – 20% Indicate all possible methods 1. Using the distributive law multiply ½ into the binomial. Take the numbers onto the RHS, find a LCM, multiply out and solve for y. 2. Find an LCM which is 2. Multiply out and solve for y. 3. Substitution of 13 on a trial and error basis. In order to get this answer learners must know: 1. Learners need to know\how to determine a lowest common denominator (LCD) and/or LCM. 2. Learners need to understand the properties of an equation in solution of the unknown variable. 3. Learners need to know the procedure of solving an equation involving the RHS and LHS operations. Part D: Analysis of learners’ errors For answer D – 33% Indicate all possible methods ½(y+3) = 8 ½ ½ y+3 =4 y =1 Possible reasons for errors: Learners do not understand the concept of dividing fractions. They misconceive 8 divided by a half as a half of 8. The answer that they get is 4. For answer C – 28% Indicate all possible methods a. ½(y+3) = 8 ½ y+3 =4 y =7 131 or b. ½ (7+3)=8 7+3 =8+2 or c. ½ (y+3)=8 ½ y+6 =8 ½ y = 14 y =7 Possible reasons for errors: In the first method the learners made the same error of dividing by a fraction but did not change the sign when taking the terms over (LHS/RHS rule) Substitution of the value 7. In this method the learners did not multiply every term by the LCM (2). They also did not change the sign of the 6 when taking it over. For answer B – 17% Indicate all possible methods a. ½ (y+3)=8 y+6 =16 y =10 b. ½ (10+3)=8 5+3 =8 c. ½ (y+3)=8 ½ y =8-3 y =10 Possible reasons for errors: Learners have made errors in the order of operations, substitution and the distributive law. 132 Other issues with task The instruction could have read, “Solve for y” Issues for teaching Learners need to be able to check solutions of equations. Instructions to the learners in the classroom differ to the instructions of the task. Distributive Law Substitution Principles of rules/laws of equations i.e. LHS=RHS 133 Appendix 2 Template for DIPIP error analysis of ICAS task analysis (In both Round 1 and 2) (2008 error analysis – All groups Grades 3,4,5,6,7,8,9) (2010 error analysis – One group from Grades 3,4,5,6,8) Part A: Details Grade Question Content/Description Area/LO Part B: Learner responses (%) Anticipated achievement (Before you check the actual achievement) Expectation: Comment on any particular distractors: A B C D Correct answer Difficulty Analysis Part C: Analysis of correct answer The way(s) of thinking that the learner might use in order to get the question correct (fill in more than one possible way of thinking, if necessary): Indicate for each method whether you think it is at the expected level for the grade OR higher OR lower than the group would expect. 134 1. 2. 3. 4. 5. Part D: Analysis of learners’ errors Analysis of each of the distractors, in order, from the one selected most often to the one selected least often. You may trial these with your learners – please report on this if you do so. For each distractor, think about what learners might have done to obtain the answer in each incorrect choice: For choice – % The way(s) of thinking that the learners might have used in order to get this particular incorrect choice using this particular method (fill in more than one possible way of thinking, if necessary) ie. Possible reasons for errors/Possible misconceptions 1. 2. 3. 4. For choice – % The way(s) of thinking that the learners might have used in order to get this particular incorrect choice using this particular method (fill in more than one possible way of thinking, if necessary) ie. Possible reasons for errors/Possible misconceptions 1. 135 2. 3. 4. For choice – % The way(s) of thinking that the learners might have used in order to get this particular incorrect choice using this particular method (fill in more than one possible way of thinking, if necessary) ie. Possible reasons for errors/Possible misconceptions 1. 2. 3. 4. Other issues with task not discussed above Issues to consider when teaching (for example, how to over come some of the problems that have been raised.) 136 Appendix 3 Template for DIPIP error analysis of own test task analysis (In Round 2) (2010 error analysis – Grades 3,4,5,6,7,9) Part A: Details Grade Question: Please transcribe the full question here Content/Description: Please describe the mathematical content of the question and any other relevant detail. Area/LO Part B: Learner responses Anticipated achievement Expectation: Sample number (How many tests do you have? Count and record this number. You will use all of your tests every time in your analysis) Number of learners in sample that got the item correct (count the number of learners in your sample who got the item perfectly correct) Number of learners in the sample that got the item partially correct (count the number of learners in your sample who did some partially correct working but did not complete the item correctly) Number of learners in the sample that got the item wrong (count the number of learners in the test who were awarded a 137 “zero” for the item) Difficulty (work out the percentage of the sample who got the item right. You will do this for each item. Then place them on a relative scale for your test of easiest to hardest item.) Part C: Analysis of correct answer Go through your sample of learner tests to find as many different ways of getting the question correct. Then record these ways and write about the way(s) of thinking that the learner might use in order to get the question correct (fill in more than one possible way of thinking, if necessary): USE YOUR LEANER TESTS to generate this list. Indicate for each method whether you think it is at the expected level for the grade OR higher OR lower than the group would expect. 1. 2. 3. 4. 5. Part D: Analysis of learners’ errors Go through your sample of learner tests to find as many different ways of getting the question incorrect. Then record these ways and write about the way(s) of thinking that the learner might use in order to get the question incorrect, i.e. Possible reasons for errors/Possible misconceptions (fill in more than one possible way of thinking, if necessary): USE YOUR LEANER TESTS to generate this list. 1. 2. 3. 138 4. 5. Other issues with task not discussed above Issues to consider when teaching (for example, how to overcome some of the problems that have been raised.) 139 Appendix 4 Error analysis coding template – final version (exemplar with one Grade 3 items) Grade Grade 3 Grade 3 Ite m 1 1 Maths content area Numbe r Numbe r Distr actor C B text no Text Proc edur al Conc eptu al Awa renes s of error Diag nosti c reaso ning each item of text from template Categ. Categ. Categ. Categ. 1 They counted the books on the shelf. To do this they must know how to count correctly, be able to identify a book as a unit and they must distinguish between the ends of a shelf and a book so that they do not mistake the ends of a shelf as another book. 2 They looked at shelf and saw 5 books on the left and 2 on the right and added them together. 3 Counted every book that had a label on it. 1 Subtracted the books that was standing up straight on the left and right and counted the three books that are slanting in the middle. They mistook the books Use of ever yday Disti nct expla natio n Comments F/N 140 on the left and right as part of the shelf. 2 Focused on the three thicker books. The thicker books are more visible. 141 Appendix 5 Error analysis – Template criteria for coding Criteria Explanations of the correct solution Procedural The emphasis of this code is on the teachers’ procedural explanation of the solution. Teaching mathematics involves a great deal of procedural explanation which should be done fully and accurately for the learners to grasp and become competent in working with the procedures themselves. Conceptual The emphasis of this code is on the teachers’ conceptual explanation of the procedure/other reasoning followed in the solution. Mathematical reasoning (procedural/other) needs to be unpacked and linked to the concepts to which it relates in order for learners to understand the mathematics embedded in the activity. Explanations of the incorrect solution (error in distractor) Awareness of error The emphasis of this code is on teachers’ explanation of the actual mathematical error and not on learners’ Category descriptors: Error analysis Full Partial Inaccurate Not Present No mathematica l procedural explanation is given Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. The procedure is accurate and includes all of the steps in the procedure. The procedure is accurate but it does not include all of the steps in the procedure. Teacher’s use of procedure is inaccurate or incomplete to the extent that it could be confusing. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. The explanation illuminates conceptually the background and process of the activity. The explanation includes some but not all of the conceptual links which illuminate the background and process of the activity. The explanation includes incorrect or poorly conceived conceptual links and thus is potentially confusing. Full Partial Inaccurate Teachers explain the mathematical error made by the learner. Teachers explain the mathematical error made by the learner. Teachers explain the mathematical error made by the learner. The explanation of the particular error is mathematically sound The explanation of the particular error is mathematically sound The explanation of the particular error is mathematically No conceptual links are made in the explanation. Not present No explanation is given of the mathematica l of the particular 142 reasoning. and suggest links to common misconceptions or errors but does not link to common misconceptions or errors inaccurate. error. Diagnostic reasoning Teachers describe learners’ mathematical reasoning behind the error. Teachers describe learners’ mathematical reasoning behind the error. Teachers describe learners’ mathematical reasoning behind the error. It describes the steps of learners’ mathematical reasoning systematically and hones in on the particular error. The description of the learners’ mathematical reasoning is incomplete although it does hone in on the particular error. The description of the learners’ mathematical reasoning does not hone in on the particular error. No attempt is made to describe learners’ mathematica l reasoning behind the particular error Teachers’ explanation of the learners’ mathematical reasoning behind the error appeals to the everyday. Teachers’ explanation of the learners’ mathematical reasoning behind the error appeals to the everyday. Teachers’ explanation of the learners’ mathematical reasoning behind the error appeals to the everyday. The idea of error analysis goes beyond identifying a common error and misconception. The idea is to understand the way teachers go beyond the actual error to try and follow the way the learners were reasoning when they made the error. The emphasis of this code is on teachers’ attempt to provide rationale for how learners were reasoning mathematically when they chose the distractor. Use of everyday knowledge Teachers often explain why learners make an error by appealing to everyday experiences that learners draw on and confuse with the mathematical context of the question. The emphasis of this code is on the quality of the use of everyday, judged by the links made to the mathematical understanding teachers try to advance. Multiple explanations of error One of the challenges in error analysis is for learners to hear more than one explanation of the error. This is because some explanations are more accurate or more accessible than others. This code examines the teachers’ Teachers’ use of the ‘everyday’ enables mathematical understanding by making the link between the everyday and the mathematical clear Teacher’s use of the ‘everyday’ is relevant but does not properly explain the link to mathematical understanding Multiple mathematical explanations are provided. Multiple mathematical and general explanations are provided All of the explanations (two or more) are mathematically feasible/convincing At least two of the mathematical explanations feasible/convincing No discussion of everyday is done. Teacher’s use of the ‘everyday’ dominates and obscures the mathematical understanding, no link to mathematical understanding is made Multiple mathematical and general explanations are provided One mathematically feasible/convincing explanation provided No mathematica lly feasible/conv incing explanation provided 143 explanation(s) of the error itself rather than the explanation of learners’ reasoning. This is coded F/N (mathematically feasible/not) for each new and different explanation offered by the teacher. The final code is assigned according to the level descriptors above. 144 Appendix 6 Procedural explanations The literature emphasises that the quality of teachers’ explanations depends on the balance they achieve between explaining the procedure required for addressing a mathematical question and the mathematical concepts underlying the procedure. This criterion aims to grade the quality of the teachers’ procedural explanations of the correct answer. The emphasis in the criterion is on the quality of the teachers’ procedural explanations when discussing the solution to a mathematical problem through engaging with learner test data. Teaching mathematics involves a great deal of procedural explanation which should be done fully and accurately for the learners to grasp and become competent in working with the procedures themselves. The four categories, which capture the quality of the procedural explanations demonstrated by a teacher/group, are presented in Table A1 below. Table A1: Category descriptors for “procedural explanations” Procedural Full Partial The emphasis of this code is on the teachers’ procedural explanation of the solution. Teaching mathematics involves a great deal of procedural explanation which should be done fully and accurately for the learners to grasp and become competent in working with the procedures themselves. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. The procedure is accurate and includes all of the key steps in the procedure. Inaccurate Not present Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. No procedural explanation is given The procedure is accurate but it does not include all of the key steps in the procedure. Teacher’s use of procedure is inaccurate and thus shows lack of understanding of the procedure. Exemplars of coded texts of correct answer texts for procedural explanations: The first set of exemplars relates to a single ICAS item to show the vertical differentiation of the correct answer codes in relation to one mathematical concept under discussion. These exemplar texts present explanations that received the same level 145 coding on the two criteria of procedural and conceptual explanations (see Table A5). This would not always be the case. We have chosen to present such exemplars because of the high correlation between procedural and conceptual explanations. Table A2: Criterion 1 – Procedural explanation of the choice of the correct solution in relation to one item Test Item Criterion wording (Grade 8 item 8 ICAS 2006) Procedural Which row contains only square numbers? (A) 2 4 8 16 (B) 4 16 32 64 (C) 4 16 36 64 (D) 16 36 64 96 The emphasis of this code is on the teachers’ procedural explanation of the solution. Teaching mathematics involves a great deal of procedural explanation which should be done fully and accurately for the learners to grasp and become competent in working with the procedures themselves. Selection – correct answer - C Text Code Category descriptor 1² = 1; 2² = 4; 3² = 9; 4² = 16; 5² = 25; 6² = 36 ; 7² = 49; 8² = 64; Therefore the row with 4; 16; 36; 64 only has square numbers. To get this right they need to know what “square numbers” mean and to be able to calculate or recognize which of the rows consists only of square numbers. Full Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. Learners would illustrate squares to choose the right row. Partial The procedure is accurate and includes all of the key steps in the procedure. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. The procedure is accurate but it does not include all of the key steps in the procedure. Learners could calculate the square roots of all the combinations in order to discover the correct one. To get this learners need to know how to use the square Inaccurate Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. Teacher’s use of procedure is inaccurate and thus shows lack of understanding of the procedure. 146 root operation. Learners understood the question and chose the right row. Not present No procedural explanation is given The next set of exemplars relate to various ICAS item to show the differentiation of the correct answer codes in relation to a spread of mathematical concepts. Table A3: Criterion 1 – Procedural explanation of the choice of the correct solution in relation to a range of items Criterion wording Procedural The emphasis of this code is on the teachers’ procedural explanation of the solution. Teaching mathematics involves a great deal of procedural explanation which should be done fully and accurately for the learners to grasp and become competent in working with the procedures themselves. Item Category descriptor Text: Full explanation ICAS 2007 Grade 9 Item 24 Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. Learners will take the 4 800 revolutions and divide it by 60 to convert the answer to seconds and obtain an answer of 80 they then take this answer and multiply it by 360 for the number of degrees in a circle to obtain the correct answer of 28 800. Selected correct answer – C Content area: Geometry The procedure is accurate and includes all of the key steps in the procedure. 147 ICAS 2006 Grade 7 Item 3 Selected correct answer – A Content area: Number Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. Text: Partial explanation Learners used the division line correctly inside the circle and counted the pieces correctly. They fully understood that this division in a circle happens from the centre, and that thirds mean that there are three equal parts. The procedure is accurate but it does not include all of the key steps in the procedure. ICAS 2007 Grade 6 Item 8 Selected correct answer – B Content area: Measurement Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes demonstration of procedure. Text: Inaccurate explanation 1,3 + 1,3 = 2,6 +1,3 = 3,9 +1,3 = 5,2 +1,3 = 6,5 +1,3 =7,8 +1,3 = 9,1 + 1,3 = 10,4 (that’s the closest ) Teacher’s use of procedure is inaccurate and thus shows lack of understanding of the procedure. 148 ICAS 2006 Grade 5 Item 12 Selected correct answer – B Content area: Data Handling No procedural explanation is given Text: Mathematical explanation not present Learners managed to get the selected distractor by focusing on the key 149 Appendix 7 Conceptual explanations The emphasis in this criterion is on the conceptual links made by the teachers in their explanations of the learners’ mathematical reasoning in relation to the correct answer. Mathematical procedures need to be unpacked and linked to the concepts to which they relate in order for learners to understand the mathematics embedded in the procedure. The emphasis of the criterion is on the quality of the teachers’ conceptual links made in their explanations when discussing the solution to a mathematical problem through engaging with learner test data. The four level descriptors for this criterion, which capture the quality of the conceptual explanations demonstrated by a teacher/group, are presented in Table A4 below. Table A4: Category descriptors for “conceptual explanations” Conceptual Full Partial The emphasis of this code is on the teachers’ conceptual explanation of the procedure/other reasoning followed in the solution. Mathematical reasoning (procedural/other) needs to be unpacked and linked to the concepts to which it relates in order for learners to understand the mathematics embedded in the activity. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. The explanation illuminates conceptually the background and process of the activity. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. The explanation includes some but not all of the key conceptual links which illuminate the background and process of the activity. Inaccurate Not present Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. No conceptual links are made in the explanation . The explanation includes poorly conceived conceptual links and thus is potentially confusing. Exemplars of coded correct answer texts for conceptual explanations The first set of exemplars relates to a single ICAS item to show the vertical differentiation of the correct answer codes in relation to one mathematical concept under discussion. These exemplar texts present explanations that received the same level coding on the two criteria of procedural (see Table A2) and conceptual explanations. 150 This would not always be the case. We have chosen to present such exemplars because of the high correlation between procedural and conceptual explanations. Table A5: Criterion 2 – Conceptual explanation of the choice of the correct solution in relation to one item Test Item Criterion wording (Grade 8 item 8 ICAS 2006) Which row contains only square numbers? (a) 2 4 8 16 (b) 4 16 32 64 (c) 4 16 36 64 (d) 16 36 64 96 Conceptual The emphasis of this code is on the teachers’ conceptual explanation of the procedure/other reasoning followed in the solution. Mathematical reasoning (procedural/other) needs to be unpacked and linked to the concepts to which it relates in order for learners to understand the mathematics embedded in the activity. Selection – correct answer - C Text: Number Code Category descriptor 1² = 1; 2² = 4; 3² = 9; 4² = 16; 5² = 25; 6² = 36 ; 7² = 49; 8² = 64; Therefore the row with 4; 16; 36; 64 only has square numbers. To get this right they need to know what “square numbers” mean and to be able to calculate or recognize which of the rows consists only of square numbers. Full Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. Learners would illustrate squares to choose the right row. Partial The explanation illuminates conceptually the background and process of the activity. Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. The explanation includes some but not all of the key conceptual links which illuminate the background and process of the activity. Learners could calculate the square roots of all the combinations in order to discover the correct one. To get this learners need to know how to use the square root operation. Inaccurate Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. The explanation includes poorly conceived conceptual links and thus is potentially confusing. 151 Learners understood the question and chose the right row. Not present No conceptual links are made in the explanation. Table A6: Criterion 2 – Conceptual explanation of the choice of the correct solution in relation to a range of items Criterion Conceptual wording The emphasis of this code is on the teachers’ conceptual explanation of the procedure/other reasoning followed in the solution. Mathematical reasoning (procedural/other) needs to be unpacked and linked to the concepts to which it relates in order for learners to understand the mathematics embedded in the activity. Item Level descriptor “Own test” Grade 4 Item Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. Content area: Measurement The explanation illuminates conceptually the background and process of the activity. B) Mpho has 2 l of milk. How many cups of 100ml can he fill ? __________ (2) Text: Full explanation Learner wrote 1 100 100 2 100 100 3 6 100 7 16 100 100 100 12 = 500ml 1000ml =1l 5 8 100 11 100 4 100 100 100 100 9 100 13 = 500ml 10 = 500ml 1000ml =1l 100 14 15 = 500ml 100 18 100 19 100 20 100 17 Learner then counted the number of groups of 100 to calculate how many cups and got 20 152 ICAS 2006 Grade 5 Item 30 Selected correct answer – B Content area: Pattern and Algebra Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. Text: Partial explanation Sequence of counting numbers were used matching them with odd numbers, so that for every blue block there are two pink blocks added. The explanation includes some but not all of the key conceptual links which illuminate the background and process of the activity. ICAS 2007 Grade 9 Item 8 Selected correct answer – D Content area: Number Teachers’ explanation of the learners’ mathematical reasoning behind the solution includes conceptual links. The explanation includes poorly conceived conceptual links and thus is potentially confusing. Text: Inaccurate explanation The learner will use the information that 6 slices of pizza make a whole and therefore the denominator for this fraction is 6. They will then take the 16 slices eaten and make that the numerator and hence ascertain how many whole pizzas had been eaten. 153 ICAS 2006 Grade 3 Item 14 Selected correct answer – D Content area: Data handling No conceptual links are made in the explanation. Text: Mathematical explanation not present Learners are attracted to yellow as the brightest colour. Young learners often attracted to the brightest colour. Learners didn’t understand the question. 154 Appendix 8 Awareness of error The emphasis in this criterion is on the teachers’ explanations of the actual mathematical error (and not on the learners’ reasoning). The emphasis in the criterion is on the mathematical quality of the teachers’ explanations of the actual mathematical error when discussing the solution to a mathematical problem. The four level descriptors for this criterion, which capture the quality of the awareness of error demonstrated by a teacher/group, are presented in Table A7 below. Table A7: Category descriptors for “awareness of mathematical error” Awareness of error Full Partial Inaccurate The emphasis of this code is on teachers’ explanation of the actual mathematical error and not on learners’ reasoning. Teachers explain the mathematical error made by the learner. Teachers explain the mathematical error made by the learner. The explanation of the particular error is mathematically sound and suggests links to common misconceptions or errors. The explanation of the particular error is mathematically sound but does not link to common misconceptions or errors. Teachers explain the mathematical error made by the learner. The explanation of the particular error is mathematically inaccurate or incomplete and hence potentially confusing. Not present No mathematic al explanation is given of the particular error. Exemplars of error answer texts for awareness of error criterion: As with the exemplars for procedural explanations, the first set of exemplars relates to a single ICAS item to show the vertical differentiation of the correct answer codes in relation to one mathematical concept under discussion. These exemplar texts present explanations that received the same level coding on the two criteria of awareness of the error and diagnostic reasoning (see Table A11). This would not always be the case. We have chosen to present such exemplars because of the high correlation between awareness of the error and diagnostic reasoning. Table A8: Criterion 3 – Awareness of the error embedded in the incorrect solution in relation to one item Test Item Criterion wording (Grade 7 item 26 ICAS 2006) 155 Awareness of error The emphasis of this code is on teachers’ explanation of the actual mathematical error and not on learners’ reasoning. Selection – incorrect answer - B Text Code Category descriptor The pupils started from the highlighted 3 and then counted in groups of 3 and then just added 1 because the next box being 7 was highlighted. They then continued this pattern throughout and found that it worked well. The pattern worked and so they assumed that this was the correct answer but ignored the word multiples. They just counted in three’s (plus 1) to get the answer. Full Teachers explain the mathematical error made by the learner. Perhaps the pupils looked at the 3 and the 1 in the answer, decided that 3 + 1 = 4 and thus started from the highlighted 3 and then counted in four’s and thus got the numbers of 3; 11; 15; 19; etc. Partial The pupils have a poor understanding of the word multiple, as they just counted in groups of three’s and not in multiples of three’s. Inaccurate The pupils did not work out the actual sum, they just read Not The explanation of the particular error is mathematically sound and suggest links to common misconceptions or errors Teachers explain the mathematical error made by the learner. The explanation of the particular error is mathematically sound but does not link to common misconceptions or errors Teachers explain the mathematical error made by the learner. The explanation of the particular error is mathematically inaccurate. No mathematical explanation is given of the particular error. 156 the question, it looked similar to the answer and thus chose (multiples of 3) + 1. present Table A9: Criterion 3 – Awareness of the error embedded in the incorrect solution in relation to a range of items Criterion wording Awareness of error The emphasis of this code is on teachers’ explanation of the actual mathematical error and not on learners’ reasoning. Item Level descriptor Text: Full explanation ICAS 2007 Grade 5 Item 17 Teachers explain the mathematical error made by the learner. After the learner got the white square area which was 64 he/she did not subtract the dark square area from the area of the white square. This is another halfway step to finding the answer. Selected distractor – D Content area: Measurement The explanation of the particular error is mathematically sound and suggest links to common misconceptions or errors ICAS 2006 Grade 6 Item 5 Selected distractor – C Teachers explain the mathematical error made by the learner. Content area: Number Text: Partial explanation Subtracted or added the 42 and the 6 because they misinterpreted the division sign. The explanation of the particular error is mathematically sound but does not link to common misconceptions or errors 157 ICAS 2007 Grade 4 Item 15 Selected distractor – C Teachers explain the mathematical error made by the learner. Content area: Number The explanation of the particular error is mathematically inaccurate or incomplete and hence potentially confusing. ICAS 2006 Grade 9 Item 20 Selected distractor – A Content area: Measurement No mathematical explanation is given of the particular error. Text: Inaccurate explanation The learners might have considered the last digit of the product (37) and decided to choose this answer because it has 7x7. Text: Mathematical explanation not present They did not answer the question, yet chose to answer anther: How deep is the anchor under water? 158 Appendix 9 Diagnostic Reasoning The idea of error analysis goes beyond identifying the actual mathematical error. The idea is to understand how teachers go beyond the mathematical error and follow the way learners may have been reasoning when they made the error. The emphasis in the criterion is on the quality of the teachers’ attempt to provide a rationale for how learners were reasoning mathematically when they chose a distractor. The four level descriptors for this criterion, which capture the quality of the diagnostic reasoning demonstrated by a teacher/group, are presented in Table A10 below. Table A10: Category descriptors for “diagnostic reasoning” Diagnostic reasoning Full Partial The idea of error analysis goes beyond identifying a common error and misconception. The idea is to understand the way teachers go beyond the actual error to try and follow the way the learners were reasoning when they made the error. The emphasis of this code is on teachers’ attempt to provide rationale for how learners were reasoning mathematically when they chose the distractor. Teachers describe learners’ mathematical reasoning behind the error. Teachers describe learners’ mathematical reasoning behind the error. The description of the steps of learners’ mathematical reasoning is systematic and hones in on the particular error. The description of the learners’ mathematical reasoning is incomplete although it does hone in on the particular error. Inaccurate Not present Teachers describe learners’ mathematical reasoning behind the error. No attempt is made to describe learners’ mathematical reasoning behind the particular error The description of the learners’ mathematical reasoning does not hone in on the particular error. Exemplars of coded error answer texts for diagnostic reasoning: As with the exemplars for procedural explanations, the first set of exemplars relates to a single ICAS item to show the vertical differentiation of the correct answer codes in relation to one mathematical concept under discussion. These exemplar texts present explanations that received the same level coding on the two criteria of awareness of the error (see Table A8) and diagnostic reasoning. This would not always be the case. We 159 have chosen to present such exemplars because of the high correlation between awareness of the error and diagnostic reasoning. Table A11: Criterion 4 – Diagnostic reasoning of learner when selecting the incorrect solution in relation to one item Test Item Criterion wording (Grade 7 item 26 ICAS 2006) Diagnostic reasoning The idea of error analysis goes beyond identifying a common error and misconception. The idea is to understand the way teachers go beyond the actual error to try and follow the way the learners were reasoning when they made the error. The emphasis of this code is on teachers’ attempt to provide rationale for how learners were reasoning mathematically when they chose the distractor. Selection – incorrect answer - B Text Code Category descriptor The pupils started from the highlighted 3 and then counted in groups of 3 and then just added 1 because the next box being 7 was highlighted. They then continued this pattern throughout and found that it worked well. The pattern worked and so they assumed that this was the correct answer but ignored the word multiples. They just counted in three’s (plus 1) to get the answer. Full Teachers describe learners’ mathematical reasoning behind the error. Perhaps the pupils looked at the 3 and the 1 in the answer, decided that 3 + 1 = 4 and thus started from the highlighted 3 and then Partial It describes the steps of learners’ mathematical reasoning systematically and hones in on the particular error. Teachers describe learners’ mathematical reasoning behind the error. The description of the learners’ mathematical reasoning is 160 counted in four’s and thus got the numbers of 3 – 11 – 15 – 19 – etc… incomplete although it does hone in on the particular error. The pupils have a poor understanding of the word multiple, as they just counted in groups of three’s and not in multiples of three’s. Inaccurate The pupils did not work out the actual sum, they just read the question, it looked similar to the answer and thus chose (multiples of 3) + 1. Not present Teachers describe learners’ mathematical reasoning behind the error. The description of the learners’ mathematical reasoning does not hone in on the particular error. No attempt is made to describe learners’ mathematical reasoning behind the particular error Table A12: Criterion 4 – Diagnostic reasoning of learner when selecting the incorrect solution in relation to a range of items Criterion wording Diagnostic reasoning The idea of error analysis goes beyond identifying a common error and misconception. The idea is to understand the way teachers go beyond the actual error to try and follow the way the learners were reasoning when they made the error. The emphasis of this code is on teachers’ attempt to provide rationale for how learners were reasoning mathematically when they chose the distractor. Item ICAS 2007 Grade 3 Item 16 Selected distractor – C Content area: Measurement Category descriptor Teachers describe learners’ mathematical reasoning behind the error. It describes the steps of learners’ mathematical reasoning systematically and hones in on the particular error. Text: Full explanation Learners started at the cross moved to Green Forest South (down) and then must have moved East and not West, and therefore ended at white beach because they got did not know or confused with direction between East/West. Learners may have moved towards the "W" for West on the drawing of the compass which is given on the map. This compass drawing might also have confused them and made them go to the east when they should have gone west, though they did go "down" for South initially in order to end up at 161 White Beach. ICAS 2006 Grade 4 Item 20 Selected distractor – C Content area: Geometry Teachers describe learners’ mathematical reasoning behind the error. The description of the learners’ mathematical reasoning is incomplete although it does hone in on the particular error. Text: Partial explanation This looks most like tiles we see on bathroom walls, even though it is a parallelogram, so even though it is NOT rectangular, learners selected this as the correct answer. They did not realize that this shape would go over the edges or create gaps at the edges if it was used. 162 ICAS 2007 Grade 5 Item 14 Selected distractor – B Content area: Measurement Teachers describe learners’ mathematical reasoning behind the error. Text: Inaccurate explanation The distractor selected most was B. The likelihood is that the learners could choose 600 because they could have seen six years in the question. The description of the learners’ mathematical reasoning does not hone in on the particular error. ICAS 2006 Grade 6 Item 7 Selected distractor – C Content area: Number No attempt is made to describe learners’ mathematical reasoning behind the particular error Text: Mathematical explanation not present Not thinking systematically to devise a correct formula to obtain the number of cats. This learner is perhaps feeling pressurised to produce a formula describing the situation, rather than manually grouping the number of legs for each cat. There is a question here of whether the learner understands the meaning of the numbers in the formula s/he has used to calculate the number of cats. 163 Appendix 10 Use of the everyday Teachers often explain why learners make mathematical errors by appealing to everyday experiences that learners draw on and confuse with the mathematical context of the question. The emphasis in this criterion is on the quality of the use of everyday knowledge, judged by the links made to the mathematical understanding that the teachers attempt to advance. The four level descriptors for this criterion, which capture the quality of the use of everyday knowledge demonstrated by a teacher/group, are presented in Table A13 below. Table A13: Category descriptors for “use of everyday” Use of everyday Full Partial knowledge Teachers often explain why learners make an error by appealing to everyday experiences that learners draw on and confuse with the mathematical context of the question. The emphasis of this code is on the quality of the use of everyday, judged by the links made to the mathematical understanding teachers try to advance Teachers’ explanation of the learners’ mathematical reasoning behind the error appeals to the everyday. Teachers’ use of the ‘everyday’ enables mathematical understanding by making the link between the everyday and the mathematical clear Inaccurate Not present Teachers’ explanation of the learners’ mathematical reasoning behind the error appeals to the everyday. Teachers’ explanation of the learners’ mathematical reasoning behind the error appeals to the everyday. No discussio n of everyday is done Teacher’s use of the ‘everyday’ is relevant but does not properly explain the link to mathematical understanding Teacher’s use of the ‘everyday’ dominates and obscures the mathematical understanding, no link to mathematical understanding is made “Use of the everyday” was used to measure the extent to which teachers used everyday contexts to enlighten learners about mathematical concepts. In the exemplar tables, a variety of different items where teachers used everyday contexts to add to their mathematical explanation of the reasoning behind an error are given. The code “no discussion of the everyday” includes all explanations which did not make reference to the everyday whether it was appropriate or not. Further more detailed 164 analysis of the kinds of contextualized explanations which were offered and whether they were missing when they should have been present could be carried out since it was not within the scope of this report. Exemplars of coded error answer texts for everyday criterion Table A14: Use of the everyday exemplars Criterion wording Use of everyday knowledge Teachers often explain why learners make an error by appealing to everyday experiences that learners draw on and confuse with the mathematical context of the question. The emphasis of this code is on the quality of the use of everyday, judged by the links made to the mathematical understanding teachers try to advance. Item Category descriptor Text: Full explanation ICAS 2006 Grade 9 Item 6 Teachers’ explanation of the learners’ mathematical reasoning behind the error appeals to the everyday. He draws on his frame of reference of how he perceives a litre to be e.g. a 1,25l of cold drink of a 1l of milk or a 2l of coke etc. Selected distractor – D Content area: Measurement Teachers’ use of the ‘everyday’ enables mathematical understanding by making the link between the everyday and the mathematical clear 165 ICAS 2006 Grade 7 Item 9 Selected distractor – A Content area: Data Handing Teachers’ explanation of the learners’ mathematical reasoning behind the error appeals to the everyday. Teacher’s use of the ‘everyday’ is relevant but does not properly explain the link to mathematical understanding ICAS 2007 Grade 3 Item 20 Selected distractor – B Content area: Measurement Teachers’ explanation of the learners’ mathematical reasoning behind the error appeals to the everyday. Text: Partial explanation Here learners did not read the question attentively and misinterpreted or expected that the most popular sport would be the question – this is football which has a frequency of 7 (the highest). This could be the case because of the way in which teachers use these bar graphs in class. The most popular choice of question is always the question about the most or least popular choice as indicated on the graph. Text: Inaccurate explanation In the last option the circle/ ball managed to pull the cylinder. Teacher’s use of the ‘everyday’ dominates and obscures the mathematical understanding, no link to mathematical understanding is made 166 ICAS 2006 Grade 5 Item 11 Selected distractor – B Content area: Geometry No discussion of everyday is done. Text: Mathematical explanation not present Some could have guessed just to have an answer. 167 Appendix 11 Multiple explanations One of the challenges in the teaching of mathematics is that learners need to hear more than one explanation of the error. This is because some explanations are more accurate or more accessible than others and errors need to be explained in different ways for different learners. This criterion examines the teachers’ ability to offer alternative explanations of the error when they are engaging with learners’ errors through analysis of learner test data. The four level descriptors for this criterion, which capture the quality of the multiple explanations of error demonstrated by the group, are presented in Table A15 below. Table A15: Category descriptors for “multiple explanations” Multiple explanations Full Partial of error One of the challenges in error analysis is for learners to hear more than one explanation of the error. This is because some explanations are more accurate or more accessible than others. This code examines the teachers’ explanation(s) of the error itself rather than the explanation of learners’ reasoning. Multiple mathematical explanations are provided. All of the explanations (two or more) are mathematically feasible/convincing Inaccurate Not present Multiple mathematical and general explanations are provided Multiple mathematical and general explanations are provided No mathematicall y feasible/convin cing explanation provided At least two of the mathematical explanations are feasible/convincing One mathematically feasible/convinci ng explanation provided (with/without general explanations) (combined with general explanations) “Multiple explanations” was used to evaluate the range of explanations given by groups of learners’ thinking behind the incorrect answer. This was looking to see whether teachers offer a number of different explanations in relation to the concept. For multiple explanations we again give a variety of different items where teachers gave more than one mathematically feasible explanation of the reasoning behind an error. Exemplars of coded error answer texts for multiple explanations: Table A16: Multiple explanations exemplars 168 Criterion wording Multiple explanations of error One of the challenges in error analysis is for learners to hear more than one explanation of the error. This is because some explanations are more accurate or more accessible than others. This code examines the teachers’ explanation(s) of the error itself rather than the explanation of learners’ reasoning. Item Category descriptor ICAS 2007 Grade 7 Item 14 Multiple mathematical explanations are provided. Selected distractor – B Content area: Measurement Text: Full explanation 1. 2. All of the explanations (two or more) are mathematically feasible/convincing ICAS 2006 Grade 3 Item 4 Selected distractor – A Content area: Geometry Multiple mathematical and general explanations are provided At least two of the mathematical explanations feasible/convincing Learners said 6 x 3 = 18. They took the length given and multiplied be the three sides of the triangle. They confused the radius and diameter. They halved 6cm to get 3cm and then multiplied by 6, because the triangle is made up of six radii. Text: Partial explanation 1. 2. 3. The learners did not focus on the handle, but on the splayed out bristles or on the broom as a whole. See the bristles as the most important part of a broom. Did not read or understand the question. The learners focused on the broom as a whole. Because learners have difficulty separating the handle from the broom in their minds eye 169 ICAS 2007 Grade 6 Item 14 Selected distractor – A Content area: Data and Chance Multiple mathematical and general explanations are provided One mathematically feasible/convincing explanation provided Text: Inaccurate explanation If they read from 0 to 1000 there are 3 species of fish. They are reading the bar not the line. (they are distracted by the 2-D nature of the diagram ) 170 ICAS 2006 Grade 4 Item 2 Selected distractor – Content area: Number No mathematically feasible/convincing explanation provided Text: Mathematical explanation not present 1. 2. 3. They might be distracted by the word “playing” and choose particular dogs according to what they identify as a dog that is playing. They might not identify all of the animals as dogs, and hence not count all of them. We see these as issues of interpretation, of the language and of the diagram, rather than mathematical misconceptions. 171