Measuring confidence to teach statistics to middle & high school grades: The development & validation of the SETS instruments The Research Team Supported in part by CAUSE (under NSF DUE #0618790) Leigh M. Harrell-Williams Virginia Tech Georgia State University Rebecca L. Pierce Ball State University M. Alejandra Sorto Texas State University Lawrence M. Lesser The University of Texas at El Paso Teri J. Murphy Northern Kentucky University Why measure middle and high school grades pre-service teachers’ Self-Efficacy to Teach Statistics? • Teachers who are prepared to teach mathematics are expected to teach statistics as outlined by the PreK-12 Guidelines for Assessment and Instruction in Statistics Education (GAISE) (Franklin et al., 2007) and the Common Core State Standards for Mathematics (CCSSM) (National Governors’ Association, 2010). • Self-efficacy is task specific; any instrument that assesses self-efficacy needs to be task specific as well. Self-Efficacy, Attitudes, and Beliefs Teacher efficacy affects: - teacher motivation Existing Instruments: - willingness to use more - attitude towards innovative techniques statistics (SATS, ATS) - student achievement - efficacy for learning/doing statistics - time spent teaching (CSSE, SELS) certain concepts - statistical knowledge (Czerniak, 1990; (SCI) Riggs & Enochs, 1990; Wenta, 2000). No prior instrument measures self-efficacy for future teachers. GAISE PreK-12 Curriculum Framework Process Components 1. 2. 3. 4. A Five Objectives Formulate Question Collect Data Analyze Data Interpret Results “Although these three levels may parallel grade levels, they are based on development in statistical literacy, not age. Thus, a middle-school student who has had no prior experience with statistics will need to begin with Level A concepts and activities before moving to Level B.” (p. 13) Three Developmental Levels C B Nine Objectives Six Objectives CCSSM Framework Grade-Specific Standards for Each Grade Level VERSUS Topic-Specific Standards Similar to the 4 processes in the GAISE, there are 8 mathematical practices that are threaded throughout the CCSSM. Development Process for Middle Grades SETS Instrument Identified representative behaviors from GAISE items Determined alignment of GAISE report to state standards Spring/Summer 2008 Draft items created for instrument using language aligned with GAISE and state standards Development Process for Middle Grades SETS Instrument Revised item wording based on input from practicing elementary & middle school teachers Fall 2008/Spring 2009 Pilot Study 2009 Data Collection Study for Validation Purposes 2010 - 2011 The Middle Grades SETS Instrument 26 Likert scale items in this format: Please rate your confidence in teaching middle grades students the skills necessary to complete the following tasks successfully: Scale of 1 to 6 14 Demographic items Statistics Standards in the Common Core, Grades 2 - 6 Selected SETS Items based on GAISE Level-A Grades 2 - 5: Represent and interpret data Collecting measurements and creating line plot and bar graphs 12 items • Collect data to answer a posed statistical question in contexts of interest to middle school students. Grade 6: Develop understanding of statistical variability: • Recognize statistical question • Data has distribution with specific center/variation/shape Summarize and describe distributions • Create boxplots/histograms • Summarize data numerically • Recognize that there will be natural variability between observations for individuals. • Select appropriate graphical displays and numerical summaries to compare individuals to each other and an individual to a group. Statistics Standards in the Common Core, Grades 7 - 8 Grades 7: Use random sampling to draw inferences about a population. • Inferences can be made from sample about population if sample is representative. • Generate multiple samples to gauge accuracy. Draw informal comparative inferences about two populations. • Use graphs to estimate differences. • Use numerical values to assess differences. Grade 8: Investigate patterns of association in bivariate data. • Construct and interpret scatterplots. • Informally fit, assess and interpret a linear relationship. • Use contingency table to evaluate relationship. Selected SETS Items based on GAISE Level-B 15 items • Recognize the role of sampling error when making conclusions based on a random sample taken from a population. • Recognize that a sample may or may not be representative of a larger population. • Recognize sampling variability in summary statistics such as the sample mean and the sample proportion. • Use interquartile range, five-number summaries, and boxplots for comparing distributions. • Interpret measures of association. 2010 – 2011 Validation Study • Four US public institutions of higher education with significant proportion of students pursuing degrees in education • 309 participants enrolled in either an intro statistics course or a math education course Validation Study - Methods • Confirmatory Factor Analysis • Item Analysis • Rating Scale Analysis • Reliability Validation Study - CFA • Compared unidimensional and two-dimensional factor structure using Multidimensional Random Coefficient Multinomial Logit Model as implemented in Conquest software Two dimensions: (Friel, Curcio, & Bright, 2001) • – – • • Efficacy to Teach “Reading the Data” Efficacy to Teach “Reading Between the Data” AIC and BIC confirmed two dimension structure 0.85 between-dimension correlation Validation Study – Item Analysis Index Statistic Composite Reading the Data Reading Between The Data rpoint-polyserial Mean 0.66 0.67 0.69 MSweighted SD Min Max Nitems Mean SD Min Max NiMS > 1.5 0.05 0.55 0.73 26 - 0.06 0.59 0.73 11 1.00 0.13 0.71 1.17 0 0.05 0.62 0.77 15 1.00 0.23 0.71 1.52 1 MSunweighted Mean - 1 0.98 SD Min Max NiMS > 1.5 - 0.12 0.78 1.18 0 0.21 0.72 1.47 0 Validation Study – Rating Scale Analysis Criterion Met? Essential? N > 10 for each response category Yes Yes Unimodal distribution for each response category Yes Yes Average measures increase with category Yes Yes Outfit MNSQ < 2 for each category Yes Yes Category thresholds increase with category Yes No Measure implies Category & Category implies Measure (Coherence) No No Category thresholds increase by 0.81 logits No No Category thresholds don’t increase by more than 5 logits Yes No Validation Study - Reliability Reliability of Separation - Analogous to Cronbach’s Alpha Composite Score: 0.94 Subscales: “Reading the Data” (Level A): 0.87 “Reading Between the Data” (Level B): 0.91 The High School Grades SETS Instrument • Items completed Spring 2012 • Based on both GAISE (all levels) and two of the four strands for CCSSM for High School Statistics & Probability • Interpreting Categorical & Quantitative Data • Making Inferences & Justifying Conclusions • Data collection • In-service: Summer 2012 • Pre-service: Fall 2012 & Spring 2013 • Analysis during Spring/Summer 2013 The High School Grades SETS Instrument 44 Total items: • 26 items from Middle Grades SETS instrument • 18 “new” items based on level C of GAISE and the two strands of the CCSSM Item format: • Please rate your confidence in teaching high school students the skills necessary to complete the following tasks successfully: • Scale of 1 to 6 14 Demographic items CCSSM “Making Inferences & Justifying Conclusions” Strand Selected SETS High School Items Based on GAISE Level C and CCSSM Understand and evaluate random processes underlying statistical experiments • • S-IC.1. Understand statistics as a process for making inferences about population parameters based on a random sample from that population. S-IC.2. Decide if a specified model is consistent with results from a given data-generating process, e.g., using simulation. Make inferences and justify conclusions from sample surveys, experiments, and observational studies • • • • S-IC.3. Recognize the purposes of and differences among sample surveys, experiments, and observational studies; explain how randomization relates to each. S-IC.4. Use data from a sample survey to estimate a population mean or proportion; develop a margin of error through the use of simulation models for random sampling. S-IC.5. Use data from a randomized experiment to compare two treatments; use simulations to decide if differences between parameters are significant. S-IC.6. Evaluate reports based on data. • Evaluate whether a specified model is consistent with data generated from a simulation. • Explain the role of randomization in surveys, experiments and observational studies. • Estimate a population mean or proportion using data from a sample survey. • Evaluate how well the conclusions of a study are supported by the study design and the data collected. Potential Uses • Research • Assessment of Teacher Preparation Programs • Analysis of Need for In-Service Professional Development Programs Current Projects Using SETS Instruments Jean Linner Lassiter High School & GaDOE/GCTM Academy “Recognizing that low teacher efficacy can inhibit effective teaching as well as student learning, we hope to use the SETS instrument for high school in a longitudinal study to identify and target those professional learning experiences that increase teacher efficacy for teaching statistics.” Current Projects (continued) Stephanie Casey & Andrew Ross Eastern Michigan University Scores from the SETS for pre-service secondary mathematics teachers enrolled in a statistical methods course were used to assess the pre-service teachers' confidence in teaching statistical concepts in general. Additionally, scores were used to compare the treatment (reform-oriented instruction) section and the control (traditional instruction) section. The results will be used to inform the instruction of the course and assess whether there was a significant difference between the treatment and control groups with respect to their self-efficacy for teaching statistics. Current Projects (continued) Salome Martínez & Eugenio Chandía Universidad de Chile & Pontificia Universidad Católica de Chile The government of Chile is funding a curriculum project for the preparation of future elementary teachers. The curriculum consists of a series of textbooks for all the content areas, including the Data and Chance strand of the national standards. The SETS instrument will be used to evaluate the impact of the implementation of the textbooks in 10 Chilean institutions of teacher preparation. Information on Using SETS Instruments E-mail Rebecca Pierce at rpierce@bsu.edu to request “Terms of Use” form Thank you for attending our webinar! We’d like to open the floor to discuss the following: • How would you envision using one of the SETS instruments? • What else would you like to know about the SETS instruments? Related Papers Hilton, S., Kaplan, J., Hooks, T., Harrell, L. M., Fisher, D., & Sorto, M. A. (2008). Collaborative projects in statistics education. In Proceedings of the 2008 Joint Statistical Meetings, Section on Statistical Education (pp. 752-756). Alexandria, VA: American Statistical Association. Harrell, L. M., Pierce, R. L., Sorto, M. A., Murphy, T. J., Lesser, L. M., & Enders, F. B. (2009). On the importance and measurement of pre-service teachers’ efficacy to teach statistics: Results and lessons learned from the development and testing of a GAISE-based instrument. In Proceedings of the 2009 Joint Statistical Meetings, Section on Statistical Education (pp. 3396-3403). Alexandria, VA: American Statistical Association. Sorto, M. A., Harrell, L. M., Pierce, R. L., Murphy, T. J., Enders, F. B., & Lesser, L. M., (2010). Experts’ perceptions in linking GAISE guidelines to the self-efficacy to teach statistics instrument. In Proceedings of the 2010 Joint Statistical Meetings, Section on Statistical Education (pp. 4289-4294). Alexandria, VA: American Statistical Association. Harrell-Williams, L. M., Sorto, M. A., Pierce, R. L., Lesser, L. M., & Murphy, T. J. (Under Review). Validation of Scores from a New Measure of Pre-service Teachers’ Self-efficacy to Teach Statistics in the Middle Grades. REFERENCES Czerniak, C. M. (1990). A study of self-efficacy, anxiety, and science knowledge in preservice elementary teachers. Paper presented at the National Association for Research in Science Teaching, Atlanta, GA. Enochs, L.G., Smith, P.L., Huinker, D. (2000). Establishing factorial validity of the mathematics teaching efficacy beliefs instrument. School Science and Mathematics, 100, 194-202. Finney, S. J., & Schraw, G. (2003). Self-efficacy beliefs in college statistics courses. Contemporary Educational Psychology, 28, 161-186. Franklin, C., Kader, G., Mewborn, D., Moreno, J., Peck, R., Perry, M., & Scheaffer, R. (2007). Guidelines for Assessment and Instruction in Statistics Education (GAISE) Report: A Pre-K-12 Curriculum Framework. Alexandria, VA: American Statistical Association. (Also available at http://www.amstat.org/education/gaise/) Friel, S. N., Curcio, F. R., & Bright, G. W. (2001). Making sense of graphs: Critical factors influencing comprehension and instructional implications. Journal for Research in Mathematics Education, 32(2), 124-158. REFERENCES (continued) Lutzer, D., Rodi, S., Kirkman, E., & Maxwell, J. Statistical Abstract of Undergraduate Programs in the Mathematical Sciences in the United States, Fall 2005, CBMS Survey, American Mathematical Society, Providence, R.I., 2007. National Governors Association (2010). Common Core State Standards for Mathematics. http://www.corestandards.org/assets/CCSSI_Math%20Standards.pdf Riggs, I. M., & Enochs, L. G. (1990). Toward the development of an elementary teacher's science teaching efficacy belief instrument. Science Education, 74(6), 625-637. Watson, J. (2001). Profiling Teachers’ Competence and Confidence to Teach Particular Mathematics Topics: The Case of Chance and Data. Journal of Mathematics Teacher Education, 4, 305–337. Wenta, R. G. (2000). Efficacy of preservice elementary mathematics teachers. Unpublished doctoral dissertation, Indiana University.