Syllabus PIE 2030 – Experimental Design Instructor: Clement A. Stone Office: 5920 Posvar Hall; 624-9359; email: cas@pitt.edu Description The course is designed to introduce students to different methods in experimental design. Topics include characteristics of experimental research, steps for implementing an experiment, design issues as they related to internal and external validity, classification of experimental designs, sampling, and design techniques such as blocking, analysis of covariance, and assessment of change. Other research methods will be discussed including survey research, meta-analysis, and quasiexperimental designs. General course objectives include: 1) Understand the goals of experimental design and causal inference 2) Understand how experimental designs differ from other research 3) Understand what specific experimental design techniques are intended to accomplish 4) Understand connections between specific experimental design techniques and statistical analyses 5) Understand different quasi-experimental designs, what the designs are intended to accomplish 6) Be able to recognize strengths and weaknesses of various research design practices including experimental research, survey research, meta-analysis, and quasi-experimental research Course Prerequisites Introduction to Research Methodology (PIE2001 or equivalent) Introductory Statistics course (PIE2018 or equivalent) Students should have an understanding of simple statistical concepts (mean, variance, normal distribution, correlation) as well as an understanding of basic methods such as linear regression, t-test, and hypothesis testing. Course Readings Required: Shadish, Cook, & Campbell. Experimental and Quasi-experimental Designs for Generalized Causal Inference. Publisher: Wadsworth, Cengage Learning (2002). Recommended: Christensen, Burke, & Turner. Research Methods, Design, and Analysis, 11th Edition Publisher: Pearson (2010). Course Evaluation Homework (contributing 20% to the final grade) will be assigned for each topic area. For each day that answers to an exercise are submitted late, 20% of points possible will be deducted. There will be two examinations (midterm, final each contributing 40% to the final grade). Each exam will consist of short answer and multiple-choice type questions. Exams are closed book. Class participation in discussions is expected. Schedule (tentative) Shadish et. al. Readings in Christenson et. al. Ch. 1-3 Ch. 1, 2, 6, Reading Readings in Week 1- Introduction to course – review of concepts Review of Statistical Concepts (Ch. 14-15) 2- Experimental design 3- MLK Day 4- Experimental design (cont.) 5- Issues in analyzing data from experimental designs 6- Issues in analyzing data from experimental designs (cont.) 7- Issues in analyzing data from experimental designs (cont.) 8- Midterm 9- Meta-analysis Ch. 8 Ch. 7,8 Reading Ch. 15 Ch. 13, Readings pg. 17 11 - Introduction to sampling Ch. 9 Ch. 5 12 - Determining sample size Reading Part of Ch. 9 13 - Quasi-experimental research and causal inference revisited pgs. 13-17, Ch. 4-7 Ch 11 Ch. 10 Reading Ch. 12 10 – Spring Break 14 - Quasi-experimental research (cont.) 15 - Survey research 16 - Final Exam Other Useful Readings American Educational Research Association. (2006). Standards for reporting on empirical social science research in AERA publications. Educational Researcher, 35(6), 33 – 40. Aivazidis, C., Lazaridou, M., & Hellden, G. F. (2006). A comparison between a traditional and an online environmental educational program. The Journal of Environmental Education, 37(4), 45 – 54. Batiuk, M. E., Boland, J. A., & Walker, N. (2004). Project trust: Breaking down barriers between middle school children. Adolescence, 39(155), 531 – 538. 7P:220. Briggs, D. C., Ruiz-Primo, M. A., Furtak, E., & Shepard, L. (2012). Meta-analytic methodology and inferences about the efficacy of formative assessment. Educational Measurement: Issues and Practice, 31(4), 13 – 17. de Anda, D. (2006). Baby think it over: Evaluation of an infant simulation intervention for adolescent pregnancy prevention. Health & Social Work, 31(1), 26 – 35. Feinman, R. D. (2009). Intention-to-treat: What is the question? Nutrition & Metabolism, 6(1). Han, H., Lee, J., Kim, J., Hedlin, H. K., Song, H., & Kim, M. T. (2009). A meta-analysis of interventions to promote mammography among ethnic minority women. Nursing Research, 58(4), 246 – 254. Harwell, M., & LeBeau, B. (2010). Student eligibility for a free lunch as an SES measure in education research. Educational Researcher, 39(2), 120-131. Hughes, E. K., Gullone, E., Dudley, A., & Tonge, B. (2010). A case-control study of emotion regulation and school refusal in children and adolescents. Journal of Early Adolescence, 30(5), 691 – 706. Kingston, N., & Nash, B. (2011). Formative assessment: A meta-analysis and a call for research. Educational Measurement: Issues and Practice, 30(4), 28 – 37. Kingston, N., & Nash, B. (2012). How many formative assessment angels can dance on the head of a meta-analytic pin: .2.. Educational Measurement: Issues and Practice, 31(4), 18 – 19. Leake, M., & Lesik, S. A. (2007). Do remedial English programs impact first-year success in college? An illustration of the regression-discontinuity design. International Journal of Research & Method in Education, 30(1), 89 – 99. Liu, Y. (2007). A comparative study of learning styles between online and traditional students. Journal of Educational Computing Research, 37(1), 41 – 63. Miller, G. A., & Chapman, J. P. (2001). Misunderstanding Analysis of covariance. Journal of Abnormal Psychology, 110(1), 40 – 48. Roberts, J. V., & Gepotys, R. J. (1992). Reforming rape laws: Effects of legislative change in Canada. Law and Human Behavior, 16(5), 555 – 573. Robinson, T. N., Wilde, M. L., Navracruz, L. C., Haydel, K. F., & Varady, A. (2001). Effects of reducing children’s television and video game use on aggressive behavior. Archives of Pediatric and Adolescent Medicine, 155, 17 – 23. Russo, M. W. (2007). How to review a meta-analysis. Gastroenterology & Hepatology, 3(8), 637 – 642. Shavelson, R. J., & Towne, L. (Eds.). (2002). Scientific research in education. Washington, DC: National Academy Press. Simmons, J. P., Nelson, L. D., & Simonsohn, U. (2011). False-positive psychology: Flexibility in data collection and analysis allows presenting anything as significant. Psychological Science, 22(11), 1359 – 1366. Steiner, P. M., Wrolewski, A., & Cook, T. D. (2009). Randomized experiments and quasiexperimental research designs in educational research. In K. E. Ryan, & J. B. Cousins (Eds.), The Sage International Handbook on Educational Evaluation. Thousand Oaks, CA: Sage. Sterner, W. R. (2011). What is missing in counseling research? Reporting missing data. Journal of Counseling & Development, 89, 56 – 62. Swern, A. S. (2010). A story of evidence-based medicine: Hormone replacement therapy and coronary heart disease in postmenopausal women. Chance, 23(3), 52 – 56. Tracz, S. M., Nelson, L. L., Newman, I., & Beltran, A. (2005). The misuse of ANCOVA: The academic and political implications of type IV errors in studies of achievement and socioeconomic status. Multiple Linear Regression Viewpoints, 31(1), 16 – 21. West, S. G., & Thoemmes, F. (2010). Campbell’s and Rubin’s perspectives on causal inference. Psychological Methods, 15(1), 18 – 37. Other Sources American Educational Research Association. (2000). Ethical Standards of the American Educational Research Association. Washington, DC: Author. American Psychological Association. (2010). Publication manual of the American psychological association (6th ed.). Washington, DC: Author. American Psychological Association. (2002). Ethical principles of psychologists and code of conduct. American Psychologist, 57, 1060 – 1073. Armijo-Olivo, S., Warren, S., & Magee, D. (2009) Intention to treat analysis, compliance, dropouts and how to deal with missing data in clinical research: A review. Physical Therapy Reviews, 14(1), 36 – 49. Berk, R. A., & Freedman, D. (2005). Statistical assumptions as empirical commitments. In T. G. Blomberg & S. Cohen (Eds.), Punishment, and Social Control: Essays in Honor of Sheldon Messinger (2nd ed.; pp. 235 – 254). Piscataway, NJ: Aldaine. Campbell, D. T., & Fiske, D. W. (1959). Convergent and discriminant validation by the multitrait-multimethod matrix. Psychological Bulletin, 56(2), 81 – 105. Cohen, J. (1990). Things I have learned (so far). American Psychologist, 45(12), 1304 – 1312. Cohen, J. (1994). The earth is round (p < .05). American Psychologist, 49(12), 997 – 1003. Cole, J. C. (2008). How to deal with missing data: Conceptual overview and details for implementing two modern methods. In J. W. Osborne (Ed.), Best practice is quantitative methods (pp. 214 – 238). Thousand Oaks, CA: Sage. Cook, T. D. (2002). Randomized experiments in educational policy research: A critical examination of the reasons the educational community has offered for not doing them. Educational Evaluation and Policy Analysis, 24(3), 175 – 199. Cook, T. D. (2008). Waiting for life to arrive: A history of the regression-discontinuity design in psychology, statistics and economics. Journal of Econometrics, 142, 636 – 654. Cowles, M., & Davis, C. (1982). On the origins of the .05 level of statistical significance. American Psychologist, 37(5), 553 – 558. DeVaney, T. A. (2001). Statistical significance, effect size, and replication: What do the journals say?. The Journal of Experimental Education, 69(3), 310 – 320. Freedman, D. (1999). From association to causation: Some remarks on the history of statistics. Statistical Science, 14(3), 243 – 258. 7P:220. Hall, N. S. (2007). R. A. Fisher and his advocacy of randomization. Journal of the History of Biology, 40, 295 – 325. Hall, M. V., Wilson, L. A., & Sanger, M. J. (2012). Student success in intensive versus traditional introductory college chemistry courses. Journal of Chemical Education, 89, 1109 – 1113. Holland, P. W. (1986). Statistics and causal inference. Journal of the American Statistical Association, 81(396), 945 – 960. Judd, C. M., McClelland, G, H, & Culhane, S. E. (1995). Data analysis: Continuing issues in the everyday analysis of psychological data. Annual Review of Psychology, 46, 433 – 465. Kenny, D. A., & Judd, C. M. (1986). Consequences of violating the independence assumption in analysis of variance. Psychological Bulletin, 99(3), 422 – 431. Lehrer, J. (2010, December 13). The truth wears off: Is there something wrong with the scientific method? The New Yorker. Retrieved from http://www.newyorker.com/reporting/2010/12/13/101213fa_fact_lehrer. Lohman, D. F., & Korb, K. A. (2006). Gifted today but not tomorrow: Longitudinal changes in ability and achievement during elementary school. Journal for the Education of the Gifted, 29(4), 451 – 484. Lord, F. M. (1967). A paradox in the interpretation of group comparisons. Psychological Bulletin, 68(5), 304 – 305. Lord, F. M. (1969). Statistical adjustments when comparing preexisting groups. Psychological Bulletin, 72(5), 336 – 337. Messick, S. (1995). Validity of psychological assessment: Validation inferences from persons’ responses and performances as scientific inquiry into score meaning. American Psychologist, 50(9), 741 – 749. Osborne, J. W. (2000). Advantage of hierarchical linear modeling. Practical Assessment, Research & Evaluation 7(1). Retrieved January 23, 2008 from http://PAREonline.net/getvn.asp?v=7&n=1. Pearl, J. (2009). Causal inference in statistics: An overview. Statistics Surveys, 3, 96 – 146. Robison, W., Boiisjoly, R., Hoeker, D., & Young, S. (2002). Representation and misrepresentation: Tufte and the Morton Thiokol engineers on the Challenger. Science and Engineering Ethics, 8, 59 – 81. Rodriguez, M. C. (2005). Three options are optimal for multiple-choice items: A meta-analysis of 80 years of research. Educational Measurement: Issues and Practice, 24(2), 3 – 13. Rubin, D. B. (2008). For objective causal inference, design trumps analysis. The Annals of Applied Statistics, 2(3), 808 – 840. Schneider, B. (2004). Building a scientific community: The need for replication. Teachers College Record, 106(7), 1471 – 1483. Schooler, J. (2011). Unpublished results hide the decline effect. Nature, 470, 437. Schreiber, J. B., & Griffin, B. W. (2004). Review of multilevel modeling and multilevel studies in The Journal of Educational Research (1992-2002). The Journal of Educational Research, 98(1), 24 – 33. Shadish, W. R. (1995). Philosophy of science and the quantitative-qualitative debates: Thirteen common errors. Evaluation and Program Planning, 18(1), 63 – 75. 7P:220. Shadish, W. R., Clark, M. H., & Steiner, P. M. (2008). Can nonrandomized experiments yield accurate answers? A randomized experiment comparing random and nonrandom assignments. Journal of the American Statistical Association, 103(484), 1334 – 1356. Shadish, W. R., Cook, T. D. (1999). Design rules: More steps toward a complete theory of quasiexperimentation. Statistical Science, 14, 294 – 300. Shadish, W. R., & Cook, T. D. (2009). The renaissance of field experimentation in evaluating interventions. Annual Review of Psychology, 60, 607 – 629. Shadish, W. R., Galindo, R., Wong, V. C., Steiner, P. M., & Cook, T. D. (2011). A randomized experiment comparing random and cutoff-based assignment. Psychological Methods, 16(2), 179 – 191. Shea, C. (2011, November 13). Fraud scandal fuels debate over practices of social psychology. The Chronicle of Higher Education. Retreived from http://chronicle.com/article/As-DutchResearch-Scandal/129746/. Sheiner, L. B., & Rubin, D. (1995). Intention-to-treat analysis and the goals of clinical research. Clinical Pharmacology & Therapeutics, 57(1), 6 – 15. Shields, S. P., Hogrebe, M. C., Spees, W. M., Handlin, L. B., Noelken, G. P., Riley, J. M., & Frey, R. F. (2012). A transition program for underprepared students in general chemistry: Diagnosis, implementation, and evaluation. Journal of Chemical Education, 89, 995 – 1000. Smith, M. L., & Glass, G. V. (1977). Meta-analysis of psychotherapy outcomes. American Psychologist, 32, 752 – 760. Thistlewaite, D. L., & Campbell, D. T. (1960). Regression-discontinuity analysis: an alternative to the ex post facto experiment. The Journal of Educational Psychology, 51(6), 309 – 317. Wainer, H. (1992). Understanding graphs and tables. Educational Researcher 21(1), 14 – 23. Wainer, H. (1999a). One cheer for null hypothesis significance testing. Psychological Methods, 4(2), 212 – 213. Wainer, H. (1999b). Is the Akebono school failing its best students? A Hawaiian adventure in regression. Educational Measurement: Issues and Practice, 18(3), 26 – 31. Wainer, H., & Brown, L. M. (2004). Two statistical paradoxes in the interpretation of group differences: Illustrated with medical school admission and licensing data. The American Statistician, 58(2), 117 – 123. Wainer, H., & Robinson, D. H. (2003). Shaping up the practice of null hypothesis significance testing. Educational Researcher, 32(7), 22 – 30. Wilkinson, J., & the Task Force on Statistical Inference. (1999). Statistical methods in psychology journals. American Psychologist, 54, 594 – 604. Other Useful Books Campbell, D. T., & Stanley, J. C. (1963). Experimental and quasi-experimental designs for research. Boston, MA: Houghton Mifflin. Booth, W. C., Colomb, G. G., & Williams, J. M. (2003). The craft of research (2nd ed.). Chicago, IL: The University of Chicago Press. 7P:220. Cook, T. D, & Campbell, D. T. (1979). Quasi-experimentation: Design and analysis for field settings. Boston, MA: Houghton Mifflin. Freedman, D. (2010). Statistical models and causal inference: A dialogue with the social sciences (D. Collier, J. S. Sekhin, & P. B. Stark; Eds.). New York, NY: Cambridge. Murnane, J. R., & Willett, J. B. (2011). Methods Matter: Improving Causal Inference in Educational and Social Science Research. New York, NY: Oxford University Press. Strunk, W., & White. (2000). The elements of style (4th ed.). Needham Heights, MA: Allyn & Bacon. Tufte, E. R. (2001). The visual display of quantitative information (2nd ed.). Cheshire, MA: Graphics Press.