Day 6: Non-Experimental & Experimental Design Where are the beakers?? What kind of research is considered the “gold standard” by the Institute of Education Sciences? A. B. C. D. Descriptive Causal-Comparative Correlational Experimental Why? Why does most educational research use non-experimental designs? What is the purpose of non-experimental designs? Causal-Comparative Example Green & Jaquess (1987) – Interested in the effect of high school students’ part-time employment on their academic achievement. – Sample: 477 high school juniors who were unemployed or employed > 10 hours/wk. Causal-Comparative Design A study in which the researcher attempts to determine the cause, or reason, for preexisting differences in groups of individuals At least two different groups are compared on a dependent variable or measure of performance (called the “effect”) because the independent variable (called the “cause”) has already occurred or cannot be manipulated Causal-Comparative Design Ex-post facto – Causes studied after they have exerted their effect on another variable. Causal-Comparative Design Drawbacks – Difficult to establish causality based on collected data. – Unmeasured variables (confounding variables) are always a source of potential alternative causal explanations. Some Thought Questions… Correlational Design Determines whether and to what degree a relationship exists between two or more quantifiable variables. Correlational Design The degree of the relationship is expressed as a coefficient of correlation Examples – Relationship between math achievement and math attitude – Relationship between degree of a school’s racial diversity and student use of stereotypical language – Your topics? Correlation coefficient… -1.00 strong negative 0.00 +1.00 strong positive no relationship Advantages of Correlational Design Analysis of relationships among a large number of variables in a single study Information about the degree of the relationship between the variables being studied Cautions A relationship between two variables does not mean one causes the other (Think about the reading achievement and body weight correlations) Possibility of low reliability of the instruments makes it difficult to identify relationships Cautions Lack of variability in scores (e.g. everyone scoring very, very low; everyone scoring very, very high; etc.) makes it difficult to identify relationships Large sample sizes and/or using many variables can identify significant relationships for statistical reasons and not because the relationships really exist (Avoid shotgun approach) Cautions Need to identify your sample to know what is actually being compared. If using predictor variables, time interval between collecting the predictor and criterion variable data is important. Correlational Designs Guidelines for interpreting the size of correlation coefficients – Much larger correlations are needed for predictions with individuals than with groups Crude group predictions can be made with correlations as low as .40 to .60 Predictions for individuals require correlations above .75 Correlational Designs Guidelines for interpreting the size of correlation coefficients – Exploratory studies Correlations of .25 to .40 indicate the need for further research Much higher correlations are needed to confirm or test hypotheses Correlational Designs Criteria for evaluating correlational studies – Causation should not be inferred from correlational studies – Practical significance should not be confused with statistical significance – The size of the correlation should be sufficient for the use of the results (individuals vs groups) Think… If you were going to take your action research topic, and create a causalcomparative study, what would it look like? --OR-If you were going to take your action research project, and create a correlational study, what would it look like? Experimental Design The Gold Standard? To Review Why is most educational research comprised of non-experimental research designs? To Review What is the purpose of non-experimental research? To Review How does the independent variable function in non-experimental research? To Review Can non-experimental research claim causality? An example Read the example given in class and in pairs respond to the questions Experimental Research Purpose – To make causal inferences about the relationship between the independent and dependent variables Characteristics – Direct manipulation of the independent variable – Control of extraneous variables Experimental Designs Single Group Post-test Single Group Pre-test Post-test Non-Equivalent Groups Post-test Quasi-Experimental Design Randomized Post-test only Randomized Pre-test Post-test Factorial Examples Experimental Validity Internal validity – The extent to which the independent variable, and not other extraneous variables , produced the observed effect on the dependent variable External validity – The extent to which the results are generalizable Internal Validity Threats that reduce the level of confidence in any causal conclusions Key Question: Is this a plausible threat to the internal validity of the study? Threats to Internal Validity History – Extraneous events have an effect on the subjects’ performance on the dependent variable – Ex - The crash of the stock market, 9-11, the invasion of Iraq, etc. Selection – Groups that are initially not equal due to differences in the subjects in those groups – Ex - Positive and negative attitudes, high and low achievers, etc. Threats to Internal Validity Maturation – Changes experienced within the subject over time Pretesting – The effect of having taken a pretest Instrumentation – Poor technical quality (i.e. validity, reliability) or changes in instrumentation Threats to Internal Validity Subject attrition – Differential loss of subjects from groups Statistical regression – The natural movement of extreme scores toward the mean Diffusion of treatment – The treatment is given to the control group Experimenter effects – Different characteristics or expectations of those implementing the treatments across groups Threats to Internal Validity Subject effects – The effects of being aware that one is involved in a study – Four types Hawthorne effect John Henry effect Resentful demoralization Novelty effect Internal Validity Key Point: Ultimately, validity is a matter of judgment. Ask if it is reasonable that possible threats are likely to affect the results. External Validity The extent to which results can be generalized from a sample to a particular population. Question – Why would really good internal validity often result in poor external validity? External Validity Factors affecting external validity – Subjects Representativeness of the sample in comparison to the population Personal characteristics of the subjects – Situations - characteristics of the setting Specific environment Special situation Particular school External Validity Importance of explanation of sampling procedures Experimental Designs Single Group Post-test Single Group Pre-test Post-test – Libby, Deb Non-Equivalent Groups Post-test – Mary, Cheryl Quasi-Experimental Design – Pete, Laura Randomized Post-test only – Amanda, Nicole, Tam Randomized Pre-test Post-test – Karen, Jen, Justin Examples Your Task Based on the topic of your proposal, design an experimental study using the design you were assigned. – Write a research question and hypothesis. – Sketch out the methods. Identify strengths and weaknesses of each design. Experimental Designs Notation – R indicates random selection or random assignment – O indicates an observation Test Observation score Scale score – X indicates a treatment – A, B, C, ... indicates a group Pre-Experimental Designs No pre-experimental design controls internal validity threats well Single group pretest only –A X O – Internal validity threats History, maturation, attrition, experimenter effects, subject effects, and instrumentation are viable threats Useful only when the research is sure of the status of the knowledge, skill, or attitude being changed and there are no extraneous variables affecting the results Pre-Experimental Designs Single group pretest post-test – AOXO – Internal validity threats Maturation and pretesting are threats History and instrumentation are potential threats – Useful when subject effects will not influence the results, history effects can be minimized, and multiple pretests and post-tests are used Pre-Experimental Designs Non-equivalent groups post-test only –A X O B O – Internal validity threats Definite Threat: Selection Potential Threats: History, maturation, and instrumentation – Useful when groups are comparable and subjects can be assumed to be about the same at the beginning of the study Quasi-Experimental Designs Types – Non-equivalent pretest/post-test, experimental control groups AOXO BO O – Non-equivalent pretest/post-test, multiple treatment groups A O X1 O B O X2 O Useful when subjects are in pre-existing groups (e.g. classes, schools, teams, etc.) Quasi-Experimental Designs Threats to internal validity – Selection is the major concern – Likely to control for most other threats, provided the groups are not significantly different from one another – See Table 9.2 for specific threats related to each design True Experimental Designs Important terminology – Random assignment Subjects placed into groups by random Ensures equivalency of the groups – Random selection of subjects Subjects chosen from population by random Ensures generalizability to the population from which the subjects were selected (i.e. external validity) True Experimental Designs Types – Randomized post-test only experimental control groups RA RB X O O – Randomized post-test only multiple treatment groups R A R B X1 O X2 O True Experimental Designs Types (continued) – Randomized pretest/post-test multiple treatment groups R A O X1 O R B O X2 O – Randomized pretest/post-test experimental control groups R A O X O R B O O True Experimental Designs Threats to internal validity – Controls for selection, maturation, and statistical regression – Likely to control for most other threats – See Table 9.2 for specific threats related to each design Evaluating Experimental Designs Criteria for evaluating experimental research – The primary purpose is to test causal hypotheses – There should be direct manipulation of the independent variable – There should be clear identification of the specific research design Evaluating Experimental Designs Criteria for evaluating experimental research – The design should provide maximum control of extraneous variables – Treatments are substantively different from one another – The number of subjects is dependent on or equal to the number of treatment replications