Experimental Studies - Villanova University

advertisement
Educational Research:
Experimental Studies
EDU 8603
Educational Research
Richard M. Jacobs, OSA, Ph.D.
Research...

The systematic application of a
family of methods employed to
provide trustworthy information
about problems
…an ongoing process based on many
accumulated understandings and
explanations that, when taken together
lead to generalizations about problems
and the development of theories
The basic steps of research...
Scientific and disciplined inquiry is an
orderly process, involving:
 recognition and identification of a
topic to be studied (“problem”)
 description and execution of
procedures to collection information
(“method”)
 objective data analysis
 statement of findings (“results”)
Research methods...
Quantitative…
…collects and analyzes numerical data
obtained from formal instruments
Quantitative methods...
descriptive research (“survey research”)
 correlational research
 causal-comparative research
(“ex post facto research”)
 experimental research


experimental research
…the researcher selects participants
and divides them into two or more
groups having similar characteristics
and, then, applies the treatment(s) to
the groups and measures the effects
upon the groups
Conducting an experimental study…
1. select the problem
2. select participants and instrument
3. selection and execution of a
research plan
4. data analysis and formulation of
conclusions
Types of experimental comparison…
1. comparison of two different approaches
(A versus B)
2. comparison of an existing approach
to a new approach (A and ~ A)
3. comparison of differing
amounts of a single approach
(A and a or a and A)
where:
A – experimental (“treatment”) group
B – control (“no treatment,”
“nonmanipulated”) group
Variable...
…a concept (e.g., intelligence, height,
aptitude) that can assume any one of
a range of values
Research variables...
Independent…
…an activity of characteristic believed
to make a difference with respect to
some behavior
…(syn.) experimental variable, active
variable, cause, treatment
Dependent…
…the change or difference occurring a
a result of the independent variable
…(syn.) criterion variable, assigned
variable, effect, outcome, posttest
Confounding…
…the fact that the effects of the
independent variable may intertwine
with extraneous variables, such that
it is difficult to determine the unique
effects of each variable
…(syn.) criterion variable, assigned
variable, effect, outcome, posttest
Group experimental designs…
1. single-variable
2. factorial

single-variable designs
…involve one manipulated independent
variable
pre-experimental
quasi-experimental
true experimental

types of pre-experimental designs
one-shot case study
X O
…a single group exposed to a
treatment ( X ) and then
posttested ( O )
one-group pretest-posttest design
O X O
…a single group is pretested ( O ),
exposed to a treatment ( X ) and,
then, is posttested ( O )
static group comparison
X1 O
X2 O
…involves at least two groups ( X ), one
receiving a new, or experimental
treatment ( X1 ) and another receiving a
traditional, or control treatment ( X2 )
and, then, are posttested ( O )

types of quasi-experimental designs
nonequivalent control group
design
O X O
…random assignment of intact groups
that are pretested ( O ), exposed to a
treatment ( X ) and then posttested
(O)
time-series design
O O O O X O O O O
…a single group is pretested ( O )
repeatedly until pretest scores are
stable, exposed to a treatment ( X )
and, then, is repeatedly posttested
(O)
counterbalanced design
X1 O X2 O X3 O
X3 O X1 O X2 O
X2 O X3 O X1 O
…all of the groups receive all
treatments but in a different order;
the number of groups and
treatments must be equal

types of true experimental designs
pretest-posttest control group design
R O X1 O
R O X2 O
…at least two groups are formed by
random assignment ( R ),
administered a pretest ( O ), receive
different treatments ( X1, X2 ), are
administered a posttest, and posttest
scores are compared to determine
effectiveness of treatments
posttest-only control group design
R
X1 O
R
X2 O
…at least two groups are formed by
random assignment ( R ), receive
different treatments ( X1, X2 ), are
administered a posttest, and posttest
scores are compared to determine
effectiveness of treatments
Solomon four-group design
R O X1 O
R O X2 O
R
X1 O
R
X2 O
…four groups are formed by random
assignment ( R ) of participants, two
groups are pretested ( O ) and two are not,
one pretested and one unpretested group
receive the experimental treatments ( X1,
X2 ), each group is are administered a
posttest on the dependent variable, and
posttest scores are compared to
determine effectiveness of treatments

factorial designs
…involve two or more independent
variables with at least one
independent variable being
manipulated by the researcher

examples of factorial designs
two-by-two factorial design (four cells)
2 X 2
…two types of factors (e.g., method of
instruction) each of which has two
levels (e.g., traditional vs. innovative)
A 2 X 2 factorial design…
Dependent Variable
manipulated
Independent Variable
A
B
O
Group #1
Group #2
O
Group #3
Group #4
not manipulated
Cells
A 2 X 2 factorial design…
A
No interaction
between factors
B
A 2 X 2 factorial design…
No interaction
between factors
A
B
A 2 X 2 factorial design…
A
Interacting
factors
B
A 2 X 2 factorial design…
B
Interacting
factors
A
two-by-three factorial design (six cells)
2 X 3
…two types of factors (e.g., motivation;
interest) each of which has three levels
(e.g., high, medium, low)
Single-subject experimental designs…
1. A – B – A withdrawal
2. multiple baseline designs
3. alternating treatments designs

simple A – B design
…baseline measurements ( O ) are
repeatedly made until stability is
established, then the treatment ( X ) is
introduced and an appropriate number
of measurements ( O ) are made during
treatment implementation

simple A – B design
O O O
X O X O X
baseline
phase
treatment
phase
A
|
B
O

A – B – A withdrawal designs
…baseline measurements ( O ) are repeatedly
made until stability is established, then the
treatment ( X ) is introduced and an
appropriate number of measurements ( O )
are made during treatment implementation,
followed by an appropriate number of
baseline measurements ( O ) to determine
stability of treatment ( X )

A – B – A withdrawal designs
O O O
X O X O X
O O O
baseline
phase
treatment
phase
baseline
phase
A
|
B
|
A

multiple-baseline designs
…used when a return to baseline
conditions is difficult or impossible
since treatment effects oftentimes do
not disappear when a treatment is
removed
…“multiple” refers to the study of more
than one behavior, participant, or
setting
…instead of collecting baseline data on
one specific behavior, data are
collected on: (1) several behaviors for
one participant, (2) one behavior for
several participants, or (3) one
behavior and one participant in several
settings
…then, over a period of time, the
treatment is systematically applied to
each behavior (or participant, or
setting) one at a time until all
behaviors (or participants or settings)
have been exposed to the treatment

multiple baseline design
Example: one treatment for three behaviors in three settings
A
behavior 1
behavior 2
behavior 3
B
O O OXOXOXOXOXOXOXOXOXOXOXO
O O O O O
OXOXOXOXOXOXOXO
O O O O O
O O O
OXOXOXO
the baseline remains
the same…
setting #1
setting #2
setting #3
…while the treatment
is applied at other
settings

alternating treatments design
…the relatively rapid and random
alternation of treatments ( T1 T2 …Tn )
for a single subject but not
necessarily within fixed intervals of
time
T1 - T2 - T1 - T1 - T2 - T1 - T1 - T2
The concept of validity…
…the experiment tests the variable(s)
that it purports to test
Threats to validity…
…internal: factors other than the
independent variable that affect the
dependent variable
…external: factors that affect the
generalizability of the study to groups
and settings beyond those of the
experiment
Threats to internal validity…
1. history
2. maturation
3. testing
4. instrumentation
5. statistical regression
6. differential selection of participants
7. mortality
8. selection-maturation interaction

history
…the occurrence of events that are not
part of the experimental treatment
but that occur during the study and
affect the dependent variable

maturation
…the physical, intellectual, and
emotional changes that occur
naturally in a study’s participants
over a period of time

testing
…refers to improved scores on a
posttest as a result of having taken a
pretest

instrumentation
…the unreliability or lack of
consistency in measuring
instruments that can result in an
invalid assessment of performance

statistical regression
…the tendency of participants who
score highest on a test to score
lower on a second, similar test and
vice versa

differential selection of participants
…the outcome when already formed
groups are compared raising the
possibility that the groups were
different before a study even begins

mortality
…the case in which participants drop
out of a study which changes the
characteristics of the groups and
may significantly affect the study’s
results

selection-maturation interaction
…if already-formed groups are used in
a study, one group may profit more
(or less) from a treatment or have an
initial advantage because of
maturation, history, or testing factors
Threats to external validity…
1. pretest-treatment interaction
2. selection-treatment interaction
3. multiple treatment interference
4. specificity of variables
5. treatment diffusion
6. experimenter effects
7. reactive effects

pretest-treatment interaction
…the situation when participants
respond or react differently to a
treatment because they have been
pretested

multiple-treatment interference
…the situation when the same
participants receive more than one
treatment in succession

selection-treatment interference
…the situation when participants are
not randomly selected for treatments

specificity of variables
…the situation when a study is
conducted with (1) a specific kind of
participant; (2) is based on a
particular operational definition of
the independent variable; (3) uses
specific dependent variables; (4)
transpires at a specific time; and, (5)
under a specific set of circumstances

treatment diffusion
…the situation when different treatment
groups communicate with and learn
from each other

experimenter effects
…the situation when the researchers
present potential threats to the
external validity of their own studies

reactive arrangements
…the situation when a number of
factors associated with the way in
which a study is conducted interacts
with or shapes the feelings and
attitudes of the participants involved
Types of reactive arrangements…
…Hawthorne effect: any situation in
which participants’ behavior is
affected not by the treatment per se
but by their knowledge of participating
in a study
…compensatory rivalry: the control
group is informed that they will be the
control group for a new, experimental
study (“John Henry effect”)
…placebo effect: the situation in which
half of the participants receive no
treatment but believe they are
…novelty effect: the situation in which
participant interest, motivation, or
engagement increases simply because
they are doing something different
Controlling for extraneous
(confounding) variables…
1. randomization
2. matching
3. comparing homogeneous
groups or subgroups
4. using participants as their own
controls
5. analysis of covariance (ANCOVA)

randomization
…the process of selecting and
assigning participants in such a way
that all individuals in the defined
population have an equal and
independent chance of being
selected for the sample

matching
…a technique for equating groups on
one or more variables, usually the
ones highly related to performance
on the dependent variable (e.g.,
pairwise matching)

comparing homogeneous groups or
subgroups
…a technique to control an extraneous
variable by comparing groups that
are similar with respect to that
variable (e.g., stratified sampling)

using participants as their own controls
…exposing a single group to different
treatments one treatment at a time

analysis of covariance (ANCOVA)
…a statistical method for equating
randomly formed groups on one or
more variables by adjusting scores
on a dependent variable for initial
differences on some other variable
Data analysis and interpretation…

for single-subject research
…a visual inspection and analysis of
graphical presentations of results
…focuses upon: adequacy of the
design; an assessment of treatment
effectiveness (clinical vs. statistical
significance)
Mini-Quiz…

True and false…
…experimental researcher is the only
type of research that can test
hypotheses to establish cause-andeffect relationships
True

True and false…
…the manipulation of the independent
variable is the one characteristic that
differentiates correlational research
from other types of research
False

True and false…
…an experimental study is guided by
at least one hypothesis that states
an expected causal relationship
between two variables
True

True and false…
…the “experiment” in an experimental
study is conducted to conform or
disconfirm the research problem
False

True and false…
…experimental research has both
random selection and assignment
whereas causal-comparative
research has random selection not
assignment
True

True and false…
…it is correct to state that a control
group always receives no treatment
in experimental studies
False

True and false…
…in an experimental study, the
researcher makes every effort to
ensure that the two groups start as
equivalently as possible on all
variables except the independent
variable
True

True and false…
…in an experimental study, the
researcher collects data on the
independent variable from the
groups to determine whether there is
a significant difference between their
performances
False

True and false…
…in an experimental study, the
researcher endeavors to make the
groups as similar as possible so that
the only major difference between
them is the treatment variable(s)
manipulated by the researcher
True

True and false…
…even though experimental research
is the only type of research that can
truly establish cause-effect
relationships, it is not universally
appropriate for all research
problems or studies
True

True and false…
…an experiment is valid if results
obtained are due only to the control
variable and if they are generalizable
to individuals or contexts beyond the
experimental setting
False

True and false…
…internal validity focuses upon
threats or rival explanations that
influence the outcomes of an
experimental study but are not part
of the independent variable
True

True and false…
…if a study’s results cannot be
replicated in other settings by other
researchers, then the study has low
external reliability
False

True and false…
…to conduct a valid experiment, all the
researcher has to do is to maximize
internal and external validity
False

True and false…
…true experimental designs control for
all sources of internal and external
validity
False

True and false…
…the more narrowed and controlled
a research situation is, the less
realistic and generalizable it
becomes
True

True and false…
…a study that is not internally valid
is worthless
True

True and false…
…the longer a study lasts, the more
likely it is that history will be a
threat to the study’s internal
validity
True

True and false…
…the way a researcher controls for
selection-maturation interaction is
to select a design that controls for
potential problems or to make
every effort to determine if
selection-maturation interaction is
operating in the study
True

True and false…
…pretest-treatment interaction
makes a study’s results
generalizable only to other
pretested groups
True

True and false…
…the seriousness of pretest-treatment
interaction threat is dependent upon
the research participants, the nature
of the independent and dependent
variables, and the duration of the
study
True

True and false…
…even if intact groups are randomly
selected, the possibility exists that
the experimental group is in some
important way different from the
control group and/or from the larger
population
True

True and false…
…selection-treatment interaction is an
uncontrolled variable in research
designs involving randomization
True

True and false…
…it is possible to know what a
researcher means by the terms
assigned to variables without clear
operationalized descriptions
False

True and false…
…good designs control for many
sources of invalidity, poor designs
control few
True

True and false…
…ANCOVA is used to “correct” or
adjust posttest scores for initial
pretest differences
True

True and false…
…ANCOVA is assumes that the
relationship between the
independent and covariate variables
is curvilinear
False

True and false…
…a researcher should not totally
accept or reject a research design
because of how it controls (or fails to
control) for sources of invalidity
True

True and false…
…a researcher should decide which
research design is most appropriate
not only by the controls provided but
also by the nature of the study and
the setting in which it is to be
conducted
True

True and false…
…in a static-group comparison design,
each group serves as a control or
comparison group for the other
group(s)
True

True and false…
…in a counterbalanced design, the
average performance of the groups
on each treatment are calculated and
compared
True

True and false…
…in a true experimental designs,
participants should be randomly
selected or randomly assigned
False

True and false…
…in the pretest-posttest control group
design, if groups are essentially the
same on the dependent variable at
the start of a study, a t-test can be
used for analysis; if not, ANCOVA is
used
True

True and false…
…the more dissimilar the intact groups
in a study are, the stronger the study
is
False

True and false…
…when describing and symbolizing
factorial designs, the control variable
is placed first
False

True and false…
…the purpose of a factorial design is to
determine whether the effects of an
independent variable are
generalizable across all levels or are
specific to particular variables
True

True and false…
…factorial designs are very effective
for testing research hypotheses
that cannot be tested with a singlevariable design
True

True and false…
…for all types of research, the more
a study’s results are replicated, the
less confidence there is in the
procedures that produced those
results
False

Fill in the blank…
…the two main ways that researchers
equate groups participating in
experimental research
simple random sampling
stratified random sampling

Fill in the blank…
…the researcher decides what
treatment(s) will constitute the
independent variable and which
group will get which treatment
manipulation

Fill in the blank…
…the researcher’s efforts to remove
the influence of any extraneous
variable that might affect scores on
the dependent variable
control

Fill in the blank…
…the researcher’s concern with
threats or factors other than the
independent variable that affect the
dependent variable
internal validity

Fill in the blank…
…the researcher’s concern with the
extent to which the study results
can be generalized to groups and
settings beyond those of the
experiment
external validity

Fill in the blank…
…a threat to validity involving the
tendency for scores to more
toward a average or expected
score
statistical regression

Fill in the blank…
…a threat to internal validity that may
be overcome by testing for initial
equivalence between groups
differential selection of participants

Fill in the blank…
…the population from which a
researcher endeavors to extract a
sample
target population

Fill in the blank…
…the actual population from which a
researcher extracts a sample
accessible population

Fill in the blank…
…the influences on the part of the
experimenter over which the
experimenter has no direct control
but may affect study procedures
passive experimenter effects
experimenter personal-attributes effects

Fill in the blank…
…the situation in which a
researcher's expectation of the
study affect the researcher’s
behavior and the researcher
outcomes
active experimenter effect
experimenter bias effect

Fill in the blank…
…an external variable that effects
the independent variable and
intertwines with other extraneous
variables such that it is difficult
to determine the unique effects
of each
confounding variable

Fill in the blank…
…the term referring to a design that
has more than one independent
variable
factor(ial)

Fill in the blank…
…a study to compare the achievement
of two groups: one group taught in
an anxiety-producing environment
and one group taught in an anxietyreducing environment
experimental

Fill in the blank…
…the nonmanipulated variable in a
factorial design
control variable

Fill in the blank…
…a second study conducted by a
different investigator, with different
participants, behaviors, and settings
systematic replication

Fill in the blank…
…additional studies that add
confidence to the procedures that
produced those results
replication

Fill in the blank…
…a second study involving the
development of treatment packages,
composed of two or more
interventions that have been found
to be effective individually, designed
for persons with complex behavior
disorders
clinical replication

Fill in the blank…
…a second study conducted on a
number of participants with the
same problem, location, or time
simultaneous replication
This module has focused on...
experimental studies
...which test hypotheses to establish
cause-and-effect relationships
The next module will focus on...
descriptive statistics
…the set of mathematical
procedures for describing,
synthesizing, analyzing, and
interpreting quantitative data
Download