Four Primary Areas of Quantitative Research Measurement

advertisement
Four Primary Areas of
Quantitative Research
Measurement – Construct Validity
Sampling – External Validity
Design – Internal Validity
Statistics – Statistical Conclusion Validity
Scales of Measurement
Nominal: categorical measurement. We use this when we group observations
of the same attribute together, for example, putting people into religious
categories (p. 138)
Ordinal: goes one step beyond nominal level in that we not only group like
observations together but also rank them. For example, the question “ Would
you rate your social support as very good, satisfactory, or very poor?” asks for
an ordinal judgment. Ordinal measures may use numbers for convenience in
coding answers, for example, very good (3), satisfactory (2), and very poor (1).
However, such numbers for ordinal categories may serve only as handy,
abbreviations or labels, and you could not conclude that someone answering
with a “3” had three times as much social support as someone answering
with a “3” (p318)
Interval: not only group equivalent observations together in ordered
categories, but also consider the interval between adjacent categories as
equal, but no natural zero. Such as temperature, differences make sense, but
ratios do not. You can think of the meaning of interval level measurement if
you think of it as “equal-internal” measurement (p.329)
Ratio: includes all of the characteristics of the interval level plus the
additional one of having a true zero point that allows dividing one measure by
another (p.329)
Measurement –
Primary Concepts
Reliability: refers to the degree to which observed scores are “free from
errors of measurement” (APA, 1985, p. 19). We can gauge reliability by the
consistency of scores (p. 76) – CONSISTENCY
Construct Validity: refers to the appropriateness, meaningfulness and
usefulness of the specific inferences made from the measures (APA, 1985,
p.9) (p. 76) - we use the score and how will it reflects the construct
Sampling –
Primary Concepts
Population: Collection of all elements (3rd grade teachers) to whom survey
results are to be generalized (US). (p. 348).
Sample: Subset of individuals selected from a larger population (p. 350)
Sampling Methods
Probability sampling: The actual selection of elements from the frame must
give the elements in the frame an equal probability of selection. Random
sampling provides the best way of achieving equal-probability sampling (p.
128). BETTER FOR GENERALIZATIONS; MORE RIGOROUS
Non-probability sampling: includes any method in which the elements have
unequal chances of being selected. One such method, called convenience
sampling, depends on the availability of respondents. In this procedure,
subjects select themselves (p. 129)
External Validity
Generalizability of the study’s findings to other populations, places or times
(p. 345)
Design
True Experimental Designs
(Random Assignment to Control/Comparison Group). More desirable.
Control group is randomly assigned. Random selection from population into
sample; random assignment of control/treatments groups.
Quasi-Experimental Designs
(Control/Comparison Group without Random Assignment)
Pre-Experimental Designs (No Control/Comparison Group)
Internal Validity
Truthfulness of the assertion that the observed effect is due to the
independent variables in the study (p. 346). Judge the relationship between
the independent and dependent variables. RELATED TO DESIGN
Descriptive Statistics
Describes the sample. A set of methods to describe data that we have
collected (mean, standard deviation)
Inferential Statistics
Make inferences about population. Tests probability of sample data being
drawn from the population defined by hypotheses, this is a set of methods
use to make a generalization, estimate, predication or decision.
Statistical Conclusion
Validity
Refers to inferences about whether it is reasonable to presume that a
relationship exists between variables (introduction presentation)
FERPA
Family Educational rights and privacy act (1974) is a federal law that
protects the privacy of student education records (ethics presentation)
National Research Act and
Institutional Review Board
(IRB):
Committees established by us federal regulations at each research
institution to protect human subjects from abuses through prior review of
research proposals
Risk
Exposure to the possibility of physical, psychological or social injury as a
result of the study
Coercion Anonymous
Coercing people to participate require the participants cannot be identified
Confidential
Require the confidentiality of the participants
Reliability
Refers to the degree to which observed scores are “free from errors of
measurement” (APA, 1985, p. 19). We can gauge reliability by the
consistency of scores (p. 76). Reliability is a necessary, but NOT sufficient
condition for validity, reliability can also create a tension with validity.
Reliability Coefficient
Test-retest: short time between tests, stability (error due to time)
Alternate forms: forms match better in content and stat properties,
equivalence (error due to form or internal consistency)
Inter-rater: standardizing procedure including training, rubrics and
monitoring. Scorer or rater consistency.
Split-half: reliability is higher when forms are matched in content and stat
properties, internal consistency (consistency across items – content
sampling error, flawed items)
Internal consistency: longer tests: wide range of individual variability on
construct; freedom from distractions, misunderstandings, use of items of
medium difficulty on cognitive measures (two primary methods for increasing
test reliability), consistency across items – content sampling error, flawed
items.
Standard Error of
Measurement (SEM)
Alternative index for reporting random error based on confidence intervals
(reliability presentation)
Uses and Advantages
Interpretation – 68% of the time the true score is in the interval of x +/- SEM)
95% of the time the true score is in the interval X +/- 1.96*SEM), it is sample
from estimation also important (Reliability, presentation)
Construct Validity
Definition-Construct Validity-“refers to the appropriateness, meaningfulness,
and usefulness of the specific inferences” made from the measures (p.76)
Construct Validity
(Sources of Evidence)
Content-analysis of the relationship between the test’s content and the
construct of interest. It refers to the themes, wording, and format of the
items tasks, or question on a test, as well as the procedural
Substantive /Response Process: Theoretical and empirical analyses of the
response processes of examinees are used to determine the fit between the
construct and detailed nature of the examinees actual performance or
responses.
Internal: Analysis of the internal structure of a test can indicate the degree to
which the relationship among test items and test components conform to the
construct on which the proposed test score interpretations are based
(construct validity – presentation)
Relationships with External Variables: Analysis of relationship of test scores
to variables external to the test (construct validity – presentation)
Consequences: Appraises the value implications of score interpretation as a
basis for action as well as the intended and unintended consequences of test
use, especially in regard to sources of invalidity relation to issues of bias,
fairness and distributive justice (construct validity, presentation)
Construct Irrelevant
Variance: “surplus
construct irrelevancy”
(construct validity,
presentation)
Construct irrelevant difficulty: Aspects of the task are extraneous to the focal
construct – can make the test more difficult for some individuals or groups
Construct
Underrepresentation:
The test is too narrow and fails to include important dimensions of the
construct (construct validity, presentation). For example, one test is one test
is trying to measure the mathematics curriculum including probability,
algebra, measurement, data analysis, etc. however, if the test only includes
the terms relating algebra.
Internal Validity
Definition – Isolation; Refers to whether or not the relationship between two
variables; refers to the truthfulness of the claim that one variable causes
another
Construct irrelevant easiness – when extraneous clues in the time or format
results in correct responses
Threats to Validity
History: Refers to the threat that some coincidental event outside the study
caused the observed change; can control by reducing time between measures
The situation in which some specific event occurs during the study, which in
turn affects the results. The specific event is an effect other than the
experimental treatments. To prevent this threat use a control group, shorten
time between testing, select a dependent variable less prone to history
effects, or insulate participants.
Maturation: refers to any naturally occurring process within individuals as a
function of time per se that may cause a change in their performance.
Processes include fatigue, boredom, growth, or intellectual maturation. Time
threat to internal validity in which internal or normal developmental
processes cause the observed change. To prevent this threat, conduct the
study over shorter period of time or use a control group with a comparable
maturation rate.
Testing: refers to a response to the pretest that causes the observed change
in the outcome variable; can reduce by disguising the pretest or dropping it
completely
Instrumentation: when observed changes results from shifts in the way
measures are collected; can be controlled by carefully standardizing and
monitoring the measurement procedures
Regression: Comes from the tendency of scores from unreliable measures to
move toward the mean on retest; affects one group studies that select
subjects for their extreme scores; To prevent avoid the selection on the basis
of extreme scores or create a control group of extreme scores. Can also use
highly reliable measures.
Mortality: subject attrition from pretest to posttest, which casts doubt on
validity of the study; protection of this threat cannot be provided by a control
or random assignment
Selection
Selection Interactions
Design
True Experimental Designs
Role of control/comparison groups
Detect and explain possible confounding variables
Role of random assignment
Increases isolation by removing the relationship between the
treatment and other confounding variables. Results in
equivalent groups
Experimental Designs
Designs that involve some form of control/comparison group but do
NOT use random assignments. Generally, selection and selectioninteraction are the greatest type of threats.
Pre-Experimental Designs
Experimental design without control/comparison group. Weak design,
but useful when pretests are not possible (i.e., pilot testing)
Random Assignment (Internal Validity): method of placing subjects in
different conditions so that each subject has an equal chance of being in any
group to avoid systematic subject differences between the groups (p. 349)
Random Selection (External Validity): drawing a representative group from a
population by a method that gives every member of the population an equal
chance of being drawn (p. 349)
Matching: Exact equivalence on select variables – uncertainty on all others
Statistical Adjustment: Argue theoretical model – each model leads to a
different solution (Experimental design, presentation)
Download