about the template

advertisement
EXCEL Template for Obtaining Internally Consistent Subsets of Items
This EXCEL template is intended to help delete items from a measure to produce two or
more sets of items that "fit the data" (are internally consistent).
The citation for these instructions is Ping, R.A. (2006). "More about the template for obtaining
an internally consistent set of items." http://home.att.net/~rpingjr/weeding1.doc.
New measures will almost never "fit the data" using a single construct measurement
model without dropping items to attain model-to-data fit. In addition, most well established
measures developed before covariant structure analysis (LISREL, AMOS, etc.) became popular
also will not fit the data without item weeding.
It turns out that measures used with covariant structure analysis are limited to about six
items (see discussions in Anderson and Gerbing 1984, Gerbing and Anderson 1993, Bagozzi and
Heatherton 1994, and Ping 2008). One explanation is that correlated measurement errors,
ubiquitous in survey data but customarily not specified in covariant structure analysis, eventually
overwhelm model-to-data fit in single-construct and full measurement models as indicators are
added to the specification of a construct. And, that usually happens with about 6 items per
construct.
There are ways around item weeding, such as various item aggregation techniques (see
Bagozzi and Heatherton 1994), but many reviewers in the Social Sciences do not like these
approaches. Unfortunately, reviewers also may not like dropping items from measures because
of concerns over face- or content validity (how well the items "tap" the conceptual and
operational definitions of their target construct). One "compromise" is to show the full measure's
items in the paper, and assuming the full measure does not fit a single construct measurement
model, show one submeasure that does fit the data and is maximally "equivalent" to the full
measure in face or content validity. However, to do that, several submeasures are usually
required, and finding even one is frequently a tedious task.
This template will assist in finding at least two subsets of items from the target measure
that fit the data in a single construct measurement model of the items. The process is as follows.
First, exploratory (common) factor analyze the target measure with its items using Maximum
Likelihood estimation and varimax rotation. If the measure is multidimensional, start with the
Factor 1 items. The other factors and the full measure can be used later.
Next, estimate a single construct (confirmatory) measurement model using the Factor 1
items (if the measure is unidimensional Factor 1 is the full measure). If the first measurement
model fits the data item omission is not required. If this measurement model does not fit the data,
find the "First Order Derivatives" in the output. (I will assume LISREL 8, which requires "all"
on the OU line to produce First Order Derivatives. As far as I know, most other estimation
packages produce statistics equivalent to First Order Derivatives). Paste the lower triangle of
First Order Derivatives for "THETA-EPS" into the template making sure you retain the item
names so you can figure out which item to drop (see the example on the template). Then find the
largest value in the "Overall Sum" column--it will be the same as the "Max =" value in the lower
right corner of the matrix.
Now, reestimate the measurement model with the item having the largest "Overall Sum"
omitted (call this Reestimation 1). Record the Chi Square and RMSEA values on the spreadsheet
for reference. If they are acceptable, use the items in this measurement model as submeasure 1.
There is no agreement on acceptable single construct measurement model fit. I use either
a Chi Square that is slightly nonzero for single construct measurement models (e.g., 1E-07, not
0), or an RMSEA that is .08 or slightly below, but many authors would suggest much stronger fit
criteria for single construct measurement models.1
If the unomitted items do not fit the data, find the "First Order Derivatives" for "ThetaEps" in the Reestimation 1 output. Paste these into the second matrix in the template, record the
Chi-Square and RMSEA values, and reestimate the single construct measurement model
(Reestimation 2).
Repeating this process, eventually Chi Square will become nonzero, and after that
RMSEA will decline to 0.08 or less (the recommended minimum for fit in full measurement and
structural models--see Brown and Cudeck 1993, Jöreskog 1993). This should happen with about
7 or 8, down to about 5, remaining items. If acceptable fit does not happen by about 4 items, an
error has probably been made, usually by omitting the wrong item.
Each subset after Chi Square becomes non zero is a candidate subset for "best," but
because items are disappearing with each step, these smaller subsets are usually less face valid,
and thus the first acceptable subset is usually the preferred one.
To find another subset of items, repeat the above process using "Modification Indices"
for "Theta Epsilon." The theory behind Modification Indices is different from First Derivatives,
and a different subset usually results.
Another subset of items usually can be found using reliability. The reliability of all the
Factor 1 items is computed using SAS, SPSS, etc., the item that contributes least to reliability is
deleted, and the reliability of the remaining items is computed. This process is continued until
deleting any item reduces reliability. The remaining items usually will fit the data in a single
construct measurement model.
If the full measure was multidimensional, there may be several more subsets found by
repeating the above procedures using the full measure's items instead of the Factor 1 items, then
using the reliability procedure just mentioned. Experience suggests these subsets are smaller, but
they frequently include items from Factor 2, etc. and thus they may be more face valid. This
process can also be used on any Factor 2 items, Factor 3, etc.
There are many more subsets that can be found by omitting the next largest "Overall
Sum" item instead of the "Max =" item. Specifically, the second largest item in Reestimation 1
could be omitted in place of the largest. Then, continuing as before omitting the largest "Overall
Sum" items, The result is frequently a different subset of items that fits the data. Another subset
can usually be found using this "Second Largest" approach using modification indices instead of
first derivatives. Others can be found omitting the second largest overall sum item in
Reestimation 2, instead of Reestimation 1, etc., with or without deleting the second largest in
Reestimation 1. This "Second Largest" strategy can also be used on the full set of items.
Experience suggests that there are about N-things-taken-6-at-a-time combinations of
items with real world data that will fit the data, where N is the number of items in the full
measure (more, if 5, 4 and 3 item subsets are counted). For example, if the original measure has
8 items, with real world data there are about 8!(8-6)!/6! = 112 6-tem subsets of items that might
fit the data. While the above strategies will not find all of them, experience suggests they should
identify several two subsets that are usually attractive because they are comparatively large
(again however, usually with about 6 items) and they should appear to tap the target construct
comparatively well.
The above spreadsheet approaches may not always identify the highest reliability subsets
of items, but experience suggests the resulting subsets are usually larger and as, or more, face
valid than those produced by other approaches. However, with low reliability measures, even
though the "First Derivative" or "Modification Indices" subsets should be only a few points
lower in reliability than a subset found by, for example, dropping items that contribute lest to
reliability, the higher reliability subset may be preferred to a higher face validity subset.
It may be instructive to (re)submit all the subsets found to an item-judging panel for their
selection of the "best" subset for each construct.
Other comments: There are exceptions to several of the assertions made above, but this is
probably not the place for an exhaustive exposition on item deletion strategies. For emphasis, the
template assumes lower triangular matrices. There is an additional example in Appendix E of the
monograph, Testing Latent Variable Models..., on the web site.
REFERENCES
Anderson, James C. and David W. Gerbing (1984), "The Effect of Sampling Error on
Convergence, Improper Solutions, and Goodness of Fit Indices for Maximum Likelihood
Confirmatory Factor Analysis," Psychometrika, 49, 155-73.
Bagozzi, Richard P. and Todd F. Heatherton (1994), "A General Approach to Representing
Multifaceted Personality Constructs: Application to Self Esteem," Structural Equation
Modeling, 1 (1), 35-67.
Browne, Michael W. and Robert Cudeck (1993), "Alternative Ways of Assessing Model Fit," in
Testing Structural Equation Models, K. A. Bollen et al. eds, Newbury Park CA: SAGE
Publications.
Gerbing, David W. and James C. Anderson (1993), "Monte Carlo Evaluations of Goodness-ofFit Indices for Structural Equation Models," in Testing Structural Equation Models, K. A.
Bollen and J. S. Long, eds., Newbury Park, CA: SAGE Publications.
Jöreskog, Karl G. (1993), "Testing Structural Equation Models," in Testing Structural Equation
Models, Kenneth A. Bollen and J. Scott Long eds., Newbury Park, CA: SAGE.
_______ (2004), "On Assuring Valid Measures for Theoretical Models Using Survey Data,"
Journal of Business Research, 57 (2), 125-41.
ENDNOTES
1
In my opinion, some authors go too far in real world data with single construct measurement model fit, resulting in
unnecessarily small submeasures. There are several issues here, including model fit versus face or content validity,
and experience suggests that with real-world data, "barely fits" in single construct measurement models is almost
always sufficient to attain full measurement model fit. Thus, in real world data, subsets of items that each produce a
comparatively small but nonzero Chi Square or an RMSEA that is just below .08 are usually "consistent enough" to
later produce a full measurement model that fits the data. I prefer the RMSEA criterion because it seems to produce
fewer problems later. Again, however, many authors would not agree with this strategy. Later, if it turns out that the
full measurement model does not adequately fit the data, simply estimate the next item weeding single construct
measurement model and drop the next largest "Overall Sum" items to improve full measurement model fit.
Download