Lying with Statistics Assignment

advertisement
Lying With Statistics Assignment
RSCH 6110
Guidelines for the Class Presentation and Paper on Critiquing a Use or Misuse of Statistics
Dr. Richard Lambert
UNC Charlotte
3135 Colvard
704-547-3735
email: rglamber@email.uncc.edu
website: http://education.uncc.edu/rglamber
“There are three kinds of lies: lies, damned lies, and statistics.”
Benjamin Disraeli
“You can weigh manure with a jewelers scale,
You can slice bologna with a laser,
But in the end
All you have is manure and bologna.”
Find an example of a use of statistics in the popular or professional press. Ideally find an example that
includes some use of graphical presentation of data. Try to make the example relevant to your field. Feel
free to select an example from a journal or trade publication in your discipline. If you are having trouble
finding something to critique, feel free to use one of the “stat. boxes” from the front page of USA Today
just like the ones we have been critiquing in class. They are archived on the web at
http://www.usatoday.com
Simply click on the “stat. box” at the bottom of the front page and you will be linked to the archive. Please
check with me at least one week in advance of your presentation date to verify that the example you have
selected is appropriate. Please prepare a 10-minute presentation to the class that addresses the seven
questions listed below as well as any other deceptive or misleading aspects of the example. You will need
to project your example on the screen for the whole class to see. If you need help scanning it see me before
class and we can convert it to an electronic format that can be displayed on the projection panel. Your
paper should be 3-5 pages in length and should also include your answers to the questions listed below.
Questions to Ask as You Critique an Example of the Use of Statistics
in the Popular or Professional Press
1.
What statistics were used?
2.
What graphical presentation techniques were used?
3.
Were the statistics and graphical presentation techniques appropriate given the nature of the data
represented?
4.
Toward what conclusions are the presenters trying to point the reader? Are there any causal inferences
that the are being made?
5.
Are these conclusions or inferences really justifiable?
6.
What additional information would you want to have in order to properly interpret the data? Would
you run any additional statistical procedures to enhance interpretation?
7. What qualifications or limitations do you think should be added to the presentation of results?
Use the following pages to take notes as we critique examples from USA Today throughout the
semester.
Some Issues to Consider when Evaluating the Use of Statistics in the Popular or Professional Press
Does the study have defined objectives?
Does the study have well defined outcome measures or variables?
Were the outcomes measured well or without measurement error?
Have appropriate adjustments been made to the statistics presented to enhance interpretation?
For example:
Rates to adjust for unequal time periods
Rates to adjust for unequal population or sample sizes
Rates or Percentages rather than raw counts or totals
Fixed and clearly explained time periods of equal length
Statistical adjustment for other relevant factors
Can you judge the precision of the results presented? Are you given information about sampling error or
confidence intervals?
Was there a defined target population for the purpose of making conclusions or inferences?
Can you judge whether a representative sample was used?
What about the influence of the wording of survey questions?
What about the possible influence of Response Bias?
What about the possible influence of Non-Response Bias?
Are their Order Effects, that is does something about the order in which the questions were asked present a
possible alternative explanation for the results?
Sponsorship / Funding / Motivation Issues
Incomplete Information. What is not provided to the reader that would enhance interpretation?
Can you check the match between the study’s defined goals and what was actually done? Are you really
looking at serendipitous findings?
Consider all the issues related to the graphical presentation of data from the handout that we covered in
class. Have any of these guidelines been violated?
Is premature peeking at the data, or the lack of blinding influencing the results?
Are the writers implying causality from correlational or observational data?
If the study presented itself as an experimental study, did it conform to the characteristics of a Randomized
Clinical Trial according to the guidelines from the handout that we covered in class?
If the study was quasi-experimental in nature, remember to consider Cook & Campbell’s Threats to the
Internal, External, Construct, and Statistical Conclusion Validity of a study from the handout we covered in
class. For example:
Attrition / Differential Attrition
Pre-Existing Differences Between Groups / Non-Equivalence of Groups
History
Maturation
Practice Effect
Hawthorne Effect
Implementation Effects
Experimenter Expectancies, etc.
Have the writers implied that statistical significance directly translates into Practical or Clinical
Significance?
Remember that the best a statistical significance test can do for you is tell you that the observed
results are unlikely to have been the result of sampling error under the condition that the null
hypothesis, as you have specified it, is exactly true. It can never tell you that you have an
important result.
Remember the questions a p value can not answer:
Is this finding important?
Was a statistical significance test necessary at all?
Was the right test used?
Were the assumptions of the test met?
What is the effect size?
What about cost benefit or cost effectiveness issues related to the treatment?
What are the side effects of the treatment?
What confounding variables were present in the study?
Things project officers say about Federal program evaluations that the public never gets to hear.
“The treatment really wasn’t completely implemented as designed.”
“All the subjects didn’t really get the same treatment.”
“They didn’t really do anything meaningful with their money over in that part of the country.”
“The contractor started collecting the data too late in the game.”
“Nobody told the evaluator about that issue.”
“We really don’t know how to measure the effects we’re looking for.”
“No one will ever get a grant to go back and analyze the data the right way.”
“We can always explain away findings that don’t show the program in a positive light.”
Download