EVIDENCE BASED PRACTICE Evaluating the Quality of Evidence

advertisement
EVIDENCE BASED PRACTICE
Evaluating the Quality of Evidence
Client Assessment and Risk Evaluation (CARE)
The CARE:
o Was developed to assist practitioners in weighing the quality of assessment procedures.
o Is designed to compute a single index of the quality of a client assessment procedure that reflects the assessment’s
ease of use, reliability (particularly inter-rater), and predictive validity.
o Higher scores indicate a stronger analysis.
o Provides and index of what to look for in a practical, useful, assessment procedure.
o Instructions:
o Please read the Explanation for each criteria on the CARE form with an eye to applying the criterion to any
assessment procedure. The form was intended to rate the quality of any assessment procedure, regardless of
whether the procedure follows a published measure, an interview guide, agency procedure, or common
practice. Give one point for each check mark. The form is based on the assumption that any assessment
procedure ought to be simple to apply, reliably scored, and of predictive value regarding what clients will do
in the future or against a more valid measure. Scores can range from 0 to 100. This is only an ordinal scale,
meaning that a score of 20 is higher than a score of 10 but not necessarily twice as high. No norms exist for
the CARE form.
o The CARE assumes background that does not appear in most practice and research texts. If items on the
CARE form appear unfamiliar to you, please read the Detailed Explanation for CARE Criteria that appears
following this form. Criteria can be rated on the CARE form from documentation that accompanies
assessment procedures without understanding specifics in the Detailed Explanation. Standards for risk
assessment and for judging the validity of an assessment against a more valid assessment procedure follow
the same pattern here, but discussion concerns risk for consistency.
Source in APA format:__________________________________________________________________
_____________________________________________________________________________________
Criterion
Points
(1 Point
for Each
Criterion
Checked)
Utility of Assessment Procedure for Practice.
1. Assessment procedure easy to
learn.
2. Assessment in less than 10
minutes.
3. Assessment’s scoring less
than five minutes.
Explanation
Physically examine the assessment’s procedure and its scoring
procedures to rate this. Estimate whether you and your coworkers
could do an assessment and score it without confusion simply by
following the procedure’s instructions.
The assessment’s administration would take less than 10 minutes of
additional time above what the client contact would generally take. To
estimate time, do a trial with a few actual cases or with a role-played
interview to actually time how long the assessment requires, or rely on
published reports.
Try scoring a few assessments to see how long the scoring takes.
Allow for experience as a way to shorten scoring time. Scoring should
take less than 5 minutes. You may rely on published reports of time
required to score.
1
With permission from Gibbs, L. (2003). Evidence-based practice for the helping professions: A practical guide with integrated
multimedia. Pacific Grove, CA: Brooks/Cole--Thomson Learning.
EVIDENCE BASED PRACTICE
Evaluating the Quality of Evidence
Client Assessment and Risk Evaluation (CARE)
Reliability (i.e., Consistent with Each Administration Over Time, Across Raters, or Internally Across the
Instrument’s Items)
4. Assessment procedure was
This means that two or more raters arrived at their assessment without
checked for inter-rater
conferring at all with the other raters. Give no points unless the authors
reliability.
state explicitly that assessments were done independently. Interobserver, cross-observer, across-raters mean inter-rater.
5. Some (any) inter-rater
Any inter-rater coefficient will do here so long as the assessments
reliability coefficient
were made independently and any coefficient of agreement was
computed.
computed.
6. Kappa coefficient of interThe authors must both compute a Kappa to rate the agreement of
rater reliability for assessment
assessments by independent workers, and the Kappa must exceed 0.70.
exceeds 0.70.
Since decisions in practice are binary (act/do not act), inter-rater
reliability and its most appropriate statistic (Kappa) are criteria here.
7. Assessment procedure
This criterion is met if the authors check for reliability using any
checked for form of reliability
procedure other than inter-rater reliability.
other than inter-rater
reliability (e.g., test-retest,
split half, internal
consistency).
8. Reliability coefficient
Give the points here if the authors compute a coefficient of reliability
computed other than Kappa
other than Kappa (e.g., Pearson r, Cronbach’s alpha, Kuderabove 0.70 or 70%.
Richardson formula 20) and value above 0.70.
Predictive Validity (The client’s assessment demonstrates that it can actually predict how the client will perform in
the future. The following discussion refers to risk, which is probability of an undesirable behavior, but the same
principles for this discussion can apply to other standards for judging accuracy against a more valid criterion).
9. Those who developed the
Look here for a tabular literature review tat lists studies and which
assessment procedure did a
indicators were of predictive value in each study. Give no points if you
systematic review of studies
cannot find such a table in the report.
to isolate indicators that
might have predictive value
to estimate risk.
10. The authors clearly describe
Merely stating the client type (e.g., suicidal or depressed persons) is
criteria for including clients
not enough. Authors must state the specific criterion or measure (e.g.,
of a particular type in their
specifically defined prior suicidal behaviors, Zung Self-Rating
risk-assessment study.
Depression Scale) for including subjects in the study. Knowing
inclusion criteria allows practitioners to judge whether study findings
apply to their clients.
11. The risk assessment study’s
This means that indicators of risk were collected; then clients were
results were collected
followed to see what they would do, and then the indicators were
prospectively.
evaluated for predictive efficiency against what the clients actually did
or against another gold standard.
2
With permission from Gibbs, L. (2003). Evidence-based practice for the helping professions: A practical guide with integrated
multimedia. Pacific Grove, CA: Brooks/Cole--Thomson Learning.
EVIDENCE BASED PRACTICE
Evaluating the Quality of Evidence
Client Assessment and Risk Evaluation (CARE)
12. The risk assessment study
was done prospectively and
the study resulted in greater
than 80% being contacted at
follow-up.
13. During the data analysis,
those who recorded the
subject’s actual behavior
were blind to what each
subject’s risk-assessment
score had been.
14. The risk-assessment
measure’s predictive accuracy
was checked in at least one
validation study.
15. The risk-assessment scale’s
positive predictive value
(PPV) was higher than the
prevalence rate (base rate,
prior probability) by at least
10%.
16. PPV is greater than .80.
17. NPV is greater than .80.
18. Using the same subjects, the
authors compared positive
predictive value (PPV) for
practitioners’ prediction
against PPV for the riskassessment scale’s
predictions, and the latter is
higher.*
19. The authors state specifically
that they have used a receiver
operating curve (ROC)
analysis to establish the risk
assessment’s cutoff or
division criteria (e.g.,
dividing point between
high/low risk categories).
Divide the number who were contacted at the end of the study
regarding their actual behavior by the number who took the riskassessment measure at the beginning of t study, and multiply by 100.
This analysis will compare the risk assessment’s earlier results against
what actually happened later to judge whether the assessment was
accurate. Give a point only if the authors state that those who
recorded the predicted behavior were blind to what the prediction had
been.
Risk scales may predict well where they were developed but
sometimes do not elsewhere. To meet this criterion, the measure’s
accuracy needs to be tested on a sample other than where it was
developed.
Applying a risk-assessment procedure that will not predict better than
chance (prevalence rate) makes no sense.
The study computed positive predictive value, or gives sufficient data
to do so, and PPV is greater than .80. If more than one computation of
PPV, then the average PPV is greater than .80.
The study computed negative predictive value, or gives sufficient data
to do so, and NPV is greater than .80. If more than one computation of
PPV, then the average PPV is greater than .80.
This kind of study pits the predictive accuracy of practitioners’
assessments against a risk-assessment scale. This kind of evaluation
assumes that the practitioners do not know the risk-assessment scale’s
score when they make their judgment.
Any risk-assessment scale involves a trade-off. If you want to
maximize your instrument’s sensitivity to detect the positives, you will
also increase our number of false positives. ROC analysis allows
practitioners to make an informed judgment about where best to set the
scale’s division point(s). For a detailed description, consult MedCalc
at www.medcald.be/roccman.html.
3
With permission from Gibbs, L. (2003). Evidence-based practice for the helping professions: A practical guide with integrated
multimedia. Pacific Grove, CA: Brooks/Cole--Thomson Learning.
EVIDENCE BASED PRACTICE
Evaluating the Quality of Evidence
Client Assessment and Risk Evaluation (CARE)
Total
number
checked
(19
possible)
Score =
(number
checked/19
x 100)
Summary Statistics for Assessment Procedure
Inter-rater reliability Kappa for assessment procedure
Positive predictive value for assessment procedure
Negative predictive value
* Some studies report only sensitivity, specificity, and prevalence. You can still compute PPV with Bayes’s Theorem as
follows:
(prevalence)(sensitivity)
[(prevalence)(sensitivity)]/[(1 - prevalence)(1 – specificity)]
4
With permission from Gibbs, L. (2003). Evidence-based practice for the helping professions: A practical guide with integrated
multimedia. Pacific Grove, CA: Brooks/Cole--Thomson Learning.
Download