Assessment and Progress Monitoring in School

advertisement
Assessment and Progress
Monitoring in School-Based
Mental Health
A collaboration between Public Health of Seattle-King Country, the Seattle
Public Schools, and the University of Washington School of Medicine/Seattle
Children’s Hospital
Aaron Lyon, PhD (UW/SCH)
Jessica Knaster Wasse, MPH (PHSKC)
Kristy Ludwig, PhD (UW/SCH)
October 16, 2012
Acknowledgments
Collaborators:
•
•
•
•
•
•
•
•
•
Ann Vander Stoep
Elizabeth McCauley
Eric Bruns
Evalynn Romano
Jane Koltracht
Kelly Thompson
Seattle Public Schools
TJ Cosgrove
School-based practitioners!
Funding:
• National Institute of Mental
Health (K08MH095939)
• Institute of Education Sciences
(R305A120128)
• American Psychological
Foundation
• Public Health of Seattle & King
County
• Bill and Melinda Gates
Foundation
Dr. Lyon is an investigator with the Implementation Research Institute (IRI), at the George Warren
Brown School of Social Work, Washington University in St. Louis; through an award from the National
Institute of Mental Health (R25MH080916-01A2) and the Department of Veterans Affairs, Health
Services Research & Development Service, Quality Enhancement Research Initiative (QUERI).
Overview
1. Background & goals of standardized
assessment / progress monitoring initiative
in Seattle’s school-based health centers
(SBHCs)
2. Public Health of Seattle & King County’s
(PHSKC) interest in / support for progress
monitoring in SBHCs
3. Current training approach / activities
4. Future directions
School-Based Health Centers (SBHCs)
• Integrated care clinics situated in schools
• Operate in nearly 2,000 schools in the US (NASBH,
2008)
• Typically provide primary care and mental health
services (Brown & Bolen, 2003)
• Well-substantiated as a mechanism to increase
service accessibility to underserved and
under/uninsured (Gance-Cleveland & Yousey, 2005; Kaplan
et al., 1999; Wade et al., 2008)
Evidence-Based Practice in SBMH
• EBP utilized inconsistently in SBMH (Evans & Weist, 2004)
• Implementation of EBP in SBMH should focus on lowcost, high-yield practice improvement targets
• Disseminate / Implement key competencies instead of full Tx
packages (Beidas et al., 2011; Rotheram-borus et al., 2012)
• Evidence-based assessment may be the “low-hanging
fruit” of EBP implementation
• School-based providers unlikely to use standardized assessment
and progress monitoring routinely
• Previous work has demonstrated good uptake of SA following
training (Lyon et al., 2011)
Evidence-Based Assessment
Evidence-Based Assessment (EBA): “Assessment
methods and processes that are based on empirical
evidence in terms of both their reliability and validity
as well as their clinical usefulness for prescribed
populations and purposes” (Mash & Hunsley, 2005, p.364)
• Evidence-based interventions commonly rely on EBA
Standardized assessment (SA), or the use of
measurement tools with empirical support for their
reliability, validity, etc, is at the core of EBA (JensenDoss & Hawley, 2005)
Value of Standardized Assessment
Initial Assessment
• Rating scales can increase the ease and accuracy of
clinical diagnosis (e.g., Jenkins et al., 2011; Youngstrom et al.,
2005)
• Psychological assessment carries positive, clinically
meaningful effects (Posten & Hanson, 2010)
Progress Monitoring
• Clinicians are often not able to detect client
deterioration (Hannan et al., 2005)
• Providing assessment results to clinicians can result
in improved outcomes (Bickman et al., 2011; Lambert et al.,
2003)
Steps of Standardized Assessment Use
1.
2.
3.
4.
5.
6.
7.
8.
Selection
Administration
Scoring
Interpretation
Feedback to clients
Treatment planning
Monitoring (over time w/ feedback)
Treatment plan adjustment
SA / Progress Monitoring: SBMH v. CMH
SBHC
(n = 16)
CMH
(n = 58)
Percent of Caseload:
0%
1-39%
0%
1-39%
Gave a SA measure in initial 1-2 meetings
47%
27%
16%
32%
Gave a SA measure at termination
63%
36%
52%
23%
Gave a SA measure
38%
44%
42%
42%
Gave feedback about a SA measure
50%
31%
46%
30%
Changed Tx plan based on SA data
69%
31%
64%
32%
Changed indiv. session plan based on SA
56%
38%
64%
27%
Progress Monitoring & Feedback
• Progress monitoring may use SA tools or ideographic
monitoring targets
• Value of ideographic targets (Weisz et al., 2011)…
• Add specificity to ID’d problems
• Focus on client concerns / priorities
• Give clients a voice
• Enhance rapport / alliance
• Provide foci for ongoing assessment
• Provide info about treatment termination
• Easy to fit into everyday practice
• Combination of SA and ideographic is likely optimal
Progress Monitoring & Feedback
• Feedback Intervention Theory (Kluget & DeNisi, 1996): Bx is
regulated by comparisons of feedback to goals
• Two levels (at least!): (1) feedback to the clinician & (2)
feedback to the client
• Most research has focused on feedback to clinicians
• Feedback is associated with higher rapport with provider,
client self-understanding (Allen et al., 2003), and increased
engagement in therapy (Ackerman et al., 2000)
Monitoring School Outcomes in Tx
• Academic outcomes are an essential domain for
youth success, but rarely included in MH
research (Atkins et al., 2010; Franklin et al., 2009; Hoagwood et
al., 2007)
• Youth w/ combined MH and academic risk have
a greater likelihood of long-term negative
outcomes (Roeser et al., 1999)
• School-based clinicians rate school attendance
as the most clinically-useful academic indicator
(Lyon, Ludwig, & McCauly, in prep)
Potential Assessment / Monitoring Targets
Mental Health
• Depression symptoms
• Disruptive behavior
• Self-injurious behavior
School Engagement
• Attendance
• Homework completion
• Class participation
• School connectedness
Academic Outcomes
• Grades
• Credits earned
• Standardized test scores
Social Functioning
• Interpersonal conflicts
• Positive social experiences
• Disciplinary events
Physical Health
• Sleep
• Diet & Exercise
Services
• Satisfaction with treatment
• Engagement in intervention
Current Goals & Activities
1. Engage stakeholders in planning /
adapting a computerized measurement
feedback system (MFS) to support SBMH
providers
2. Provide training to SBMH providers in the
use of standardized and ideographic
assessment / progress monitoring
3. Implement, evaluate, and iterate the MFS,
incorporating the ability to monitor
academic outcomes
Engaging Stakeholders in the
Adaptation of a Measurement
Feedback System for SBHCs
SBHCs in Seattle / King County
Policy/funding context
• Seattle Families & Education Levy
• Property tax levy, renewed for 4th 7 year term starting 2012
• SBHC funding managed by PHSKC
• Levy performance goals are academic in nature
• Successfully lobbied for Levy $ for MH enhancement
•
Goal: Enhance the academic impact of Levy health investments
by improving the quality of school mental health services
• Increase use of evidence-based practice, with focus on
standardized assessment and outcome monitoring
SBHCs in Seattle / King County
Vision: Collaborative stepped care model
• SA data used to support clinical decision making
• Track clinical improvement and identify patients who are not
improving as expected
Aligns with Seattle Public Schools’
Response to Intervention (RtI) strategy
Graphic from the
National Center on
Response to
Intervention
Stakeholder Committee
• “When developing an evaluation system, two of the most
challenging components are getting buy-in from all
partners involved and then selecting the measurement
tools” (Sander et al., 2011, p.15).
• Gates Foundation funding for one-year planning process
• Stakeholder committee comprised of: school-based MH
providers, supervisors, evaluators, program managers,
computer programmers, school district
Stakeholder Committee
Monthly meetings designed to:
• Build initial consensus about data fields and other
•
•
•
•
user-interface issues
Identify initial battery of standardized assessment
measures
Identify SA toolkit for various MH domains
Develop and refine implementation / evaluation plan
Enhance provider buy-in
Stakeholder Committee
Challenges/hurdles
• Differing documentation requirements
• Duplicate data entry EHR and MFS
• Varying degrees of technical knowledge
• Evaluation design: impact of randomization
Stakeholder Committee
Results/Successes
•
•
•
•
Data sharing agreement with Seattle Public Schools
Specifications completed on schedule
Negotiated acceptable evaluation plan
Providers engaged and excited
Training in Standardized
Assessment and Progress
Monitoring
Training in SA & Progress Monitoring
In promoting data-based decision making,
SBMH professionals often struggle to (Kelly,
2011)…
1. Clearly define the reason for collecting the data
2. Ensure that data-collection procedures are
manageable / user-friendly
3. Map out a clear system to use the data for continuous
improvement
Training in SA & Progress Monitoring
• Over the past 2 years, our group has provided
a training and consultation series to local
SBHC MH providers covering…
• Standardized assessment principles, psychometrics,
and normative data
• Initial assessment using standardized tools
• Feedback to students on the basis of assessment
• Identifying and tracking progress monitoring targets
Training in SA & Progress Monitoring
Principles of progress monitoring:
• Monitoring of some type is a good idea for ALL youth
receiving services
• Only track those targets that are meaningful to the
specific case/match treatment goals
• Graphical feedback increases
understanding and makes the
feedback more memorable (Kluger &
DiNisi, 1996; Miller & Watkins, 2010)
• Can be used as a tool to validate changes in a student’s
experience
Training in SA & Progress Monitoring
Moving from the WHAT to WHY…
• Progress monitoring is the first step and gives you the
WHAT
• Next steps include identifying the WHY (i.e., why is a
target changing or not changing?)
• Moving beyond just “success” or “failure”
• The WHY is important to the kid feeling successful
• Kid’s feelings of agency and control (therapy isn’t being done to
them or for them but something they are willfully and actively
participating in)
Assessment Flow Chart
SA
Student
Presentation
Assess
1) Decrease anxiety
symptoms
2) Improved school
engagement
3) Improved sleep
hygiene
Treatment
Goal ID
Ideographic
SA: MASC-10
IG: Anxiety experienced when
arriving at school?
SA: School Connectedness Scale
IG: # on-time arrivals for 1st period
SA: None?
IG: Nights/wk w/o FB or texting
after 8pm
Current Findings and Future
Evaluation Directions
Focus Groups
• Focus groups conducted to evaluate the context in which
the MFS will be implemented.
• Transcribed and evaluated using a conventional content
analysis approach.
• EMERGING THEME: Factors which facilitate or inhibit the
utility of a technology in SBMH (Lyon et al., in prep)
• Accessibility
• Impact on Youth Engagement
• Confidentiality
• Knowledge or Attitudes (Provider)
• Integration of Aspects of Care
• Relevance to MH Care
• Speed of the Technology
Pilot of the Un-adapted MFS
• Anonymous surveys of youth receiving SBMH
services supported by the un-adapted MFS (n = 34)
• 71% of youth respondents stated that answering questionnaires
about emotions made it easier to tell their providers how they
were feeling
• 55% said answering the questionnaires (and receiving feedback)
made the time spent with their SBMH providers either a little or
much easier/better.
• 45% said it did not affect their interactions (no students
reported that it made it worse)
Many Questions Remain…
Selected findings from Key Informant individual
interviews from a related project…
• Who should provide feedback about progress?
• Teachers, students, school teams
• Who should receive / review feedback?
• In addition to clinician, principal, teachers, school staff, the
academic counselor, school teams
• Allow student to determine who will review
• Other challenges
• Unique data systems and data sharing rules
• May require individualized approach district-by-district
• Providers not used to collecting/using data systematically
• Transitioning from health/MH to academic outcomes
Next Steps
1. Complete initial adaption of the MFS
2. Implement the MFS with randomly-selected providers
3. Compare provider behavior (e.g., use of standardized or
ideographic assessment/monitoring) between MFS and
non-MFS groups
4. Conduct independent evaluations of youth outcomes by
recruiting from the caseloads of MFS and non-MFS
providers
Questions and Discussion
Download