EEP Handouts

advertisement
CORE VIPP EVALUATION:
EEP MEETING FEB 2012, DOCUMENT CONTENTS
This document contains 3 sections: (1) A table that includes the CORE BIC Performance Measures, possible indicators to assess if states are
reaching those measures, and questions from the current State of the States Questionnaire that link to the measures; (2) Section 22 of the States
of the States Questionnaire, and (3) an Evaluation Question Bank of State of the States questions that are not in Section 22 but are in other
sections of the questionnaire.
GOALS:
1) To generate new questions and/or modify existing questions on the State of the State Questionnaire to measure the state
health departments’ injury and violence evaluation capacity. The question items must:
1) Improve data collection;
2) Improve feedback and reporting; and/or
3) Align with the BIC performance measure indicators.
2) Where necessary, develop additional BIC indicators.
CORE Basic Integration Component (BIC) Performance Measure Indicators
Page 2
 This section shows the BIC Performance Measures that are listed in Funding Opportunity Announcement (FOA) and the
Program Consultants brainstormed indicators that elaborate on the performance measures. Safe State/SAVIR Evaluation
Team is using these Performance Measure Indicators as a guide in generating or strengthening questions that show capacity
building overtime. It is important that the 2011 State of the States Questionnaire captures appropriate BIC Performance
Measures Indicators. Please note that not all indicators will be best captured in the survey, but may be capture in other data
collection tools like progress reports, annual reports, etc.
We invite the EEP to brainstorm new indicators and evaluation capacity building questionnaire items that will align
with the listed performance measures and that will useful in the BIC evaluation and reporting. Blank spaces are
provided for new entries for both BIC Performance Measures Indicators and Questionnaire Items.
State of the States Questionnaire – Section 22: Prevention Programs and Evaluation
Page 4
 This section is designed to capture the health departments’ programmatic and evaluation activities for each injury and violence
focus area. The questions presented are updated items from 2009 State of the States. We would like for the EEP to comment
and recommend changes to the question items that will strengthen and improve the data collection and reporting of the
question items and (if appropriate) capture important indicators that are related to the BIC performance measure indicators
(page 6).
Note: Items in blue font were asked in previous State of the States Questionnaires and reported in aggregate to
national state health departments.
Evaluation Question Bank
Page 7
 This section shows two question items that we would like to introduce into Section 22: Prevention Programs and Evaluation.
We would like for the EEP to recommend new question items that will 1) strengthen the data collection and reporting within this
section; and 2) capture important indicators that are related to the BIC grantees performance measure indicators (page 6).
1
BIC PERFORMANCE MEASURES INDICATORS & BRAINSTORMED QUESTIONNAIRE ITEMS
[Evaluation Capacity Area]
Goal /
Performance Measure
BIC Indicator
Time
3. Support and Evaluate Program and Policy Interventions / 5. Program Evaluation
YEAR 1
3.y1.1
Used surveillance findings and literature
Development of recommendations on most
review to inform decision making about the
appropriate best practice or evidence based
use of best practices for injury and violence
strategy
prevention
Evidence of substantial support for evaluation of 4
identified strategies
5.a1
Used surveillance findings to guide injury
prevention and control activities
3.y1.2
Through the ICPG: Selected four strategies
(two program interventions and two policy
strategies) that are best practice or evidencebased and align with priorities identified by the
ICPG
Determination of 4 best practice strategies (at least
two policy strategies)
3.y1.3
Through the ICPG: Developed a thorough and
specific plan for implementation for at least
two topic specific program interventions and
two policy strategies
Creation of a specific implementation plan for
selected strategies, including measureable goals
and objectives with timeframes
Through the ICPG: Developed a thorough and
specific evaluation plan
Evaluation plan
Health Impact measures
Development of objective/quantitative measures of
effectiveness
Specific strategies identified by plan
ICPG partner roles and responsibilities defined
SHD roles and responsibilities defined (contribution)
3.y1.4
Brainstormed SOTS Questionnaire Items Related
to Evaluation Capacity
*22.3 <weak>
2
BIC Performance Measures (continued)…
Goal /
Time
Performance Measure
BIC Indicator
3. Support and Evaluate Program and Policy Interventions / 5. Program Evaluation
YEARS 2- 5
3.[y2-y5].1
Collected and analyzed data to evaluate the
Health Impact measures
effectiveness of intervention programs and
Impact Assessment
policy strategies
Focus of SHD Injury Program evaluations
Data used to support evaluation; Baseline HI data
reported in Year 1; Annual HI data reported in
Years 2-5
3.[y2-y5].2
Used existing CDC evaluation framework
and other tools
Used existing CDC evaluation framework and other
tools
Brainstormed SOTS Questionnaire Items Related
to Evaluation Capacity
*22.4
*22.5
*22.4
Evaluation plan
5.y1.1
Used existing CDC evaluation framework
and other tools in the development of a logic
model. Links:
3.[y2-y5].3
5.a2
Used evaluation findings to make changes in
programs, evaluations, and policies
ANNUALLY
3.a1
Worked with CDC-appointed technical
5.a4
advisor and CDC-named evaluation
contractor
 Was a logic model developed as part of your
program evaluation? [Yes/No/Don’t Know]
 If yes, did you use the CDC Evaluation Framework
to develop the logic model? [Yes/No/Don’t Know]
Use of SHD Injury Program evaluation results
Updated action plans; Submission of Annual Data
report
 Are you using the results of the evaluation findings
to make changes to your program and strategies?
Technical assistance reports
3
2011 STATE OF THE STATES QUESTIONNAIRE
22. PREVENTION PROGRAMS AND EVALUATION
This section primarily focuses on the areas of injury/violence addressed by IVPP programs in 2011.
Questions marked with an asterisk (*) require an answer.
22. PREVENTION PROGRAMS
*22.1. This question attempts to assess the level of effort the state IVPP gives to each injury/violence area, whether or not it is lead by or located
within the state IVPP.
Please select from the drop down menu to indicate the level of program focus during 2011 in each injury/violence area that the state IVPP
addresses through its programs and activities.
Choices include primary focus, secondary focus, minimal focus, or did not focus on in 2011. You can choose multiple areas of primary or
secondary focus.
For ease of data entry, this question continues with additional Prevention Programs in Question #2 below.
Part 1
All Terrain Vehicle (ATV) injury
Program Focus
 Primary Area of Focus
 Secondary Area of Focus
 Minimal Focus on this Topic
 Did not address in 2011
Child abuse/neglect
Child passenger safety
Domestic/intimate partner violence
Elder Abuse
Fall injuries
Fire and burns injury
Firearm injury
Homicide
Mass trauma/disaster-related
Motor vehicle injury
Motorcycle/motorized scooter injury
Occupational injury
4
2011 STATE OF THE STATES QUESTIONNAIRE
*22.2. CONTINED FROM QUESTION #1: This question attempts to assess the level of effort the state IVPP gives to each injury/violence area,
whether or not it is lead by or located within the state IVPP.
Please select from the drop down menu to indicate the level of program focus during 2011 in each injury/violence area that the state IVPP
addresses through its programs and activities.
Choices include primary focus, secondary focus, minimal focus, or did not focus on in 2011.You can choose multiple areas of primary or
secondary focus.
Part 2
Pedestrian injury
Program Focus
 Primary Area of Focus
 Secondary Area of Focus
 Minimal Focus on this Topic
 Did not address in 2011
Poisoning
Rural/agricultural injury
School-based injury
Sexual assault/rape
Spinal cord injury (SCI)
Submersion injuries/drowning
Suicide/self-inflicted
Suicide attempts
Teen Dating Violence
Traumatic brain injury (TBI)
Injuries to children
Injuries to adolescents
Injuries to elderly
Injuries to a racial/ethnic group
Injuries to gender-specific group
Other #1
Other #2
Other #3
Other #4
Other #5
5
2011 STATE OF THE STATES QUESTIONNAIRE
*22.3. For the primary areas of focus you indicated in the previous questions above (#1 and #2), please select how the IVPP determined that those
areas of focus were primary. Choose all that apply.
 Data
 Funding directives
 Needs assessment
 Political influence
 State mandates
 Other (please specify)
 2011 State of the States Questionnaire
*22.4. In questions #1 and #2 above, you indicated the areas of primary focus for your state IVPP in 2011. Please tell us more about how you
evaluated activities in these primary focus areas. For the areas indicated as primary focus in 2011, for how many of them did you complete the
following types of evaluation? The total number entered for each type of evaluation should not exceed the total number of primary focus areas.
For example, if you had four (4) areas of primary focus in 2011, the total number of formative evaluations completed should not be more than four
(4).
o
o
o
o
o
Definitions are provided below for each type of evaluation:
Formative Evaluation – testing program plans, messages, materials, strategies, or modifications for weaknesses and strengths before they
are put into effect
Process Evaluation – testing whether the program’s procedures for reaching the target population are working as planned
Impact Evaluation – assessing the program’s progress toward its goals (i.e., measuring the immediate changes brought about by the
program in the target population such as changes in knowledge, attitudes and beliefs that may lead to changes in health behavior)
Outcome Evaluation – measuring whether your program met its ultimate goal of reducing morbidity and mortality due to injury
Citation: Thompson NJ, McClintock HO. Demonstrating Your Program’s Worth: A Primer on Evaluation for Programs To Prevent Unintentional
Injury. Atlanta: Centers for Disease Control and Prevention, National Center for Injury Prevention and Control, 1998.




Formative
Process
Impact
Outcome
*22.5. Did your state IVPP use the CDC "Framework for Program Evaluation in Public Health" to plan and implement program evaluations in 2011?
 Yes
 No
*22.6. Does your IVPP have access to a program evaluator?
 Yes, within IVPP program
 Yes, within state health department
 Yes, by consultant
 No
 Other (please specify)
6
EVALUATION QUESTION BANK
(These questions appear on the SOTS in other sections, and we can consider adding them to the evaluation section.)
*25.2. How were these programs or activities evaluated in 2011? Check all that apply.
 Number of participants enrolled
 Pre- and post-surveys of participants
 Emergency department or hospital admissions
 Media reports
 Policy changes
 Program costs
 Other (please specify)
*28.2. What outcomes did you measure to evaluate these programs and activities in 2011? Please check all that apply.
 Process indicators (e.g. number of participants, number of materials distributed, etc.)
 Anecdotes/informal measures (e.g., word of mouth; interviews)
 Institutional/policy change
 Media reports/presence
 Likeability/acceptability
 Knowledge gain/retention/refresh
 Behavior change/maintenance
 Long term outcomes (health impact measures like injuries, hospitalizations, fatalities, etc.)
 Policy change
 Environment change (e.g. new crosswalk, new program, home modifications, etc.)
 None
 Other (please specify)
7
8
Download