NEW YORK UNIVERSITY ROBERT F. WAGNER GRADUATE SCHOOL OF PUBLIC SERVICE

advertisement
NEW YORK UNIVERSITY
ROBERT F. WAGNER GRADUATE SCHOOL OF PUBLIC SERVICE
PADM-GP 2171/GPH-GU 2371 Section 2
Program Analysis and Evaluation
Spring 2014 - Tuesday 4:55pm – 6:35pm
12 Waverly Place, Room L120
Professor Rachel Swaner
Contact Information: Email: RLS200@nyu.edu
Phone: 917-445-7219
Office hours: by appointment
Course Pre-requisites
Students must have completed (or waived) P11.1011 (Statistical Methods) and P11.1022 (Introduction
to Public Policy), as this course builds on these introductory courses.
Course Description and Objectives
Program evaluation is a critical component in designing and operating effective programs. Evaluations
supply information to program managers and policymakers that can assist them in making decisions
about which programs to fund, modify, expand or eliminate. Evaluation can be an accountability tool
for program managers and funders. This course serves as an introduction to evaluation methodology
and evaluation tools commonly used to assess publicly funded programs. Students will become
familiar with the concepts, methods, and applications of evaluation research; learn how to read
evaluation research critically; understand how to use evaluation results to anticipate or improve
program performance; and be able to propose an appropriate evaluation plan to assess the
implementation and effectiveness of a program.
Course Structure
The class includes lecture, readings, and discussion. There is no specific policy or sector focus to this
course, as evaluation tools are used in all policy areas and by public (government) and private
(foundation) funders as well as by public and private sector program managers. Students are
encouraged to relate the general material of the course to their specific policy interests.
Readings
The required textbook for this course is:
Carol H. Weiss (1998) Evaluation: Methods for Studying Programs & Policies 2nd edition. Prentice
Hall
In addition to the required text, you will have additional readings, which are mostly articles (case
studies). Most of the articles are available through Bobst electronic journals, ones that aren’t will be
posted on NYU Classes. There are also additional optional readings, all of which can be downloaded.
There is a sizable body of literature, which deals with program evaluation and policy analysis. The
journal Evaluation Review (previously Evaluation Quarterly) is an especially rich source on the
subject, as is the Evaluation Studies Review Annual (Sage, more or less annually). Evaluation
Practice, Evaluation and Program Planning, New Directions for Program Evaluation, and Journal of
Policy Analysis and Management are also recommended. There are also evaluation journals for
1
specific fields, including Evaluation and the Health Professions, Evaluation in Education, and
Evaluation and Human Services.
Course requirements
Class preparation and participation are important for this course. Students need to read required text
and articles in advance and be prepared to participate in class discussion. In addition to class
participation, students will write two brief memos (with an optional third for extra credit), complete
one take-home midterm exam, and write a final evaluation design paper. Note: The following
descriptions are not enough to complete the assignments adequately. More detailed instructions for
each assignment are posted on NYU Classes.
Midterm Examination
There will be a take-home essay style examination due March 25th.
Program Statement
February 25th
Students will submit a short description of a selected program, indicating the problem to be addressed
by the intervention, the intended beneficiaries or targets of the program, the intended benefits, and the
causal model/program theory underlying the program. This memo is a preliminary step in writing the
final design paper.
Measurement Memo
April 8th
Using the program model developed in the first memo, students will specify the research questions,
operational definitions, and specific measures they would use in an evaluation of the program.
OPTIONAL Evaluation Review (for extra credit)
May 6th
It is important to become a good consumer of evaluations, if not a good evaluator oneself. Review one
of three selected evaluation articles. In 3 pages, students will summarize the type of evaluation
described, its design and methods, and write a critique of the evaluation.
Final Paper: Outcome Evaluation Design
May 13th
The final paper builds on earlier assignments. Students will design a comprehensive evaluation plan
for their chosen programs. The proposal will focus on outcome or impact evaluation but will include a
brief section on process evaluation as well. Students will work in groups.
Relative Weight of Assignments
Midterm Exam
Final Paper
40%
40%
Two memos
Class Participation
Final Grades
94.5+
A
78.5 – 82.4
B-
88.5 – 94.4
A-
76.5 – 78.4
C+
85.5 – 88.4
B+
73.5 – 76.4
C
82.5 – 85.4
B
68.5 – 73.4
C2
10%
10%
Course Schedule
Part I:
Planning and Implementation
Jan 28
Class 1: Introduction to the course and the field of program evaluation; purposes
and stakeholders.
 Weiss Chapters 1 & 2
 Optional: Mercier C. (1997). “Participation in a stakeholder-based evaluation: A
case study.” Evaluation and Program Planning, 20(4): 467-475.
Feb 4
Class 2: Pre-program evaluation activities: needs assessment
 Review Weiss Chapter 2
 Witkin BR. (1994). “Needs Assessment Since 1981: The state of the practice.”
Evaluation Practice, 15(1): 17-27.
 Berberet HM. (2006). “Putting the pieces together for queer youth: a model of
integrated assessment of need and program planning.” Child Welfare, 85(2): 361384.
 Optional: Ma GX, and Thompson B. (1999). “Needs for youth substance abuse and
violence prevention in schools and communities.” Journal of Primary Prevention,
20(2): 93-105.
 Optional: Dietze PM, Rumbold GR, Cvetkovski S, Hanlin KJ, Laslett A, and Jonas
HA. (2000). “Using population-based data on alcohol consumption and related
harms to estimate the relative need for alcohol services in Victoria, Australia.”
Evaluation and Program Planning, 23: 429-436.
Feb 11
Class 3: Explicating and assessing program theory
 Weiss Chapter 3
 Chen, Wang & Lin. (1997). “Evaluating the process and outcome of a garbage
reduction program in Taiwan.” Evaluation Review, 21(1): 27-42.
 Epstein D, and Klerman JA. (2012). “When is a Program Ready for Rigorous
Impact Evaluation? The Role of a Falsifiable Logic Model.” Evaluation Review,
36(5): 375-401.
 Optional: Cooksy, Gill & Kelly. (2001). “The program logic model as an integrative
framework for a multimethod evaluation.” Evaluation and Program Planning, 24:
119-128.
 Optional: Unrau YA. (2001). “Using client interviews to illuminate outcomes in
program logic models: a case example.” Evaluation and Program Planning, 24:
353-361.
Feb 18
Class 4: Process evaluation, formative evaluation and implementation analysis
 Coulon SM, Wilson DK, Griffin S, St. George SM, Alia KA, Trumpeter NN,
Wandersman AK, Forthofer M, Robinson S, and Gadson B. (2012). “Formative
Process Evaluation for Implementing a Social Marketing Intervention to Increase
Walking Among African Americans in the Positive Action for Today’s Health
Trial.” American Journal of Public Health, 102(12): 2315-2321.
 Curran A, Gittelsohn J, Anliker J, Ethelbah B, Blake K, Sharma S, & Cabellero B.
(2005). “Process evaluation of a store-based environmental obesity intervention on
two American Indian reservations.” Health Education Research, 20(6): 719-729.
3



Heinz LC, & Grant PR. (2003). “A process evaluation of a parenting group for
parents with intellectual disabilities.” Evaluation and Program Planning, 26(3):
263-274.
Optional: Dewa, Horgan, Russell & Keates (2001). “What? Another form? The
process of measuring and comparing service utilization in a community mental
health program model.” Evaluation and Program Planning, 24: 239-247.
Optional: Katayoun Rabiei K, Kelishadi R, Sarrafzadegan N, Ali Abedi H, Alavi M,
Heidari K, Bahonar A, Boshtam M, Zare K, & Sadeghi S. (2009). “Process
evaluation of a community-based program for prevention and control of noncommunicable disease in a developing country: The Isfahan Healthy Heart Program,
Iran.” BMC Public Health, 9: 57.
Feb 25
Class 5: Program Memo Presentations
 Program memo due (send electronically)
Part II:
Measuring the Impacts of Programs
Mar 4
Class 6: Outcome/Impact evaluation: design, internal and external validity
 Weiss Chapter 8
 Bornstein D. (2012). “The Dawn of the Evidence-Based Budget.” NY Times, May
30, 2012. Available at: http://opinionator.blogs.nytimes.com/2012/05/30/worthy-ofgovernment-funding-prove-it/
 Program memo returned
Mar 11
Class 7: Outcome/Impact evaluation: randomized experimental design
 Weiss Chapter 9
 Seron C, Ryzin GV, Frankel M, & Kovath J. (2001). “The impact of legal counsel
on outcomes for poor tenants in New York City’s housing court: results of a
randomized experiment.” Law & Society Review, 35(2): 419-434.
 Kim JS, Capotosto L, Hartry A, and Fitzgerald R. (2011). “Can a mixed-method
literacy intervention improve the reading achievement of low-performing
elementary school students in an after-school program?” Educational Evaluation
and Policy Analysis, 33(2): 183-201.
 Bakalar N. (2012). “Omega-3’s Don’t Cut Diabetics’ Heart Risks.” NY Times, June
18, 2012. Available at http://well.blogs.nytimes.com/2012/06/18/omega-3s-dontcut-diabetics-heart-risks/
 Rosenberg T. (2012). “Out of Jail, and Into a Job.” NY Times, March 28, 2012.
Available at http://opinionator.blogs.nytimes.com/2012/03/28/out-of-jail-and-intojobs
 Optional: Lewin A. (2001). “Changing work ethic and welfare dependence through
welfare reform: the 100-hour waiver experiment for AFDC-U.” Evaluation Review,
25(3): 370-388.
 Potential midterm questions handed out
Mar 18
No Class – Spring Break
Mar 25
Class 8: Outcome/Impact evaluation: quasi-experimental designs
 Ballart X, & Riba C. (1995). “Impact of legislation requiring moped and motorbike
riders to wear helmets.” Evaluation and Program Planning, 18: 311-320.
4







Elbel B, Kersh R, Brescoll BL, and Dixon LB. (2009). “Calorie Labeling and Food
Choices: A First Look at the Effects on Low-Income People in New York City.”
Health Affairs, 28(6): w1110-w1121.
Connelly BS, Sackett PR, and Waters SD. (2013). “Balancing Treatment and
Control Groups in Quasi-Experiments: An Introduction to Propensity Scoring.”
Personnel Psychology, 66(2): 407-442.
Optional: Avery-Leaf, Cascardi, O’Leary & Cano. (1997). “Efficacy of a dating
violence prevention program on attitudes justifying aggression.” Journal of
Adolescent Health, 21(1): 11-17.
Optional: Wiener RL, Baron-Donovan C, Gross K, & Block-Lieb S. (2005).
“Debtor education, financial literacy, and pending bankruptcy legislation.”
Behavioral Sciences and the Law, 23: 347-366.
Optional: Cumberland P, Edwards T, Hailu G, Harding-Esch E, Andreasen A,
Mabey D, & Todd J. (2008). “The impact of community level treatment and
preventative interventions on trachoma prevalence in rural Ethiopia.” International
Journal of Epidemiology, 37: 549–558.
Optional: Chemin M. (2008). “The benefits and costs of microfinance: evidence
from Bangladesh.” Journal of Development Studies, 44(4): 463-484.
Midterm answers due (send electronically)
Apr 1
Class 9: Formulating Research Questions and Measurement
 Weiss, Chapter 6
 Beebe TJ, Harrison PA, Sharma A, & Hedger S. (2001). “The Community
Readiness Survey: Development and Validation.” Evaluation Review, 25(1): 55-71.
 Boyd, D. and Marwick A. (2011). “Bullying as True Drama.” NY Times.
http://www.nytimes.com/2011/09/23/opinion/why-cyberbullying-rhetoric-missesthe-mark.html
 Litwin MS. (2003). How to assess and interpret survey psychometrics, 2nd edition,
Chapters 2 and 3: 5-43. (on NYU Classes)
 Optional: Scherer M, Debra Furr-Holden C, and Voas RB. (2013).“Drug Use
Disorder (DUD) Questionnaire: Scale Development and Validation.” Evaluation
Review, 37: 35-58.
 Optional: Dufrene RL. (2000). “An evaluation of a patient satisfaction survey:
validity and reliability.” Evaluation and Program Planning, 23: 293-300.
Apr 8
Class 10: Full coverage and reflexive designs; Sampling
 Weiss, review Chapter 8 pp. 191-199
 Pettifor A, Taylor E, Nku D, Duvall S, Tabala M, Mwandagalirwa K, Meshnick S,
& Behets F. (2009). “Free distribution of insecticide treated bed nets to pregnant
women in Kinshasa: an effective way to achieve 80% use by women and their
newborns.” Tropical Medicine and International Health, 14(1): 20–28.
 Babbie ER. (2013). Chapter 7: The logic of sampling. In The Practice of Social
Research, 13th Edition. (on NYU Classes)
 Optional: Bickman & Hamner. (1998). “An evaluation of the Yad Vashem
Holocaust Museum.” Evaluation Review, 22(4): 435-446.
 Optional: Cook C. (2002). “The effects of skilled health attendants on reducing
maternal deaths in developing countries: testing the medical model.” Evaluation and
Program Planning, 25: 107-116.
5




Optional: Veney JE. (1993). “Evaluation applications of regression analysis with
time-series data.” Evaluation Practice, 14(3): 259-274.
Optional: Gorman DM, Huber Jr JC, & Corozza SE. (2006). “Evaluation of the
Texas 0.08 BAC law.” Alcohol & Alcoholism, 41(2): 193-199.
Measurement memo due (send electronically)
Midterm exams returned
Apr 15
Class 11: Class Cancelled
 Measurement memo returned
Apr 22
Class 12: Participatory Evaluation; Evaluation and Research Ethics
 Weiss, Chapter 14
 The New York Times Editorial Board. (2013). “An Ethical Breakdown.” NY Times.
http://www.nytimes.com/2013/04/16/opinion/an-ethical-breakdown-in-medicalresearch.html
 GreenMills LL, Davison KK, Gordon KE, et al. (2013). “Evaluation of a Childhood
Obesity Awareness Campaign Targeting Head Start Families: Designed by Parents
for Parents.” Journal of Health Care for the Poor and Underserved, 24(2)(S): 2533.
 Richardson L. (2013). “Putting the Research Boot on the Policymakers’ Foot: Can
Participatory Approaches Change the Relationship between Policymakers and
Evaluation?” Social Policy & Administration, 47(4): 483-500.
 Optional: Norris N. (2005). “The politics of evaluation and the methodological
imagination.” American Journal of Evaluation, 26(4): 584-586.
 Optional: Shaw IF. (2003). “Ethics in qualitative research and evaluation.” Journal
of Social Work, 3(1): 9-29.
 Optional: Bluestein J. (2005). “Toward a more public discussion of the ethics of
federal social program evaluation.” Journal of Policy Analysis and Management,
24(4): 824-852.
 Optional: O’Sullivan RG. (2012). “Collaborative Evaluation within a framework of
stakeholder-oriented evaluation approaches.” Evaluation and Program Planning,
35(4): 518-522.
Apr 29
Class 13: Qualitative Methods; Final Presentations
 Weiss, Chapter 11
May 6
Class 14: Final Presentations (continued)
 OPTIONAL Evaluation review memo due (for extra credit)
May 13
Final Paper Due
6
Download