Evaluation Research Kodi D. Havins AED 615 Fall 2006 Dr. Franklin Evaluation Research • Definitions: is the systematic assessment of the worth or merit of some object” – “Evaluation is the systematic acquisition and assessment of information to provide useful feedback about some object” – “Evaluation (Trochim, 2006) Evaluation Research • Basically is used to provide feedback on an event, organization, program, policy, technology, person, activity, etc. – All of these are called “object(s)” • It assesses information about the “object” • Used heavily in social sciences and government • agencies Usually outcomes influence decision-making process Research Designs • 4 Major Types: 1. 2. 3. 4. Survey Research Case Study Field Experiment Secondary Data Analysis (Garson, n.d.) Procedures • 4 Strategies: 1. Scientific-Experimental Models 2. Management-Oriented Systems Models 3. Qualitative/Anthropological Models 4. Participant-Oriented Models (Trochim, 2006) 1. Scientific-Experimental Models • Take values and methods from the sciences (especially social sciences) • Prioritize on the desirability of impartiality, accuracy, objectivity, and validity of the information generated (Trochim, 2006) 1. Scientific-Experimental Models • Examples: Tradition of experimental and quasiexperimental designs Objectives-based research that comes from education Econometrically-oriented perspectives including cost-effectiveness and costbenefit analysis Recent articulation of theory-driven evaluation (Trochim, 2006) 2. Management-Oriented Systems • 2 of the most common: – – Program Evaluation and Review Technique (PERT) Critical Path Method (CPM) (Trochim, 2006) 2. Management-Oriented Systems Examples: • Framework (“Logframe”) Model – Used widely by U.S. Agency for International Development (Trochim, 2006) 3. Qualitative/Anthropological Models • Emphasis on: – – – The importance of observation The need to retain the phenomenological quality of the evaluation context The value of subjective human interpretation in the evaluation process (Trochim, 2006) 3. Qualitative/Anthropological Models Examples: – – – – The approaches known in evaluation as naturalistic or “Fourth Generation” Evaluation The various qualitative schools Critical Theory and art criticism approaches The “Grounded Theory” approaches of Glaser and Strauss (Trochim, 2006) 4. Participant-Oriented Models • Emphasize the central importance of the evaluation participants (Trochim, 2006) 4. Participant-Oriented Models Examples: – Client-Centered Research – Stakeholder Approaches – Consumer-Oriented Research (Trochim, 2006) So, how do you decide which one to use??? Things to keep in mind: • Great evaluators use ideas from each of the 4 strategies as they need them • Methodologies needed will and should be varied – This should not be an easy process • There is no simple answer!!! (Trochim, 2006) Types of Evaluations: • There are many different types of evaluations Types of Evaluations: • 2 Important Types: 1. Formative Evaluation 2. Summative Evaluation (Trochim, 2006) 1. Formative Evaluation • Purpose: – To strengthen or improve the object being evaluated (Trochim, 2006) • In other words, to provide useful feedback to for the greater good of the “object” 1. Formative Evaluation • Includes several Evaluation Types – – – – – Needs Assessment Evaluablitiy Assessment Structured Conceptualization Implementation Evaluation Process Evaluation (Trochim, 2006) 1. Formative Evaluation • Questions generally asked: – What is the definition and scope of the problem or issue, or what’s the question? – Where is the problem and how big or serious is it? – How should the program or technology be delivered to address the problem? – How well is the program or technology delivered? (Trochim, 2006) 2. Summative Evaluation • Purpose: – To examine the effects or outcomes of some object (Trochim, 2006) • In other words, to review the results/conclusions of an “object” 2. Summative Evaluation • Can be subdivided into different categories: – – – – – Outcome Evaluations Impact Evaluation Cost-Effectiveness and Cost-Benefit Analysis Secondary Analysis Meta-Analysis (Trochim, 2006) 2. Summative Evaluation • Questions generally asked: – What type of evaluation is feasible? – What was the effectiveness of the program or technology? – What is the net impact of the program? (Trochim, 2006) Limitations • Limitations of this type of research include the level of expertise and knowledge of the expert panels – If the evaluation is only as good as the knowledge of the expert panel Journal Article A Review of Subject Matter Topics Researched in Agricultural and Extension Education By: Rama B. Radhakrishna & Wenwei Xu A Review of Subject Matter Topics Researched in Agricultural and Extension Education Retrieved from: Journal of Agriculture Education http://pubs.aged.tamu.edu/jae/search/defa ult.asp Purpose of Study • The purpose of the study was to “examine subject matter topics researched in agricultural education over a ten year period” Objectives • The objectives of the study were to: 1. Identify subject matter topics researched in agricultural extension education in the last decade (1986 – 1996) 2. Categorize subject matter topics published in the Journal of Agricultural Education and proceedings of the National Agricultural Education Research Meeting over a ten year period (1986-1996) How the Evaluation Method was used • 2 sources were used to gain data – Journal of Agricultural Education – Papers Presented at the National Agricultural Education Research meetings (NAERM) from 1986-1996 How the Evaluation Method was used • Resulted in 402 journal articles and 451 papers How the Evaluation Method was used • Each paper was given a code number • Each was reviewed and categorized into relevant subject matter categories – 3 Criteria were used: • Title of Study • Central Theme and Focus • Findings and conclusions How the Evaluation Method was used • 25 subject-matter topics were identified • The list of the 25 topics was given to a panel of experts for review and validation How the Evaluation Method was used • Panel of Experts – Asked to comment on: 1. The appropriateness of categories 2. Add or identify categories they thought that were 3. left out or delete categories they felt did not fit Suggest whether some categories could be combined How the Evaluation Method was used • Revised list of subject-matter categories contained 30 subject matter topics • Data was then summarized using frequencies and percentages Some of the Findings Top 5 Studies: • 75 Secondary Agriculture Programs Studies • 70 Styles/Theory and Cognition Studies • 43 Professionalism Studies • 42 Extension Studies • 38 Agriculture Mechanics/Safety Studies Conclusions of Study “Study provided information on subject matter topics investigated by agricultural and extension educators, which in turn provides perspectives about the research efforts of the agricultural and extension education profession.” Summary of Evaluation Research Method • The evaluation method was used in this study to assess what topics were researched in the agricultural education field during a 10 year period. Summary of Evaluation Research Method • The outcomes of this study will provide useful feedback to those in the profession References • Radhkrishna, R.B. & Xu, W. (1997). A review of subject • matter topics researched in agricultural and extension education. Journal of Agricultural Education. 38(3), 5969. Trochim, W.M. (2006) Research methods knowledge base. Drake University. Retrieved from: http://www.socialresearchmethods.net/kb/intreval.htm on 10/26/06. Any Questions??? Thank You