Summing Up: Research Process Revisited Research Methods for Public Administrators Dr. Gail Johnson Dr. G. Johnson, www.ResearchDemystified.org 1 What is Research? A systematic search for answers to questions. Search: to uncover, examine, find by exploration, to investigate, to inquire. Research: "the systematic inquiry into a subject in order to discover or revise facts, theories Dr. G. Johnson, www.ResearchDemystified.org 2 Deductive Vs. Inductive Logics Deductive: Sherlock Holmes Testing theory Inductive: Building theory from the ground up Dr. G. Johnson, www.ResearchDemystified.org 3 Planning Phase 1. Determining your questions 2. Identifying your measures and measurement strategy 3. Selecting a research design 4. Developing your data collection strategy A. Methods B. Sample 5. Identifying your analysis strategy 6. Reviewing and testing your plan Dr. G. Johnson, www.ResearchDemystified.org 4 Planning: An Iterative Process Not as linear as its presented The process is like going through a funnel It feels like going around in circles but things get very narrow and focused-at the end There are no shortcuts Dr. G. Johnson, www.ResearchDemystified.org 5 Step 1: What’s the Question? Identify the issue or concern What’s already known? Literature review, talk with experts Is there a theory? Engage the stakeholders Clarify the issues What matters most? Choosing the question is deceptively difficult and takes way longer than people expect Dr. G. Johnson, www.ResearchDemystified.org 6 Types of Questions Descriptive: what is Normative: comparison to a target Impact/causal: cause-effect Tough to answer and require: Logical theory Time-order Co-variation Eliminates all other rival explanations Dr. G. Johnson, www.ResearchDemystified.org 7 Step 2: Developing a Measurement Strategy Identify and define all the key terms. Do tax cuts cause economic stability? What tax cuts? What is economic stability? Develop operational definitions that translates a concept into something that can be concretely measured. How much money? How is economic stability measured? Will need to think about natural variation too. Dr. G. Johnson, www.ResearchDemystified.org 8 Step 2: Developing a Measurement Strategy Boundaries. Scope of the study: who’s included and excluded, time frame, geographic location(s). Tax cuts during what time frame? Are we talking U.S.? Unit of analysis. If talking about U.S., then we need to look at national data. If looking at a state, then we need to look at state data Dr. G. Johnson, www.ResearchDemystified.org 9 Levels of Measurement Nominal: names and categories Ordinal: has an order to it but not real numbers Interval and Ratio: real numbers Remember: different levels of measurement use different analytical techniques Dr. G. Johnson, www.ResearchDemystified.org 10 Validity and Reliability Valid measures: the exact thing you want to measure The number of books in a library is not a good measure of the quality of the school Infant mortality rate—often used as an indicator of the quality of the health care system Reliable measures: the exact thing measured in the exact same way every time A steel ruler rather than an elastic ruler Dr. G. Johnson, www.ResearchDemystified.org 11 Reliability and Validity Remember: The poverty measure is reliable because it is measured the same way every time But many question its validity because it does not accurately measure actual living costs and various financial benefits that some people might receive Dr. G. Johnson, www.ResearchDemystified.org 12 Step 3: Research Design The Xs and Os framework Experimental Random assignment to treatment or control (comparison) groups Quasi-experimental Non-random assignment to groups Non-experimental The one-shot design: implement a program and then measure what happens Dr. G. Johnson, www.ResearchDemystified.org 13 Question-design Connection One-shot designs make sense for descriptive questions and normative questions but are weakest for cause-effect questions Dr. G. Johnson, www.ResearchDemystified.org 14 Question-Design Connection The best design for a cause-effect question is the classic experimental design But: quasi-experimental designs, including using statistical controls, are more typically used in public administration research But: sometimes a one-shot design is as good as it gets in public administration research Sophisticated users exercise great caution in drawing cause-effect conclusions Dr. G. Johnson, www.ResearchDemystified.org 15 Research Design: Other Common Research Approaches Secondary Data Analysis Evaluation Synthesis Content Analysis Survey Research Case Studies Cost-Benefit Analysis Dr. G. Johnson, www.ResearchDemystified.org 16 Step 4: Data Collection Options The decision depends upon: What you want to know Numbers or stories Where the data resides Environment, files, people Resources available; time, money, staff to conduct the research Dr. G. Johnson, www.ResearchDemystified.org 17 Data Collection Methods Locate sources of information Data collection methods: Available data Archives, documents Data collection instruments (DCIs) Observation Interviews, focus groups Surveys: mail, in-person, telephone, cyberspace Dr. G. Johnson, www.ResearchDemystified.org 18 Multiple Methods Quantitative-Qualitative war is over Neither is inherently better Law of the situation rules Each work well in some situations, less well in others Dr. G. Johnson, www.ResearchDemystified.org 19 Multiple Methods Quantitative and qualitative data collection often used together Available data with surveys Surveys with observations Observations with available data Surveys with focus groups Dr. G. Johnson, www.ResearchDemystified.org 20 Nonrandom Sampling Useful in qualitative research Sometimes a nonrandom sample is the only choice that makes sense Weakness: potential selection bias Do they hold a particular point of view that suits the agenda of the researchers? Limitations: reflects only those included Results are never generalizable Dr. G. Johnson, www.ResearchDemystified.org 21 Nonrandom Sampling Non-random sample options Quota Accidental Snow-ball Judgmental Convenience Face validity: does the choice make sense? Size is not important Dr. G. Johnson, www.ResearchDemystified.org 22 Random Sample Based on probability: ever item in the population (people, files, roads, whatever) has an equal chance of being selected Size matters: must be large enough—sample size table is needed Advantages: Ability to make inferences or generalizations about the larger population based on what we learn from the sample Eliminates selection bias Dr. G. Johnson, www.ResearchDemystified.org 23 Random Sample Challenge: To locate a complete listing of the entire population from which to select a sample Analysis requires inferential statistics There is a calculable amount of error in any random sample Confidence level, confidence intervals and sampling error (also called margin of error) Tests of statistical significance Dr. G. Johnson, www.ResearchDemystified.org 24 Step 5: Analysis Plan Analysis techniques vary based on: Level of data collected Sampling choices The analysis plan links the data collection instruments, the questions and the planned data analyses Check to make sure all the needed data are collected to answer the questions Check to make sure unneeded data are not collected Dr. G. Johnson, www.ResearchDemystified.org 25 Step 6: Test Your Plan Test all data collection instruments and data collection plans to make sure they work the way expected Pre-test in real settings Expert review Cold-reader review Revise and re-test Finalize the research proposal Dr. G. Johnson, www.ResearchDemystified.org 26 The Design Matrix A tool that helps pull on the pieces of the research plan together Helps focus on all the details to make sure everything connects It is a visual Focus is on content not writing style It is a living document Planning is an iterative process This is generic format Dr. G. Johnson, www.ResearchDemystified.org 27 “Begin with the end in mind.” It is worth the time planning and testing your plan. “If you do not know where you are going, you can wind up anywhere.” It is hard to correct mistakes after the data has been collected Remember: no amount of statistical wizardry will correct planning mistakes Dr. G. Johnson, www.ResearchDemystified.org 28 Doing Phase Gathering the data Preparing data for analysis Analyzing and interpreting the data Dr. G. Johnson, www.ResearchDemystified.org 29 Doing Phase Collect the data Accuracy is key Prepare data collected for analysis Data entry, test error rate Analyze the data Qualitative approaches Control for bias Quantitative approaches Numeric analysis Interpretation in the English language Dr. G. Johnson, www.ResearchDemystified.org 30 Analysis Techniques Descriptive Frequency, percents, means, medians, modes, rates, ratios, rates of change, range, standard deviation Bi-Variate Cross-tabs, comparison of means Dr. G. Johnson, www.ResearchDemystified.org 31 Analysis Techniques Relationships Correlation and measures of association Association does not mean the variables are causally related The closer 0, the weaker the relationship, the closer to 1, the stronger the relationship No defined rules: .2 to .3, something to look at, . 4 to .5 moderately strong, and above .5 is strong It is rare to get correlations above .9 Signal to take a more careful look at the measures Dr. G. Johnson, www.ResearchDemystified.org 32 Analysis Techniques Inference To infer something to the larger population based on the results of a random sample. Confidence intervals (standard is 95% precision) Confidence levels (standard is 95% confidence) – I am 95% certain (confident) that the true average salary in the population is between $45,000 and $50,000. Sampling error (standard is plus/minus 5%) – Think polling data Dr. G. Johnson, www.ResearchDemystified.org 33 Analysis Techniques Inference Statistical Significance: How likely are we to have gotten these results from chance alone? Many different tests to meet specific situations but interpretation is always the same Standard practice: If there is a 5% chance or less that the results are due to chance, researchers will conclude that the results are statistically significant. Dr. G. Johnson, www.ResearchDemystified.org 34 Reporting Phase What’s Your Point? Major message/story? Who is your audience? Reporting Options Executive summary Written Reports Oral presentations Use of charts and tables Dr. G. Johnson, www.ResearchDemystified.org 35 Communication Guidelines Present what matters to your audience The goal is to illuminate, not impress What’s your hook? Grab your audience’s attention Use clear, accurate and simple language Use graphics to highlight points Avoid jargon Dr. G. Johnson, www.ResearchDemystified.org 36 Communication Guidelines Organize around major themes or research questions Decide on your message and stick to it Leave time for reviews by experts and cold readers Leave time for the necessary revisions Dr. G. Johnson, www.ResearchDemystified.org 37 Communication Guidelines Provide information about your research methods so others can judge its credibility Always provide the limitations of this study For reports: Place technical information in an appendix Provide an executive summary for busy readers Dr. G. Johnson, www.ResearchDemystified.org 38 Oral Presentations Consider the needs of the audience Consider the requirements of the situation High tech or low tech? Formal or informal? Powerpoints as appropriate: simple, clear, large font, no distracting bells and whistles Not too many, not too few: just right Handouts as needed Dr. G. Johnson, www.ResearchDemystified.org 39 Visual Display Of Data Tables: better for presenting data Graphs/charts: more effective in communicating the message. Impact Increases audience acceptance Increases memory retention Shows big picture and patterns Visual relief from narrative. Dr. G. Johnson, www.ResearchDemystified.org 40 Ethics and Values Tell the truth--always. Interpretations can have spin: may not be agreed criteria for what is “good.” Build in checks to assure accuracy. Be honest about the limitations of your research. Do no harm to subjects of your research. Dr. G. Johnson, www.ResearchDemystified.org 41 Ethics and Values Don’t take cheap shots at other people’s research Don’t accuse of wrong-doing without evidence Be careful not to harm people who benefit from program by concluding a program does not work when all you know is that your study was not able to find an impact. Dr. G. Johnson, www.ResearchDemystified.org 42 Creative Commons This powerpoint is meant to be used and shared with attribution Please provide feedback If you make changes, please share freely and send me a copy of changes: Johnsong62@gmail.com Visit www.creativecommons.org for more information Dr. G. Johnson, www.ResearchDemystified.org 43