Topic Outlines - Michigan State University

advertisement
PRR 844 Research Methods in Parks, Recreation & Tourism
Daniel J. Stynes
TOPIC HANDOUTS
TOPIC 1. INTRODUCTION TO PHILOSOPHY OF SCIENCE
2
TOPIC 2. RECREATION, TOURISM & LEISURE AS RESEARCH AREAS.
7
TOPIC 3. RESEARCH PROPOSAL
10
TOPIC 4. RESEARCH DESIGN OVERVIEW/ RESEARCH PROCESS
11
TOPIC 5. DEFINITION & MEASUREMENT
16
TOPIC 6. SAMPLING
25
TOPIC 7A. SURVEY METHODS
27
TOPIC 7B. EXPERIMENTAL DESIGN
29
TOPIC 7C. SECONDARY DATA DESIGNS
31
EXAMPLES: USING SECONDARY DATA TO ESTIMATE
32
TOPIC 7D. OBSERVATION & OTHER METHODS
32
TOPIC 8. DATA GATHERING, FIELD PROCEDURES AND DATA ENTRY
(INCOMPLETE)
33
TOPIC 9. DATA ANALYSIS AND STATISTICS
34
TOPIC 10. RESEARCH ETHICS : ACCEPTABLE METHODS & PRACTICES
38
TOPIC 11. RESEARCH WRITING & PRESENTATIONS
38
TOPIC 12. SUPPLEMENTAL MATERIAL
39
APPLYING RESEARCH
39
EVALUATING RECREATION SURVEYS - A CHECKLIST
40
PRR 844 Topic Outlines
Page 2
TOPIC 1. Introduction to Philosophy of Science
A. Science can be viewed as **A body of knowledge that is:
systematic: propositions related within body of theory
abstract: doesn't explain everything, simplifications, assumptions.
general: aim for general laws, not explanation of isolated events.
parsimonious: prefer simpler explanation
OR
** A method of inquiry that is:
logical: hypothetico-deductive system, deduction, induction
self-corrective: iterative search for knowledge
empirical: ultimate test of truth requires testing with real world observations
no single scientific method: many different approaches-must fit method to the purpose and characteristics of the inquiry.
B. Alternatives to the scientific method -- C.S. Pierce - methods of fixing belief:
Tenacity: repeat belief until accepted
Authority: rely on an accepted or noted authority
Intuition: rely on common sense or intuition
Scientific method : systematic and objective gathering of empirical data to support or refute ideas.
Example : beliefs about the benefits of recreation or leisure. Can base these on repetition to convince people of
the benefits, reliance on a noted authority (NRPA, Academy of Leisure Science, or endorsement by sports
figure or scholar) or tradition (we’ve known this for many years), our intuition that recreation is good for you,
or studies to measure what the benefits actually are.
What is the basis for the following beliefs?
Recreation reduces juvenile delinquency.
Tourism is beneficial for a community.
Recreation facilities should be accessible to those with disabilities.
Notice that the last statement is a value judgment rather than a statement of fact that can be empirically tested.
Beliefs including those about management practices rest on a mix of tradition, intuition, authority and science.
A good scientist asks for empirical support for propositions - this means making objective observations in real
world and letting the facts determione "truth".
C. Epistemology is a branch of philosphy dealing with the theory of knowledge. Philosophy of science is a part of
epistemology. Major contributors include Plato, Aristotle, Descartes, Bacon, Locke, Hume, Berkeley, Carnap,
Russell, Pierce, Dewey, and James. See any introductory philosophy text or set of readings.
Ontology: what can be known about the world
Epistemology: philosophy of how we come to know things, relations between knower and known
Methodology : practical aspects of how we know, the science of methods
Current philosophy directly relevant to research methods divides into two camps that roughly parallel
quantitive vs qualitative methods. Quantitatve methods come from traditional science and the positivist
philosophy. Positivism was a reaction to metaphysical explanations of the natural world. Positivists subscribe
to Lord Kelvin’s statement that “measurement is the beginning of scence”. If we can’t measure something we
really can’t study it scientifically. The stereotypical scientist in a lab coat conducting experiments and making
measurements is a positivist. Post-positivism is a term that refers to modern views of science. These range
from modest updates of positivism that more fully recognize the fallibility of scientific methods to interpretive
paradigms that completely reject the notion of absolute truth. Qualitative methods are grounded in
phenomenology, hermaneutics, and related epistemologies.
2
PRR 844 Topic Outlines
Page 3
D. Qualitative vs Quantitative Methods
We will primarily cover what are known as quantitative research methods in this course. These are based in logical
positivism and encompass "traditional science". A somewhat distinct set of methods are based in phenomonology
and other non-positivist theories of knowledge. These are loosely termed "qualitative research methods". With roots
in sociology and anthropology, qualitative methods are now used in almost all sciences. They have become
increasingly popular in leisure science within the past ten years.
Qualitative methods include: ethnography, focus groups, in-depth interviews, case studies, historical analysis,
participant observation and related techniques.
There are three key differences between qualitative (QL) and quantitative (QN) methods:
1. Purposes: QN seeks general laws, testing & verification, and tends to focus on behavior. QL studies
particular instances, seeks understanding (verstehen), and focuses more on intentions and meanings.
2. Perspective: QN takes an outsider (scientist's perspective). The researcher is an objective observer. QL tries
to measure and study phenomona from an insider perspective- observe from the subject or actor's frame of
reference.
3. Procedures : QN uses standardized and structured procedures, operational definitions, probability sampling.
QN tends to be reductionist (reduces things to their most important aspects). Interpretation is separate from
analysis. QL uses unstandardized procedures, actor defined concepts, and non-probability samples. QL
tends to be holistic. Interpretation tends to be interlinked with procedures and analysis and hence somewhat
inseparable from them.
Both sets of methods are useful, although usually for quite different purposes. The key is matching the methods with
the purpose. Qualitative methods tend to be used more for exploratory research, although this is not their only use.
Ideally QL and QN methods should be integrated to study a particular problem. One may start with a qualitative
approach talking with key informants observing as a participant, etc. From this are generated hypotheses that can be
tested formally within a QN approach, maybe a survey. Results of QN suggest additional study for a more in-depth
understanding of the phenomonon. This may suggest further QL type research, maybe followed again by QN. While
QN and QL are often presented as competing scientific paradigms, they are complementary more than competitive.
QL research, being less standardized, is more difficult for the novice researcher. QL data is more difficult to analyze
and to write up. It is harder to publish. Some QL work is more properly labeled philosophy than research. Good QL
research requires first a good understanding of QN methods. There is a tendency for those researchers who are
uncomfortable with statistics and mathematics to gravitate toward QL methods, while people who are good at
mathematics and analytical thinking favor QN.
Examples of recreation research problems QL is well-suited to:
1) Evaluating effects of a unique recreation program on a specific group. eg. Has pet therapy been successful in
my nursing home. May be no need here to generalize beyond a given home or group of patients.
2). Focus groups to understand people's motivations for recreation, the benefits they obtain from particular
products or experiences, etc. Results might yield a list of motivations, attributes or benefits for a structured
questionnaire. Focus groups widely used to evaluate advertising.
3). Understanding small group dynamics in recreation settings
4). Case studies of particular programs or events.
5). Participant observation by a researcher prior to a QN investigation. eg. spend a day in a wheelchair before
designing a study about handicapper accessibility.
6). Studying the meaning of leisure to people.
Remember that most of the basic information commonly used to support management, planning, marketing and
evaluation is inherently quantititaive and requires careful, objective measurements, i.e. numbers of visitors, days of
recreation, costs, spatial distribution of supply and demand, market characteristics, spending, etc. Qualitative
methods add to but do not substitute for quantitative type analyses.
3
PRR 844 Topic Outlines
Page 4
SELECTED REFERENCES ON QUALITATIVE METHODS: need to update, particularly applications move to
later, substitute general science/philos references here.
Bogdan, R.C. & Biklen, S.K. 1992. Qualitative research for education: An introduction to theory and methods.
Boston: Allyn & Bacon.
Creswell, John W. 1994. Research Desugn: Qualitative and Quantitative Approaches. Thousand Oaks, CA: Sage.
Dezin, N.KJ. and Lincoln, Y.S. (eds). 1994. Handbook of qualitative research. Thousand Oaks, CA: Sage.
Glaser and Strauss. 1967. The Discovery of Grounded Theory.Chicago: Aldine.
Henderson, Karla. 1991. Dimensions of Choice: A qualitative approach to recreation, parks and leisure. State
College, PA: Venture.
Howe, C.Z.1985. Possibilities of using a qualitative approach in the sociological study of leisure. Journal of
Leisure Research 17(3): 212-224. - a review paper on qualitative methods.
Lincoln, Y.S. and Guba, E.G. 1985. Naturalistic inquiry. Beverly Hills, CA: Sage.
Maanen (Ed). 1983. Special issue of Administrative Science Quarterly on qualitative methods.
Marshall, C. and Rossman, G.B. Designing qualitative research. Newbury Park, CA: Sage.
Miles, M. and Huberman, A. 1984. Qualitative data analysis. Beverly Hills, CA: Sage.
Patton, M.Q. 1987. How to use qualitative methods in evaluation. Newbury Park, CA: Sage.
Rosenau, P.M. 1992. Post-modernism and the social sciences. Princeton NJ: Princeton Univ. Press. - philosophical
treatment of post-modern theories of knowledge.
Schwartz and Jacobs. 1979. Qualitative Sociology. Free Press: NY.
Strauss, A. and Corbin, J. 1990. Basics of Qualitative Research. Newbury Park, CA: Sage.
Chirban, J.T. 1996. Interviewing in depth. Thousand Oaks, CA: Sage.
Krueger, R.A. 1994. Focus Groups; A practical guide for applied research. 2nd edition. Thousand Oaks, CA:
Sage.
Rubin, H.J. & Rubin, I.S. 1995. Qualitative interviewing. Thousand Oaks, CA: Sage.
Yin, R.K. 1989. Case Study Research: Design and Methods. Revised edition. Newbury Park, CA: Sage
SELECTED RECREATION AND TOURISM APPLICATIONS
Brandmeyer JLR 1986 18 (1): 26-40 - baseball fantasy camp
Glancy JLR 1986 18(2): 59-80. - participant observation in a recreation setting
Glancy, M. 1988. JLR 20(2):135-153. Play world setting of the auction
Hartmann, R. 1988. Combining field methods in tourism research. Annals of Tourism Research 15: 88-105.
Henderson, K.A. 1987. Journal of Experiential Education 10(2): 25-28. woman’s work experience
.1988. Leisure Sciences 10(1); 41-50 . Farm women & leisure
Hummell JLR 1986 18(1): 40-52.
Shaw, S. 1985. Leisure Sciences 7(1): 1-24. Meaning of leisure.
.
Also see works of Denzin, Filstead, Sage Series on Qualitative methods, methods texts in sociology and
anthropology, books on ethnography, participant observation, focus group interviewing, etc.
Exercise : Find and read a research article using a qualitative approach. Write up a brief summary (one page or less)
including :
a). Complete citation in APA format
b). Brief summary of topic, study objectives
c). Description of methods used
d). Principal conclusions
e). Your observations on the study and pros and cons of the qualitative approach.
f). Also hand in a copy of the article.
4
PRR 844 Topic Outlines
Page 5
E. Sciences may be divided into the natural sciences (physics, chemistry, biology), social sciences (sociology,
psychology, economics, geography, political science, communication, anthropology) and a variety of applied
and management sciences (e.g., management, marketing, social work, forestry, engineering, urban planning,
recreation and tourism). Over time, the basic disciplines have subdivided into specialty areas and formed
hybrids such as economic geography and social psychology. Many of the applied sciences are interdisciplinary,
borrowing from the more basic disciplines while also developing their own unique bodies of theory and
methods. Recreation, leisure and tourism are applied sciences that borrow both from basic disciplines as well as
other applied sciences.
F. Research is the application of scientific methods, the search for knowledge, controlled inquiry directed toward the
establishment of truth. Purposes of research (Stoltenberg):
1. answer questions of fact arising during management (problem solving)
2. develop new alternatives for management (developmental research)
3. answer questions of fact arising during research (methods)
(4.) answer questions for sake of knowing (pure research, theory)
G. Pure vs applied research -as a continuum
-pure or basic research: for its own sake, more general, longer time to use
-applied research: to solve a problem, for client, immediate application.
Most recreation and tourism research is applied research although some basic research is conducted, for
example on the nature and meaning of leisure. Think of basic and applied research as endpoints of a
continuum. An individual study may be placed along this continuum based on the degree to which it is done
for its own sake or to solve a problem, the generality of the research and the client and timeframe for
application. The use of complex models or statistical tools does not make a study basic or applied, as both
simple and complex analysis tools may be used in either type of study. Theories and models are equally
useful in applied and basic research, although theory development itself tends toward the basic end of the
continuum.
H. Stages in the development of a science
exploratory to descriptive to explanatory to predictive research
definition, measurement, quantification, theory building
Thomas Kuhn - Structure of Scientific Revolutions examines how sciences evolve and change over time. He
stresses the importance of paradigms to help structure and guide research. A paradigm is an accepted set of
assumptions, theories, and methods within a given science. In the pre-paradigm phase of a science
investigation is somewhat random and disorganized. Once a set of methods, assumptions and theories are
accepted, research proceeds in a more organized and cumulative fashion as the paradigm identifies the most
important variables to measure and suggests the appropriate methods, while also providing the structure into
which results may be integrated. Scientific revolutions occur when a competing paradigm is adopted and
“overthrows” the old ideas. In astronomy the switch from an earth centered solar system to the present
model was a paradigm shift. What paradigms guide recreation and tourism research?
I.
Research communication chain. As applied fields of study, recreation and tourism research involve
considerable interaction between scientists and practitioners . MSU, as a land grant school, is grounded in a
philosophy of using research to help solve problems. This requires a chain of communication between research
centers and the field. In the simple model, problems flow up from the field to the University where they are
addressed by researchers. Solutions are then extended back to the field where the results are applied. Today
one more frequently finds researchers working directly with clients in the field to help identify and solve
problems.
pure research---applied research----extension---practitioner
5
PRR 844 Topic Outlines
Page 6
J. Some mistaken notions
1. Theoretical means not useful - theories play very important roles both in science and management.
Theories help to organize knowledge and direct further investigation. Management decisions, like research, are often
based on "theories" of people's behavior, how markets work, etc. Almost all actions and decisions are based on
some theory or set of assumptions. Question is whether theories/assumptions are well-founded and useful.
2. Some things are measureable and others are not. Measurement is as much a function of our creativity and
success in developing measurement instruments than characteristics inherent in the things being measured. Individual
measures almost always capture only a narrow aspect of the thing being measured and there are always many
different potential measures of any characteristic. For example, the usual measure of age captures only the time that
has passed since birth. While this measure may be correlated with physical health, attitudes etc., it doesn't measure
these traits. Don't criticize a measure for not measuring something it doesn't purport to measure. Suggest a different
measure. Many things we measure today were deemed "unmeasureable" in the past until suitable instruments and
procedures were developed, e.g. the thermometer to measure temperature.
Classification is a form of meaurement. When we place a person into a religious, racial or ethnic group or
classify people based on hair color, we are making a measurement of the underlying attribute. We regularly measure
"qualities" with quantitative measurements, e.g. personality, intelligence, beauty. You can argue how good these
measures are or which dimensions of the underlying attributes they capture, but arguing they can't be measured
suggests either an anti-scientific bias or "giving up". Remember that no measure is perfect or complete.
Science and application progress by improving our measurements and using these to better understand relationships
between variables.
3. Can we measure the beauty of a sunset? Yes, if we have a clear understanding of what the concept
"beauty" means and have devised suitable instruments/procedures to measure it. If the concept rests on people's
perceptions, then attitude surveys and rating techniques can be used --just as we "rate" the performance of a gymnast
or ice skater, we could rate sunsets. Perhaps one can identify the characteristics that define the beauty of a sunset
and develop rating guidelines that "experts" can apply, as they do in evaluating skaters. Or if beauty can be defined
by physical features, perhaps an instrument/formula can be devised based on wavelengths of light emitted.
Remember that prior to themometer's, temperature was measured by people's perceptions of hot/cold. Only when we
understood the relationship between temperature and expansion of mercury could a more precise and standard
quantitative measure be developed. Is the notion of warmth inherently more measureable than beauty?
6
PRR 844 Topic Outlines
Page 7
TOPIC 2. Recreation, Tourism & Leisure as Research Areas.
In this section, we explore the nature of recreation, parks, leisure, and tourism as research areas and also as sciences.
Are these sciences and if so, what kind of science(s)? Before we can discuss these questions, we will need to become
familiar with the research literature, some of the history of recreation and tourism research, and some of the
important research topics, theories and methods.
A. Characteristics of the recreation field
1. Relatively new field for research-start about 1960 with the ORRRC studies in US.
2. Interdisciplinary field.
3. Applied Field
4. Divisions between resource-oriented, program oriented; urban-rural, recreation, parks, tourism
5. Status of recreation as research area is low
6. Few highly trained researchers
7. Little theory of its own
8. Heavy reliance on surveys and descriptive studies
B. Characteristics of tourism as a "science"
1. Lacks clear academic home, cuts across business, planning, recreation, geography,…
2. Divisions between international-domestic, marketing-social-environmental, hospitality-resource-based
3. More & better research outside U.S.
4. Heavy dose of consulting and commercial studies relative to academic ones
5. Strong political influence on research
6. Marketing studies and advertising evaluation dominate
7. Convergence of tourism and recreation research since 1985.
C. WHO is doing recreation & tourism research:
1. UNIVERSITIES (North America): About a dozen recreation/park/leisure studies/tourism curricula making major
contributions: Michigan State, Texas A&M, Illinois, Waterloo, Oregon State, Penn State, Clemson, Indiana.,
Colorado State. One or two individuals at many other institutions. Scattered, but significant contributions from
individuals in disciplines (geography, sociology, psychology, economics,...) or related applied fields (HRI, forestry,
fish & wildlife, business, marketing, agric. economics, health and medicine,...) within Universities.
2. GOVERNMENT:
a. Federal: Over 30 agencies involved in recreation-related research. e.g. USDA, Forest Service, National
Park Service, Sports fisheries & Wildlife, US Army Corps of Engineers, Bureau of Land Mgmt, Bureau of
Reclamation, NOAA-Sea Grant Program, Dept. of Commerce, Health, Education, Welfare, NIH, NSF,
Transportation. Federal role in tourism research mainly periodic surveys in Transportation (BTS) and in-flight
surveys of international travelers.
b. State: Considerable planning-related research as part of SCORP (recreation) and other planning. Also as
part of forestry, fish & wildlife, water resource, land use, tourism, and economic development. State travel offices a
primary funder of travel research, some from CVB and industry associations.
c. Local & municipal: mostly planning studies with small research component. Larger efforts in major
metropolitan park systems like Hennepin county, MN, Huron Clinton and Cleveland metroparks. Metro CVB's.
3. FOUNDATIONS and Non-Profit Organizations: Ford, Rockefeller, Resources for the Future, Wilderness Society,
Sierra Club, Appalachian Mt. Club, NOLS etc. Tourism - TIA, WTO, WTTC.
4. NRPA, National Recreation & Park Assoc: A little & growing. TTRA = Travel & Tourism Research Assoc.
7
PRR 844 Topic Outlines
Page 8
5. COMMERCIAL/PVT SECTOR: Large market surveys, industry sponsored research, small scale in-house
research. Use of consultants. Examples from campground, recreation vehicle, ski, fishing equipment, boating,
lodging, and other recreation/tourism industries. Numerous consulting companies, particularly in tourism (D.K.
Shifflet, Smith Travel, …) conduct large scale travel surveys and custom research.
6. INTERNATIONAL: Canada, Australia, United Kingdom, Netherlands have research similar to US programs; UK,
Netherlands focus on land use & planning. Europe: sociology of sport, cultural studies. USSR, Eastern Europe: time
budget studies. Tourism research more common globally.
Reference: Stanley Parker. A review of leisure research around the world. WLRA and International Handbook of
Leisure Studies and Research. See Graefe and Parker's Recreation & Leisure; An Introductory Handbook or
Parker's chapter in Barnett (1988), Research About Leisure. Ritchie & Goeldner. Travel, Tourism & Hospitality
Research.
D. WHAT - Selected Recreation Research Themes by Problem (with key contributors
prior to 1990)
1. Landscape aesthetics, visual quality: Daniel, Zube, Schroeder, Anderson, Brown, Buhyoff, Westphal, Vining.
See Taho conference -Our National Landscape.
2. Use estimation, demand modeling, valuation. economists: Clawson, Knetsch, VK Smith, GL Peterson,
Randall, Brookshire, Walsh, Dwyer, Cicchetti, Loomis, Sorg, Hoehn, Stoll, Talhelm, Brown, See PCAO...
geographers: Ewing, Baxter, Cesario, Fesenmaier, Timmermans.
3. Costs/Supply : Reiling, Gibbs, Echelberger, S. Daniels, Cordell, Harrington. See 1985, 1990 Trends Symposia,
Harrington RFF book.
4. Forecasting: Van Doren, Stynes, Bevins, Moeller, Shafer, BarOn, Guerts, Witt, Calantone, Sheldon, See
Stynes in PCAO.
5. Marketing : Crompton, Mahoney, Snepenger, Woodside, Etzel, Goodrich, Perdue. See JTR, Perdue chapter in
Barnett (1988).
6. Carrying capacity, crowding & satisfaction : Stankey, Lime, Lucas, Graefe, Heberlein, Schreyer, Shelby,
Manning book, Graefe review in LS, Stankey in PCAO.
7. Environmental. psych : Knopf, Hammitt, Fridgen, Williams, Schreyer
8. Recreation Motivations: Driver, Knopf, Brown et al.
9. Leisure theory: Tinsley, Pierce, Iso-Ahola, Kleiber, Ragheb, Mannell, Kelly, Godbey, Goodale.
10. Play & Human development: Barnett, Klieber, M. Ellis
11. Therapeutic Recreation: C. Peterson, Austin, Dickason, Witt, G. Ellis. See TRJ.
12. Methodology/Statistics: J. Christensen, Tinsley, Stynes, Samdahl, Fesenmeier
13. Tourism: Broad category emcompassing most of the others in context of travel-related use of leisure. Perdue,
Crompton, Gitelson, Fridgen, Hunt, Burke, Gartner, Uysal, Goeldner, Ritchie, Archer, Rovelstad, Jafari,
O'Leary, Becker, Var, Sheldon. See Ritchie & Goeldner, PCAO.
14. Evaluation of recreation programs : Howe, Theobald, van der Smissen, McKinney, Russell
15. Gen'l forest recreation mgmt: Brown, Driver, Lucas, Lime, Stankey, Harris
16. Depreciative behavior: Clark, Christiansen, Westover
17. Interpretation/communication: Roggenbuck, McDonough, Machlis, Ham
18. Social psychology of leisure. Iso-Ahola, Mannell, Csikszentmihalyi, Weissinger, Fridgen,...
19. Modeling : VK Smith, J. Ellis, Stynes, Levine, GL Peterson, Cesario, Fesenmaier, Timmermans, Louviere,
Ewing.
20. Economic impact : Archer, Propst, Alward, Schaffer, Maki, Stevens. See Propst (1984).
21. Rec Choice : Dwyer, Stynes, Peterson, Louviere, Timmermans, Vaske, see Choice Symposium.
22. Environmental impacts : Cole, van Wagtendonk, see PCAO and Wilderness Mgmt Proceedings.
23. Leisure time use: Robinson in PCAO
24. Pricing: Driver in 1985 Trends, Manning,
25. Others: Policy processes, Sports, Arts & culture, Social impacts, public involvement.
26. Benefits of leisure . See book edited by Driver, Brown & Peterson.
27. Constraints to Leisure - see book and artilcles by Edgar Jackson
8
PRR 844 Topic Outlines
Page 9
Major Tourism Research Themes: marketing, conversion studies, advertising evaluation, social, environmental
and economic impacts, segmentation, destination images, information search & processing, destination choice,
community attitudes. - See Ritchie and Goeldner, JTR, Annals, Tourism Management.
E. Selected Theories from disciplines:
Economics
Geography
Psychology
Sociology
Anthropology
Political Science
Communications
Biological
Marketing
Utility theory, choice, exchange, market behavior, decisionmaking, risk, uncertainty, and
value.
Spatial behavior, location theories, envir. perception, central place theory.
Choice, perception and cognition, learning, cognitive dissonance, attitude & behavior,
personality, locus of control, Flow
Social structure and change, group behavior, institutions, status, norms, conflict
Norms, institutions, cultural change.
Collective action, gaming, small group decisionmaking, power and influence.
Verbal and nonverbal behavior, information processing, mass media and interpersonal
communication, diffusion theories
Territoriality, predator-prey, adaptation, migration, population growth
Theories that combine economic, psychology, communications and geographic theories to
explain market behavior.
E. IMPORTANT DATES AND EVENTS IN RECREATION RESEARCH (prior to 1990)
1. Samuel Dana Problem analysis in forest recreation research, 1957
2. ORRRC Commission Report 1962
3. Conference on research in Ann Arbor, 1963
4. Economics of Outdoor Recreation, Clawson & Knetsch 1966
RFF Program on outdoor recreation research during 1960's.
5. National Acad. of Sciences, Program for outdoor rec. research, 1969
6. Journal of Leisure Research founded, 1969
7. First Nationwide Outdoor Recreation Plan 1973
8. Harper's Ferry Research Needs Workshop, 1974
9. NAS Report Assessing Demand for Outdoor Recreation, 1975
10. CORDS Reports, 1976
11. Leisure Sciences founded 1977
12. First NRPA/SPRE Leisure Research Symposium 1977
13. First National Outdoor Recreation Trends Symposium, 1980
14. NRPA/BOR research agenda project, 1982
15. PCAO Reports 1986
16. Wilderness research Conference 1985
17. Anniversary Leisure Research Symposium - Barnett (1988) book
18. Benchmark conference, RPA planning 1988
19. Recreation Benefits Workshop 1989.
20. 1990 Trends Symposium
Regular Symposia with published Proceedings/Abstracts
NRPA/SPRE Leisure Research Symposia; Since 1977; Abstracts published since 1981.
Canadian Leisure Research Symposia: 1975 (Laval); 1979 (Toronto);1983 (Ottawa); 1987 (Halifax); 1990
(Waterloo); 1994 (Winnepeg).
Southeast region symposia (SERR): since 1979
Northeast region symposia:since 1989
Travel and Tourism Research Assoc. since about 1973
CenStates, TTRA: Last few years
World Leisure and Recreation Assoc. (WLRA).
Research in the National Parks, since 1979.
IUFRO, SAF, Assoc. of American Geographers, Recreation Working groups
Trends every five years since 1980; Social Sci & Res Mgmt every other year since 1986?
9
PRR 844 Topic Outlines
Page 10
TOPIC 3. RESEARCH PROPOSAL
A. Purpose of research proposal
1. focus your research efforts, organize it into blocks.
2. structure final paper or report
3. communicate to client exactly what will be done. View proposal as a contract to be carried out.
4. Allow client or reviewers to make suggestions, review design, point out potential problems, oversights, etc.;
decide if this study can and should be done.
B. Suggested Proposal Format
1. Problem Statement:
a. put project in context, work from broad problem area to specific part you will tackle. Identify the
client, key dependent and independent variables.
b. establish importance of problem, justify the study, potential importance and uses of results.
c. delimit the problem: narrow problem down, define your terms.
d. Brief sketch of proposed approach.
e. Anticipated products and uses/users of results.
2. Objectives
a. Be specific, a one, two, three listing is preferred, itemizing objectives in some logical order.
b. State as testable hypotheses or answerable questions
c. Objectives should flow clearly from problem statement.
3. Review of Literature
a. Provide background theory and evidence of your knowledge of the subject area. Show you have
done homework, are capable in given area.
b. Show relationship of your study to other studies (completed and on-going). Link other research
specifically to your study in a logical way. Show how you will fill a gap, advance knowledge.
c. Help in defining problem, justifying importance, and identifying best approach.
d. This isn't an annotated bibliography. Review only most relevant studies and pinpoint linkages to this
one. A handful of best and most closely related studies usually sufficient.
4. Procedures : What are you going to do, when, how, to whom, and why?
a. Design: independent, dependent variables, how control for threats to validity and reliability.
b. Define variables, measures, instrumentation.
c. Define Population, sampling plan, sample size.
d. Data collection, field procedures.
e. Analysis; data processing, planned analysis and statistical tests.
5. Attachments: Time sequence of activities, Budget, Vita of personnel, Instruments
6. Overall Considerations
a. All sections should fit together like a puzzle and be presented in a logical fashion.
b. Be preecise, concise, and to the point.
c. Write for the intended audience
Identifying Research questions/ problems
To be added
10
PRR 844 Topic Outlines
Page 11
TOPIC 4. RESEARCH DESIGN OVERVIEW/ RESEARCH PROCESS
Five basic steps in a research study.
1. Define the research problem/ objectives of the study.
2. Choose appropriate research designs for a given problem
3. Collect data (secondary or primary)
4. Process and analyze data to meet each objective
5. Communicate results with clients/publics
1. Defining the problem is the most difficult and most important step in any study. Good research results
from good questions/problems that are clearly understood. This is especially true of applied research, which must
begin from clear problems that have been translated into one or more specific research questions or hypotheses.
Given that research resources are limited, it is important that priorities be established and resources are directed at
questions where the expected benefit-cost ratio is high. The most difficult task for both students and managers is
often that of defining clear, do-able research problems.
2. Once a problem has been defined, the research design task may begin. Most problems may be addressed
in a variety of ways, each offering advantages and disadvantages. The research design task is to choose the most
promising design, based upon a clear understanding of different approaches, the particular characteristics of the
problem at hand, and resource constraints. Almost all research design questions involve tradeoffs between costs and
accuracy. Choose the design to fit the problem and resources at hand, not necessarily based on what has been done in
past or your personal preferences.
3. Data collection tends to be a tedious task, but requires strict adherence to procedures established in the
design stage. Your analysis will be no better than the data you gather so careful attention to details here will pay off
later. In primary data collection interviewers must be trained and supervised. Procedures and data should be checked
as you proceed to identify and solve problems that may occur. When gathering secondary data, sources must be
carefully documented and the validity and reliability of all data must be assessed and hopefully cross-checked.
4. Data processing includes coding, data entry, cleaning and editing, file creation & documentation, and
analysis. In the first stage, data is transferred from surveys or records into the computer. Begin with the identification
and naming of variables, develop a codebook (including missing codes) and a data entry form/format. Clean and edit
data as it is entered. Eventually, you will create a special "system file" suitable for use by whatever statistical
package you are using. It is a good practice to write out the first 10 and last 10 cases prior to starting your formal
analysis to verify that data have been correctly entered or converted. Check that you have the correct number of
cases and variables.
Next run some exploratory data analysis procedures to get a feel for the data and further check for possible
problems. Run frequencies on nominal variables and other variables assuming a limited set of values. Run
descriptive statistics on interval scale variables. Check these results carefully before beginning any bivariate or
multivariate analysis. Analysis should proceed from simple univariate description to bivariate descriptive tables
(means by subgroup or two way tables) to explanation, hypothesis testing and prediction.
5. Identify the audience and their needs before preparing oral or written reports of results. In most studies
you should plan out a series of reports which might include





a comprehensive technical report or series of such reports,
an executive summary
journal articles for research audiences
reports and recommendations for management
articles and information for the general public/lay audiences
Large studies often contain many pieces that must be packaged for different purposes and audiences.
11
PRR 844 Topic Outlines
Page 12
Research Approaches
1. Primary data gathering:
a. Surveys (self-administered, telephone, personal interview): description, exploring relationships among variables.
b. Experiments (field, lab): test cause-effect relationships, evaluating interventions, impacts.
c. Observation: restricted to data that is accessible to observation, describing observable phenomona.
d. Qualitative (focus groups, in-depth interviews, partic. observation): explore interpret, and discover, study
meanings and intentions, describe unique phenomona.
2. Secondary data: Use of any data collected for some purpose other than that for which the data were originally
gathered. Original source may be any of the above. Most commonly, data are: Surveys: re-analysis of survey data
Government statistics: analysis of various population and economic censuses, or other regularly gathered social,
economic and environmental indicators.
Agency data : use, registration data, budget, personnel, client/patient records, etc.
Documents and written material : content analysis, literature reviews.
3. Measurement instruments:
Questionnaires-- elicit verbal or written reports from verbal or written questions or statements.
Observation -- one or more observers (people) record information
Physical instruments-- a variety of measuring devices from yardsticks to thermometers, to electronic counting
devices, cameras, tape recorders, barometers, various sensors, instruments to measure various components of
water, air, and environmental quality, ...
4. Other dimensions of research designs
a. Cross sectional or longitudinal (Time) - study phenomona at one point in time or more than one point. Trend,
cohort, and panel designs are examples of longitudinal approaches.
b. Structured vs unstructured approaches (formal vs informal)
c. Direct vs indirect approaches
d. Exploratory vs confirmatory: explore vs test
e. Descriptive, explanatory, predictive
f. Correlational vs cause-effect relationships
12
PRR 844 Topic Outlines
Page 13
Alternative Research Designs
Where Data are Collected
How Data are
Gathered
Household
On-Site
Laboratory
Othera
Personal
Interview
Surveys
Surveys & Field
Experiments
Focus Groups
Surveys & Field
Experiments
Telephone or
Computer
Inteview
Survyes
Computer
Interviews
Computer
Interviews
Surveys & Field
Experiments
Experiments
Surveys
SelfAdministered
Questionnaire
Observation &
Traces
NA
Observable
Characteristics
Observable
Characteristics
Observable
Characteristics
Secondary
Sources
NA
Internal Records
NA
Gov't , Industry &
Other external
sources
a. Other locations include highway cordon studies, mall intercept surveys, surveys at outdoor shows & other special
events, etc.
A few simple guidelines on when to use different methods:
1. Describing a population - surveys
2. Describing users/visitors - on-site survey
3. Describing potential users or general population - household survey
4. Describing observable characteristics of visitors - on-site observation
5. Measuring impacts, cause-effect relationships - experiments
6. Anytime suitable secondary data exists - secondary data
7. Short, simple household studies - telephone surveys
8. Captive audience or very interested population - self-administered surveys
9. Testing new ideas - experimentation or focus groups
10. In-depth study - in-depth personal interviews, focus groups.
13
PRR 844 Topic Outlines
Page 14
STEPS IN RESEARCH PROCESS
1. Define the Problem - Problem Analysis
a. Identify problem area
b. Immerse yourself in it
Literature: Relevant concepts, theory, methods, previous research
Talk with people involved
Observe
c. Isolate more specific research problem(s)
Identify key variables (dependent, independent), hypothesized relationships
Identify relevant populations
Identify key decisionmakers/clients/users or research
d. Specify research objectives and/or hypotheses
2. Select Research Design - Develop Research Proposal
a. Define key elements of the research problem
Information needed, variables
Population(s) to be studied
Resources available (time, expertise,money)
b. Evaluate alternative designs
Cross sectional, longitudinal, panel study,...
Qualitative or quantitative approach
Survey, experiment, secondary data, ...
Mail, telephone, personal interview, on-site or household
Instruments, questionnaires, observation, traces...
Census, sample, probability?, stratify?, cluster?
c. Specify measurement procedures
Define concepts, variables, measurement of each variable
Measurement scales & instrumentation
Assess reliability & validity of measures
d. Specify population and sampling design
Define population, identify sampling frame
Choose sampling approach
Choose sample size
e. Specify analysis approach
Data processing
Statistical analysis
Descriptive
Inferential - hypotheses to test
Intended tables, figures, format of anticipated results
f. Assess threats to reliability & validity, possible errors
g. Assess feasibility of the design - time, costs, expertise
h. Assess ethical issues, human subjects review
3. Implement the research design - Data gathering/Field procedures
a. Pre-testing of instruments and procedures
b. Choose sample and administer measurement procedures
c. Monitor the process, solve problems as they occur
4. Analysis and Reporting - Results
a. Data entry - coding, cleaning, ...
b. Preliminary analysis
c. Descriptive analysis
d. Hypothesis testing
e. Preparation of tables and figures
f. Writing/presenting results
5. Put it all together
a. Final report(s) and articles
b. Drawing conclusions, assessing limitations
c. Applying the results, implications for intended users
14
PRR 844 Topic Outlines
Page 15
POTENTIAL SOURCES OF ERROR IN A RESEARCH STUDY
Whether you are designing research or reading and evaluating it, it is useful to approach the task as one of
controlling for or looking for errors. The following is a list of the types of errors to watch for when designing research or
reading research reports. The principles and methods for research design are largely to control for error.
1. Problem Definition: Conceptualization of research problem may not adequately or accurately reflect the real
situation.
- use of a theory or assumptions that are faulty or do not apply
- research problem doesn't address the management questions
- reductionism - omission of key variables
2. Surrogate information error: variation between the information required to solve the problem and the
information sought by researcher.
3. Measurement error: variation between information sought and information produced by the measurement
process. (reliability and validity)
4. Population specification error: variation between the population required to provide needed information and the
population sought by the researcher. (rule for clearly defining the study population)
5. Frame error: variation between the population as defined by the researcher and list of population elements used
by the researcher.
6. Sampling error: variation between a representative sample and the sample generated by a probability sampling
method (sampling error estimates, checking for representativeness).
7. Selection error: variation between a representative sample and the sample obtained by a nonprobability sampling
method. (check for representativeness)
8. Nonresponse error: variation between the sample that was selected and the one that actually participated in the
study. (evaluating non-response bias).
9. Experimental error: variation between the actual impact of treatment and the impact attributed to it based on an
experimental design (pre-measurement, interaction, selection, history, maturation, instrumentation, mortality,
reactive error, timing, surrogate situation - will define later when we cover experimental design). (experimental
design)
10. Data processing errors: errors in coding and handling of data. (cleaning)
11. Analysis errors: Covers a variety of errors including violation of assumptions of statistical procedures, use of
inappropriate or incorrect procedures, mis-handling of missing values, calculation errors, and faulty
interpretation of results.
12. Reporting and Communication errors: Errors made in preparing oral or written reports including both
typographic and logical errors. Faulty interpretation of results made by users. (editing)
13. Application errors: Inappropriate or faulty application of the research results to a management problem. Overgeneralizing the results to situations where they may not apply is a common error in applying research results.
15
PRR 844 Topic Outlines
Page 16
TOPIC 5. DEFINITION & MEASUREMENT
1. TYPES OF DEFINITIONS
Nominal or Conceptual definitions define concepts in terms of other concepts.
Operational definitions define concepts in terms of a set of measurement procedures for generating the concept.
EXAMPLE 1. Conceptual definition - Length of table is the distance from one end to the other. Operational definition - Length of table
is the number that results from the following procedure : Place a yardstick along one edge of the table, place additional yardsticks end to
end until one extends over the other edge. Read the numeric marking for inches on the final yardstick at the point where it is exactly over
the other edge of the table (call this X inches). Count the number of yardsticks you have used (call this N). Compute (N-1) * 36 + X.
Repeat this process on the edge perpendicular to this one. The length of the table in inches is the larger of these two numbers.
EXAMPLE 2. Age- Conceptual definition- the number of years that have passed since a person's date of birth. Operational definitions:
A. Ask an individual for his name, place of birth (county and state), and names of both parents. Go to the county records office for the
given county and find the record for the person with the given name. If there are more than one such records, check for names of parents.
Identify the date of birth from this record. Subtract the year of birth from 1991. If the month and day of birth has not yet been reached in
the current year, subtract 1. DEF B. Give the subject a slip of paper with the following question. ENTER YOUR AGE IN YEARS AT
YOUR LAST BIRTHDAY ______. The number written in the blank is the person's age.
2. LEVELS OF MEASUREMENT
NOMINAL: classify objects into categories with no ordering, the special case of two categories is termed a
dichotomous variable. examples: Religion, State of Birth.
ORDINAL: classify objects into categories with an ordering (less than, equal to, or greater than) but not neccesarily
equal distances between levels. examples; high, medium, low ; small, medium, large; hardness scale for minerals.
INTERVAL: An ordered scale where distances between categories are meaningful. For example there is 10 years
difference in age between someone age 10 and age 20, or someone age 70 and age 80. Examples - anything measured
using the real number system is an interval scale, income in dollars, age in years, temperature in degrees Celsius.
RATIO: An interval scale with a "natural zero". A natural zero is the total absence of the attribute being measured. eg.
Kelvin temperature scale is a ratio scale (0 degrees Kelvin = absence of any molecular motion), while Farenheit and
Celsius scales are interval, but not ratio scales. A ratio scale is required to make statements like "X is twice as warm as
Y”. Otherwise, ratio and interval scales have similar properties.
Examples: Measures of income at each level of measurement.
NOMINAL-Middle income if income is between $30,000 and $50,000
Not middle if less than $30 or more than $50.
ORDINAL -
LOW is less than 30
MID if between 30 and 50
HIGH if greater than 50
RATIO -
Income in dollars from Line 17 of 1995 tax return.
INTERVAL but not RATIO scale - Income in dollars + $10,000. This contrived measure doesn't have a natural
zero, but it is an interval scale. Note how the shift changes interpretations like "twice or half as large".
16
PRR 844 Topic Outlines
Page 17
3. STATISTICS APPROPRIATE TO THE LEVEL OF MEASUREMENT
Descriptive
Inferential
NOMINAL
mode, frequency tables, percentages
Chi square
ORDINAL
median, percentile,
range, interquartile deviation
Rank order statistics
INTERVAL
mean
standard deviation
T-test, ANOVA, etc.
A dichotomous variable measured as 0 or 1 can be considered to be any of these scales and all of the above statistics can
be applied.
4. RELIABILITY AND VALIDITY OF MEASURES
RELIABILITY is the absence of random error in a measure, the degree to which a measure is repeatable - yields the
same answer each time we measure it. We assess reliability by test-retest, split half method for indexes & scales
(Cronbach's alpha), and alternative forms.
VALIDITY is the absence of systematic error (bias) in a measure, the degree to which we are measuring what we
purport to measure. Types of validity are content (or face), criterion-related and construct validity
ACCURACY of a measure typically implies the absence of both systematic and random error, i.e a measure that is both
reliable and valid is called accurate. Also note the distinction between precision (fineness of distinctions, number of
decimal places) vs accuracy (how close the measure is to the "true" value". The "true" measure is something we can
never know with certainty. Scientists therefore refer to reliability and validity of measures vs accuracy.
The differences between reliability and vailidity are illustrated by looking at the targets below. Think of taking 100
independent measures (shots at a target here) and displaying the results graphically. Note that validity and reliability are
two independent concepts. Reliability is indicated by the repeatability of the measure (a tight shot group) while validity
refers to whether the measures center on the true measure (or target).
17
PRR 844 Topic Outlines
Page 18
5. Assessing Reliability and Validity of a measure
a. Reliability
1. test-retest
2. alternative forms
3. split half/ Cronbach's alpha
b. Validity
1. content or face validity
2. criterion-related validity
a. concurrent
b. predictive
3. construct validity
6. Souces of measurement error
a. Sources of bias
1. Due to researcher or measurement instrument; expectations, training, deception
2. Due to subjects, e.g. Hawthorne effect, reactivity
3. Due to research design – biased samples, nonresponse
b. Sources of noise (random error)
1. Differences among people
2. Fuzzy definitions, criteria, procedures yielding inconsistent interpretations
3. Mixing processes
18
PRR 844 Topic Outlines
Page 19
QUESTIONNAIRE DESIGN
A. Kinds of Information
1. Demographics, socioeconomics and other personal attributes: age, race, education, gender, income,
family size/structure, location of residence, years living in the area, place where person was born,
grew up, personality, ...
2. Cognitive - what do they know, beliefs (what do they think is true or false)
3. Affective - attitudes, how do they feel about something, preferences, likes and dislikes, evaluations,...
4. Behavioral - what have they done in past, are doing in present, expect to do in future (behavioral
intentions).
B. Question structures
1. Open ended
2. Close ended with ordered choices, eg. Likert scale
3. Close ended with unordered choices
4. Partially close ended, ("other" option)
C. Question content
1. Is this question necessary? useful?
2. Are several questions needed on this subject? Double barrelled questions.
3. Do respondents have information to answer the question? Should a filter question be included. How
precise can subjects answer? Is question too demanding?
4. Does question need to be more concrete, specific and related to subject's personal experience? Is a time
referent provided?
5. Is question sufficiently general? Do you want recent behavior or "typical behavior"?
6. Do replies express general attitudes or specific ones?
7. Is content loaded or biased
8. Are subjects willing to answer?
9. Can responses be compared with existing information?
D. Question wording
1. Will words be uniformly understood? Simple language. Beware of technical phrases, jargon and
abbreviations.
2. Does question adequately express the alternatives?
3. Is the question misleading due to unstated assumption or unseen implications.
4. Is wording biased, emotional, or slanted?
5. Will wording be objectionable to respondents?
6. Use more or less personalized wording.
7. Ask in more direct or more indirect way?
E. Form of Response
1. Open or closed
2. If closed, ordered or unordered; number of categories, type of queue,forced or unforced choice
3. Be sure categories are mutually exclusive.
F. Sequencing of questions
1. Will this question influence responses to others?
2. Is question led up to in a natural way?
3. Placement to create interest, improve response rate.
4. Branching, Skipping, and transitions on questionnaires.
19
PRR 844 Topic Outlines
G. Most Common Problems in Questionnaire Design
Page 20
1. Lack of specificity.
a.. Not indicating a timeframe for questions about behavior (within past day, week, month, etc)
b. Asking for what people do in general or on average vs what they did on a particular trip or during a
particular week. The concern is usually that the last trip or a particular day or trip may not be a
"typical one". Proper sampling usually covers this potential problem. Remember that we are usually
not interested in a particular subject's response, but will likely report averages or percentages across
population subgroups. If we randomly sample trips, chances are that if for one person the trip we
choose happens to be longer than their usual trip, for others the random choices will yield the
opposite. For a large enough sample and proper random sampling no single observation is
representative, but collectively the sample will be.
c. Too much aggregation: to measure complex phenomona we usually must break them into component
parts. e.g. How many days of recreation participation, hours of leisure last week? Concepts too vague
and aggregate to yield reliable information. Ask for individual activities and then add up responses.
2. Not using filter/branch/contingency questions for questions that may not apply to all subjects. Don't assume
everyone is aware of everything or fulfills the requirements for answering a particular question.
a. Question doesn't apply : e.g. if not stay overnight don't ask number of nights or lodging type
b. Subject not informed adequately about a topic : e.g. don’t ask attitudes or opinions about things without
first assessing awareness.
3. Use of technical terms or complex language. Don't use technical terms like carrying capacity, sustained
yield, ADA, etc. on non-technical audiences. What you put in research proposal must be translated
and phrased for the population being surveyed.
4. Asking questions subjects likely cannot answer.
a. Recall of details from distant or even recent past. "How many fish did you catch in 1985? "How many
people were in the pool this morning?
b. Looking for deep motivations and attitudes on topics subjects really haven't thought about very much or
don't know much about. " To visitors - Should we use more volunteers or regular park staff at the
visitor center?"
5. Trying to handle complex matters in a single question vs breaking it down into smaller pieces. Often termed
"double barreled" questions.
6. Using questionnaires to gather data that is better collected via observation, physical instruments, or from
available records. e.g. What was the temperature or weather conditions today, in personal interview are you
male or female, for a park manager - how many visitors did you have last week (without opportunity to look it
up)..
7. Implied alternatives: State them. e.g. Do you feel Lansing should charge $2 for entrance to the zoo OR
should the zoo be free of charge.
8. Unstated assumptions: How far did you drive to reach the park? Assumes subject arrived by car.
9. Frame of reference: particularly important when asking for attitudes or evaluations.
e.g. How would you evaluate the performance of the Lansing Parks Department?
- as a user, taxpayer, parent,
- Am I satified?, Do I think most people are satisfied, Are they doing best job possible
with resources they have
20
PRR 844 Topic Outlines
Page 21
Sample Question Formats
Kinds of variables typically measured in surveys
1. Demographic and socioeconomic characteristics
2. Behavior
3. Attitudes, Interests,Opinions, Preferences, Perceptions
a. Cognitive : Beliefs and knowledge
b. Affective : Feelings, emotional responses
c. Behavioral intentions
1. Simple fill in the blank. Obtaining a straightforward number or other easily understood response.
How old are you? ___________.
years
In what county is your permanent residence? __________
county
How much money did you spend on this trip?
$ __________________.
2. Open ended: To avoid leading subject, to obtain wide range of responses in subject’s own words, or when you
don’t know kinds of responses to expect.
What is your primary reason for visiting the park today? _______________________________________.
3. Partially closed ended. List major response categories while leaving room for others. If you have exhaustive
set of categories question can be completely “closed-ended”. Usually check a single reponse, but can also allow
multiple responses (see checklist below).
Which of the following community recreation facilities do you most frequently use? (check one).

neighborhood parks/playgrounds

swimming pools

community centers

natural areas

tennis courts

other (please specify) ___________________
4. Checklists: Allow subjects to check multiple responses. Categories exhaustive & mutually exclusive
Which of the following winter recreation activities have you participated in during the past month?(check all that
apply)

Cross-country skiing

Downhill skiing

Snowmobiling

Ice Skating

Sledding or Toboganning
5. Likert Scales: Versatile format for measuring attitudes. Can replace “agree” with “importance” “satisfaction”,
“interest” “preference” and other descriptors to fit the attitude you wish to measure.
Please check the box that best represents your level of agreement or disagreement with each of the following
statements about downhill skiing:
Downhill skiing is exciting
Downhill skiing is dangerous
Downhill skiing is expensive
Strongly agree



Agree



21
Neutral



Disgaree Strongly disagree






PRR 844 Topic Outlines
Page 22
6. Rank Ordering: To measure preferences/priorities . Limit to short lists.
Rank the following states in terms of your interest as possible travel destinations for a summer vacation trip.
Place a 1 beside the state you would most like to visit, place a 2 besides your second choice, and a 3 beside your
third choice.
______ Michigan
______ Wisconsin
______ Minnesota
7. Filter Question. To screen for eligibility or knowledge prior to asking other questions. Make sure each
question applies to all subjects or use filters and skips to direct respondents around questions that don’t apply.
Did you stay overnight on your most recent trip?
 NO
 YES
 IF YES, How many nights did you spend
away from home? ______
8. Semantic Differential scale. Measure perception or image of something using a set of polar adjectives.
For each of the characteristics listed below, mark an X on the line where you feel downhill skiing falls with respect
to that characteristic. (Could repeat with cross country ski, snomobiling and compare perceptions) (or Coke and
Pepsi).
exciting _____
______ ______
______
______
______
dull
expensive _____
______ _____
______
______
______
inexpenive
safe
_____
______ _____
______
______
______
dangerous
SOME EXAMPLES OF BAD QUESTIONS (THINGS TO AVOID).
1. Loaded questions
Most people are switching to brand X. Have you switched yet?
Do you agree with (famous person or authority) that our parks are in terrible shape?
How satisfied were you with our service: Very satisfied Quite Satisfied Satisfied
2. Double-barrelled (two questions in one)
Do you use the city parks or recreation centers?
Should Lansing build a new baseball stadium by raising taxes?
3. Not specific enough
Do you downhill ski? - Have you downhill skiied within the past 12 months.
Have you made a trip within the past three months? - Have you taken any overnight trips of 50 miles or
more (one-way)...
How much did you spend on this trip?- Please estimate the amount of money spent by you or any
member of your travel
party within 30 miles of this park.
How many hours of leisure did you have last week. (How define leisure?)
4. Subject able to answer?.
Do you think the American’s with Disabilities Act (ADA) has been effective?
How would you rate the job done by our new Parks & Recreation Director.
How much time do your teenagers spend on homework.
How many fish did you catch last year, by species, location, month, ....
How many trips would you make to MI State Parks if entrance fees were eliminated?
What is zipcode of your travel destination?
5. Sensitive questions
Have you committed any crimes within the past week? Yes No.
When did you stop beating your wife?
REFERENCES:
Sudman, S. and Bradburn, N.M. (1982). Asking Questions: A Practical Guide to Questionnaire Design. San
Francisco: Jossey-Bass.
Hogarth, R.M. (ed). (1982). Question framing and response consistency. San Francisco: Jossey-Bass.
22
PRR 844 Topic Outlines
Page 23
Indices and Scales
Indices and scales are defined and used in a variety of ways. Generally they are ordinal measures of some construct that
are based on a composite of items. They are most often used to measure a construct or latent (vs manifest) variable that
is not directly observable.
Babbie distinguishes an index from a scale by
index = simple sum of items
scale = based on patterns of response
Devillis distinguishes them based on causal relationships
index= cause indicator - the items determine level of the construct, e.g. SES
scale = effect indicator - all items result from a common cause, e.g. alienation
Indicators are often measured via an index ( e.g. economic index of leading indicators, quality of
life indicators, ...)
Indices and scales may be uni-dimensional (one underlying construct) or multi-dimensional (two
or more distinct facets or dimensions of the construct , e.g (verbal, quantitative &
analytical parts of GRE score.)
Off-the-shelf index/scale vs “home-built” : There are thousands of scales/indices measuring almost any concept
you are interested in. (See Miller or Bearden et. al. for numerous examples). Existing scales have the
advantage of having their reliability & validity tested (most of them anyway). They may however be
too long, or not exactly what you need. In that case you develop your own scale or index or modify an
existing one, but then you must do your own assessment of validity & reliability.
STEPS FOR DEVELOPING AN INDEX
1. Define the concept /construct-- theory, specificity, what to include in the measure
2. Identify items that measure it - Generate an item pool to choose from. Items should reflect the scale’s purpose:
Can include some redundancy of items
Tips on Good & bad items
avoid lengthy items
aim at reading level of audience/population
avoid multiple negatives, avoid double barrelled items
watch for ambiguous pronouns, misplaced modifiers, adjective forms
reversing polarity of some items avoids agreement bias but may confuse
Determine format for measures
Thurstone - items to represent each level of the attribute
Guttman - progressively more difficult to pass
Likert, Standard formats/issues - number of cues, balanced?, weights
Decide whether a unidimensional or multidimensional scale is appropriate
3. Evaluate face validity of items
Have the item pool reviewed by experts - face validity & item clarity/wording
Consider including some validation items in the scale (predictive or criterion-related)
4. Administer scale/index to a test sample and evaluate the scale
a. Item analysis: means, variances, correl, coefficient alpha,
b. Check bivariate relationships between items - correlations
c. Check multivariate relationships
d. split samples to check stability across samples
5. Finalize the scale - assign scores to items, optimize scale length- alpha depends on co-variaton among
items and number of items, handle missing data if necessary
23
the
PRR 844 Topic Outlines
INDEX EVALUATION - RELIABILITY AND VALIDITY OF THE SCALE
Page 24
1. Reliability (SPSS Reliability analysis computes alpha and inter-item correlations): Internal consistency
homogeneity, item analysis
or
a. Coefficient alpha measures the proportion of total variance in a set of items that is due to the
latent variable (in common across the scale items).
 = [k/(k-1)] * [1 - (I 2 / y 2)],
where k= number of items in scale, I 2 =sum of variances of each item, and y
2
= total variance of scale
b. Spearman Brown formula -- reliability = k r / [ 1+(k-1) r], where r = the average
inter-item correlation and k = number of items.
Note that longer scales automatically are more reliable. example.1: suppose r for a set of items is
.5. Then a scale with 3 items has = .75. For n =10 and r=.5, = .91. Example 2: Suppose r
=.25, then for n=3, = .5 and for n=10 =.77.
Good scale if reliability above .7, .8 better , if above .9 consider shortening the scale.
c. Other types of reliability
Alternative forms reliability
Temporal stability - test-retest reliablity
Generalizability Theory
2. Validity
Content or Face Validity
Criterion-related, validity
Construct validity
Multi-trait, multi-method validition - MTMM
3. Factor Analysis is a multivariate procedure often used in developing scales. It can assess how many latent variables
(dimensions) underlie a set of items, condensing info, & helping define the content or meaning of each dimension
SELECTED REFERENCES
Ajzen, I. & Fishbein, M. 1980. Understanding attitudes and predicting social behavior. Englewood Cliffs, NJ: PrenticeHall.
Bearden, W.O., Netemeyer, R.G., & Mobley, M.F. 1993. Handbook of Marketing Scales. Newbury Park, CA.: Sage.
- whole book of marketing-related scales
Devellis, R.F. 1991. Scale Development: Theory & Applications. Newbury Park, CA.: Sage.
Eagly, A.H. & Chaiken, S. 1993. The Psychology of Attitudes. Fort Worth, TX: Harcourt Brace Jovanovich College
Publishers.
Henerson, M., Morris, L.L. and Fitz-Gibbon, C.T. 1987. How to Measure Attitudes. Newbury Park, CA.: Sage.
Miller, D.C. 1991. Handbook of Research Design & Social Measurement 5th ed. Newbury Park, CA.: Sage. Chapter 6 is guidebook to social scales
Nunnally, J.C. 1978. Psychometric Theory. 2nd edition. New York: McGraw Hill.
Robinson, J.P., Shaver, P.R, & Wrightsman, L.S. 1991. Criteria for scale selection and evaluation. in Measures of
Personality and Social Psychological Attitudes. Ann Arbor: Univ. of Michigan Survey Reseasrch Center Institute
for Social Research.
24
PRR 844 Topic Outlines
Page 25
TOPIC 6. SAMPLING
1. Census vs sample
2. Steps in sampling process
a. Define study population
b. Specify sampling frame
c. Specify sampling unit
d. Specify sampling method
e. Determine sample size
f. Specify sampling plan
g. Choose sample
2. Study population: Define population in terms of element, sampling unit, extent and time.
Element
Adults 12 years of age and older
Sampling unit
In vehicles
Extent
Entering Yogi Bear Park
Time
Between July 1 and August 31, 1993
3. Sampling frame : a perfect one lists every element of the population once
4. Sampling unit is element or set of elements considered for selection at some stage in sampling
5. Sampling method
Probability - each element has a known chance of selection, can estimate sampling error when probability
sampling is used.
Non-probability - don't know probabilities & can't estimate sampling errors. Examples: Judgement,
convenience, quota, purposive, snowball
Probability sampling methods:
Simple random sample (SRS),
Systematic sample.
Stratified vs Cluster sample
Proportionate vs disproportionate
Single vs multistage
6. Stratified samples : Stratify to a) ensure enough samples in designated population subgroups and to increase
efficiency of sample by taking advantage of smaller subgroup variances. In stratified sample you divide population
into subgroups (strata) and sample some elements from each subgroup. Subgroups should be formed to be
homogeneous - people in same group are similar to each other on variables you intend to measure and people in
different groups are different on these variables.
7. Cluster or Area Sample : Cluster to reduce costs of gathering data. Cluster samples are less efficient than simple
random samples in terms of the sampling error for a given sample size.
In cluster sampling, you divide
population into subgroups (clusters) and sample elements only from some of the clusters. When cluster sampling,
form heterogeneous subgroups so that you will not miss any particular type of person/element because you didn't
select a particular cluster. Generally groups are formed based on geographic considerations in cluster sampling and
it is therefore also called area sampling.
8. Disproportionate sampling. Note that all elements need NOT have an EQUAL chance of selection to be a probability
sample, only a KNOWN probability. However, to have a representative sample each population subgroup should
be represented proportionately to their occurence in the population. The best way to assure this is for each element
to have equal chance of selection. If sampling is disproportionate, you must weight the resulting sample to adjust
for different probabilities of selection. See attached example.
9. Determine sample size: Sample size is based on a) budget, b) desired accuracy (confidence level/interval), and c)
amount of variance in population on variable(s) being measured. It is also affected by whether or not you wish to
do subgroup analyses. Tables provide the sampling errors associated with different sample sizes for sampling from
a binomial distribution. More on this later.
25
PRR 844 Topic Outlines
Page 26
TOPIC 6A. Sampling populations of visitors, trips and nights - Weighting
Recreation and travel studies often sample populations of visitors using household and on-site sampling designs. Care
must be exercised in such studies to avoid a number of potential biases in the sample that can result from the unequal
sampling probabilities caused by variations in the length of stay or frequency of visitation across population subgroups.
Length of stay bias is a common problem in on-site studies. For example, if motel visitors are sampled by randomly
choosing occupied rooms each night, visitors staying short periods will be less likely to be chosen than visitors staying
for a longer time. Someone staying two nights would have twice the chance of being chosen as someone staying only
one night. This bias in the sample can be corrected by weighting cases in inverse proportion to their probability of being
chosen. For example, two night stay visitors would be weighted 1/2 that of single night visitors to adjust for the unequal
selection probabilities.
The existence of such sampling biases depends on how the population has been defined and the sampling frame chosen.
Three common populations may be defined for these kinds of recreation and travel surveys:
(1) Population of visitors = any individual visiting a site for one or more days during a given time period.
(2) Population of trips or visits = defined as either person or party visits, the population in this case is the visit or
trip, which may consist of several days/nights.
(3) population of nights (person or party) = an individual or party staying one night. Someone staying 3 nights on a
trip would be treated as three distinct members of the population of nights.
Biases enter when a sampling frame appropriate for one definition of the population is used to generate a sample for a
different population definition. For example, if the population of interest is visitors (definition 1) and we sample nights
or trips, repeat and longer stay visitors will be overrepresented in the sample. Similar biases can result when household
surveys are used to generate a sample of trips or nights. For example, travel surveys that ask households to report their
most recent trip and then analyze trip characteristics, will overrepresent the kinds of trips that less frequent travelers
take.
If the researcher is aware of the problem , it can usually be corrected easily by asking length of stay and frequency of
visit questions and then applying appropriate weights based on these measures. A simple example helps illustrate.
Note that in the example we KNOW the full population(s) and therefore any bias in the sample is readily evident by
simply comparing a sample estimate with the population value.
General rule for weighting cases is
1. Identify unequal probabilities of selection for population subgroups.
2. Weight cases in inverse proportion to their selection probabilities. e.g. if one type of visitor is twice as likely to be
chosen as another, assign the former a weight of 1/2 relative to the latter.
3. It is sometimes also desirable to normalize weights so that the weighted sample is the same size as the original one.
Do this by making the weights sum to one.
EXAMPLE: Population of 1,000 Visitors divided equally between two types (say non-resident=Type A and
resident=Type B). Assume TYPE A visitors take 1 trip per year, averaging 4 nights per trip
. TYPE B visitors
take 4 trips per year averaging 2 nights per trip. We can completely specify the population of visitors, trips and nights as
follows:
TYPE A
TYPE B
TOTAL
Popl n of Vi s i t or s
N
Pc t
500
50%
500
50%
1000
Popl n of Tr i
N
500
2000
2500
Suppose we want to Draw sample to estimate the following:
1. Percent of population who are Type A
2. Percent of Trips by Type A visitors
3. Percent of Nights by Type A visitors
26
ps
Pc t
20%
80%
Popl n of Ni ght s
N
Pc t
2000
33%
4000
67%
6000
PRR 844 Topic Outlines
Page 27
OPTION 1. CHOOSE RANDOM SAMPLE OF 200 PEOPLE USING HOUSEHOLD SURVEY, respondents report
their last trip. This gives us a sample of 200 visitors and 200 trips.
If random sampling works reasonably, we will get a sample of roughly 100 TYPE A's and 100 Type B's, with one trip
reported for each visitor. Note that the sampling frame is for the population of visitors. Be careful if you start doing
analyses with trips (or nights) as the unit of analysis. You won’t necessarily have a representative sample of trips or
nights. Let’s look:
From this sample, the percent of trips by TYPE A = 1/2, a biased (wrong) answer, since we know in the population of
trips that only 20% are by Type A’s. Problem is we do not have a representative sample of trips - a given trip by a Type
A is four times more likely to be chosen as a given trip of a Type B person.
We can choose trips, by first choosing person (these are equal), but gathering one trip per visitor introduces a bias since
it gives different kinds of trips unequal probabilities of being sampled. Correct by weighting the sample to adjust for
unequal probability of selection. .
Weight cases inversely proportional to their probability of selection.
Prob of choosing given trip = prob of choosing person x prob trip is selected given person
Type A trip prob = 100/500 * 1/1
Type B trip prob = 100/500 * 1/4
= 1/5
= 1/20
Weight inversely proportional to these probabilities, e.g. since probability ratio is 4:1 make weights 1:4. This reflects
fact that type A trips are four times more likely to be chosen.
Type A Weight = 1
Type B Weight =4
NOW WEIGHT CASES IN SAMPLE IN ORDER TO GET A REPRESENTATIVE SAMPLE OF TRIPS
TYPE A
TYPE B
TOTAL
SAMPLE TRIPS X
100
X
100
X
200
Weight = Corrected sample
1
= 100
4
= 400
500
Pct of trips
20%
80%
Note the percents are the correct values as observed in population above. If we had not weighted, we would irroniously
estimate half of trips are Type A. Normalized weights would be 1/5 and 4/5 - two weights in proprotions of 1:4 that add
to one.
OPTION 2. RANDOM SAMPLE OF NIGHTS. Suppose instead of sampling households (visitors) we take our sample
on-site by randomly selecting every 10th occupied site/room in motels or campgrounds. This sampling frame is for the
population of nights. Unweighted estimates of Type A nights would be correct, but estimates of Type A visitors or trips
would be biased. Use same procedures as above to calculate appropriate weights. Check your result to see if you obtain
the correct percentages - in this case
you know these from the population figures above.
TOPIC 7A. SURVEY METHODS
1. Survey research attempts to measure things as they are. The researcher wishes to measure phenomona without
intruding upon or changing the things being measured. This is in contrast with experiments, which intentionally
manipulate at least one variable (the treatment) to study its effects on another (dependent variable or effect).
2. Survey research generally means gathering data from people by asking questions - self administered, phone, or in
person personal interviews (Interview Surveys). See Babbie, Trochim or any methods text for strengths and
weaknesses of these three approaches. Key is choosing the most appropriate approach in a given situation. This will
depend on a) budget, b) turnaround time, c) content and number of questions, and d) population being measured.
3. Survey designs best suited to describe a population- demographic & socioeconomic characteristics, knowledge &
beliefs, attitudes, feelings & preferences, and behaviors and behavioral intentions.
27
PRR 844 Topic Outlines
Page 28
4. Types of survey designs
a. Cross sectional : study phenomonon at a single point in time (snapshot of the study population).
b. Longitudinal : study phenomona at more than one point in time.
i.. Trend study: Measure same general population at two or more times.
ii. Cohort Study : Measure same specific population at two or more times
iii. Panel study: Measure same individuals (same sample) at two or more times.
c. Approximating longitudinal study with cross sectional designs.
i. replication of previous studies.
ii. using recall or expectation questions
iii. using cohorts as surrogates for a temporal dimension
5. Key Survey Research Issues
a. Study population: identifiable, reachable, literacy, language
b. Sampling, obtaining representative samples, sample size, sampling unit
c. Choice of mail, interview, phone; household, on-site, internet, etc.
d. Questionnaire development & testing
e. Follow-up procedures & non-response bias e. Interviewer/instrument effects and other biases
f. Field procedures, interviewer training, costs, facilities, time, personnel
g. Data handling & analysis
6. Steps in Survey
1. Assess Feasibility
2. Set objectives, identify variables & population
3. Choose approach/design
4. Develop instruments (write questions, format, sequencing, cover letters…)
5. Pretest instruments and procedures, train interviewers
6. Sampling plan and sample size
7. Gather data/execute survey
8. Data entry, processing and analysis
9. Reporting
28
PRR 844 Topic Outlines
Page 29
TOPIC 7B. EXPERIMENTAL DESIGN
A. Introduction
Experiments are designed to test for relationships between one or more dependent variables and one or more
independent variables. They are uniquely qualified to determine cause-effect relationships.
Consider experimental or quasi-experimental approaches whenever you are interested in studying the effects of
one variable on another as contrasted with studying the relationship between the two variables.
Experiments are distinguished from surveys in that the researcher consciously manipulates the situation in order to study
a cause-effect relationship. Survey procedures measure things as they are and try to avoid any researcher caused changes
in the objects under study.
FOR EXAMPLE : A pollster asking who people intend to vote for in the next election is conducting a survey -trying to measure what the result woud be if the election were held today. His questioning is designed to
measure voter preference without influencing the choice in any way. This research could become an
experiment if the study is investigating how pre-election polling influences voting patterns. In this situation, the
act of polling is an intervention or treatment, consciously introduced for the purpose of manipulating the
situation and then measuring the effects on voters.
Typical examples of experiments in recreation & tourism:
1. Effects of an information or promotion programs on knowledge, attitudes, or behavior. eg. Is a promotional
program successful in increasing brand awareness, image, or sales? Which media, messages, etc. are more
effective with which subgroups of potential customers? What is the effect of interpretation, environmental
education, or outdoor adventure programs on environmental knowledge, attitudes, or behaviors? Does an outdoor
adventure program increase a person's self-concept, improve family bonding, increase social cohesion, etc.
2. Consumer reaction to price changes. Bamford, Manning et. al. (JLR 20,4, 324-342) a good example. Response of
consumers to product changes or location/distributional approaches. With promotion above, this covers 4 P's of
marketing.
3. Effectiveness of various TR interventions.
4. Impacts of tourism on community attitudes; social, economic, and environmental impacts.
5. Positive and negative consequences of recreation and tourism - physical health, mental health, family bonding,
economic impacts, learning, etc. Consequences for individuals, families, social groups, communities, societies,
global consequences.
6. Experiments have been used in studying people's preferences for landscapes and more generally to measure the
relative importance of different product attributes in consumer choices. e.g. conjoint analysis (see Tull &
Hawkins pp. 359-370.
B. There are many alternative experimental designs, but the basic principles behind an experiment can be illustreated
with the Classic Pre-test, post-test with control (also called before after with control).
This experimental design involves two groups (experimental and control) and measurements before and after the
"treatment". It can be diagrammed as follows:
Before
R
MB1
R
MB2
After
X
MA1
Experimental group
MA2
Control Group
R denotes random assignment to experimental and control groups
X denotes administration of the treatment
Other entries denote measures made before (B) or after (A) the administration of the treatment.
Measure of the effect = (MA1-MB1) - (MA2-MB2)
with
without ;
NOT simply after (MA1) - before (MB1)
29
PRR 844 Topic Outlines
Page 30
Note the effect is the change in experimental group adjusted for any change in the control group. Without the adjustment
for the control group, you have a before-after measure rather than a with vs without measure. The experimental group
change is the change with the treatment, while control group change is the change without the treatment. The difference
in the two is a "with minus without" measure. This design controls for all major internal validity errors except
interaction (see below).
In program evaluation, the program is the treatment or "cause" and program outcomes or impacts are the "effects" to be
measured. Critical to understanding experiments is understanding of the role of the control group and the potential
sources of error that experiments control for (covered in D below).
C. Characteristics of a true experiment
1. Sample equivalent experimental and control groups
2. Isolate and control the treatment
3. Measure the effect
In the classic Pre-test -Post-test with control, design above, note how these characteristics are met:
1. random assignment to groups to assure "equivalence"
2. treatment administered to experimental group and withheld from control group
3. Effect is measured by change in experimental group - change in control group.
Quasi-experimental designs fail on one or more of these characteristics:
A simple before-after design lacks a control group.
Ex post facto designs form groups "after the fact". These usually do not involve "equivalent" groups
and may not isolate or control the intended "treatment". People who volunteer for a program
may be different than those who do not in many respects.
D. Sources of Experimental Error - these are the errors that experiments tryto control for in evaluating a causeeffect relationship.
Internal Validity: errors internal to a specific experiment, designs control for these
*1. Premeasurement (Testing) : effect of pre-measurement on dependent variable (post-test)
*2. Selection: nonequivalent experimental & control groups, (statistical regression a special case)
*3. History: impact of any other events between pre- and post measures on dependent variable
*4. Interaction: alteration of the “effect” due to interaction between treatment & pre-test.
5. Maturation: aging of subjects or measurement procedures
6. Instrumentation: changes in instruments between pre and post.
7. Mortality: loss of some subjects
External validity: errors in generalizing beyond the specific experiment
8. Reactive error - Hawthorne effect - artificiality opf experimental situation
9. Measurement timing - measure dependent variable at wrong time
10. Surrogate situation: using popln, treatment or situation different from “real” one.
E. OTHER EXPERIMENTAL DESIGNS
1. After Only with Control : This design is used widely in evaluating changes in knowledge or attitudes due to an
information, education, or advertising program. It is used instead of pre-test post-test with control due to the likelihood
of interaction affects when testing knowledge or attitudes. It omits the pre-test and relies on large samples or careful
assignment to groups to achive “equivalent” experimental and control groups. No pre-measure avoids interaction effects.
Sacrifice is weaker control over selection errors.
R
R
X
MA1
MA2
30
PRR 844 Topic Outlines
2. Other Designs
Quasi Experimental
a. Ex Post Facto
Page 31
b..After Only (no control)
Experimental
a. Simulated before-after with control
R
R
c. Before-After
(no control)
MB
X
MA
controls for pre-meas/interaction; an alternative to after-only with control - simply varies timing of
measurement on control group (possible history & selection errors)- this design avoids contaminating control
group with the treatment, say in evaluating a large mass media campaign).
b. Solomon 4 group - controls for everything (internal) , but expensive, two control & two experimental groups.
3.Advanced designs
Random Block Designs (RBD) - one major intervening variable, block (stratify) on this variable.
Latin Squares- two non-interacting extraneous variables
Factorial Design - effects of two or more indep. variables (multiple treatments), including interactions
F. LABORATORY vs Field (Natural) experiments, tradeoff between external and internal validity
LAB - high internal validity, but may not generalize to real worl settings
FIELD - high external validity, but may be problems obtaining adequate controls (internal validity).
G. Steps in Conducting an Experiment
1. Identify the relevant study population, the primary "treatment(s)" or independent variable (s) and the measures of the
effect or dependent variables.
2. Choose a suitable design based on an analysis of the potential threats to validity in this situation.
3. Form groups and assign subjects to groups - remember you want to form "equivalent groups".
Random assignment
Matching characteristics
4. Develop measurement procedures/instruments, measures of the effect
5. Run the experiment - make pre-measurements, administer treatment to experimental group, make post-measurements
6. Compute measure of the effect
G. Quasi- Experimental Designs - Examples (Find potential flaws here):
1. Travel Bureau compares travel inquiries in 1991 and 1994 to evaluate 1992 promotion efforts.
2. To assess effectiveness of an interpretive exhibit, visitors leaving park are asked if they saw exhibit or not, Two
groups are compared relative to knowledge, attitudes etc.
TOPIC 7C. Secondary data designs
Secondary data analysis : use of data gathered by someone else for a different purpose – reanalysis of existing data.
See methods links page for links to secondary sources of data about recreation & tourism
Sources:
Government agencies: e.g. Population, housing & economic censuses, tax collections, traffic counts, employment,
environmental quality measures, park use, …
Internal records of your organization – sales, customers, employees, budgets…
Private sector - industry associations often have data on size and characteristics of industry
Previous surveys – as printed reports or raw data, survey research firms sell data
Library & Electronic sources – the WWW, on-line & CD-ROM literature searches, …
Previously published research – reports have data in summary form, original data often available from author.
Issues in using secondary data.
1) data availability – know what is available & where to find it
31
PRR 844 Topic Outlines
Page 32
2) relevance – data must be relevant to your problem & situation
3) accuracy – need to understand accuracy & meaning of the data
4) sufficiency – often must supplement secondary data with primary data or judgement to completely address
the problem
Since you did not collect the secondary data it is imperative that you fully understand the meaning and accuracy of the
data before you can intelligently use it. This usually requires you to know how it was collected and by whom. Find
complete documentation of the data or ask about details from source of data. At Gov. Info site choose INFO links to see
documentation of each data source.
Examples: using secondary data to estimate
Trends – Compare surveys in different years or plot time series data - many tables in Spotts Travel &
Tourism Statistical Abstract, Michigan county tourism profiles, economic time series at BLS site, REIS
data at Gov Info Clearing House.
b) Spatial variations - gather data across spatial units, map the result
c) Recreation participation – apply rates from national, state and local surveys to local population data from
Census, rates at NSGA, ARC web pages (Roper-Starch study), NSRE 1994-95 survey
d) Tourism spending – Stynes estimates tourism spending by county in Michigan using a variety of secondary
data and some judgement – lodging room use taxes, motel, campground and seasonal home inventories,
occupancy rates by region, average spending by segment and statewide travel counts. See my economic
impact web site. Also see Leones paper on measuring tourism activity.
a)
TOPIC 7D. OBSERVATION & OTHER METHODS
Unobtrusive Measurement – content analysis, physical instruments, and observation (See Webb et. al. Unobtrusive
Measures book.
Observation : gathering data using human observers. There are a range of observational methods from highly structured
quantitative counts to qualitative participant observation. Babbie distinguishes a continuum of approaches:
complete participant -- participant as observer -- observer as participant-- complete observer
Quantitative versions of observation employ probability sampling techniques (time or event sampling) and
quantitative measures of the behaviors being observed (generally by means of a highly structured observation
form). Studies may employ multiple observers and evaluate the reliability of observations by comparing
independent observers. E.g. measuring use of a park or trail by counting visitors and observing their
characteristics during randomly selected periods.
Qualitative forms of observation are more interpretive with the observer making field notes and interpreting what
they observed.
Content analysis is a special set of technique for analyzing documentary evidence (letters, articles, books, legal
opinions, brochures, comments, TV programs, …). It is used quite widely in communication and media studies to study
media and messages in human communication. (See Babbie Chapter 12; A Sage monograph by Weber, Robert Philip.
1990. Basic content analysis. Newbury Park, CA: Sage and a book by Ole Holsti . 1969. Content analysis for the social
sciences and humanities. Reading,Mass: Addison-Wesley Publ, Co. are good references
Content analysis is the application of scientific methods to documentary evidence. Makes inferences via objective and
systematic identification of specified characteristics of messages.
REFS: Holsti, Ole R. 1969. Content analysis for the social sciences and humanities. Reading Mass: Addison Wesley.
Also see Babbie's Chapter 12.
Applications in recreation to coding open ended survey questions, analyzing public involvement, analysis of societal (or
individual) values, opinions and attitudes as reflected in written documents, etc.
32
PRR 844 Topic Outlines
Page 33
TOPIC 8. Data Gathering, Field Procedures and Data Entry (incomplete)
There are many tedious but important procedures involved in gathering data. These should be clearly thought
out in advance, tested, and included in a study proposal.
For surveys: mailing options, follow-ups, interviewer training,..
Data entry options: CATI systems,
Coding and cleaning
33
PRR 844 Topic Outlines
Page 34
TOPIC 9. DATA ANALYSIS AND STATISTICS
1. Functions of statistics
a. description: summarize a set of data
b. inference: make generalizations from sample to population. parameter estimates, hypothesis tests.
2. Types of statistics
i. descriptive statistics: describe a set of data
a. central tendency: mean, median (order statistics), mode.
b. dispersion: range, variance & standard deviation,
c. Others: shape -skewness, kutosis.
d. EDA procedures (exploratory data analysis).
Stem & leaf display: ordered array, freq distrib. & histogram all in one.
Box and Whisker plot: Five number summary-minimum,Q1, median, Q3, and maximum.
Resistant statistics: trimmed and winsorized means,midhinge, interquartile deviation.
ii. inferential statistics: make inferences from samples to populations.
iii. Parameteric vs non-parametric statistics
a. parametric : generally assume interval scale measurements and normally distributed variables.
b. nonparametric (distribution free statistics) : generally weaker assumptions: ordinal or nominal
measurements, don't specify the exact form of distribution.
3.Steps in hypothesis testing.
1. Make assumptions & choose the appropriate statistic. Check measurement scale of variables.
2. State null hypothesis; and the alternative
3. Select a confidence level for the test. Determine the critical region - values of the statistic for which you will
reject the null hypothesis.
4. Calculate the statistic.
5. Reject or fail to reject null hypothesis.
6. Interpret results.
Type I error: rejecting null hypothesis when it is true.
Prob of Type I error is 1-confidence level.
Type II error: failing to reject null hypothesis when it is false.
Power of a test = 1-prob of a type II error.
4. Null hypotheses for simple bivariate tests.
a. Pearson Correlation
rxy =0.
b. T-Test
x =y
c. One Way ANOVA
M1=M2=M3=...=Mn
d. Chi square : No relationship between x and y. Formally, this is captured by the "expected table", which
assumes cells in the X-Y table can be generated completely from row and column totals.
5. EXAMPLES OF T-TEST AND CHI SQUARE
(1) T-TEST. Tests for differences in means (or percentages) across two subgroups. Null hypothesis is mean of Group 1
= mean of group 2. This test assumes interval scale measure of dependent variable (the one you compute means for) and
that the distribution in the population is normal. The generalization to more than two groups is called a one way analysis
of variance and the null hypothesis is that all the subgroup means are identical. These are parametric statistics since they
assume interval scale and normality.
(2) Chi square is a nonparametric statistic to test if there is a relationship in a contingency table, i.e. Is the row variable
related to the column variable? Is there any discernible pattern in the table? Can we predict the column variable Y if we
know the row variable X?
The Chi square statistic is calculated by comparing the observed table from the sample, with an "expected" table derived
under the null hypothesis of no relationship. If Fo denotes a cell in the observed table and Fe a corresponding cell in
expected table, then
2
Chi square (  ) =
2
 (Fo -Fe) /Fe
34
PRR 844 Topic Outlines
Page 35
cells
The cells in the expected table are computed from the row (nr ) and column (nc ) totals for the sample as follows:
Fe =nr nc / n
.
CHI SQUARE TEST EXAMPLE: Suppose a sample (n=100) from student population yields the following observed
table of frequencies:
Male
GENDER
Female
Total
20
30
50
40
10
50
60
40
100
IM-USE
Yes
No
Total
EXPECTED TABLE UNDER NULL HYPOTHESIS (NO RELATIONSHIP)
Male
GENDER
Female
Total
30
20
50
30
20
50
60
40
100
IM-USE
Yes
No
Total
2
2
2
2
2
 = (20-30) /30 + (40-30) /30 + (30-20) /20 + (10-20) /20
100/30 + 100/30 + 100/20 +100/20 = 13.67
Chi square tables report the probability of getting a Chi square value
this high for a particular random sample, given that there is no
relationship in the population. If doing the test by hand, you would
look up the probability in a table. There are different Chi square tables
depending on the number of cells in the table. Determine the number of
degrees of freedom for the table as (rows-1) X (columns -1). In this
case it is (2-1)*(2-1)=1. The probability of obtaining a Chi square of
13.67 given no relationship is less than .001. (The last entry in my table gives 10.83 as the chi square value
corresponding to a probability of .001, so 13.67 would have a smaller probability).
If using a computer package, it will normally report both the Chi square and the probability or significance level
corresponding to this value. In testing your null hypothesis, REJECT if the reported probability is less than .05 (or
whatever confidence level you have chosen). FAIL TO REJECT if the probability is greater than .05.
For the above example : REVIEW OF STEPS IN HYPOTHESIS TESTING:
(1) Nominal level variables, so we used Chi square.
(2) State null hypothesis. No relationship between gender and IM-USE
2
(3) Choose confidence level. 95%, so alpha = .05, critical region is  > 3.84 (see App E -p. A-28).
2
(4) Draw sample and calculate the statistic;  = 13.67
(5). 13.67 > 3.84, so inside critical region, REJECT null hypothesis. Alternatively, SIG= .001 on computer printout,
.001<.05 so REJECT null hypothesis.
Note we could have rejected null hypothesis at .001 level here.
WHAT HAVE WE DONE? We have used probability theory to determine the likelihood of obtaining a contingency
table with a Chi square of 13.67 or greater given that there is no relationship between gender and IMUSE. If there is no
35
PRR 844 Topic Outlines
Page 36
relationship (null hypothesis is true), obtaining a table that deviates as much as the observed table does from the
expected table would be very rare - a chance of less than one in 1000. We therefore assume we didn't happen to get this
rare sample, but instead our null hypothesis must be false. Thus we conclude there is a relationship between gender and
IMUSE.
The test doesn't tell us what the relationship is, but we can inspect the observed table to find out. Calculate row or
column percents and inspect these. Percents below are row percents obtained by dividing each entry on a row by the row
total.
Row percents:
Male
GENDER
Female
Total
.33
.75
.50
.67
.25
.50
1.00
1.00
1.00
IM-USE
Yes
No
Total
To find the "pattern" in table, compare row percents for each row with the "Totals" at bottom. Thus, half of sample are
men, whereas only a third of IMusers are male and three quarters of nonusers are male. Conclusion - men are less likely
to use IM.
-------------------------------------------------------------Column Percents: Divide entries in each column by column total.
GENDER
Male
Female
Total
IM-USE
Yes
.40
.80
.60
No
.60
.20
.40
Total
1.00
1.00
1.00
PATTERN: 40% of males use IM, compared to 80% of women. Conclude women more likely to use IM. Note in this
case the column percents provide a clearer description of the pattern than row percents.
6. BRIEF NOTES AND SAMPLE PROBLEMS
a. Measures of strength of a relationship vs a statistical test of a hypothesis. There are a number of statistics that
measure how strong a relationship is, say between variable X and variable Y. These include parametric statistics like the
Pearson Correlation coefficient, rank order correlation measures for ordinal data (Spearman's rho and Kendall's tau), and
a host of non-parametric measures including Cramer's V, phi, Yule's Q, lambda, gamma, and others. DO NOT confuse a
measure of association with a test of a hypothesis. The Chi square statistic tests a particular hypothesis. It tells you little
about how strong the relationship is, only whether you can reject a hypothesis of no relationship based upon the
evidence in your sample. The problem is that the size of Chi square depends on strength of relationships as well as
sample size and number of cells. There are measures of association based on chi square that control for the number of
cells in table and sample size. Correlation coefficients from a sample tell how strong the relationship is in the sample,
not whether you can generalize this to the population. There is a test of whether a correlation coefficient is significantly
different from zero that evaluates generalizability from the sample correlation to the population correlation. This tests
the null hypothesis that the correlation in the population is zero.
b. Statistical significance versus practical significance. Hypothesis tests merely test how confidently we can
generalize from what was found in the sample to the population we have sampled from. It assumes random
sampling-thus, you cannot do statistical hypothesis tests from a non-probability sample or a census. The larger the
sample, the easier it is to generalize to the population. For very large sample sizes, virtually ALL hypothesized
relationships are statistically significant. For very small samples, only very strong relationships will be statistically
significant. What is practically significant is a quite different matter from what is statistically significant. Check to see
how large the differences really are to judge practical significance, i.e. does the difference make a difference?.
c. SOME SAMPLE/SIMPLE STATISTICAL PROBLEMS:
36
PRR 844 Topic Outlines
Page 37
1. Calculate mean, median, standard deviation, variance from a set of data.
2. Compute Z scores to find areas under normal distribution.
3. Find an alpha (95%) percent confidence interval for the mean.(alpha given). Must estimate the standard error of
mean, 95% CI = 2 S.E.'s either side of mean.
4. Given a confidence level, accuracy desired, and estimate of variance, determine the required sample size for a
survey. n=Z22/ e2 , where Z is number of standard errors assoc. with confidence level,  is an estimate of
standard deviation of variable in the population, and e is size of error you can tolerate.
5. Similar problems for proportions rather than means. Simply replace standard deviation by sqrt(p(1-p)), the
standard deviation of a binomial distribution with probability p. n=Z2p(1-p)/ e2
6.Chi square test of relationship between two nominal scaled variables.
7. Brief Summary of Multivariate Analysis Methods. SPSS procedure in Capitals.
1. Linear Regression: Estimate a linear relationship between a dependent variable and a set of independent
variables. All must be interval scale or dichotomous (dummy variables).(See Babbie p 437, T&H, p 619, Also
JLR 15(4). Examples: estimating participation in recreation activities, cost functions, spending. REGRESSION.
2. Non-linear models : Similar to above except for the functional form of the relationship. Gravity models, logit
models, and some time series models are examples. (See Stynes & Peterson JLR 16(4) for logit, Ewing Leisure
Sciences 3(1) for gravity. Examples: Similar to above when relationships are non-linear. Gravity models widely
used in trip generation and distribution models. Logit models in predicting choices. NLR
3. Cluster analysis : A host of different methods for grouping objects based upon their similarity across several
variables. (See Romesburg JLR 11(2) & book review same issue.) Examples: Used frequently to form market
segments or otherwise group cases. See Michigan Ski Market Segmentation Bulletin #391 for a good
example.CLUSTER QUICK CLUSTER
4. Factor analysis. Method for reducing a large number of variables into a smaller number of independent
(orthogonal) dimensions or factors. (See Kass & Tinsley JLR 11(2); Babbie p 444, T&H pp 627). Examples:
Used in theory development (e.g. What are the underlying dimensions of leisure attitudes?) and data reduction
(reduce number of independent variables to smaller set). FACTOR.
5. Discriminant analysis: Predicts group membership using linear "discriminant" functions. This is a variant of
linear regression suited to predicting a nominal dependent variable. (See JLR 15(4) ; T&H pp 625). Examples:
Predict whether an individual will buy a sail, power, or pontoon boat based upon demographics and
socioeconomics. DISCRIMINANT
6. Analysis of Variance (ANOVA): To identify sources of variation in a dependent variable across one or more
independent variables. Tests null hypothesis of no difference in means of dependent variable for three or more
subgroups (levels or categories of independent variable). The basic statistical analysis technique for experimental
designs. (See T&H pp 573, 598). Multivariate analysis of variance (MANOVA) is the extension to more
complex designs. (See JLR 15(4)). ANOVA, MANOVA.
7. Multidimensional scaling (MDS): Refers to a number of methods for forming scales and identifying the structure
(dimensions) of attitudes. Differ from factor analysis in employing non-linear methods. MDS can be based on
measures of similarities between objects. Applic in recreation & tourism- mapping images of parks or travel
destinations. Identifying dimensions of leisure attitudes.RELIABILITY (See T&H pp 376.)
8.Others: Path analysis (LISREL) (Babbie p. 441), canonical correlation, conjoint analysis (See T& H 359, App C),
multiple classification analysis, time series analysis (Babbie p. 443), log linear analysis (LOGLINEAR
HILOGLINEAR) linear programming, simulation.
37
PRR 844 Topic Outlines
Page 38
Topic 10. Research Ethics : acceptable methods & practices
1. Voluntary Participation : Informed Consent
2. No Harm to Subjects, Deception, Right to Privacy
3. Anonymity & Confidentiality
4. Open & Honest reporting
5. Client confidentiality
6. Don't use research as guise for selling something or for other purposes
Examples in Babbie: Trouble in Tearoom & Human Obedience Study
Political and social issues: substance & use of research
1. Objectivity of scientist
2. Client-researcher relationships
3. Understanding range of stakeholders and the likely uses/misuses of research results
4. Scientific truth & knowledge as goal vs other practical matters
5. Social & political factors particularly important in evaluation research
6. Keeping clients & stakeholders informed
7. Political correctness & science
8. Researcher & organizational reputations
9. Relationships within and outside the organization
Topic 11. Research Writing & Presentations
1. Research style more impersonal, objective, concise and to the point, avoid embellishments
2. Complex subject requires clear organization, careful definition of terms, effective use of tables, graphs, charts.
3. Use a standard style guide, e.g. APA or Chicago Manual of Style. Follow style of outlet paper or talk will appear in.
4. Literature review, citing others work, plagiarism
5. Major headings of research article or report are ABSTRACT or EXEC SUMMARY, INTRO., PROBLEM,
OBJECTIVES, LITERATURE REVIEW, METHODS, RESULTS, DISCUSSION or IMPLICATIONS,
LIMITATIONS & SUGGESTIONS FOR FURTHER RESEARCH.
6. Typical Sub-headings in Methods Section: Study population, sampling procedures; definition of concepts,
measurement, special sections for questionnaire design, experimental design where appropriate, field procedures
including pretesting, data gathering, follow-ups etc., plan of analysis, hypotheses and hypothesis testing
procedures if appropriate.
7. General guidelines same as any communications:
a. Know your subject
b. Know your audience
c. Choose most effective way of reaching audience
d. Multiple outlets and formats for most research
e. Rewrite or practice, have paper/talk reviewed, rewrite again.
f. Learn to edit/criticize your own work
g. Use tools you have available - e.g. spell checkers, enlargement for overheads
h. Make effective use of tables, diagrams, charts, bulleted lists, etc.
38
PRR 844 Topic Outlines
Page 39
Topic 12. Supplemental Material
APPLYING RESEARCH
In some cases, one has a problem in search of solution, in other cases a research result in search of an application, in
others some of both.
PROBLEM TO SOLUTION:
1. Know what to look for. First must define what the problems/questions are. Identify key parts of problem. General and
specific information needs.
* If specific information need, consider how you might generalize the question. You are more likely to find
research relevant to a more generally stated question. Questions that are very specific, particularly to time and
place seldom have answers in existing research. Information would either exist locally (in your organization or
closely related place) or you would need to begin your own study.
Example : Need to know where my visitors come from. Possible answers found in a) your organizations
records (maybe registration forms or visitor log book). b) Design a study to find out, or c) Is there research at nearby
facilities or similar ones that measures visitor origins or how far visitors travel for this type of facility or activity.
2. Where to find it? Know sources of research information/data. Internal and external. Familiar with recent and on-going
studies in the organization/area, related organizations or nearby areas. Who to contact about research relevant to the
problem. People in organization, in area, elsewhere.
3. How to evaluate it? Is it good research? Evaluate the research report/article. See research review/evaluation
checklists.
4. Does it apply to my situation? This involves comparing the study situation with your own. Generalizability over:
a) Time: Have things changed substantially since the study was done? or is this a reasonably stable phenomona?
Can I adjust for time using simple indices, eg. price or cost index.
b) Space: Is the site/area similar in respects that would alter the results?
c) Population: Is study population similar to the one to which I intend to apply the results.
d) Situation/setting : Is the situation/setting similar? Identify any other variables that might change the results. Are
these variables similar enough between the study and the application to validly apply the results? Does the
research identify relationships or models that let me adjust the results to my situation? What assumptions
underlie these models and do they hold in this situation?
5. How to apply it.
a. Understand the research and its limits.
b. Understand the situation.
c. Consider factors not addressed in research that might yield unintended consequences.
d. Will changes/decisions be politically acceptable and accepted by all relevant stakeholders.
e. Know how to best implement different types of decisions within your organization/situation/market.
f. Establish an implementation plan.
g. Consider alternatives. Research seldom "makes" a decision. It may suggest several alternatives.
39
PRR 844 Topic Outlines
Page 40
EVALUATING RECREATION SURVEYS - A CHECKLIST
1. CLEAR PROBLEM AND OBJECTIVES: The study should have a clearly defined problem that leads to specific
study objectives. The objectives should identify the questions to be answered by the survey. There should be
methods and results for each stated objective.
a. What is the general topic? Motivation or reason for study?
b. Who conducted the study?
c. Is purpose primarily exploratory, description, explanation, prediction? If this is an evaluation
study, what is being evaluated using what criteria?
2. APPROPRIATE METHODS: The methods should be appropriate for the study's purposes in light of cost and time
constraints.
a. STUDY POPULATION should be defined in terms of content, units, extent and time Does sampling frame
represent/cover this population?
b. SAMPLE. If a sample is drawn, probability samples are needed to make inferences from the sample to the
study population. If not non-probability samples may be appropriate.
1. Evaluate if sample is representative of population
a. By comparisons with known characteristics of population
b. By evaluating sampling methods
c. Checking for non-response bias
2. Is sample size adequate? Estimate sampling error by using tables of sampling error for given
sample sizes. Are confidence intervals or estimates of sampling error reported? Watch for small
samples, if study reports results for popln subgroups.
3. What is the response rate? Does author address possible non-response bias? Is it likely to be large
or small? Were/did some groups have higher response than others?
4. Were sampling procedures carried out properly? What is time and place the data represent?
c. MEASUREMENT
1. VARIABLES : Are variables appropriate to questions/objectives and have they been
operationally defined? Has study ignored any important variables?
2.. Evaluate reliability of measures.
3. Evaluate validity of measures. Start with content or face validity. Then look for
consistency among different measures or with similar studies.
4. How were specific questions worded? Any possible question sequencing effects?
5. Is a telephone, personal interview or self-administered approach used? Is this
appropriate? What possible errors might occur from data gathering procedures?
d. ANALYSIS. Are appropriate analysis procedures used and are results reported clearly?
1. Are statistics appropriate to measurement scales
2. Are all Tables and Figures clear and easy to follow. Does text match tables?
3. Is the analysis appropriate to the study objectives? Does report present some basic descriptive
results before launching into more complex analyses.
4. If hypothesis testing is done, does author distinguish between statistical and practical significance
(importance)? Watch sample sizes in hypothesis tests.
5. Have any important variables been ignored in drawing conclusions about relationships between
variables? How strong are the relationships?
e. REPEATABILITY: Are methods presented in enough detail that you could repeat this study?
3. CONCLUSIONS AND IMPLICATIONS
a. Does author draw meaningful implications for managers, researchers, or whomever the intended
audience may be.
b. Do conclusions stay within the empirical results or does the author go beyond the findings?
c. How generalizable are results and to what kinds of situations?
d. Are study limitations, problems and errors clearly noted and discussed?
40
PRR 844 Topic Outlines
e. Is report well written, objective, and at the level of intended audience?
Page 41
Steps in a Research or Evaluation Study (Alternative Handout)
1. Identify Purpose and Constraints
a. define purpose
b. clarify purposes and information needs with decision-makers/clients
c. determine time and cost constraints and error tolerance
d. obtain agreement on desired information and study purposes
2. Develop Research Plan (includes steps 3-6)
a. identify & evaluate alternatives
b. identify available/existing information (literature & data)
c. clarify research objectives
d. write research proposal
e. have proposal reviewed and evaluated, including human subjects review
f. final plan/proposal
3. Choose Overall Approach
a. review alternatives & select approach(s)
b. obtain agreement and plan steps 4-6
4. Develop Sampling Plan
a. define study population
b. identify sampling frame (if applicable)
c. determine sample size taking into account response rates
d. evaluate alternative sampling designs & choose best
e. draw the sample
5. Develop Measurement Instruments and Procedures
a. identify information needed
b. develop operational measures for each variable
c. design questions or choose measurement instruments
d. assemble questionnaire (if applicable)
e. have instruments reviewed & evaluated, including pre-testing
f. revise and repeat evaluation of instruments as needed
6. Plan the Analysis
a. identify planned analyses, prepare "dummy" tables
b. choose statistical procedures/tests
c. evaluate if intended analysis will meet study objectives
d. review sample and measurement instruments for compatibility with the analysis (e.g. measurement scales suited
to chosen statistics, adequate samples for subgroup analyses & desired accuracy).
7. Administer the Study
a. assemble personnel and materials, assign responsibilities
b. carry out data collection procedures including follow-ups
c. manage personnel and data
d. solve problems & make adjustments to procedures as needed
8. Conduct Data Processing and Analysis
a. develop codebook
b. data entry, and cleaning, selected data checking analyses
c. carry out planned analysis
d. prepare tables and figures
e. double check results for consistency
f. data file documentation
9. Prepare Reports
a. identify audience(s) and plan report(s)
b. develop outlines for each report
c. assemble tables and figures
d. prepare text
e. reviews, editing & rewriting
f. final proofing
g. printing and distribution
h. oral presentations
10. Document the Study and File Key Reports and Materials
41
PRR 844 Topic Outlines
a. document & file final reports, codebooks, and copies of data files
b. assure confidentiality/anonymity of individual subject's records
42
Page 42
Download