Uploaded by Chiccopsianyinda

560.long paper

advertisement
ZIMBABWE OPEN UNIVERSITY
IN COLLABORTATION WITH
THE UNIVERSITY OF ZAMBIA
MASTER OF EDUCATION IN ADMINISTRATION, PLANNING
AND POLICY STUDIES
NAME: PHIRI, CHICCO
COURSE: MDEA 560
COMPUTER: 717819407
Question:
discuss the survey method and show its relevance to educational
research.
Page 1 of 16
The quality of the educational system and infrastructure is central to every nation’s economic,
social, political and technological well-being. The quality of education depends and builds on the
quality, rigor and relevance of available educational research. It is therefore of critical
importance to secure and raise the standards for conducting research in order to improve
education. “Educational research is a systematic approach of gaining a better understanding of
educational processes, generally with a view of improving educational efficiency whilst applying
scientific methods to the study of educational problems,” (Munroe, 2000: 14 )
In Zambia, the Research Council holds a critical position when it comes to organizing and
funding educational research. The Research Council has been funding educational research
programmes since the mid-1990s, starting with the research programme Competence, Learning
Processes and Value Creation in Work life, and the evaluation of the 1976 educational reforms.
However, it has been reported that all research initiatives within the educational sciences have
lacked a long-term perspective, sufficient volume for funding and infrastructures that paid
attention to processes of synthesizing and accumulating research within the education sector. It
was therefore a huge step forward when the Ministry of Education together with the Research
Council of Zambia launched the new research programme Educational Research towards 2020;
dubbed UTDANNING2020. The UTDANNING2020 research programme is designed to address
and challenge scientific merits, multidisciplinarily, rigor and relevance in educational research.
In order to promote scientific quality and merits of educational research, UTDANNING 2020
uses different tools and strategic actions. For this reason, funding of high quality research
relevant to the educational sciences holds a key position in this tool kit. Through a rich portfolio
of varied and intriguing research projects the programme aims to contribute to new insight,
accumulate knowledge, support methodological awareness and growth and contribute to
fostering research capacity within the educational sciences.
Annual seminars and conferences as mechanisms for knowledge exchange and knowledge
building are examples of other activities meant to foster quality in educational research. Within
the programme, these seminars and conferences are targeting different groups and audiences like
policymakers and stakeholders, the teaching profession, researchers and other knowledge
brokers.
Page 2 of 16
Therefore, against the above background the aim of this paper is to discuss the survey methods
and show their relevance to educational research.
Educational research has long recognized that every method of scientific inquiry is subject to
limitations and that choosing among research methods inherently involves trade-offs. With the
control of a laboratory experiment, for example, comes an artificiality that raises questions about
the generalizability of results. And yet the 'naturalness" of a field study or an observational study
can jeopardize the validity of causal inferences. The inevitability of such limitations has led
many methodologists to advocate the use of multiple methods and to insist that substantive
conclusions can be most confidently derived by triangulating across measures and methods that
have no overlapping strengths and weaknesses (Ibid).
Survey research involves the collection of data about a research issue from the target population
to answer questions that have been raised. Survey researchers have a variety of techniques at
their disposal for evaluating survey questions (Presser et al. 2004). These range from cognitive
interviews (Willis: 2005), to the conventional pretests recommended in many questionnaire
design texts (Converse and Presser: 1986), to behavior coding of various types (Maynard et al:
2002), to question wording experiments (Fowler: 2004), to the application of statistical
procedures, such as latent class analysis (Biemer: 2004) or structural equation modeling that
provide quantitative estimates of the level of error in specific items. These different evaluation
techniques do not bear a close family resemblance. Although they all share the general goal of
helping question writers to evaluate survey questions, they differ in their underlying
assumptions,the data collection methods they use, the types of problems they identify, the
practical requirements for carrying them out, the type of results they generate, and so on.
To illustrate the differences across techniques, consider cognitive interviewing and the use of
latent class modeling methods for evaluating survey items. Cognitive interviewing is a practice
derived from the protocol analyses used in the work of Simon and his collaborators. Loftus
(1984) first pointed out the potential relevance of Simon’s work to the testing of survey items
more than 25 years ago. The key assumptions of protocol analysis and its latter-day survey
descendant, cognitive interviewing, are that the cognitive processes involved in answering
survey questions leave traces in working memory (often intermediate products of the process of
formulating an answer) and that respondents can verbalize these traces with minimal distortion
Page 3 of 16
(Ericsson and Simon 1980). Cognitive interviews added several techniques to the think-aloud
methods introduced by Simon and his colleagues, especially the use of specially designed
probes, or follow-up questions; responses to these probes are also thought to provide important
clues about how respondents come up with their answers and about potential problems with
those processes. Some researchers see later developments of cognitive interviews as a departure
from the original paradigm proposed by Ericsson and Simon, and argue that the use of probes
has largely supplanted the use of think-aloud methods in cognitive interviewing (for example,
Schaeffer and Presser 2003, p. 82; see also Beatty and Willis 2007; Gerber 1999). Cognitive
interviews are rarely subjected to formal analyses; instead, the questionnaire testing personnel,
often staff with advanced degrees, draw conclusions about the questions from their impressions
of the verbal reports produced by respondents during the cognitive interviews (Willis 2005 for a
thorough discussion of cognitive interviewing).
On the other end of the continuum stands the application of quantitative methods, such as latent
class modeling, to assess problems in survey questions. In a series of papers, Biemer and his
colleagues (Biemer 2004) have used latent class models to estimate error rates in survey items
designed to assess such categorical constructs as whether a person is employed or not. Latent
class analysis is sometimes described as the categorical analogue to factor analysis
(McCutcheon: 1987). It is used to model the relationships among a set of observed categorical
variables that are indicators of two or more latent categories (for example: whether one is truly
employed or unemployed). In contrast to cognitive interviews, latent class analysis is a statistical
technique that yields quantitative estimates. It uses maximum likelihood methods to estimate
parameters that represent the prevalence of the latent classes and the probabilities of the different
observed responses to the items conditional on membership in one of the latent classes.
Given the large number of different evaluation methods and the large differences between them,
it is an important theoretical question whether the different methods should yield converging
conclusions and, if not, whether they should be used alone or in combination with each other. In
practice, the choice between techniques is often dictated by considerations of cost and schedule,
and it is an important practical question whether clear conclusions will result even if different
methods are adopted
Page 4 of 16
The answers to the questions of whether converging conclusions should be expected and how to
cope with diverging conclusions about specific items depend in part on how researchers conceive
of the purpose of the different evaluation methods. Much work on question evaluation and
pretesting tends to treat question problems as a binary characteristic – the question either has a
problem or it does not. Question evaluation methods are used to identify the problems with an
item and group questions into two categories – those with problems that require the item to be
revised, and those without such problems. Under this conceptualization, all of the question
evaluation methods flag some items as problematic and others as non-problematic, even though
the methods may differ in which items they place in each category. Of course, questions may
have problems that differ in seriousness, but ultimately questions are grouped into those that
require revision and those that do not. Different question evaluation methods are, then, compared
on their success in identifying question problems and correctly placing items into one of the two
categories. Presser and Blair’s (1994) study is a classic example of such a conceptualization.
Implicit in such work is the assumption that if any method reveals a problem with an item, that
problem should be addressed. That assumption has been challenged recently by Conrad and Blair
(2004; see also Conrad and Blair 2009), who argue that the “problems” found in cognitive
interviews may well be false alarms.
Recently, the field of question evaluation and pretesting has seen a shift towards a more general
conceptualization of question problems and goals of question evaluation methods (Miller 2009).
Survey questions measure the construct they are supposed to measure more or less well (Saris
and Gallhofer 2007). Thus, it is possible to conceive of question problems as a matter of degree,
and the purpose of question evaluation methods is to determine the degree of fit between
questions and construct (Miller 2009).
There is limited empirical work comparing different question evaluation methods, especially
work comparing qualitative methods (like cognitive interviews and expert reviews) with
quantitative methods (like measurements of reliability and validity). The few prior studies that
have been done seem to suggest that the consistency between the different methods is not very
high, even at the level of classifying items as having problems or not (Presser and Blair 1994;
Rothgeb et al. 2001; and Willis et al. 1999, for examples).
Page 5 of 16
Most researchers use the term 'sample" to refer both to (a) the set of elements that are sampled
from the sampling frame from which data ideally will be gathered and (b) the final set of
elements on which data actually are gathered. Because almost no survey has a perfect response
rate, a discrepancy almost always exists between the number of elements that are sampled and
the number of elements from which data are gathered. Lavrakas (1993) suggests that the term
'sampling pool' be used to refer to the elements that are drawn from the sampling frame for use in
sampling and that the term 'sample" be served for that subset of the sampling pool from which
data are gathered.
Moreover, survey researchers can make use of non-probability sampling methods. Such methods
have been used frequently in studies inspired by the recent surge of interest among social
psychologists in the impact of culture on social and psychological processes (Kitayama, 1996).
In a spate of articles published in top journals, a sample of people from one country has been
compared with a sample of people from another country, and differences between the samples
have been attributed to the impact of the countries' cultures. In order to convincingly make such
comparisons and properly attribute differences to culture, of course, the sample drawn from each
culture must be representative of it. And for this to be so, one of the probability sampling
procedures described above must be used.
Alternatively, one might assume that cultural impact is so universal within a country that any
arbitrary sample of people will reflect it. However, hundreds of studies of Americans have
documented numerous variations between subgroups within the culture in social psychological
processes, and even recent work on the impact of culture has documented variation within
nations (Cohen, 1996). Therefore, it is difficult to have much confidence in the presumption that
any given social psychological process is universal within any given culture, so probability
sampling seems essential to permit a conclusion about differences between cultures based on
differences between samples of them.
On the other hand, nearly all recent social psychological studies of culture have employed
nonprobability sampling procedures. These are procedures where some elements in the
population have a zero probability of being selected or have an unknown probability of being
selected. For example, Heine and Lehman (1995) compared college students enrolled in
psychology courses in a public and private university in Japan with college students enrolled in a
Page 6 of 16
psychology course at a public university in Canada. Rhee et al. (1995) compared students
enrolled in introductory psychology courses at New York University with psychology majors at
Yonsei University in Seoul: Korea. Han and Shavitt (1994) compared undergraduates at the
University of Illinois. With students enrolled in introductory communication or advertising
classes at a major university in Seoul. And Benet and Waller (1995) compared students enrolled
at two universities in Spain with Americans listed in the California 'Ibvin Registry.
In all of these studies, the researchers generalized the findings from the samples of each culture
to the entire cultures they were presumed to represent. For example, after assessing the extent to
which their two samples manifested self-enhancing biases, Heine and Lehman (1995) concluded
that 'people from cultures representative of an interdependent construal of the self," exhibited by
the Japanese students, "do not self-enhance to the same extent as people from cultures
characteristic of an independent self," exhibited by the Canadian students. Yet, the method of
recruiting potential respondents for these studies rendered zero selection probabilities for large
segments of the relevant populations. Consequently, it is impossible to know whether the
obtained samples were representative of those populations, and it is also impossible to estimate
sampling error or to construct confidence intervals for parameter estimates. As a result, the
statistical calculations used in these articles to compare the different samples were invalid,
because they presumed simple random sampling.
More important, their results are open to alternative interpretations, as is illustrated by Benet and
Waller's (1995) study. One of the authors' conclusions is that in contrast to Americans,
"Spaniards endorse a 'radical' form of individualism". Justifying this conclusion, ratings of the
terms "unconventional," "peculiar," and "odd" loaded in a factor analysis on the same factor as
ratings of "admirable" and "high-ranking" in the Spanish sample, but not in the American
sample. However, Benet and Waller's American college student sample was significantly
younger and more homogeneous in terms of age than their sample of California twins (the
average ages were 24 years and 37 years, respectively; the standard deviations of ages were 4
years and 16 years, respectively). Among Americans, young adults most likely value
unconventionality more than older adults, so what may appear in this study to be a difference
between countries attributable to culture may instead simply be an effect of age that would be
apparent within both cultures.
Page 7 of 16
Even researchers who recognize the value of survey methodology in educational research inquiry
are sometimes reluctant to initiate survey research because of misconceptions regarding the
feasibility of conducting a survey on a limited budget. And indeed, the cost of prominent largescale national surveys conducted by major survey organizations are well outside of the research
budgets of most schools. But survey methodologists have recently begun to rekindle and expand
the early work of Hansen and his colleagues (Hansen & Madow, 1953) in thinking about survey
design issues within an explicit cost-benefit framework geared towards helping researchers make
design decisions that maximize data quality within the constraints of a limited budget. This
approach to survey methodology, known as the 'total survey error" perspective can provide
educational research with a broad framework and specific guidelines for making decisions to
conduct good surveys on limited budgets while maximizing data quality, (Dillman: 1978,
Fowler: 198, Groves: 1989 & Lavrakas: 1993).
The total survey error perspective recognizes that the ultimate goal of survey research is to
accurately measure particular constructs within a sample of people who represent the population
of interest. In any given survey, the overall deviation from this ideal is the cumulative result of
several sources of survey error. Superficially, the total survey error perspective disaggregates
overall error into four components: coverage error, sampling error, non-response error, and
measurement error. Coverage error refers to the bias that can result when the pool of potential
survey participants from which a sample is selected does not include some portions of the
population of interest. Sampling error refers to the random differences that invariably exist
between any sample and the population from which it was selected. Non-response error is the
bias that can result when data are not collected from all of the members of a sample. while
measurement error refers to all distortions in the assessment of the construct of interest,
including systematic biases and random variance that can be brought about by respondents' own
behavior (such as misreporting true attitudes, or failing to pay close attention to a question),
interviewer behavior (manifested in misreporting responses, or providing clues that lead
participants to respond in one way or another), and the questionnaire (when the instrument has
ambiguous or confusing questions, or biased question wording or response options). The total
survey error perspective explicitly advocates taking into consideration each of these sources of
error and making decisions about the allocation of finite resources with the goal of reducing the
sum of the four. In the sections that follow, we consider each of these potential sources of survey
Page 8 of 16
error and their implications for psychologists seeking to balance pragmatic budget considerations
against concerns about data quality.
Surveys offer the opportunity to execute studies with various designs, each of which is suitable
for addressing particular research questions of long-standing interest to social psychologists. In
this section, we will review several standard designs, including cross-sectional, repeated crosssectional, panel, and mixed designs, and discuss when each is appropriate for social
psychological investigation. We will also review the incorporation of experiments within
surveys.
Cross-sectional surveys involve the collection of data at a single point in time from a sample
drawn from a specified population. This design is most often used to document the prevalence of
particular characteristics in a population. For example, cross-sectional surveys are routinely
conducted to assess the frequency with which people perform certain behaviors or the number of
people who hold particular attitudes or beliefs. However, documenting prevalence is typically of
little interest to social psychologists, who are usually more interested in documenting isolations
between variables and the causal processes that give rise to those associations. Cross-sectional
surveys do offer the opportunity to assess relations between variables and differences between
subgroups in a population. Although many scholars believe their value ends there, this is not the
case. Cross-sectional data can be used to test causal hypotheses in a number of ways. For
example, using statistical techniques such as two-stage least squares regression, it is possible to
estimate the causal impact of variable A on variable B, as well as the effect of variable B on
variable A (Blalock: 1972). Such an analysis rests on important assumptions about causal
relations among variables, but these assumptions can be tested and revised as necessary (James
& Singh: 1978).
Furthermore, path analytical techniques can be applied to test hypotheses about the mediators of
causal relations, thereby validating or challenging notions of the psychological mechanisms
involved. Moreover, cross-sectional data can be used to identify the moderators of relationships
between variables, thereby also shedding some light on the causal processes at work, (Kenny:
1979).
In a panel survey, data is collected from the same people at two or more points in time. Perhaps
the most obvious use of panel data is to assess the stability of psychological constructs and to
Page 9 of 16
identify the determinants of stability. At the same time, with such data, one can test causal
hypotheses in at least two ways. Firstly, one can examine whether individual-level changes over
time in an independent variable correspond to individual-level changes in a dependent variable
over the same period of time. So, for example, one can ask whether people who experienced
increasing interracial contact manifested decreasing racial prejudice, while at the same time,
people who experienced decreasing interracial contact manifested increasing racial prejudice.
Secondly, one can assess whether changes over time in a dependent variable can be predicted by
prior levels of an independent variable. Hence for example, do people who have had the highest
amounts of interracial contact at Time 1 manifest the largest decreases in racial prejudice
between Time 1 and Time 2? Such a demonstration provides relatively strong evidence
consistent with a causal hypothesis, because the changes in the dependent variable could not
have caused the prior levels of change in the independent variable (Blalock, 1995).
One application of this approach occurred in a study of a long-standing social psychological idea
called the projedm hypothesis. Rooted in cognitive consistency theories, it proposes that people
may overestimate the extent to which they agree with others whom they like, and they may
overestimate the extent to which they disagree with others whom they dislike. By the late 19805,
a number of cross-sectional studies by political psychologists yielded correlations consistent with
the notion that people's perceptions of the policy stands of presidential candidates were distorted
to be consistent with attitudes toward the candidates (Granberg: 1985, Kinder: 1978). However,
there were alternative theoretical interpretations of these correlations, so an analysis using panel
survey data seemed in order. Krosnick (1999) did such an analysis exploring whether attitudes
toward candidates measured at one time point could predict subsequent shifts in perceptions of
presidential candidates' issue stands. And he found no projection at all to have occurred, thereby
suggesting that the previously documented correlations were more likely due to other processes
(such as deciding how much to like a candidate based on agreement with him or her on policy
issues).
Furthermore, with each additional wave of panel data collection, it becomes increasingly difficult
to locate respondents to re-interview them, because some people move to different locations,
some die, and so on. This may threaten the representativeness of panel survey samples if the
members of the first-wave sample who agreed to participate in several waves of data collection
Page 10 of 16
differ in meaningful ways from the people who are interviewed initially but do not agree to
participate in subsequent waves of interviewing.
Also, participation in the initial survey may sensitize respondents to the issues under
investigation, thus changing the phenomena being studied. As a result, respondents may give
special attention or thought to these issues, which they have an impact on subsequent survey
responses. For example, Bridge et al. (1977) demonstrated that individuals who participated in a
survey interview about health subsequently considered the topic to be more important. And this
increased importance of the topic can be translated into changed behavior. For example, people
interviewed about politics are subsequently more likely to vote in elections. Even answering a
single survey question about one's intention to vote, increases the likelihood that an individual
will turn out to vote on Election Day (Greenwald & Young: 1987).
Additional evidence of causal processes can be documented in surveys by building in
experiments. If respondents are randomly assigned to 'treatment' and 'control" groups,
differences between the two groups can then be attributed to the treatment. Every one of the
survey designs described above can be modified to incorporate experimental manipulations.
Some survey respondents (selected at random) can be exposed to one version of a questionnaire,
whereas other respondents are exposed to another version. Differences in responses can then be
attributed to the specific elements that were varied.
Many social researchers are aware of examples of survey research that have incorporated
experiments to explore effects of question order and question wording. Less salient are the
abundant examples of experiments within surveys that have been conducted to explore other
social psychological phenomena, (Presser: 1981).
What is the benefit of doing these experiments in representative sample surveys? Couldn't they
instead have been done just as well in laboratory settings with college undergraduates? Certainly,
the answer to this latter question is yes; they could have been done as traditional social
psychological experiments. But the value of doing the studies within representative sample
surveys is at least three-fold. Firstly, survey evidence documents that the phenomena are
widespread enough to be observable in the general population. This boosts the apparent value of
the findings in the eyes of the many no psychologists who instinctively question the general
liability of laboratory findings regarding undergraduates.
Page 11 of 16
Secondly, estimates of effect seen from surveys provide a more accurate basis for assessing the
significance that any social psychological process is likely to have in the course of daily life.
Effects that seem large in the lab (perhaps because undergraduates are easily influenced) may
actually be quite small and thereby much less socially consequential in the general population.
Lastly, general population samples allow researchers to explore whether attributes of people that
are homogeneous in the lab but vary dramatically in the general population (for instance: age and
educational attainment) moderate the magnitudes of effects or the processes producing them
(Sanders: 1990).
Carefully monitoring ongoing data collection permits early detection of problems and seems
likely to improve data quality. In self-administered surveys, researchers should monitor
incoming data for signs that respondents are having trouble with the questionnaire. In face-toface or telephone surveys, researchers should maintain running estimates of each interviewer's
average response rate, level of productivity, and cost per completed interview. The quality of
each interviewer's completed questionnaires should be monitored, and if possible, some of the
interviewees themselves should be supervised. When surveys are conducted by telephone,
monitoring the interviews is relatively easy and inexpensive and should be done routinely. When
interviews are conducted face-to-face, interviewers can tape record some of their interviews to
permit evaluation of each aspect of the interview.
When data collection occurs from a single location (like telephone interviews that are conducted
from a central phone bank), researchers can be relatively certain that the data are authentic.
When data collection does not occur from a central location (in a case of face-to-face interviews
or telephone interviews conducted from interviewers' homes), researchers might be less certain.
It may be tempting for interviewers to falsify some of the questionnaires that they turn in, and
some occasionally do. To guard against this, researchers often establish a procedure for
confirming that a randomly selected subset of all interviews did indeed occur (recontacting some
respondents and asking them about whether the interview took place and how long it lasted).
Perhaps a first step for many scholars in this direction might be to make use of the many
archived survey data sets that are suitable for secondary analysis. So even when the cost of
conducting an original survey is prohibitive, survey data sets have a lot to offer educational
researchers. It is hoped that more educational researchers will take advantage of the opportunity
Page 12 of 16
to collect and analyze survey data in order to strengthen our collective enterprise. Doing so may
require somewhat higher levels of funding than has been evidenced from the pas. It is anticipated
that theoretical richness, scientific edibility, and impact across disciplines are likely to grow as a
result, in ways that are well worth the price.
In conclusion, over the last several decades, educational researchers have come to more fully
appreciate the complexity of social thought and behavior. In domain after domain, simple 'main
effect" theories have been replaced by more sophisticated theories involving moderated effects.
Statements such as 'People
Learners behave like this" are being replaced by 'Certain types of learners behave like this under
these circumstances." More and more, educational researchers are recognizing that learning
processes apparent in one class may operate differently among learners in other classes and that
personality factors, social identity, and cultural norms can have a profound impact on the nature
of many educational problems being researched on.
Despite this, the bulk of research in the education continues to be conducted with a very narrow,
homogeneous base of participants. As a result, some scholars have come to question the
generalization of educational research findings, and some disciplines look at the findings with
skepticism.
In this paper, survey method has been presented as an overview research methodology that
allows educational research to address pertinent problems in education. Surveys enable scholars
to explore educational research phenomena with samples that accurately represent the population
about whom generalizations are to be made. And the incorporation of experimental
manipulations into survey designs offers special scientific power. It is the writer’s anticipation
that the discussed advantages of survey research make a valuable addition to the methodological
arsenal available to educational research. The incorporation of this methodology into a full
program of research enables researchers to triangulate across measures and methods, providing
more compelling evidence of educational research phenomena than any single methodological
approach can offer.
Page 13 of 16
REFERENCES
Adomo, T. W., Frenkel-Brunmrik, E., Levinson, D. J., 6. Sanford, R. N. (1950). The
authoritarian personally. New York: Harper 6. Row.
Ahlgren, A. (1983). Sex differences in the correlates of cooperative and competitive school
attitudes. Devr&pmrntalPsychology, 19,881-888.
Alwin, D. F., Cohen, R. L., 6. Newcomb, T. M. (1991). The women of Bmningtmr: A of political
orientations over the life span. Madison: University of Wiscansin Press.
Alwin, D. F., 6. Jackson, D. J. (1982). Adult values for children: An application of factor
analysis to ranked preference data. In K. F. Schuessler (Ed.), Sociological methodology 1980.
San PransW Jossey-Bass.
Babbie, E. R. (1990). Survcy research methods. Belmont, CA: Wadsworth.
Bardn, R. M., 6. Kenny, D. A. (1986). The moderator mediator variable distinction in social
psychological research: Conceptual, strategic, and statistical considerations.
Journal of Personalily and Soda1 Psychology, 51.1 17 3- 1182.
Bass, R. T., 6 Tortora, R. D. (1988). A comparison of centralized CAT1 facilities for an
agricultural labor survey. In R. M. Groves, P. P. Beimer, L. E. Lyberg, J. T. Massey, W. L.
Nicholls, 6. J. Waksberg (Eds.), Trlephonesumymethodow(pp; 497-508). New York: Wiley.
Bearden, W. Q., Netemeyer, R. G., 6. Mobley, M. F. (1993). Handbook of marketing scales.
Newbury Park, CA: sage.
Ban, D. J., 6 McConneU, H. K. (1970.). Testing the self-perception explanation of dissonance
phenomena: On the salience of pre-manipulation attitudes. Journal of Pmmral- 3
ilyandSodalPrydtology, 1123-31.
Page 14 of 16
Benet, V., 6. Waller, N. G. (1995). The big seven factor model of personality description:
Evidence for its cross-cultural generality in a Spanish sample. Journal of Personality and soda1
Prydtology, 69,701-718.
Bischoping, K. (1989). An evaluation of interviewerdebriefingin survey pretests. In C. F.
Cannell, (pp. 15-29). Ann Arbor, MI:Survey Research Center.
Bishop, G. F., 6 Fisher, B. S. (1995). 'Seaet ballots" and self -reports in an exit-poll experiment.
Public Opinion Quartn&, 59, 568-588.
Bradburn, N. M., Sudman, S., 6. Associates. (1 981). Improving interringmethodand
qwstt'onnairc&sign. San Francisco: Jossey-Bass.
Brehm, J. (1993). The phantomrrspondntts. Ann Arbor: University of Michigan Press.
Campbell, D. T., 6 Stanley, J. C. (1963). Experimental and quasi-experimental designs for
research. In N. L. Gage (Ed.), Handbook of March on teaching (pp. 171-246). Chicago: Rand
McNally
Cohen, D., Nisbett, R. E., Bowdle, B. P., 6 Schwm, N. (1996). Insult, aggression, and the
southern culture of honor: An 'experimental ethnography.." Journal of P.-ality and Soda1
Psychology, 70,945-960.
Granberg, D., 6 Holmberg, S. (1992). The Hawthorne effect in election studies: The impact of
survey participationon voting. British Journal of PoliticulScim, 22, 240- 247.
Greenwald, A. G., Carnot, C. G., Beach, R., 6 Young, B. ( 1987). Increasing voting behavior by
asking people if they expect to vote. Journal of Applied Psychology, 72, 31 5- 318.
Hansen, M. H., 6 Madow, W. G. (1953). Survey methodsand theme. New York: Wiley.
Miethe, T. D. (1985). The validity and reliability of value measurements. Journal of Pmonaliv,
119,441453.
Presser, S., 6 Blair, J. (1994). Survey pretesting: Do different methods produce different results?
In P. V. Marsden (Ed.), Socblogicul methodology, 1994 (pp. 73-104). Cambridge, MA:
Blackwell.
Page 15 of 16
Schumau, H., 6 Presser, S. (1981). Qucsrions and ansum in attit& sum. San Diego, CA:
Academic Press.
Weisberg, H. F., Krosnick, J. A., & Bowen, B. D. (1996). An introduction to survey march,
polling, and hta analysis (3rd ed.). Newbury Park, CA: Sage.
Page 16 of 16
Download