Uploaded by Andrew Sweeney

PAF 501 - Textbook Chapter Notes

advertisement
PAF 501 - Chapter 1: Science and Scientific Research
Natural Science
This science of natural occurring objects or phenomena, such as light, objects,
matter, earth, celestial bodies, or the human body. Three categories of Natural
Science: Physical Science (physics), Earth Science (geology), Life Science (biology).
Precise, accurate, deterministic (ex. Physics testing speed of sound)
Social Science
This science of people or collections of people, such as groups, firms, societies, or
economics, and their individual or collective behaviors. Three categories of Social
Sciences: Psychology (study of human behavior), sociology (study of social
groups), economics (study of firms, markets, economics).
Less accurate, less deterministic, and have a high degree of measurement error
(ex. Measuring someone’s happiness)
Basic Science (Pure
Science)
Explains the most basic objects and forces, relationships between them, and laws
governing them. (ex. Physics, mathematics, and biology).
Applied Science
(Practical Science)
Are sciences that apply scientific knowledge from basic sciences in a physical
environment. (ex. Engineering)
Scientific Knowledge
Refers to a generalized body of laws and theories to explain a phenomenon or
behavior of interest that are acquired using the scientific method.
Laws
Are observed patterns of phenomena or behaviors (ex. Newton’s Three Laws of
Motion)
Theories
Are systematic explanations of the underlying phenomenon or behavior
Scientific Evidence
Only based on two things, logic (theory) and evidence (observations). The two are
interconnected as theories provide meaning and significance to what we observe,
and observations help validate or refine existing theories or construct new ones.
Theoretical Level
Concerned with developing abstract concepts about a natural or social
phenomenon and relationships between those concepts.
Empirical Level
Concerned with testing the theoretical concepts and relationships to see how well
they reflect our observations of reality, with the goal of ultimately building better
theories.
Inductive Research
(Theory-Building)
The goal of a researcher is to infer theoretical concepts and pattens from
observed data. Most useful when there are few prior theories or explanations.
(Generalization of observations)
Deductive Research
(Theory-Testing)
The goal of the researcher is to text concepts and patterns known from theory
using new empirical data. Not only are you testing the theory, but you’re hoping
to possibly refine, improve, and extend it. Most useful when there are many
competing theories of the same phenomenon. (Test hypothesis)
Methodological
“Know how” to conduct the right type of scientific research
Theoretical
“Know what’ to measure and look at for scientific research
Scientific Method
Refers to a standardized set of techniques for building scientific knowledge, such
as how to make valid observations, how to interpret results, and how to
generalize those results. Must satisfy four characteristics:
-
Logical: Scientific inferences must be based on logical principles of
reasoning
Confirmable: Inferences derived must match with observed evidence
Repeatable: Other scientists should be able to independently replicate or
repeat a scientific study and obtain similar, if not identical results.
Scruntinizable: The procedures used and in the inferences derived must
withstand critical scrutiny (peer review) by other scientists.
Exploratory Research
Often conducted in new areas of inquiry, where the goals of research are (1) to
scope out the magnitude or extent of a particular phenomenon problem, or
behavior, (2) to generate some initial ideas about the phenomenon or, (3) to test
the feasibility of undertaking a more extensive study regarding that phenomenon.
Descriptive Research
Is directed at making careful observations and detailed documentation of a
phenomenon of interest. These observations must be based on the scientific
method. (ex. Population growth, demographic statistics). Looking for the “what”,
“where”, “when” of the phenomenon.
Explanatory Research
Seeks explanations of observed phenomena, problems, or behaviors. Looking for
the “why” and “how” of the phenomenon.
Rationalism
Coined by Greek philosophers, it suggests that the fundamental nature of being
and the world can be understood more accurately through a process of
systematic logical reasoning.
Empiricism
Coined by Francis Bacon, suggested that knowledge can only be derived from
observations in the real world. Eventually led to the scientific method
Natural Philosophy
The idea of fusing Rationalism and Empiricism together. Focused specifically on
understanding nature and the physical universe. (Galileo and Sir Isaac Newton).
German Idealism
Argued that experience is purely subjective and processing them using pure
reason without first delving into the subjective nature of experiences will lead to
theoretical illusions. (Immanuel Kant)
Positivism
Suggested that theory and observations have circular dependence on each other.
While theories may be created via reasoning, they are only authentic if they can
be verified through observations. Was usually tied to quantitative research
methods. (Auguste Comte)
Antipositivism
Employed qualitative methods such as unstructured interviews and participant
observation.
Postpositivism
Amends positivism by suggesting that it is impossible to verify the truth although
it is possible to reject false beliefs, though it retains the positivist notion of an
objective truth and its emphasis on the scientific method.
Critical Research
(Critical Theory)
Retains similar ideas of critiquing and resolving social inequality and adds that
people can and should consciously act to change their social and economic
circumstances, although their ability to do so is constrained by various forms of
social, cultural, and political domination.
PAF 501 - Chapter 2: Thinking Like a Researcher
Unit of Analysis
Refers to the person, collective, or object that is the target of the investigation.
(Individuals, groups, organizations, countries, technologies, and objects).
Concepts
Generalizable properties or characteristics associated with objects, events, or
people.
Construct
An abstract concept that is specifically chosen or created to explain a given
phenomenon.
Broken down into several underlying concepts:
-
Unidimensional: ex. Someone’s weight
Multi-Dimensional: ex. Someone’s communication skills
Operational
Definitions
Define constructs in terms of how they will be empirically measured (ex. If you’re
measuring temperature, are you using Celsius, Fahrenheit, or Kelvin?).
Variable
Is a measurable representation of an abstract construct.
Nomological Network:
-
Independent Variable: A variable that explains another variable (the cause)
Dependent Variable: Variables that are explained by other variables. (the
effect)
Mediating Variables: A variable that are explained by independent
variables while also explain dependent variables.
Moderating Variables: A variables that influences the relationship between
independent and dependent variables.
Control Variables: A variable that must be controlled for study because if
not it has the ability to impact other variables.
Proposition
Is a tentative and conjectural relationship between constructs that is stated in a
declarative form. (ex. An increase in student intelligence causes an increase in
their academic achievement) (stated in an theoretical plane)
Hypotheses
The empirical formulation of propositions, stated as a relationship between
variables. (ex. An increase in students’ IQ score causes an increase in their grade
point average) (stated in an empirical plane)
Theory
A set of systematically interrelated constructs and propositions intended to
explain and predict a phenomenon or behavior of interest, within certain
boundary conditions and assumptions. (explains the phenomenon)
Model
Is a representation of all or part of a system that is constructed to study that
system. . . e.g. how the system works or what triggers the system. (represents the
phenomenon)
Deduction
Is the process of drawing conclusions about a phenomenon or behavior based on
theoretical or logical reasons and an initial set of premises. (ex. A bank enforces a
strict code of ethics for its employees and Jamie is an employee at that bank, then
Jamie can be trusted to follow ethical practices)
Induction
Is the process of drawing conclusions based on facts or observed evidences. (ex. If
a firm spent a lot of money on a promotional campaign, but the sales did not
increases, then possibly the promotion campaign was poorly executed).
PAF 501 - Chapter 3: The Research Process
Paradigms
The mental models or frames we use to organize our reasoning and observations
(belief systems)
Ontology
Refers to our assumption about how we see the world e.g., does the world consist
mostly of social order or constant change. S
Epistemology
Refers to our assumptions about the best way to study world, e.g., should we use
an objective or subjective approach to study social reality.
Functionalism
View the world consisting of social order and seek patterns of ordered events or
behaviors and believe that the best way to study the world is using objective
approach by using standardized data collection like surveys. (Objectivism/Social
Order)
Interpretivism
Study social order through subjective interpretation of participants involved, such
as interviewing different participants and reconciling differences among their
responses using their own subjective perspectives (Subjectivism/Social Order)
Radical Structuralism
Believe that the world consists of radical changes and seek to understand or enact
change using an objectivist approach (Objectivism/Radical Change)
Radical Humanism
Wish to understand social change using subjective perspectives of the
participants involved (Subjectivism/Radical Change)
Research Process
Exploration
1) Observation: We observe a natural or social phenomenon, event, or
behavior that interest us.
2) Rationalization: We try to make sense of or the observed phenomenon,
event or behavior by logically connecting the difference pieces of the
puzzle that we observe, which in some cases, may lead to the construction
of a theory.
3) Validation: We test our theories using scientific method through a process
of data collection and analysis and in dog so, possibly modify or extend
our initial theory.
The first phase of research. This phase includes exploring and selecting research
questions for further investigation, examining the published literature in the area
of inquiry to understand the current state of knowledge in the area, and
identifying theories that may help answer the research questions of interest.
1) Research Questions: Identifying one or more questions dealing with a
specific behavior, event, or phenomena of interest that you wish to seek
answers for in your research. (Research questions are more appealing if
they can be for a broader population).
2) Literature review: three-fold purpose (1) to survey the current state of
knowledge in the area of inquiry (2) to identify key authors, articles,
theories, and finding in that area and (3) to identify gaps of knowledge in
the research area.
3) Theories: Will help identify which of these constructs is logically relevant to
the target phenomenon and how.
Research Design
This process is concerned with creating a blueprint of the activities to take in
order to satisfactorily answer the research questions identified in the exploration
phase.
1) Operationalization: Is the process of designing precise measure for
abstract theoretical constructs.
2) Research Method: What way are you going to collect data to address their
research question(s) of interest.
3) Sampling: Choose the target population from which you wish to collect
data. Closing linked to the unit of analysis in a research problem.
Research Proposal
Details all of the decisions made in the preceding stages of the research process
and the rationale behind each decision. What’s your question? Why? What’s the
prior knowledge on it? What theories do you wish to employ? What hypothesis
do you want to test? How can it be measured? What’s the research method?
Why? What’s the desired sampling structure? Etc.
Research Execution
Arrive here once you decide who to study (subjects), what to measure (concepts),
and how to collect data (research method). Now you include pilot testings the
measurement instruments, data collection, and data analysis.
1) Pilot Testing: Testing on usually a small subset of the target population
that you’re seeing if there’s any problems in your research design.
2) Data Collection: Once your pilot test is successful you can start collecting
your qualitative or quantitative data.
3) Research Analysis: Analyze and interpret the data for the purpose of
drawing conclusions regarding your research question.
4) Research Report: Document the entire research process and its findings in
the form of a research paper, dissertation, or monograph. This should
outline in detail all the choices made during the research process and why,
and what the outcomes at each phase were.
PAF 501 - Chapter 4: Theories in Scientific Research
Science
Knowledge represented as a collection of “theories” derived using the scientific
method.
Idiographic
Explanations
Are those that explain a single situation or event in idiosyncratic detail. (ex. You
did poorly on an exam because (1) you forgot tht you had an exam on that day,
(2) you arrived late to the exam due to a traffic jam, (3) you panicked midway
through that exam).
Nomothetic
Explanations
Seek to explain a class of situations or events rather than a specific situation or
event. (ex. Students do poorly in exams do so because they did not spend
adequate time preparing for exams or that they suffer from nervousness,
attention-deficit, or some other medical disorder). Generalized across situations,
events, or people . . . less precise, less complete, less detailed.
Theory
Four Building Block of Theory: Constructs, propositions, logic, and boundary
conditions/assumptions.
1) Concepts: Capture the “what” of theories. Reminder concepts can be
unidimensional (embody a single concept . . . height, weight, age) or they
can be multi-dimensional (embody multiple underlying concepts . . .
culture, personality).
2) Propositions: Capture the “how” of theories. Associations postulated
between constructs on deductive logic. Should ideally indicate a causeeffect relationship (if X happens, then Y with follow). Exists only on the
theoretical plane (Construct A -> Proposition -> Construct B)
o Hypothesis: Is the empirical formulation of a proposition. Exists only
on the empirical place. (Independent Variable -> Hypothesis ->
Dependent Variable).
3) Logic: Represents the “why” of theories. Provides the basis for justifying
the propositions as postulated. Acts as like the glue that connects the
theoretical constructs and provides meaning and relevance to the
relationships between these constructs.
4) Boundary Conditions/Assumptions: Examines the “who, when, and where”
of theories. Govern where the theory can be applied and where it cannot
be applied. (ex. Many economic theories assume that human beings are
rational and employ utility maximizations based on cost and benefit
expectations as a way of understanding human behavior.
How can we evaluate
the goodness of a
given theory?
-
How do we build a
theory?
-
Logical Consistency: Are the theoretical constructs, propositions, boundary
conditions, and assumptions logically consistent with each other?
Explanatory Power: How much does a given theory explain (or predict)
reality?
Falsifiability: Theories cannot be theories unless they can be empirically
testable to see if their falsifiable.
Parsimony: Examines how much of a phenomenon is explained with how
few variables.
Grounded Theory Building: Observe patterns of events and behaviors.
Bottom-Up Conceptual Analysis: Identify different sets of predictors
relevant to the phenomenon of interest using a predefined framework.
Agency Theory
(Principal-Agent
Theory)
Classic theory in the organizational economics literature that explains two-party
relationships. (ex. Such as those between an employer and its employees). The
core assumptions of this theory are that human beings are self-interested
individuals, boundedly rational, and risk-averse. (Consists of a principal and agent)
Theory of Planned
Behavior
Presumes that individual behavior represents conscious reasoned choice and is
shaped by cognitive thinking and social pressures.
Innovative Diffusion
Theory
Explains how innovations are adopted within a population of potential adopters.
Four key elements in this theory are innovation, communications channels, time,
and social system.
Elaboration Likelihood
Model
Dual-process theory of attitude formation or change in psychology literature.
Explains how individuals can be influenced to change their attitude toward a
certain object, events, or behavior and the relative efficacy of such change
strategies.
General Deterrence
Theory
Examines why certain individuals engage in deviant, anti-social, or criminal
behaviors.
PAF 501 - Chapter 16: Research Ethics
Ethics
Conformance to the standards of conduct of a given profession or graph.
-
-
Example: scientists should not manipulate their data collection, analysis,
and interpretation procedures in a way that contradicts the principles of
science or scientific method or advances their personal agenda.
It’s an moral distinction between right and wrong, not necessarily what’s
legal and illegal.
Voluntary
Participation and
Harmlessness
Subjects in a research project must be aware that their participation in the study is
voluntary, that they have the freedom to withdraw from the study at any time
without any unfavorable consequences, and they are not harmed as a result of
their participation or non-participation in the project.
Informed Consent
Form
Clearly describes their right to not participate and right to withdraw, before their
responses in the study can be recorded.
-
If under 18, form must be sign by legal guardian.
Must retain these forms for often times up to three years after the study
concludes.
Legal Condition
Anonymity
Implies that the researcher or readers of the final research report or paper cannot
identify a given response with a specific respondent.
Confidentiality
Which the researcher can identify a person’s responses but promises not to
divulge that person’s identify in any report, paper, or public forum.
Disclosure
Usually the obligation to provide some information about their study to potential
subjects before data collection to help them decide whether or not they wish to
participate in the study.
Analysis & Reporting
Researchers have ethical obligations to the scientific community on how data is
analyzed and reported in their study.
Institutional Review
Boards
The IRB reviews all research proposal involving human subjects to ensure that the
principles of voluntary participation, harmlessness, anonymity, confidentiality.
1) Exempt:
o Protected subject identity
o Work already done
o Normal course of events
2) Expedited
o Social Science Data
o Medical but not dangerous
3) Full Review
o Invasive
o Pain / harm
Association of
Information Systems
(AIS)
The global professional association of researchers in the information systems
discipline,
-
-
Category 1: Includes serious transgressions such as plagiarism and
falsification of data, research procedures, or data analysis, which may lead
to expulsion from the association, dismissal from employment, legal
action, and fatal damage to professional reputation.
Category 2: Includes less serious transgression such as not respecting the
rights of research subjects, misrepresenting the originality of research
projects, and using data published by others without acknowledgement.
In-Person Notes (8/30/21):
-
-
Asch Experiment: Taught us about group conformity
o College student asked to say which line length was closest to the baseline: A, B, C
 Confederate in group selected the wrong answer.
Milgram Experiment: Authority figure. Will you bend your standards to adhere to authority.
o Experimenter directs the teacher to give electric shocks to a student subject.
Online Lecture Video:
-
-
Fraud: Drop outliers, bad reporting, plagiarism/theft, set up experiments with known results.
Plagiarism:
o Reference all material whether it is in print or in-line
Human Subject Use: Today’s biological and social scientist are required to ensure the safety and dignity
of human subjects.
o 18th Century: Edward Jenner: smallpox vaccinations on son and neighborhood children
o 20th Century: mustard gas experiment on Irish / Indian soldiers, WW2 Nazi experiments.
Nuremburg Code: Demonstrate that research has a purpose, avoid all unnecessary harm, inform subject
of potential harm, and subject must volunteer, be provide informed consent, and be empowered to
stop the experiment. (The Code is not Law, just a recommendation)
-
The 1979 Belmont Report:
o Autonomy: Respect for persons – each person is given respect, time, and opportunity to make a
decision
o Beneficence: Obligation of the researcher to protect the subject and to maximize their benefit
from the research
o Justice: Risk vs. Benefits.
PAF 501 - Chapter 5: Research Design
Research Design
Comprehensive plan for data collection in an empirical research project. It is the
“blueprint” for empirical research aimed at answering specific research questions
or testing specific hypotheses, and must specify at least three processes
1) The data collection process
2) The instrument development process
3) The sampling process
Positive Methods
Such as laboratory experiments and survey research, are aimed at theory (or
hypotheses) testing, while interpretive methods, such as action research and
ethnography, are aimed at theory building.
-
Interpretive Methods
Employ an inductive approach that starts with data and tries to derive a theory
about the phenomenon of interest from the observed data.
-
Mixed-Mode Designs
Employ a deductive approach to research, starting with a theory and
testing theoretical postulates using empirical data.
Conducting multiple interviews and then coming up with a theme from
them.
A combination of qualitative and quantitative data
Quality of Research
-
-
-
-
Internal Validity (Causality): Examines whether the observed change in a
dependent variable is indeed caused by a corresponding change in
hypothesized independent variables, and not by variables extraneous to
the research context.
o Causality requires (1) covariation of cause and effect, (2) temporal
precedence: cause must proceed effect in time, (look at the
sequence. This happens than that happens afterwards. (3) no
plausible alternative explanation.
External Validity (Generalizability): Refers to whether the observed
associations can be generalized from the sample to the population
(population validity), or to other people, organizations, contexts, or time
(ecological validity).
Construct Validity: Examines how well a given measurement scale is
measuring the theoretical construct that it is expected to measure.
o Example: Construct validity must assure that a measure of empathy
is indeed measuring empathy and not compassion, which may be
difficult since theses
o constructs are somewhat similar In meaning.
Statistical Conclusion Validity: Examines the extent to which conclusions
derived using a statistical procedure is valid. (Not applicable for
interpretive research designs).
Controls required to
assure internal validity
(causality) of research
designs
Experimental Studies
1) Manipulation: The researcher manipulates the independent variables in
one or more levels (called treatments) and compares the effect of the
treatments against a control group where subject do not receive the
treatment.
2) Elimination: Technique relies on eliminating extraneous variables by
holding them constant across treatments, such as by restricting the study
to a single gender or a single socio-economic status.
3) Inclusion Technique: The role of the extraneous variables is considered by
including them in the research design and separately estimating their
effects on the dependent variable. (Allows for greater generalizability, but
requires larger samples).
4) Statistical Control: Extraneous variables are measured and used as
covariates during the statistical testing process.
5) Randomization: Is aimed at canceling out the effects of extraneous
variables through a process of random sampling, if it can be assured that
these effect are of a random (non-systematic) nature.
o Random Selection: Where a sample is selected randomly from a
population
o Random Assignment: Where subjects selected in a non-random
manner are randomly assigned to treatment groups.
Are intended to test cause-effect relationship (hypotheses) in a tightly controlled
setting by separating that cause from the effect in time, administrating the cause
to one group of subjects (the “treatment group”) but not to another group
(“control group”), and observing how the mean effects vary between subjects in
these two groups.
-
True Experimental Designs: subjects must be randomly assigned between
each group.
-
Quasi-Experimental Designs: When random assignment between each
group isn’t followed.
Video Lecture Notes:
-
Studies:
o Level of Analysis: What is the focus of the study?
o Unit of Analysis: What is being measured? (people, nations, households cells, mRNA, protein,
bones
-
Variables
o Level of Measurement: Distinguishes between quantitative and qualitative variables (nominal,
ordinal, interval, ratio)
 Nominal: Categorical data. (Male or Female)
 Ordinal: Order of Variables (Surveys such as Very Happy, Happy, Fine, Unhappy, and Very
Unhappy) (Try to always have 5 categories
 Interval: Real number data. (Difference between 1 and 2, is the same as 3 and 4. Example:
Temperature)
 Ratio: Must have a clear definition of 0.
o

Unit of Measurement: The specific unit such as dollars, cents, kilograms, pounds.
In-Person Notes (8/30/21):
-
PAF 501 - Chapter 6: Measurement of Constructs
Conceptualization
Is the mental process by which fuzzy and imprecise constructs (concepts) and
their constituent components are defined in concrete and precise terms.
Unidimensional
Are constructs that are expected to have a single underlying dimension. Example:
person’s weight, wind speed, and probably even complex constructs like selfesteem.
Multidimensional
Are constructs that consist of two or more underlying dimensions. Example: If we
conceptualize a person’s academic aptitude as consisting of two dimensions, such
as mathematical and verbal abilities.
Operationalization
Occurs after a theoretical construct is defined. Refers to the process of developing
indicators or items for measuring these constructs.
-
Constructs: Exist on the theoretical level
Indicators: Exist on the empirical level
Variable
A combination of indicators at the empirical level.
-
Each indicators may have several attributes (or levels) and each attribute
represents a value. Example: Gender variable has two attributes, male and
female.
o Values of attributes may be qualitative (non-numeric) or
quantitative (numeric)
Reflective Indicator
Is a measure that “reflects” an underlying construct. Example: If religiosity is
defined as a construct that measure how religious a person is, then attending
religious service may be a reflective indicator of religiosity. (Construct causes the
change in the measurement items (indicators))
Formative Indicator
Is a measure that “forms” or contributes to an underlying construct. Such
indicators may represent different dimensions of the construct of interest.
Example: If religiosity is defined as composing of a belief dimension, a devotional
dimension, and a ritual dimension, then indicators chosen to measure each of
these different dimensions will be considered formative indicators. (Measurement
items (indicators) cause the change in the construct)
Levels of
Measurement
(Rating Scale)
Refer to the values that an indicator can take (by says nothing about the
indicators itself). Example, gender could be measured as M and F or 1 and 2.
Scale
Central
Tendency
Statistics
Transformations
Nominal
Mode
Chi-square
Ordinal
Median
Interval
Arithmetic
mean,
range,
standard
deviation
Geometric
mean,
harmonic
mean
Percentile,
nonparametric
statistics
Correlation,
regression,
analysis of
variance
One-to-One
(equality)
Monotonic
Increasing
(order)
Ratio
Nominal (Categorical
Scale)
Positive linear
(affine)
Coefficient of Positive
variation
similarities
(multiplicative,
logarithmic
Measures categorical data. These scales use used for variables or indicators that
have mutually exclusive attributes. Example: Gender (two values: male and
female). Nominal scale merely offer names or labels for different attribute values.
Ordinal Scales
Measure rank-ordered data, such as the ranking of students in a class as first,
second, third, and so forth. However, the actual or relative values of attributes or
difference in attribute values cannot be assessed. For instance, ranking of students
in class says nothing about the actual GPA or test scores of the students.
Interval Scales
Are those where the values measured are not only rank-ordered, but are also
equidistant from adjacent attributes. Example, the temperature scale where that
difference between 30 and 40 degree F is the same as the difference between 80
and 90 F. Allows us to examine “how much more” is one attribute when compared
to another, which is not possible with nominal or ordinal scales.
Ratio Scales
Are those that have all the qualities of nominal, ordinal, and interval scales, and in
addition, also have a “true zero” point (where the value zero implies lack or nonavailability of the underling construct. Example: A firm of size zero means that it
has not employees or revenues.
Binary Scales
Are nominal scales consisting of binary items that assumed one of two possible
values, such as yes or no.
Video Lecture Notes:
-
Quality of Measurement:
o The indicator represents the concept.
-
Measuring Concepts: It is usually too complex to represent a construct with only one indicator. Usually a
collection of indicators to represent a concept is usually referred to as a scale or latent variable.
o Example of multiple indicators making up a construct . . .
 Social Economic Status (Construct)
1) Indicator: Income
2) Indicator: Education
3) Indicator: Occupation
-
Quality of Measurement Continued
- The indicator is measured properly. (All indicators include a small amount of error)
 Measurements should be. . .
1. Accurate
2. Unbiased
3. Reliable (Consistency)
- Measurement error is random. Statistics can control measurement error.
In-Person Notes (8/30/21):
-
Likert Scales: Strongly agree, strongly disagree, etc
PAF 501 - Chapter 7: Scale Reliability and Validity
Reliability
Is the degree to which the measure of a construct is consistent or dependable. In
other words, if we use this scale to measure the same construct multiple times, do
we get pretty much the same result every time, assuming the underlying
phenomenon is not changing?
-
Note, reliability implies consistency but not accuracy.
Inter-rater Reliability
(Inter-observer
reliability)
Is a measure of consistency between two or more independent raters (observers)
of the same construct. Example: If these are two rates rating 1—observations into
one of three possible categories, and their rating match from 75% of the
observations, then the inter-rater reliability is 0.75.
Test-retest Reliability
Is a measure of consistency between two measurements (tests) of the same
construct administered to the same sample at two difference points in time. The
time interval between the two tests is critical
Split-half Reliability
Split-half reliability is a measure of consistency between two halves of a construct
measure.
True Score Theory
This is a psychometric theory that examines how measurement works, what it
measures, and what it does not measure.
X (Observed Score = T (True Score) + E (Error)
Measurement Errors
Random Error: Is the error that can be attributed to a set of unknown and
uncontrollable external factors that randomly influence some observations but
not others. Adds variability (standard deviation) to the distribution but does not
affect its central tendency (mean)
Systematic Error: Is an error that is introduced by factors that systematically affect
all observations of a construct across an entire sample in a systematic manner.
Affects the central tendency but not the variability, this reduces the validity of the
measurement.
X = T + Er + Es
Assessment of Validity
PAF 501 - Chapter 8: Sampling
Sampling
Is the statistical process of selecting a subset of a population of interest for
purposes of making observations and statistical inferences about that population.
Sampling Process
Population: The group you want to generalize to -> Sampling Frame: A list from
where you can draw your sample -> Sample: the actual units selected for
observation.
Probability Sampling
Is technique in which every unit in the population has a chance (non-zero
probability) of being selected in the sample, and this chance can be accurately
determined.
Simple Random
Sampling
In this technique, all possible subset of the population (more accurately, of a
sampling frame) are given an equal probability of being select.
Systematic Sampling
In this technique, the sampling frame is ordered according to some criteria and
elements are selected at regular intervals through that ordered list.
Stratified Sampling
The sampling frame is divided into homogeneous and non-overlapping
subgroups (called “strata”), and a simple random sample is drawn within each
subgroup. Layered sample.
Cluster Sampling
If you have a population dispersed over a wide geographic region, it may not be
feasible to conduct a simple random sampling of the entire population. In such
case, it may be reasonable to divide the population into cluster, randomly sample
a few clusters, and measure all units with that cluster.
Match-pairs Sampling
Sometimes research may want to compare two subgroups within one population
based on a specific criterion.
Multi-state sampling
The probability sampling techniques described previously are all examples of
single stage sampling techniques. Depending on your sampling needs, you may
combine these single-stage techniques to conduct multi-state sampling.
Nonprobability
Sampling
A sampling technique in which some units of the population have zero chance of
selection or where the probability of selection cannot be accurately determined.
Does not allow the estimation of sampling errors, and may be subjected to
sampling bias.
Convenience Sampling
Also called accidental or opportunity sampling, this is a technique in which a
sample is drawn from that part of the population that is close to hand, readily
available, or convenient.
Quota Sampling
In this technique the population is segmented into mutually-exclusive subgroups
)just as in stratified sampling), and then a non-random set of observations is
chosen from each subgroup to meet a predefined quota.
-
Proportional Quote Sampling: The proportion of respondents in each
subgroup should match that of the population.
Non-proportional Quote Sampling: IS less restrictive in that you don’t have
to achieve a proportional representation, but perhaps meet a minimum
size in each subgroup.
Expert Sampling
This is a technique where respondents are chosen in a non-random manner based
on their expertise on the phenomenon being studied.
Snowballing Sampling
You start by identifying a few respondents that match the criteria for inclusion in
your study, and then ask them to recommend others they know who also meet
your selection criteria.
Response
Is a measurement value provided by a sampled unit
Frequency
Distribution
Graphed responses from difference respondents to the same item or observation
Normal Distribution
Large number of responses in a sample where the frequency distribution tends to
resemble a bell-shaped curve. Can be used to estimate overall characteristics of
the entire sample, such as sample mean or standard deviation. These estimates
are called sample statistics.
Population
Parameters
Since the entire population can never be sampled, population characteristics are
always unknown and called this
Sampling Error
Differences between Sample Statistics and Population Parameters.
Standard Error
The variability or spread of sample statistics in a sampling distribution. (The
standard deviation of a sampling statistic).
Confidence Interval
Is the estimated probability that a population parameter lies within a specific
interval of sample statistic values. All normal distributions tend to follow a 68-9599 precent rule, which says that over 68% of the cases in the distribution lie within
on standard deviation of the mean, over 95% of the cases in the distribution lie
within two standard deviations of the mean value, over 99% of the cases in the
distribution lie within three standard deviations of the mean value.
Video Lecture Notes:
What is Sampling?
-
Theory of Sampling: To represent a population of interest as a smaller cohort (or grouping) of individuals within a research
experiment to test a hypothesis.
-
Sampling: Demonstrating Causation
-
Inference: To draw conclusion about a population from a sample
o How do we do this? Through statistics.
Types of Sampling:
-
-
Random Sampling (Probability)
o Simple Random Sampling
 Every potential subject has an equal chance of being selected
 Drawing numbers out of hat
 Random number generator
 Advantages
 Findings are very generalizable
o Systemic Random Sampling
o Stratified Random Sampling
o Proportionate Random Sampling
o Disproportionate
o Cluster
o Multi-stage sampling
o Catch / re-catch
Non-probability Sampling (Not Random)
o Purposive Sampling
o Accidental, Haphazard
o Quota
o Snowball sampling
Steps for Picking a Sample:
PAF 501 - Chapter 10: Experimental Research
Experimental Research
Often considered the gold standard in research designs. In this design, one of
more independent variables are manipulated by the researcher (as treatments),
subjects are randomly assigned to different treatment levels (random
assignment), and the results of the treatments on outcomes (dependent variables)
are observed.
-
Internal Validity: The unique strength of experimental research is its
internal validity (causality) due to its ability to link cause and effect
through treatment manipulation, while controlling for the spurious effect
of extraneous variable.
Laboratory
Experiments
Conducted in a lab (artificial) setting tend to be high in internal validity, but this
comes at the cost of low external validity (generalizability) because the artificial
lab setting in which the study is conducted may not reflect the real world
Field Experiments
Conducted in field settings such as in a real organization, and high in both
internal and external validity. These studies are rare because of the difficulties
associated with manipulating treatments and controlling for extraneous effects in
a field setting.
Treatment / Control
Groups
In order to test the effects of a new drug intended to treat a certain medical
condition like dementia, if a sample of dementia patients is randomly divided into
three groups, with the first group receiving a high dosage of the drug, the second
group receiving a low dosage, and the third group receives a placebo such as a
sugar pill (control group), the first two groups are experimental groups, and the
third group is a control group.
Treatment
Manipulation
Pretest Measures: Conducted before the treatment is administered
Random Selection vs.
Random Assignment
Random Selection: Is the process of randomly drawing a sample from a
population or sampling frame. (Usually employed in survey research because it
assures each unit in the population has a positive chance of being selected into
the sample). High external validity (generalizability).
Posttest Measures: Conducted after the treatment is administered
Random Assignment: Is a process in true experimental research to ensure that
treatment groups are similar (equivalent) to each other and to the control group,
prior to treatment administration. High internal validity.
Threats to Internal
Validity
-
-
History Threat: A historic event, not related to the experiment mat
influence outcomes
Maturation Threat: Duration or aging can affect a post-test or outcome
- Maturity
- Boredom
- Strength
Testing Threat: A pre-test can affect a post-test (sensitive to issue to a
subject response to please the experimenter)
Instrumentation Threat: Reliability (or appropriateness) of the tool or
qualification of the person used to evaluate the outcome
Mortality Threat: Sample reduction that may not be independent from the
dependent variable (outcome)
- Not actual, mortality, this refers to drop out rate for human
subjects, but can also refer to “rodent” death in animal
experiments.
- Reflects the same idea as attrition, but is considered a group threat
because it reveals that the two “groups” might have been different
in ways not recognized. Attrition can occur for studies with only
one group as some participants leave over time.
-
Regression Threat: Regression toward the mean (sample chosen for
extreme scores – there is nowhere to go but away from the extreme)
In-person: time and group effects
Discussion post is done on this – find an actual study that control or didn’t control
for these threat on internal validity
Two Group
Experiments:
Pretest-Posttest
Control Group Designs
Two Group
Experiments:
Subjects are randomly assigned to treatment and control groups, subjected to an
initial (pretest) measurement of dependent variables of interests, the treatment
group is administered a treatment (representing the independent variable of
interest), and the dependent variables measured again (posttest). The treatments
effect is measured as the difference in the posttest and pretest scores between
the treatment and control groups.
This design is simpler version of the pretest-posttest design where pretest
measurements are omitted. The treatments effect is measured simply as the
difference in the posttest scores between the two groups.
Posttest Only Control
Group Design
Two Group
Experiments:
Covariance Designs
Sometimes, measures of dependent variables may by influenced by extraneous
variables called covariates. Covariates are those variables that are not of central
interest to an experimental study, buy should nevertheless by controlled in an
experimental design in order to eliminate their potential effect on the dependent
variables.
Covariance Design is a special type of pretest posttest control group design
where the pretest measure is essentially a measurement of the variants of interest
rather than that of the dependent variables. This effect is measured as the
difference in the posttest scores between the treatment and the control group.
Factorial Designs
Enable the researcher to examine not only the individual effect of each treatment
on the dependent variables (called main effects), but also their joint effect (called
interaction effects).
-
-
Main Effect: Is said to exist if the dependent variable shows a significant
difference between multiple levels of on factors, at all levels of other
factors.
Interaction Effect: Exists when the effect of differences in one factor
depends upon the level of a second factor.
Hybrid Experimental
Designs:
This is a variation of the posttest-only or pretest-[posttest control group design
where the subject population can be grouped intro relatively homogeneous
subgroups (called blocks) within which the experiment is replicated.
Randomized Block
Design
Hybrid Experimental
Designs:
Solomon Four-Group
Design
Hybrid Experimental
Designs:
Switch Replication
Design
In this design, the sample is divided into two treatment groups and two control
groups. One treatment group and one control group receive the pretest and the
other two groups do not. This design represents a combination of post-test-only
and pretest-posttest control group designs, and is intended to test for potential
biasing effect of pretest measurement on posttest measures that tend to occur in
pretest-posttest designs but not in posttest only designs.
Two-group design implemented in two phases with three waves of measurement.
The treatment group in the first phase serves as the control group in the second
phases, and the control group in the first phase serves because the treatment
group in the second phases.
Video Lecture Notes: Research Continuum
-
Level of Analysis: What are you studying?
- Macro Whole System (In Vivo), Midlevel (In Situ), Micro Control (In Vitro)
Video Lecture Notes: Causality
-
-
-
-
-
-
Slide 1
- Ideas are theories only when you can test them to be falsifiable
 We do this testing through a hypothesis
1. Test for hypothesis should be able to be replicated
Slide 2:
- Theories are statements about relationships between concepts or constructions. A statement about a
cause-and-effect association between variables.
 They are a statement about an association between variables: + or –
 Remember: a theory is about an association not necessarily a cause.
1. A theory can be tested repeatedly
2. Evidence can accept or refute a hypothesis about an association which strengthens or
weakens a theory.
Slide 3:
- Variables: may be modifiable (weight, calorie intake) or non-modifiable (race, ethnic origin)
- Constant the opposite of a variable, these stay the same.
- Variables can have multiple groups and in those groups you can have different constants
Slide 4:
- Independent Variable: Measure of the cause
 This variable, condition, or “thing” that is manipulated (or changed) by the research in an
experiment
- Dependent Variable: Measure of the effect
 This is the variable that is observed (for changes) following change or manipulations in the
independent variable.
Slide 5
- How do we show causation? (Demonstrating Causation)
 Co-variation (must be correlated – either + or -)
 Cause prior to effect
 Absence of a plausible rival hypothesis
Slide 6:
- Covariation: A change in one variable (X – manipulation/treatment – what your change) corresponds with
a change in another variable (y – outcome/observation)
-
-
Slide 7:
- Cause and Effect Names

Slide 8:
- Reserve Causality


-
Slide 9:
- Pretest-Posttest – Insuring Cause Happens Before Effect

-
-
Slide 10:
- Plausible Rival Hypothesis (Empirical Example)

Slide 11:
- Hypothesis Testing: Operationalization of the cause and effect

Video Lecture Notes: The Experimental Method (Random Control Trials)
-
Slide 1:
- Whether a “study” is an experiment depends on its “design”
 “Design” implies methods including group set-up, required techniques/strategies, and analysis.
-
Slide 2:
- How to demonstrate cause in Random Control Trial Experiment
-

Slide 3:
- Random Control Trials



Random Assignment vs Random Sampling


Random Assignment vs Matching

-
-
1.
Matching can be influenced by your bias
Slide 4:
- Elements of a Random Control Trial

Slide 5:
- Pretest -> Posttest

-
-
-
Slide 6:
- RCT Design Graphic

Slide 7:
- RCTS and Causality

Slide 8:
- Downside of RCT
 Dealing with group threats through random assignment and time threats through
manipulations/treatment
 Contamination
1.
a.
Expectancy: The experimenter might expect something and influence the results
1) Fix this by utilizing a blind experiment.

Too Focused – control is achieved through the sacrifice of realism
-
1.
Slide 9: Design Notation
-
Slide 10: Notation
-
Notation Continued
-

Slide 11: The goal of hypothesis testing
-
In-Person Notes:
-
Lab experiments (high internal validity, lower external validity/generalizability), field experiments high in both
internal and external validity, treatment/control groups, pretest/posttest measures, random select
Inferential Statistics Video Lecture
-
Inferential Statistics: provide a means to generalize the data that you do not have
- Underlying Assumption: your sample in unbiased
 The specifics differ for each statistic
- p is the probability that your results are “real” and not by accident (show representative the sample is of
the population)
-
Distributions
- A distribution is a description of frequency of occurrences of events, for a range of events.
- Histograms, box plots, stem and leaf plot are methods for visualizing empirical distributions
- “Normal” distributions are balanced, and the mean=median=mode
-
Making Apples into Oranges
- Means and standard deviations are expressed in original units
-
Z scores are a conversion to “standard” units representing values as a function of their distance from the
normalized mean of zero
-
Z-Test
- Z tests are the first inferential statistical tests taught
- Z tests are used to compare a sample to a population with a known mean (or proportion) and variance
(standard deviation)
- Few studies use Z tests, but they are useful for introducing hypothesis testing, because the Z / Normal /
Gaussian distributions have nice properties.
 Useful to illustrate how to determine if a sample value is significantly different than the
population value
 Depends only on “value”, spread, and sample size
-
Hypothesis for T-Test
- General Null = There is NO difference between the groups (Equal to)
- General Alternative = There is a (significant) difference between groups (Not Equal to)
-
T-Test Comparing Means Assumptions
- Distributions are normal
- Variances are equal
- Random and Independent Samples
-
Hypothesis Testing
- Hypothesis testing is about rejecting the null hypothesis
- If your inferential statistics indicate that you can have confidence in rejecting the null (your p value is <.05)
your result is significant.
 a is the probability that your rejection of the null is false
 If you do reject the null hypothesis when you should not, you are making a Type 1 (a) error.
 Setting your criteria to alpha 0.01 makes it less likely you will accidently publish finding due to
chance.
- There is also a risk that you might falsely retain the null hypothesis
 THis occurs when your data analysis result are not statistically significant, but the alternative
hypothesis is true
1. This is called a Type II (B) error
Types of Error
-
-
-
Sample Size and Power
- As a sample size increase, it becomes easier to correctly reject a null hypothesis
- Power: Is the probability of correctly rejecting the null hypothesis when the alternative hypothesis is true
- If a researcher can reject the null, the he/she can conclude that the results generalize from the sample to
the population and can publish a paper
Steps to perform a T-test
-
-
fc
-
Standard Error
- Standard Error is the standard deviation of the sampling distribution
 How likely you are to be wrong
PAF 501 - Chapter 11: Case Research
Case Research
Intensively studying a phenomenon over time within its natural setting.
Strengths:
-
-
Can be theory testing or theory building
Research questions can be modified during the research process if the
original questions are found to be less relevant
Can help drive richer, more contextualized, and more authentic
interpretation of the phenomenon of interest than most other research
methods.
The phenomenon of interest can be studied from the perspectives of
multiple participants and using multiple level of analysis (individual and
organizational)
Weaknesses:
-
-
Key Decisions in Case
Research
Interpretive Case
Research
-
No experimental control, internal validity of inferences remain week.
Quality of inferences derived from case research depends heavily on the
integrative power of the researcher. (Expert notices pattens vs novice miss
them)
Because the inferences are heavily contextualized, it may be difficult to
generalize inferences from case research to other contexts or other
organizations.
Is the right method for the research questions being studied?
What is the appropriate unit of analysis for a case research study?
Should the researcher employ a single-case or multiple-case design?
What sites should be chosen for case research?
What techniques of data collection should be used in case research?
In an inductive technique where evidence collected from one or more case sites is
systematically analyzed and synthesized to allow concepts and patterns to
emerge for the purpose of building new theories or expanding existing ones.
Conducting Case
Research
-
Define research question
Select case sites
Create instruments and protocols
Select respondents
Start data collection
Conduct within-case data analysis
Conduct cross-case analysis
Build and test hypotheses
Write case research report
PAF 501 - Chapter 13: Qualitative Analysis
Qualitative Analysis
Heavily dependent on the researcher’s analytic and integrative skills and personal
knowledge of the social context where the data is collected. The emphasis in
qualitative analysis is “sense making” or understanding a phenomenon, rather
than predicting or explaining.
Grounded Theory
An inductive technique of interpreting recorded data about a social phenomenon
to build theories about that phenomenon.
Coding Techniques for
Analyzing Text Data
-
-
-
Integration
Techniques
-
Open Coding: is a process aimed at identifying concepts or key ideas that
are hidden within textual data, which are potentially related to the
phenomenon of interest.
- Categories: Tend to be broad and generalizable, and ultimately
evolve into constructs in a ground theory. Categories are needed to
reduce the amount of concepts the researcher must work with and
to build a “big picture” of the issues salient to understanding a
social phenomenon.
Axial Coding: The categories and subcategories are assembled into causal
relationship or hypotheses that can tentatively explain the phenomenon of
interest. (little spheres of coded concepts and here you are drawing bigger
circles around that).
Selective Coding: Which involves identifying a central category or a core
variable and systematically and logically relating this central category to
other categories. (Looking for central organized principle to explain the
phenomenon and linking it to other categories).
Storylining: Categories and relationships are used to explicate and/or
refine a story of the observed phenomenon.
-
-
Memoing: Is the process of using these memos to discover patterns and
relationships between categories using two-by-two tables, diagrams or
figures.
Concept Mapping: Is a graphical representation of concepts and
relationships between those concepts (Using boxes and arrows).
In-Person Notes: 9/27/21
-
Theoretical Saturation: Coding/data refinement until when additional data does not yield marginal change in core
categories/relationships.
-
Bernard (Participation Observation):
-
Bernard: (Direct / Indirect Observation): Direct (watching people on the spot) and indirect (archeology of human
behavior)
- roundedtive/nonreactive:
- Continuous monitoring/spot sampling
- Unobtrusive observations
Video Lecture Notes: Unobtrusive Methods
-
Unobtrusive Methods: Studies that do not “inconvenience” the subject matter
- Do not take them out of their element
- Do not place a burden
- Do not put subjects at risk (except perhaps a risk of loss of anonymity)
- Researchers study them in their natural environment
-
Problems with Obtrusive Studies:
- People change when they know someone is watching
-
Common Problems with Obtrusive Studies
- Reactivity
 Hawthorne Effect
1. Western Electric Company: If changing the lights in the factory improve productivity, they
found out that it did but rather than the lights causing the change it was found that
productivity increased because the observers were watching the people.
 Response set: People responding and behaving as they think the way the researchers would want
 Placebo effect
- The Observer’s Paradox
 Heisenberg’s Uncertainty Principle
1. It is impossible to measure systems (in quantum physics) without affecting those systems
2. Position and momentum
3. Schrodinger’s Cat

-
When you’re conducting researcher and directly observing them, you are changing the
environment you are studying.
Unobtrusive versus Qualitative


Observation is unobtrusive.
Unobtrusive is a mix of Qualitative and Quantitative
Ethnography
-
-
Ethology: Scientific and objective study of animal behavior (under natural conditions)
- Jane Goodall (Chimps)
- Dian Fossey (Gorillas)
Ethnography: Systematic study of people and culture (write and “people”)
- Critical versus Realist (purpose)
- You can ask questions
- Trust building
Participant Observation:
-
Complete emersion in new social setting: fly on the wall.
- Completely immersive
Observation (behavioral)
Open ended
Exploratory (not confirmatory)
Time consuming
All consuming
Chance of “going native”
- So immersed that you can’t be objective
Unstructured does not equal Undisciplined
-
Researcher must pay close attention and take copious notes
Must refrain from bringing in preconceptions
-
Must refrain from becoming too sympathetic
Unstructured does not equal Without Standards
-
Inductive does not mean anything goes:
- Good work conforms to strict standards:
 Gaining access / Rapport
1. Key informants
2. Gatekeepers
 Explicit Awareness (mindfulness)
 Building memory
 Naivety and Objectivity (value neutrality)
- Field notes
 Jotting / diary / log / formal field notes
Theoretical Approach
-
Inductive (usually)
Grounded Theory
- Also a “re” search but not for falsifying based on evidence but for gaining additional, deeper and broader
knowledge
- “Discover new worlds, boldly go where no one has gone before”
 See questions from new perspectives
 Ask different questions
- Use the new data to bootstrap the theory
 Completely inductive ^
Theoretical Approach Continued
-
-
Coding
Evaluation Standards:
Internal Validity
External Validity (Generalizability)
Reliability
Objectivity
Creditable / Richness of Data / Free from bias
Transferability / Indicator of elements of the population to know limit
Dependable / Reliable insofar as context is constant
Theoretical Orientation / Detail / Confirmability
Caution
-
Analytical Induction
- Includes a search for falsifying evidence
- Flawed approach: no example can either prove or disprove a rule
-
Fallacy of Composition (exception of composition): Thinking you can make generalized statements just because
you studied something or someone. (Individual -> Group)
Ecological Fallacy: You know on average what a group does something so you general that to an individual.
(Group -> Individual)
-
Finding out Why
-
When studying humans we cannot only investigate what they do, behavior, we can find out WHY
Researchers can ask people questions about
- Motivation
- Attitudes
- Feelings
- Methods to find out why are almost always obtrusive
Content Analysis
-
-
-
Two processes:
- Specification
- Application of rules
Level / Units of analysis
Recording Unit (word, terms, themes, characters, paragraphs, items)
- Theme is a sentence (subject / predicate)
- Items can be books, articles.
- Semiotic focuses on symbols other than “words”
Context Unit
- Context in which recording unit is found
-
What does the word mean in the sentence
Content Analysis
-
-
-
Common categories
- Who
- What
- Where
- When
- How
Manifest vs Latent
- Qualitative
- Quantitative (Manifest – counting the number of times a certain word appears)
Inductive or Deductive
Requires
- Coding (themes)
- Grouping (frames)
- Linking to theory
Archives
-
Government records
Commercial records
Any recorded artifact at any level
Data bit and papers
Level and Unit of Analysis matter
Traces:
-
Methods used to studying animals can be used to study humans
Physical traces of human activity
- Where people go (where and wear)
 Museum exhibits
 Where to pace
- What people eat
- What people waste
Conclusion:
-
Reactivity makes it difficult to study people when they are aware they are being studied
People behave differently when not in their natural environment
To find out what people REALLY do, it is important to be able to observe / record / infer what they really do
- Not just ask
- Not just assume that what we see is all there is
Bernard (RM – Participant Observation):
-
fieldwork: complete participant/participant observer/complete observer; going native; rapid assessments;
validity: impact on researcher on phenomenon (want to minimize reactivity); hanging out/rapport;
practice objectivity; complete immersion: becoming the phenomenon; objectivity does not mean value
neutrality; indigenous research: objectivity challenge; initial contact/culture shock; mid-fieldwork break
important to get distance/objectivity
Bernard (RM – Direct and Indirect Observation):
-
direct (watching people on the spot) and indirect (archeology of human behavior) observation;
reactive/nonreactive; continuous monitoring/spot sampling; unobtrusive observations; indirect
observations (trace/archival research);
Bernard (RM – Text Analysis):
-
Unobtrusive methods Lecture Slides: unobtrusive (collection methodology) but
qualitative/quantitative is the analysis technique; ethology (animal study in natural conditions) &
ethnography (systematic study of people/culture), participant observation (become part of the group and
observe from the inside and their perspective); unstructured not equal to undisciplined (notes, data) or
standards (access/mindfulness/naivety & objectivity; grounded theory (use data inductively to create
new theory); inductive grounded theory: code field notes, code content (software), text to themes/themes
to theory/validate models; evaluation standards: internal (free from bias)/external (generalizability)
validity, reliability, objectivity; caution in generalizing to larger groups; ecological fallacy: assuming
individual characteristics based on general population characteristics; fallacy of composition:
generalizing to larger groups based on smaller group characteristics; content analysis (who, what,
where, when, how): level/unit of analysis, recording unit (word, term, theme, character, paragraph,
item), manifest (quantitative, i.e. word frequencies) and latent (qualitative); archived data of any type is
unobtrusive data gathering; traces: physical traces of human activity; reactivity: difficult to study people
when they are aware of it (reactive);
Statistic Videos
RSM4 Bivariate
Objects of Measurement
-
Level of Measurement: Distinguishes between quantitative and qualitative variables (nominal, ordinal, interval,
ratio)
Unit Measurement: The specific unit such as dollar, cents, kg, or ibs, etc.
Unit of Analysis: What you are measuring, such as people, nations, households, cells, mRNA, protein, bones.
Level of Analysis: The comparison is at what level: individual, city, national.
Levels of Measurement:
-
Nominal Data:
-
Binary Data:
-
Any data can be converted into a binary variable:
- Interval ratio data -> Threshold
- Nominal -> 2 groups
- Binary “nominal” data can be used in quantitative analysis coded as 0 or 1
 This is sometimes referred to as “dummy coded”
Ordinal Data:
-
-
Order of variables is important but not the difference between variables
- Utilizes Likert Scales
 Ex. Very happy, unhappy, fine, happy, very happy
Survey creation note: The fact that we have an odd number of possible answer helps our post-hoc analyses to
interpret on the central tendency (mean, median, mode) easier/better than with an even number of answers.
Statistics Matter
-
Quality vs. Quantity
- Univariates: one variable
- Bivariate: two variables
 Correlation
 Regression
 Cross tabulations (contingency tables)
- Distributions
- Probability (Inferential Statistics
Causation Review:
-
Requires three things
- Correlation / Covariation
 Covariation:
1. When X goes up, Y goes up (positive correlation).
2. When X goes down, Y goes down (positive correlation)
3. When X goes up, Y goes down (negative correlation)
- Cause prior to effect
- Absence of a plausible rival hypothesis
Bi-Variables: Nominal or Ordinal
-
-
Cross Tabular Data
- 2-way frequencies
- Contingency tables
- Bivariate analysis of categorical data
Survey Analysis
- Quantification -> crosstabs
- Are 2 variables related?
 Does knowing the first provide you with information about the second?
Bivariates: Interval
-
Variance -> co-variance
How do you compute variance and what does it mena?
- (X – mean deviation from the mean)
Correlation Is the standardized covariances
- Covariation: Is a measure of the degree two variables vary relative to one another, but the two variables
can be at different scales
- Correlation: Provides a statistic of covariance for the “standardized” variables.
- Standardizing: Converts both variables to the same scale (Z scales)
 To standardize:
1. Divide each deviation by the variable’s standard deviation
2. Then sum and divide by n-1
Regression:
-
IF a relationship is linear (as one variables increases so does the second) it can be represented as a line.
Review
-
3 types of bivariates
- Crosstabs
 Used for nominal or ordinal
 Two-way frequency/contingency table
- Correlation (Covariation)
 Used for interval or ratio

- Regression
1. Shows not only how two variables correlate or covary but also slope. (where that zero
value sits)
 Slope
 Intercept
- PAF 501 – Bernard Chapter 9: Unstructured and Semistructured
Informal Interviewing
Categorized by a total lack of structure or control. The researcher tries to
remember conversations heard during the course of a day in the field. This
requires constant jotting and daily sessions in which you sit at a computer, typing
away, unburdening your memory, and developing field notes.
Unstructured
Interviewing
There is nothing at all informal about unstructured interviewing and nothing
deceptive, either. You sit down with another person and hold an interview. Period.
Both of you know what you’re doing and there is no shared feeling tht you’re just
engaged in pleasant chitchat.
Used when you have lots and lots of time – like when you are doing long-term
fieldwork
It is used in studies that requires on textual data and in studies that requires both
textual and numerical data.
This methodology is great for lived experiences (what was it like to survive handto-hand combat, etc)
Semistructured
Interviewing
It has much of the freewheeling quality of unstructured interviewing, and requires
all the same skills but semi-structured interviewing is based on the use of an
interview guide.
-
Interview guide: Is a written list of questions and topics that need to be
covered in a particular order.
The best method in situations where you won’t get more than one chance to
interview someone. Also, semistructured interviewing works very well in projects
where you are dealing with high-level bureaucrats and elite members of a
community.
Structured
Interviewing
People are asked to respond to as nearly identical a set of stimuli as possible.
One variety of structured interviews involves use of an interview schedule.
-
Interview Schedule: An explicit set of instructions to interviewers who
administer questionnaires orally.
Video Lecture: RM-05 Interviews
Finding Out Why
-
-
Obsevation is a great way to discover WHAT people do . . .
When studying humans we cannot only investigate what they do (behavior) we can find out WHY.
Researchers can ask people QUESTIONS about
- Motivation
- Attitudes
- Feelings
Methods to find out why are almost always obtrusive
- Although it can be a part of an unobtrusive method like participant observation
Way To Ask Questions
-
There are many techniques for asking questions of respondents to learn about their opinions, attitudes, and
behavior.
- What sets these techniques apart is how structured they are.

1.
2.
3.
All of these methods requires the researcher attempt to stay “neutral” and objective as
possible
An “interview guide” directs the topics and questions
“Interview schedules” are instructions regarding what questions to ask \
The Many Level of Measurement
-
Researchers have several options for how to ask questions varying in:
- The level of “structure”
- Whether the focus is on individuals or groups

Context
-
Validity
-
Measurement:
-
Ethics
-
-
-
For the most part asking questions is
- Non-invasive
- Obvious (so there is no deception)
So IRB’s provide Exempt or Expedited approvals
- The only real concern -> confidentiality
 This is especially true when the objective of the study is to discover information about personal,
illegal, or non-normative attitudes or behaviors
Still requires consent of respondent
Standard (expectations for good studies)
-
Asking questions well requires researchers
- Establish rapport (but not the degree of introducing bias)
- Record responses
- Do not lead, or inject yourself into, the conversation
 Listen and probe to elicit a more complete response
 Stay as neutral as possible to avoid biases responses
 Stay as objective as possible to avoid biasing the interview.
Qualitative Evaluation Standards
-
Internal, external, reliability, objectivity.
More on Threats to Validity
-
-
Respondent accuracy
- Memory (people forget)
 Techniques to jog memory are useful to reduce this
Interaction introduced biases (Reactivity . . . again)
- Characteristics or presence of the interviewer inhibit the responses
 Response effects
1. Respondent want to please the interviewer (response set bias)
2. Deference effect (respondent says what he/she think you want to hear)
3. Social desirability effect (people want to look good so they lie or exaggerate positive and
hide negatives).
 Expectancy effects (Expectations of the interviews “lead” responses}
A survey is a carefully designed set of questions. Good survey design requires the researcher get 3 things right:
-
Ask questions in a manner that does not threaten validity.
Organize the questions in a manner that does not threaten validity.
Administer the sample in a way that does not threaten validity.
Conclusion
-
When studying people asking questions can provide useful information
HOWEVER
-
Regardless of level of structure, there techniques are vulnerable to error
Respondent may not be truthful for a variety of reasons
- They do not recall or remember accurately
- They want to present an image of themselves that is not accurate
-
The more naturalistic studies are likely to be high on ecological validity, can be high or low on internal validity, but
may not be generalizable
The more structured interviews (surveys) are likely to be generalizable (if the sample was properly obtained) but
ecological validity is low.
Video Lecture: RM-05 Survey Research 1
-
Why Survery
- Contrasted with unstructured or semi-structured interviews, surveys provide a way to “standardize” asking
questions
 Mainly confirmatory (deductive)
 Rigid in structure
1. Time
2. Questions
- Easily replicated
- Easier to analyze
-
Survey Pros
- Used to collect data on the feeling, perceptions and motivations of people
- Prepared and rehearsed in advance
- Respondent options are limited
- Can focus on the topic of interest directly
- While “obtrusive” not usually too bad (e.g. questions are not usually personally invasive)
- Easy to get a good (random) sample
- If well-constructed amenable to statistical analysis
- Informed consent
-
Survey Cons
- Not as sensitive to the broad range of feeling as more opened ended approaches are.
- Limited by the preconceptions of designers
 Respondent option are limited (con as well as pro)
 Can miss important elements of a topic
Types of Surveys
-
Longitudinal
-
Panel
- Same respondents many times
Trend
- Different respondents many times
Cohort
- Trend study (different respondents) focusing on change during the life course
 Required to separate the effect of age, cohort and time.
Cohort vs. Age
-
-
Everybody is 18 once
- Is a characteristic the result of “age” or the shared life experience that occurred to the group that were
born a certain year
Liberal youth in the 1960s
- Was it their age? (Are you 18 year olds liberal?)
- Was it their cohort (Are people born in 1948 liberal when they are 18 and 50?)
- Is it a combination of age and cohort?
Video Lecture: RM-05 Survey Research 2
Why Survey?
-
-
Contrasted with unstructured or semi structured interviews surveys provide a way to “standardize” asking
questions
- Mainly confirmatory (deductive)
- Rigid in structure
 Time
 Questions
Easily replicated
Easier to analyze
Survey Challenges
-
-
All types of surveys:
- Completion rates (surveys not completed)
- Refusal rates (survey’s refused by some)
 May or may not be random (introduce bias)
- People lie / people forget / people misunderstand
 Response set
 Response style
Panel Studies have additional issues
- Attrition (People drop out)
- Pretest sensitization (people learn how to answer).
Threats to Good Surveys
-
-
Biases
- Sample bias
- Collection bias
- Respondent / Subject bias
Question wording
- Bad questions (garbage in / garbage out)
Survey order
Sample Bias
-
Sampling procedure is not representative
- Only the rich, only the sick, only the ones who return the survey, etc.
Over-sampling is sometimes important
- To get information on underrepresented groups
- Or because the researcher is only interested in that one group
Data Collection Bias
-
2 Types
- Observation Bias:
 What you ask
1. How the questions are chosen / worded / ordered
 How you collect the data (phone vs face to face / interviewer characteristics)
- Non-observation bias
 Who / what you DON’T ask = what you miss
Subject Bias
-
-
Response Style
- Measures the style of the subject rather than the construct of interest
 Some people will never give anything a 10 out of 10
 Other always exaggerate the positive
1. Such extreme answers can be corrected for with statistics.
Response Set
- Social desirability bias
- What is “acceptable” rather than what is a true response
- Answers seek to please the investigator or fall within societal norms.
2 Main Challenges
-
Challenge 1: Getting the questions right
- Be clear and unambiguous: short, simple, to the point
 Define “difficult” or ambiguous terms (don’t assume a mutual understanding)
 Avoid double barrel items (ask one question at a time
 Favor closed ended questions with mutually exclusive and exhaustive categories
- Avoid biasing answers
 Avoid leading questions
 Provide unbiased response option
 Be cautious of Order Effects
- Ask questions so that they can be quantified

Seven Deadly Sins
1) Leading words / questions
1. Example: Would you agree with responsible parents that car
seats should be required for infants?
2) No “opt out” option (prefer not to answer)
1. What are your religious beliefs? (Too personal).
3) Ambiguous: Not asking direct questions
1. What should government do to improve commute time?
4) Not giving mutually exclusive choices
5) Failing to provide an exhaustive set of options
6) Unbalanced (biased) scales
7) Double (triple) barreled questions
-
Challenge 2: Getting the question order right
-
Vetted Social Science Scale Types
-
Scales refer to “multiple indicators” to describe a construct
- Types include
 Binary
1. Capturing attitudes as binary
a. Yes / No (1 / 0)
2. Multiple indicators can be added together (summed) to create a quantitative measure
 Likert
1. A likert scale is the aggregation of many likert items
2. A likert item expand questions (items) to a 3-9 (odd number) point scale
a. Each question is composed of a “stem” with corresponding anchor
b. Anchors are coded as ascending or descending numbers
-
-
c.
3. Is a summative scale
 Semantic Differential
1. Multiple pair of polar opposites
2. Respondent is asked to select where they are on the continuum
3. Same anchor, different types of contrasts
4. Summative
 Guttman
1. Cumulative scale of tolerances
2. Assumptions: Preferences on item are ordinal
a. If you agree with on you will agree with all that follow
b. Listing of “tolerances” from least to most extreme
c. Value of first “yes” indicates attitude of tolerance
3. Example: Bogardus Social Distance scale
 Thurstone
 Bogardus
2 element scales
 Question structure
 Scaling -> Consolidation of multiple questions
This is a means for quantifying qualitative data.
Semantic Differential
-
Multiple pairs of polar opposites presented
Respondent is asked to select where they are on the continuum
Same anchor, different types of contrasts
Summative
-
Scale Validation (Thurston Scale)
-
There are many ways to validate scales, these include
- Criteria validity (comparison with a validated measure)
- Correlation with onther similar concepts
- Expect Judges
 Example: Thurstone 11 point scale construction
1. Clear conceptual definition
2. Generation of 80-100 candidate questions
3. Panel of judges rank questions 1 (poor) to 11 (good) depending on closeness to concept
4. Statistical analysis of judges ranks result in 11 point scale
4 main types of Validity
-
-
-
Face: common sense
- Does the question make sense using common sense
Content: used mainly when questions wording is the concern – subjective judgement of content
- Used mainly when question wording is the concern – need to ensure that all the words being used in a
survey (content) mean the same to all people being survery
- Deals with question design
Criterion: Tested against a standard
- Ask a new question (e.g. variation on a theme) that can be compared to a standard (other questions or
physical attribute) that has been already and deemed to be a valid standard.
Construct: Closeness to target (statistical)
- Relation among items on test
- Relation of the test to other supposed measures
 Like criterion validity, but neither measure is assumed to be a “standard”
- Relation to indictors of other construct of theory.
- Idea the cannot be measured with one variable so construct validity is a series of statistical test that
determine how close your question(s) are to the idea you are measure (sometimes referred to as
“closeness to target”).
Question Ordering
-
Not as well studied as question wording the order of questions matters as well
- Framing
- Prompting
-
Be aware of the order of questions and how the order of topics might influence how subjects respond.
10/18/21: In-Person Notes
.
10/29/21: Research Deigns Revisited
Study Design:
-
Fancy techniques in a laboratory do not constitute an experiment
What you can conclude from a “study” depends on its “design”
“Design” implies methods, which includes, sampling, group set-up, required techniques/strategies, and analyses.
Causation
-
X -> Y
- Three factors to demonstrate causation
 Co-variation
 Cause prior to effect
 Absence of a plausible rival hypothesis.
Random Control Trials
-
-
One of the ways to demonstrate causality
- Random: Assignment/distribution of samples you are testing
- Control: Interventions/treatments
- Trials: Repeat trials (not just an n=1 for each group)
Elements of the RTC
- Random Assignment of 9 into groups
 Control vs. Treatment (Experimental Group)
- Pre-test: Pre-treatment measurements between groups
 Both groups should be indistinguishable
 Ensures cause (treatment) precedes the effect (outcome)
 Provide time-specific evidence that the effect occurred after the treatment
 No difference between pretest groups is further evidence that nothing else differs between the
groups.
- Explicit treatment given to only one group
 The introduction of a treatment, controlling for everything else, ensures that the treatment is the
reason for the difference between groups
- Post-test: Measurable difference of outcome / effect between treatment and control groups following the
treatment.
True Experimental Design (RCT)
-
Group 1: No Y -> Treatment X -> Y
Group 2: No Y -> _____________ -> No Y
Components of an Experiment
-
R = Random Assignment
X = Exposure to a Treatment or Event
O = Observation
Column = Time ordered from left to right; two events in the same column happen at the same time
Row = Experimental Unit or Subject
Experimental Designs: Post-Tests
-
-
Random Control Trial: (O = O . . . Null Hypothesis) (O does not equal O . . . Alternative Hypothesis)
- ROXO
- RO_O
One Shot (Least experimental design)
- XO
Designs: Main Types
-
-
-
One Shot: Correlation. Measures dependent variable. (post-hoc analysis)
- Not control
- Not treatment
- No pre-test
One Group: Look for changes within each subject
- Manipulate dependent variable
- Pre-test / post-test
- Advantages:
 Control of event
 O measured before and after event
- Disadvantages:
 O may have occurred anyway
 Something other than X may have occurred to generate O
Static Group: Look for change between difference subjects or groups
- The difference between individuals or groups might be the result of something other than what is being
studied (plausible rival hypothesis).
- Advantages:
 Comparison of the presence and absence of X
- Disadvantages:
-
 Differences not measured may have accounted for O
 Differential recruitment or mortality)
Experimental (RCT).
- Advantages:
 Control, comparison, pretest/posttest, random assignment
 Called the Solomon 4
Quasi-Experiment (NEGD) (many policy questions use non-equilivinant group designs (quasi))
-
NOXO
NO_O
Interrupted Time Series
-
OXO...OX...O
Downside to RTC
-
-
Experimenter expectancy
- Self-fulfilling prophesy
 Solved by having a blind experimenter
- Demand characteristics (response set)
 Treatment group demonstrates and effect that is not real (placebo)
 Control group is impacted by experiment in unanticipated ways
1. Compensation
2. Exaggerated contamination
RTC focuses on “control” at the expense of realism in many ways
- IT is always important to ask if A is
 Necessary or
 Sufficient to cause B
10/29/21: Figures
-
Figures
- Highlight and must be referred to by the text. They are useful if . . .
 Only a picture can tell the story
 There is too much information to include in a table
 What is subtle in a table is obvious in a picture
- Things to consider to include:
 Volume of data you are presenting
 Time intervals
 Qualitative or Quantitative data
 Comparisons
 Raw vs. summarized
- In the end ask yourself
 Does the table/graph help the reader?
 Does it help you (the writer)
-
Things to consider for your figure
- Do no duplicate the information from a Table
- Size . . . make sure it fits well in the paper
- Color
10/29/21: Table
-
Tables
- Highlight and must be referred to by the test. They are useful if:
 There is too much information to include in a paragraph
 There are too many dimensions to a concept to describe clearly in a paragraph
- Thing to consider include:
 Volume of data you are presenting
 Time intervals or continuous (depends on your experiment)
 Space
 Qualitative or Quantitative
 Comparisons
 Raw vs. summarized
In the City Lights and American Red Cross Nutrition Behavior Study researchers recognized a shared issue with
their sample. What was the main, common concern?
-
Baseline (N = 1461)
Posttest (N=1031)
Another name for a Pretest-Posttest Comparison Group Design is:
- Both RTC and Non-equivalent group design can be described as pretest-posttest
What is the main difference between an experiment and a quasi-experiment?
-
Doesn’t sampling fall under assignment procedures? So it could be either or.
Hey Professor,
I hope you’re having a good weekend so far!
I’m reaching out this morning to hopefully get some clarification on a few questions I missed on this week’s
quiz.
1) Another name for a Pre-test-Posttest Comparison Group Design is:
- My Answer: Random Control Trial
- Correct Answer: Non-equivalent Group Design
- I was a little confused here because both the RCT and Non-equivalent Group Design utilize the pre-testposttest method.
2) What is the main difference between an experiment and a quasi-experiment?
- My Answer: Sampling
- Correct Answer: Assignment Procedures
- I was going back and forth between both sampling and assignment procedures because in my mind
sampling and assignment procedures fall underneath the same bucket of describing the subject not being
randomly selected.
11/1/21 : Module 7
11/15/21:
Evidence leading up to a policy recommendation
Policy issue/problem, this is how I organized it, these are the solutions <- Put that all in executive summary.
Important element(s)
Conclusion: Here’s what we talked about and here’s why all these matter
Download