Uploaded by Hiba Salameh

final exam 2

advertisement
Week4 Research Designs
Research Design
What is a research design?
A research design provides a framework for the collection and analysis of data.
A choice of research design reflects decisions about the priority being given to
a range of dimensions of the research process. These include the importance
attached to:
 expressing causal connections between variables;
 Generalizing to larger groups of individuals than those actually forming
part of the investigation;
 understanding behavior and the meaning of that behaviour in its specific
social context;
 having a temporal (that is, over time) appreciation of social phenomena
and their interconnections.
Research Method
What is a research method?
A research method is simply a technique for collecting data. It can involve a
specific instrument, such as a self-completion questionnaire or a structured
interview schedule, or participant observation whereby the researcher listens
to and watches others.
Criteria in social research
Reliability–are measures consistent?
Replication/replicability–is study repeatable?
Validity–are conclusions well-founded
Types of validity
Measurement (or construct) validity–do measures reflect concepts?
Internal validity–are causal relations between variables real?
External validity–can results be generalized beyond the research setting?
Ecological validity–are findings applicable to everyday life?
Alternative criteria in qualitative research
Trustworthiness (Lincoln and Guba (1985):
Credibility, parallels internal validity -i.e. how believable are the findings?
Transferability, parallels external validity -i.e. do the findings apply to other
contexts?
Dependability, parallels reliability -i.e. are the findings likely to apply at other
times?
Confirmability, parallels objectivity -i.e. has the investigator allowed his or her
values to intrude to a high degree?
Variable
What is a variable?
A variable is simply an attribute on which cases vary. 'Cases' can obviously
be people, but they can also include things such as households, cities,
organizations, schools, and nations.

If an attribute does not vary, it is a constant.

If all manufacturing organizations had the same ratio of male to female
managers, this attribute of such organizations would be a constant and not a
variable.

Constants are rarely of interest to social researchers.

It is common to distinguish between different types of variable.

The most basic distinction is between independent variables and
dependent variables. The former are deemed to have a causal influence on the
latter.

In addition, it is important to distinguish between variables-whether
independent or dependent-in terms of their measurement properties.

This is an important issue in the context of quantitative data analysis

Types of research design
1. Experimental
2. Cross-sectional
3. Longitudinal
4. Case study
5. Comparative
Experimental design elements
1.
2.
3.
4.
5.
Random assignment of subjects to experimental and control groups
Pre-testing of both groups
Independent variable manipulated; all other variables held constant
Post-testing of both groups
Computation and analysis of group differences
Threats to validity
1. Other (non-experimental) events may have caused the changes observed
(‘history’)
2. Subjects may become sensitized to ‘testing’
3. People change over time in any event (‘maturation’)
4. Non-random ‘selection’ could explain differences
5. ‘Ambiguity about the direction of causal influence’ because sometimes the
temporal sequence is unclear
Cross-sectional design
“A cross-sectional design entails the collection of data on more than one case
(usually quite a lot more than one) and at a single point in time in order to
collect a body of quantitative or quantifiable data in connection with two or
more variables (usually many more than two), which are then examined to
detect patterns of association.”




A sample of cases
At a single point in time
Qualitative or quantitative data
Patterns of association
Evaluating cross-sectional research
 Reliability and Measurement Validity are not connected to the design as
such
 Replicability will be high as long as the researcher specifies all the
procedures
 Internal Validity is weak, because co-relations are much more likely to be
found than causality
 External Validity will be strong if the sample is truly random
 Ecological Validity may be compromised by the instruments used
Longitudinal design
 Survey of the same sample that is repeated on more than one occasion
over time
 Typically used to map change in social research
 In panel study(e.g. TUIK IncomeandLivingConditionsPanel Survey)
 Or a cohort study(e.g. NCDS –National Child Development Study –sample
of children born in 1958)
 Surveys like the Crime Survey for England and Wales, and the British Social
Attitudes Survey should not be considered longitudinal as they do not involve
the same people –they are better thought of as repeated cross-sectional
designs
Evaluating longitudinal research
 Attrition, because people die, or move home, or withdraw from the study.
 Knowing when is the right time for the next wave of data collection.
 The first round may have been badly thought out or not originally
designed as longitudinal, which leaves the later rounds in a bit of a mess.
 A panel conditioning effect may creep into the research
Case study design
 Detailed and intensive analysis of one case e.g. a single community, school,
family, person, event, or organization
 Often involves qualitative research, but not necessarily so
 Approach is usually idiographic

Types of case: critical, unique, exemplifying, revelatory, longitudinal e.g.
Hold away(1982, 1983): an ethnography of occupational culture in a particular
police force (Overall, the thesis demonstrates the continuing dominance of
lower police ranks who maintain freedom from legal and organizational
constraints to practice police work as they define it)
Evaluating case-study research
 The biggest issue concerns external validity because it is impossible to
generalize the findings.
 Many case-writers argue, though, that the point of the research is to
examine particulars rather than attempt to generalize. Instead, the quality of
theoretical reasoning is given precedence.
 Cases may be extended longitudinally or through a comparative design.
Comparative design
 Using the same methods to compare two or more meaningfully contrasting
cases
 Can be qualitative or quantitative
 Often cross-cultural comparisons Gallie’s(1978) study of the impact of
automation on industrial workers in England and France (It shows that national
traditions of industrial relations were more important than technology in
explaining worker attitudes and management–worker relations
 Problem of translating research instruments and finding comparable
samples Includes multiple case studies
Evaluating comparative design
 The comparative design is essentially two or more cross-sectional studies
carried out at the same point in time, so the problems are very similar.
 Familiar issues of reliability, validity, and replicability then apply.
 Observed differences between cases need to be treated with caution –and
further research –as they might not be attributable to the distinguishing
features of the case. Causality can not be assumed.
However, comparing two or more cases can show circumstances in which a
particular theory will or will not hold.
Week5
The Nature of Quantitative Research
What is a concept?
Concepts are:



Building blocks of theory
Labels that we give to elements of the social world
Categories for the organization of ideas and observations
Concepts are useful for:



Providing an explanation of a certain aspect of the social world
Standing for things we want to explain
Giving a basis for measuring variation
Why measure?
To delineate fine differences between people, organizations, or any other
unit of analysis

To provide a consistent device for gauging distinctions

To produce precise estimates of the degree of the relationship between
concepts

Indicators
What is an indicator?
It is worth making two distinctions here. First, there is a distinction between an
indicator and a measure. The latter can be taken to refer to things that can be
relatively unambiguously counted, such as personal income, household
income, age, number of children, or number of years spent at school.
Measures, in other words, are quantities. If we are interested in some of the
causes of variation in personal income, the latter can be quantified in a
reasonably direct way. We use indicators to tap concepts that are less directly
quantifiable. If we are interested in the causes of variation in job satisfaction,
we will need indicators that will stand for the concept. These indicators will
allow job satisfaction to be measured, and we can treat the resulting
quantitative information as if it were a measure. An indicator, then, is
something that is devised or already exists and that is employed as though it
were a measure of a concept. It is viewed as an indirect measure of a concept,
like job satisfaction. We see here a second distinction between direct and
indirect indicators of concepts. Indicators
A Multiple-Indicator measure of a concept
A multiple-indicator measure of a concept
The research on the effects of redundancy by Westergaard et al. (1989), which
was referred to in Chapters 2 and 3, was conducted by structured interview
with 378 steel workers who had been made redundant. One of the authors'
interests was whether their respondents' commitment to work varied
according to whether they were still unemployed at the time of the interview
or had found work or had retired. In order to measure commitment to
employment, the authors gave their respondents ten statements and asked
them to indicate their level of agreement or disagreement on a seven-point
scale running from 'Yes, I strongly agree' to 'No, I strongly disagree. There was
a middle point on the scale that allowed for a neutral response. This approach
to investigating a cluster of attitudes is known as a Likert scale, though in many
cases researchers use a five-point rather than a seven-point scale for
responses. See Key concept 7.2 for a description of what a Likert scale entails.
The ten statements were as follows.
A Multiple-Indicator measure of a concept
1. Work is necessary but rarely enjoyable.
2. Having a job is not very important to me.
3. I regard time spent at work as time taken away from the things I want to do.
4. Having a job is/was important to me only because it brings in money. 5. Even
if I won a great deal of money on the pools I'd carry on working.
6. If unemployment benefits were really high, I would still prefer to work. 7. I
would hate being on the dole.
8. I would soon get bored if I did not go out to work.
9. The most important things that have happened to me involved work.
10. Any feelings I've had in the past of achieving something worthwhile have
usually come through things I've done at work.
The authors found that their respondents' replies did not differ a great deal in
terms of whether they had found work since being made redundant or were
still unemployed or had taken retirement.
Likert scale
What is a Likert scale?
The investigation of attitudes is a prominent area in much survey research. One
of the most common techniques for conducting such an investigation is the
Likert scale, named after Rensis Likert, who developed the method. The Likert
scale is essentially a multiple-indicator or multiple-item measure of a set of
attitudes relating to a particular area. The goal of the Likert scale is to measure
intensity of feelings about the area in question. In its most common format, it
comprises a series of statements (known as 'items') that focus on a certain
issue or theme. Each respondent is then asked to indicate his or her level of
agreement with the statement. Usually, the format for indicating level of
agreement is a five-point scale going from 'strongly agree' to 'strongly
disagree', but seven-point scale and other formats are used too. There is
usually a middle position of 'neither agree nor disagree' or 'undecided'
indicating neutrality on the issue. Each respondent's reply on each item is
scored, and then the scores for each item are aggregated to form an overall
score. Normally, since the scale measures intensity, the scoring is carried out so
that a high level of intensity of feelings in connection with each indicator
receives a high score (for example, on a five-point scale, a score of 5 for very
strong positive feelings about an issue and a score of 1 for very negative
feelings). The measure of commitment to work referred to in Research in focus
7.2 is an example of a Likert scale. Variations on the typical format of indicating
degrees of agreement are scales referring to frequency (for example, 'never'
through to 'always') and evaluation (for example, 'very poor' through to 'very
good").
There are several points to bear in mind about the construction of a Likert
scale. The following are particularly important.
• The items must be statements and not questions.
• The items must all relate to the same object (job, organization, ethnic groups,
unemployment, sentencing of offenders, etc.).
• The items that make up the scale should be interrelated (see the discussion
of internal reliability in this chapter and Key concept 7.3).
It is useful to vary the phrasing so that some items imply a positive view of the
phenomenon of interest and others a negative one. Thus, in the example in
Research in focus 7.2, some items imply a negative view of work (for example,
'Having a job is not very important to me') and others a positive view of work
(for example, "I would soon get bored if I did not go out to work'). This
variation is advised in order to identify respondents who exhibit response sets
(see the sections on 'Response sets' in Chapters 9 and 10).
Reliability
What is reliability?
Reliability refers to the consistency of a measure of a concept. The following
are three prominent factors involved when considering whether a measure is
reliable:
• Stability. This consideration entails asking whether a measure is stable over
time, so that we can be confident that the results relating to that measure for a
sample of respondents do not fluctuate. This means that, if we administer a
measure to a group and then readminister it, there will be little variation over
time in the results obtained. In February 2010, the then Shadow Home
Secretary, Chris Grayling, was roundly criticized by the UK Statistics Authority
for comparing Home Office statistics from the late 1990s with current figures
to suggest that there had been a big increase in violent crimes since Labour
took office in 1997. The reason for the criticism was that there had been a
change to the definition of violent crime, which had produced an immediate 35
per cent increase in the crime. In this case, the measure of violent crime was
not reliable from the point of view of inferring a change over time. For this
story, see 'Chris Grayling Accused of Damaging Public Trust over Crime Figures',
www.thetimes.co.uk/tto/news/politics/article2030815.ece (accessed 9 August
2010).
• Internal reliability. The key issue is whether the indicators that make up the
scale or index are consistent- in other words, whether respondents' scores on
any one indicator tend to be related to their scores on the other indicators.
Inter-observer consistency. When a great deal of subjective judgement is
involved in such activities as the recording of observations or the translation of
data into categories and where more than one 'observer' is involved in such
activities, there is the possibility that there is a lack of consistency in their
decisions. This can arise in a number of contexts, for example: in content
analysis where decisions have to be made about how to categorize media
items; when answers to open questions have to be categorized; or in
structured observation when observers have to decide how to classify subjects'
behaviour.
What does validity mean?
Does the indicator measure the concept?
It does if it has
Face validity (right for the concept?)

Concurrent validity(supported by a relevant criterion today?)

Predictive validity(likely to be supported by a relevant criterion
tomorrow?)

Construct validity(are useful hypotheses produced?)

Convergent validity(supported by results from other methods?

Assessing the internal reliability and the concurrent and predictive validity of a
measure of organizational climate
1. human relations model: feelings of belonging and trust in the organization
and the degree to which there is training, good communication, and
supervisory support;
2. internal process model: the degree of emphasis on formal rules and on
traditional ways of doing things;
3. open systems model: the extent to which flexibility and innovativeness are
valued;
4. rational goal model: the degree to which clearly defined objectives and the
norms and values associated with efficiency, quality, and high performance are
emphasized.
A factor analysis (see Key concept 7.6) was conducted to explore the extent to
which there were distinct groupings of items that tended to go together. This
procedure yielded seventeen scales, such as autonomy, involvement,
innovation and flexibility, and clarity of organizational goals.
The internal reliability of the scales was assessed using Cronbach's alpha,
showing that all scales were at a level of 0.73 or above. This suggests that the
measure's constituent scales were internally reliable.
Concurrent validity was assessed following semi-structured interviews, with
each company's managers in connection with their organization's practices.
The interview data were coded to provide criteria against which the validity of
the scales could be gauged. In most cases, the scales were found to be
concurrently valid. For example, the researcher examined the correlation
between a scale designed to measure the emphasis on tradition and the
degree to which practices associated with the 'new manufacturing paradigm'
(Patterson et al. 2005: 397) were adopted, as revealed by the interview data.
The correlation was -0.42, implying that those firms
that were perceived as rooted in tradition tended to be less likely to adopt new
manufacturing practices. Here the adoption of new manufacturing practices
was treated as a criterion to assess the extent to which the scale measuring
perceptions of tradition really was addressing tradition. If the correlation had
been small or had been positive, the concurrent validity of the scale would
have been in doubt.
To assess predictive validity, the researchers asked a senior key informant at
each company to complete a questionnaire one year after the main survey had
been conducted. The questionnaire was meant to address two of the
measure's constituent scales, one of which was the innovation and flexibility
scale. It asked the informants to assess their company in terms of its
innovativeness in a number of areas. For example, the correlation between the
innovation and flexibility scale and informants' assessments of their companies
in terms of innovativeness with respect to products achieved a correlation of
0.53. This implies that there was indeed a correlation between perceptions of
innovativeness and flexibility and a subsequent indicator of innovativeness.
The main preoccupation of quantitative researchers
Measurement
Causality



Explanation of why things are the way they are
Direction of causal influence between DV&IV
Confidence in the researcher's causal inferences
Sampling
Basic terms and concepts: 1





Population: the universe of units from which the sample is to be selected
Sample: the segment of population that is selected for investigation
Sampling frame: list of all units
Representative sample: a sample that reflects the population accurately
Sample bias: distortion in the representativeness of the sample
Basic terms and concepts: 2
Probability sample:sample selected using random selection

Non-probability sample:sample selected not using random selection
method

Sampling error: difference between sample and population

Non-sampling error:findings of research into difference between sample
and population

Non-response: when members of sample are unable or refuse to take part

Census:data collected from entire population

Sampling error
Difference between sample and population

Biased sample does not represent population

some groups are over-represented; others are under-represented

Sources of bias

non-probability sampling, inadequate sample frame, non-response

Probability sampling reduces sampling error and allows for inferential
statistics

4 types of probability sample




Simple random sample
Systematic sample
Stratified random sample
Multi-stage cluster sample
Simple random sampling





Each unit has an equal probability of selection
Sampling fraction: n/N
where n = sample size and N = population size
List all units and number them consecutively
Use random numbers table to select units
Systematic sampling
Select units directly from sampling frame

From a random starting point, choose every nth unit (e.g. every 4th name)

Make sure sampling frame has no inherent ordering –if it has, rearrange it
to remove bias

Stratified random sampling
Starting point is to categorize population into ‘strata’ (relevant divisions, or
departments of companies for example)

So the sample can be proportionately representative of each stratum

Then, randomly select within each category as for a simple random sample

Multi-stage cluster sampling
Useful for widely dispersed populations

First, divide population into groups (clusters) of units, like geographic
areas, or industries, for example

Sub-clusters (sub-groups) can then be sampled from these clusters, if
appropriate

Now randomly select units from each (sub)cluster

Collect data from each cluster of units, consecutively

Qualities of a probability sample
Representative -allows for generalization from sample to population

Inferential statistical tests

Sample means can be used to estimate population means

Standard error (SE): estimate of discrepancy between sample mean and
population mean

95% of sample means fall between +/-1.96 SE from population mean

Sample size




Absolute size matters more than relative size
The larger the sample, the more precise and representative it is likely to be
As sample size increases, sampling error decreases
Important to be honest about the limitations of your sample
Factors affecting sample size: 1
Time and cost

After a certain point (n=1000), increasing sample size produces less
noticeable gains in precision

very large samples are decreasingly cost-efficient (Hazelrigg, 2004)

Non-response

response rate = % of sample who agree to participate (or % who provide
usable data)

responders and non-responders may differ on a crucial variable

Factors affecting sample size: 2
Heterogeneity of the population

the more varied the population is, the larger the sample will have to be
 Kind of analysis to be carried out

Some techniques require a large sample (e.g. contingency table; inferential
statistics)

Types of non-probability sampling: 1
1. Convenience sampling



the most easily accessible individuals
useful when piloting a research instrument
may be a chance to collect data that is too good to miss
2. Snowball sampling


researcher makes initial contact with a small group
these respondents introduce others in their network
Week6
Data Collection Techniques-I
What is a structured interview?
 Useful tool of quantitative research
 Often used in social surveys
 Standardized interview schedule
 Each interviewee gets the same questions, in the same way, in the same
order
 Closed, pre-coded or fixed choice questions
 Minimizes variation between interviews
Advantages of structured interviewing: 1

Reduces error due to interviewer variability:

differences in responses are due to ‘true variation’, not inconsistencies in
the conduct of interviews
potential sources of error are reduced by standardization (question
wording, memory, misunderstanding)

reduces intra-interviewer and inter-interviewer variability

greater validity of data?

Advantages of structured interviewing: 2

Accuracy and ease of data processing:

closed ended, pre-coded or fixed choice questions (limited choice of
possible answers)

interviewer does not interpret responses before recording them

standardized coding frame reduces variability in coding procedure

reduces intra-coder and inter-coder variability
Conducting structured interviews: 1


Know your way around the schedule
Introduce the research

Spoken or written rationale

identify yourself, your employer, purposes of research and procedure of
interview

ethical issues: anonymity, confidentiality, right to withdraw

opportunity for interviewee to ask questions
Conducting structured interviews: 2
Building rapport
can be difficult if limited time and little opportunity for discussion (closed
questions)

Asking questions

keep to the schedule: even small variations in wording can affect responses
Recording answers
write exact words used by interviewee, or use fixed choice questions
Conducting structured interviews: 3
Clear instruction


some questions are not relevant to every interviewee
filter questions help interviewer navigate the schedule
Question order





every interviewee must get questions in the same order
general questions before specific questions
earlier questions may affect salience of later ones
first questions should be directly related to the topic
potentially embarrassing or sensitive questions towards the end
Conducting structured interviews: 4
Probing
when respondent does not understand question or gives insufficient
answer

non-directive probes: “mmm”, “can you say a bit more about that?”

repeat fixed choice alternatives

Prompting


interviewer suggests possible answers
show cards
Leaving the interview
Training and supervision

necessary if researcher hires interviewer(s)
Problems with structured interviewing: 1
Characteristics of interviewers


gender, age, ethnicity, class (rapport)
can evoke socially desirable responses
Response sets
people may respond in consistent but irrelevant ways, by acquiescence
(agreeing or disagreeing to all questions) or for reasons of social desirability
(interviewees reflect on the way their answers might be perceived)

What is a self-administered questionnaire?






Also called a self-completion questionnaire
No interviewer present
Respondent writes answers on form
Returned to researcher or deposited for collection
Usually postal questionnaires
Can be distributed in person or by email
Self-completion questionnaires have advantages



Cheaper and quicker to administer (to widely dispersed populations)
No interviewer variability
Convenience for respondents
but many disadvantages



Cannot probe or prompt
Can only ask salient questions
Few open-ended or complex questions
Respondent can see the whole questionnaire before answering (question
order effects)

Cannot ensure that the ‘right’ person answers

Cannot collect additional data

Respondent fatigue if too many questions

Excludes people with limited literacy skills

Greater risk of missing data

Lower response rates

Online social surveys
There has been a huge growth in online administration of surveys. They come
in different forms
Email Surveys
Embedded –in the body of an email –using x or deleting alternatives, typing
in answers etc.

Easy to manage, higher response rate than attached survey

Attached –as an email attachment and the email explains it –again
respondents type answers.

Web surveys
-Respondents are invited to visit a website where the questionnaire can be
completed.

It can be widely embellished in appearance, filter questions can be used,
answers can be programmed to download into a database eliminating the
coding aspect

Week7
Data Collection Techniques-II
Asking Questions-Open questions
Advantages



Respondents answer in their own terms
Allow for new, unexpected responses
Exploratory -generate fixed answer questions
Disadvantages




Time-consuming for interviewer and respondent
Difficult to code
More effort required from respondent
Interviewer variation in recording answers
Asking Questions-Closed questions
Advantages
Quicker and easier to complete (better response rate and less missing
data)

Easy to process data (pre-coded)

Easy to compare answers (inter-coder reliability)

Disadvantages



Restrictive range of answers: no spontaneity
Difficult to make fixed choice answers exhaustive
Respondents may interpret questions differently
Designing questions: general rules



Remember your research questions
Decide exactly what you want to find out
Imagine yourself as a respondent


How would you answer the questions?
Identify any vague or misleading questions
Things to avoid…..
Ambiguous terms: ‘often’, ‘regularly’, ‘frequently’

Long questions

Double-barreled questions: may be different answers to each part(e.g.
How satisfied are you with the pay and conditions in your job?)

Very general questions: because they lack a frame of referance

Leading questions: hinting at a preferred response

Asking two questions in one

Negative terms: ‘not’, ‘never’-especially double negatives

Technical terms:(jargon and acronyms)


Using existing questions







Common practice in survey research
Questions have already been piloted
Known properties of reliability and validity
Helps you to draw comparisons with other studies
‘Question banks
’Repositories of questions used in previous surveys
ConsultTURKSTAT (https://www.tuik.gov.tr/)
What is structured observation?
A method of systematically observing people’s behavior

By grouping behavior into categories

Direct observation, not reports or inference, make it an alternative to
survey methods

Aggregates and compares behavior of everyone in the sample


Uses a cross-sectional research design
The observation schedule
Must have a clear focus and be easy to use

Categories must be inclusive (cover all options) and mutually exclusive

It must be easy to operate

Specifies categories of behaviour to be observed and how to allocate
behaviour to a category

Clear guidelines needed so that observers can distinguish between
behaviour categories

A pilot study is useful to iron out any problems

Strategies for Observing Behaviour & Sampling












Record incidents, events, and interventions
Observe for short periods of time, repeatedly
Observe for long periods of time, continuously
Time sampling, record whatever is happening every x minute
Ad libitum sampling
Record whatever is happening at the time
Focal sampling (most common form)
Observe a specific individual for a set period of time
Scan sampling Record
behaviour of everyone in the group at regular time intervals
Behavior sampling
Observe whole group to see who was involved in a particular behaviour
What is secondary analysis?




Primary data is data we collect ourselves and Secondary data is that collected by others.
Secondary analysis is done on secondary data In other words, someone else gathered the data
for their own purposes
and then we analyse it for our own purposes
Advantages of secondary analysis
• Saves money and time
• Offers high quality data
• Gives an opportunity for longitudinal analysis
• Allows subgroup or subset analysis
• Gives an opportunity for cross-cultural studies
• Allows more time for data analysis
• Enables the application of recent theory to old data
• Gets more value from the original data
…but there is a down-side…
• You need to become familiar with how the data was collected, coded and managed
• The data can be very large and complex
• The quality of the data should never be taken for granted
• Variables important to your analysis might be missing
Official statistics
• Collected by agencies of the state, in the course of their business – e.g. TURKSTAT compiles data for the
level of unemployment
• Advantages over quantitative data from surveys
– reduced time and cost
– no problem of reactivity
– cross-sectional and longitudinal analysis
– cross-cultural analysis
• Disadvantages of official statistics
- Only reveal ‘tip of the iceberg’: the ‘dark figure’ of unrecorded events e.g. unemployed people who do
not claim benefits are not officially listed as unemployed
- The process used for data collection needs interpretation
Week 7 Quantitative Data Analysis – I
Descriptive and Inferential Statistics Two branches of statistics:
• Descriptive statistics
– Graphical and numerical procedures to summarize and process data
• Inferential statistics
– Using data to make predictions, forecasts, and estimates to assist decision making
Descriptive Statistic
Collect data – e.g., Survey
• Present data – e.g., Tables and graphs
• Summarize data – e.g., Sample mean =
Inferential Statistics
Estimation  e.g., Estimate the population mean weight using the sample mean weight  Hypothesis
testing  e.g., Test the claim that the population mean weight is 140 pounds
Inference is the process of drawing conclusions or making decisions about a population based on sample
results
Classification of Variables
Measurement Levels
Graphical Presentation of Data
• Data in raw form are usually not easy to use for decision making
• Some type of organization is needed
• Table
• Graph
• The type of graph to use depends on the variable being summarized
Graphical Presentation of Data
Univariate analysis (analysis of one variable at a time)
Frequency tables
– Number of people or cases in each category
– Often expressed as percentages of sample – Interval/ratio data needs to be grouped
• Diagrams – Bar chart or pie chart (nominal or ordinal variables) – Histogram (interval/ratio variables)
Describing Data Numerically


The arithmetic mean (mean) is the most common measure of central tendency
Median Not affected by extreme values
Mode
• A measure of central tendency
• Value that occurs most often
• There may may be no mode
• There may be several modes
Measures of Variability
1.
Measures of variation give information on the spread or variability of the data values.
Population Variance
• Average of squared deviations of values from the mean
Sample Variance
• Average (approximately) of squared deviations of values from the mean
Population Standard Deviation
• Most commonly used measure of variation
• Shows variation about the mean
• Has the same units as the original data
A
i
FINAL
Week 9 Quantitative Data Analysis – II
Bivariate analysis (analysis of two variables at a time)




Explores relationships between variables
Searches for co-variance and correlations
Cannot establish causality
Can sometimes infer the direction of a causal relationship
o
If one variable is obviously independent of the other
Contingency tables

Connects the frequencies of two variables

Helps you identify any patterns of association
Pearson’s r : relationship between two interval/ratio variables






Coefficient shows the strength and direction of the relationship
Lies between -1 (perfect negative relationship) and +1 (perfect positive relationship)
Relationships must be linear for the method to work, so, plot a scatter diagram first
Coefficient of determination
found by squaring the value of r
Shows how much of the variation in one variable is due to the other variable?
Multivariate analysis(three or more variables)


The relationship between two variables might be spurious
Each variable could be related to a separate, third variable



There might be an intervening variable
A third variable might be moderating the relationship
e.g. correlation between age and exercise could be moderated by gender
Statistical significance

How confident can we be that the findings from a sample can be generalized to the population as a
whole?

How risky is it to make this inference?

Only applies to probability samples
Testing procedure for statistical significance
1.
Set up a null hypothesis -suggesting no relationship between examined variables in the population
from which the sample was drawn
2.
Decide on an acceptable level of statistical significance
3.
Use a statistical test
4.
If acceptable level attained
Reject null hypothesis
If acceptable level not attained
Accept it
but we might be wrong to accept or reject the null hypothesis
Independent Samples t-test

Independent samples t-test is used to decide whether the mean differences of the same variable
from two different independent samples are statistically significant or not.
Correlation and Significance


How confident can we be about a relationship between two variables?
Whether a correlation coefficient is statistically significant depends on:
o
the size of the coefficient (the higher the better)
o
the size of the sample (the larger the better)
Result and Interpretation:

The relationship between communication quality and job satisfaction is tested through correlation
analysis. It shows that there is a statistically significant and positive relationship between communication
and job satisfaction(r (155) = ,73; p < ,01). H1 is supported as a result of correlation test. The variables
move together, meaning while better communication quality result in higher job satisfaction scores.
Comparing means

Statistical significance of relationship between two variables means

Total variation in dependent variable: error variance (variation within subgroups of IV)

explained variance (variation between subgroups of IV)

F statistic
expresses amount of explained variance in relation to amount of error variance
Regression Analysis

Regression analysis is used when you want to predict a continuous dependent variable from a
number of independent variables.

The independent variables used in regression can be either continuous or dichotomous.

Regression analysis is usually used with naturally-occurring variables, as opposed to experimentally
manipulated variables
Simple Regression Analysis

Simple linear regression is when you want to predict values of one variable, given values of another
variable.
Week11 Data Sources in Qualitative Research:
Observation–Interview–Focus Groups-Documents
What is ethnography?

Researcher immersed in the social setting for an extended period.

Participant observation, often with interviews and documentary methods.

Participant observation is usually used with respect to the method of investigation only.

Understanding the culture, norms and values of the group.

It is both a research method and the written product of research.
Gaining Access & Key informants

Gaining access to the field requires permission and authorization in advance for the researcher to be
present.

Use friends, colleagues and contacts to help you.

Be clear and honest about your aims and methods.

Need time to gain trust of wary participants.

Key informants are sponsors and/or gatekeepers who help ethnographer to collect data. They
provide advice on playing the role of insider.
Different roles of observers
Taking field notes

Write down notes as quickly as possible.

Write up full field notes at the very latest at the end of the day.

Use a tape recorder if you prefer.

Notes must be vivid and clear -you should not have to ask at a later date ‘what did I mean by that?’

You need to take copious notes, so, if in doubt, write it down.
Types of field notes

Mental notes–particularly useful when it is inappropriate to be seen taking notes

Jotted notes(also called Scratch notes) –very brief notes written down on pieces of paper or in
small notebooks to jog one's memory about events that should be written up later

Full field notes–as soon as possible make detailed notes, which will be your main data source

Methodological notes –notes devoted to methodological decisions, that help to document the
process of research
Key Points
Interviews-Differences between structured and qualitative interviews

Qualitative interviews…are less structured/standardized,

take the participant’s viewpoint,

encourage ‘rambling’ off the topic,

are more flexible,

seek rich, detailed answers,

aim to understand rather than to generalize.
Preparing an interview guide

Have a logical but flexible order of topics.

Focus on research questions: ‘what do I need to know about?’

Use comprehensible and relevant language.

Avoid leading questions.

Record ‘facesheet’ information (name, age, gender, position etc.).
Kinds of questions

Introducing (“Tell me about…”)

Follow-up (“What do you mean by that?”)

Probing (“Could you tell me more about…?)

Specifying, (“What happened next?”)

Direct (“Do you think that…?”)

Indirect (“What do most people think about…?”)

Structuring (“Let’s move on to…”)

Silence

Interpreting (“Do you mean that…?”)
Recording and transcription

Audio-recording and transcribing:Researcher is not distracted by note-taking.

Can focus on listening and interpreting.






Corrects limitations of memory and intuitive glosses
Detailed and accurate record of interviewee’s account.
Opens data to public scrutiny.
Good quality digital recorders are now widely available.
Transcriber or transcription software?
Selective transcription saves time.
Getting it taped and transcribed: an illustration of two problem
Online interviewing

Online personal interviews for qualitative research

Textual in nature: email exchanges, direct messaging, forums.

Asynchronous/synchronous.

Delivery of questions/answers made one-at-a-time, small batches, all-at-once?

Editing –issues of reliability.

‘Spamming’ –issues of validity.
Participant observation vs Interviews









Advantages of participant observation over qualitative interviewing
Seeing through others’ eyes
Learning the native language
Taken for granted ideas more likely to be revealed
Access to deviant or hidden activities
Sensitivity to context of action
Flexibility in encountering the unexpected
Naturalistic emphasis
Embodied nature of the experience









Advantages of qualitative interviewing over participant observation
Finding out about issues resistant to observation
Interviewees reflect on past events / life course
More ethically defensible
Fewer reactive effects
Less intrusive
Longitudinal research (follow-up interviews)
Greater breadth of coverage
Specific focus
What is a focus group?






Form of group interview
Several participants and a moderator
Individuals in a group/social context
Discussion of specific issue
Study interaction between group members
How are opinions expressed and modified through group discussions?
Uses of Focus Groups & Recording

To examine the way people (collectively) construct and organize knowledge.

To understand why people hold (and retain) certain views. Participants probe and challenge each
other.

To elicit a wide range of views.

Keep a tape/digital record rather than take notes of how the topic was discussed(nuances of
language, collective defining etc)

Need to study not only what people say but who says what
How many groups?

More than one but not too many

Continue until theoretical saturation point reached

Socio-demographic characteristics

Use stratifying criteria (age, gender, etc)

Larger number of groups needed to represent a diverse range of viewpoints

Running more groups increases volume of data and complexity of analysis
Size of focus groups

Average 6-10 members per group

Over-recruit in anticipation of ‘no-shows’

Use smaller sized groups when:

The topics are sensitive or controversial

Each person will have plenty to say

You want to glean personal, detailed accounts

Use larger groups when:

You want to hear numerous brief suggestions
Level of moderator involvement

Try to be unobtrusive and non-directive

Ask small number of general questions to stimulate discussion

Allow a ‘free rein’ to participants

Discover what they define as important

Intervene if the discussion wanders ‘off track’ or there is a long silence

Respond to potentially interesting points that are not picked up by the participants
Selecting participants

Usually on the basis of a shared experience or characteristic

Socio-demographic factors -stratifying criteria

Look for systematic variation between groups

Strangers versus ‘natural groups

’Which makes it easier to discuss the topic?

Natural groups use taken for granted assumptions
Limitations of focus groups

Researcher has less control over proceedings

Data are difficult to analyse

Produce large volume of data

Need to analyse forthemes in what people say as well as patterns of interaction

Difficult to organize / risk of no-shows

Very time-consuming to transcribe

Tendency for more agreement than disagreement in discussions

Potential for causing participant discomfort

Longitudinal research (follow-up interviews)

Greater breadth of coverage
Documents available for research
These are ‘documents’ that…..

can be read (but not just text –visuals as well!)

have not been produced specifically for the purposes of research

are preserved so that they become available for analysis

are relevant to the concerns of the social researcher

Types of the documents Personal-Official State documents-Official company documents-Mass
media outputs-Virtual documents
Personal documents: diaries, letters, autobiographies, photographs

Could be the primary source of data within a qualitative study

Could be used as adjuncts to other methods, such as interviews or participant observation

Perhaps to trace the history of an organization

Diaries may also be used as a method of data collection –in turn, available to other researchers
Official documents
State documents:

From the state, in the form of legislation, reports of public enquiries

Can be a very time-consuming form of research

Question of credibility because the original report may be biased
Official company documents
In the public domain:

annual reports

mission statements

reports to shareholders

transcripts of chief executives’ speeches

press releases, advertisements

public relations material in printed form and on the internet
Not in the public domain:

company newsletters

organizational charts

external consultancy reports

minutes of meetings

memos

internal and external correspondence

manuals for new recruits

policy statements

company regulations
Mass media outputs




Newspapers, magazines, radio, TV, film
Problem with authenticity when authorship unclear
Problem with credibility because of the possibilities of image distortion
Much contextual awareness is often required
The reality of documents




Documents exist in their own right
We should examine them in terms of their context and intended readership
Inter-textuality:documents are linked to other documents
Researchers are usually more interested in the content of documents but must keep the document’s
original purpose in mind
Week12 Qualitative Data Analysis
What is content analysis?



Approach to analysis of documents and texts
Quantifies content in terms of predetermined categories
Systematic and reliable




Different from semiotics and ethnographic content analysis
Most commonly focus on mass media communications
Decide which dimensions of texts to quantify
E.g. news coverage of an issue: who, what, how much, where and why?
Selecting a sample

If working with the mass media (most likely case), two dimensions are possible for random
sampling:

The media itself






Which types of text?
Printed or visual data? Documents? Mass media?
If mass media, which kind?
TV, radio, newspapers, magazines…
More than one type?
For each type of text, which examples? E.g. tabloid or broadsheet newspapers


Dates of publication
Starting dates may be predetermined by an historical event ( like the terrorist attack on the World
Center on 11 September 2001 )

More open if the study is an going, general phenomenon

End dates can be a matter of judgment
What is to be counted?

Significant actors: protagonists and alternative voices

Words: frequency of words or phrases (e.g. ‘hooligan’ or ‘non-smoker’); can use computer software
for analysis

Subjects and themes

Dispositions: values, bias and ideology
The coding schedule

A tabular form onto which coded data will be entered

Each column represents a dimension to be analysed

Each row represents a unit of analysis (item of text)

Codes are written into blank cells in table

E.g. Crime reporting in a local newspaper
Portions of a coding manual
Advantages of content analysis

Transparency

Ease of longitudinal analysis

Flexibility

Ease of access
Disadvantages of content analysis

Questions of authenticity, credibility and representativeness of documents

Interpretation by coders

Inability to answer `why?’ questions

Has an a theoretical approach
Qualitative data analysis
1.
General strategies:

Analytic induction

Grounded theory
2.


Coding:
Steps, considerations, problems
Thematic analysis
Analytic induction

A rigorous search for universal explanation of phenomena:
1. Rough definition of research question
2. Hypothetical explanation
3. Data collection (examination of cases)
4. If any deviant cases found, redefine or
reformulate hypothesis
5. Continue until all cases fit hypothesis
Grounded theory

Theory is derived from the data, which are systematically
gathered and analysed

Iterative process

Repetitive interplay between data collection and analysis / theory building

Developments in grounded theory

Straussian model more prescriptive
Considerations in developing codes

Of what general category is this item of data an instance?

What does this item of data represent?

What is this item of data about?

Of what topic is this item of data an instance?

What question about a topic does this item of data suggest?

What sort of answer to a question about a topic does this item of data imply?

What is happening here?

What are people doing?


What do people say they are doing?
What kind of event is going on?
Steps and considerations in coding
1. Code as soon as possible
2. Read through your initial set of transcripts
3. Do it again!!
4. Review your codes
5. Consider more general theoretical ideas in relation to codes and data
6. Any one item or slice of data can and often should be coded in more than one way
7. Do not worry about generating what seem to be too many codes
8. Keep coding in perspective –it is not analysis
Problems with coding

Losing the context of what was said (extracting sections of data)

Fragmentation of data –loss of narrative flow

Narrative analysis as solution

Risk of only providing descriptive account of data rather than theorizing
Thematic analysis

One of the most common approaches to qualitative data analysis

Not an approach to analysis that has an identifiable heritage or that has been outlined in terms of a
distinctive cluster of techniques e.g. Disney study
Narrative analysis

Concerned with how people ‘make sense’ and emphasises the storied nature of human recounting
of lives and events Elicited personal narratives

Life history / biographical approach

Secondary data

Explores narrative accounts produced in interviews and other forms of data (i.e. documents)
Narrative analysis of extent material

Deliberately asking certain kinds of question that tend to elicit a narrative
Week13 Research Ethics and Referencing
What is “ethical” research?

No harm should come to research participants

They should agree to participate and know what the research is about

Their privacy should not be invaded

They should not be lied to or cheated
How could you harm research participants?






Physically
By damaging their development or self-esteem
By causing stress
By hurting their career prospects or employment opportunities
By breaking confidentiality
By revealing their identity
Invasion of privacy


Privacy is very much linked to the notion of informed consent
The research participant does not abrogate the right to privacy entirely by providing informed
consent

Covert methods are usually deemed to be violations of the privacy principle

The issue of privacy is invariably linked to issues of anonymity and confidentiality in the research
process
Informed consent
Lies, deception and research

Deception usually means we represent our research as something other than it is, so that
participants will respond more naturally

This means it is quite a widespread practice

So deception means not telling the whole truth, while not actually telling a lie

We must be vigilant in keeping deception to a minimum, and when it is necessary to the research,
mitigating its degree and effects as much as possible

Apart from moral objections to deception, research participants and funders can become wary of
being fooled, or tricked into providing data
What does your own institution require?

Most colleges and universities have developed their own guidelines for ethical research. Some of
the questions they might pose are as follows:
1. Is the study funded?
2. Is the research compromised by the source of funding?
3. Are there potential conflicts of interest in the financial or organisational arrangements?
4. Will confidentiality be maintained appropriately at all stages of enquiry?
5. Will human rights and dignities be actively respected?
6. Will highly personal, intimate, or other private or confidential information be sought?
7. Will there be any harm, discomfort, physical, or psychological risks?
8. Will participants be involved whose ability to give informed voluntary consent may be limited?
9. Will the study involve obtaining or processing personal data relating to living individuals?
Citation–APA style

APA Style uses the author–date citation system, in which a brief in-text citation directs readers to a
full reference list entry. The in-text citation appears within the body of the paper (or in a table, figure,
footnote, or appendix) and briefly identifies the cited work by its author and date of publication. This
enables readers to locate the corresponding entry in the alphabetical reference list at the end of the
paper.

Each work cited must appear in the reference list, and each work in the reference list must be cited
in the text (or in a table, figure, footnote, or appendix).

Both paraphrases and quotations require citations.
Appropriate Level of Citation

The number of sources you cite in your paper depends on the purpose of your work. For most
papers, cite one or two of the most representative sources for each key point. Literature review papers,
however, typically include a more exhaustive list of references.

Provide appropriate credit to the source (e.g., by using an in-text citation) whenever you do the
following:
 paraphrase (i.e., state in your own words) the ideas of others
 directly quote the words of others
 refer to data or data sets
Plagiarism

Plagiarism is the act of presenting the words, ideas, or images of another as your own; it denies
authors or creators of content the credit they are due. Whether deliberate or unintentional, plagiarism
violates ethical standards in scholarship.

Writers who plagiarize disrespect the efforts of original authors by failing to acknowledge their
contributions, stifle further research by preventing readers from tracing ideas back to their original
sources, and unfairly disregard those who exerted the effort to complete their own work.
Paraphrasing

A paraphrase restates another’s idea (or your own previously published idea) in your own words.
Paraphrasing allows you to summarize and synthesize information from one or more sources, focus on
significant information, and compare and contrast relevant details.
Quotations

A direct quotation reproduces words verbatim from another work or from your own previously
published work. It is best to paraphrase sources rather than directly quoting them because paraphrasing
allows you to fit material to the context of your paper and writing style.

Use direct quotations rather than paraphrasing:
 when reproducing an exact definition
 when an author has said something memorably or succinctly
 when you want to respond to exact wording
Basic Principles of Reference List Entries

A reference list entry generally has four elements: the author, date, title, and source. Each element
answers a question:
 author: Who is responsible for this work?
 date: When was this work published?
 title: What is this work called?
 source: Where can I retrieve this work?

Consistency in reference formatting allows readers to understand the types of works you consulted
and the important reference elements with ease.
Journal article
Whole book
Whole edited book
Chapter in an edited book
Aron, L., Botella, M., & Lubart, T. (2019). Culinary arts: Talent and their development. In R. F. Subotnik, P.
Olszewski-Kubilius, & F. C. Worrell (Eds.),The psychology of high performance: Developing human
potential into domain-specific talent(pp. 345–359). American Psychological
Association.https://doi.org/10.1037/0000120-016
Dillard, J. P. (2020). Currents in the study of persuasion. In M. B. Oliver, A. A. Raney, & J. Bryant
(Eds.),Media effects: Advances in theory and research(4th ed., pp. 115–129). Routledge.


Parenthetical citations: (Aron et al., 2019; Dillard, 2020)
Narrative citations: Aron et al. (2019), Dillard (2020))
Unpublished Dissertation or Thesis References
Report by a Government Agency References
Week14 Writing Up the Thesis& Abstract



Issues in writing up social research
Do quantitative and qualitative research reveal divergent approaches?
 The similarities are frequently more striking and apparent than the differences
Many people find writing up research more difficult than carrying it out
 You must convince your readers about the credibility of the knowledge claims you make
 Good writing is to do with developing your style so that it is persuasive and convincing
Typical ways of constructing an argument
Checklist for writing up your research: 1.

Have you clearly specified your research questions?

Have you clearly indicated how the literature you have read relates to your research questions?

Is your discussion of the literature critical and organized so that it is not just a summary of what you
have read?

Have you clearly outlined your research design and your research methods?
Checklist for writing up your research: 2.

Have you presented your data in a manner that relates to your research questions?

Does your discussion of your findings show how they relate to your research questions?

And on how they shed light on the literature that you presented?

Are the interpretations of the data that you offer fully supported with tables, figures, or segments
from transcripts?
Checklist for writing up your research: 3.

Does your writing avoid sexist, racist, and disablist language?

Have you included all appendices that you need (interview schedule, communications with research
participants)?

Have you checked that your list of references includes all the items referred to in your text?

And that it follows precisely the style that your institution requires?
Checklist for writing up your research: 4.

Have you ensured that your institution’s requirements for submitting projects are fully met in terms
of word length (neither too long nor too short) and whether an abstract and table of contents are
required?

Have you ensured that you do not quote excessively when presenting the literature?

Have you fully acknowledged the work of others so that you cannot be accused of plagiarism?
Download