Uploaded by N.F.Heukelom

Stuvia-1136765-summary-introduction-to-statistics-including-examples-and-pictures

advertisement
Summary Introduction to Statistics including examples and pictures
written by
VdeBresser
www.stuvia.com
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Statistics
Questions worth doing again: p.96 Q5, p.129 Q17, p.191 Q25, p.222 Q17, p.265 Q15, alle
opdrachten week 9
 best way to make sense of great loads of data
= a set of mathematical procedures for organizing, summarizing, and interpreting
information
Chapter 1
Descriptive statistics
Statistical procedures used to summarize, organize, and simplify data
- About describing the data
- Through summary statistics (organized in a table or computing the average)
- The wealthiest 1% own 50% of the goods
Inferential statistics
Consist of techniques that allow us to study samples and then make generalizations about
the populations from which they were selected
- To make an inference from something to something else
 make a5513n inference from a sample to the population
(= Select a few from your population and test these, then make an inference from
these few to the population)
Parameter: a value (usually numerical) that describes a population, it is usually derived from
measurements of the individuals in the population
Statistic: a value (usually numerical) that describes a sample, it is usually derived from
measurements of the individuals in the sample
Data differs
- Height (in cm)
- Annual income (EUR)
- Smoker vs. non-smoker  no value/ number
- Pet (dog, cat, hamster, bunny)  no value/ number
- How much you support Trump (from -5 to +5)  scale
Research methods
Correlational method: two different variables are observed to determine whether there is a
relationship between them
- Relationship between wake up time and academic performance in students
- Scores are often of categorical, not numerical, value
- Shows a correlation, but never a cause-and-effect relationship
Statistics used for comparing two (or more) groups of scores: examining descriptive statistics
that summarize and describe the scores in each group, and we use inferential statistics to
determine whether differences between the groups can be generalized to the entire
population
- Calculating averages and comparing them
- Computing proportions and comparing them
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Experimental method: one variable is manipulated (independent variable) while another
variable is observed and measured (dependent variable). To establish a cause-and-effect
relationship between the two variables, an experiment attempts to control all other
variables to prevent them from influencing the results
- Participant variables (= age, gender, intelligence) and environmental variables (=
lightning, time of day, weather) must not differ between groups
 random assignment, matching (each group has equal males and females), and
holding variables constant (only using boys)
- Using two groups, control group and experimental group
Control group gets no treatment, or neutral/placebo treatment
Experimental group gets the experimental treatment
Non-experimental methods
- Nonequivalent groups design: groups are already divided; the researcher has no
ability to control
- Pre-post design: the researcher has no control over other variables that change with
time, like the light and weather
- Quasi-independent variable: the independent variable that is used to create the
different groups of scores. It cannot be manipulated by the researcher
Dimensions of data
Constructs vs operationalizations
Construct: internal attributes or characteristics that cannot be directly observed but are
useful for describing and explaining behavior
- Interest in relationship between construct A and B
 needs operationalizations
- Interest in relationship between intelligence and uni performance
 operationalization intelligence: IQ test
 operationalization uni performance: Average grade
Discrete vs continuous variables
Discrete variables:
Variables that only consist of a limited number of categories
- Gender, eye color, native language, how many pets/ siblings
 there cannot be a value between two neighboring categories
 there cannot be a value between 1 and 2 pets
Continuous variables:
Variables that can take all values between two points
- Income, height, weight, speed
- Height can be 1,76 m, 176 cm, or 1,751356344 m
 it is very rare to obtain identical measurements for two different individuals, there
is an infinite number of possible values
 a continuous variable is actually an interval, when two people claim to weigh 150
pounds, they do not weigh exactly the same but are both around 150 pounds. Thus
150 is not a specific point on the scale but an interval. To differentiate 150 from 149
there need to be specific boundaries (real limits) and are placed exactly halfway
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
between adjacent scores  X=150 is between 149,5 (lower real limit) and 150,5
(upper real limit)
The nominal scale
Named categories (dog, cat, hamster)
- No quantitative distinction between them  you cannot say ‘a dog is more than a
cat’/ room 109 is not 9 bigger than room 100
- No true zero
The ordinal scale
Consists of a set of categories that are organized in an ordered sequence
Ranked named categories (1st, 2nd, 3rd)
- No equal distance between ranks (1st place was much faster than 2nd and 3rd)
- No true zero
The interval scale
Consists of ordered categories of equally sized intervals between values
 Each unit has the same size
- Temperature (from 21C to 26C, or from 1C to 6C)
 both have the same difference
- No true zero (arbitrarily chosen, can have the value 0)
0 degrees does not mean that there is no temperature, it can even go lower
The ratio scale
Consists of equally sized intervals between values
 each unit has the same size
- There is an absolute zero!!!
- Distance: a distance of 0 means your bike has not moved
- 0 gallons means that your tank is empty
Statistical notations
X = raw scores  when having two scores Y is also indicated as raw scores
N = number of scores in a population
n = number of scores in a sample
 = summation  10,6,4,7 X=27 and N=4
- The  is always followed by a symbol or mathematical expression, which identifies
exactly which values are to be added
Order of mathematical operations
1. Any calculation within parentheses
2. Squaring (or raising other exponents)
3. Multiplying and/or dividing
4. Summation
5. Any other addition and/or subtraction
Chapter 2
Representing data
Frequency distribution
Organized tabulation of the number of individuals located in each category on the scale of
measurement
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
-
Places the disorganized scores in order from highest to lowest, grouping together
individuals who all have the same score
How many pets do you have?
- Ask 10 people the number of pets that currently live in their main
household
- Construct: number of pets
Operationalization: number of pets that currently live in a
person’s main household
Add structure
- Count how often each option occurs
 how many people have 0, 1, 2… pets
 Frequencies of values
- by adding up the frequencies, you obtain the total number of
individuals  f=N
The structured table is then called a frequency distribution table
Grouped frequency distributions
Bundle some value ranges together = class intervals
- 10 intervals is the general guide
- The width should be a relatively simple number (2,5,10,20)
- Low (0-25000)
Middle (25001-50000)
Upper middle (50001-75000)
High (75000+)
- Frequency distributions for continuous variables
 because we cannot put them together in normal frequency
distributions
Higher sample number
Pet example: instead of n=10  n=10000
- Frequency goes up
- The large dataset differs much from the small data set
 proportion: p = f/n
- All proportions add up to 1!!!
 percentage: p = f/n * 100
- All percentages add up to 100%
Frequency distribution graphs
Graphs for interval or ratio data
Histograms
- Numerical scores along the x-axis and bars above each X
value in which the height corresponds to the frequency for
that category on the y-axis.
- Ability to compare n=10 to n=10000 when put into
proportions
- No spaces between the bars
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Polygons
- Numerical scores along the x-axis, then a dot is centered above each score so that
the vertical position of the dot corresponds to the frequency for the category, then a
continuous line is drawn from dot to dot
- With class intervals the dot is drawn directly above the midpoint of the class interval
In a class interval of 20-29, the dot is drawn above 24,5
Graphs for nominal or ordinal data
Bar graph
- the same as a histogram, except that spaces are left between adjacent bars
 separate distinct categories and the categories are not all the same size (the bars
are)
Graphs for population distributions
- it is impossible to obtain an exact count of the number of people in a large
population. A frequency distribution can be constructed when they have relative
frequencies and smooth curves
relative frequencies
you don’t know the exact score but based on past census data and general trends you know
certain numbers are close (like how many males and females there are) so you draw the
outnumbering category slightly above the latter category
smooth curves
you are not connecting a series of dots, but you are showing the relative changes that occur
from one score to the next
- normal curve
the shape of a frequency distribution
frequency distribution graphs all have the same three characteristics
1. shape
2. central tendency (= where the center of the distribution is located)
3. variability (= the degree to which the scores are spread over a wide range or are
clustered together)
symmetrical distribution: possibility to draw a vertical line through the middle so that one
side of the distribution is a mirror image of the other
skewed distribution: the scores tend to pile up toward one end of the scale and taper off
gradually at the other end
tail of the distribution: the section where the scores taper off toward one end of a
distribution
positively/negatively skewed: positively = tail on the right
side because the tail points toward the positive end of
the x-axis, negatively = tail on the left side
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Locating data points
We may want to find where a value lies relative to the whole data
- Are 3 pets a lot or not?
- (percentile) rank: rank of a score that is defines as the percentage of individuals in
the distribution with scores at or below a particular value
 locate points based on the frequency distribution
Percentiles
Score that is identified
1. Sort the frequency table (X, f, prop, perc)
2. We calculate a cumulative percentage
c%= (cf/N) x100
Option 0 and 1 cover 60,48% of the data
3. We locate our data point of interest (having 3 pets)
3 pets correspond to a cumulative percentage of
95,17%
‘3 pets’ has a percentile rank of 95,17%
‘3 pet’s are the 95th percentile
Interpolation
A method of finding intermediate values
1. A single interval is measured on two separate scales, the endpoints of the interval are
known for each scale
2. You are given an intermediate value on one of the scales, the problem is to find the
corresponding intermediate value on the other scale
3. The interpolation process requires 4 steps
a. Find the width of the interval on both scales
b. Locate the position of the intermediate value in the interval, this position
corresponds to a fraction of the whole interval
Fraction = distance from the top of the interval/ interval width
c. Use the same fraction to determine the corresponding position on the other
scale, first use the fraction to determine the distance from the top of the interval
Distance = (fraction) x (width)
d. Use the distance from the top to determine the position on the other scale
Question 1: find the percentile rank for X = 9
Table:
X
C%
10-14
75%
5-9
60%
0-4
10%
Step 1: find the width of the interval on both scales, this is the top of interval 5-9  9,5 and
the top of interval 0-4  4.5
9 falls between these intervals
top
9.5
60%
9
?
bottom
4.5
10%
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Step 2: Locate the position of the intermediate value in the interval, this position
corresponds to a fraction of the whole interval
Fraction = distance from the top of the interval/ interval width
This is the top (9.5) – the value (9) = 0.5
To get the fraction you do 0.5 (the distance from top of the interval) / 5 (total interval width
 9.5-4.5 = 5) = 0.1
Step 3: Use the same fraction to determine the corresponding position on the other scale,
first use the fraction to determine the distance from the top of the interval
Distance = (fraction) x (width)
The distance between both intervals is 50 (60% - 10%)
To calculate the distance from 9 to 9.5 but in percentages  0.1 x 50 = 5
Step 4: Use the distance from the top to determine the position on the other scale
60% - 5 = 55 %
 so the percentage of X=9 is 55%
Stem and leaf display
A way to represent data in a different way than the
frequency table
1. Take all first numbers that are unique (1,2,3,4,5…)
2. Take the second numbers and place them behind the
line
3. You can flip it so you can see it as bars (you see that
most answers are in the 30’s)
Chapter 3
Central tendency
A statistical measure to determine a single score that defines the center of a distribution.
The goal of central tendency is to find the single score that is most typical or most
representative of the entire group (average or typical individual)
- Characterizes what is typical for a large population and in doing so makes large
amounts of data more digestible
It is difficult to have one way of defining a center in different distributions because they all
look different
 solution: the mean, median and the mode
The mean
The arithmetic average  computed by adding all the scores in the distribution and dividing
them by the number of scores
- The mean of a population is defined by the μ  μ = ΣX/N
The mean of a sample is defined by M  M = ΣX / n
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Alternative definitions
- The mean as the amount each individual receives when the total is divided equally
among all the individuals
- The mean as a balance point for the distribution. When you have 7 points below the
mean and 7 points above the mean, the mean is the balance point.
The weighted mean
When combining two sets of scores you need to
find the overall mean for the combined group,
you need:
1. The overall sum of the scores for the
combined group
2. The total number of scores in the
combined group (n)
 M = ΣX1 + ΣX2 / n1 + n2
Characteristics of the mean
Every score in the distribution contributes to the value of the mean
- Changing the value of any score will change the mean
- Adding or removing a score will change the mean, unless the score is located exactly
at the mean
-
If a constant value is added/removed to every score in a distribution, the same
constant will be added/removed from the mean
If every score in a distribution is multiplied/divided by a constant value, the mean will
change in the same way
The median
The goal is to locate the midpoint of the distribution
If the scores in a distribution are listed in order from smallest to largest, the median is the
midpoint of the list. It is the point on the measurement scale below which 50% of the scores
in the distribution are located
- No special symbols or notation, just median
To calculate the median, you need to put all scores from smallest to largest, you start with
the smallest and count the scores as you move up the list. The median is the first point you
reach that is greater than 50% of the scores in the distribution
- With this set of scores: 3, 5, 8, 10, 11  the median is 8
- With this set of scores: 3, 5, 8, 10  the median is 5 + 8 / 2 = 6.5
Finding the precise median for a continuous variable
A continuous variable consists of categories that can be split into an infinite number of
fractional parts. It is not the same as the discrete variables shown above. When
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
encountering a set of scores like 1,2,3,4,4,4,4,6 the median of continuous variables cannot
be 4.  X=4 corresponds to an interval from 3.5-4.5
to find the precise median, we first observe that the distribution contains 8 scores,
represented by 8 boxed in the graph. to make it equal that both sides of the median contain
4 boxes you need to split the 4  ¼ belongs with 1,2,3 and ¾ belongs with 6
- Fraction = number needed to reach 50% / number in the interval
Fraction = 1/4 = 0.25
 median is 3,75
 You can also use interpolation by trying to find a score of 50%
The mode
In a frequency distribution, the mode is the score or category that has the greatest
frequency, it is used to determine the typical or most frequent value of any scale of
measurement
- No special symbols or notation, just median
- Can be used when using category’s like restaurants
In a questionnaire 42 out of 100 students chose Luigi’s to be their favorite restaurant,
this was the most frequently chosen restaurant  the mode is Luigi’s
- In a distribution there can only be one mean, and one median, but there can be
multiple modes  when multiple scores have the same frequency
You call this a bimodal or multimodal distribution
- The mode is technically the score with the highest frequency, but when two scores
are close to each other and are both peaks they are both modes  bimodal
distribution with a major and minor mode
Selecting a measure of central tendency
The goal is to find the single value that best represents the entire distribution  using the
mean is most common
When to use the median
- Extreme scores or skewed distributions
Unable to use the mean because one or two extreme values can have a large
influence and cause the mean to be displaced  these outliers can be spotted by
taking context into account
- Undetermined values (when participants never finish the experiment)
Unable to use the mean because one value is unknown
- Open-ended distributions = when there is no limit (number of pizza’s eaten with
largest category ‘5 or more’)
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Unable to use the mean because you do not have an exact total score
- Ordinal scale, it allows you to determine direction but do not allow you to determine
distance  median is the best way, this shows half of the scores below and half of
the scores above the median
When to use the mode
- Nominal scales: the categories that make up a nominal scale are differentiated only
by name, they do not measure quantity
Unable to use the mean or median
- Discrete variables: exist only in whole, indivisible categories
Unable to use the mean because it would calculate categories that do not exist (1,6
children and 5,3 rooms instead of when using the mode 2 children and 5 rooms)
- Describing shape: it does not need calculation so it can be used as an indication of
the mean or median
Using graphs to present means and medians
A graph allows several means (or medians) to be shown simultaneously so it is possible to
make quick comparisons between groups or treatment conditions
- X-axis: groups and treatment conditions
- Y-axis: values for the dependent variable
- The means (or medians) are displayed using a line graph, histogram or bar graph
Mean, median and modes in distributions
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Chapter 4
Variability: provides a quantitative measure of the differences between scores in a
distribution and describes the degree to which the scores are spread out or clustered
together
The purpose is to obtain and objective measure of how the scores are spread out in a
distribution  describe the data
A good measure of variability has:
1. It describes a distribution; it tells whether the scores are clustered together or are
spread out over a large distance  defined in terms of distance
2. It measured how well an individual score represents an entire distribution it
provides information about how much error to expect if you are using a sample to
represent a population
 to measure variability you have the range, standard deviation and variance
The range
The distance covered by the scores in a distribution
- Calculated by measuring the difference between the largest score and the smallest
score  Range = Xmax – Xmin
- Values 1-5 have a range of 4
Values ¼, ½, ¾ have a range of 1  between 0 and 1
- When the measurements are continuous values the range is defined by the upper
real limit for the largest score, and the lower real limit for the smallest score
Values 1-5 have a range of 5,5 – 0,5 = 5
- when the scores are all whole numbers the range is  Range = Xmax – Xmin + 1
the range is rarely used because it does not consider all scores in the distribution, outliers
make a whole other range than no outliers  unreliable
Standard deviation and variance
Most commonly used and most important variability measure. It used the mean of the
distribution as a reference point and measures variability by considering the distance
between each score and the mean
Step 0: calculate the mean
Step 1: determine the deviation/distance from the mean, for each individual score
- deviation score = X – μ
with a μ = 50 and a score X = 53  deviation is 3
- when adding up all deviation scores in a distribution you should arrive at 0, because
the total of distances above the mean are equal to the total of
distances below the mean
- the mean of the deviations is also always 0
Step 2: because the mean of the deviation is always 0, you need to get
rid of all the – /+ signs  square each deviation score, add them up
and calculate the mean
- compute the mean squared deviation  variance
- variance = Σ (X – μ)2 / N
Step 3: to get rid of the squared deviation you need to take the root
- standard deviation: the square root of the variance and provides a measure of the
standard, or average distance from the mean  expressed by σ
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
-
standard deviation = √ variance
Standard deviation and variance in a population
The concepts of standard deviation are the same for both samples and populations, however
the calculations slightly differ
Variance = mean squared deviation = sum of squared deviations (= SS)/ number of scores
To compute the SS (= sum of squares) you need two formulas
1. The definition formula: SS = Σ (X – μ)2
first, find each deviation score, then square each
deviations core and then add the squared
deviations
2. The computational formula: SS = ΣX2 – (ΣX)2 / N
First, square each score and then add the values,
then find the sum of the scores and square this
total, divide this by N, and finally subtract the
second part from the first
- Used when the mean is not a whole number, the deviations all contain decimals or
fractions, and the calculations become difficult
Population variance:
Represented by the symbol σ2 and equals the mean squared distance from the mean,
population variance is obtained by dividing the sum of squared by N
- Variance (σ2) = SS/N
Population standard deviation
Represented by the symbol σ and equals the square root of the population variance
- Standard deviation (σ) = √ SS/√ N
Standard deviation and variance in a sample
A few extreme scores in a population can make the variability quite large, however the
chance that these scores are included in a sample are very small, which makes the sample
variability relatively small  biased estimate of population variability
Calculations to fix this problem
- Calculating the sum of squared deviations is the same, except the notations
Definition formula: SS = Σ (X – μ)2  SS = Σ (X – M)2
Computational formula: SS = ΣX2 – (ΣX)2 / N  SS = ΣX2 – (ΣX)2 / n
- Calculating the sample variance and standard deviation is different
Sample variance
Represented by the symbol s2 and equals the mean squared distance from the mean, sample
variance is obtained by dividing the sum of squared by n – 1
- Variance (s2) = SS / n – 1
Sample standard deviation
represented by the symbol s and equal the square root of the sample variance
- Standard deviation (s) = √ SS/ √ n-1
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Degrees of freedom
For a sample of n scores the degrees of freedom (df) for the sample variance are defined as
df = n – 1. The degrees of freedom determine the number of scores in the sample that are
independent and free to vary.
Sample variance as an unbiased statistic
Unbiased: a sample statistic is unbiased if the average value of the statistic is equal to the
population parameter (the average value of the statistic is obtained from all the possible
samples for a specific sample size, n)
- divided by n-1
Biased: a sample statistic is biased if the average value of the statistic either underestimates
or overestimates the corresponding population parameter  correct this by using the sum
of squares
- divided by n
Transformations of scale
Occasionally a set of scores is transformed by adding a constant to each score or by
multiplying each score by a constant value. The easiest way to determine the effect of a
transformation is to remember that the standard deviation is a measure of distance. If you
select any two scores and see what happens to the distance between them, you will also find
out what happens to the standard deviation
1. Adding a constant to each score does not change the standard deviation, the
distance between two scores will stay the same. It does change the mean, which will
also increase or decrease by the same number
2. Multiplying each score by a constant causes the standard deviation to be multiplied
by the same constant, the distance between the scores will also multiply by the same
amount, as does the mean
Chapter 5
The individual scores within a distribution
z-score / standard score: identify and describe the exact location of each score in a
distribution
- A score by itself does not necessarily provide much information about its position
within a distribution
- Raw scores: original, unchanged scores that are the direct result of measurement
The process of transforming X values into z-scores serves the purpose of:
1. Each z-score tells the exact location of the original X value within the distribution
2. The z-scores form a standardized distribution that can be directly compared to other
distributions that also have been transformed into z-scores
The z-score
Location within a distribution
The z score accomplishes to describe the exact location by transforming each X value into a
signed number (+ or -)  the sign tells whether the score is located above or below the
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
mean, and the number tells the distance between the score and the mean in terms of the
number of standard deviations
- In a distribution with μ=100, σ=15, in a score of X=130  z= +2,00
- Formula: z = (X – μ) / σ
 X – μ is a deviation score, it measures the distance in points between X and μ and
indicates whether X is located above or below the mean
Determining raw scores from a z-score
In general, it is easier to use the definition of a z-score, rather than a formula, when you are
changing z-scores into X values. But you can use a formula:
- X = μ + (z x σ)
- μ = 60, σ = 8, which value of X corresponds to z = -1,50
determine the distance corresponding 1,5 standard deviations  1,5 x 8 = 12
then find the value of X located below the mean by 12 points  60 – 12 = 48
Standardized distribution
a composition of scores that have been transformed to create predetermined values for μ
and σ, standardized distributions are used to make dissimilar distributions comparable
You can transform every value of X to its corresponding z-score, this new distribution of
z-scores has characteristics that make the z-score transformation a very useful tool:
1. Shape: the distribution of z-scores will have exactly the same shape as the original
distribution of scores
2. The mean: the z-score distribution will always have a mean of zero, which makes the
mean a convenient reference point
3. The standard deviation: the distribution of z-scores will always have a standard
deviation of 1, the numerical value of a z-score is exactly the same as the number of
standard deviations from the mean
Many people find z-score distributions not nice because they contain negative values and
decimals, so it is common to standardize a distribution by transforming the scores into a new
distribution with a predetermined mean and standard deviation that are whole round
numbers  create a new distribution that has simple values for the mean and standard
deviation, but does not change any location within the distribution = standardized scores
This is a two-step process:
1. The original raw scores are transformed into z-scores
2. The z-scores are transformed into new X values so that the specific μ and σ are
attained
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Z scores are most commonly used in a population, but the same principles can be used to
identify individual locations within a sample. You can use the exact same method, as long as
you use the sample mean and the sample standard deviation
Researchers use z-scores when comparing results of a sample to the original population, to
determine if the individuals in the sample are noticeably different you need the z-score to
decide whether a sample is noticeably different
Chapter 6
Probability in inferential statistics:
- You start with developing probability as a bridge from populations to samples, this
involves identifying the types of samples that probably would be obtained from a
specific population. Then you select from which population you should sample to it is
as representative as possible
Probability: a fraction of a proportion of all the possible outcomes  probability of A =
number of outcomes classified as A / total number of possible outcomes
- The chance of getting a king in a card deck  4/52 (4 out of 52)
- Symbol: p
- Identifies probability as a fraction or a proportion
- Probability values range from 0—1
- When calculating multiple probabilities, you need to multiply them or put one to the
power of how many predictions.
P(correct) x P(correct) x etc.  P(correct)10
Joint probability
Two events occurring together is less probable than each event happening individually:
P(A) = 0.6
P(B) = 0.7
P(C)  A + B  0.6 x 0.7 = 0.42
Conditional probability
The probability of how likely something is, given that something else is true
P (X I A)  probability of something, given A
(simple) random sample: each individual in the population has an equal chance of being
selected
Independent random sample: each individual has an equal chance of being selected and the
probability of being selected stays constant from one selection to the next if more than one
individual is selected
 used the most, is a required component in statistical applications
Requirements:
- There is no bias in the selection process, you select randomly in your population
- Sample with replacement: to keep the probabilities from changing from one selection
to the next, it is necessary to return each individual to the population before you
make the next selection
Normal distribution
Bell shaped, but not all bell-shaped distribution is a normal distribution
Function: gives you the probability density  formula:
Standard normal: mean of 0 and standard deviation of 1
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Using probability: what is the probability of randomly selecting an individual from a
population with a score greater than 700?
1. All scores are the population distribution, from which the mean is 500, so the score
700 is right to the mean. We are interested in scores greater than 700 to we need to
be at the right of the 700 mark.
2. To identify the exact position, you need to compute the z-score  z = (X – μ) / σ
 (700 – 500) / 100 = 2.00
The score 700 is exactly 2 standard deviations
above the mean
3. What is the probability?  use the standard
percentages of a normal distribution
P(z > + 2.00) = 2.28%
The unit normal table
A complete listing of z-scores and proportions (instead of a graph, which shows just a few
z-score values)
1. The body always corresponds to the larger part
of the distribution, no matter on which side it is.
The tail is always the smaller section
2. The normal distribution is symmetrical, so the
proportions on the right-hand side are exactly
the same as the proportions on the left-hand
side. The table does not list negative z-score
values
3. The z-score values change signs, but the
proportions are always positive
In calculations you should always sketch a distribution,
locate the mean with a vertical line and shade in the portion you are trying to determine,
this will help you avoid making errors
Most of the time you need to find probabilities (/ proportions, they mean the same) for
specific X values, so the first step you need to do is transform the X values into z-scores, and
then you can use the unit normal table to look up the proportions corresponding to the zscore values
 examples of calculations on pages 173 – 177
The binomial distribution
Binomial data: categorical variables, only a few options
A variable that is measured on a scale consisting of exactly two categories, the resulting data
are called binomial. This occurs when a variable naturally exists with only two categories
(male and female; head or tails)
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
The researcher often knows the probabilities associated with each of the two categories  a
coin, p(heads) = p(tails) = ½
Notation:
1. The two categories are identified as A and B
2. The probabilities associated with each category are identifies as
p(A) = p = the probability of A
p(B) = q = the probability of B
 p + q = 1.00
3. The number of individuals or observations in a sample is identified as n
4. The variable X refers to the number of times category A occurs in the sample
The binomial distribution tends to approximate a normal distribution, particularly when n is
large. Then it has the following parameters:
- Mean: μ = pn
- Standard deviation: σ = √ npq
- Z-score: z = (X – μ) / σ  z = (X – pn) / √ npq
Binomial values are discrete (coin tosses, 2 or 3, not 2,5), normal values are continuous
Chapter 7
A z-score value near zero indicates a central, representative sample; a z-value beyond +2.00
or -2.00 indicates an extreme sample
The difficulty with working with samples is that a sample provides an incomplete picture of
the population. A sample almost certainly misses some segments of the population that are
not included.
Sampling error: the natural discrepancy, or amount of error, between a sample statistic and
its corresponding population parameter
The distribution of sample means
The collection of sample means for all the possible random samples of a particular size (n)
that can be obtained from a population
 two separate samples will be different even though they are taken from the same
population
- Sampling distribution: a distribution obtained by selecting all the possible samples of
a specific size from a population
- you can obtain the distribution of sample means by selecting a random sample of a
specific size (n) from a population, calculate the sample mean, and place the sample
mean in the distribution. Do this with every random sample
 you can change the specific size (n) of a sample, and how many times you take a
sample of that size. When you increase both the distribution will start to look more
and more like a normal distribution and the sampling error will reduce
characteristics:
- Samples are not expected to be perfect, but they are representative of the
population, so they should be relatively close to the population mean
- The pile of sample means should tend to form a normal-shaped distribution
- The larger the sample size, the closer the sample means should be to the population
mean
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Central limit theorem
Way of determining exactly what the distribution of sample means looks like without taking
hundreds of thousands of samples
For any population with the mean μ and the standard deviation σ, the distribution of sample
means for a sample size n will have a mean of μ and a standard deviation of σ/ √ n and will
approach a normal distribution as n approaches infinity
- It describes the distribution of sample means for any population
- The distribution of sample means approaches a normal distribution very rapidly,
when the number of scores in each sample is relatively large and when the
population is a normal distribution
- The rule of thumb is that n needs to be at least 30 and then it will rapidly approach
the normal distribution
The mean
The mean of the distribution of sample means is always identical to the population mean, it
is called the expected value of M. it is identified by the symbol μM or μ
The standard error
The standard deviation for the distribution of sample means, it is identified by the symbol σM
and is called the standard error of M
1. It describes the distribution of sample means, provides a measure of how much
difference is expected from one sample to another
2. It measures how well an individual sample mean represents the entire distribution. It
provides a measure of how much distance is reasonable to expect between a sample
mean and the overall mean for the distribution of sample means
The magnitude of the standard error is determined by:
1. The size of the sample
Law of large numbers: the larger the sample size, the more probable it is that the
sample mean will be close to the population mean
2. The standard deviation of the population from which the sample is selected
Bigger samples have smaller error, and smaller samples have bigger error
- When n=1, σM = σ (standard error = standard deviation)
- Standard error (SE) = σM = σ/ √ n
- The population variance and the population standard deviation are directly related,
and it is easy to substitute variance into the equation of standard error
 standard error (SE) = σM = σ/ √ n = √ σ2/n
Probability
The primary use of the distribution of sample means is to find the probability associated with
any specific sample. Because the distribution of sample means represents the entire set of
possible samples, you can use proportions to determine probabilities
1. Determine the standard error of the population by σM = σ/ √ n
2. Determine the z-score and the mean and use it to locate the exact point given
3. Then you can look up the probability in the normal unit table
We are finding a location within the distribution of sample means: by finding the location for
sample mean M rather than a score X, and the standard deviation for the distribution of
sample means is the standard error
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
 the z formula for the z-score of sample means is z = M- μ / σM
Standard error vs sampling error
When you are working with a sample mean, you must use the standard error
Sampling error: a sample typically will not provide a perfectly accurate representation of its
population, typically there is some error between a statistic in a sample and the population
Standard error: most sample means are close to the population mean these samples provide
a fairly accurate representation of the population. Some means are far from the population
mean and are not representative. For each sample you can measure the error (or distance)
between the sample mean and the population mean
- When sample size increases, the standard error gets smaller and the sample means
tend to get closer to the population mean
- With increasing n, we decrease the standard error SE, and thereby reduce the
sampling error
The standard error measures how much discrepancy you should expect, between a sample
statistic and a population parameter. Statistical inference involves using sample statistics to
make a general conclusion about a population parameter, thus standard error plays a crucial
role in inferential statistics
Chapter 8
Hypothesis testing
a statistical procedure that allows researchers to use sample data to draw inferences about
the population of interest, by using sample data to evaluate a hypothesis about a population
- It is one of the most commonly used procedures in inferential statistics
- The details of a hypothesis test change from one situation to another, but the general
process remains constant
The process:
1. State a hypothesis about a population
American adults gain an average of μ = 7 pounds a year
2. Predict the characteristics the sample should have
The sample should have a mean around 7 pounds
3. Obtain a random sample from the population
4. Compare the obtained sample data with the prediction that was made from the
hypothesis
The unknown population
The researcher begins with a known population (the group before the treatment), which a
certain standard deviation and a mean. The purpose of the study is to determine what
happens to the population after the treatment. With any effect you can add or subtract a
certain number from each individual score ( this changes the mean, but not the shape of
the distribution or the standard deviation)
- To do this research you use a sample as your population because the research wants
to generalize it, but it is impossible to test a whole population
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
The four steps of a hypothesis test
Step 1: state a hypothesis about a population
The null hypothesis
Known as H0, states that in the general population there is no change, no difference, or no
relationship. In the context of an experiment, H0 predicts that the independent variable
(treatment) has no effect on the dependent variable (scores) for the population
The alternative hypothesis
Known as H1 (or HA), states that there is a change, a difference, or a relationship for the
general population. In the context of an experiment, H1 predicts that the independent
variable (treatment) does have an effect on the dependent variable
 data will evaluate the credibility of the hypotheses
According to the null hypothesis, the distribution of sample means is divided into two
sections:
1. Sample means that are likely to be obtained if H0 is true: that is, sample means that
are close to the null hypothesis = high probability
2. Sample means that are very unlikely to be obtained if H0 is true: that is, sample
means that are very different from the null hypothesis = low probability
 null hypothesis significance testing: the null hypothesis is likely or unlikely
Step 2: predict the characteristics the sample should have/ set the criteria for a decision
Alpha level
To find the boundaries that separate the high-probability
samples from the low-probability samples, we must define
what exactly is meant by high and low probability  level
of significance / alpha level
The alpha (α) value is a small probability that is used to
identify the low-probability samples
- Alpha = .05 (5%), alpha = .10 (10%)  separate the
most unlikely 5% of the sample means, from the
most likely 95%
- The extremely unlikely values make up the critical
region
- The extremely unlikely values can also be called the
convincing evidence that the treatment has an effect
To find the critical region you use the unit normal table. The boundaries separate the
extreme 5% from the middle 95%. The 5% is split up in two tails, so there is exactly 2.5% in
each tail  in the unit normal table you look up .0250 in column C and you find the z-score
boundary
Step 3: collect data and compute sample statistics
The collection of data happens after the hypothesis is set, so the researcher makes an
honest, objective evaluation of the data
1. The raw data is summarized, for example with the sample mean.
2. Then the data is compared with the hypothesis. This is done by computing a z-score
exactly where the sample mean is located relative to the hypothesized population
mean from H0  z = M – μ / σM
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
The value of the sample mean M is obtained from the sample data, the value of μ is
obtained from the null hypothesis
Step 4: make a decision by comparing the obtained sample data with the prediction that
was made from the hypothesis
The researcher uses the z-score value obtained in step 3 to make a decision about the null
hypothesis according to the criteria established in step 2, there are two possible outcomes:
1. The sample data are located in the critical region, the sample is not consistent with
H0 and the null hypothesis is rejected
The z-score is 2.25, this falls within the boundary of 1,96 (alpha level .05)
2. The sample data re not in the critical region, the sample mean is reasonably close to
the population mean specified in the null hypothesis, so we fail to reject the null
hypothesis, the treatment has no effect
The z-score is 0.75, this does not fall within the critical region (1.96)
The z-score
The z-score statistic used in the hypothesis test is a test statistic: the sample data are
converted into a single, specific statistic that is used to test the hypothesis
- The z-score as a recipe: you follow the instructions and use the right ingredients and
you produce a z-score
- The z-score as a ratio: it indicates a large discrepancy between sample data and the
null hypothesis
Errors in hypothesis testing
Type 1 errors
Occurs when a researcher rejects a null hypothesis that is actually true. In a typical research
situation, a type 1 error means the researcher concludes that a treatment does have an
effect when in fact it has no effect
- Occurs when a researcher unknowingly obtains an extreme, nonrepresentative
sample
- The alpha level determines the probability of a type 1 error, it determines the
probability of obtaining sample data in the critical region even though the null
hypothesis is true
Type 2 errors
Occurs when a researcher fails to reject a null hypothesis that is really false. In a typical
research situation, a type 2 error means that the hypothesis test failed to detect a real
treatment effect
- Occurs when the sample mean is not in the critical region, even though the
treatment has an effect on the sample, this often happens when the effect of the
treatment is relatively small
- The probability depends on a variety of factors and therefore is a function, it is
represented by the Greek letter β
Selecting an alpha level
To avoid type 1 errors, setting an alpha level is very important. The largest value is .05, but to
be more careful many researchers use an alpha level of .01 or .001 to reduce the risk of a
false report
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
-
Though, a lower alpha level means less risk of a type 1 error, but it also means that
the hypothesis test demands more evidence from the research results
Statistical significance
A result is said to be (statistically) significant if it is very unlikely to occur when the null
hypothesis is true. That is, the result is sufficient to reject the null hypothesis, thus a
treatment has a significant effect
- The p (probability) of obtaining a sample mean in the critical region is extremely
small (less than .05) if there is no treatment effect
 the p-value must be smaller than .05 to be considered statistically significant
Factors that influence a hypothesis test
The variability of scores
High variability can make it very difficult to see any clear patterns in the results from a
research study. In a hypothesis test, higher variability can reduce the chances of finding a
significant treatment effect
- Increasing the variability of the scores produces a larger standard error and a smaller
value for the z-score
Number of scores in the sample
Increasing the number of scores in the sample produces a smaller standard error and a
larger value for the z-score. If all factors held constant, the larger the sample size is, the
greater the likelihood of finding a significant treatment effect
Other factors that can influence the hypothesis test are considered to be constant, these
are:
- Random sampling
- Independent observation
- The value of σ is unchanged by the treatment
- Normal sampling distribution (in order to use the unit normal table)
One tailed vs two tailed hypothesis tests
What is discussed before is in a two-tailed hypothesis test  the critical region is divided
between the two tails of the distribution
One-tailed test
Also called a directional hypothesis test, where the statistical
hypotheses (H0 and H1) specify an increase of a decrease in the
population mean. That is, they make a statement about the
direction of the effect
Alcohol is expected to slow reaction times
Step 1: state the statistical hypothesis
H0: alcohol does not slow reaction times  μ </= the
population mean
H1: alcohol does slow reaction times  μ > the population mean
Step 2: define the critical region
If the prediction is that the treatment will produce an increase in scores, then the critical
region is located entirely in the right-hand tail of the distribution
- You are allowed to use an alpha level of .05, for just one tail of the distribution
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Chapter 9 (9.1, 9.2, 9.4)
The problem with using a z-score is that the z-score formula requires more information than
is usually available, especially the value of the population standard deviation
- Circularity problem: we make an inference about a population based on a population
parameter, if you do not know the population parameter you cannot calculate the zscore
solution: the t-statistic = estimating the standard deviation
You can estimate the standard error: estimated standard error
- the estimated standard error (sM) is used as an estimate of the real standard error σM
when the value of σ is unknown. It is computed from the sample variance or sample
standard deviation and provides an estimate of the standard distance between a
sample mean M and the population mean μ
- variance = s2 = SS/ (n – 1) = SS / df
- df = degrees of freedom (sample of 400 has 399 as the degrees of freedom)
 the estimated value is computed from sample data (s) rather than from the actual
population parameter (σ)
- SM = s/√ n, or
SM = √ s2/n
Estimated standard error = √ sample variance / sample size
 we use the variance instead of the standard deviation because it is unbiased, and the tstatistic needs the variance instead of the standard deviation
The t-statistic
Used to test hypotheses about an unknown population mean (μ) when the value of σ is
unknown. The formula for the t statistic has the same structure as the z-score formula,
except that the t statistic used the estimated standard error in the denominator
- t = (M – μ) / sM

t = (M – μ) /√ s2/n
- to determine how well a t statistic approximates a z-score, you must determine how
well the sample variance approximates the population variance  degrees of
freedom = df = n – 1
The t-distribution
a complete set of t values computed for every possible random sample for a specific sample
size (n) or a specific degree of freedom (df). The t distribution approximates the shape of a
normal distribution
- if df gets large, the t distribution gets closer in shape to a normal z-score distribution
- it has an estimated standard error so it is flatter  you can’t put all the weight in the
middle because it’s an estimation
- you do not have one t-distribution but the t-family because the distribution depends
on the sample size used
The t-distribution table
Given the degrees of freedom and the tail proportions (alpha level), locate the critical tstatistic
- The two rows at the top of the table show proportions of the t distribution contained
in either one or two tails, depending on which row is used
- The first column lists the degrees of freedom for the t statistic
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
-
The numbers in the body are the t values that mark the boundary between the tails
and the rest of the t distribution
 when your df value is not listed in the table (53), you loop up the critical t value for the
numbers surrounding it (40 and 60) and you use the smaller t-value!
Hypothesis tests with the t-statistic
Using the same procedure as with the z-score, the hypothesis test would result in a formula:
t = (sample mean, from the data) – (population mean, hypothesized from H 0) / estimated
standard error, computed from the sample data
- It works the same as with the z-score, if we obtain a large score then we can say that
the data is not consistent with the H0 hypothesis.
- You calculate the critical region by taking the degree of freedom: n-1
To calculate the t statistic, you need 3 steps:
1. Calculate the sample variance (the population variance is unknown, and you must
use the sample value)
 s2 = SS / (n – 1)
2. Use the sample variance and the sample size to compute the estimated standard
error
 sM = √ s2 / n
3. Compute the t statistic
 t = (M – μ) / sM
Assumptions of the t Test
- The values in the sample must consist of independent observations (no relationship)
- The population sampled must be normal
The t-statistic one tailed hypothesis test works the same as the z-score one tailed hypothesis
test
Chapter 10
General categories of two sets of data
1. independent measures research design / between subjects design: a research design
that uses a separate group of participants for each treatment condition (comparing
men and women)
2. Repeated measures research design / within subjects design: One group of people
before and after treatment
When you have two different groups in a study you also need to notate the data differently.
The number of scores in sample 1 would be n1 and the number of scores in sample two
would be n2
Independent-measures hypotheses
The goal is to evaluate the mean difference between two populations. The mean for the first
population is μ1 and for the second population is μ2  mean difference is μ1 – μ2
The null hypothesis means that there is no difference  H0: μ1 – μ2 = 0
The alternative hypothesis means that there is a difference, so H 1: μ1 – μ2 = not 0
Independent measures t-statistic
the basic structure of the t-statistic is the same for both the independent-measures and the
single-sample hypothesis test
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
- t = (M – μ) / sM
the independent measures t is basically a two sample t that doubles all the elements of the
single sample t formulas
- t = ((M1 – M2) – (μ1 – μ2) / S(M1 – M2)  t = sample mean difference / estimated standard
error
The estimated standard error
Can be interpreted two ways:
1. it measures the standard distance between (M1 – M2) and (μ1 – μ2)
2. it measures the standard, or average size of (M1 – M2) if the null hypothesis is true.
That is, it measures how much difference is reasonable to expect between the two
sample means
calculation:
1. each of the two sample means represents its own population mean, but in each case
there is some error
M1 approximates μ1 with some error  same with M2
2. the amount of error associated with each sample mean is measured by the estimated
standard error of M, the estimated standard error for each sample mean is computes
as:
For M1, SM = √ s12 / n1  same with M2
3. for the independent-measures t statistic we want to know the total amount of error
involved in using two sample means to approximate two population means. To do
this, we will find the error from each sample separately and then add the two errors
together 
This formula only works when two sample sizes are exactly the same size. When two sample
variances are not equally good, they should not be treated equally  the variance obtained
from a large sample is a more accurate estimate of σ2 than the variance obtained from a
small sample
Pooled variance: method for correcting the bias in the standard error by combining the two
sample variances into a single value. This is obtained by averaging the two sample variances
using a procedure that allows the bigger sample to carry more weight in determining the
final value
- With one sample the sample variance is s2 = SS/df
- The pooled variance is:
 when in doubt always used the pooled variance
With this information you can now calculate an unbiased standard error for a sample mean
difference, this formula will be:
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Hypothesis test with the independent-measures t-statistic
For a two-tailed test
1. State the hypotheses and select the alpha level
H0: μ1 – μ2 = 0
H1: μ1 – μ2 ≠ 0
α = .05
2. Determine the critical region by calculating the degrees of freedom . It is an
independent measures design, so the calculation is:
df = df1 + df2
 look up boundaries in t-distribution table
3. Obtain the data and compute the test statistic
First, find the pooled variance
Second, use the pooled variance to compute the estimated standard error
Third, compute the t-statistic
4. Make a decision
Assumptions underlying the independent-measures t-formula
There are three assumptions that should be satisfied before you use the independentmeasures t-formula for hypothesis testing
1. The observations within each sample must be independent
2. The two populations from which the samples are selected must be normal
3. The two populations from which the samples are selected must have equal variances
- If the two sample variances are estimating different population variances, then the
average is meaningless
- If two population variances are equal, then the two sample variances should be very
similar, when the two sample variances are close, then the homogeneity assumption
is satisfied. When one variance is three or four times larger than the other, then
there is concern
Chapter 11
Repeated measures / within-subjects design
The dependent variable is measured two or more times for each individual in a single
sample, the same group is used in all of the treatment conditions
- Measured before and after treatment
 no risk for participant in one group are very different than the other group
Matched-subjects design
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Each individual in one sample is matched with an individual in the other sample. The
matching is done so that the two individuals are
equivalent with respect to a specific variable that
the researcher would like to control
- Study on learning, make sure people of all
different IQ scores are in one group. Match
one person with 120 with another person
with 120
The t-statistic for repeated-measured design
It is essentially the same as the single-sample t statistic,
but instead of using raw scores you use difference
scores
Because every person is measured twice you need to
calculate the difference score:
D = X2 – X1
 shows the direction of change for each person
Hypothesis:
H0: μD = 0  there is no difference between the time-points
H1: μD ≠ 0
For repeated-measures the t-statistic is almost the same as a single-sample t-statistic:

-
The estimated standard error is exactly the same as in a single-sample t-statistic
-
For the mean you calculate the mean by dividing the added difference scores by the
number of participants
The standard deviation is calculated by
dividing the squared difference scores by
the number of participants  it is the
standard deviation of the difference, not
the difference in standard deviations
-
Effect size
Whenever a treatment effect is found to be
statistically significant, it is recommended that
you also report a measure of the absolute
magnitude of the effect  Cohen’s d and r2
Cohen’s d
A standardized measure of the mean difference between treatments
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
For a repeated measures study, the population mean and standard deviation are unknown,
so you use the sample mean difference divided by the sample mean deviation
The percentage of variance accounted for = r2
Computed using the obtained t value and the df value 
The answer is the percentage of the variance in the scores
that is explained by the effect
Differences repeated measures and independent measures
1. Number of subjects: a repeated-measures design requires fewer subjects
2. Study changes over time: the repeated-measures design is especially well suited for
studying learning, development, or other changes that take place over time
3. Individual differences: a repeated-measures design eliminates individual differences
 the disadvantage of repeated measures is:
- Order effects, so you can solve it by counterbalancing: participants are randomly
divided into groups and they do both treatments but switched. Group one does first
A, and then B. Group two does first B, and then A
- Participant loss (in longitudinal designs)
- Sometimes they are practically difficult
Chapter 8.5
There have been some concerns for hypothesis testing
1. The focus of a hypothesis test is on the data rather than the hypothesis. When the
null hypothesis is rejected, we are making a statement about the sample data, not
about the null hypothesis
2. Demonstrating a significant treatment effect does not necessarily indicate a
substantial treatment effect. Statistical significance does not provide any real
information about the absolute size of a treatment effect
 the test is making a relative comparison: the size of the treatment effect is being
evaluated relative to the standard error (a very small standard error can still be large
enough to be significant; a significant effect does not mean a big effect)
Effect size
Intends to provide a measurement of the absolute magnitude of a treatment effect,
independent of the size of the sample(s) being used
Cohen’s d:
 does not include the sample size, so it is not influenced by the number of scores in a
sample
Chapter 9.3
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
In most situations the population values are not known, so you must substitute the
corresponding sample values in their place, it is the same with Cohen’s d; which becomes,
An estimated d of 1, indicates that the size of the treatment effect is equivalent to one
standard deviation
Confidence intervals
An interval, or range of values centered around a sample statistic. The logic behind a
confidence interval is that a sample statistic, such as a sample mean, should be relatively
near to the corresponding population parameter
- Normally we report findings using point estimates (M=5, SD=3)
- An interval accepts that there is uncertainty around an estimated value  instead of
saying M = 5, but say that the mean lies in an interval from A to B
- An interval has a certain width and is defined by a lower and upper boundary
 95% of the 95% confidence interval will contain the true value of M
Constructing a confidence interval
1. The observation that every sample mean has a corresponding t value defined by the
equation
2. You know M and SM, but t and μ are unknown, however you can estimate t. when you
know the sample has n scores, your df is n-1. If you want to be 80% confident that
the sample mean corresponds to a t value in a certain interval you can use a two
tailed proportion of 0.20 combined.
3. You then plug the estimated t value into the equation, and you can calculate the
value of μ.
The calculation is:
 it’s two values
Example:
Factors affecting width of a confidence interval
1. When changing the level of confidence, the width of the interval changes. To have
more confidence, the width must be increased. And when you have a smaller more
precise interval, you must give up confidence
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
2. The width of the interval also changes if you change sample size, a bigger sample size
has a smaller interval
Chapter 8.6
Revision errors
Type 1 error
- False positive, we conclude that there is an effect when there isn’t
 keep the error low by an alpha level, with an alpha level of .01 1% of the values under
the null lie in that area, thus in 1% of the cases, we will incorrectly conclude that there is
an effect
Type 2 error
- Missed effects, we conclude that there is no difference, but in reality there is one
- This error term is β
- Keep these low as well  you can’t set them all low, one is always a compromise of
the other
- The higher your β (type 2 error), the lower you statistical power
 if we make α stricter (decreasing), we increase β, so we decrease statistical power (1 – β)
 if we make 1 – β higher (increasing), we decrease β, so we increase type 1 error α
Solutions (increasing statistical power)
- Increase sample size  statistical power increases (less overlap between
distributions)
- Increase the effect size of interest
- Increase α
Statistical power
Instead of measuring effect size, you can measure the power of the statistical test
 the power is the probability that the test will correctly reject a false null hypothesis. That
is, power is the probability that the test will identify a treatment effect if one really exists
- Whether a treatment has an effect has one of two outcomes, reject or accept the null
hypothesis. This should together add up to 1, failing to reject H0 when there is an
effect is a type 2 error, with a probability of p= β  second outcome is 1 – β
- Power of a hypothesis test is equal to 1 – β
- When the power of the test is 70% (1- β) then the probability of a type 2 error is 30%
(β)
- The power is directly related to the sample size, one reason to compute power is to
determine what sample size is necessary to achieve a reasonable probability for a
successful research study.
When comparing two normal distributions, one without treatment (μ=80) and one with
treatment (μ=88). You take an alpha level of .05 in a two-sided hypothesis test. You will see
that the distribution with treatment lies almost
entirely in the critical region, which means that
the treatment is close to a 100% power.
Calculation
Locate the exact boundary for the critical region
- SD x z-score  2 x 1.96 = 3.92
Add the distance to the sample mean
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
 80 + 3.92 = 83,92
Determine what proportion of the treated samples are greater than M=83,92, calculate the
z- score
Look up the z-score in the unit normal table (-2.04) which corresponds to p=0.9793
 if the treatment has an 8-point effect, 97.93% of all the possible sample means will be in
the critical region and we will reject the null hypothesis
Chapter 15
Correlation
= a numerical value that describes and measures three characteristics of the relationship
between X and Y
1. The direction of the relationship: the sign of the correlation, positive of negative,
describes the direction of the relationship
 positive correlation: the variables tend to change in the same direction
 negative correlation: the variables tend to change in the opposite direction
2. Form of the relationship: linear form = points tend to cluster around a straight line
3. Strength or consistency: how the points fall on/around the line. A perfect correlation
is marked as 1.00, a correlation with no consistency is marked as 0.00
Why correlations are used
1. Prediction
2. Validity: checking if a test gives the result it was expected to, the psychologist could
measure the correlation between the new test and the other measures to
demonstrate that the new test is valid
3. Reliability: demonstrating that a test produces stable, consistent measurements
4. Theory verification: the prediction of a theory could be tested by determining the
correlation between two variables
Correlation and…
Causation
 there may be a causal relationship between two variables, but the existence of a
correlation does not prove it
Restricted range
A sample only shows a limited range of scores, a calculation with these scores could be
totally different from a calculation with a full range of scores
 To be safe you should not generalize any correlation beyond the range of data
represented
Outliers
Extreme data points, most easily noticed in a scatter plot
The coefficient of determination (r2)
= it measures the proportion of variability in one variable that can be determined from the
relationship with the other variable
- Used to measure the size of strength of the correlation
Pearson correlation (r)
= measures the degree and direction of the linear relationship between two variables
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
- The correlation coefficient ranges from -1.00 to +1.00
Correlation coefficient
- Population notation: ρ
- Sample notation: r
- Effect size: R2  expresses absolute magnitude of the effect (by squaring r)
- Values approaching +/- 1.00 indicate a strong relationship, values around 0 indicate a
weak relationship
Sum of products (SP)
To calculate the Pearson correlation, you need the sum of products (SP)
This is calculated by:
or
- MX is the mean for X scores
- MY is the mean for the Ys
Steps:
1. Find the X deviation and the Y deviation for each individual
2. Find the product of the deviations for each individual
3. Add the products
Pearson correlation
- SP is used to measure the degree to which X and Y vary together (covariability)
- Both SS of X and Y are used to measure how they vary separately
Calculation:
Z-scores
Each X and Y score can be transformed into a z-score, so you can express the Pearson
correlation in terms of z-scores
Hypothesis test
 null hypothesis: there is no correlation, which is zero
 alternative hypothesis: there is a correlation, which is nonzero
T-statistic:
 sample value = calculated r
 population value = ρ (often 0)
Standard error:
Complete t statistic:
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
Degrees of freedom: n – 2
Chapter 17
Associations between variables
Most calculations are done with numerical data, but often data are not numerical  could
be problematic
For example: experiment of extra lessons vs no extra lessons and pass vs fail
 2 x 2 table
 is there an association between the factor extra lessons (yes – no) and
the result (pass – fail)
 unable to use all calculations used before
The chi-square test for goodness of fit
= Uses sample data to test hypotheses about the shape or proportions of a population
distribution. The test determines how well the obtained sample proportions fit the
population proportions specified by the null hypothesis
- Present the scale of measurement as a series of boxes, with each box corresponding
to a separate category on the scale
- Symbol: χ2
Hypotheses
Null hypothesis: specifies the proportion (or percentage) of the population in each category
 50% of lawyers are men and 50% are women
1. No preference, equal proportions: no difference among the categories, the
population is equally divided among the categories
 alternative hypothesis would say that the population distribution has a different
shape
2. No difference from a known population: the proportions for one population are not
different from the proportions known to exist for another population
 alternative hypothesis states that the population proportions are not equal to the
values specified by the null hypothesis
Data
No need to calculate, just select a sample of n individuals
and count how many there are in each category
 the resulting values are the observed frequencies (fo) = the number of individuals from
the sample who are classified in a particular category, each individual is counted in one and
only one category
When you have a certain population in your null hypothesis that is distributed in a certain
way in percentages (group 1 is 25%, group 2 is 50% and group 3 is 25%), to find the exact
frequency expected for each category you multiply the sample size (n) by the proportion (or
percentage) from the null hypothesis
 these values are the expected frequencies (fe) = for each category the frequency value
that is predicted from the proportions in the null hypothesis and the sample size (n). the
expected frequencies define an idea, hypothetical sample distribution that would be
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
obtained if the sample proportions were in perfect agreement with the proportions specified
in the null hypothesis
Chi square statistic
Purpose is to determine whether the sample data support or refute a hypothesis about the
population
Steps:
1.
2.
3.
4.
Find the difference between fo (the data) and fe (the hypothesis) for each category
Square the difference  makes all values positive
Divide the squared difference by fe
Sum the values from all the categories
Chi square distribution
the set of chi-square values for all the possible random samples when H 0 is true,
characteristics are:
1. The formula involves squaring, so all values are zero or larger  hypothesis is
positively skewed
2. When H0 is true, you expect the data (fo ) to be close to the hypothesis (fe) values 
we expect chi-square values to be small when H0 is
true
Degrees of freedom
Df = C – 1
 C is the amount of categories
Critical region
 chi square distribution table
Chi-square for goodness fit hypothesis test example
Data: n=50
Step 1: state the hypotheses and alpha level
H0: there is no preference, so all 4 categories are equal  each one has 25%
H1: one or more of the orientations is preferred over the others
α = .05
Step 2: locate the critical region
Degrees of freedom is C – 1  4 – 1 = 3
For df = 3 and alpha is .05, the critical value is 7.81
Step 3: calculate the chi square statistic
Compute the expected frequencies from H0
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
 null hypothesis says that each category has 25%, the
sample is n=50
 fe = pn = ¼ x 50 = 12.5
Calculate the value of the chi square statistic
Step 4: make a decision
The chi square value lies in the critical region, so H0 is rejected  the four categories are not
equal
Chi square test for independence
= Uses the frequency data from a sample to evaluate the relationship between two variables
in a population. Each individual in the sample is classified on both of the two variables,
creating a two-dimensional frequency distribution matrix. The frequency distribution for the
sample is then used to test hypotheses about the corresponding frequency distribution in
the population
- Example:
Hypotheses
1: H0: for the general population, there is no relationship between A and B
 The value obtained for one variable is not related to the value for the second variable
2: H0: In the population, the proportions in the distribution of A are not different from the
proportions in the distributions of B, the two distributions have the same shape
 the data consists of two (or more) separate samples that are being used to test
differences between two (or more) populations
Data
Same logic as the chi-square test for goodness of fit.
1. Sample is selected, and each individual is classified or categorized
 two variables, so every individual is classified on both variables
 the frequencies in the sample distribution are the observed frequencies (fo)
2. To find the expected frequencies (fe) you determine the overall distribution of A, and
then apply this distribution to both categories of B
Using the example used above:
The total sample consists of 200 people, the proportion selecting red is 100 out of 200  50%
Using this you get:
- 50% prefer red, 10% prefer yellow, 20% prefer green, 20% prefer blue
According to the null hypothesis, both personality groups should have the same proportions
for color preferences. When calculating the introverts and red you do: 50% out of 50 = 25
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
When you do this everywhere you get:
Instead of extensive calculations you
have a formula for the
expected frequencies:
-
Fc is the frequency total for the column
Fr is the frequency total for the row
Example: the red introvert  100 x 50 / 200 = 25
Statistic and degrees of freedom
The chi-square statistic is exactly the same:
The degrees of freedom is different, because you’re not only dealing with categories or
columns  df = (R – 1) x (C – 1)  R is row, C is column
Chi-square for independence hypothesis test example
Data: n = 200
Step 1: hypothesis and level of significance
Version 1:
H0: in the general population, there is no relationship between parents’ rule for alcohol use
and the development of alcohol related problems
H1: there is a relationship
Version 2:
H0: in the general population, the distribution of alcohol-related problems has the same
proportions for teenagers whose parents permit drinking and for those whose parents do
not
H1: there are different proportions
Alpha level: .05
Step 2: degrees of freedom and critical region
Degrees of freedom: (R – 1) x (C – 1)  (2 – 1) x (2 – 1) = 1
 critical value of 3.84
Step 3: determine the expected frequencies and calculate the chi-square statistic
- 160 out of 200 is 80%
40 out of 200 is 20%
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
 null hypothesis states these are the same for both groups
For the teens who were allowed to drink:
- No: 64  80% out of 80
- Yes: 16  20% out of 80
For the teens who were not allowed to drink:
- No: 96  80% out of 120
- Yes: 24  20% out of 120
Chi-square statistic:
Step 4: make a decision
Reject the null hypothesis
Effect sizes of the Chi square test
Aim: “corrected” indication of the effect
Phi Coefficient
the square root of the chi-square value divided by the sample size
Cramer’s V
 beyond 2 by 2 tables
Square root of chi-square divided by n, multiplied with the degrees of freedom with an
asterix  the smaller of (R – 1) and (C – 1)
Chapter 18
Binomial test
= uses sample data to evaluate hypotheses about the values of p and q for a population
consisting of binomial data
- Binomial data: data classified in two distinct categories, in which each observation is
classified in one of these two categories, and the sample data consist of the
frequency or number of individuals in each category
Hypotheses
H0 falls in one of two categories
1. Just chance: the null hypothesis states that the two outcomes A and B occur in the
population with the proportions that would be predicted simply by chance
 50%  p = ½
2. No chance or no difference: when you know the proportions for one population and
want to determine if the same proportions apply to a different population
Z-score formula
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Stuvia.com - The Marketplace to Buy and Sell your Study Material
When pn and qn (both categories x sample size) are both equal or greater than 10, the
binomial distribution resembles a normal distribution
- The standard deviation is:
- The mean is p x n
the z-score calculation is:
 X/n is the proportion of individuals in the sample who are classified in
category A
 P is the hypothesized value for the proportion of individuals in the
population who are classified into category A

is the standard error of the sampling distribution of X/n and
provides a measure of the standard distance between the sample
statistic (X/n) and the population parameter
Direct formula
 when the data is not at least 10
Example
10 multiple choice questions  each answer has 4 options (25%)
Goal is to have 5 correct by guessing
Downloaded by: giuliagasbarro8 | giulia.gasbarro8@gmail.com
Distribution of this document is illegal
Powered by TCPDF (www.tcpdf.org)
Download