Statistical Significance R.Raveendran Why should we test significance? Heart rate (bpm) Mean ± SEM n In men - 73.34 ± 5.82 In women - 80.45 ± 6.13 10 10 The difference between means (80.45-73.34) = 7.11 We do not need a stat test of significance, if only : a. the data from all subjects in a group are IDENTICAL b. we can collect data from all subjects in a population Why should we test significance? We test SAMPLE to draw conclusions about POPULATION If two SAMPLES (group means) are different, can we be certain that POPULATIONS (from which the samples were drawn) are also different? Is the difference obtained TRUE or SPURIOUS? Will another set of samples be also different? What are the chances that the difference obtained is spurious? The above questions can be answered by STAT TEST. Stat test Tests whether two groups are statistically different from each other Statistically different? = Truly different? Not just apparently different You do not need statistics to say these two are truly different. Do you? But statistics does help us determine which group of trees is taller You do not need statistics to say these two are truly different. Do you? How to find statistical difference? How does a Stat test work? Stat test analyses the data (numbers) submitted (by the researcher) to calculate the chances of obtaining a difference when there is none i.e. probability of obtaining a spurious difference. It does not indicate (a) (b) (c) (d) whether your design is right or wrong whether the type of data is correct or wrong the magnitude of the difference whether the difference will be practically useful All it can point out is whether the obtained difference between two groups is REAL or FALSE What does a Stat test infer? Stat test Data P value When p<0.05, it shows that the chances of obtaining a false difference is less than 5% (1 in 20) [p<0.01 – 1 in 100; p<0.001 – 1 in 1000] Since we consider 5% P is small, we conclude that the difference between groups is TRUE Truth is something which is most likely to be true and 100% certainty is impossible. How to test statistical significance? State Null hypothesis Set alpha (level of significance) Identify the variables to be analysed Identify the groups to be compared Choose a test Calculate the test statistic Find out the P value Interpret the P value Calculate the CI of the difference Calculate Power if required Thank you Null hypothesis Null hypothesis (statistical hypothesis) states that there is no difference between groups compared. Alternative hypothesis or research hypothesis states that there is a difference between groups. e.g. New drug ‘X’ is an analgesic - (Research hypothesis) New drug ‘X’ is not an analgesic – (Null hypothesis) Alpha / type 1 error / level of significance The level of significance is to be set It is generally set at 0.05 (5%) and not above. If the P value is less than this limit then null hypothesis is rejected i.e. the difference between groups is not due to chance. Choosing a stat test Why should we choose a test? Choosing a stat test………. Why should we choose a test? There are many tests The selection of test varies with the type of data, analysis, study design, distribution & no. of groups Choosing a stat test……… Parametric Non-parametric Student’s t test paired t unpaired t Wilcoxon Pearson’s correlation Spearman’s rank correlation ANOVA way signed rank test rank sum test One – way Kruskal-Wallis two Friedman Chi square test Kolomogorov-Smirnov test Choosing a stat test…… Determine : Aim of the study – Parameter to be analysed Data type- [Continuous, Discrete, Rank, Score, Binomial] Analysis type- [Comparison of means, Quantify association, Regression analysis] No. of groups to be analysed No. of data sets to be analysed Distribution of data - [normal or non-normal] Design - [paired or unpaired] With the above information, one can decide the suitable test using the table given. Choosing a stat test…… 1. Data type 2. Distribution of data 3. Analysis type (goal) 4. No. of groups 5. Design Table downloaded from www.graphpad.com Table downloaded from www.graphpad.com Calculating test statistic difference between group means variability of groups e.g. t test t = XT - XC SE(XT - XC) Determining P Find out the degrees of freedom (df) Use t and df to find out P using a formula or ‘critical values table’ How to interpret P? If P < alpha (0.05), the difference is statistically significant If P>alpha, the difference between groups is not statistically significant / the difference could not be detected. If P> alpha, calculate the power If power < 80% - The difference could not be detected; repeat the study with more ‘n’ If power ≥ 80 % - The difference between groups is not statistically significant. Degrees of Freedom It denotes the number of samples that a researcher has freedom to choose. The concept can be explained by an analogy : X + Y = 10 X+ Y+Z = 15 df = 1 df = 2 For paired t test df = n-1 For unpaired t test df= N1+N2 - 1