2. Overview
ā¢ Introduction.
ā¢ Why ANOVA instead of multiple t-tests?
ā¢ One way ANOVA.
ā¢ Assumptions of One way ANOVA.
ā¢ Steps in One way ANOVA.
ā¢ Example.
ā¢ Conclusion.
3. Introduction
ā¢ ANOVA is an abbreviation for the full name of the method:
Analysis Of Variance.
ā¢ Invented by R.A. Fisher in the 1918.
ā¢ ANOVA is used to test the significance of the difference
between more than two sample means.
ā¢ Name āANOVAā is a misnomer as it compares mean to check
variance between group.
4. Summary Table of Statistical tests
Level of
Measurement
Sample Characteristics
Correlation
1 Sample
2 Sample K Sample (i.e., >2)
Independent Dependent Independent Dependent
Categorical or
Nominal
Ī§2 or
bi-
nomina
l
Ī§2 Macnarmarās
Ī§2
Ī§2 Cochranās Q
Rank or
Ordinal
Mann
Whitney U
Wilcoxin
Matched
Pairs Signed
Ranks
Kruskal
Wallis H
Friedmanās
ANOVA
Spearmanās
rho
Parametric
(Interval &
Ratio)
z test
or
t test
t test
between
groups
t test
within
groups
1 way ANOVA
between
groups
1 way ANOVA
(within or
repeated
measure)
Pearsonās
r
Factorial (2 way) ANOVA
Ī§2
5. Why ANOVA instead of multiple t-tests?
ā¢ If you are comparing means between more than two groups,
we can choose two sample t-tests to compare the mean of one
group with the mean of the other groups?
:Before ANOVA, this was the only option available to
compare means between more than two groups.
ā¢ The problem with the multiple t-tests approach is that as the
number of groups increases, the number of two sample t-tests
also increases.
ā¢ As the number of tests increases the probability of making a
Type I error also increases.
6. One way ANOVA
ā¢ One way ANOVA (=F test) compares the mean of two or more
group whenever there is one independent variable is involved.
ā¢ It finds out whether there are any statistical significance
difference between their group means.
ā¢ If more then one independent variable is involved then it is
called as N way ANOVA.
ā¢ One way ANOVA specifically tests the null hypothesis.
ā¢ H0 = u1 = u2 = u3 = uk , u = group mean & k= no. of groups.
ā¢ If One way ANOVA shows a statistical significant result it
means HA is true.
7. Variables In One way ANOVA
ā¢ In an ANOVA, there are two kinds of variables: independent
and dependent
ā¢ The independent variable is controlled or manipulated by the
researcher.
ā¢ It is a categorical (discrete) variable used to form the
groupings of observations.
8. ā¢ There are two types of independent variables: active and
attribute.
ā¢ If the independent variable is an active variable then we
manipulate the values of the variable to study its affect on
another variable.
ā¢ For example, anxiety level is an active independent variable.
ā¢ An attribute independent variable is a variable where we do not
alter the variable during the study.
ā¢ For example, we might want to study the effect of age on
weight. We cannot change a personās age, but we can study
people of different ages and weights.
9. ā¢ In the One-way ANOVA, only one independent variable is
considered, but there are two or more (theoretically any finite
number) levels of the independent variable.
ā¢ The independent variable is typically a categorical variable.
ā¢ The independent variable (or factor) divides individuals into two
or more groups or levels.
ā¢ The procedure is a One-way ANOVA, since there is only one
independent variable.
10. ā¢ The (continuous) dependent variable is defined as the variable
that is, or is presumed to be, the result of manipulating the
independent variable.
ā¢ In the One-way ANOVA, there is only one dependent variable ā
and hypotheses are formulated about the means of the groups on
that dependent variable.
ā¢ The dependent variable differentiates individuals on quantitative
(continuous) dimension.
11. Assumptions of One way ANOVA
1) All populations involved follow a normal distribution.
2) Homogeneity of variances: The variance within each group
should be equal for all groups.
3) Independence of error: The error (variation of each value
around its own group mean) should be independent for each
value.
4) Only ONE independent variable should be checked whether
it produces a significant difference between the groups.
12. Example of One way ANOVA
Group A Group B Group C
160,110,118,124,13
2
122,136,124,126,12
0,138
148,126,124,128,14
0
N1= 5 N2= 6 N3= 5
Mean=128.8 Mean= 127.66 Mean=133.2
EXAMPLE: A study conducted to assess & compare the
effect of Treatment A vs Treatment B vs Treatment C
on SBP in a specified population.
13. ANOVA
One way ANOVA Three way ANOVA
Effect of Drugs on SBP
Two way ANOVA
Effect of Diet &
Drugs on SBP
Effect of Exercise,
Drugs, Diet on SBP
14. Steps in One way ANOVA
2. State Alpha
3. Calculate degrees of Freedom
4. Calculate test statistic
- Calculate variance between samples
- Calculate variance within the samples
- Calculate F statistic
1. State null & alternative hypotheses
15. Example- one way ANOVA
Example: A investigator wants to find out the
analgesic effect of aspirin vs diclofenac vs
ibuprofen in a group of population with equal
variances.
Aspirin Diclofenac Ibuprofen
1 5 9
4 10 3
7 2 2
9 1 4
3 7 2
16. Steps Involved
1.Null hypothesis ā
No significant difference in the means of 3 samples
2. State Alpha i.e 0.05
3. Calculate degrees of Freedom
k-1 & n-k = 2 & 12
4. State decision rule
Table value of F at 5% level of significance for d.f 2 & 12 is
3.88
The calculated value of F > 3.88 , H0 will be rejected
5. Calculate test statistic
17. One way ANOVA: Table
Source of
Variation
SS (Sum of
Squares)
Degrees of
Freedom
MS (Mean
Square)
Variance
Ratio of F
Between
Samples
SSB k-1 MSB=
SSB/(k-1)
MSB/MSW
Within
Samples
SSW n-k MSW=
SSW/(n-k)
Total SS(Total) n-1
18. Calculating variance BETWEEN samples
1. Calculate the mean of each sample.
2. Calculate the Grand mean.
3. Take the difference between means of various samples &
grand average.
4. Square these deviations & obtain total which will give sum
of squares between samples (SSC)
5. Divide the total obtained in step 4 by the degrees of freedom
to calculate the mean sum of square between samples (MSC).
20. Variance BETWEEN samples (M1=4.8, M2=5,M3=4)
Sum of squares between samples (SSC) =
n1 (M1 ā Grand avg)2 + n2 (M2ā Grand avg)2 + n3(M3ā Grand avg)2
5 ( 4.8 - 4.6) 2 + 5 ( 5 - 4.8) 2 + 5 ( 4.6 - 4.8) 2 = 0.6
Calculation of Mean sum of squares between samples (MSB)
=0.6/2 = 0.3
k= No of Samples, n= Total No of observations
21. Calculating Variance WITHIN Samples
1. Calculate mean value of each sample.
2. Take the deviations of the various items in a sample from the
mean values of the respective samples.
3. Square these deviations & obtain total which gives the sum
of square within the samples (SSE)
4. Divide the total obtained in 3rd step by the degrees of
freedom to calculate the mean sum of squares within samples
(MSE).
22. Variance WITHIN samples (M1= 4.8, M2= 5,M3= 4)
X1 (X1 ā M1)2 X2 (X2ā M2)2 X3 (X3ā M3)2
1 14.4 7 4 2 4
4 0.64 5 0 9 25
7 4.84 10 25 3 1
9 17.64 2 9 2 4
3 3.24 1 16 4 0
40.76 54 34
Sum of squares within samples (SSE) = 40.76 + 54 +34 =128.76
Calculation of Mean Sum Of Squares within samples (MSW)
= 128.76/12 = 10.73
23. The mean sum of squares
1ļ
ļ½
k
SSC
MSC
kn
SSE
MSE
ļ
ļ½
Calculation of MSC-
Mean sum of Squares
between samples
Calculation of MSE
Mean Sum Of
Squares within
samples
k= No of Samples, n= Total No of observations
24. Calculation of F statsitics
groupswithinyVariabilit
groupsbetweenyVariabilit
F ļ½
Compare the F-statistic value with F(critical) value which is
obtained by looking for it in F distribution tables against
degrees of freedom. The calculated value of F > table value
H0 is rejected
25. ā¢ F Value = MSB/MSW = 0.3/10.73 = 0.02.
The Table value of F at 5% level of significance for d.f 2 & 12 is 3.88
The calculated value of F < table value
H0 is accepted. Hence there no is significant difference in sample
means
26. Within-Group
Variance
Between-Group
Variance
Between-group variance is large relative to the
within-group variance, so F statistic will be
larger & > critical value, therefore statistically
significant .
Conclusion ā At least one of group means is
significantly different from other group means
28. Post-hoc Tests
ā¢ Used to determine which mean or group of means is/are
significantly different from the others (significant F)
ā¢ Depending upon research design & research question:
ļ¼ Bonferroni (more powerful)
Only some pairs of sample means are to be tested
Desired alpha level is divided by no. of comparisons
ļ¼ Tukeyās HSD Procedure
when all pairs of sample means are to be tested
ļ¼ Scheffeās Procedure (when sample sizes are unequal)
29. Application of ANOVA
ā¢ ANOVA is designed to detect differences among
means from populations subject to different treatments.
ā¢ ANOVA is a joint test, the equality of several
population means is tested simultaneously or jointly.
ā¢ ANOVA tests for the equality of several population
means by looking at two estimators of the population
variance (hence, analysis of variance).
30. Conclusion
ā¢ The one-way analysis of variance is used where there is a
single factor that will be set to three or more levels.
ā¢ t is not appropriate to analyse such data by repeated t-tests as
this will raise the risk of false positives above the acceptable
level of 5 per cent.
ā¢ If the ANOVA produces a significant result, this only tells us
that at least one level produces a different result from one of
the others.
ā¢ Follow-up tests needs to be carried out to find out which group
differs from each other.