Which of the following are assumptions underlying independent-measures analysis of variance?

To use the ANOVA test we made the following assumptions:

  • Each group sample is drawn from a normally distributed population
  • All populations have a common variance
  • All samples are drawn independently of each other
  • Within each sample, the observations are sampled randomly and independently of each other
  • Factor effects are additive

The presence of outliers can also cause problems. In addition, we need to make sure that the F statistic is well behaved. In particular, the F statistic is relatively robust to violations of normality provided:

  • The populations are symmetrical and uni-modal.
  • The sample sizes for the groups are equal and greater than 10

In general, as long as the sample sizes are equal (called a balanced model) and sufficiently large, the normality assumption can be violated provided the samples are symmetrical or at least similar in shape (e.g. all are negatively skewed).

The F statistic is not so robust to violations of homogeneity of variances. A rule of thumb for balanced models is that if the ratio of the largest variance to smallest variance is less than 3 or 4, the F-test will be valid. If the sample sizes are unequal then smaller differences in variances can invalidate the F-test. Much more attention needs to be paid to unequal variances than to non-normality of data.

We now look at how to test for violations of these assumptions and how to deal with any violations when they occur.

  • Testing that the population is normally distributed (see Testing for Normality and Symmetry)
  • Testing for homogeneity of variances and dealing with violations (see Homogeneity of Variances)
  • Testing for and dealing with outliers (see Outliers in ANOVA)

Hypotheses: In ANOVA we wish to determine whether the classification (independent) variable affects what we observe on the response (dependent) variable. In the example, we wish to determine whether Temperature affects Learning.

In statistical terms, we want to decide between two hypotheses: the null hypothesis (Ho), which says there is no effect, and the alternative hypothesis (H1) which says that there is an effect.

In symbols:

Which of the following are assumptions underlying independent-measures analysis of variance?

Note that this is a non-directional test. There is no equivalent to the directional (one-tailed) T-Test.

The t test statistic for two-groups:

Recall the generic formula for the T-Test:
Which of the following are assumptions underlying independent-measures analysis of variance?

For two groups the sample statistic is the difference between the two sample means, and in the two-tail test the population parameter is zero. So, the generic formula for the two-group, two-tailed t-test can be stated as:
Which of the following are assumptions underlying independent-measures analysis of variance?

(We usually refer to the estimated standard error as, simply, the standard error).

The F test statistic for ANOVA:

The F test statistic is used for ANOVA. It is very similar to the two-group, two-tailed T-test. The F-ratio has the following structure:
Which of the following are assumptions underlying independent-measures analysis of variance?

Note that the F-ratio is based on variance rather than difference.

But variance is difference: It is the average of the differences of a set of values from their mean.

The F-ratio uses variance because ANOVA can have many samples of data, not just two as in T-Tests. Using the variance lets us look at the differences that exist between all of the many samples.

  • The numerator
  • : The numerator (top) of the F-ratio uses the variance between the sample means. If the sample means are all clustered close to each other (small differences), then their variance will be small. If they are spread out over a wider range (bigger differences) their variance will be larger. So the variance of the sample means measures the differences between the sample means.

  • The denominator
  • : The denominator (bottom) of the F-ratio uses the error variance, which is the estimate of the variance expected by chance. The error variance is just the square of the standard error. Thus, rather than using the standard deviation of the error, we use the variance of the error. We do this so that the denominator is in the same units as the numerator.

Independent Variable:
Temperature (Farenheit)
Treatment 1
50-F
Treatment 2
70-F
Treatment 3
90-F
0
1
3
1
0
4
3
6
3
4
1
2
2
0
0
Mean=1Mean=4Mean=1

The most obvious thing about the data is that they are not all the same: The scores are different; they are variable.

The heart of ANOVA is analyzing the total variability into these two components, the mean square between and mean square within. Once we have analyzed the total variability into its two basic components we simply compare them. The comparison is made by computing the F-ratio. For independent-measures ANOVA the F-ratio has the following structure:

Which of the following are assumptions underlying independent-measures analysis of variance?

or, using the vocabulary of ANOVA,
Which of the following are assumptions underlying independent-measures analysis of variance?

For the data above:
Which of the following are assumptions underlying independent-measures analysis of variance?

(Note: The book says 11.28, but this is a rounding error. The correct value is 11.25.)

Degrees of Freedom: Note that the exact shape depends on the degrees of freedom of the two variances. We have two separate degrees of freedom, one for the numerator (sum of squares between) and the other for the denominator (sum of squares within). They depend on the number of groups and the total number of observations. The exact number of degrees of freedom follows these two formulas (k is the number of groups, N is the total number of observations):

Which of the following are assumptions underlying independent-measures analysis of variance?

A Conceptual View of ANOVA

Conceptually, the goal of ANOVA is to determine the amount of variability in groups of data, to determine where it comes from, and to see if the variability is greater between groups than within groups.

We can demonstrate how this works visually. Here are three possible sets of data. In each set of data there are 3 groups sampled from 3 populations. We happen to know that each set of data comes from populations whose means are 15, 30 and 45.

We have colored the data to show the groups. We use

  1. Red for the group with mean=15
  2. Green for the group with mean=30
  3. Blue for the group with mean=45
With each visualization we present the corresponding F-Test value and its p value.
  1. For the first example the populations each have a variance of 4.
    Which of the following are assumptions underlying independent-measures analysis of variance?

    F=854.24, p<.0001.

  2. For the second example the outer two populations still have a variance of 4, but the middle one has a variance of 64, so it overlaps the outer two (though they are still fairly well separated).
    Which of the following are assumptions underlying independent-measures analysis of variance?

    F=11.66, p<.0001.

  3. For the third example the three populations have a variance of 64, so they all overlap a lot.
    Which of the following are assumptions underlying independent-measures analysis of variance?

    F=1.42, p=.2440.
Note that in these examples, the means of the three groups haven't varied, but the variances have. We see that when the groups are well separated, the F value is very significant. On the other hand, when they overlap a lot, the F is much less significant.

Post Hoc Tests

You will recall, that in ANOVA the null and alternative hypotheses are:
Which of the following are assumptions underlying independent-measures analysis of variance?

When the null hypothesis is rejected you conclude that the means are not all the same. But we are left with the question of which means are different:

Post Hoc tests help give us an answer to the question of which means are different.
Post Hoc tests are done "after the fact": i.e., after the ANOVA is done and has shown us that there are indeed differences amongst the means. Specifically, Post Hoc tests are done when:
  1. you reject Ho, and
  2. there are 3 or more treatments (groups).
A Post Hoc test enables you to go back through the data and compare the individual treatments two at a time, and to do this in a way which provides the appropriate alpha level.

T-Tests can't be used: We can't do this in the obvious way (using T-Tests on the various pairs of groups) because we would get too "rosy" a picture of the significance (for reasons I don't go into). The Post Hoc tests gaurantee we don't get too "rosy" a picture (actually, they provide a picture that is too "glum"!).

Two Post Hoc tests are commonly used (although ViSta doesn't offer any Post Hoc tests):

  • Tukey's HSD Test (thats HSD for Honestly Significant Difference). This test can be used only when the groups are all the same size. It determines a single value that is the minimum difference between a pair of groups that is needed for the difference to be significant at a specific alpha level.
  • Scheffe's Test is very conservative. It involves computing an F-Ratio that has a numerator that is a mean-square that is based on only the two groups being compared (the denominator is the regular error variance term).

  • Example
  • We look at hypothetical data about the effect of drug treatment on the amount of time (in seconds) a stimulus is endured. We do an ANOVA following the formal hypothesis testing steps. Note that the books steps are augmented here to reflect current thinking about using visualizations to investigate the assumptions underlying the analysis.
    1. State the Hypotheses
    2. :
      The hypotheses, for ANOVA, are:
      Which of the following are assumptions underlying independent-measures analysis of variance?

    3. Set the Decision Criterion

    4. We arbitrarily set
      Which of the following are assumptions underlying independent-measures analysis of variance?

    5. Gather the Data
    6. :
      The data are obtained from 60 subjects, 20 in each of 3 different experimental conditions. The conditions are a Placebo condition, and two different drug conditions. The independent (classification) variable is the experimental condition (Placebo, DrugA, DrugB). The dependent variable is the time the stimulus is endured.

      Here are the data as shown in ViSta's data report:

      Which of the following are assumptions underlying independent-measures analysis of variance?

      The data may be gotten from the ViSta Data Applet. Then, you can do the analysis that is shown below yourself.

    7. Visualize the Data

    8. We visualize the data and the model in order to see if the assumptions underlying the independent-measures F-test are met. The assumptions are:

      1. The observations within each sample must be independent (this assumption is satisfied by the nature of the experimental design).
      2. The populations from which the samples are selected must be normal (the data and model visualizations can inform us about this).
      3. The populations from which the samples are selected must have equal variance (the data and model visualizations can inform us about this also). This is called homogeneity of variance.

      The data visualization is shown below. The boxplot shows that there is somewhat more variance in the "DrugA" group, and that there is an outlier in the "DrugB" group. The Q plots (only the "DrugB" Q-Plot is shown here) and the Q-Q plot show that the data are normal, except for the outlier in the "DrugB" group.

      Which of the following are assumptions underlying independent-measures analysis of variance?

    9. Evaluate the Null Hypothesis

    10. We use ViSta to calculate the observed F-ratio, and the observed probability level. The report produced by ViSta is shown below. The information we want is near the bottom:

      Which of the following are assumptions underlying independent-measures analysis of variance?


      We note that F=4.37 and p=.01721. Since the observed p < .05, we reject the null hypothesis and conclude that it is not the case that all group means are the same. That is, at least one group mean is different than the others.

      Here is the F distribution for df=2,57 (3 groups, 60 observations). I have added the observed F=4.37:

      Which of the following are assumptions underlying independent-measures analysis of variance?

    11. Visualize the Model

    12. Finally, we also visualize the ANOVA model to see if the assumptions underlying the independent-measures F-test are met. The boxplots are the same as those for the data. The partial regression plot shows that the model is significant at the .05 level of significance, since the curved lines cross the horizontal line. The residual plot shows the outline in the "DrugB" group, and shows that the "DrugA" group is not as well fit by the ANOVA model as the other groups. Here is the model visualization:
      Which of the following are assumptions underlying independent-measures analysis of variance?

    What are the assumptions for the independent measures ANOVA?

    There are three primary assumptions in ANOVA: The responses for each factor level have a normal population distribution. These distributions have the same variance. The data are independent.

    Which of the following is an assumption underlying a repeated measures analysis of variance ANOVA )?

    A repeated measures ANOVA assumes sphericity – that variances of the differences between all combinations of related groups must be equal.