Proportion of Variance Explained

Author(s)

David M. Lane

Prerequisites

Analysis of Variance, Partitioning Sums of Squares, Multiple Regression

Learning Objectives
  1. State the difference in bias between η2 and ω2
  2. Compute η2 Compute ω2
  3. Distinguish between ω2 and partial ω2
  4. State the bias in R2 and what can be done to reduce it

Effect sizes are often measured in terms of the proportion of variance explained by a variable. In this section, we discuss this way to measure effect size in both ANOVA designs and in correlational studies.

ANOVA Designs

Responses of subjects will vary in just about every experiment. Consider, for example, the "Smiles and Leniency" case study. A histogram of the dependent variable "leniency" is shown in Figure 1. It is clear that the leniency scores vary considerably. There are many reasons why the scores differ. One, of course, is that subjects were assigned to four different smile conditions and the condition they were in may have affected their leniency score. In addition, it is likely that some subjects are generally more lenient than others, thus contributing to the differences among scores. There are many other possible sources of differences in leniency ratings including, perhaps, that some subjects were in better moods than other subjects and/or that some subjects reacted more negatively than others to the looks or mannerisms of the stimulus person. You can imagine that there are innumerable other reasons why the scores of the subjects could differ.

Histogram of leniency

Figure 1. Distribution of leniency scores.

One way to measure the effect of conditions is to determine the proportion of the variance among subjects' scores that is attributable to conditions. In this example, the variance of scores is 2.794. The question is how this variance compares with what the variance would have been if every subject had been in the same treatment condition. We estimate this by computing the variance within each of the treatment conditions and taking the mean of these variances. For this example, the mean of the variances is 2.649. Since the mean variance within the smile conditions is not that much less than the variance ignoring conditions, it is clear that "Smile Condition" is not responsible for a high percentage of the variance of the scores. The most convenient way to compute the proportion explained is in terms of the sum of squares "conditions" and the sum of squares total. The computations for these sums of squares are shown in the chapter on ANOVA. For the present data, the sum of squares for "Smile Condition" is 27.535 and the sum of squares total is 377.189. Therefore, the proportion explained by "Smile Condition" is:

27.535/377.189 = 0.073.

Thus, 0.073 or 7.3% of the variance is explained by "Smile Condition."

An alternative way to look at the variance explained is as the proportion reduction in error. The sum of squares total (377.189) represents the variation when "Smile Condition" is ignored and the sum of squares error (377.189 - 27.535 = 349.654) is the variation left over when "Smile Condition" is accounted for. The difference between 377.189 and 349.654 is 27.535. This reduction in error of 27.535 represents a proportional reduction of 27.535/377.189 = 0.073, the same value as computed in terms of proportion of variance explained.

This measure of effect size, whether computed in terms of variance explained or in terms of percent reduction in error, is called η2 where η is the Greek letter eta. Unfortunately, η2 tends to overestimate the variance explained and is therefore a biased estimate of the proportion of variance explained. As such, it is not recommended (despite the fact that it is reported by a leading statistics package).

An alternative measure, ω2 (omega squared), is unbiased and can be computed from
omega squared formula

where MSE is the mean square error and k is the number of conditions. For this example, k = 4 and ω2 = 0.052.

It is important to be aware that both the variability of the population sampled and the specific levels of the independent variable are important determinants of the proportion of variance explained. Consider two possible designs of an experiment investigating the effect of alcohol consumption on driving ability. As can be seen in Table 1, Design 1 has a smaller range of doses and a more diverse population than Design 2. What are the implications for the proportion of variance explained by Dose? Variation due to Dose would be greater in Design 2 than Design 1 since alcohol is manipulated more strongly than in Design 1. However, the variance in the population should be greater in Design 1 since it includes a more diverse set of drivers. Since with Design 1 the variance due to Dose would be smaller and the total variance would be larger, the proportion of variance explained by Dose would be much less using Design 1 than using Design 2. Thus, the proportion of variance explained is not a general characteristic of the independent variable. Instead, it is dependent on the specific levels of the independent variable used in the experiment and the variability of the population sampled.

Table 1. Design Parameters

Design Dose Population
1 0.00
0.30
0.60
All Drivers between 16 and 80 Years of Age
2 0.00
0.50 1.00
Experienced Drivers between 25 and 30 Years of Age

Factorial Designs

In one-factor designs, the sum of squares total is the sum of squares condition plus the sum of squares error. The proportion of variance explained is defined relative to sum of squares total. In an A x B design, there are three sources of variation (A, B, A x B) in addition to error. The proportion of variance explained for a variable (A, for example) could be defined relative to the sum of squares total (SSQA + SSQB + SSQAxB + SSQerror) or relative to SSQA + SSQerror.

To illustrate with an example, consider a hypothetical experiment on the effects of age (6 and 12 years) and of methods for teaching reading (experimental and control conditions). The means are shown in Table 2. The standard deviation of each of the four cells (Age x Treatment combinations) is 5. (Naturally, for real data, the standard deviations would not be exactly equal and the means would not be whole numbers.) Finally, there were 10 subjects per cell resulting in a total of 40 subjects.

Table 2. Condition Means

  Treatment
Age Experimental Control
6 40
42
12 50 56

The sources of variation, degrees of freedom, and sums of squares from the analysis of variance summary table as well as four measures of effect size are shown in Table 3. Note that the sum of squares for age is very large relative to the other two effects. This is what would be expected since the difference in reading ability between 6- and 12-year-olds is very large relative to the effect of condition.

Table 3. ANOVA Summary Table

Source df SSQ η2 partial
η2
ω2 partial
ω2
Age 1 1440 0.567 0.615 0.552 0.586
Condition 1 160 0.063 0.151 0.053 0.119
A x C 1 40 0.016 0.043 0.006 0.015
Error 36 900
Total 39 2540

First, we consider the two methods of computing η2, labeled η2 and partial η2. The value of η2 for an effect is simply the sum of squares for this effect divided by the sum of squares total. For example, the η2 for Age is 1440/2540 = 0.567. As in a one-factor design, η2 is the proportion of the total variation explained by a variable. Partial η2 for Age is SSQAge divided by (SSQAge + SSQerror), which is 1440/2340 = 0.615.

As you can see, the partial η2 is larger than η2. This is because the denominator is smaller for the partial η2. The difference between η2 and partial η2 is even larger for the effect of condition. This is because SSQAge is large and it makes a big difference whether or not it is included in the denominator.

As noted previously, it is better to use ω2 than η2 because η2 has a positive bias. You can see that the values for ω2 are smaller than for η2. The calculations for ω2 are shown below:

omega squared factorial
omega square formula

where N is the total number of observations.

The choice of whether to use ω2 or the partial ω2 is subjective; neither one is correct or incorrect. However, it is important to understand the difference and, if you are using computer software, to know which version is being computed. (Beware, at least one software package labels the statistics incorrectly).

Correlational Studies

In the section "Partitioning the Sums of Squares" in the Regression chapter, we saw that the sum of squares for Y (the criterion variable) can be partitioned into the sum of squares explained and the sum of squares error. The proportion of variance explained in multiple regression is therefore:

SSQexplained/SSQtotal

In simple regression, the proportion of variance explained is equal to r2; in multiple regression, it is equal to R2.

In general, R2 is analogous to η2 and is a biased estimate of the variance explained. The following formula for adjusted R2 is analogous to ω2 and is less biased (although not completely unbiased):

adjusted r square formula

where N is the total number of observations and p is the number of predictor variables.

Please answer the questions:
correct feedback