A simple ANOVA program containing only class and model statements will return the ANOVA table detailing source, degrees of freedom, Type I and III sums of squares, mean square errors, F-values for individual tests, and their associated p-values. While this is all useful information, it only tells us whether or not the means are equal across all of the classes of interest. If the F-test shows that the means are not all equal, we need to know how to detect which means are different and to find meaningful conclusions from this information.
Using a means statement in the glm procedure will provide the mean value of the dependent variable at each treatment level. Additionally, we can use post hoc tests as options in this statement to perform a variety of multiple comparisons that provide information about the differences between treatment levels. To do multiple comparisons, type a slash (/) after the means statement and then indicate which post hoc test you want SAS to perform. Detailed descriptions of these tests begin on page 253 of your textbook and on page 179 of your course notes. For this lab, we will perform all tests so that you can get a feel for the differences between them. Another set of options available in the means statement allows us to test for homogeneity of variance using a variety of tests. Again after the slash, type in hovtest= and then bartlett, bf, levene, or obrien welch. Descriptions of these tests can be found on page 179 of your course notes.
The lsmeans statement provides adjusted least-square means for the main effects. In balanced designs these are the same as the raw means, but in unbalanced designs you need to use lsmeans instead. Options available after the slash (/) in the lsmeans statement include the standard error of each mean (stderr), probabilities for pairwise differences (pdiff), and which post-hoc test, if any, you want SAS to use (adjust=). Contrasts can be used to test other meaningful comparisons, and are discussed at length beginning on page 242 of your textbook.
Below is a sample SAS program and output for you to become familiar with before writing and interpreting your own.
dm 'log;clear;output;clear';
options nodate nocenter nonumber ls=100 ps=100;
title 'Example of 1-Way ANOVA';
data one;
input c$ y @@;
cards;
l 1.59 l 1.73 l 3.64 l 1.97
m 3.36 m 4.01 m 3.49 m 2.89
h 3.92 h 4.82 h 3.87 h 5.39
;
proc print;
title2'Insulin Release as a function of glucose concentrations';
run;
proc glm;
title2'GLM Example';
class c;
model y=c;
output out=two p=yhat residual=resid;
means c/hovtest=bartlett hovtest=bf hovtest=levene hovtest=obrien welch lsd tukey bon scheffe duncan;
lsmeans c/stderr pdiff adjust=tukey;
contrast 'Average of Low & Med vs. Average of High' c 1 1 -2;
run;
proc univariate data=two normal plot;
title2'Tests of Assumptions';
var resid;
run;
options ls=70 ps=40;
proc plot data=two;
plot resid*yhat;
run;
quit;
A.
Dependent Variable: y
Sum of
Source DF Squares Mean Square F Value Pr > F
Model 2 10.29665000 5.14832500 9.31 0.0064
Error 9 4.97935000 0.55326111
Corrected Total 11 15.27600000
B.
Example of 1-Way ANOVA
GLM Example
The GLM Procedure
Levene's Test for Homogeneity of y Variance
ANOVA of Squared Deviations from Group Means
Sum of Mean
Source DF Squares Square F Value Pr > F
c 2 0.5407 0.2703 0.91 0.4352
Error 9 2.6626 0.2958
O'Brien's Test for Homogeneity of y Variance
ANOVA of O'Brien's Spread Variable, W = 0.5
Sum of Mean
Source DF Squares Square F Value Pr > F
c 2 0.9612 0.4806 0.58 0.5771
Error 9 7.3961 0.8218
Brown and Forsythe's Test for Homogeneity of y Variance
ANOVA of Absolute Deviations from Group Medians
Sum of Mean
Source DF Squares Square F Value Pr > F
c 2 0.2056 0.1028 0.38 0.6975
Error 9 2.4670 0.2741
Bartlett's Test for Homogeneity of y Variance
Source DF Chi-Square Pr > ChiSq
c 2 1.2700 0.5299
C.
Bonferroni (Dunn) t Tests for y
NOTE: This test controls the Type I experimentwise error rate, but it generally has a higher Type
II error rate than REGWQ.
Alpha 0.05
Error Degrees of Freedom 9
Error Mean Square 0.553261
Critical Value of t 2.93332
Minimum Significant Difference 1.5428
Means with the same letter are not significantly different.
Bon
Groupi
ng Mean N c
A 4.5000 4 h
A
B A 3.4375 4 m
B
B 2.2325 4 l
D.
The GLM Procedure
Least Squares Means
Adjustment for Multiple Comparisons: Tukey
Standard LSMEAN
c y LSMEAN Error Pr > |t| Number
h 4.50000000 0.37190762 <.0001 1
l 2.23250000 0.37190762 0.0002 2
m 3.43750000 0.37190762 <.0001 3
Least Squares Means for effect c
Pr > |t| for H0: LSMean(i)=LSMean(j)
Dependent Variable: y
i/j 1 2 3
1 0.0050 0.1629
2 0.0050 0.1085
3 0.1629 0.1085
E.
Dependent Variable: y Contrast DF Contrast SS Mean Square F Value Pr > F Average of Low & Med vs. Average of High 1 0.01353750 0.01353750 0.02 0.8792
| Instr 1 | Instr 2 | Instr 3 | Instr 4 | Instr 5 |
| 11.6 | 8.5 | 14.5 | 12.3 | 13.9 |
| 10.0 | 9.7 | 13.3 | 12.9 | 16.1 |
| 10.5 | 6.7 | 14.5 | 11.4 | 14.3 |
| 10.6 | 7.5 | 14.8 | 12.4 | 13.7 |
| 10.7 | 6.7 | 14.4 | 11.6 | 14.9 |
2. Test the hypothesis that the mean scores of all five classes are equal. Remember to ouptut the residuals and test for normality. Report your conclusion, along with F-value, p-value, and degrees of freedom.
3. If the means are not equal, use pairwise comparisons to determine which means are different. Try both means and lsmeans, along with several different methods of testing. Report the appropriate conclusions along with support from your SAS output.
4. Construct a contrast to test the hypothesis that the average of Classes 3 & 4 equals the average of Classes 1 & 5 and report your conclusions.