Checking assumptions in regression

Chong Ho Yu, Ph.D. 
Several assumptions for the data should be met in order to apply a valid regression model. Wonnacott and Winacott (1981) argued that if the assumptions of linearity, normality and independence are upheld, additional assumptions such as fixed values of X are not problematic. Berry & Feldman (1995) pointed out that most regression assumptions are concerned with residuals. This writeup will demonstrate how different assumptions of regression are examined.
Residuals have constant variance (homoescedasticity)
When the error term variance appears constant, the data are considered homoscedastic, otherwise, the data are said to be heteroscedastic. You can use either SAS's command syntax or SAS/Insight to check this assumption.

To use SAS/Insight:
 Choose Solutions/Analyze/Interactive Data Analysis
 Open the data set from the SAS library
 Choose Fit[Y X]

To use SAS's syntax, type and submit the following:
proc reg;
model y = x1 x2;
output out=two
p=y_hat
r=y_res;
proc gplot data=two; plot y_res * y_hat;

Both of the above methods produce a plot of residuals versus predicted values. If the residuals variance is around zero, it implies that the assumption of homoscedasticity is not violated. If there is a high concentration of residuals above zero or below zero, the variance is not constant and thus a
systematic error exists.
Independence of Residuals
A regression model requires independence of error terms. Again, a residuals plot can be used to check this assumption. Random, patternless residuals imply independent errors. Even if the residuals are even distributed around zero and the assumption of constant variance of residuals is satisfied, the regression model is still questionable when there is a pattern in the residuals as shown in the following figure.
Normality of Residuals
It is important to note that for regression the normality test should be applied to the residuals rather than the raw scores. There isn't a general agreement of the best way to test normality. In SAS, there are four test statistics for detecting the presence of nonnormality, namely, the ShapiroWilk (Shapiro & Wilk, 1965), the KolmogorovSmirnov test, Cramer von Mises test, and the AndersonDarling test. According to the SAS manual, if the sample size is over 2000, the Kolmgorov test should be used. If the sample size is less than 2000, the Shapiro test is better. The null hypothesis of a normality test is that there is no significant departure from normality. When the p is more than .05, it fails to reject the null hypothesis and thus the assumption holds.
NCSS (NCSS Statistical Software, 2007) provides more normality tests in addition to the Shapiro test and the Kolmgorov test (see the following table). According to NCSS, the Shapiro and the Anderson Darling tests are the best. The Kolmogorov test is included just because of its historical popularity, but is bettered in almost every way by other tests.
However, some researchers argue that the Shapiro test was originally constructed to test a sample size carrying up to 50 subjects. To examine normality for a sample size between 51 to 1999, other tests such as AndersonDarling, MartinezLglewicz, and D'Agostino tests are recommended.
Normality Test Section
Test Name

Test Value

Prob Level

10% Critical Value

5% Critical Value

Decision (5%)

ShapiroWilk W

0.937

0.21



Accept Normality

AndersonDarling

0.443

0.29



Accept Normality

MartinezIglewicz

1.026


1.216

1.357

Accept Normality

KolmogorovSmirnov

0.148


0.176

0.192

Accept Normality

D'Agostino Skewness

1.037

0.299

1.645

1.960

Accept Normality

D'Agostino Kurtosis

.786

0.432

1.645

1.960

Accept Normality

D'Agostino Omnibus

1.691

0.429

1.645

1.960

Accept Normality

Although many authors recommended using skewness and kurtosis for examining normality (Looney, 1995), Wilkinson (1999) argued that skewness and kurtosis often fail to detect distributional irregularities in the residuals. By this argument, the D'Agostino skewness and the D' Agostino Kurtosis test may be less useful.
Also, statistical tests depend on sample size, and as sample size increases, the tests often will reject innocuous assumptions. Most normality tests have small statistical power (probability of detecting nonnormal data) unless the sample size is large. If the null is rejected, the data are definitely nonnormal. But if it fails to reject the null, the conclusion is uncertain. All you know is that there was not enough evidence to reject the normality assumption (NCSS Statistical Software, 2007). To rectify the sample size problem, Hair, Anderson, Tatham and Black (1992) suggested that in a small sample size it is safer to use both a normal probability plot and test statistics to ensure the normality. In SAS submitting the following command syntax returns both normal probability plot and test statistics.
proc univariate normal plot data=two; var y_res;
histogram y_res/normal;
QQplot y_res; run;


The same results can also be obtained from SAS/Insight. To use SAS/Insight:
 Choose Solutions/Analyze/Interactive Data Analysis
 Open the data set from the SAS library
 Choose Distribution
 Choose the options as shown in the figure.

In a normal probability plot, the normal distribution is represented by a straight line angled at 45 degrees. The standard residuals are compared against the diagonal line to show the departure. If the residuals follow along the straight line, it means that the departure from normality is slight.
Residuals have mean as zero
The above PROC UNIVARIATE statement returns the mean. One can also use PROC MEANS to get the same result. However, when the mean value carries many
decimals, the SAS system will use Enotation. In the following example, the
decimal point should shift 15 positions to the left, and thus the mean value is near zero (.000000000000001862483).
Linearity
To examine the assumption of linearity, one can apply a scatterplot matrix showing all Xs against Y in a pairwise manner. However, this option is not available in SAS and SPSS's scatterplot matrix is not interactive. I recommend using an interactive scatterplot matrix, which is a feature of DataDesk. In a DataDesk's scatterplot matrix, one can assign colors to the
data points for detecting clusters in different relationships. Take the following graphs as an example, the assumption of linearity seems to be violated because it appears that there are two clusters within the subjects. A linear fit to all
data points is not the best fit.
Fox (1991) suggested that although it is useful to plot y against each x for the examination of linearity, these plots are inadequate because they only tell the partial relationship between y and each x, controlling for the other xs. Therefore, it is desirable to use residual plots against y. How to output residuals has been illustrated in the section of homoescedasticity. One can modify those methods by replacing the predicted values (Y hat) with the observed values (Y).
The absence of multicollinearity
Multicollinearity will cause the variances to be high. These inflated variances are quite detrimental to regression because some variables add very little or even no new and independent information to the model (Belsley, Kuh & Welsch, 1980). Although Schroeder, Sjoquist and Stephen (1986) asserted that there is no statistical test that can determine whether or not multicollinearity really is a problem, there are still several ways for detecting multicollinearity such as a matrix of bivariate correlation and the regression of each independent variable in the equation on all other independent variables (Berry and Feldman, 1985). Nonetheless, the former approach lacks sensitivity to multiple correlations while the latter cannot tell much about the influence of regressors to variances. A better approach is using the Variance Inflation Factor (VIF). The detail of detecting multicollinearity is in the writeup Multicollinearity, variance inflation factor, and orthogonalization.
Put all things together
It is more efficient to check all of the preceding issues at the same time. The
following simple SAS macros was written for this purpose.
/*
name = name of the data set
dv = dependent variable,
x1 = first independent variable,
xlast = last independent variable */
%macro reg (name, dv, x1, xlast);
proc reg data=&name; model &dv = &x1  &xlast/vif;
output out=two
p=y_hat
r=y_res;
title "Check
multicollinearity using VIF";
proc gplot data=two; plot y_res * y_hat;
title "Check
homoescedasticity and independence of residuals";
proc univariate normal plot data=two; var y_res;
histogram y_res/normal;
probplot y_res;
QQplot y_res;
title "Check the
mean and normality of residuals";
proc gplot data=&name; plot &dv * (&x1  &xlast);
title "Check
linearity";
run;
%mend reg;
/*
To invoke the macros, enter the actual names of the data set,
DV, fist IV, and the last IV in the following.
*/
%reg(name, dv, x1, xlast)

References
 Belsley, D. A., Kuh, E. & Welsch, R. E. (1980). Regression Diagnostics: Identifying influential data and sources of collinearity. New York: John Wiley.
 Berry, W. D. & Feldman, S. (1985). Multiple regression in practice. London: Sage Publications.
 Fox, J. (1991). Regression diagnostics. London: Sage Publications.
 Looney, S. W. (1995). How to use tests for univariate normality to assess multivariate normality. American Statistician, 49, 6470.
 NCSS Statistical Software. (2007). NCSS. [Computer Software] Kaysville, UT:
Author.
 Schroeder, L. D., Sjoquist, D. L. & Stephan, P. E. (1986). Understanding regression analysis. Beverly Hills, CA: Sage Publications.
 Shapiro, S. S., & Wilk, M. B. (1965). An analysis of variance test for
normality. Biometrika, 52, 591611.
 Wilkinson, L, & Task Force on Statistical Inference. (1999). Statistical methods in psychology journals: Guidelines and explanations. American Psychologist, 54,
594604.
 Wonnacott, T. H. & Wonnacott, R. J. (1981). Regression: A second course in statistics. New York: Wiley.
Last updated: 2010
Navigation
Index
Simplified Navigation
Table of Contents
Search Engine
Contact
