Skip to main content icon/video/no-internet

Serial correlation, or autocorrelation, is defined as the correlation of a variable with itself over successive observations. It often exists when the order of observations matters, the typical scenario of which is when the same variable is measured on the same participant repeatedly over time. For example, serial correlation is an important issue to consider in any longitudinal designs.

Serial correlation has mainly been considered in multiple regression and time-series models. Multiple regression models are designed for independent observations, where the existence of serial correlation is undesirable. So the main focus in multiple regression is on testing whether serial correlation exists. Conversely, the purpose of time-series analysis is to model the serial correlation to understand the nature of time dependence in the data. The pattern of serial correlation is essential for identifying the appropriate model. This presentation on serial correlation is around regression and time series.

Multiple Regression Model

Let the multiple regression model be

None

where y i is the response and xi is a 1 × (k + 1) vector consisting of a 1 and the values of the k predictors in the ith observation. The assumptions of this model are (a) the expectation of the error εi is 0; (b) εi has constant variance γ2; and (c) εi and εj are uncorrelated if ij. Let

None
be the least squares (LS) estimator of β and SSE be the sum of squared errors. When the assumptions are valid,
None
is the best linear unbiased estimator and s2 = SSE/(nk − 1) is an unbiased estimator of δ2. When the errors are correlated (violating assumption c), although
None
is still unbiased, s2 and the estimated standard error of
None
are biased. Consequently, the F or t statistic in testing the significance of
None
is misleading. Therefore, it is important to test for the presence of serial correlation. The most widely used test is the Durbin—Watson d test, which tests for first-order serial correlation ρ = Corr(εi, εi-1) using the autoregressive model
None

Obviously, ρ = 0 if η = 0. To test the null hypothesis H0 = 0, the test statistic d is formulated as

None

where

None
i are the LS residuals with fitting Equation

1. Let

None
be the LS estimate of η in Equation 2. It follows from
None
that d ≈(1-
None
) Thus, if there is no serial correlation, d ≈ 2; if the serial correlation is close to 1, d ≈ 0; and if the serial correlation is close to −1, d ≈ 4. The critical values of d (denote the lower bound as dL and the upper bound as dU) depend on n, k, and the significance level of the test. Tables of dU and dL can be found in the Appendix of Arnold Studenmund's book. The appropriate decision rules of testing for positive serial correlation are
None

The decision rules of testing for negative serial correlation are

None

The inconclusive region is one main disadvantage of the Durbin—Watson d test. Moreover, the test ignores serial correlation beyond the first order. It does not allow earlier observed y to predict later y in the regression model either.

...

  • Loading...
locked icon

Sign in to access this content

Get a 30 day FREE TRIAL

  • Watch videos from a variety of sources bringing classroom topics to life
  • Read modern, diverse business cases
  • Explore hundreds of books and reference titles

Sage Recommends

We found other relevant content for you on other Sage platforms.

Loading