This article was automatically translated from the original Turkish version.
Autocorrelation (or serial dependence) refers to the situation in which the error terms (residuals) in a time series are correlated with each other. Statistically, autocorrelation is defined as the correlation of a random variable with its own lagged values. This concept is particularly important in regression analysis, as it indicates a violation of one of the classical regression assumptions—that the error terms are independent of each other.
Autocorrelation is commonly observed in time series data structures; for example, it is typical to find relationships between consecutive observations in economic, financial, or meteorological data collected over time.
One of the assumptions of the Classical Linear Regression Model (CLRM) is that the error terms are independent and uncorrelated (white noise). When this assumption is violated—that is, when there is serial correlation among the error terms—the model’s predictive power, reliability, and interpretability can be seriously compromised.
Specific problems that arise include:
Autocorrelation typically arises due to the following reasons:
Autocorrelation can be detected using various statistical methods. The most common approaches are:
1. Graphical Methods
2. Durbin-Watson (DW) Test: This is the most widely used method for testing the presence of autocorrelation. The test statistic ranges between 0 and 4. A value of 2 indicates no autocorrelation; values close to 0 suggest positive autocorrelation, while values close to 4 suggest negative autocorrelation.
3. Breusch-Godfrey (BG) Test: Due to certain limitations of the DW test—such as its ability to detect only first-order autocorrelation—the more general BG test is often preferred. The BG test can detect higher-order autocorrelation and accommodates models with lagged independent variables.
If autocorrelation is detected, the following alternative techniques may be used instead of the Classical OLS method:
1. Generalized Least Squares (GLS): The GLS method accounts for the covariance structure of the error terms, yielding efficient and consistent estimates.
2. Cochrane-Orcutt Procedure: This iterative method models the autocorrelation structure of the error terms and re-estimates the regression coefficients accordingly.
3. Prais-Winsten Method: Similar to the Cochrane-Orcutt procedure, but it preserves the first observation, avoiding its loss.
4. Yule-Walker Equations and AR (Autoregressive) Models: Error terms are modeled using autoregressive processes such as AR(1) or AR(p), explicitly accounting for their temporal structure. This approach is widely used in time series modeling.
The Importance of Autocorrelation in Regression Models
Causes of Autocorrelation
Detecting Autocorrelation
Alternative Approaches When Autocorrelation Is Present