Autocorrelation in statistics is a mathematical tool that is usually used for analyzing functions or series of values, for example, time domain signals. In other words, autocorrelation determines the presence of correlation between the values of variables that are based on associated aspects. In a way, it is the cross-correlation of a signal with itself. Most of the statistical models are based upon the assumption of instance independence, which gets desecrated by autocorrelation. Autocorrelation is very useful in activities where there are repeated signals. Autocorrelation is used to determine the periodic signals that get obscured beneath the noise, or it is used to realize the fundamental frequency of a signal that does not have that frequency as a component but applies it there along with many harmonic frequencies.
For a free consultation on autocorrelation or other statistical tests, click here.
In statistics, the correlation of a method against a time-shifted version is the autocorrelation of a discrete time series method. There are many applications where autocorrelation is very useful. Autocorrelation is used for the measurement of the optical spectra, and for the measurements of minutely lived light pulses that are produced by lasers. This is done with the help of optical autocorrelators. Autocorrelation is used in optics, where the normalized autocorrelations and cross correlations together give the degree of coherence of an electromagnetic field. Autocorrelation is also used for signal processing. Autocorrelation can help you get information about repetitive events like musical beats or pulsar frequencies, but it can still not give the position in time of the beats. Thus, autocorrelation can be used for identifying non-randomness in data and to classify an appropriate time series model for the non random data.
In other words it can be said that autocorrelation is a correlation coefficient, where the correlation is not between the two different variables. Rather, it is between the different values of the same variables. When autocorrelation is used to spot non–randomness, it is mostly the autocorrelation of the first lag that is taken into consideration. Autocorrelation used to determine a proper time series model is conventionally plotted for various lags.
Thus, we can say that autocorrelation is useful for answering two main questions:
1. Was what was generated taken from a random process or a non–random process?
2. Which model is a more appropriate model for the generated data?
Autocorrelation has many properties and it is used for diverse studies and diverse dimensions. The properties of autocorrelation are interchangeable in different dimensions. The most important quality of autocorrelation is symmetry and consistency. Autocorrelation for periodic functions is also periodic, with a similar period. If the functions are completely uncorrelated, then their autocorrelation is the sum of the autocorrelations of each function separately. Autocorrelation is a unique kind of cross correlation, and thus it bears all the properties of cross correlation. The autocorrelation function is easily available is general statistical software programs, thus one can easily access autocorrelation.
Request
To request a blog written on a specific topic, please email James@StatisticsSolutions.com with your suggestion. Thank you!
Showing posts with label autocorrelation. Show all posts
Showing posts with label autocorrelation. Show all posts
Wednesday, June 3, 2009
Thursday, May 7, 2009
Autocorrelation
In cross sectional studies, data often involves households (in a consumption function analysis) or firms (in an investment study analysis), so if by chance an error term of a particular household or firm gets correlated with some other household or firm, then such correlations are termed as autocorrelation. Autocorrelation also occurs in the case of time series data. In the case of time series data, if the observations exhibit intercorrelations, especially when the time interval between the successive observations are short, then those intercorrelations are nothing but autocorrelation.
So, the term autocorrelation is defined as the correlation between the members of the series of the observations that are ordered with respect to time. Let us discuss two cases based on cross sectional and time series data to explain autocorrelation in a much better manner. In the case of cross sectional data, if a change in the income of a particular person affects the consumption expenditure of another household (other than his), then autocorrelation is present in the data. In the case of time series data, if an output is low in one quarter due to a labor strike, and if the data shows low output continues in the next quarter as well, then autocorrelation is present in the data.
Autocorrelation can be defined as the lag correlation of a given series with itself, lagged by a number of time units. On the other hand, serial autocorrelation defines the lag correlation between the two series in time series data.
There are certain patterns of autocorrelation, thus autocorrelation comes with various patterns. One example would be showing a discernible pattern among the residual errors. Autocorrelation exists when the residual error shows a cyclical pattern or an upward or downward trend in the disturbances, etc.
One of the major reasons for autocorrelation is the inertia or sluggishness caused in the time series data. The usage of an incorrect functional form also becomes a reason for an autocorrelation to occur.
The manipulation involving the extrapolation and the interpolation in the data also becomes a reason for autocorrelation. In this, the time series data is averaged so that smoothness occurs in the data, and this smoothness in the data exhibits a systematic pattern which in turn, introduces autocorrelation in the data.
A non stationarity property in the time series data also gives rise to the phenomenon of autocorrelation. Therefore, in order to make the time series free of autocorrelation, the researcher should make the data stationary.
Researchers should know that autocorrelation can be positive as well as negative. Economic time series generally exhibits positive autocorrelation as the series move in an upward or downward pattern. If the series move in a constant upward and downward movement, then autocorrelation is negative.
The major consequence of using ordinary least square (OLS) in the presence of autocorrelation is that it simply makes the estimator inefficient. As a result, the hypothesis testing procedures will give inaccurate results due to the presence of autocorrelation.
There is a popular test called the Durbin Watson test that detects the presence of autocorrelation. This test is conducted under the null hypothesis that there is no autocorrelation in the data. A test statistic called‘d’ is computed. “d” is defined as the ratio between the sum of the square of the difference in the residuals with ith and (i-1) time and the square of the residual in ith time. If the upper critical value of the test comes out to be less than the value of ‘d,’ then there is no autocorrelation. If the lower critical value of the test is more than the value of ‘d’ then there is autocorrelation.If one detects autocorrelation in the data, then the first thing a researcher should do is to try to find whether or not the autocorrelation is pure. If it is pure autocorrelation, then one can transform it into the original model, which is free from pure autocorrelation.
So, the term autocorrelation is defined as the correlation between the members of the series of the observations that are ordered with respect to time. Let us discuss two cases based on cross sectional and time series data to explain autocorrelation in a much better manner. In the case of cross sectional data, if a change in the income of a particular person affects the consumption expenditure of another household (other than his), then autocorrelation is present in the data. In the case of time series data, if an output is low in one quarter due to a labor strike, and if the data shows low output continues in the next quarter as well, then autocorrelation is present in the data.
Autocorrelation can be defined as the lag correlation of a given series with itself, lagged by a number of time units. On the other hand, serial autocorrelation defines the lag correlation between the two series in time series data.
There are certain patterns of autocorrelation, thus autocorrelation comes with various patterns. One example would be showing a discernible pattern among the residual errors. Autocorrelation exists when the residual error shows a cyclical pattern or an upward or downward trend in the disturbances, etc.
One of the major reasons for autocorrelation is the inertia or sluggishness caused in the time series data. The usage of an incorrect functional form also becomes a reason for an autocorrelation to occur.
The manipulation involving the extrapolation and the interpolation in the data also becomes a reason for autocorrelation. In this, the time series data is averaged so that smoothness occurs in the data, and this smoothness in the data exhibits a systematic pattern which in turn, introduces autocorrelation in the data.
A non stationarity property in the time series data also gives rise to the phenomenon of autocorrelation. Therefore, in order to make the time series free of autocorrelation, the researcher should make the data stationary.
Researchers should know that autocorrelation can be positive as well as negative. Economic time series generally exhibits positive autocorrelation as the series move in an upward or downward pattern. If the series move in a constant upward and downward movement, then autocorrelation is negative.
The major consequence of using ordinary least square (OLS) in the presence of autocorrelation is that it simply makes the estimator inefficient. As a result, the hypothesis testing procedures will give inaccurate results due to the presence of autocorrelation.
There is a popular test called the Durbin Watson test that detects the presence of autocorrelation. This test is conducted under the null hypothesis that there is no autocorrelation in the data. A test statistic called‘d’ is computed. “d” is defined as the ratio between the sum of the square of the difference in the residuals with ith and (i-1) time and the square of the residual in ith time. If the upper critical value of the test comes out to be less than the value of ‘d,’ then there is no autocorrelation. If the lower critical value of the test is more than the value of ‘d’ then there is autocorrelation.If one detects autocorrelation in the data, then the first thing a researcher should do is to try to find whether or not the autocorrelation is pure. If it is pure autocorrelation, then one can transform it into the original model, which is free from pure autocorrelation.
Subscribe to:
Posts (Atom)