One of the assumptions of the classical linear regression model is that there is no heteroscedasticity. Breaking this assumption means that the
Gauss–Markov theorem does not apply, meaning that
OLS estimators are not the
Best Linear Unbiased Estimators (BLUE) and their variance is not the lowest of all other unbiased estimators. Heteroscedasticity does
not cause ordinary least squares coefficient estimates to be biased, although it can cause ordinary least squares estimates of the variance (and, thus, standard errors) of the coefficients to be biased, possibly above or below the true of population variance. Thus, regression analysis using heteroscedastic data will still provide an unbiased estimate for the relationship between the predictor variable and the outcome, but standard errors and therefore inferences obtained from data analysis are suspect. Biased standard errors lead to biased inference, so results of hypothesis tests are possibly wrong. For example, if OLS is performed on a heteroscedastic data set, yielding biased standard error estimation, a researcher might fail to reject a null hypothesis at a given
significance level, when that null hypothesis was actually uncharacteristic of the actual population (making a
type II error). Under certain assumptions, the OLS estimator has a normal
asymptotic distribution when properly normalized and centered (even when the data does not come from a
normal distribution). This result is used to justify using a normal distribution, or a
chi square distribution (depending on how the
test statistic is calculated), when conducting a
hypothesis test. This holds even under heteroscedasticity. More precisely, the OLS estimator in the presence of heteroscedasticity is asymptotically normal, when properly normalized and centered, with a variance-covariance
matrix that differs from the case of homoscedasticity. In 1980, White proposed a
consistent estimator for the variance-covariance matrix of the asymptotic distribution of the OLS estimator. The
F test can still be used in some circumstances. However, it has been said that students in
econometrics should not overreact to heteroscedasticity. In addition, another word of caution was in the form, "heteroscedasticity has never been a reason to throw out an otherwise good model." With the advent of
heteroscedasticity-consistent standard errors allowing for inference without specifying the conditional second moment of error term, testing conditional homoscedasticity is not as important as in the past. Yet, in the context of binary choice models (
Logit or
Probit), heteroscedasticity will only result in a positive scaling effect on the asymptotic mean of the misspecified MLE (i.e. the model that ignores heteroscedasticity). As a result, the predictions which are based on the misspecified MLE will remain correct. In addition, the misspecified Probit and Logit MLE will be asymptotically normally distributed which allows performing the usual significance tests (with the appropriate variance-covariance matrix). However, regarding the general hypothesis testing, as pointed out by
Greene, "simply computing a robust covariance matrix for an otherwise inconsistent estimator does not give it redemption. Consequently, the virtue of a robust covariance matrix in this setting is unclear." == Correction ==