âšī¸ Skipped - page is already crawled
| Filter | Status | Condition | Details |
|---|---|---|---|
| HTTP status | PASS | download_http_code = 200 | HTTP 200 |
| Age cutoff | PASS | download_stamp > now() - 6 MONTH | 0.4 months ago |
| History drop | PASS | isNull(history_drop_reason) | No drop reason |
| Spam/ban | PASS | fh_dont_index != 1 AND ml_spam_score = 0 | ml_spam_score=0 |
| Canonical | PASS | meta_canonical IS NULL OR = '' OR = src_unparsed | Not set |
| Property | Value |
|---|---|
| URL | https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties |
| Last Crawled | 2026-04-01 02:46:51 (11 days ago) |
| First Indexed | 2016-05-05 07:46:26 (9 years ago) |
| HTTP Status Code | 200 |
| Meta Title | Properties of the OLS estimator | Consistency, asymptotic normality |
| Meta Description | Learn what conditions are needed to prove the consistency and asymptotic normality of the OLS estimator. |
| Meta Canonical | null |
| Boilerpipe Text | In this lecture we discuss under which assumptions the OLS (Ordinary Least
Squares) estimator has desirable statistical properties such as consistency
and asymptotic normality.
Table of contents
The regression model
Matrix notation
The estimator
Writing the estimator in terms of sample means
Consistency of the OLS estimator
Assumption 1 - Convergence of sample means to population means
Assumption 2 - Full rank
Assumption 3 - Orthogonality
Proof of consistency
Asymptotic normality of the OLS estimator
Assumption 4 - CLT condition
Proof of asymptotic normality
Consistent estimation of the variance of the error terms
Assumption 5 - Regularity of error terms
Proof of consistency
Consistent estimation of the asymptotic covariance matrix
Proof of consistency
Consistent estimation of the long-run covariance matrix
Assumption 6 - No serial correlation
More explicit formulae for the long-run covariance
Proof of consistency under Assumption 6
Formula for the covariance matrix of the OLS estimator under Assumption 6
Assumption 7 - Conditional homoskedasticity
Proof of consistency under Assumption 7
Formula for the covariance matrix of the OLS estimator under Assumptions 6 and 7
Weaker assumptions
Hypothesis testing
References
The regression model
Consider the
linear
regression
model
where:
the outputs are denoted by
;
the associated
vectors of inputs are denoted by
;
the
vector of regression coefficients is denoted by
;
are unobservable error terms.
Matrix notation
We assume to observe a sample of
realizations, so that the vector of all outputs
is
an
vector, the
design
matrix
is
an
matrix, and the vector of error
terms
is
an
vector.
The estimator
The OLS estimator
is the vector of regression coefficients that minimizes the sum of squared
residuals:
As proved in the lecture on
Linear
regression
, if the design matrix
has full rank, then the OLS estimator is computed as
follows:
Writing the estimator in terms of sample means
The OLS estimator can be written as
where
is
the
sample mean
of the
matrix
and
is
the sample mean of the
matrix
.
Consistency of the OLS estimator
In this section we are going to propose a set of conditions that are
sufficient for the
consistency
of the OLS estimator, that is, for the convergence in probability of
to the true value
.
Assumption 1 - Convergence of sample means to
population means
The first assumption we make is that the sample means in the OLS formula
converge to their population counterparts, which is formalized as follows.
Assumption 1 (convergence)
: both the sequence
and the sequence
satisfy sets of conditions that are sufficient for the
convergence in probability
of their sample means
to the population means
and
,
which do not depend on
.
For example, the sequences
and
could be assumed to satisfy the conditions of
Chebyshev's Weak Law of Large Numbers for
correlated sequences
, which are quite mild (basically, it is only required
that the sequences are
covariance stationary
and
that their auto-covariances are zero on average).
Assumption 2 - Full rank
The second assumption we make is a rank assumption (sometimes also called
identification assumption).
Assumption 2 (rank)
: the square matrix
has
full rank
(as a
consequence, it is
invertible
).
Assumption 3 - Orthogonality
The third assumption we make is that the regressors
are orthogonal to the error terms
.
Assumption 3 (orthogonality)
: For each
,
and
are orthogonal, that
is,
Proof of consistency
It is then straightforward to prove the following proposition.
Proposition
If Assumptions 1, 2 and 3 are satisfied, then the OLS estimator
is a consistent estimator of
.
Proof
Asymptotic normality of the OLS estimator
We now introduce a new assumption, and we use it to prove the asymptotic
normality of the OLS estimator.
Assumption 4 - CLT condition
The assumption is as follows.
Assumption 4 (Central Limit Theorem)
: the sequence
satisfies a set of conditions that are sufficient to guarantee that a Central
Limit Theorem applies to its sample
mean
For a review of some of the conditions that can be imposed on a sequence to
guarantee that a Central Limit Theorem applies to its sample mean, you can go
to the lecture on the
Central Limit
Theorem
.
In any case, remember that if a Central Limit Theorem applies to
,
then, as
tends to
infinity,
converges
in distribution
to a
multivariate normal
distribution
with mean equal to
and covariance matrix equal
to
Proof of asymptotic normality
With Assumption 4 in place, we are now able to prove the asymptotic normality
of the OLS estimator.
Proposition
If Assumptions 1, 2, 3 and 4 are satisfied, then the OLS estimator
is asymptotically multivariate normal with mean equal to
and asymptotic covariance matrix equal
to
that
is,
where
has been defined above.
Proof
As in the proof of consistency, the
dependence of the estimator on the sample size is made explicit, so that the
OLS estimator is denoted by
.
First of all, we have
where,
in the last step, we have used the fact that, by Assumption 3,
.
Note that, by Assumption 1 and the Continuous Mapping theorem, we
have
Furthermore,
by Assumption 4, we have
that
converges
in distribution to a multivariate normal random vector having mean equal to
and covariance matrix equal to
.
Thus, by Slutski's theorem, we have
that
converges
in distribution to a multivariate normal vector with mean equal to
and covariance matrix equal to
Consistent estimation of the variance of the error terms
We now discuss the consistent estimation of the variance of the error terms.
Assumption 5 - Regularity of error terms
Here is an additional assumption.
Assumption 5
: the sequence
satisfies a set of conditions that are sufficient for the convergence in
probability of its sample
mean
to
the population mean
which
does not depend on
.
Proof of consistency
If this assumption is satisfied, then the variance of the error terms
can be estimated by the sample variance of the
residuals
where
Proposition
Under Assumptions 1, 2, 3, and 5, it can be proved that
is a consistent estimator of
.
Proof
Consistent estimation of the asymptotic covariance matrix
We have proved that the asymptotic covariance matrix of the OLS estimator
is
where
the long-run covariance matrix
is defined
by
Usually, the matrix
needs to be estimated because it depends on quantities
(
and
)
that are not known.
Proof of consistency
The next proposition characterizes consistent estimators of
.
Proposition
If Assumptions 1, 2, 3, 4 and 5 are satisfied, and a consistent estimator
of the long-run covariance matrix
is available, then the asymptotic variance of the OLS estimator is
consistently estimated
by
Proof
Thus, in order to derive a consistent estimator of the covariance matrix of
the OLS estimator, we need to find a consistent estimator of the long-run
covariance matrix
.
How to do this is discussed in the next section.
Consistent estimation of the long-run covariance matrix
The estimation of
requires some assumptions on the covariances between the terms of the sequence
.
Assumption 6 - No serial correlation
In order to find a simpler expression for
,
we make the following assumption.
Assumption 6
: the sequence
is
serially
uncorrelated
, that
is,
and
weakly stationary
, that is,
does
not depend on
.
Remember that in Assumption 3 (orthogonality) we also ask
that
More explicit formulae for the long-run covariance
We now derive simpler expressions for
.
Proposition
Under Assumptions 3 (orthogonality), the long-run covariance matrix
satisfies
Proof
This is proved as
follows:
Proposition
Under Assumptions 3 (orthogonality) and 6 (no serial correlation), the
long-run covariance matrix
satisfies
Proof
The proof is as
follows:
Proof of consistency under Assumption 6
Thanks to assumption 6, we can also derive an estimator of
.
Proposition
Suppose that Assumptions 1, 2, 3, 4 and 6 are satisfied, and that
is consistently estimated by the sample
mean
Then,
the long-run covariance matrix
is consistently estimated
by
Proof
Formula for the covariance matrix of the OLS
estimator under Assumption 6
When the assumptions of the previous proposition hold, the asymptotic
covariance matrix of the OLS estimator
is
As a consequence, the covariance of the OLS estimator can be approximated
by
which
is known as
heteroskedasticity-robust
estimator
.
Assumption 7 - Conditional homoskedasticity
A further assumption is often made, which allows us to further simplify the
expression for the long-run covariance matrix.
Assumption 7
: the error terms are
conditionally
homoskedastic
:
Proof of consistency under Assumption 7
This assumption has the following implication.
Proposition
Suppose that Assumptions 1, 2, 3, 4, 5, 6 and 7 are satisfied. Then, the
long-run covariance matrix
is consistently estimated
by
Proof
Formula for the covariance matrix of the OLS
estimator under Assumptions 6 and 7
When the assumptions of the previous proposition hold, the asymptotic
covariance matrix of the OLS estimator
is
As a consequence, the covariance of the OLS estimator can be approximated
by
which
is the same estimator derived in the
normal
linear regression model
.
Weaker assumptions
The assumptions above can be made even weaker (for example, by relaxing the
hypothesis that
is uncorrelated with
),
at the cost of facing more difficulties in estimating the long-run covariance
matrix.
For a review of the methods that can be used to estimate
,
see, for example, Den and Levin (1996).
Hypothesis testing
The lecture entitled
Linear
regression - Hypothesis testing
discusses how to carry out
hypothesis tests
on the coefficients of a linear regression model in the cases discussed above,
that is, when the OLS estimator is asymptotically normal and a consistent
estimator of the asymptotic covariance matrix is available.
References
Haan, Wouter J. Den, and Andrew T. Levin (1996). "Inferences from parametric
and non-parametric covariance matrix estimation procedures." Technical Working
Paper Series, NBER.
How to cite
Please cite as:
Taboga, Marco (2021). "Properties of the OLS estimator", Lectures on probability theory and mathematical statistics. Kindle Direct Publishing. Online appendix. https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties. |
| Markdown | 
[StatLect](https://www.statlect.com/)
[Index](https://www.statlect.com/) \> [Fundamentals of statistics](https://www.statlect.com/fundamentals-of-statistics/)
# Properties of the OLS estimator
by [Marco Taboga](https://www.statlect.com/about/#author), PhD
In this lecture we discuss under which assumptions the OLS (Ordinary Least Squares) estimator has desirable statistical properties such as consistency and asymptotic normality.


Table of contents
1. [The regression model](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid2)
2. [Matrix notation](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid3)
3. [The estimator](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid4)
4. [Writing the estimator in terms of sample means](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid5)
5. [Consistency of the OLS estimator](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid6)
1. [Assumption 1 - Convergence of sample means to population means](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid7)
2. [Assumption 2 - Full rank](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid8)
3. [Assumption 3 - Orthogonality](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid9)
4. [Proof of consistency](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid10)
6. [Asymptotic normality of the OLS estimator](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid11)
1. [Assumption 4 - CLT condition](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid12)
2. [Proof of asymptotic normality](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid13)
7. [Consistent estimation of the variance of the error terms](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid14)
1. [Assumption 5 - Regularity of error terms](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid15)
2. [Proof of consistency](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid16)
8. [Consistent estimation of the asymptotic covariance matrix](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid17)
1. [Proof of consistency](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid18)
9. [Consistent estimation of the long-run covariance matrix](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid19)
1. [Assumption 6 - No serial correlation](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid20)
2. [More explicit formulae for the long-run covariance](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid21)
3. [Proof of consistency under Assumption 6](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid22)
4. [Formula for the covariance matrix of the OLS estimator under Assumption 6](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid23)
5. [Assumption 7 - Conditional homoskedasticity](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid24)
6. [Proof of consistency under Assumption 7](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid25)
7. [Formula for the covariance matrix of the OLS estimator under Assumptions 6 and 7](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid26)
8. [Weaker assumptions](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid27)
10. [Hypothesis testing](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid28)
11. [References](https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties#hid29)
## The regression model
Consider the [linear regression](https://www.statlect.com/fundamentals-of-statistics/linear-regression) model![\[eq1\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)where:
- the outputs are denoted by ;
- the associated  vectors of inputs are denoted by ;
- the  vector of regression coefficients is denoted by ;
-  are unobservable error terms.
## Matrix notation
We assume to observe a sample of  realizations, so that the vector of all outputs
![\[eq2\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)is an  vector, the [design matrix](https://www.statlect.com/glossary/design-matrix)![\[eq3\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)is an  matrix, and the vector of error terms![\[eq4\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)is an  vector.
## The estimator
The OLS estimator  is the vector of regression coefficients that minimizes the sum of squared residuals:![\[eq5\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
As proved in the lecture on [Linear regression](https://www.statlect.com/fundamentals-of-statistics/linear-regression), if the design matrix  has full rank, then the OLS estimator is computed as follows:![\[eq6\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
## Writing the estimator in terms of sample means
The OLS estimator can be written as ![\[eq7\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)where ![\[eq8\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)is the [sample mean](https://www.statlect.com/glossary/sample-mean) of the  matrix  and ![\[eq9\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)is the sample mean of the  matrix .
## Consistency of the OLS estimator
In this section we are going to propose a set of conditions that are sufficient for the [consistency](https://www.statlect.com/glossary/consistent-estimator) of the OLS estimator, that is, for the convergence in probability of  to the true value .
### Assumption 1 - Convergence of sample means to population means
The first assumption we make is that the sample means in the OLS formula converge to their population counterparts, which is formalized as follows.
**Assumption 1 (convergence)**: both the sequence ![\[eq10\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) and the sequence ![\[eq11\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) satisfy sets of conditions that are sufficient for the [convergence in probability](https://www.statlect.com/asymptotic-theory/convergence-in-probability) of their sample means to the population means ![\[eq12\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) and ![\[eq13\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==), which do not depend on .
For example, the sequences ![\[eq14\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) and ![\[eq15\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) could be assumed to satisfy the conditions of [Chebyshev's Weak Law of Large Numbers for correlated sequences](https://www.statlect.com/asymptotic-theory/law-of-large-numbers#cheby2), which are quite mild (basically, it is only required that the sequences are [covariance stationary](https://www.statlect.com/glossary/covariance-stationary) and that their auto-covariances are zero on average).
### Assumption 2 - Full rank
The second assumption we make is a rank assumption (sometimes also called identification assumption).
**Assumption 2 (rank)**: the square matrix ![\[eq16\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) has [full rank](https://www.statlect.com/matrix-algebra/rank-of-a-matrix) (as a consequence, it is [invertible](https://www.statlect.com/matrix-algebra/inverse-matrix)).
### Assumption 3 - Orthogonality
The third assumption we make is that the regressors  are orthogonal to the error terms .
**Assumption 3 (orthogonality)**: For each ,  and  are orthogonal, that is,![\[eq17\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
### Proof of consistency
It is then straightforward to prove the following proposition.
Proposition If Assumptions 1, 2 and 3 are satisfied, then the OLS estimator  is a consistent estimator of .
Proof
Let us make explicit the dependence of the estimator on the sample size and denote by ![\[eq18\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) the OLS estimator obtained when the sample size is equal to  By Assumption 1 and by the [Continuous Mapping theorem](https://www.statlect.com/asymptotic-theory/continuous-mapping-theorem), we have that the probability limit of ![\[eq19\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) is ![\[eq20\]](https://www.statlect.com/images/OLS-estimator-properties__47.png)Now, if we pre-multiply the regression equation![\[eq21\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)by  and we take expected values, we get![\[eq22\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)But by Assumption 3, it becomes![\[eq23\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)or![\[eq24\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)which implies that![\[eq25\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
## Asymptotic normality of the OLS estimator
We now introduce a new assumption, and we use it to prove the asymptotic normality of the OLS estimator.
### Assumption 4 - CLT condition
The assumption is as follows.
**Assumption 4 (Central Limit Theorem)**: the sequence ![\[eq26\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) satisfies a set of conditions that are sufficient to guarantee that a Central Limit Theorem applies to its sample mean![\[eq27\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
For a review of some of the conditions that can be imposed on a sequence to guarantee that a Central Limit Theorem applies to its sample mean, you can go to the lecture on the [Central Limit Theorem](https://www.statlect.com/asymptotic-theory/central-limit-theorem).
In any case, remember that if a Central Limit Theorem applies to ![\[eq26\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==), then, as  tends to infinity,![\[eq29\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) [converges in distribution](https://www.statlect.com/asymptotic-theory/convergence-in-distribution) to a [multivariate normal distribution](https://www.statlect.com/probability-distributions/multivariate-normal-distribution) with mean equal to  and covariance matrix equal to![\[eq30\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
### Proof of asymptotic normality
With Assumption 4 in place, we are now able to prove the asymptotic normality of the OLS estimator.
Proposition If Assumptions 1, 2, 3 and 4 are satisfied, then the OLS estimator  is asymptotically multivariate normal with mean equal to  and asymptotic covariance matrix equal to![\[eq31\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)that is,![\[eq32\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)where  has been defined above.
Proof
As in the proof of consistency, the dependence of the estimator on the sample size is made explicit, so that the OLS estimator is denoted by ![\[eq33\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==). First of all, we have ![\[eq34\]](https://www.statlect.com/images/OLS-estimator-properties__67.png)where, in the last step, we have used the fact that, by Assumption 3, ![\[eq35\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==). Note that, by Assumption 1 and the Continuous Mapping theorem, we have![\[eq36\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)Furthermore, by Assumption 4, we have that![\[eq37\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)converges in distribution to a multivariate normal random vector having mean equal to  and covariance matrix equal to . Thus, by Slutski's theorem, we have that![\[eq38\]](https://www.statlect.com/images/OLS-estimator-properties__73.png)converges in distribution to a multivariate normal vector with mean equal to  and covariance matrix equal to ![\[eq39\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
## Consistent estimation of the variance of the error terms
We now discuss the consistent estimation of the variance of the error terms.
### Assumption 5 - Regularity of error terms
Here is an additional assumption.
**Assumption 5**: the sequence ![\[eq40\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) satisfies a set of conditions that are sufficient for the convergence in probability of its sample mean![\[eq41\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)to the population mean ![\[eq42\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)which does not depend on .
### Proof of consistency
If this assumption is satisfied, then the variance of the error terms  can be estimated by the sample variance of the residuals![\[eq43\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)where ![\[eq44\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
Proposition Under Assumptions 1, 2, 3, and 5, it can be proved that ![\[eq45\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) is a consistent estimator of .
Proof
Let us make explicit the dependence of the estimators on the sample size and denote by ![\[eq46\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) and ![\[eq47\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) the estimators obtained when the sample size is equal to  By Assumption 1 and by the [Continuous Mapping theorem](https://www.statlect.com/asymptotic-theory/continuous-mapping-theorem), we have that the probability limit of ![\[eq47\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) is ![\[eq49\]](https://www.statlect.com/images/OLS-estimator-properties__89.png)where: in steps  and  we have used the Continuous Mapping Theorem; in step  we have used Assumption 5; in step  we have used the fact that ![\[eq50\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)because ![\[eq51\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) is a consistent estimator of , as proved above.
## Consistent estimation of the asymptotic covariance matrix
We have proved that the asymptotic covariance matrix of the OLS estimator is![\[eq52\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)where the long-run covariance matrix  is defined by![\[eq53\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
Usually, the matrix ![\[eq54\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) needs to be estimated because it depends on quantities ( and ![\[eq16\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)) that are not known.
### Proof of consistency
The next proposition characterizes consistent estimators of ![\[eq56\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==).
Proposition If Assumptions 1, 2, 3, 4 and 5 are satisfied, and a consistent estimator  of the long-run covariance matrix  is available, then the asymptotic variance of the OLS estimator is consistently estimated by![\[eq57\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
Proof
This is proved as follows![\[eq58\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)where: in step  we have used the Continuous Mapping theorem; in step  we have used the hypothesis that  is a consistent estimator of the long-run covariance matrix  and the fact that, by Assumption 1, the sample mean of the matrix  is a consistent estimator of ![\[eq59\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==), that is![\[eq60\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
Thus, in order to derive a consistent estimator of the covariance matrix of the OLS estimator, we need to find a consistent estimator of the long-run covariance matrix . How to do this is discussed in the next section.
## Consistent estimation of the long-run covariance matrix
The estimation of  requires some assumptions on the covariances between the terms of the sequence ![\[eq61\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==).
### Assumption 6 - No serial correlation
In order to find a simpler expression for , we make the following assumption.
**Assumption 6**: the sequence ![\[eq62\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) is [serially uncorrelated](https://www.statlect.com/fundamentals-of-statistics/autocorrelation), that is,![\[eq63\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)and [weakly stationary](https://www.statlect.com/glossary/covariance-stationary), that is, ![\[eq64\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)does not depend on .
Remember that in Assumption 3 (orthogonality) we also ask that![\[eq65\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
### More explicit formulae for the long-run covariance
We now derive simpler expressions for .
Proposition Under Assumptions 3 (orthogonality), the long-run covariance matrix  satisfies![\[eq66\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
Proof
This is proved as follows:![\[eq67\]](https://www.statlect.com/images/OLS-estimator-properties__128.png)
Proposition Under Assumptions 3 (orthogonality) and 6 (no serial correlation), the long-run covariance matrix  satisfies![\[eq68\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
Proof
The proof is as follows:![\[eq69\]](https://www.statlect.com/images/OLS-estimator-properties__131.png)
### Proof of consistency under Assumption 6
Thanks to assumption 6, we can also derive an estimator of .
Proposition Suppose that Assumptions 1, 2, 3, 4 and 6 are satisfied, and that ![\[eq70\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) is consistently estimated by the sample mean![\[eq71\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)Then, the long-run covariance matrix  is consistently estimated by![\[eq72\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
Proof
We have![\[eq73\]](https://www.statlect.com/images/OLS-estimator-properties__137.png)where in the last step we have applied the Continuous Mapping theorem separately to each entry of the matrices in square brackets, together with the fact that ![\[eq74\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)To see how this is done, consider, for example, the matrix![\[eq75\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)Then, the entry at the intersection of its \-th row and \-th column is![\[eq76\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)and![\[eq77\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
### Formula for the covariance matrix of the OLS estimator under Assumption 6
When the assumptions of the previous proposition hold, the asymptotic covariance matrix of the OLS estimator is![\[eq78\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
As a consequence, the covariance of the OLS estimator can be approximated by![\[eq79\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)which is known as [heteroskedasticity-robust estimator](https://www.statlect.com/glossary/heteroskedasticity-robust-standard-errors).
### Assumption 7 - Conditional homoskedasticity
A further assumption is often made, which allows us to further simplify the expression for the long-run covariance matrix.
**Assumption 7**: the error terms are [conditionally homoskedastic](https://www.statlect.com/glossary/heteroskedasticity):![\[eq80\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
### Proof of consistency under Assumption 7
This assumption has the following implication.
Proposition Suppose that Assumptions 1, 2, 3, 4, 5, 6 and 7 are satisfied. Then, the long-run covariance matrix  is consistently estimated by![\[eq81\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
Proof
First of all, we have that![\[eq82\]](https://www.statlect.com/images/OLS-estimator-properties__149.png)But we know that, by Assumption 1, ![\[eq83\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) is consistently estimated by![\[eq84\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)and by Assumptions 1, 2, 3 and 5, ![\[eq85\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) is consistently estimated by![\[eq86\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)Therefore, by the Continuous Mapping theorem, the long-run covariance matrix  is consistently estimated by ![\[eq87\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
### Formula for the covariance matrix of the OLS estimator under Assumptions 6 and 7
When the assumptions of the previous proposition hold, the asymptotic covariance matrix of the OLS estimator is![\[eq88\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)
As a consequence, the covariance of the OLS estimator can be approximated by![\[eq89\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)which is the same estimator derived in the [normal linear regression model](https://www.statlect.com/fundamentals-of-statistics/normal-linear-regression-model).
### Weaker assumptions
The assumptions above can be made even weaker (for example, by relaxing the hypothesis that ![\[eq90\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==) is uncorrelated with ![\[eq91\]](data:image/gif;base64,R0lGODlhAQABAIAAANvf7wAAACH5BAEAAAAALAAAAAABAAEAAAICRAEAOw==)), at the cost of facing more difficulties in estimating the long-run covariance matrix.
For a review of the methods that can be used to estimate , see, for example, Den and Levin (1996).
## Hypothesis testing
The lecture entitled [Linear regression - Hypothesis testing](https://www.statlect.com/fundamentals-of-statistics/linear-regression-hypothesis-testing) discusses how to carry out [hypothesis tests](https://www.statlect.com/fundamentals-of-statistics/hypothesis-testing) on the coefficients of a linear regression model in the cases discussed above, that is, when the OLS estimator is asymptotically normal and a consistent estimator of the asymptotic covariance matrix is available.
## References
Haan, Wouter J. Den, and Andrew T. Levin (1996). "Inferences from parametric and non-parametric covariance matrix estimation procedures." Technical Working Paper Series, NBER.
## How to cite
Please cite as:
Taboga, Marco (2021). "Properties of the OLS estimator", Lectures on probability theory and mathematical statistics. Kindle Direct Publishing. Online appendix. https://www.statlect.com/fundamentals-of-statistics/OLS-estimator-properties.
The books
Most of the learning materials found on this website are now available in a traditional textbook format.
[Probability and statistics](https://www.statlect.com/about/book)[Matrix algebra](https://www.statlect.com/about/linear-algebra-book)
Featured pages
- [Poisson distribution](https://www.statlect.com/probability-distributions/Poisson-distribution)
- [Uniform distribution](https://www.statlect.com/probability-distributions/uniform-distribution)
- [Multinomial distribution](https://www.statlect.com/probability-distributions/multinomial-distribution)
- [Beta function](https://www.statlect.com/mathematical-tools/beta-function)
- [Student t distribution](https://www.statlect.com/probability-distributions/student-t-distribution)
- [Wishart distribution](https://www.statlect.com/probability-distributions/wishart-distribution)
Explore
- [Chi-square distribution](https://www.statlect.com/probability-distributions/chi-square-distribution)
- [Bayes rule](https://www.statlect.com/fundamentals-of-probability/Bayes-rule)
- [Exponential distribution](https://www.statlect.com/probability-distributions/exponential-distribution)
Main sections
- [Mathematical tools](https://www.statlect.com/mathematical-tools/)
- [Fundamentals of probability](https://www.statlect.com/fundamentals-of-probability/)
- [Probability distributions](https://www.statlect.com/probability-distributions/)
- [Asymptotic theory](https://www.statlect.com/asymptotic-theory/)
- [Fundamentals of statistics](https://www.statlect.com/fundamentals-of-statistics/)
- [Glossary](https://www.statlect.com/glossary/)
About
- [About Statlect](https://www.statlect.com/about/)
- [Contacts](https://www.statlect.com/about/contacts)
- [Cookies, privacy and terms of use](https://www.statlect.com/about/cookies-privacy-policy-and-terms-of-use)
Glossary entries
- [Factorial](https://www.statlect.com/glossary/factorial)
- [Alternative hypothesis](https://www.statlect.com/glossary/alternative-hypothesis)
- [Almost sure](https://www.statlect.com/glossary/almost-sure)
- [Type II error](https://www.statlect.com/glossary/Type-II-error)
- [Discrete random variable](https://www.statlect.com/glossary/discrete-random-variable)
- [Integrable variable](https://www.statlect.com/glossary/integrable-random-variable)
Share
- To enhance your privacy,
- we removed the social buttons,
- but **don't forget to share**. |
| Readable Markdown | null |
| Shard | 126 (laksa) |
| Root Hash | 3586688910177265926 |
| Unparsed URL | com,statlect!www,/fundamentals-of-statistics/OLS-estimator-properties s443 |