next up previous
Next: Covariance and Correlation. Up: Theory: Covariance & Correlation Previous: Theory: Covariance & Correlation

Review of Mathematical Expectation.

Suppose X is a r.v. (random variable) with density tex2html_wrap_inline593 . If its distribution is discrete, then the density is also known as the probability mass function and in fact gives the ``point'' probabilities:

displaymath577

Probabilities of sets with multiple points are obtained by summing:

displaymath578

If the distribution of X is continuous, then all point probabilities are 0 and the density gives probabilities of intervals through integration:

displaymath579

For The mathematical expectation of a function of a r.v. is defined to be

displaymath580

The formulae displayed here illustrate the general principle: summations for discrete r.v.'s and integration for continuous r.v.'s. Also, whenever the limits of integration or summation are not shown, it is assumed that they are over the entire ``space,'' which effectively means all values of x where tex2html_wrap_inline611 .

Technical Note: To define E[h(X)], it is usually required that either (i) tex2html_wrap_inline615 so that the summands or integrand is never negative, and then E[h(X)] may possibly be tex2html_wrap_inline619 , or (ii) tex2html_wrap_inline621 , which means effectively that the summation or integration converges absolutely. We will generally not bother ourselves with such details. We always asssume that the integral or summation satisfies whatever mathematical properties are needed for things to make sense. There are very few practical situations where problems of infinite expectation arise.

The connection between mathematical expectation and data comes through the notion of long run averages: If tex2html_wrap_inline623 , tex2html_wrap_inline625 , tex2html_wrap_inline627 , tex2html_wrap_inline629 is a sample of realized values of the r.v. X, then as tex2html_wrap_inline633 the sample mean tex2html_wrap_inline635 tends to E[h(X)]. The precise mathematical formulation of this ``principle'' is the Law of Large Numbers, which makes certain assumptions on how the sample is generated (e.g. that the sample are realized values of independent and identically distributed (abbreviated i.i.d.) random variables with the same distribution as X).

Of course, there are certain mathematical expectations which are of most interest, namely the mean and variance of the r.v.:

eqnarray193

Some useful properties of these mathematical ``operators'' are summarized in the next proposition. Note that if X and Y are r.v.'s, then so are h(X) and g(X,Y) for any appropriately defined real valued functions h(x) and g(x,y).

  proposition195

Proof. Part (i) is proved in Hogg & Craig. For part (ii), assuming X is a continuous r.v. we have

eqnarray203

For (iii) we apply the version of Chebyshev's inequality that says if X is a nonnegative r.v. and c > 0 then P[X > c] tex2html_wrap_inline687 E[X]/c. See Hogg & Craig. Since E[X] = 0, it follows that P[X > c] = 0 for all c > 0. In particular, P[X > 0] = tex2html_wrap_inline701 tex2html_wrap_inline687 tex2html_wrap_inline705 = 0. Since we know tex2html_wrap_inline711 we have 1 = tex2html_wrap_inline717 = P[X = 0] + P[X > 0] = P[X=0].
tex2html_wrap_inline727

  proposition214

Proof. Since the r.v. tex2html_wrap_inline741 is nonnegative, it follows that tex2html_wrap_inline743 = tex2html_wrap_inline747 tex2html_wrap_inline749 0 by part (ii) of Proposition 1. Continuing with the fact that tex2html_wrap_inline741 tex2html_wrap_inline749 0, if 0 =t tex2html_wrap_inline743 = tex2html_wrap_inline747 , then by part (iii) of Proposition 1 it follows that the r.v. tex2html_wrap_inline741 = 0 with probability 1, i.e.\ tex2html_wrap_inline777 with probability 1. Since tex2html_wrap_inline781 is a constant, the completes the proof of part (i) of Proposition 2.

For part (ii), note from part (i) of Proposition 1 that E[a X + b] = a E[X] + b, so

eqnarray226

Part (iii) is already proved in Hogg & Craig, right after the definition of variance.
tex2html_wrap_inline727

The corresponding ``sample'' quantities are given by

  eqnarray228

An ``alternative'' sample variance is sometimes considered::

  equation239

The difference between the two sample variances is unimportant when n is large.


next up previous
Next: Covariance and Correlation. Up: Theory: Covariance & Correlation Previous: Theory: Covariance & Correlation

Dennis Cox
Tue Jan 21 09:20:27 CST 1997