User:Jmath666/Conditional probability and expectation

From Wikipedia, the free encyclopedia

Elementary description[edit]

If are events such that , the conditional probability of the event given is defined by

If is fixed, the mapping is a conditional probability distribution given the event .

If also , then also

and so

which is known as the Bayes theorem.

Conditioning of discrete random variables[edit]

If is a discrete real random variable (that is, attaining only values , ), then the conditional probability of an event given that is

The mapping defines a conditional probability distribution given that .

Note that is a number, that is, a deterministic quantity. If we allow to be a realization of the random variable , we obtain conditional probability of the event given random variable , denoted by , which is a random variable itself. The conditional probability attains the value of with probability .

Now suppose and are two discrete real random variables with a joint distribution. Then the conditional probability distribution of given is

If we allow to be a realization of the random variable , we obtain the conditional distribution of random variable given random variable . Given , the random variable that attains the value with probability .

The random variables and are independent when the events and are independent for all and , that is,

Clearly, this is equivalent to

The conditional expectation of given the value is

which is defined whenever the marginal probability

This is a description common in statistics [1]. Note that is a number, that is, a deterministic quantity, and the particular value of does not matter; only the probabilities do.

If we allow to be a realization of the random variable , we obtain conditional expectation of random variable given random variable , denoted by . This form is closer to the mathematical form favored by probabilists (described in more detail below), and it is a random variable itself. The conditional expectation attains the value with probability .

Conditioning of continuous random variables[edit]

For continuous random variables , with joint density , the conditional probability density of given that is

where

is the marginal density of . The conventional notation is often used to mean the same as , that is, the function of two variables and . The notation , often used in practice, is ambigous, because if and are substituted for by something else (like specific numbers), the information what means is lost.

The continuous random variables are independent if, for all and , the events and are independent, which can be proved to be equivalent to

This is clearly equivalent to

The conditional probability density of given is the random function . The conditional expectation of given the value is

and the conditional expectation of given is the random variable

dependent on the values of .

Warning[edit]

Unfortunately, in the the literature, esp. more elementary oriented statistics texts, the authors do not always distinguish properly between conditioning given the value of a random variable (the result is a number) and conditioning given the random variable (the result is a random variable), so, confusingly enough, the words “ given the random variable\textquotedblright can mean either.

Mathematical synopsis[edit]

This section follows [2]. In probability theory, a conditional expectation (also known as conditional expected value or conditional mean) is the expected value of a random variable with respect to a conditional probability distribution, defined as follows.

If is a real random variable, and is an event with positive probability, then the conditional probability distribution of given assigns a probability to the Borel set . The mean (if it exists) of this conditional probability distribution of is denoted by and called the conditional expectation of given the event .

If is another random variable, then the conditional expectation of given that the value is a function of , let us say . An argument using the Radon-Nikodym theorem is needed to define properly because the event that may have probability zero. Also, is defined only for almost all , with respect to the distribution of . The conditional expectation of given random variable , denoted by , is the random variable .

It turns out that the conditional expectation is a function only of the sigma-algebra, say , generated by the events for Borel sets , rather than the particular values of . For a -algebra , the conditional expectation of given the -algebra is a random variable that is -measurable and whose integral over any -measurable set is the same as the integral of over the same set. The existence of this conditional expectation is proved from the Radon-Nikodym theorem. If happens to be -measurable, then .

If has an expected value, then the conditional expectation also has an expected value, which is the same as that of . This is the law of total expectation.

For simplicity, the presentation here is done for real-valued random variables, but generalization to probability on more general spaces, such as or normed metric spaces equipped with a probability measure, is immediate.

Mathematical prerequisites[edit]

Recall that probability space is , where is a -algebra of subsets of , and a probability measure with measurable sets. A random variable on the space is a -measurable function. is the sigma algebra of all Borel sets in . If is a set and a random variable, or are common shorthands for the event

Probability conditional on the value of a random variable[edit]

Let be probability space, a -measurable random variable with values in , (i.e., an event not necessarily independent of ), and . For and , the conditional probability of given is by definition

We wish to attach a meaning to the conditional probability of given even when . The following argument follows Wilks [3], who attributes it to Kolmogorov [4]. Fix and define

Since is -measurable, the set function is a measure on Borel sets . Define another measure on by

Clearly,

\newline and hence implies . Thus the measure is absolutely continuous with respect to the measure and by the Radon-Nykodym theorem, there exists a real-valued -measurable function such that

We interpret the function as the conditional probability of given ,

Once the conditional probability is defined, other concepts of probability follow, such as expectation and density.

One way to justify this interpretation is as the conditional probability of given the limit of probability conditioned on the value of being in a small neighborhood of . Set (a neighborhood of with radius ) to get

and using the fact that , we have

so

for almost all in the measure .\footnote{I do not know how to prove that without additional assumptions on , like continuous. [3] claims the limit a.e. “ can\textquotedblright be proved, though he does not proceed this way, and neglects to mention a.e. is in the measure .}

As another illustration and justification for understanding as the conditional probability of given , we now show what happens when the random variable is discrete. Suppose attains only values , , with . Then

Choose and as a neighborhood of with radius so small that does not contain any other , . Then for any ,

by the definition of , and from the definition of by Radon-Nykodym derivative,

This gives, for ,

by definition of conditional probability. The function is defined only on the set . Because that's where the variable is concentrated, this is a.s.

Expectation conditional on the value of a random variable[edit]

Suppose that and are random variables, integrable. Define again the measures on generated by the random variable ,

and a signed finite measure on ,

Here, is the indicator function of the event , so if and zero otherwise. Since

and , we have that , so is absolutely continuous with respect to . Consequently, there exists Radon-Nikodym derivative such that

The value is conditional expectation of given and denoted by . Then the result can be written as

for almost all in the measure generated by the random variable .

This definition is consistent with that of conditional probability: the conditional probability of given is the same as the conditional mean of the indicator function of given . The proof is also completely the same. Actually we did not have to do conditional probability at all and just call it a special case of conditional expectation.

Expectation conditional on a random variable and on a -algebra[edit]

Let be conditional expectation of the random variable given that random variable . Here is a fixed, deterministic value. Now take random, namely the value of the random variable , . The result is called the conditional expectation of given , which is the random variable

So now we have the conditional expectation given in terms of the sample space rather than in terms of , the range space of the random variable . It will turn out that after the change of the independent variable, the particular values attained by the random variable do not matter that much; rather, it is the granularity of that is important. The granularity of can be expressed in terms of the -algebra generated by the random variable , which is

By substitution, the conditional expectation satisfies

which, by writing

is seen to be the same as

It can be proved that for any -algebra , the random variable exists and is defined by this equation uniquely, up to equality a.e. in [5]. The random variable is called the conditional expectation of given the -algebra . It can be interpreted as a sort of averaging of the random variable to the granularity given by the -algebra [6].

The conditional probability of a an event (that is, a set) given the -algebra is obtained by substituting , which gives

An event is defined to be independent of a -algebra if and any are independent. It is easy to see that is independent of -algebra if and only if

that is, if and only if a.s. (which is a particularly obscure way to write independence given how complicated the definitions are).

Two random variables , are said to be independent if

which is now seen to be the same as

Properties of conditional expectation[edit]

To be done.

Conditional density and likelihood[edit]

Now that we have for an arbitrary event , we can define the conditional probability for a random variable and Borel set . Thus we can define the conditional density as the Radon-Nikodym derivative,

where is the Lebesgue measure. In the conditional density , and are random variables that identify the density function, and and are the arguments of the density function.

Note that in general is defined only for almost all (in Lebesgue measure) and almost all (in the measure generated by the random variable ).\textbf{ }Under reasonable additional conditions (for example, it is enough to assume that the joint density is continuous at , and ), the density of conditional on satisfies

Note that this density is a deterministic function.

Density of a random variable conditional on a random variable is

It is a function valued random variable obtained from the deterministic function by taking to be the value of the random variable .

A common shorthand for the conditional density is

This abuse of notation identifies a function from the symbols for its arguments, which is incorrect. Imagine that we wish to evaluate the value of the conditional density of at given ; then becomes , which is a nonsense.

When the value of is constant, the function is a probability density function of . When the value of is constant, the function is called the likelihood function.

References[edit]

  1. ^ William Feller. An introduction to probability theory and its applications. Vol. I. Third edition. John Wiley \& Sons Inc., New York, 1968.
  2. ^ Wikipedia. Conditional expectation. Version as of 18:29, 28 March 2007 (UTC), 2007.
  3. ^ a b Samuel S. Wilks. Mathematical statistics. A Wiley Publication in Mathematical Statistics. John Wiley \& Sons Inc., New York, 1962.
  4. ^ A. N. Kolmogorov. Foundations of the theory of probability. Chelsea Publishing Co., New York, 1956. Translation edited by Nathan Morrison, with an added bibliography by A. T. Bharucha-Reid.
  5. ^ Claude Dellacherie and Paul-Andr{\'e} Meyer. Probabilities and potential, volume 29 of North-Holland Mathematics Studies. North-Holland Publishing Co., Amsterdam, 1978.
  6. ^ S. R. S. Varadhan. Probability theory, volume 7 of Courant Lecture Notes in Mathematics. New York University Courant Institute of Mathematical Sciences, New York, 2001.