Processing math: 100%

MathJax

Sunday, September 2, 2012

Probability 2 - Conditional expectation

Second post in the series.

Conditional expectation

Given a random variable X over probability space of (Ω,F,P) with E[X]< we  define conditional expectation Y=E[X|F1] where F1F as follows:

  • It is random variable measurable on F1
  • The following holds for any AF1,E[Y1A]=E[X1A]

What does those conditions imply?

Lets begin with the first condition. The measurability on F1 is the requirement that the level sets, {ωE[X|F1](ω)α;αR}, of the random variable in a question, are in F1.

To get a feeling of what’s going on, I think of E[X|F1] as a steps function over sets in F1. The justification comes from the fact that any measurable function can be approximated as the limit of a sequence of steps functions. When I say “steps function” I refer to a function defined as AFαA1A usually called “simple function” (that notion should be familiar from Lebesgue integration theory). 

To build the intuition for conditional expectation I will define the notion of minimal sets in F1 . Minimal set is one that can’t be further divided into smaller sets inside F1. look at the very simple case below, where F1 consists only of A and its completion AC. Since A and AC are the only sets in F1 (except the whole Ω), they are obviously the minimal sets there. The random variable Y on F1 can change only on the boundaries of the minimal sets. It is easy to see why, suppose Y did change inside, obviously, that would create level sets which are not in F1 in contradiction to Y being measurable on F1. So minimal sets dictate the best resolution of the random variables that live on F1.

image

Now I will turn my attention to the second condition. It requires that the expectation of both the X and the Y, agree when calculated over sets in F1.

This means that conditional expectation Y is a coarse estimation of the original random variable X. Fires note that the minimal sets in F1 are not in general minimal in F. It is possible to further divided them to get still smaller sets in F. Consequently the resolution of random variable on F is greater then on F1. Thus the variable in F may appear smooth while variable in F1 may appear coarse in comparison. 

Now think of X as step function that is defined on so fine structure of F that it just appears as smooth function over Ω, while Y is confined to the very course structure of F1F:

image

 

In order to satisfy condition 2 the average of X and Y over sets in F1 should agree. Y is constant on the minimal sets of F1 so it has to be equal to the average of X over minimal set of F1. If you still follow me, you should see by now why Y is a sort of course version of X by courser resolution of F1.

What is the relation to the “undergraduate” notion of conditional expectation?

In introduction to the probability theory course we defined conditional expectation differently. Given random variable X and event A conditional expectation , E[X|A],  is AXdω , so how this definition is related to the “graduate” conditional expectation?

First of all while E[X|A] is real number the “graduate” definition talks about a random variable.

All the rest is pretty straight forward, lets define F1 to be (A,AC,,Ω) then Y=E[X|F1] equals to:

E[X|F1]={E[X|A],ωAE[X|AC],ωAC

That’s it!

Technorati Tags: ,,,

Windows Live Tags: Intuition Probability,Probability,Conditional expectation,expectation

WordPress Tags: Intuition Probability,Probability,Conditional expectation,expectation

Blogger Labels: Intuition Probability,Probability,Conditional expectation,expectation

No comments: