Second post in the series.
Conditional expectation
Given a random variable X over probability space of (Ω,F,P) with E[X]<∞ we define conditional expectation Y=E[X|F1] where F1⊂F as follows:
- It is random variable measurable on F1
- The following holds for any A∈F1,E[Y1A]=E[X1A]
What does those conditions imply?
Lets begin with the first condition. The measurability on F1 is the requirement that the level sets, {ω∣E[X|F1](ω)≤α;∀α∈R}, of the random variable in a question, are in F1.
To get a feeling of what’s going on, I think of E[X|F1] as a steps function over sets in F1. The justification comes from the fact that any measurable function can be approximated as the limit of a sequence of steps functions. When I say “steps function” I refer to a function defined as ∑A∈FαA1A usually called “simple function” (that notion should be familiar from Lebesgue integration theory).
To build the intuition for conditional expectation I will define the notion of minimal sets in F1 . Minimal set is one that can’t be further divided into smaller sets inside F1. look at the very simple case below, where F1 consists only of A and its completion AC. Since A and AC are the only sets in F1 (except the whole Ω), they are obviously the minimal sets there. The random variable Y on F1 can change only on the boundaries of the minimal sets. It is easy to see why, suppose Y did change inside, obviously, that would create level sets which are not in F1 in contradiction to Y being measurable on F1. So minimal sets dictate the best resolution of the random variables that live on F1.
Now I will turn my attention to the second condition. It requires that the expectation of both the X and the Y, agree when calculated over sets in F1.
This means that conditional expectation Y is a coarse estimation of the original random variable X. Fires note that the minimal sets in F1 are not in general minimal in F. It is possible to further divided them to get still smaller sets in F. Consequently the resolution of random variable on F is greater then on F1. Thus the variable in F may appear smooth while variable in F1 may appear coarse in comparison.
Now think of X as step function that is defined on so fine structure of F that it just appears as smooth function over Ω, while Y is confined to the very course structure of F1⊂F:
In order to satisfy condition 2 the average of X and Y over sets in F1 should agree. Y is constant on the minimal sets of F1 so it has to be equal to the average of X over minimal set of F1. If you still follow me, you should see by now why Y is a sort of course version of X by courser resolution of F1.
What is the relation to the “undergraduate” notion of conditional expectation?
In introduction to the probability theory course we defined conditional expectation differently. Given random variable X and event A conditional expectation , E[X|A], is ∫AXdω , so how this definition is related to the “graduate” conditional expectation?
First of all while E[X|A] is real number the “graduate” definition talks about a random variable.
All the rest is pretty straight forward, lets define F1 to be (A,AC,∅,Ω) then Y=E[X|F1] equals to:
E[X|F1]={E[X|A],∀ω∈AE[X|AC],∀ω∈AC
That’s it!
Technorati Tags: Intuition Probability,Probability,Conditional expectation,expectationWindows Live Tags: Intuition Probability,Probability,Conditional expectation,expectation
WordPress Tags: Intuition Probability,Probability,Conditional expectation,expectation
Blogger Labels: Intuition Probability,Probability,Conditional expectation,expectation
No comments:
Post a Comment