2.6 Conditioning (1)
Conditional PMFs, given the value of another random variable
Properties of conditional PMFs.
Contents
Conditioning (1)
- conditioning a random variable on an event
- conditioning one random variable on another
- summary of facts about conditional PMFs
Conditioning (2)
- summary of facts about conditional expectation
- total expectation theorem
+ mean and variance of the geometric (example 2.17.)
+ The Two-Envelopes Paradox (example 2.18. - 이건 따로 다룰 예정)
Conditioning a Random Variable on an Event
conditional PMF
The conditional PMF of a random variable \(X\), conditioned on a particular event \(A\) with \(P(A) > 0\), is defined by
$$p_{X|A} (x) = P(X = x|A) = \frac{P({X=x} \cap A)}{P(A)}.$$
Note that the events \({X=x} \cap A\) are disjoint for different values of x, their union is A, therefore,
$$P(A) = \sum_x P({X=x} \cap A).$$
Combining the above two formulas, we see that $$\sum_x p_{X|A} (x) = 1$$, so \(p_{X|A}\) is a legitimate PMF. (정당한 PMF다~)
Conditioning one Random Variable on Another Random Variable
Let X and Y be two random variables associated with the same experiment.
If we know that the value of Y is some particular y [with \(p_Y (y) >0\), this provides partial knowledge about the value of X.
-> conditional PMF \(p_{X|Y}\) of X given Y, which is defined by specializing the definition of \(p_{X|A}\) to events A of the form \({Y=y}\):
$$p_{X|Y} (x|y) = P(X=x|Y=y).$$
Using the definition of conditional probabilities, we have
$$p_{X|Y} (x|y) = \frac{P(X=x|Y=y)}{P(Y=y)} = \frac{p_{X,Y} (x,y) }{p_{Y}(y)}.$$
Let us fix some y with \(p_{Y}(y) > 0\) and consider \(p_{X|Y} (x|y)\) as a function of x.
-> a valid PMF for X: it assigns nonnegative values to each possible x, and these values add to 1.
Furthermore, this funciton of x has the same shape as \(p_{X,Y} (x,y)\) except that it is divided by \(p_{Y}(y)\) which enforces the normalization property
$$\sum_x p_{X|Y} (x|y) =1.$$
Counterpart
To calculate the marginal PMF
we have by using the definitions,
This formula provides a divide-and-conquer method for calculating marginal PMFs.
It is in essence identical to the total probability theorem given in Chapter 1, but cast in different notation.
Summary of Facts About Conditional PMFs
Let X and Y be two random variables associated with the same experiment.
- Conditional PMFs are similar to ordinary PMFs, but pertain to a universe where the conditioning event is known to have occurred.
- The conditional PMF of X given an event A with P(A) > 0, is defined by $$p_{X|A} (x) = P(X=x|A)$$ and satisfies $$\sum_x p_{X|A}(x) = 1.$$
- If \(A_1 , . . . ,A_n\) are disjoint events that form a partition of the sample space, with \(P(A_i) > 0 \) for all \(i\), then $$p_{X} = \sum_{i=1}^{n} P(A_i)p_{X|A_i}(x).$$
(This is a special case of the total probability theorem.) - Furthermore, for any event B, with \(P(A_i \cap B) > 0\) for all \(i\), we have $$p_{X|B}(x) = \sum_{i=1}^{n} P(A_i|B)p_{X|A_i \cap B}(x).$$
- The conditional PMF of \(X\) given \(Y = y\) is related to the joint PMF by $$p_{X, Y} (x, y) = p_Y(y) p_{X|Y}(x|y).$$
- The conditional PMF of \(X\) given \(Y\) can be used to calculate the marginal PMF of \(X\) through the formula $$p_X(x) = \sum_y p_Y(y) p_{X|Y}(x|y).$$
- There are natural extensions of the above involving more than two random variables.
References
Bertsekas, D. P., Tsitsiklis, J. N. (2008). Introduction to Probability Second Edition. Athena Scientific.