Chapter 0Review
0.1 Definitions
Probability Space and Function. We consider a finite set \(\Omega=\{\omega_i\}_{i=1}^N\) as our space of all plausible events, and an additive probability function \(P:2^\Omega\to[0,1]\) that assigns likelihood to any subset of \(\Omega\), which we denote by \(2^\Omega\). Namely, we require that for any \(A\subset\Omega\),
so in particular, we have that if \(A\cap B=\emptyset\), then
Moreover, we need that the probability of any event occurring is exactly 1
Example. Rolling two independent dice, we have the plausible pairs
and the natural probability distribution
Random variable. A random variable (or observable) is a function
Example. For two dice, the typical example is the sum of the two dice:
which has the following probability distribution:
| x | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 | 11 | 12 |
|---|---|---|---|---|---|---|---|---|---|---|---|
| P(X=x) | 1/36 | 2/36 | 3/36 | 4/36 | 5/36 | 6/36 | 5/36 | 4/36 | 3/36 | 2/36 | 1/36 |
Simplest case,
Next, we consider the non-linear random variables. Now we consider \(Y:\Omega\to \mathbb{R}\) and \(Z:\Omega\to \mathbb{R}\) given by
which have probability distributions given by:
| y | 1 | 2 | 3 | 4 | 5 | 6 |
|---|---|---|---|---|---|---|
| P(Y=y) | 11/36 | 9/36 | 7/36 | 5/36 | 3/36 | 1/36 |
| z | 1 | 2 | 3 | 4 | 5 | 6 |
|---|---|---|---|---|---|---|
| P(Z=z) | 1/36 | 3/36 | 5/36 | 7/36 | 9/36 | 11/36 |
Knowing the probability distributions of \(Y\) and \(Z\), we cannot actually figure out the joint distribution, which is given by:
| y \ z | 1 | 2 | 3 | 4 | 5 | 6 |
|---|---|---|---|---|---|---|
| 1 | 1/36 | 2/36 | 2/36 | 2/36 | 2/36 | 2/36 |
| 2 | 0 | 1/36 | 2/36 | 2/36 | 2/36 | 2/36 |
| 3 | 0 | 0 | 1/36 | 2/36 | 2/36 | 2/36 |
| 4 | 0 | 0 | 0 | 1/36 | 2/36 | 2/36 |
| 5 | 0 | 0 | 0 | 0 | 1/36 | 2/36 |
| 6 | 0 | 0 | 0 | 0 | 0 | 1/36 |
Note: Knowing the individual distributions, we do not know the joint distributions.
Conditional Probability. How does the probability of an event \(B\subset \Omega\) get affected by the knowledge that we know, or suppose, that another event \(A\subset \Omega\) happened?
or equivalently the multiplication rule
We can use this with the Law of total probability: Assume \(A_1,\dots, A_N\subset\Omega\) satisfy \(\bigcup_{i=1}^N A_i=\Omega\) and \(A_i\cap A_j=\emptyset\) for \(i\ne j\), then for any \(B\subset \Omega\),
Example. Assume we have two urns with balls inside them. Urn 1 has 10 Blue balls and 30 Red Balls, and Urn 2 has 40 Blue balls and 20 Red balls. Assume a ball is picked with the following algorithm:
- Pick an Urn at random.
- Pick a ball from that Urn at random.
Assume that you win if the ball picked is Blue. What is the probability that the ball is Blue?
In the background, we have a Probability space
and probability function
Question: Assume that you can arrange the balls yourself, how do you maximize your chances of winning?
Conditioning random variables
How does knowing an outcome affect the distribution of the other random variables. For instance, let's look back at the example of \( Y\,\, Z\) the min and the max of two independent dice. If
| z | 1 | 2 | 3 | 4 | 5 | 6 |
|---|---|---|---|---|---|---|
| P(Z=z|Y=2) | 0 | 1/9 | 2/9 | 2/9 | 2/9 | 2/9 |
| z | 1 | 2 | 3 | 4 | 5 | 6 |
|---|---|---|---|---|---|---|
| P(Z=z|Y=5) | 0 | 0 | 0 | 0 | 1/3 | 2/3 |
We say that two variables are independent if the information of one random variables gives no useful information about the other. For dependence, think of the temperature value today to tomorrow, given the temperature today, you can give an informed guess of the next couple of days. Total independence of random variables means the output of one does not affect the other. There are several ways to define This
Note: Overusing the independence assumption is blamed for the collapse of the mortgage bubble in 2008.
Expectation: