5 Joint distributions
5.1 Definitions
Let (\Omega, \mathcal H, \mathbb P) be a probability space.
5.1.1 Random variable
A random variable is a measurable function: (\Omega, \mathcal H) \to (E, \mathcal E). The distribution of a random variable is the pushforward measure \mathbb P \circ X^{-1}.
\mathbb P(X \in A) = \mathbb P(X^{-1} (A))
\mu is the distribution of X \iff \forall f \in \mathcal E_+, \mathbb E[f\circ X] = \mu[f]
5.1.2 Expectation
Lebesgue integral of a random variable wrt a probability measure. \mathbb E[X] = \mathbb P[X]
Some handy relations between expectation and probability: * \mathbb P(A) = \mathbb E [1_A] * For non-negative rv X \mathbb E[X] = \int \mathbb P(X > t) dt
Proof
\begin{align*} \int \mathbb P(X > t) dt = \int_{t = 0}^\infty \int_{x} \mathbb I(x > t) dP_X(x) dt \quad \text{ Fubini's}\\ = \int_x \int_{t = 0}^x dt dP_X(x) \\ = \int_x x dP_X(x) \end{align*}
5.1.3 Moment generating function
For some random variable X,
M_X(s) = \mathbb E[\exp sX]
For random vectors X \in \mathbb R^d, \forall t \in R^d:
M_X(t) = \mathbb E[\exp t^\top X]
Properties:
- For independent RVs X and Y, M_{X+Y}(s) = M_X(s) M_Y(s)
- If the MGF is finite in a neighbourhood of 0, then \mathbb E|X^n| < \infty for all n and M_X(s) = \sum_{n=0}^\infty \frac{s^n}{n!} \mathbb E [X^n].
- Thus, the r-th derivative at zero gives M_X^{(r)}(0) = \mathbb E[X^r].
Proof
\begin{align*} M_X(s) &= \mathbb E[\exp sX] = \mathbb E[ \sum_{n=0}^\infty \frac{(sX)^n}{n!} ]\\ &= \mathbb E[ \lim_{N \to \infty} \sum_{n=0}^N \frac{(sX)^n}{n!} ]\\ &= \lim_{N \to \infty} \mathbb E[ \sum_{n=0}^N \frac{(sX)^n}{n!} ]\\ &= \sum_{n=0}^\infty \frac{s^n}{n!} \mathbb E[X^n] \end{align*}
For the third step, we need to swap limits and integrals. We can use DCT with the dominating function \exp |sX|. Now this is integrable because by assumption for some s_0 > 0 and |s| < s_0, M_X(s) < \infty
\mathbb E[ \exp |sX|] < \mathbb E[ e^{sX} + e^{-sX}] = M_X(s) + M_X(-s)
5.1.4 Characteristic function
5.2 Product spaces
Let (E, \mathcal E) and (F, \mathcal F) be two measurable spaces.
5.2.1 Product sigma algebra
A measurable rectangle is a Cartesian product of a measurable sets. The sigma algebra generated by all possible measurable rectangles is called the product sigma algebra.
\mathcal E \otimes \mathcal F = \sigma \{ A \times B: A \in \mathcal E, B \in \mathcal F \}
This can be easily extended to a product of finitely many sigma algebras \mathcal E_1, \dots \mathcal E_n. Note that the set of all measurable rectangles is a p-system.
\mathcal E_1 \otimes \dots \otimes \mathcal E_n = \sigma \{ A_1 \times A_2 \dots \times A_n: A_1 \in \mathcal E_1, B \in \mathcal E_n \}
The concatenation of measurable functions f_i: \Omega \to E_i each measurable with respect to some sigma algebra \mathcal E_i is measurable with respect to the product sigma algebra.
5.2.2 Infinite product spaces
Let T be an arbitrary index set such that for all t \in T, (E_t, \mathcal E_t) is a measurable space. A measurable rectangle is defined as a Cartesian product of the form \times_{t \in T} A_t = \{ x \in \times_{t \in T} E_t : x_t \in A_t \forall t \in T \}
where only finitely many A_t \subsetneq E_t.
The sigma algebra generated by all measurable rectangles is the product sigma algebra.
Again, the arbitrary “concatenation” of measurable functions f_t: \Omega \to E_t each measurable with respect to some sigma algebra \mathcal E_t is measurable with respect to the product sigma algebra.
5.3 Sigma algebra generated by random variables
Let X be \mathcal H$E$ measurable, then: \sigma X = X^{-1} \mathcal E = \{X^{-1} A : A \in \mathcal E\}
\sigma X is the smallest sigma algebra that X is measurable with respect to.
For an arbitrary collection of random variables X_t: \Omega \to E_t, t \in T:
\sigma \{X_t: t \in T \} = \sigma \left( \cup_{t \in T} \sigma X_t \right)
5.3.1 Measurability with random variable sigma algebra
A function V is measurable with respect to \sigma X if and only if V = f \circ X for some f \in \mathcal E.
For a stochastic process, V is measurable with respect to \sigma \{X_t: t \in T \} if and only if there exists a sequence (t_n) \in T and a function f \in \otimes_n \mathcal E_{t_n} such that:
V = f(X_{t_1}, X_{t_2}, \dots)