Conditional Expectation

(Most of these equalities are related to ANOVA/regression.)

Basic properties:

  • Linearity: $\mathbb{E}[aX+bY|Z=z] = a \mathbb{E}[X|Z=z] + b \mathbb{E}[Y|Z=z]$
  • $\mathbb{E}[Y \phi(X)|X=x] = \phi(x) \mathbb{E}[Y|X=x]$

Tower property (Law of Iterated Expectations): $\mathbb{E} [ \mathbb{E}(Y|X) ] = \mathbb{E}Y$

Conditional variance identity: $\mathrm{Var}Y = \mathbb{E}[\mathrm{Var}(Y|X)] + \mathrm{Var}[ \mathbb{E}(Y|X) ]$

Generalization of conditional variance identity: $\mathbb{E} [Y-\eta(X)]^2 = d_2^2(\mathbb{E}(Y|X),\eta(X)) + \mathbb{E}[\mathrm{Var}(Y|X)]$

Hierarchical models

The advantage of hierarchical models is that complicated processes may be modeled by a sequence of relatively simple models placed in a hierarchy, and hierarchical models often make calculations easier.

Theoretically, any more complicated hierarchy can be treated as a two-stage hierarchy.

Hierarchical models lead to mixture distribution.

See Structured Probabilistic Model.

Independence as an assumption and simplification

Independence originates from the idea that for the two events considered, the state of either event doesn't affect the other. In the perspective of probability theory, independence means additional knowledge of either event does not affect your knowledge of the other. Symbolically, $P(A|B) = P(A)$ and $P(B|A) = P(B)$, which is equivalent to $P(A \cap B) = P(A) \cdot P(B)$. The latter equation is then generalized to the standard definition of statistical independence (which is weaker than the original idea): $f(x,y) = f(x) \cdot f(y)$.

Formally, dependence refers to any situation in which random variables do not satisfy the condition of statistical independence.

Mutual Independence

Random vectors $\mathbf{X}_1, \cdots, \mathbf{X}_n$ have joint PDF (or PMF) $f(\mathbf{x}_1, \cdots, \mathbf{x}_n)$ and marginal PDFs (or PMFs) $f_{\mathbf{X}_i} (\mathbf{x}_i)$. They are called mutually independent random vectors if $f(\mathbf{x}_1, \cdots, \mathbf{x}_n) = \prod_{i=1}^n f_{\mathbf{X}_i}(\mathbf{x}_i)$.

Random vectors $\mathbf{X}_1, \cdots, \mathbf{X}_n$ are mutually independent, iff there exist functions $g_1, \cdots, g_n$, s.t. $f(\mathbf{x}_1, \cdots, \mathbf{x}_n) = \prod_{i=1}^n g_i(\mathbf{x}_i)$.

For mutually independent random vectors $\mathbf{X}_1, \cdots, \mathbf{X}_n$ and any functions $g_1, \cdots, g_n$, random variables (or vectors) $g_1(\mathbf{X}_1), \cdots, g_n(\mathbf{X}_n)$ are also mutually independent.

For mutually independent random vectors $\mathbf{X}_1, \cdots, \mathbf{X}_n$ and any functions $g_1, \cdots, g_n$, $\mathbb{E}( \prod_{i=1}^n g_i(X_i) ) = \prod_{i=1}^n (\mathbb{E} g_i(X_i))$.

Covariance and correlation

Covariance of multiple random variables is the generalization of variance of a real random variable: $\mathrm{cov}(X,Y) = \mathbb{E}[(X-\mu_X)(Y-\mu_Y)]$.

Correlation of two random variables indicates the average tendency for the conditioned mean of one random variable change in accord to the other: $\mathrm{cor}(X,Y) = \frac{\mathrm{cov}(X,Y)}{\sigma_X \sigma_Y}$.

The information given by correlation is not enough to define the dependence structure between random variables.


🏷 Category=Probability