Let $Pf = \mathbb{E}f$ denote the integration of random variable $f$ under probability measure $P$.

Empirical Distribution

Empirical measure (empirical probability density function, EPDF) is the discrete uniform measure on a random sample; that is, each observation of the sample is assigned the same finite probability. Empirical measure is a random measure because it is dependent on a random sample. Symbolically, the empirical measure $P_n$ of a measurable set $A$ in sample space $(Ω,Σ)$ is defined as $$P_n(A) = \frac{1}{n} \sum_{i=1}^n \mathbf{1}_A(X_i)$$

The empirical distribution function (ECDF) of a real-valued random variable is the empirical measure indexed by a class of one-sided intervals: $$P_n(x) = P_n (-\infty,x] = \frac{1}{n} \sum_{i=1}^n \mathbf{1}(X_i \leq x)$$

Empirical distribution functions have the following properties (point-wise in the parameter domain):

  1. Unbiased: $P P_n(x) = F(x)$
  2. Consistent: $P_n(x) \overset{p}{\to} F(x)$
  3. Asymptotically normal: $\sqrt{n} ( P_n(x) - F(x) ) \Rightarrow N(0, F(x)(1-F(x)) )$

Theorem: (Glivenko-Cantelli) Empirical distribution function converges uniformly to the population distribution function: $$\| P_n - F \|_{\infty} \overset{a.s.}{\to} 0$$

Method-of-Moments Estimators

For a parameter $\mathbf{h}(P \mathbf{g}(X))$ that is a function of expectations, its method-of-moments estimator (substitution estimator) [@Pearson1894] replaces the population measure with the empirical measure: $\mathbf{h}(P_n \mathbf{g}(X))$. If the parameter to be estimated is a function of moments, $h(PX, \cdots, PX^k)$, the method-of-moments estimator is $h( P_n X, \cdots, P_n X^k)$.

Empirical moments are consistent and asymptotically normal: using central limit theorem, we can show that $$\sqrt{n} [ (P_n X, \cdots, P_n X^k) - (P X, \cdots, P X^k) ] \Rightarrow N(0,Σ)$$, where $Σ_{ij} = P X^{i+j} - P X^i P X^j$. If $h(\cdot)$ is continuously differentiable at $(PX, \cdots, PX^k)$, by the Delta method, we have $$\sqrt{n} [ h( P_n X, \cdots, P_n X^k) - h(PX, \cdots, PX^k) ] \Rightarrow N(0, (∇h)'Σ(∇h))$$, where $∇h$ is evaluated at $(PX, \cdots, PX^k)$.

For Gamma distribution $Γ(α, β)$:

  • The parameters $(α, β) = g(PX, PX^2)$, where $g(x,y) = ( x^2/(y-x^2), (y-x^2)/x )$.
  • The $q$-th order moment is $PX^q = α (α+1) \cdots (α+q-1) β^q$.
  • The asymptotic variance of empirical 1st and 2nd moments is $Σ = \begin{pmatrix} α β^2 & 2 α(α+1)β^2 \\ 2 α(α+1)β^2 & 2 α(α+1)(2α+3)β^4 \end{pmatrix}$
  • Gradient of $g(x,y)$ at $(PX, PX^2)$ is $∇ g(PX, PX^2) = \begin{pmatrix} \frac{2(α+1)}{β} & -\frac{α+1}{α} \\ -\frac{1}{β^2} & \frac{1}{αβ} \end{pmatrix}$
  • Then we have the asymptotic variance for method of moments estimator $g(P_n X, P_n X^2)$: $Σ' = \begin{pmatrix} 2α(α+1) & -2(α+1)β \\ -2(α+1)β & \frac{(2α+3)β^2}{α} \end{pmatrix}$

For uniform distribution $U(0, θ)$, because $θ = 2 PX$, its method of moments estimator is $2 P_n X$. But this estimator is not very efficient.


🏷 Category=Statistics