Normal distribution - Advanced Theoretical Tools and Extensions
Understand the characteristic and moment‑generating functions (and cumulants) of the normal distribution, the Central Limit Theorem, and the multivariate normal extension.
Summary
Read Summary
Flashcards
Save Flashcards
Quiz
Take Quiz
Quick Practice
What is the characteristic function $\varphiX(t)$ of a normal distribution $X \sim N(\mu, \sigma^2)$?
1 of 10
Summary
Fourier Transform, Generating Functions, and the Normal Distribution
Introduction
When working with probability distributions, we often need tools that can capture all the information about a distribution in a compact form. Generating functions—especially the characteristic function and moment-generating function—are powerful mathematical tools that allow us to:
Extract moments (mean, variance, skewness, etc.) easily
Understand how distributions combine under addition
Prove fundamental theorems about convergence of distributions
This section introduces these tools and explores how they apply to the normal distribution and beyond.
Characteristic Function
The characteristic function of a random variable $X$ is defined as:
$$\varphiX(t) = E[e^{itX}]$$
where $i = \sqrt{-1}$ is the imaginary unit and $t$ is a real parameter.
Why Use the Characteristic Function?
The characteristic function is a fundamental tool because:
It always exists for any distribution (unlike the moment-generating function)
It uniquely determines the distribution — if two random variables have the same characteristic function, they have the same distribution
It makes theoretical work easier, especially for proving limit theorems
The Normal Case
For a normal random variable $X \sim N(\mu, \sigma^2)$, the characteristic function has a particularly elegant form:
$$\varphiX(t) = \exp\left(i\mu t - \frac{1}{2}\sigma^2 t^2\right)$$
This formula tells us that knowing just the mean $\mu$ and variance $\sigma^2$ completely determines the distribution through its characteristic function.
Moment-Generating Function (MGF)
The moment-generating function is closely related to the characteristic function:
$$MX(t) = E[e^{tX}]$$
where $t$ is a real parameter (not imaginary).
Why "Moment-Generating"?
The name comes from a key property: if you expand $MX(t)$ as a power series, the coefficients are the moments of $X$:
$$MX(t) = E[e^{tX}] = E\left[\sum{k=0}^{\infty} \frac{(tX)^k}{k!}\right] = \sum{k=0}^{\infty} \frac{E[X^k]}{k!} t^k$$
This means that $E[X^k] = \frac{d^k MX}{dt^k}\bigg|{t=0}$ — you can find any moment by taking derivatives!
The Normal Case
For $X \sim N(\mu, \sigma^2)$, the MGF is:
$$MX(t) = \exp\left(\mu t + \frac{1}{2}\sigma^2 t^2\right)$$
Important: The MGF doesn't exist for all distributions (e.g., Cauchy distribution), which is why the characteristic function is more general.
Cumulant-Generating Function (CGF)
The cumulant-generating function is simply the logarithm of the moment-generating function:
$$KX(t) = \log MX(t)$$
For a normal distribution:
$$KX(t) = \mu t + \frac{1}{2}\sigma^2 t^2$$
Understanding Cumulants
Cumulants are quantities extracted from the CGF, defined as:
$$\kappaj = \frac{d^j KX}{dt^j}\bigg|{t=0}$$
For the normal distribution, something remarkable happens: only the first two cumulants are non-zero:
First cumulant: $\kappa1 = \mu$ (the mean)
Second cumulant: $\kappa2 = \sigma^2$ (the variance)
All higher cumulants: $\kappa3 = \kappa4 = \cdots = 0$
Why This Matters
This property is special to the normal distribution. Cumulants measure higher-order properties like skewness and kurtosis, which are zero for normal distributions. This reflects the fact that the normal distribution is completely determined by just its mean and variance—nothing else matters.
The Central Limit Theorem (CLT)
The Central Limit Theorem is one of the most important results in statistics. It explains why the normal distribution appears so frequently in practice.
The Statement
Suppose $X1, X2, \ldots, Xn$ are independent, identically distributed (i.i.d.) random variables with:
Mean: $E[Xi] = 0$
Variance: $\text{Var}(Xi) = \sigma^2 < \infty$
Then as $n \to \infty$, the sum properly scaled converges in distribution to a normal distribution:
$$\frac{1}{\sqrt{n}}\sum{i=1}^{n}Xi \xrightarrow{d} N(0, \sigma^2)$$
The notation $\xrightarrow{d}$ means "converges in distribution to."
The General Version
If $Xi$ have non-zero mean $\mu$, then:
$$\frac{1}{\sqrt{n}}\sum{i=1}^{n}(Xi - \mu) \xrightarrow{d} N(0, \sigma^2)$$
Or equivalently, for the sample mean $\bar{X}n = \frac{1}{n}\sum{i=1}^{n}Xi$:
$$\sqrt{n}(\bar{X}n - \mu) \xrightarrow{d} N(0, \sigma^2)$$
Why This Matters
The CLT is remarkable because:
It applies to almost any distribution as long as the variables are i.i.d. with finite variance — the original distribution doesn't need to be normal
It justifies normal approximations in practice: sample means, test statistics, and many other quantities approximately follow normal distributions for large enough sample sizes
It explains why normal distributions are ubiquitous in nature and statistics
<extrainfo>
Convergence Rate: The CLT tells us the distribution converges to normal, but not how fast. The Berry-Esseen theorem provides bounds on the convergence rate, showing that convergence is generally faster for more symmetric distributions.
</extrainfo>
Extensions to Multivariate Settings
Multivariate Normal Distribution
So far we've discussed univariate (single-variable) normal distributions. The theory extends naturally to random vectors.
A random vector $\mathbf{X} = (X1, X2, \ldots, Xk)^\top \in \mathbb{R}^k$ follows a multivariate normal distribution if:
Every linear combination of its components is univariate normal.
Mathematically, for any vector $\mathbf{a} \in \mathbb{R}^k$, the quantity $\mathbf{a}^\top \mathbf{X} = a1 X1 + a2 X2 + \cdots + ak Xk$ must be normally distributed.
Parameters of the Multivariate Normal
A multivariate normal distribution is completely specified by two parameters:
Mean vector $\boldsymbol{\mu} \in \mathbb{R}^k$: This is simply the vector of means of each component
Covariance matrix $\boldsymbol{\Sigma} \in \mathbb{R}^{k \times k}$: This is a square matrix where entry $(i,j)$ is $\text{Cov}(Xi, Xj)$
We write: $\mathbf{X} \sim N(\boldsymbol{\mu}, \boldsymbol{\Sigma})$
Requirements on the Covariance Matrix
For a valid multivariate normal distribution, the covariance matrix must be:
Symmetric: $\boldsymbol{\Sigma} = \boldsymbol{\Sigma}^\top$
Positive-definite: All eigenvalues are strictly positive, and for any non-zero vector $\mathbf{v}$, we have $\mathbf{v}^\top \boldsymbol{\Sigma} \mathbf{v} > 0$
These conditions ensure the distribution is well-defined and non-degenerate.
Why This Generalization Is Important
Multivariate normals are crucial for:
Understanding joint distributions of related variables
Regression and linear models
Multivariate hypothesis testing
Principal component analysis and other dimension reduction techniques
Summary
The normal distribution's mathematical elegance emerges through generating functions:
The characteristic function $\varphiX(t) = \exp(i\mu t - \frac{1}{2}\sigma^2 t^2)$ captures all distribution information
The moment-generating function provides a tool for extracting moments
The cumulant-generating function reveals that normal distributions are uniquely determined by just two cumulants (mean and variance)
The Central Limit Theorem explains why normal distributions are fundamental: sums of i.i.d. random variables converge to normal distributions
The multivariate normal extends these concepts to multiple dimensions, governed by a mean vector and covariance matrix
Understanding these tools prepares you to work confidently with normal distributions in theoretical and applied contexts.
Flashcards
What is the characteristic function $\varphiX(t)$ of a normal distribution $X \sim N(\mu, \sigma^2)$?
$\exp(i\mu t - \frac{1}{2}\sigma^2 t^2)$
What is the formula for the moment-generating function $MX(t)$ of a normal distribution $X \sim N(\mu, \sigma^2)$?
$\exp(\mu t + \frac{1}{2}\sigma^2 t^2)$
What is the cumulant-generating function $KX(t)$ for a normal distribution $X \sim N(\mu, \sigma^2)$?
$\mu t + \frac{1}{2}\sigma^2 t^2$
Which cumulants of a normal distribution $N(\mu, \sigma^2)$ are non-zero?
The first cumulant (equals the mean $\mu$)
The second cumulant (equals the variance $\sigma^2$)
What does the Central Limit Theorem (CLT) state regarding the sum of many independent, identically distributed random variables?
They converge in distribution to a normal distribution.
What is the formal limit of $\frac{1}{\sqrt{n}}\sum{i=1}^{n}Xi$ as $n \to \infty$ if $Xi$ are i.i.d. with mean $0$ and variance $\sigma^2$?
$N(0, \sigma^2)$
What are two common statistics for which the Central Limit Theorem justifies normal approximations?
Sample means and test statistics.
How is a random vector $\mathbf{X} \in \mathbb{R}^k$ defined as multivariate normal in terms of its linear combinations?
Every linear combination $\mathbf{a}^\top\mathbf{X}$ is univariate normal.
What are the two primary parameters involved in the density of a multivariate normal distribution?
Mean vector $\boldsymbol{\mu}$
Covariance matrix $\mathbf{\Sigma}$
What property must the covariance matrix $\mathbf{\Sigma}$ satisfy in a multivariate normal distribution?
Symmetric positive-definite.
Quiz
Normal distribution - Advanced Theoretical Tools and Extensions Quiz Question 1: What is the characteristic function $\varphi_X(t)$ of a normal random variable $X\sim N(\mu,\sigma^{2})$?
- $\exp\!\bigl(i\mu t -\tfrac{1}{2}\sigma^{2}t^{2}\bigr)$ (correct)
- $\exp\!\bigl(\mu t +\tfrac{1}{2}\sigma^{2}t^{2}\bigr)$
- $\exp\!\bigl(i\mu t +\tfrac{1}{2}\sigma^{2}t^{2}\bigr)$
- $\exp\!\bigl(-i\mu t -\tfrac{1}{2}\sigma^{2}t^{2}\bigr)$
Normal distribution - Advanced Theoretical Tools and Extensions Quiz Question 2: What is the moment‑generating function $M_X(t)$ of $X\sim N(\mu,\sigma^{2})$?
- $\exp\!\bigl(\mu t +\tfrac{1}{2}\sigma^{2}t^{2}\bigr)$ (correct)
- $\exp\!\bigl(i\mu t -\tfrac{1}{2}\sigma^{2}t^{2}\bigr)$
- $\exp\!\bigl(i\mu t +\tfrac{1}{2}\sigma^{2}t^{2}\bigr)$
- $\exp\!\bigl(\mu t -\tfrac{1}{2}\sigma^{2}t^{2}\bigr)$
Normal distribution - Advanced Theoretical Tools and Extensions Quiz Question 3: When is a random vector $\mathbf{X}\in\mathbb{R}^{k}$ called multivariate normal?
- When every linear combination $\mathbf{a}^{\top}\mathbf{X}$ is univariate normal. (correct)
- When each component of $\mathbf{X}$ is independent and identically normal.
- When $\mathbf{X}$ has a uniform distribution over a sphere.
- When the covariance matrix of $\mathbf{X}$ is diagonal.
Normal distribution - Advanced Theoretical Tools and Extensions Quiz Question 4: What is the value of the third cumulant of a normal distribution?
- 0 (correct)
- $\mu$ (the mean)
- $\sigma^{2}$ (the variance)
- It equals the skewness, which is non‑zero
Normal distribution - Advanced Theoretical Tools and Extensions Quiz Question 5: Which condition is required for the Central Limit Theorem to hold for a sequence of i.i.d. random variables?
- Both the mean and variance must be finite (correct)
- Only the variables must be identically distributed
- Only independence is required
- The variables must have infinite variance
Normal distribution - Advanced Theoretical Tools and Extensions Quiz Question 6: In the standard form of the CLT, how is the sum $\sum_{i=1}^{n}X_{i}$ of i.i.d. zero‑mean variables with variance $\sigma^{2}$ scaled to converge to $N(0,\sigma^{2})$?
- Divide by $\sqrt{n}$ (correct)
- Divide by $n$
- Multiply by $\sqrt{n}$
- Multiply by $n$
Normal distribution - Advanced Theoretical Tools and Extensions Quiz Question 7: Which statistical quantity can be approximated by a normal distribution for large samples due to the CLT?
- The sampling distribution of the sample mean (correct)
- The exact distribution of the sample variance
- The distribution of the sample maximum
- The distribution of a single observation
Normal distribution - Advanced Theoretical Tools and Extensions Quiz Question 8: Given the cumulant‑generating function of a normal random variable $X\sim N(\mu,\sigma^{2})$ is $K_X(t)=\mu t+\tfrac{1}{2}\sigma^{2}t^{2}$, what is the value of the second derivative $K_X''(0)$?
- \sigma^{2} (correct)
- \mu
- 2\sigma^{2}
- 0
What is the characteristic function $\varphi_X(t)$ of a normal random variable $X\sim N(\mu,\sigma^{2})$?
1 of 8
Key Concepts
Functions of Random Variables
Characteristic function
Moment‑generating function (MGF)
Cumulant‑generating function (CGF)
Statistical Properties
Cumulants
Central Limit Theorem (CLT)
Multivariate normal distribution
Mathematical Foundations
Fourier transform
Definitions
Characteristic function
The Fourier transform of a probability distribution, given by ϕ_X(t)=E[e^{itX}], uniquely characterizing the distribution.
Moment‑generating function (MGF)
A function M_X(t)=E[e^{tX}] that encodes all moments of a random variable via its derivatives at t=0.
Cumulant‑generating function (CGF)
The logarithm of the MGF, K_X(t)=log M_X(t), whose derivatives at t=0 yield the cumulants of the distribution.
Cumulants
Quantities derived from the CGF; the first cumulant is the mean, the second is the variance, and higher‑order cumulants describe shape characteristics.
Central Limit Theorem (CLT)
A fundamental result stating that the normalized sum of many i.i.d. random variables with finite mean and variance converges in distribution to a normal law.
Multivariate normal distribution
An extension of the normal distribution to vectors, defined by a mean vector and a positive‑definite covariance matrix, with all linear combinations being univariate normal.
Fourier transform
A mathematical operation converting a function into its frequency components; in probability theory it underlies the definition of characteristic functions.