Random variable Study Guide
Study Guide
📖 Core Concepts
Random Variable (RV) – a measurable function \(X:\Omega\to\mathbb{R}\) that assigns a real number to each outcome of a random experiment.
Distribution (Law) – the push‑forward measure \(PX(B)=P\bigl(X\in B\bigr)\) on the real line; it tells the probabilities of all possible values of \(X\).
Cumulative Distribution Function (CDF) – \(FX(x)=P(X\le x)\); uniquely determines the distribution.
Probability Mass Function (PMF) – for discrete RVs, \(pX(x)=P(X=x)\) with \(\sumx pX(x)=1\).
Probability Density Function (PDF) – for absolutely continuous RVs, \(fX(x)\) with \(\displaystyle\int{-\infty}^{\infty} fX(x)\,dx=1\) and \(P(a\le X\le b)=\inta^b fX(x)\,dx\).
Measurability – \(X\) is measurable if \(X^{-1}(B)\in\mathcal{F}\) for every Borel set \(B\subset\mathbb{R}\); equivalently, \(\{\omega:X(\omega)\le x\}\in\mathcal{F}\) for all \(x\).
Moments – expectations of powers: \(\mathbb{E}[X]\) (mean), \(\operatorname{Var}(X)=\mathbb{E}[(X-\mathbb{E}[X])^2]\), higher moments \(\mathbb{E}[X^k]\).
Moment‑Generating Function (MGF) – \(MX(t)=\mathbb{E}[e^{tX}]\); if it exists, it uniquely characterizes the distribution.
---
📌 Must Remember
Discrete vs Continuous:
Discrete → countable support, described by PMF.
Continuous → uncountable support, described by PDF; \(P(X=x)=0\) for any single point.
CDF properties: non‑decreasing, right‑continuous, \(\displaystyle\lim{x\to-\infty}FX(x)=0,\;\lim{x\to\infty}FX(x)=1\).
Expectation formulas
Discrete: \(\displaystyle\mathbb{E}[X]=\sumx x\,pX(x)\).
Continuous: \(\displaystyle\mathbb{E}[X]=\int{-\infty}^{\infty} x\,fX(x)\,dx\).
Variance identity: \(\operatorname{Var}(X)=\mathbb{E}[X^2]-(\mathbb{E}[X])^2\).
Convolution for sums (independent RVs): \(\displaystyle f{X+Y}(z)=\int{-\infty}^{\infty} fX(x)fY(z-x)\,dx\).
Equality in distribution: \(X\stackrel{d}{=}Y\iff FX=FY\).
Almost‑sure equality: \(X=Y\) a.s. \(\iff P(X\neq Y)=0\).
Convergence hierarchy: a.s. \(\Rightarrow\) in probability \(\Rightarrow\) in distribution (but not vice‑versa).
LLN: Sample mean \(\bar Xn\) of i.i.d. variables converges in probability to \(\mathbb{E}[X]\).
CLT: \(\displaystyle\frac{\sum{i=1}^n Xi - n\mu}{\sqrt{n}\sigma}\xrightarrow{d} \mathcal N(0,1)\) for i.i.d. \(Xi\) with mean \(\mu\) and variance \(\sigma^2\).
---
🔄 Key Processes
Checking Measurability
Verify \(\{\omega:X(\omega)\le x\}\in\mathcal{F}\) for every real \(x\).
Finding the PDF of a Monotone Transform \(Y=g(X)\)
Compute inverse \(x=g^{-1}(y)\).
Apply \(fY(y)=fX\!\bigl(g^{-1}(y)\bigr)\,\bigl|\frac{d}{dy}g^{-1}(y)\bigr|\).
Finding the PDF of a Non‑Monotone Transform
Identify all pre‑images \(\{xi\}\) such that \(g(xi)=y\).
Sum contributions: \(fY(y)=\sumi fX(xi)\,\bigl|\frac{dxi}{dy}\bigr|\).
Convolution (sum of independent RVs)
Write integral (continuous) or sum (discrete) of product of densities/PMFs shifted by the argument.
Computing Moments via MGF
Differentiate \(MX(t)\): \(\displaystyle\mathbb{E}[X^k]=MX^{(k)}(0)\).
---
🔍 Key Comparisons
Discrete vs Continuous RV
Support: countable vs uncountable interval.
Probability of a point: >0 vs 0.
Descriptor: PMF vs PDF.
CDF vs PMF/PDF
CDF: cumulative probability \(P(X\le x)\).
PMF: derivative of CDF at points of mass (discrete jumps).
PDF: Radon–Nikodym derivative of CDF where it is absolutely continuous.
Equality in Distribution vs Almost‑Sure Equality
In distribution: same CDF, may differ on underlying sample space.
Almost sure: identical values with probability 1 (stronger).
Convergence Types
a.s. strongest → implies in probability → implies in distribution.
In distribution only cares about limiting CDF, not about probabilities of specific events.
---
⚠️ Common Misunderstandings
“PDF gives probability at a point” – false; \(fX(x)\) is a density, not a probability. Only integrals over intervals give probabilities.
Confusing convergence in probability with almost‑sure convergence – a.s. convergence requires the entire sequence to settle for almost every outcome, not just that the probability of large deviations shrinks.
Treating mixed‑type RVs as purely discrete or continuous – they have both a PMF component (jumps) and a PDF component (continuous part).
Assuming independence when only uncorrelated – zero covariance does not guarantee independence unless the joint distribution is known to be jointly normal.
---
🧠 Mental Models / Intuition
RV as a “black box”: input = outcome \(\omega\); output = a number. The distribution is what you see outside the box, regardless of the underlying \(\Omega\).
PDF as “mass density”: think of a thin sheet of sand; the height at \(x\) tells how much sand you’d collect over a small interval around \(x\).
Convolution = “smearing”: adding independent randomness spreads (smears) the distribution; the shape of the sum is the overlap of the two original shapes.
Transformations: monotone transforms stretch/compress the density; non‑monotone transforms fold the density and add contributions from each branch.
---
🚩 Exceptions & Edge Cases
Mixed‑type RVs: CDF has jumps (discrete mass) and a smooth part; the PDF exists only on intervals where the CDF is absolutely continuous.
Non‑invertible transformations with infinite roots: the formula with a sum over pre‑images assumes a finite or countable set of roots; infinite uncountable pre‑images require measure‑theoretic treatment beyond the simple sum.
MGF may not exist: heavy‑tailed distributions (e.g., Cauchy) have no finite MGF, yet moments of lower order may still exist.
---
📍 When to Use Which
Determine PMF vs PDF:
If the support is countable → use PMF.
If the support is an interval and \(P(X=x)=0\) for all \(x\) → use PDF.
Choose Convolution vs Direct Calculation:
Independent sum → convolution integral/sum.
Dependent sum → need joint distribution, not convolution.
Use MGF:
To find moments quickly and when the MGF exists (e.g., normal, exponential).
Otherwise compute moments directly from definition.
Select Convergence Test:
Prove a.s. convergence → use Borel‑Cantelli or almost‑sure criteria.
Prove convergence in probability → Chebyshev’s inequality or Slutsky’s theorem.
Prove convergence in distribution → show CDFs converge at continuity points or use characteristic functions.
---
👀 Patterns to Recognize
Jump in CDF ⇒ discrete mass at that point.
Flat segment in PDF ⇒ zero probability density (possible gap in support).
Symmetric PDF about 0 → mean = 0 (if it exists).
Sum of i.i.d. normals → still normal (closed under convolution).
When a transformation is monotone, only one term appears in the density‑change formula.
In CLT problems, look for “standardized sum” \((\sum Xi - n\mu)/(\sqrt{n}\sigma)\).
---
🗂️ Exam Traps
Choosing PMF for a continuous RV – answer will be wrong because \(P(X=x)=0\).
Confusing “in distribution” with “equal a.s.” – a distractor may state the two are equivalent; remember they differ.
Omitting absolute value in Jacobian when transforming PDFs – leads to negative densities.
Assuming independence for convolution – if dependence is hinted, convolution does not apply.
Using LLN for a single observation – LLN concerns averages of many i.i.d. variables, not a single draw.
Applying CLT without standardizing – the limit is standard normal only after centering and scaling.
---
or
Or, immediately create your own study flashcards:
Upload a PDF.
Master Study Materials.
Master Study Materials.
Start learning in seconds
Drop your PDFs here or
or