Saral Shiksha Yojna
Courses/Behavioral Research: Statistical Methods

Behavioral Research: Statistical Methods

CG3.402
Vinoo AlluriMonsoon 2025-264 credits
Revision Notes/Unit 3 — Probability & Distributions/Probability, Distributions, and the CLT

Probability, Distributions, and the CLT

NotesStory

Intuition

Probability is either long-run frequency (frequentist) or degree of belief (Bayesian). PDFs describe continuous variables (probability of any *exact* value = 0); CDFs give cumulative probabilities. The Central Limit Theorem is the spell that makes everything else work: sample means → Normal as grows, regardless of population shape (provided finite variance). That is why parametric tests apply broadly — they assume normality of the *sampling distribution*, not of the raw data, and the CLT guarantees the former.

Explanation

The Fischer-Taimanov puzzle. 1971: Bobby Fischer vs Mark Taimanov, 6 games played, Fischer up 3–2–1. Probability Fischer wins game 7? Naïvely 3/6 = 0.5. But should the draw count? Should we treat all 6 games as equally informative? Every statistical method makes assumptions, and the answer depends on them. Fischer ultimately won 6–0 — the fair-coin model was wrong. Statistics is the art of choosing the right model, with humility.

Probability vs Statistics — the directions are opposite. *Probability* flows from model → data: given a fair coin, P(two tails in a row) = 0.25. *Statistics* (inferential) flows from data → model: Fischer won 3/3; given that, is P(Fischer wins) really 0.5 or something higher? Most of this course is the second direction.

Frequentist probability. Probability = long-run frequency of an event in repeated sampling. 'P(heads) = 0.5' means: flip a billion times, proportion of heads → 0.5. *Pros:* objective. *Cons:* counter-intuitive for one-off events. '70% chance of rain today' under frequentism means 'across the infinite class of days similar to this one, it rains 70% of the time' — try saying that to a friend with an umbrella.

Bayesian probability. Probability = degree of subjective belief, updated as evidence arrives. 'P(Carlsen beats Nepomniachtchi) = 0.7' means you, the observer, are 70% confident based on prior knowledge. *Pros:* applies to non-repeatable events. *Cons:* not fully objective — depends on priors. Most of this course uses frequentism; Bayes returns in Unit 13.

Independence. Two events are independent if knowing one tells you nothing about the other: , equivalently . Coin flips are independent; weather and whether you carry an umbrella are not.

i.i.d. — independent and identically distributed. A sequence is i.i.d. if (a) they're all independent of each other and (b) they all come from the same distribution. i.i.d. is the assumption underlying almost every test in this course. Violations (same participant tested multiple times → repeated measures; neighbouring brain voxels → spatial autocorrelation) require special methods.

Sample vs population. Population = full set you care about (all Indian voters). Sample = subset you actually have (your 1000 respondents). You almost never observe the population; you observe a sample and try to *infer* population properties.

Sampling distribution of a statistic — the secret heart of inferential statistics. Take a sample; compute a statistic (mean, median, t-value); take another sample; compute again. The distribution of those statistics across samples is the sampling distribution. Almost every test compares an observed statistic to its sampling distribution under some null. Distinct from the population distribution AND from a single sample's distribution.

PDF / PMF / CDF formally. Discrete: PMF assigns probability to each value. Continuous: PDF ; the probability of *any exact* value is zero. Total area under the PDF = 1. CDF runs from 0 to 1; it's the integral of the PDF.

Bernoulli(p). Simplest discrete: single trial with two outcomes. (success), (failure). One parameter.

Binomial(n, p). Sum of i.i.d. Bernoulli(p) trials. PMF . Mean , variance . Example: P(exactly 6 heads in 10 flips with ) = dbinom(6, 10, 0.7) ≈ 0.200.

**Normal .** Bell-shaped, symmetric. Two parameters: mean and SD . PDF . Empirical rule (exam staple): ~68% within , ~95% within , ~99.7% within . This is why outliers beyond 2–3 SDs are flagged. Standard Normal ; any Normal can be standardised via .

t-distribution. Bell-shaped but with heavier tails than Normal — more probability in extremes. One parameter: degrees of freedom . Used when sample size is small and is unknown (estimating with adds uncertainty, captured by heavier tails). As → Normal; at still visibly fatter; at indistinguishable. Critical for t-tests (Unit 7) and confidence intervals.

**Chi-square .** Take independent standard Normals, square each, sum: . Right-skewed, always . Mean , variance . Used in tests for categorical data (Unit 9) and inside variance estimates.

F-distribution. Ratio of two scaled s: where , , independent. Right-skewed, . Two df parameters for numerator and denominator. Used in ANOVA — F values near 1 indicate similar variances; indicates the numerator chi-square is much larger. F-tests are one-tailed by construction.

R's four-letter pattern. For any distribution xxx: dxxx density / PMF; pxxx cumulative CDF ; qxxx quantile (inverse CDF); rxxx random sample. Examples: dbinom(6, 10, 0.7) = 0.200; pbinom(4, 10, 0.7) ≈ 0.047; qnorm(0.975) ≈ 1.96; rnorm(100, mean=0, sd=1) for a sample. Warning for discrete CDFs: because discrete vars only take certain values, only some percentiles are achievable — R rounds up to the next reachable quantile.

Central Limit Theorem — the most important theorem in the course. *Given a sufficiently large sample size, the sampling distribution of the sample mean approximates a Normal distribution, regardless of the original population's distribution — as long as that population has finite variance.* The population can be anything (skewed, bimodal, weird); the *sampling distribution* of the mean is approximately Normal at large . This is why most parametric tests apply broadly — they assume normality of the test statistic's sampling distribution, which CLT guarantees, *not* normality of the raw data.

Two corollaries. (1) The sample mean is an *unbiased estimator* of the population mean: across hypothetical re-runs. (2) **SEM = ** — the standard deviation of the sampling distribution. Larger → smaller SEM → more precise estimates. **The in the denominator is famous: to halve your SEM, you need *4×* as much data.** Exam fixture.

Law of Large Numbers (LLN). As , . LLN is about the *point estimate* converging to the truth. CLT is about the *shape of variability* around that truth. LLN says you're going somewhere; CLT says the path is Normal.

When CLT hasn't kicked in. If your sample is tiny AND the raw data is highly non-Normal, the sampling distribution may also be non-Normal. Then use non-parametric tests (Unit 9). Rule of thumb: per group is sufficient for moderately skewed data.

Sampling with vs without replacement. With replacement: pure i.i.d. Without replacement: strictly not independent (each draw changes what's left), but if the *population* is much larger than the sample, the dependence is negligible — treat as i.i.d. Most methods assume with-replacement; without-replacement gives essentially the same answers for reasonably-sized populations.

Definitions

  • Frequentist probabilityLong-run frequency of an event in repeated sampling. Objective but counter-intuitive for one-off events.
  • Bayesian probabilityDegree of subjective belief, updated by evidence. Intuitive for one-off events; depends on priors.
  • Independent events, equivalently . Coin flips are independent; correlated measurements are not.
  • i.i.d.Independent AND identically distributed. The bedrock assumption of most inferential tests.
  • Sample vs populationPopulation = the full set of interest. Sample = the subset you actually observe. Inference goes sample → population.
  • Sampling distributionDistribution of a statistic across many hypothetical samples. The secret heart of inferential statistics — every test compares an observed statistic to this distribution under the null.
  • PDF / PMF / CDFDensity (continuous) / mass (discrete) / cumulative . For continuous RVs .
  • Bernoulli(p)Single yes/no trial. . Mean , variance .
  • Binomial(n, p)Sum of i.i.d. Bernoulli(p) trials. PMF . Mean , variance .
  • Normal $\mathcal{N}(\mu, \sigma^2)$Bell-shaped, symmetric, two parameters. 68/95/99.7 rule. Standard Normal is .
  • t-distributionLike Normal with heavier tails; one parameter (df). Use when unknown, small samples. → Normal as df → ∞.
  • Chi-square $\chi^2_k$Sum of squared standard Normals. Right-skewed, . Mean , variance . Used in tests.
  • F-distributionRatio of two scaled chi-squares. Right-skewed, . Two df parameters. Used in ANOVA / regression.
  • Central Limit Theorem (CLT)Sampling distribution of as grows, regardless of population shape (finite variance required).
  • Law of Large NumbersSample mean → population mean as . About convergence of the point estimate.
  • Standard Error of the Mean (SEM) — SD of the sampling distribution. Measures precision of as an estimate of .
  • Empirical rule (68/95/99.7)For Normal data, ~68% within , ~95% within , ~99.7% within .
  • Sampling with vs without replacementWith replacement is pure i.i.d. Without is dependent in principle but negligibly so when population ≫ sample.
  • R four-letter patternd density / PMF, p cumulative CDF, q quantile (inverse CDF), r random sample. Works for every distribution: norm, binom, t, chisq, f, …

Formulas

Derivations

Binomial mean and variance. Let with independent. Then , . By linearity of expectation: . By independence (sum of variances): .

**Standardisation: .** If , then and . Z is Normal because linear transformations of Normals are Normal. This is why tables of standard-Normal quantiles work universally.

Sketch of CLT. Let be i.i.d. with mean and variance . Define . Then and . Standardise: . The characteristic-function of converges to (the characteristic function of ) as . By Lévy's continuity theorem, , equivalently .

Why SEM = σ/√n. (using independence). Hence . To halve SEM, increase n by 4× — root-n is famous.

Why χ²_k has mean k. with . . By linearity . The variance is (uses fourth moments of Z, which equal 3, so ).

Binomial → Normal approximation. When is large and , are both , Binomial. Continuity correction: for , use . Applies because Binomial is a sum of i.i.d. Bernoullis — CLT in action.

Examples

  • Binomial PMF. P(exactly 6 heads in 10 flips, ) = . In R: dbinom(6, 10, 0.7).
  • Binomial CDF. P(at most 4 heads in 10 fair flips) = pbinom(4, 10, 0.5) ≈ 0.377.
  • Inverse CDF. What is the 4th percentile of Binomial(10, 0.7)? qbinom(0.04, 10, 0.7) = 4.
  • Standardisation example. SAT score X = 1400, , . . From the Normal table, P(Z ≤ 1.75) ≈ 0.96 — 96th percentile.
  • CLT in practice. Population is right-skewed (reaction times: most fast, some slow). Take samples of n = 5; sample means still slightly skewed. n = 20; bell-shaped. n = 100; indistinguishable from Normal. The skew of the population determines how fast convergence happens.
  • SEM trade-off. , : SEM = 2. To halve SEM to 1, need (4×).
  • χ² sampling. Generate 1000 random values by rchisq(1000, df=3). Mean ≈ 3, variance ≈ 6 (= ). Distribution is right-skewed with a long tail.
  • Empirical rule. Heights of adult men, cm, cm. Range = [161, 189] cm captures ~95%. A man at 200 cm is ~3.6σ above mean — extreme outlier.
  • i.i.d. violation. Same participant tested in two conditions → measurements within participant are correlated (not independent) → use paired t-test, not independent t-test.

Diagrams

  • The four key sampling distributions side-by-side. Normal (bell, symmetric); t with df = 5 (bell, heavier tails); χ²₃ (right-skewed, ≥ 0); F(3, 20) (right-skewed, ≥ 0). Annotate the parameter(s) of each.
  • CLT animation. Strongly right-skewed population. As n grows from 1 → 5 → 20 → 100, the sampling distribution of transforms from the original shape to a Normal. The width also shrinks (SEM ∝ 1/√n).
  • PDF vs CDF. A bell-shaped PDF and its S-shaped CDF integral. Mark P(a ≤ X ≤ b) as an area under the PDF and as a difference of CDF values.
  • Empirical rule on a Normal. ±1σ (68%), ±2σ (95%), ±3σ (99.7%) shaded under the bell.
  • Sampling distribution illustration. Population on the left (whatever shape); many samples drawn; means computed; histogram of means forms an approximately Normal sampling distribution.
  • **Standardisation .** Slide the Normal so ; rescale so . Demonstrate that quantiles are preserved.

Edge cases

  • ** for continuous RVs** — only intervals have positive probability. Pitfall in defining 'tied' Likert vs continuous responses.
  • CLT requires finite variance. Cauchy distribution has infinite variance; sample means do NOT concentrate on any value — drawing more samples doesn't help.
  • Binomial → Normal approximation valid when AND . Skewed cases (e.g., ) need very large n.
  • t at small df is meaningfully different from Normal. has noticeably heavier tails; using Normal critical values at small n is anti-conservative (you reject too often).
  • Heavy skew + small n → CLT hasn't converged. Use non-parametric methods (Mann-Whitney, Wilcoxon, Kruskal-Wallis) or transform the data.
  • Without-replacement at small populations. If sample size approaches population size, the i.i.d. approximation breaks; use finite-population corrections.
  • Independent ≠ uncorrelated in general, but for jointly Normal variables, uncorrelated implies independent.

Common mistakes

  • Confusing PMF (discrete, 0 ≤ P(X=k) ≤ 1) with PDF (continuous, can have — density not probability).
  • Saying CLT means *raw data* are Normal at large n. CLT is about the *sampling distribution of the mean*, not raw data.
  • Using Normal critical values at small n with unknown σ — should use t.
  • Stating Binomial variance as or similar — variance is .
  • Computing P(X = 4.5) for a continuous Normal — always zero; ask for P(X ≤ 4.5) or P(4 ≤ X ≤ 5).
  • Forgetting the in SEM: writing instead of .
  • Treating dependent observations as i.i.d. — same participant measured twice violates i.i.d.; use paired tests.
  • Treating itself as Normal even when the *population* is normal — true only if data are i.i.d.
  • Confusing the sampling distribution (across hypothetical samples) with the empirical distribution of one sample.

Shortcuts

  • i.i.d. assumption underlies almost every test.
  • Bernoulli sum → Binomial(n, p). Mean , variance .
  • CLT: sample-mean distribution → .
  • SEM = σ/√n. To halve, need 4× data.
  • R prefixes: d density / PMF, p CDF, q quantile (inverse CDF), r random.
  • Empirical rule: ±1σ → 68%, ±2σ → 95%, ±3σ → 99.7%.
  • Critical Normal quantiles: (95%), (99%).
  • ** mean = k; variance = 2k.**
  • t → Normal as df → ∞. At df ≥ ~30 they're nearly indistinguishable.

Proofs / Algorithms

** (unbiasedness of sample mean).** . By linearity: . The sample mean is an unbiased estimator of the population mean, *regardless* of the population distribution (only finite mean required).

** (variance of sample mean).** — using and independence (sum of variances). Hence SEM = .

CLT via characteristic functions (sketch). Let be the characteristic function of with mean and variance . By Taylor: . For : . Sum: — characteristic function of . By Lévy's continuity theorem, convergence in distribution follows.