Pearson, Spearman, Partial, Reliability Metrics
Maya Measures Together-ness
Maya's next study isn't about turmeric milk any more. She's interested in whether students who spend more hours studying actually get better exam scores. Two continuous variables: study hours, exam scores. No experimental manipulation, just observed pairs. The natural question: do these two variables move together?
That question has a number attached to it, and the number is called a correlation coefficient. This unit covers correlation, its cousins (partial and semi-partial), the eternal warning that correlation is not causation, and then the deeper machinery of reliability — Cohen's κ, Cronbach's α, and how to handle outliers that wreck your analyses.
Correlation is NOT causation
Before any math, the warning. If you remember nothing else from this unit, remember this.
Two variables can be correlated for any of four reasons:
1. A causes B. Smoking causes lung cancer. 2. B causes A. Lung cancer causes people to smoke more for relief. 3. A third variable C causes both. Ice cream sales and drowning deaths are correlated — but neither causes the other. Hot weather causes both. 4. Coincidence / sampling noise. Especially in small samples, two genuinely unrelated variables can show a high correlation just by chance.
Showing a correlation establishes that two things move together. It does not establish why. To claim causation you need either an experiment (with manipulation and randomisation) or extremely careful causal-inference methods that rule out confounders. This is one of the most heavily examined statements in BRSM — there will almost certainly be a question that gives you a correlation and asks if you can conclude causation. The answer is always: *not from correlation alone.*
Pearson's r — the standard correlation coefficient
When both variables are continuous and approximately normally distributed, the standard measure of linear association is **Pearson's correlation coefficient **:
Equivalently, is the average product of z-scores: .
Interpretation rules
- ranges from −1 to +1.
- : perfect positive linear relationship.
- : no linear relationship.
- : perfect negative linear relationship.
- Sign = direction; magnitude = strength.
Strength benchmarks (common exam table)
| | Interpretation | | --- | --- | | < 0.1 | negligible | | 0.1–0.3 | small / weak | | 0.3–0.5 | moderate | | ≥ 0.5 | large / strong |
The coefficient of determination, r²
The square of . Proportion of variance shared between X and Y.
If , then → *"49% of the variance in Y is explained by its linear relationship with X."* A very common reporting style.
**For simple linear regression with one predictor, equals ** — we'll revisit this in Unit 12.
What Pearson's r does NOT measure
- Non-linear relationships. A perfect parabola has — Pearson is blind to curvature. Always look at the scatter plot.
- Heteroscedastic spreads. Pearson r still gives one number even when y-spread depends on x.
- Causation. Already covered.
Pearson's r is sensitive to outliers
A single extreme point can drag dramatically. Two scatter plots can both yield — one because the data really hugs a line, and the other because two extreme points anchor the slope while the rest of the data is a noisy cloud.
This is exactly why Anscombe's quartet (Unit 4) matters: identical r, four totally different stories.
Spearman's ρ — the rank-based alternative
When the data are ordinal, when the relationship is monotonic but not linear, or when outliers are wrecking Pearson's r, use Spearman's ρ.
Spearman's ρ is just Pearson's r computed on the ranks of the data, not the raw values. Replace each x with its rank within x's; each y with its rank within y's; compute Pearson r on the ranks.
Properties:
- Ranges from to , same as Pearson.
- Monotonicity, not linearity. when each variable is a perfectly increasing function of the other (linear or not).
- Robust to outliers because ranks compress extreme values to merely the highest or lowest rank.
- Works for ordinal data — no interval-scale assumption.
Pearson vs Spearman — when to use which
The exam loves this comparison:
| Situation | Use | | --- | --- | | Both vars interval/ratio + Normal + no outliers + linear | Pearson r | | Ordinal OR non-Normal OR outliers OR monotonic-non-linear | Spearman ρ |
If both are valid, they give similar values; Pearson is slightly more powerful when its assumptions hold.
A short-answer favourite: *"Why is Spearman's ρ less sensitive to outliers than Pearson's r?"* Because Spearman uses ranks; the largest value, no matter how extreme, becomes just rank . The numerical extremity gets compressed away.
Significance of correlation — strength vs significance
You compute . Is that statistically significant? It depends on sample size.
Statistical significance tests (in the population). Test statistic:
Practical consequences:
- With , you can get just from noise. Apparently strong correlations from tiny samples are unreliable.
- With , an of 0.10 — barely visible to the eye — can be statistically significant. The effect is real but tiny.
Two things to memorise:
1. Strength is shown by the *magnitude* of — close to ±1 is strong. 2. Statistical significance is shown by the *p-value*, which depends on both AND .
When n is low, the odds are high that a "good-looking" correlation occurred by chance. When n is high, even meaningless correlations may pass the significance bar. Same "statistical vs practical significance" lesson as always.
Partial correlation
Maya cares about the relationship between test scores and IQ, but she worries that study hours might be inflating their correlation. People who study more both have higher IQs *and* score higher on tests. So part of the apparent IQ–score relationship might be "just" the effect of study time on both.
To isolate the relationship while controlling for study hours, she computes a partial correlation.
Definition: the correlation between X and Y after removing the linear effect of Z from both X and Y.
Mechanically:
1. Regress X on Z; compute residuals. 2. Regress Y on Z; compute residuals. 3. Compute the Pearson correlation between the two residual sets.
Formula (one control variable):
A worked exam-style example
Given:
Partial correlation between GPA and IQ controlling for study hours:
So the GPA–IQ relationship is still strong (0.67) even after accounting for study hours. Study hours explained some of the original 0.75 correlation, but most of it remains. Maya can argue IQ has a relationship with GPA above and beyond what is explained by study hours.
Semi-partial correlation
A subtler cousin. Partial correlation removes Z from both X and Y. Semi-partial removes Z from only one of them.
For instance, Maya wants the relationship between tutoring and exam scores, removing the confounding effect of study time only on tutoring (because students who get tutored may have less study time, and she wants tutoring's *unique* contribution to exam scores).
Mechanically:
1. Regress tutoring on study time; compute residuals. 2. Compute the Pearson correlation between those residuals and the raw exam scores.
The result: the relationship between the *unique* part of tutoring (independent of study time) and exam scores in their original form.
Semi-partial correlations are central to interpreting multiple regression — the unique contribution of each predictor beyond the others (Unit 12).
Reliability deepened — quantifying agreement
Back in Unit 2 we listed the kinds of reliability. Now we'll see how each is measured.
Recall the four kinds (plus one more):
- Test-retest — same instrument, same people, two time points.
- Inter-rater — same instrument, different raters, same targets.
- Parallel forms — two equivalent versions of the same instrument.
- Internal consistency — different parts of the same instrument should agree.
- Intra-rater — same rater rating the same things at two different time points.
Each has a quantitative measure. Memorise the pairings.
Cohen's Kappa — two raters, nominal data
Two raters rate items into nominal categories. Just looking at how often they agree isn't enough — they might agree a lot by chance. Cohen's κ corrects for chance agreement.
where = observed agreement, = expected agreement by chance.
Worked example: 50 images rated yes/no by two raters.
| | R2: Yes | R2: No | | --- | --- | --- | | R1: Yes | 4 | 16 | | R1: No | 15 | 15 |
- Observed agreement: .
- Rater 1 said Yes: . Rater 2 said Yes: .
- Expected by chance: both Yes = ; both No = . Total .
- .
Negative κ means less agreement than chance — quite bad.
Cohen's κ interpretation (Landis & Koch)
| κ | Agreement | | --- | --- | | < 0 | worse than chance | | 0–0.20 | slight | | 0.20–0.40 | fair | | 0.40–0.60 | moderate | | 0.60–0.80 | substantial | | 0.80–1.00 | almost perfect |
The cousins
- Fleiss' κ — generalisation of Cohen's κ for more than two raters rating into nominal categories.
- Kendall's W (coefficient of concordance) — multiple raters ranking items rather than placing them in nominal categories. Range 0 to 1.
- Krippendorff's α — Swiss army knife: any number of raters, missing data, any measurement level.
- Intra-rater reliability with Pearson — same rater at two time points, continuous data → Pearson r.
Cronbach's alpha — internal consistency
Most common measure of internal consistency — do all items on a test measure the same underlying construct?
where k = number of items, = variance of item i, = variance of total scores.
Intuition: if items are highly intercorrelated, the variance of the total score is much larger than the sum of individual item variances → α high.
Interpretation:
| α | Quality | | --- | --- | | ≥ 0.9 | excellent (sometimes too high — items may be redundant) | | 0.8–0.9 | good | | 0.7–0.8 | acceptable | | 0.6–0.7 | questionable | | < 0.6 | poor |
Split-half reliability is a simpler historical alternative. Kuder-Richardson 20/21 is the related index for binary-item tests.
Outliers — when to worry, when to keep
An outlier is a data point unusually distant from the rest. The decision to keep, remove, or transform it is one of the most consequential choices in analysis, and there is no universal rule.
Why outliers happen
1. Measurement or execution errors — instrument malfunction, improper handling, extraction errors. 2. Data entry errors / missing-data codes — someone entered "999" for missing. 3. Data processing errors — accidental scaling, misaligned columns. 4. Sampling errors — mixed data from different sources. 5. Natural variability / novelties — sometimes data is just genuinely unusual. A real IQ prodigy. A miraculous treatment responder.
Errors should be corrected or removed; natural outliers should be respected and investigated.
Detecting outliers
Graphical: scatter plots, box plots (1.5 × IQR rule), histograms.
Numerical: 1.5 × IQR rule, 2 or 3 SDs from mean, Grubbs' test, Tietjen-Moore.
Dealing with outliers — five strategies
1. Omit — only with strong evidence of error. Document and justify. 2. Replace — winsorise to a nearest non-outlier value. 3. Use different analysis methods — non-parametric tests (Mann-Whitney, Wilcoxon, Spearman). 4. Value the outliers — keep them and investigate. Sometimes the most interesting data. 5. Transform the data — log / sqrt / Box-Cox to pull in extremes.
The default mistake students make: silently dropping outliers because they look weird. The exam will mark you down. Detect, investigate, decide based on cause, document.
What you carry into the exam
- Correlation is NOT causation — four reasons. Memorise verbatim.
- Pearson r — linear association on continuous Normal data, , sensitive to outliers and blind to non-linearity.
- Spearman ρ — Pearson r on ranks. Monotone, robust, ordinal-friendly.
- r² = proportion of variance shared.
- r is strength; p depends on r AND n. Apparently 'strong' r in small n is suspicious; weak r in large n is real but trivial.
- Partial r removes Z from both X and Y; semi-partial removes from one side only.
- Cohen's κ — chance-corrected nominal agreement, two raters. Fleiss κ for >2; Kendall W for ordinal; Krippendorff α general; Pearson r for intra-rater continuous.
- Cronbach's α — internal consistency. > 0.7 acceptable; > 0.95 may indicate redundancy.
- Outliers — detect, investigate, decide; don't silently drop.
When you're ready, send "next" and we'll move into hypothesis testing — NHST, p-values, Type I and II errors, power, Cohen's d, and the t-test family.