Skip to content

S3 Chapter 7: International Exam Review

This handout is a high-yield revision guide for S3 topics: Sampling \rightarrow Combinations of RVs \rightarrow Estimation & CI \rightarrow CLT & Mean Tests \rightarrow Correlation \rightarrow χ2\chi^2 Tests.

We will keep using HelloTea to connect ideas:

  • Population: all students (e.g. 3000).
  • Sample: e.g. n=200n=200 students chosen by a sampling method.
  • Data types: ratings (1—5), drink choice (tea/coffee/hot chocolate), screen-time, etc.

Chapter 1 Review: Sampling Methods (Getting Good Data)

Section titled “Chapter 1 Review: Sampling Methods (Getting Good Data)”

Definition: Population, Sample, Sampling Frame

  • Population: the full group of interest.
  • Sample: the selected observations from the population.
  • Sampling frame: the actual list you can sample from.
MethodRandom?How to do itMain risk / limitation
Simple Random (SRS)Yeschoose nn IDs using RNG / random number tablecan be time-consuming; may miss small subgroups by chance
SystematicPartlychoose random start, then every kkthperiodicity (hidden patterns in the list)
StratifiedYes (within strata)split into strata, SRS inside eachneed strata info beforehand; more steps
QuotaNoset quotas, then convenience within eachselection bias; no valid sampling error / inference guarantee

Common Exam Pitfalls (From Examiner Reports)

Section titled “Common Exam Pitfalls (From Examiner Reports)”
  • Missing the numbering step: Before using random numbers, you MUST explicitly state that you will “number/label the sampling frame (e.g., from 1 to NN).”
  • Systematic sampling errors: If the period is kk, students often forget that you cannot select two adjacent items.
  • Vague language: Stating a method is “more accurate” or “more representative” usually scores zero. Use precise terms like “reflects the population structure” (for stratified) or “gives every item an equal chance of selection” (for simple random).
  • Quota vs. Stratified: Quota sampling suffers from interviewer bias (the person choosing who to survey), which means no valid sampling error can be calculated.

Chapter 2 Review: Combinations of Random Variables

Section titled “Chapter 2 Review: Combinations of Random Variables”

The Biggest Exam Trap: 3X3X vs X1+X2+X3X_1+X_2+X_3

Section titled “The Biggest Exam Trap: 3X3X3X vs X1+X2+X3X_1+X_2+X_3X1​+X2​+X3​”
ScenarioNotationVariance
”3 times the weight of a randomly chosen bag"3X3XVar(3X)=32Var(X)=9Var(X)\mathrm{Var}(3X) = 3^2 \mathrm{Var}(X) = \mathbf{9\mathrm{Var}(X)}
"The total weight of 3 randomly chosen bags”X1+X2+X3X_1 + X_2 + X_3Var(X1+X2+X3)=3Var(X)\mathrm{Var}(X_1+X_2+X_3) = \mathbf{3\mathrm{Var}(X)}

Common Exam Pitfalls (From Examiner Reports)

Section titled “Common Exam Pitfalls (From Examiner Reports)”
  • Subtracting variances: Students frequently write Var(XY)=Var(X)Var(Y)\mathrm{Var}(X-Y) = \mathrm{Var}(X) - \mathrm{Var}(Y). This is WRONG! Variances always add for independent variables: Var(XY)=Var(X)+Var(Y)\mathrm{Var}(X-Y) = \mathrm{Var}(X) + \mathrm{Var}(Y).
  • Averaging variables: To find the variance of the sample mean of 5 observations A=X1+...+X55A = \frac{X_1+...+X_5}{5}, you must square the denominator: Var(A)=5Var(X)25=Var(X)5\mathrm{Var}(A) = \frac{5\mathrm{Var}(X)}{25} = \frac{\mathrm{Var}(X)}{5}. Many incorrectly divide by 5 instead of 25.
  • Difference without direction: If the question asks for the probability of a “difference” in weight being greater than 5g, you must calculate P(XY>5)=P(XY>5)+P(XY<5)P(|X-Y| > 5) = P(X-Y > 5) + P(X-Y < -5) (two tails).
  • Standardisation sign errors: When equating your standardisation Z=xμσZ = \frac{x-\mu}{\sigma} to a critical value (e.g. 1.2816), make sure the signs match. If the probability area implies xx is below the mean, ZZ must be negative!

Chapter 3 Review: Estimation, Bias, Standard Error, Confidence Interval

Section titled “Chapter 3 Review: Estimation, Bias, Standard Error, Confidence Interval”

The Three Layers: Parameter, Statistic, Value

Section titled “The Three Layers: Parameter, Statistic, Value”

Definition: Bias

Bias(θ^)=E[θ^]θ.\mathrm{Bias}(\hat{\theta}) = E[\hat{\theta}] - \theta.

Definition: Standard Error

SE(θ^)=Var(θ^).\mathrm{SE}(\hat{\theta}) = \sqrt{\mathrm{Var}(\hat{\theta})}.

Definition: Generic Form

Estimate±(critical value)×(standard error).\text{Estimate} \pm (\text{critical value}) \times (\text{standard error}).

Common Exam Pitfalls (From Examiner Reports)

Section titled “Common Exam Pitfalls (From Examiner Reports)”
  • Wrong interpretation: “95% probability that μ\mu is in this interval” (not correct phrasing). Instead say: “We are 95% confident that the true population mean lies within this interval.”
  • Mixing SD and SE: SD describes individual items; SE describes estimator variability. Don’t forget to divide by n\sqrt{n} when calculating the standard error!
  • Hypotheses notation: Always use population parameters (e.g. μ\mu) in hypotheses, never sample statistics (xˉ\bar{x}). Also, define your subscripts clearly (e.g. μA\mu_A vs μB\mu_B).
  • CIs as a Binomial process: If asked for the probability that YY out of nn calculated confidence intervals contain μ\mu, you must use the Binomial distribution YB(n,confidence level)Y \sim B(n, \text{confidence level}).

Chapter 4 Review: CLT & Inference for Means

Section titled “Chapter 4 Review: CLT & Inference for Means”

Theorem: Central Limit Theorem (usable form) If X1,,XnX_1,\ldots,X_n are i.i.d. with mean μ\mu and variance σ2<\sigma^2<\infty, then for large nn,

XˉN ⁣(μ,σ2n).\bar{X} \approx N\!\left(\mu,\frac{\sigma^2}{n}\right).

One-Sample Mean Test (Large-Sample zz-test idea)

Section titled “One-Sample Mean Test (Large-Sample zzz-test idea)”

To test H0:μ=μ0H_0:\mu=\mu_0,

Z=xˉμ0S/nN(0,1) under H0(large n).Z=\frac{\bar{x}-\mu_0}{S/\sqrt{n}} \approx N(0,1)\ \text{under }H_0\quad(\text{large }n).

Decision by critical value or pp-value.

Difference of Two Means (Independent Samples)

Section titled “Difference of Two Means (Independent Samples)”

If two independent large samples:

XˉYˉN ⁣(μXμY, σX2nX+σY2nY).\bar{X}-\bar{Y} \approx N\!\left(\mu_X-\mu_Y,\ \frac{\sigma_X^2}{n_X}+\frac{\sigma_Y^2}{n_Y}\right).

Use estimated SE with sample SDs.

Common Exam Pitfalls (From Examiner Reports)

Section titled “Common Exam Pitfalls (From Examiner Reports)”
  • Explaining the CLT: Many students lose marks by saying “the sample is normally distributed.” You must say “the sample mean is approximately normally distributed.”
  • Combining samples: When asked to treat two samples as a single combined sample, do not calculate a weighted standard deviation. Find the new overall mean and calculate the standard error for the new total size n1+n2n_1+n_2.
  • Using CLT with small nn when the population is clearly skewed/heavy-tailed.
  • Forgetting “independent samples” for the two-sample formula.
  • Mixing up one-tailed vs two-tailed critical regions.

Chapter 5 Review: Correlation & Rank Correlation

Section titled “Chapter 5 Review: Correlation & Rank Correlation”

Given paired data (xi,yi)(x_i,y_i) for i=1,,ni=1,\ldots,n,

r=SxySxxSyy,Sxy=xy(x)(y)n.r=\frac{S_{xy}}{\sqrt{S_{xx}S_{yy}}},\quad S_{xy}=\sum xy-\frac{(\sum x)(\sum y)}{n}.
  • H0:ρ=0H_0:\rho=0
  • Compare r|r| with the critical value for (n,α)(n,\alpha).

Use ranks when:

  • relationship is monotonic but not linear, or data is ordinal, or outliers break Pearson.

If no ties, shortcut:

rs=16d2n(n21).r_s = 1-\frac{6\sum d^2}{n(n^2-1)}.

Test rsr_s using Spearman critical value tables.

Common Exam Pitfalls (From Examiner Reports)

Section titled “Common Exam Pitfalls (From Examiner Reports)”
  • Tied Ranks: If there are tied ranks, you MUST use the full PMCC formula on the ranks. The 16d2n(n21)1-\frac{6\sum d^2}{n(n^2-1)} shortcut is only valid when there are no ties!
  • Alphabetical coding: When given letters (e.g., Grades A, B, C), students sometimes rank them alphabetically instead of by their actual value/order.
  • Hypotheses: Always state hypotheses in terms of ρ\rho or ρs\rho_s. Never use rr or state them just in words.
  • Contextual conclusion: Simply stating “there is correlation” is insufficient. You must include the direction and context (e.g., “there is evidence of positive correlation between age and price”).
  • Non-linear relationships: If a PMCC test shows no significant correlation, but a Spearman’s test shows significant correlation, it strongly suggests a non-linear relationship exists.

Chapter 6 Review: χ2\chi^2 Tests (Goodness of Fit & Independence)

Section titled “Chapter 6 Review: χ2\chi^2χ2 Tests (Goodness of Fit & Independence)”

χ2\chi^2 Statistic (same structure for both tests)

Section titled “χ2\chi^2χ2 Statistic (same structure for both tests)”
χ2=(OE)2E.\chi^2=\sum \frac{(O-E)^2}{E}.
  • Use when: one categorical variable, testing a specified distribution/model.
  • Hypotheses: H0H_0: model fits; H1H_1: model does not fit.
  • df: df=k1mdf=k-1-m where m=m= parameters estimated from data.
  • Use when: two categorical variables; test for association.
  • Expected: Eij=(row total)(col total)grand totalE_{ij}=\dfrac{(\text{row total})(\text{col total})}{\text{grand total}}.
  • df: (r1)(c1)(r-1)(c-1).

Common Exam Pitfalls (From Examiner Reports)

Section titled “Common Exam Pitfalls (From Examiner Reports)”
  • Frequencies, not percentages: A Chi-squared test MUST use raw frequencies (counts). If given percentages, convert them back to frequencies first.
  • Hypotheses for estimated parameters: If you estimate a parameter (e.g. λ=3.5\lambda=3.5), do NOT include the 3.5 in your hypotheses. Write ”H0H_0: A Poisson distribution is a suitable model” (not “Po(3.5)”).
  • Degrees of Freedom (mm): Students often forget to subtract mm (the number of estimated parameters) when calculating df=k1mdf = k - 1 - m for Goodness of Fit tests.
  • Pooling correctly: You pool cells to ensure Expected frequencies are 5\ge 5. Do not pool based solely on Observed frequencies!

One-Page Formula Sheet (Students Should Memorise)

Section titled “One-Page Formula Sheet (Students Should Memorise)”
  • Expectation: E(aX±bY)=aE(X)±bE(Y)E(aX \pm bY)=aE(X) \pm bE(Y)
  • Variance (Indep): Var(aX±bY)=a2Var(X)+b2Var(Y)\mathrm{Var}(aX \pm bY)=a^2\mathrm{Var}(X) + b^2\mathrm{Var}(Y)
  • Multiple items: Var(X1+..+Xn)=nVar(X)\mathrm{Var}(X_1+..+X_n)=n\mathrm{Var}(X)
  • Sample mean: Xˉ=1nXi\bar{X}=\dfrac{1}{n}\sum X_i, SE(Xˉ)=σn\mathrm{SE}(\bar{X})=\dfrac{\sigma}{\sqrt{n}}
  • Sample variance: S2=1n1(XiXˉ)2S^2=\dfrac{1}{n-1}\sum (X_i-\bar{X})^2
  • CI for mean (large nn): xˉ±zSn\bar{x}\pm z^*\dfrac{S}{\sqrt{n}}
  • CLT: XˉN ⁣(μ,σ2n)\bar{X}\approx N\!\left(\mu,\dfrac{\sigma^2}{n}\right)
  • PMCC: r=SxySxxSyyr=\dfrac{S_{xy}}{\sqrt{S_{xx}S_{yy}}}
  • Spearman (no ties): rs=16d2n(n21)r_s=1-\dfrac{6\sum d^2}{n(n^2-1)}
  • χ2\chi^2: χ2=(OE)2E\chi^2=\sum\dfrac{(O-E)^2}{E}
  • df GOF: k1mk-1-m, df independence: (r1)(c1)(r-1)(c-1)