Glossary: M09 — Parametric and Nonparametric Tests
Module: M09 Formulas: Formula Sheet
Parametric Test (detailed)
A hypothesis test that relies on assumptions about the probability distribution of the population (typically that the data are normally distributed) and tests hypotheses about specific parameters such as the mean or variance. Requires data measured on at least an interval scale.
LOS: 9.a | When to use: When underlying distribution assumptions are met and data are interval or ratio scale. | Examples: t-Test, F-Test, z-Test. | Contrast: Nonparametric Test (detailed)
Nonparametric Test (detailed)
A hypothesis test that does not rely on assumptions about the probability distribution of the population, or that tests hypotheses that are not about specific parameters. Used when data violate normality assumptions, are measured on an ordinal scale, or when sample sizes are very small.
LOS: 9.a | When to use: (1) Data are ranked (ordinal); (2) population is non-normal and sample is small; (3) hypothesis is not about a parameter (e.g., independence); (4) data contain outliers. | Advantage: Distribution-free. | Disadvantage: Less powerful than parametric tests when parametric assumptions hold.
Spearman Rank Correlation
A nonparametric measure of the monotonic relationship between two variables, based on the ranks of the observations rather than their actual values. Ranges from −1 to +1.
where = difference between the ranks of the -th pair of observations and = number of pairs.
LOS: 9.b | Test statistic (to test : ):
Key: More robust than Pearson correlation when data have outliers or are not normally distributed. | Related: Pearson Correlation
Chi-Square Test of Independence
A nonparametric test that determines whether two categorical variables are statistically independent. Uses a Contingency Table (test context) to compare observed and expected frequencies.
where = number of rows and = number of columns in the contingency table.
LOS: 9.c | Decision: Reject (independence) if . | Related: Chi-Square Test, Contingency Table
Contingency Table (test context)
A cross-tabulation matrix that displays the frequency distribution of two categorical variables simultaneously. Used as the input for the Chi-Square Test of Independence.
& B_1 & B_2 & \text{Total} \\ \hline A_1 & O_{11} & O_{12} & R_1 \\ A_2 & O_{21} & O_{22} & R_2 \\ \hline \text{Total} & C_1 & C_2 & n \end{array}$$ **LOS**: 9.c | **Related**: [[quantitative-methods/glossary/m03-statistical-measures#contingency-table|Contingency Table (descriptive)]] --- ## Expected Frequency The frequency that would be expected in each cell of a contingency table if the two variables were perfectly independent. Used in the chi-square test. $$E_{ij} = \frac{R_i \times C_j}{n}$$ where $R_i$ = row $i$ total, $C_j$ = column $j$ total, $n$ = overall sample size. **LOS**: 9.c | **Rule of thumb**: All expected frequencies should be at least 5 for the chi-square approximation to be valid. --- ## Observed Frequency The actual count of observations falling in each cell of a [[#contingency-table-test-context|Contingency Table (test context)]]. Compared to [[#expected-frequency|Expected Frequency]] in the chi-square test statistic. **LOS**: 9.c | **Notation**: $O_{ij}$ = observed frequency in row $i$, column $j$.