Chapter 2 Notation Summary

This chapter provides a summary of notation used throughout STM1001, designed to be used as a reference. For convenience, the chapter is divided into sections based on the topic for which the notation is most relevant, starting with a section containing general notation used throughout STM1001. The final section contains a consolidated table of all notation used throughout STM1001.

2.1 General notation used throughought STM1001

The table below provides a summary of general notation used throughout STM1001. Most of this notation is introduced early on in the subject, during Topics 1-4, and is then used throughout the subject.

Notation Meaning Comments
\(n\) Sample size
\(x_i\) The \(i\)th \(x\) value, usually from a list of \(n\) \(x\) values, e.g. (\(x_1, x_2, \ldots , x_n\)), where \(i\) can take any value from 1 to \(n\)
\(\displaystyle \sum\) Summation sign
\(\displaystyle \sum_{i=1}^{n}\) The sum from \(i=1\) to \(n\) See the previous chapter for an example
\(\mu\) Population mean This is a Greek letter pronounced 'm-yoo'
\(\overline{x}\) Sample mean The line on top of the \(x\) is referred to as a 'bar', so that \(\overline{x}\) is pronounced '\(x\) bar'
\(\sigma^2\) Population variance \(\sigma\) is a Greek letter pronounced 'sigma'
\(s^2\) Sample variance
\(\sigma\) Population standard deviation \(\sigma\) is a Greek letter pronounced 'sigma'
\(s\) Sample standard deviation
Q1 Quartile 1: 25% quantile
Q2 Quartile 2: 50% quantile, and also the median
Q3 Quartile 3: 75% quantile
\(\rho\) Population correlation This is a Greek letter pronounced 'rho'
\(r\) Sample correlation
\(|x|\) The 'absolute value' of some number \(x\) See the previous chapter for an example
\(X\) A random variable (discrete or continuous)
\(\text{E}(X)\) Expected value (or mean) of \(X\)
\(\text{Var}(X)\) Variance of \(X\)
\(\text{SD}(X)\) Standard deviation of \(X\)
\(\overline{X}\) Sample mean (random)
\(\pm\) Plus or minus. For example, \(x \pm a = (x - a, x + a)\)

2.2 Topics 3 and 4: Probability, Distributions and Sampling Distributions

The following table provides a summary of notation used in Topics 3 and 4.

The lecture slides for Topic 3 can be found here.

The readings for Topic 3 can be found here.

The lecture slides for Topic 4 can be found here.

The readings for Topic 4 can be found here.

Notation Meaning Comments
\(\Omega\) Sample space This is a Greek letter pronounced 'omega'
\(\emptyset\) Null event (or null set) Also sometimes referred to as 'empty set'
\(P(A)\) The probability of event \(A\)
\(A^C\) The complement of \(A\), or "not \(A\)" Sometimes denoted as \(A^\prime\)
\(X\) A random variable (discrete or continuous)
\(P(X = x)\) The probability that the random variable \(X\) takes the value \(x\). For example, let \(X\) denote the number of times you check this document. Then \(P(X=2)\) denotes the probability that you check this document exactly 2 times.
\(\text{E}(X)\) Expected value (or mean) of \(X\)
\(\text{Var}(X)\) Variance of \(X\)
\(\text{SD}(X)\) Standard deviation of \(X\)
\(\overline{X}\) Sample mean (random)
\(X \sim N(\mu, \sigma^2)\) \(X\) follows a Normal distribution with mean \(\mu\) and variance \(\sigma^2\) To define a Normal distribution, we need to know the mean and variance
\(Z \sim N(0, 1)\) \(Z\) follows the "Standard Normal distribution", that is, a Normal distribution with \(\mu = 0\) and \(\sigma^2 = 1\). The usual convention is to use \(Z\) instead of \(X\) when using the standard Normal distribution. The variance of \(Z\) is \(\sigma^2 = 1^2 = 1\). The standard deviation is also 1, since \(\sqrt{\sigma^2} = \sqrt{1^2} = 1\)
\(z\) \(z\)-score, defined as \(z = \displaystyle \frac{x - \mu}{\sigma}\) For a given value of \(x\), the corresponding \(z\)-score can be thought of as its "standardised" value
\(\pm\) Plus or minus. For example, \(x \pm a = (x - a, x + a)\)

2.3 Topics 5 and 6: Hypothesis testing and \(t\)-tests

The following table provides a summary of notation used in Topics 5 and 6.

The lecture slides for Topic 5 can be found here.

The readings for Topic 5 can be found here.

The lecture slides for Topic 6 can be found here.

The readings for Topic 6 can be found here.

Notation Meaning Comments
\(\pm\) Plus or minus. For example, \(x \pm a = (x - a, x + a)\)
\(t\)-distribution The distribution used for \(t\)-tests To define a \(t\)-distribution, we need to know the degrees of freedom
\(\text{df}\) Degrees of freedom
\(T \sim t_{\text{df}}\) \(T\) follows a \(t\)-distribution with degrees of freedom equal to \(\text{df}\). For example, if \(\text{df} = 1\), then \(T \sim t_1\)
\(H_0\) Null hypothesis
\(H_1\) Alternative hypothesis Sometimes denoted \(H_a\)
\(\mu_0\) The population mean under the null hypothesis
\(T\) Random test statistic
\(\overline{X}\) Sample mean (random)
\(\text{S}\) Estimator of the standard deviation of \(X\)
\(\text{SE}\) Estimator of the standard error, i.e. standard deviation of the sample mean Equal to \(\displaystyle\frac{S}{\sqrt{n}}\)
\(t\) Observed test statistic Sometimes called the '\(t\) value'
\(\bar{x}\) Observed sample mean
\(\text{s}\) Observed standard deviation
\(\text{se}\) Observed standard error, i.e. observed standard deviation of the sample mean Equal to \(\displaystyle\frac{s}{\sqrt{n}}\)
\(p\) \(p\)-value
\(\alpha\) Significance level, such that if \(p < \alpha\), we reject \(H_0\) Usually \(\alpha = 0.05\), but different values for \(\alpha\) can be chosen
Type I error The error that occurs when we reject \(H_0\) when \(H_0\) is true
Type II error The error that occurs when we fail to reject \(H_0\) when \(H_0\) is false
\(t_{\text{df,}1 - \alpha/2}\) The value from the \(t_{\text{df}}\) distribution such that \(P(T \leq t_{\text{df,}1 - \alpha/2}) = 1 - \alpha/2\), i.e. the \((1 - \alpha/2)\)th quantile. For example, if \(\alpha = 0.05\), \(1 - \alpha/2 = 1 - 0.05/2 = 1 - 0.025 = 0.975.\) We then have that \(P(T \leq t_{\text{df,}0.975}) = 0.975\)
\(d\) Effect size, Cohen's \(d\) We use Cohen's \(d\) for \(t\)-tests

2.4 Topic 7: One-way ANOVA

The following table provides a summary of notation used in Topic 7.

The lecture slides for Topic 7 can be found here.

The readings for Topic 7 can be found here.

Notation Meaning Comments
\(H_0\) Null hypothesis
\(H_1\) Alternative hypothesis Sometimes denoted \(H_a\)
\(p\) \(p\)-value
\(\alpha\) Significance level, such that if \(p < \alpha\), we reject \(H_0\) Usually \(\alpha = 0.05\), but different values for \(\alpha\) can be chosen
Type I error The error that occurs when we reject \(H_0\) when \(H_0\) is true
Type II error The error that occurs when we fail to reject \(H_0\) when \(H_0\) is false
ANOVA ANalysis Of VAriance
\(F\)-distribution The distribution used for ANOVA Hypothesis tests To define the \(F\)-distribution, we need to know \(d_1\) and \(d_2\)
\(d_1\) Degrees of freedom 1
\(d_2\) Degrees of freedom 2
\(N\) Total sample size (in the context of ANOVA)
\(k\) Number of groups (in the context of ANOVA)
\(F_{d_1, d_2}\) \(F\)-distribution with degrees of freedom 1 equal to \(d_1\) and degrees of freedom 2 equal to \(d_2\). For example, if \(d_1 = 3\) and \(d_2 = 45\), then our distribution is \(F_{3, 45}\)
\(\eta^2\) Effect size, 'eta squared' \(\eta\) is a Greek letter pronounced 'eta'. We use \(\eta^2\) for One-way ANOVA tests

2.5 Topic 8: Correlation and Simple Linear Regression

The following table provides a summary of notation used in Topic 8.

The lecture slides for Topic 8 can be found here.

The readings for Topic 8 can be found here.

Notation Meaning Comments
\(\rho\) Population correlation This is a Greek letter pronounced 'rho'
\(r\) Sample correlation
\(H_0\) Null hypothesis
\(H_1\) Alternative hypothesis Sometimes denoted \(H_a\)
\(p\) \(p\)-value
\(\alpha\) Significance level, such that if \(p < \alpha\), we reject \(H_0\) Usually \(\alpha = 0.05\), but different values for \(\alpha\) can be chosen
Type I error The error that occurs when we reject \(H_0\) when \(H_0\) is true
Type II error The error that occurs when we fail to reject \(H_0\) when \(H_0\) is false
\(x\) The explanatory variable, also referred to as the independent variable or predictor variable
\(y\) The response variable, also referred to as the dependent variable
\(\beta_0\) Intercept coefficient in the simple linear regression model \(\beta\) is a Greek letter pronounced 'beta'. \(\beta_0\) is pronounced 'beta nought'
\(\beta_1\) Slope coefficient in the simple linear regression model \(\beta\) is a Greek letter pronounced 'beta'. \(\beta_1\) is pronounced 'beta 1'
\(\epsilon\) Random error term in the simple linear regression model
\(\widehat{y}\) The estimated value for \(y\) based on a simple linear regression model The "\(\widehat{}\)" symbol is referred to as a 'hat' and is normally used to denote an estimate, so that \(\widehat{y}\) is pronounced '\(y\)-hat'
\(\widehat{\beta}_0\) The estimated value for \(\beta_0\) The "\(\widehat{}\)" symbol is referred to as a 'hat' and is normally used to denote an estimate, so that \(\widehat{\beta}_0\) is pronounced '\(\beta_0\)-hat'
\(\widehat{\beta}_1\) The estimated value for \(\beta_1\) The estimated value for \(\beta_0\)
\(R^2\) Coefficient of Determination. This value can be used to evaluate the fit of a simple linear regression model and is also the correlation squared.

2.6 Topic 9: Hypothesis testing for one and two sample proportions

The following table provides a summary of notation used in Topic 9.

The lecture slides for Topic 9 can be found here.

The readings for Topic 9 can be found here.

Notation Meaning Comments
\(H_0\) Null hypothesis
\(H_1\) Alternative hypothesis Sometimes denoted \(H_a\)
\(p\) \(p\)-value
\(\alpha\) Significance level, such that if \(p < \alpha\), we reject \(H_0\) Usually \(\alpha = 0.05\), but different values for \(\alpha\) can be chosen
Type I error The error that occurs when we reject \(H_0\) when \(H_0\) is true
Type II error The error that occurs when we fail to reject \(H_0\) when \(H_0\) is false
\(p\) In the context of a one-sample test of proportions, \(p\) is the proportion of a population with a certain characteristic Not to be confused with the \(p\)-value in the context of hypothesis testing
\(n\) In the context of a one-sample test of proportions, \(n\) either is the number of observations in a random sample, or the number of independent trials
\(x\) In the context of a one-sample test of proportions, \(x\) is either the number of observations in the sample that have a certain characteristics, or the number of success in \(n\) trials
\(p_0\) The population proportion under the null hypothesis
\(\widehat{p}\) The estimate of \(p\)
\(p_1\) The proportion of Population 1 (or Group 1) with a certain characteristic
\(p_2\) The proportion of Population 2 (or Group 2) with a certain characteristic
\(n_1\) The sample size from Population 1 (or Group 1)
\(n_2\) The sample size from Population 2 (or Group 2)
\(x_1\) The number of individuals in the sample from Population 1 (or Group 1) exhibiting the trait of interest
\(x_2\) The number of individuals in the sample from Population 2 (or Group 2) exhibiting the trait of interest
\(\hat{p}_1\) The estimate of \(p_1\)
\(\hat{p}_2\) The estimate of \(p_2\)

2.7 Topic 10: Chi-squared tests for categorical data

The following table provides a summary of notation used in Topic 10.

The lecture slides for Topic 10 can be found here.

The readings for Topic 10 can be found here.

Notation Meaning Comments
\(\displaystyle \sum\) Summation sign
\(\displaystyle \sum_{i=1}^{n}\) The sum from \(i=1\) to \(n\) See the previous chapter for an example
\(H_0\) Null hypothesis
\(H_1\) Alternative hypothesis Sometimes denoted \(H_a\)
\(p\) \(p\)-value
\(\alpha\) Significance level, such that if \(p < \alpha\), we reject \(H_0\) Usually \(\alpha = 0.05\), but different values for \(\alpha\) can be chosen
Type I error The error that occurs when we reject \(H_0\) when \(H_0\) is true
Type II error The error that occurs when we fail to reject \(H_0\) when \(H_0\) is false
\(\chi^2\)-distribution The distribution used for Chi-squared tests \(\chi\) is a Greek letter pronounced 'ky'
\(\chi^2_{\text{df}}\) \(\chi^2\)-distribution with degrees of freedom equal to \(\text{df}\). For example, if \(\text{df}=5\), then our distribution is \(\chi^2_5\)
\(X^2\) The random test statistic for a Chi-squared test. \(X^2 \sim \chi^2_{\text{df}}\) under \(H_0\)
\(\chi^2\) The observed test statistic for a Chi-squared test
\(O_i\) The observed frequency in the \(i\)th category in a Chi-squared goodness of fit test
\(E_i\) The expected frequency for the \(i\)th category in a Chi-squared goodness of fit test
\(k\) The number of categories in a Chi-squared goodness of fit test
\(O_{ij}\) The observed frequency in the \(i\)th row and the \(j\)th column in a Chi-squared test of independence
\(E_{ij}\) The expected frequency of the \(i\)th row and the \(j\)th column in a Chi-squared test of independence
\(r\) The number of rows in a Chi-squared test of independence Not to be confused with the sample correlation \(r\) in the context of correlation
\(c\) The number of columns in a Chi-squared test of independence

2.8 Topic 11: Statistical Power and Sample Size Calculation

The following table provides a summary of notation used in Topic 11.

The lecture slides for Topic 11 can be found here.

The readings for Topic 11 can be found here.

Notation Meaning Comments
\(H_0\) Null hypothesis
\(H_1\) Alternative hypothesis Sometimes denoted \(H_a\)
\(p\) \(p\)-value
\(\alpha\) Significance level, such that if \(p < \alpha\), we reject \(H_0\) Usually \(\alpha = 0.05\), but different values for \(\alpha\) can be chosen
Type I error The error that occurs when we reject \(H_0\) when \(H_0\) is true
Type II error The error that occurs when we fail to reject \(H_0\) when \(H_0\) is false
\(\alpha\) The probability of a Type I Error This is also the significance level
\(\beta\) The probability of a Type II Error Not to be confused with \(\beta_0\) or \(\beta_1\) in the context of simple linear regression

2.9 Complete table of all notation used throughout STM1001

The following table provides a summary of all notation used in STM1001. All notation summarised in the previous sections of this chapter are provided in the consolidated table below.

Notation Meaning Comments
\(n\) Sample size
\(x_i\) The \(i\)th \(x\) value, usually from a list of \(n\) \(x\) values, e.g. (\(x_1, x_2, \ldots , x_n\)), where \(i\) can take any value from 1 to \(n\)
\(\displaystyle \sum\) Summation sign
\(\displaystyle \sum_{i=1}^{n}\) The sum from \(i=1\) to \(n\) See the previous chapter for an example
\(\mu\) Population mean This is a Greek letter pronounced 'm-yoo'
\(\overline{x}\) Sample mean The line on top of the \(x\) is referred to as a 'bar', so that \(\overline{x}\) is pronounced '\(x\) bar'
\(\sigma^2\) Population variance \(\sigma\) is a Greek letter pronounced 'sigma'
\(s^2\) Sample variance
\(\sigma\) Population standard deviation \(\sigma\) is a Greek letter pronounced 'sigma'
\(s\) Sample standard deviation
Q1 Quartile 1: 25% quantile
Q2 Quartile 2: 50% quantile, and also the median
Q3 Quartile 3: 75% quantile
\(\rho\) Population correlation This is a Greek letter pronounced 'rho'
\(r\) Sample correlation
\(|x|\) The 'absolute value' of some number \(x\) See the previous chapter for an example
\(\Omega\) Sample space This is a Greek letter pronounced 'omega'
\(\emptyset\) Null event (or null set) Also sometimes referred to as 'empty set'
\(P(A)\) The probability of event \(A\)
\(A^C\) The complement of \(A\), or "not \(A\)" Sometimes denoted as \(A^\prime\)
\(X\) A random variable (discrete or continuous)
\(P(X = x)\) The probability that the random variable \(X\) takes the value \(x\). For example, let \(X\) denote the number of times you check this document. Then \(P(X=2)\) denotes the probability that you check this document exactly 2 times.
\(\text{E}(X)\) Expected value (or mean) of \(X\)
\(\text{Var}(X)\) Variance of \(X\)
\(\text{SD}(X)\) Standard deviation of \(X\)
\(\overline{X}\) Sample mean (random)
\(X \sim N(\mu, \sigma^2)\) \(X\) follows a Normal distribution with mean \(\mu\) and variance \(\sigma^2\) To define a Normal distribution, we need to know the mean and variance
\(Z \sim N(0, 1)\) \(Z\) follows the "Standard Normal distribution", that is, a Normal distribution with \(\mu = 0\) and \(\sigma^2 = 1\). The usual convention is to use \(Z\) instead of \(X\) when using the standard Normal distribution. The variance of \(Z\) is \(\sigma^2 = 1^2 = 1\). The standard deviation is also 1, since \(\sqrt{\sigma^2} = \sqrt{1^2} = 1\)
\(z\) \(z\)-score, defined as \(z = \displaystyle \frac{x - \mu}{\sigma}\) For a given value of \(x\), the corresponding \(z\)-score can be thought of as its "standardised" value
\(\pm\) Plus or minus. For example, \(x \pm a = (x - a, x + a)\)
\(t\)-distribution The distribution used for \(t\)-tests To define a \(t\)-distribution, we need to know the degrees of freedom
\(\text{df}\) Degrees of freedom
\(T \sim t_{\text{df}}\) \(T\) follows a \(t\)-distribution with degrees of freedom equal to \(\text{df}\). For example, if \(\text{df} = 1\), then \(T \sim t_1\)
\(H_0\) Null hypothesis
\(H_1\) Alternative hypothesis Sometimes denoted \(H_a\)
\(\mu_0\) The population mean under the null hypothesis
\(T\) Random test statistic
\(\text{S}\) Estimator of the standard deviation of \(X\)
\(\text{SE}\) Estimator of the standard error, i.e. standard deviation of the sample mean Equal to \(\displaystyle\frac{S}{\sqrt{n}}\)
\(t\) Observed test statistic Sometimes called the '\(t\) value'
\(\bar{x}\) Observed sample mean
\(\text{s}\) Observed standard deviation
\(\text{se}\) Observed standard error, i.e. observed standard deviation of the sample mean Equal to \(\displaystyle\frac{s}{\sqrt{n}}\)
\(p\) \(p\)-value
\(\alpha\) Significance level, such that if \(p < \alpha\), we reject \(H_0\) Usually \(\alpha = 0.05\), but different values for \(\alpha\) can be chosen
Type I error The error that occurs when we reject \(H_0\) when \(H_0\) is true
Type II error The error that occurs when we fail to reject \(H_0\) when \(H_0\) is false
\(t_{\text{df,}1 - \alpha/2}\) The value from the \(t_{\text{df}}\) distribution such that \(P(T \leq t_{\text{df,}1 - \alpha/2}) = 1 - \alpha/2\), i.e. the \((1 - \alpha/2)\)th quantile. For example, if \(\alpha = 0.05\), \(1 - \alpha/2 = 1 - 0.05/2 = 1 - 0.025 = 0.975.\) We then have that \(P(T \leq t_{\text{df,}0.975}) = 0.975\)
\(d\) Effect size, Cohen's \(d\) We use Cohen's \(d\) for \(t\)-tests
ANOVA ANalysis Of VAriance
\(F\)-distribution The distribution used for ANOVA Hypothesis tests To define the \(F\)-distribution, we need to know \(d_1\) and \(d_2\)
\(d_1\) Degrees of freedom 1
\(d_2\) Degrees of freedom 2
\(N\) Total sample size (in the context of ANOVA)
\(k\) Number of groups (in the context of ANOVA)
\(F_{d_1, d_2}\) \(F\)-distribution with degrees of freedom 1 equal to \(d_1\) and degrees of freedom 2 equal to \(d_2\). For example, if \(d_1 = 3\) and \(d_2 = 45\), then our distribution is \(F_{3, 45}\)
\(\eta^2\) Effect size, 'eta squared' \(\eta\) is a Greek letter pronounced 'eta'. We use \(\eta^2\) for One-way ANOVA tests
\(x\) In the context of simple linear regression, \(x\) is the explanatory variable, also referred to as the independent variable or predictor variable
\(y\) In the context of simple linear regression, \(y\) is the response variable, also referred to as the dependent variable
\(\beta_0\) Intercept coefficient in the simple linear regression model \(\beta\) is a Greek letter pronounced 'beta'. \(\beta_0\) is pronounced 'beta nought'
\(\beta_1\) Slope coefficient in the simple linear regression model \(\beta\) is a Greek letter pronounced 'beta'. \(\beta_1\) is pronounced 'beta 1'
\(\epsilon\) Random error term in the simple linear regression model
\(\widehat{y}\) The estimated value for \(y\) based on a simple linear regression model The "\(\widehat{}\)" symbol is referred to as a 'hat' and is normally used to denote an estimate, so that \(\widehat{y}\) is pronounced '\(y\)-hat'
\(\widehat{\beta}_0\) The estimated value for \(\beta_0\) The "\(\widehat{}\)" symbol is referred to as a 'hat' and is normally used to denote an estimate, so that \(\widehat{\beta}_0\) is pronounced '\(\beta_0\)-hat'
\(\widehat{\beta}_1\) The estimated value for \(\beta_1\) The estimated value for \(\beta_0\)
\(R^2\) Coefficient of Determination. This value can be used to evaluate the fit of a simple linear regression model and is also the correlation squared.
\(p\) In the context of a one-sample test of proportions, \(p\) is the proportion of a population with a certain characteristic Not to be confused with the \(p\)-value in the context of hypothesis testing
\(n\) In the context of a one-sample test of proportions, \(n\) either is the number of observations in a random sample, or the number of independent trials
\(x\) In the context of a one-sample test of proportions, \(x\) is either the number of observations in the sample that have a certain characteristics, or the number of success in \(n\) trials
\(p_0\) The population proportion under the null hypothesis
\(\widehat{p}\) The estimate of \(p\)
\(p_1\) In the context of a two-sample test of proportions, \(p_1\) is the proportion of Population 1 (or Group 1) with a certain characteristic
\(p_2\) In the context of a two-sample test of proportions, \(p_2\) is the proportion of Population 2 (or Group 2) with a certain characteristic
\(n_1\) In the context of a two-sample test of proportions, \(n_1\) is the sample size from Population 1 (or Group 1)
\(n_2\) In the context of a two-sample test of proportions, \(n_2\) is the sample size from Population 2 (or Group 2)
\(x_1\) In the context of a two-sample test of proportions, \(x_1\) is the number of individuals in the sample from Population 1 (or Group 1) exhibiting the trait of interest
\(x_2\) In the context of a two-sample test of proportions, \(x_2\) is the number of individuals in the sample from Population 2 (or Group 2) exhibiting the trait of interest
\(\hat{p}_1\) The estimate of \(p_1\)
\(\hat{p}_2\) The estimate of \(p_2\)
\(\chi^2\)-distribution The distribution used for Chi-squared tests \(\chi\) is a Greek letter pronounced 'ky'
\(\chi^2_{\text{df}}\) \(\chi^2\)-distribution with degrees of freedom equal to \(\text{df}\). For example, if \(\text{df}=5\), then our distribution is \(\chi^2_5\)
\(X^2\) The random test statistic for a Chi-squared test. \(X^2 \sim \chi^2_{\text{df}}\) under \(H_0\)
\(\chi^2\) The observed test statistic for a Chi-squared test
\(O_i\) The observed frequency in the \(i\)th category in a Chi-squared goodness of fit test
\(E_i\) The expected frequency for the \(i\)th category in a Chi-squared goodness of fit test
\(k\) The number of categories in a Chi-squared goodness of fit test
\(O_{ij}\) The observed frequency in the \(i\)th row and the \(j\)th column in a Chi-squared test of independence
\(E_{ij}\) The expected frequency of the \(i\)th row and the \(j\)th column in a Chi-squared test of independence
\(r\) The number of rows in a Chi-squared test of independence Not to be confused with the sample correlation \(r\) in the context of correlation
\(c\) The number of columns in a Chi-squared test of independence
\(\alpha\) The probability of a Type I Error This is also the significance level
\(\beta\) The probability of a Type II Error Not to be confused with \(\beta_0\) or \(\beta_1\) in the context of simple linear regression