Review the key concepts, formulae, and examples before starting your quiz.
🔑Concepts
Null and Alternative Hypotheses ( and ): The Null Hypothesis () represents the status quo or a statement of 'no effect' or 'independence'. The Alternative Hypothesis () is the claim we are testing for. Visually, is represented by the central portion of a probability distribution curve, while points toward the 'rejection regions' or tails of the distribution.
Significance Level () and P-value: The significance level (typically or ) is the probability threshold for rejecting . The p-value is the probability of observing results as extreme as the sample data, assuming is true. Visually, the p-value is the area under the curve in the tail(s); if this area is smaller than the area defined by , we reject .
Chi-squared () Test for Independence: This test evaluates whether two categorical variables are independent based on a contingency table. Visually, imagine a table comparing 'Observed' () frequencies against 'Expected' () frequencies. The distribution curve is positively skewed (starting at zero and trailing to the right), and a high value moves the result into the shaded rejection region in the right tail.
Degrees of Freedom (): This parameter defines the specific shape of the or t-distribution curve. For a test of independence, where is rows and is columns. Visually, as degrees of freedom increase, the peak of the curve shifts to the right and becomes more symmetrical, resembling a normal distribution.
One-sample and Two-sample t-tests: These tests compare means. A one-sample t-test compares a sample mean to a hypothesized population mean, while a two-sample t-test compares the means of two independent groups. Visually, the t-distribution looks like a bell-shaped Normal curve but has 'fatter tails' to account for greater uncertainty in smaller samples.
Critical Values and Rejection Regions: The critical value is the 'cut-off' point on the x-axis of a distribution. Visually, this value separates the 'fail to reject' region (the main body of the curve) from the 'rejection region' (the tails). If your calculated test statistic (t or ) falls beyond this value into the tail, the result is statistically significant.
One-tailed vs. Two-tailed Tests: A one-tailed test looks for a change in a specific direction ( or ), while a two-tailed test looks for any difference (). Visually, a one-tailed test places the entire significance level in one tail, whereas a two-tailed test splits into two equal areas of at both ends of the distribution.
📐Formulae
(for Independence tests)
(for Goodness of Fit or One-sample t-test)
(One-sample t-test statistic)
(Unbiased sample standard deviation)
💡Examples
Problem 1:
A researcher wants to test if a die is fair. It is rolled times with the following frequencies: 1 (7), 2 (12), 3 (8), 4 (11), 5 (9), 6 (13). Perform a Goodness of Fit test at the significance level.
Solution:
- Hypotheses: : The die is fair (all probabilities are ). : The die is not fair.
- Expected Frequencies: Since the total , for a fair die, each face should appear times.
- Calculate :
- Degrees of Freedom: .
- P-value/Critical Value: Using a GDC or table, for and , the -value is .
- Conclusion: Since (or critical value), we fail to reject . There is no significant evidence that the die is unfair.
Explanation:
This is a Goodness of Fit test because we are comparing observed data against a theoretical distribution (uniform distribution). We calculate the 'squared difference' for each outcome, scale it by the expected frequency, and sum them up.
Problem 2:
A study compares the exam scores of two independent groups. Group A () has a mean of and . Group B () has a mean of and . Test if Group A performed significantly better than Group B at a level (assume equal variances).
Solution:
- Hypotheses: , (One-tailed test).
- Parameters: Group A: . Group B: .
- Test Statistic: Enter data into GDC for a 2-sample t-test.
- Results: The GDC calculates and .
- Conclusion: Since , we fail to reject at the significance level.
- Interpretation: While Group A had a higher average, the difference is not statistically significant at the level.
Explanation:
This is a two-sample t-test for independent means. Because we are testing if one group is specifically 'better' than the other, it is a one-tailed test. The p-value indicates that there is a chance this difference occurred by random variation.