IB Math AA: Probability Distributions
AI-Generated Content
IB Math AA: Probability Distributions
Probability distributions are the mathematical engines that power predictions, from forecasting election results to modeling genetic inheritance. In IB Math Analysis and Approaches, mastering these distributions moves you from counting simple outcomes to quantifying real-world uncertainty with precision. This knowledge is not just exam-critical; it forms the foundational language for data science, economics, and experimental research.
Discrete Foundations: The Binomial Distribution
Many real-world scenarios involve a fixed number of independent trials, each with the same two possible outcomes: success or failure. The binomial distribution is the discrete probability model for exactly these situations. It is defined by two parameters: , the fixed number of trials, and , the constant probability of success on a single trial.
A random variable following a binomial distribution is written as . The probability of achieving exactly successes is given by the formula: where is the binomial coefficient, calculating the number of ways to choose successes from trials.
The mean, or expected value, and the variance of a binomial distribution are derived from its parameters: The standard deviation is simply the square root of the variance: . For example, if a fair die is rolled 60 times (), the expected number of fours is , with a standard deviation of .
Continuous Modeling: The Normal Distribution
While the binomial distribution handles discrete counts, many natural phenomena—like heights, exam scores, or measurement errors—are continuous and often cluster around a central value. The normal distribution models these continuous data with its iconic bell-shaped curve, symmetric about its mean.
A normal distribution is defined by its mean (the center) and its standard deviation (the spread), denoted . The total area under its probability density function is 1. To find probabilities, you must standardise the variable. This transforms any normal distribution into the standard normal distribution using the -score formula: The -score tells you how many standard deviations an observation is from the mean. You then use your GDC or -tables to find probabilities like .
Often, you need to work backwards from a probability to find a corresponding data value, which is an inverse normal calculation. For instance, if you know , you first find the -score where (approximately 1.645), and then "un-standardise" using .
A crucial application is using the normal distribution to approximate a binomial distribution when is large. This requires a continuity correction because you are approximating a discrete distribution (binomial) with a continuous one (normal). If you want for a binomial variable, you approximate it as .
HL Extension: The Poisson Distribution and Hypothesis Testing
At Higher Level, you encounter the Poisson distribution, which models the number of events occurring in a fixed interval of time or space. These events must occur independently and at a known constant average rate, . Examples include calls received by a call center per hour or typos per page. If , then the probability of events is: Its expected value and variance are both equal to : .
These distributions become powerful tools for formal hypothesis testing. This statistical method allows you to make inferences about a population parameter based on sample data. A typical test for a population mean using the normal distribution involves five steps:
- State the null hypothesis () and the alternative hypothesis ().
- Choose the significance level (commonly 5%).
- Calculate the test statistic (e.g., a -score from sample data).
- Determine the -value—the probability of obtaining results at least as extreme as the sample, assuming is true.
- Compare the -value to . If , you reject the null hypothesis.
The -value is the cornerstone of this decision. A small -value provides evidence against the null hypothesis, suggesting your sample result is unlikely to have occurred by chance alone.
Common Pitfalls
- Misidentifying the Distribution: Assuming data is binomial when trials are not independent (like drawing cards without replacement) is a frequent error. Similarly, applying Poisson to events that are not independent or whose rate is not constant will lead to incorrect probabilities. Always check the conditions before selecting your model.
- Misusing the Continuity Correction: When using a normal distribution to approximate a binomial, forgetting the continuity correction leads to significant inaccuracy. Remember: for , use ; for , use ; for , use to .
- Confusing -value and Significance Level: The significance level is a pre-set threshold for rejection (e.g., 0.05). The -value is a calculated probability from your sample data. You do not "accept" the null hypothesis; you either reject it or fail to reject it based on the comparison.
- Incorrect Inverse Normal Setup: When performing an inverse normal calculation, students often mix up which probability to input. If a question asks for the top 10%, you need to use or, equivalently, when looking up the critical -value on their GDC.
Summary
- The binomial distribution models discrete counts of success in independent trials, with and .
- The normal distribution models continuous data; standardisation via is essential for finding probabilities and performing inverse calculations.
- Hypothesis testing is a structured process using a test statistic to generate a -value, which is compared to a significance level to decide whether to reject a null hypothesis.
- HL Only: The Poisson distribution models the count of independent events occurring at a constant rate, where .
- Always verify the conditions for each distribution (independence, fixed and , constant ) and remember the continuity correction when approximating binomial with normal.