This is an extremely brief “review” of the limited probability theory we utilize in Chapter 5. It’s not even particularly fair to call this a review, since probability is not a prerequisite to this course. However, the limited amount we use is fairly straightforward and intuitive.
If a more thorough treatment is needed, then depending on your goals, there are good options available. For someone looking to explore some elementary probability theory, the introductory statistics textbook “OpenIntro Statistics” by David Diez, Christopher Barr, and Mine Çetinkaya-Rundel does a good job presenting this material. It also is an excellent introductory statistics text with labs and data available. For a calculus-based, theory heavy treatment of this text, I recommend “Probability: Lecture and Labs”.
In probability, an experiment is an occurrence with a measurable result. Each instance of an experiment is a trial. The possible results of each trial are called outcomes. The set of all possible outcomes for an experiment is the sample space for that experiment.
But what does it mean for an event \(A\) to have probability \(P(A)\text{?}\) It means that if I repeat the experiment over and over, the proportion of them where \(A\) is true should be \(P(A)\text{.}\)
So if I roll a die over and over, the proportion of them that give me a 6 over time should be \(\frac{1}{6}\text{.}\) So if we roll a dice 10000 times, we would expect one sixth of them to come up heads:
A probability distribution is, roughly speaking, a complete description of a random variable and the likelihood of each output. In the case of random variables with a finite number of possible outputs a probability distribution table is a convenient way of presenting this information.
Snow White has a basket of 10 apples, 3 are poisoned. She is going to pick 4 apples at random to eat for some reason. Let \(X\) denote the number of poisoned apples she eats.
Note that the “Expected Value” may not be a value we actually expect, that is, may not be one of the outcomes, just an average outcome. We think of this as the outcomes of \(X\text{,}\) weighted by their likelihood, so the more likely outcomes contribute more than the less likely ones.