On Monday, we talked about the Beta-Binomial model for binary outcomes with an unknown probability of success, \pi.
We will now discuss sequentality in Bayesian analyses.
Working example:
Let \pi, a random value between 0 and 1, denote the unknown proportion of recent movies that pass the Bechdel test.
Three friends (feminist, clueless, and optimist) have some prior ideas about \pi.
\begin{align*} Y|\pi &\sim \text{Bin}(n, \pi) \\ \pi &\sim \text{Beta}(\alpha, \beta) \end{align*}
\begin{align*} Y|\pi &\sim \text{Bin}(n, \pi) \\ \pi &\sim \text{Beta}(\alpha, \beta) \end{align*}
\pi | (Y=y) \sim \text{Beta}(\alpha+y, \beta+n-y)
The differing prior means show disagreement about whether \pi is closer to 0 or 1.
The differing levels of prior variability show that the analysts have different degrees of certainty in their prior information.
Questions to think about:
| Analyst | Prior | Posterior |
|---|---|---|
| the feminist | Beta(5, 11) | Beta(14, 22) |
| the clueless | Beta(1, 1) | Beta(10, 12) |
| the optimist | Beta(14, 1) | Beta(23, 12) |
In addition to priors affecting our posterior distributions… the data also affects it.
Let’s now consider three new analysts: they all share the optimistic Beta(14, 1) for \pi, however, they have access to different data.
How will the different data affect the posterior distributions?
Which posterior is the most in sync with their data?
Which posterior is the least in sync with their data?
Find the posterior distributions. (i.e., What are the updated parameters?)
|
|
|
|
|---|---|---|
| Morteza | Y=6 of n=13 | Beta(20, 8) |
| Nadide | Y=29 of n=63 | Beta(45, 35) |
| Ursula | Y=46 of n=99 | Beta(60, 54) |
Let’s now turn our thinking to - okay, we’ve updated our beliefs… but now we have new data!
The evolution in our posterior understanding happens incrementally, as we accumulate new data.
Let’s revisit Milgram’s behavioral study of obedience from Chapter 3. Recall, \pi represents the proportion of people that will obey authority, even if it means bringing harm to others.
Prior to Milgram’s experiments, our fictional psychologist expected that few people would obey authority in the face of harming another: \pi \sim \text{Beta}(1,10).
Now, suppose that the psychologist collected the data incrementally, day by day, over a three-day period.
Find the following posterior distributions, each building off the last:
Find the following posterior distributions, each building off the last:
Recall from Chapter 3, our posterior was \text{Beta}(27,24)!
In Mario Kart 8 Deluxe, item boxes give different items depending on race position. To reduce the “position bias,” only item boxes opened while the racer was in mid-pack (positions 4–10) were recorded.
You want to estimate the probability that an item box yields a Red Shell. When playing the Special Cup, only 31 red shells were seen in 114 boxes opened by mid-pack racers.
Find the posterior distribution under two priors:
How different are the posterior distributions?
Suppose that we know the individual data points for the Special Cup:
Prove sequentiality to yourself.
Analyze the data race-by-race in this order: Cloudtop Cruise, Bone-Dry Dunes, Bowser’s Castle, and Rainbow Road.
Analyze the data race-by-race in this order: Rainbow Road, Bowser’s Castle, Cloudtop Cruise, and Bone-Dry Dunes.
Today we have discussed balance and sequentiality.
Remember that order of data inclusion does not matter – we will end up with the same posterior.
We have seen that prior specification “matters” but there will not be a large difference in the posterior distribution when priors are more similar to one another.
Next week:
From the Bayes Rules! textbook:
STA6349 - Applied Bayesian Analysis - Fall 2025