When do priors matter? When don’t they matter?

6.2. When do priors matter? When don’t they matter?#

Question

What happens when enough data is collected?

Follow-ups:

  • Which prior(s) get to the correct conclusion fastest for \(p_h = 0.4, 0.9, 0.5\)? Can you explain your observations?

  • Does it matter if you update after every toss or all at once?

  • Why does the “anti-prior” work well even though its dominant assumptions (most likely \(p_h = 0\) or \(1\)) are proven wrong early on?

Different priors eventually give the same posterior with enough data. This is called Bayesian convergence. How many tosses constitute ``eventually”? Clearly it depends on \(p_h\) and how close you want the posteriors to be. How about for \(p_h = 0.4\) or \(p_h = 0.9\)?

Choosing priors: a good reference is the Stan page on Prior Choice Recommendations.