Rebel Bayes Day 3

Prior beliefs about Bayesian statistics, updated by reading Statistical Rethinking by Richard McElreath.

Duncan Garmonsway
February 20, 2019

Reading week

This week I am reading Statistical Rethinking by Richard McElreath. Each day I post my prior beliefs about Bayesian Statistics, read a bit, and update them. See also Day 1, Day 2, Day 4 and Day 5.

Prior beliefs

  1. The interaction term refers to our conviction that what we are concerned with here is the fundamental interconnectedness of all things.
  2. Just because I sympathise with you doesn’t mean you sympathise with me.
  3. If you include an interaction the you must (must not?) include the individual terms as well. R hedges its bets with syntax for either * or :.
  4. STAN is the man.
  5. Markov chains are gone to the office now that’s a good idea to have a great day and I will be in the morning …
  6. Wardrobes with high entropy suffer less from overfitting.
  7. Generalized linear models wouldn’t be such a big deal if everyone teaching undergrad stats had heeded Nelder’s and Wedderburn’s advice in the original paper to abandon t-tests, ANOVA, etc.

    We hope that the approach developed in this paper will prove to be a useful way of unifying what are often presented as unrelated statistical procedures, and that this unification will simplify the teaching of the subject to both specialists and non-specialists.

New data

7. Interactions

7.1 Building an interaction

7.2 Symmetry of the linear interaction

7.3 Continuous interactions

7.4 Interactions in design formulas

8. Markov Chain Monte Carlo

8.1 Good King Markov and His island kingdom

8.2 Markov chain Monte Carlo

8.3 Easy HMC: map2stan

8.4 Care and feeding of your Markov chain

9. Big Entropy and the Generalized Linear Model

9.1 Maximum entropy

9.2 Generalized linear models

9.3 Maximum entropy priors

Updated beliefs

  1. ✓ The interaction term refers to our conviction that what we are concerned with here is the fundamental interconnectedness of all things.
  2. ✓ Just because I sympathise with you doesn’t mean you sympathise with me.
  3. ✕ If you include an interaction the you must (must not?) include the individual terms as well. R hedges its bets with syntax for either * or :. It can be reasonable to omit the main effects, but each one can only be interpreted when the others are zero.
  4. ✕ STAN is the man. Stan was a man.
  5. ✕ Markov chains are gone to the office now that’s a good idea to have a great day and I will be in the morning … There’s more to Markov chains than predictive text. For example, fusion bombs.
  6. ✓ Wardrobes with high entropy suffer less from overfitting. They’re maximally conservative, given the constraints.
  7. ✓ Generalized linear models wouldn’t be such a big deal if everyone teaching undergrad stats had heeded Nelder’s and Wedderburn’s advice in the original paper to abandon t-tests, ANOVA, etc. They’re not a big deal in this book, which heeded the advice.

Critic’s choice

‘Pathological examples’ of things going wrong with MCMC, in Chapter 8. The intuition of Gaussian and Binomial distributions maximising entropy given constraints in Chapter 9.

Today’s chapters tended to address topics I didn’t expect and hadn’t stated prior beliefs about. For example, it continued a strong case for plots rather than tables, and explored the relative merits of Gibbs and Hamiltonion Monte Carlo.

I remain glum about inference. There have now been several mentions of the fact that enough data will wash out the priors. Today Gelman’s folk theorem of statistical computing was quoted – if modelling is hard you’re doing it wrong. I’d go further and say that if you’re modelling at all then the data isn’t convincing.

For example, there was a statistical hoo-ha a while back about whether the rate of death on New Zealand roads was increasing. Well respected statisticians did their stuff, but whatever they found couldn’t have helped make any important decisions. Better questions to ask are whether the present rate is tolerable, and whether the cost of a change of rate in either direction can be borne. Those are are largely matters of policy and economics. There’s so much data about roads and the economy that I don’t believe modelling would be necessary to make convincing arguments.

Another example, dwelt on in the book, investigated countries’ GDP and terrain. Country-level analysis will never work, because there are only a small number of countries, and they are so various – much more so than, say humans, and we know how hard it is to detect person-level effects.

Corrections

If you see mistakes or want to suggest changes, please create an issue on the source repository.

Reuse

Text and figures are licensed under Creative Commons Attribution CC BY 4.0. Source code is available at https://github.com/nacnudus/duncangarmonsway, unless otherwise noted. The figures that have been reused from other sources don't fall under this license and can be recognized by a note in their caption: "Figure from ...".

Citation

For attribution, please cite this work as

Garmonsway (2019, Feb. 20). Duncan Garmonsway: Rebel Bayes Day 3. Retrieved from https://nacnudus.github.io/duncangarmonsway/posts/2019-02-20-rebel-bayes-day-3/

BibTeX citation

@misc{garmonsway2019rebel,
  author = {Garmonsway, Duncan},
  title = {Duncan Garmonsway: Rebel Bayes Day 3},
  url = {https://nacnudus.github.io/duncangarmonsway/posts/2019-02-20-rebel-bayes-day-3/},
  year = {2019}
}