←back to thread

Bayesian Statistics: The three cultures

(statmodeling.stat.columbia.edu)
309 points luu | 3 comments | | HN request time: 0.02s | source
Show context
tfehring ◴[] No.41081746[source]
The author is claiming that Bayesians vary along two axes: (1) whether they generally try to inform their priors with their knowledge or beliefs about the world, and (2) whether they iterate on the functional form of the model based on its goodness-of-fit and the reasonableness and utility of its outputs. He then labels 3 of the 4 resulting combinations as follows:

    ┌───────────────┬───────────┬──────────────┐
    │               │ iteration │ no iteration │
    ├───────────────┼───────────┼──────────────┤
    │ informative   │ pragmatic │ subjective   │
    │ uninformative │     -     │ objective    │
    └───────────────┴───────────┴──────────────┘
My main disagreement with this model is the empty bottom-left box - in fact, I think that's where most self-labeled Bayesians in industry fall:

- Iterating on the functional form of the model (and therefore the assumed underlying data generating process) is generally considered obviously good and necessary, in my experience.

- Priors are usually uninformative or weakly informative, partly because data is often big enough to overwhelm the prior.

The need for iteration feels so obvious to me that the entire "no iteration" column feels like a straw man. But the author, who knows far more academic statisticians than I do, explicitly says that he had the same belief and "was shocked to learn that statisticians didn’t think this way."

replies(3): >>41081867 #>>41082105 #>>41084103 #
Onavo ◴[] No.41082105[source]
Interesting, in my experience modern ML runs almost entirely on pragmatic Bayes. You find your ELBO, you choose the latest latent variable du jour that best models your problem domain (these days it's all transformers), and then you start running experiments.
replies(1): >>41082455 #
1. tfehring ◴[] No.41082455[source]
I think each category of Bayesian described in the article generally falls under Breiman's [0] "data modeling" culture, while ML practitioners, even when using Bayesian methods, almost invariably fall under the "algorithmic modeling" culture. In particular, the article's definition of pragmatic Bayes says that "the model should be consistent with knowledge about the underlying scientific problem and the data collection process," which I don't consider the norm in ML at all.

I do think ML practitioners in general align with the "iteration" category in my characterization, though you could joke that that miscategorizes people who just use (boosted trees|transformers) for everything.

[0] https://projecteuclid.org/journals/statistical-science/volum...

replies(1): >>41083670 #
2. nextos ◴[] No.41083670[source]
> the model should be consistent with knowledge about the problem [...] which I don't consider the norm in ML at all.

I don't think that is so niche. Murphy's vol II, a mainstream book, starts with this quote:

"Intelligence is not just about pattern recognition and function approximation. It’s about modeling the world." — Josh Tenenbaum, NeurIPS 2021.

Goodman & Tenenbaum have written e.g. https://probmods.org, which is very much about modeling data-generating processes.

The same can be said about large parts of Murphy's book, Lee & Wagenmakers or Lunn et al. (the BUGS book).

replies(1): >>41086537 #
3. ttyprintk ◴[] No.41086537[source]
Archive for Goodman & Tenenbaum, since their site is flaky:

https://archive.ph/WKLyM