M-open, M-complete, M-closed
May 30, 2016 — July 23, 2023
Placeholder. I encountered the M-open concept and thought it was useful to state, but I have not had time to really delve into it. M-open, M-complete, and M-closed describe different relations between our hypothesis class and reality — basically, do we have the true model in my hypothesis class (spoiler: no, I do not, except with synthetic data) and if not, what does our estimation procedure get us?
Fancy people write this as \(\mathcal{M}\)-open etc., but life is too short for indulgent typography.
Le and Clarke (2017) summarises:
For the sake of completeness, we recall that Bernardo and Smith (Bernardo and Smith 2000) define M-closed problems as those for which a true model can be identified and written down but is one amongst finitely many models from which an analyst has to choose. By contrast, M-complete problems are those in which a true model (sometimes called a belief model) exists but is inaccessible in the sense that even though it can be conceptualised it cannot be written down or at least cannot be used directly. Effectively this means that other surrogate models must be identified and used for inferential purposes. M-open problems according to Bernardo and Smith (2000) are those problems where a true model exists but cannot be specified at all.
They also mention Clyde and Iversen (2013) as a useful resource.
You will note that many of the references are interested in this concept because of application to Bayesian model stacking.
Related: likelihood principle, decision-theory, black swans, misspecified models, Robust bayes …
1 Gibbs posteriors
Gibbs posteriors seem to be an attempt to address the M-open problem, by removing the need for a valid likelihood.