|
| 1 | +--- |
| 2 | +layout: proof |
| 3 | +mathjax: true |
| 4 | + |
| 5 | +author: "Thomas J. Faulkenberry" |
| 6 | +affiliation: "Tarleton State University" |
| 7 | +e_mail: "faulkenberry@tarleton.edu" |
| 8 | +date: 2020-08-26 12:00:00 |
| 9 | + |
| 10 | +title: "Encompassing Prior Method for computing Bayes Factors" |
| 11 | +chapter: "Model Selection" |
| 12 | +section: "Bayesian model selection" |
| 13 | +topic: "Bayes factor" |
| 14 | +theorem: "Computation using Encompassing Prior Method" |
| 15 | + |
| 16 | +sources: |
| 17 | + - authors: "Klugkist, I., Kato, B., and Hoijtink, H." |
| 18 | + year: 2005 |
| 19 | + title: "Bayesian model selection using encompassing priors" |
| 20 | + in: "Statistica Neerlandica" |
| 21 | + pages: "vol. 59, no. 1., pp. 57-69" |
| 22 | + url: "https://dx.doi.org/10.1111/j.1467-9574.2005.00279.x" |
| 23 | + doi: "10.1111/j.1467-9574.2005.00279.x" |
| 24 | + |
| 25 | + - authors: "Faulkenberry, Thomas J." |
| 26 | + year: 2019 |
| 27 | + title: "A tutorial on generalizing the default Bayesian t-test via posterior sampling and encompassing priors" |
| 28 | + in: "Communications for Statistical Applications and Methods" |
| 29 | + pages: "vol. 26, no. 2, pp. 217-238" |
| 30 | + url: "https://dx.doi.org/10.29220/CSAM.2019.26.2.217" |
| 31 | + doi: "10.29220/CSAM.2019.26.2.217" |
| 32 | + |
| 33 | +proof_id: "P157" |
| 34 | +shortcut: "bf-ep" |
| 35 | +username: "tomfaulkenberry" |
| 36 | +--- |
| 37 | + |
| 38 | + |
| 39 | +**Theorem:** Consider two models $m_1$ and $m_e$, where $m_1$ is nested within an [encompassing model](/D/em) $m_e$ via an inequality constraint on some parameter $\theta$, and $\theta$ is unconstrained under $m_e$. Then |
| 40 | +\[ |
| 41 | + B_{1e} = \frac{c}{d} = \frac{1/d}{1/c} |
| 42 | +\] |
| 43 | +where $1/d$ and $1/c$ represent the proportions of the posterior and prior of the encompassing model, respectively, that are in agreement with the inequality constraint imposed by the nested model $m_1$. |
| 44 | + |
| 45 | +**Proof:** |
| 46 | +Consider first that for any model $m_1$ on data $y$ with parameter $\theta$, [Bayes theorem](/P/bayes-th) implies |
| 47 | + |
| 48 | +$$ \label{eq:bayesth} |
| 49 | + p(\theta \mid y,m_1) = \frac{p(y \mid \theta,m_1) \cdot p(\theta \mid m_1)}{p(y \mid m_1)}. |
| 50 | +$$ |
| 51 | + |
| 52 | +Rearranging Equation \eqref{eq:bayesth} allows us to write the [marginal likelihood](/D/ml) for $y$ under $m_1$ as |
| 53 | + |
| 54 | +$$ \label{eq:marginal} |
| 55 | + p(y \mid m_1) = \frac{p(y \mid \theta,m_1) \cdot p(\theta \mid m_1)}{p(\theta \mid y,m_1)}. |
| 56 | +$$ |
| 57 | + |
| 58 | +Taking the ratio of the marginal likelihoods for $m_1$ and the [encompassing model](/D/em) $m_e$ yields the following [Bayes factor](/D/bf): |
| 59 | + |
| 60 | +$$ \label{eq:bayesfactor} |
| 61 | + B_{1e} = \frac{p(y \mid \theta,m_1) \cdot p(\theta \mid m_1) / p(\theta \mid y,m_1)}{p(y \mid \theta,m_e) \cdot p(\theta \mid m_e) / p(\theta \mid y,m_e)}. |
| 62 | +$$ |
| 63 | + |
| 64 | +Now, both the constrained model $m_1$ and the [encompassing model](/D/em) $m_e$ contain the same parameter vector $\theta$. Choose a specific value of $\theta$, say $\theta'$, that exists in the support of both models $m_1$ and $m_e$ (we can do this because $m_1$ is nested within $m_e$). Then, for this parameter value $\theta'$, we have $p(y \mid \theta',m_1)=p(y \mid \theta',m_e)$, so the expression for the Bayes factor (Equation \eqref{eq:bayesfactor} above) reduces to an expression involving only the priors and posteriors for $\theta'$ under $m_1$ and $m_e$: |
| 65 | + |
| 66 | +$$ \label{eq:bayesfactor2} |
| 67 | + B_{1e} = \frac{p(\theta' \mid m_1) / p(\theta' \mid y,m_1)}{p(\theta' \mid m_e) / p(\theta' \mid y,m_e)}. |
| 68 | +$$ |
| 69 | + |
| 70 | +Because $m_1$ is nested within $m_e$ via an inequality constraint, the prior $p(\theta' \mid m_1)$ is simply a truncation of the encompassing prior $p(\theta' \mid m_e)$. Thus, we can express $p(\theta' \mid m_1)$ in terms of the encompassing prior $p(\theta' \mid m_e)$ by multiplying the encompassing prior by an indicator function over $m_1$ and then normalizing the resulting product. That is, |
| 71 | + |
| 72 | +$$ \label{eq:normalize} |
| 73 | +\begin{split} |
| 74 | + p(\theta' \mid m_1) & = \frac{p(\theta' \mid m_e) \cdot I_{\theta' \in m_1}}{\int p(\theta' \mid m_e) \cdot I_{\theta' \in m_1}d\theta'}\\ |
| 75 | + & = \Biggl(\frac{I_{\theta' \in m_1}}{\int p(\theta' \mid m_e) \cdot I_{\theta' \in m_1}d\theta'}\Biggr) \cdot p(\theta' \mid m_e), |
| 76 | +\end{split} |
| 77 | +$$ |
| 78 | + |
| 79 | +where $I_{\theta' \in m_1}$ is an indicator function. For parameters $\theta' \in m_1$, this indicator function is identically equal to 1, so the expression in parentheses reduces to a constant, say $c$, allowing us to write the prior as |
| 80 | + |
| 81 | +$$ \label{eq:prior} |
| 82 | + p(\theta' \mid m_1) = c \cdot p(\theta' \mid m_e). |
| 83 | +$$ |
| 84 | + |
| 85 | +By similar reasoning, we can write the posterior as |
| 86 | + |
| 87 | +$$ \label{eq:posterior} |
| 88 | + p(\theta' \mid y,m_1) = \Biggl(\frac{I_{\theta' \in m_1}}{\int p(\theta' \mid y,m_e)I_{\theta' \in m_1}d\theta'}\Biggr)\cdot p(\theta' \mid y,m_e) = d \cdot p(\theta' \mid y,m_e). |
| 89 | +$$ |
| 90 | + |
| 91 | +This gives us |
| 92 | + |
| 93 | +$$ \label{eq:bayesfactor3} |
| 94 | + B_{1e} = \frac{c \cdot p(\theta' \mid m_e) / d \cdot p(\theta' \mid y,m_e)}{p(\theta' \mid m_e) / p(\theta' \mid y,m_e)} = \frac{c}{d} = \frac{1/d}{1/c}, |
| 95 | +$$ |
| 96 | + |
| 97 | +which completes the proof. Note that by definition, $1/d$ represents the proportion of the posterior distribution for $\theta$ under the [encompassing model](/D/em) $m_e$ that agrees with the constraints imposed by $m_1$. Similarly, $1/c$ represents the proportion of the prior distribution for $\theta$ under the [encompassing model](/D/em) $m_e$ that agrees with the constraints imposed by $m_1$. |
0 commit comments