Skip to content

Commit 827981d

Browse files
authored
added 8 definitions
1 parent 22770ee commit 827981d

File tree

8 files changed

+294
-0
lines changed

8 files changed

+294
-0
lines changed

D/prior-conj.md

Lines changed: 34 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,34 @@
1+
---
2+
layout: definition
3+
mathjax: true
4+
5+
author: "Joram Soch"
6+
affiliation: "BCCN Berlin"
7+
e_mail: "joram.soch@bccn-berlin.de"
8+
date: 2020-12-02 17:55:00
9+
10+
title: "Conjugate and non-conjugate prior distribution"
11+
chapter: "General Theorems"
12+
section: "Bayesian statistics"
13+
topic: "Prior distributions"
14+
definition: "Conjugate vs. non-conjugate"
15+
16+
sources:
17+
- authors: "Wikipedia"
18+
year: 2020
19+
title: "Conjugate prior"
20+
in: "Wikipedia, the free encyclopedia"
21+
pages: "retrieved on 2020-12-02"
22+
url: "https://en.wikipedia.org/wiki/Conjugate_prior"
23+
24+
def_id: "D120"
25+
shortcut: "prior-conj"
26+
username: "JoramSoch"
27+
---
28+
29+
30+
**Definition:** Let $m$ be a [generative model](/D/gm) with [likelihood function](/D/lf) $p(y \vert \theta, m)$ and [prior distribution](/D/prior) $p(\theta \vert m)$. Then,
31+
32+
* the [prior distribution](/D/prior) is called "conjugate", if it, when combined with the [likelihood function](/D/lf), leads to a [posterior distribution](/D/post) that belongs to the same family of [probability distributions](/D/dist);
33+
34+
* the prior distribution is called "non-conjugate", if this is not the case.

D/prior-eb.md

Lines changed: 34 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,34 @@
1+
---
2+
layout: definition
3+
mathjax: true
4+
5+
author: "Joram Soch"
6+
affiliation: "BCCN Berlin"
7+
e_mail: "joram.soch@bccn-berlin.de"
8+
date: 2020-12-02 18:19:00
9+
10+
title: "Empirical Bayes prior distribution"
11+
chapter: "General Theorems"
12+
section: "Bayesian statistics"
13+
topic: "Prior distributions"
14+
definition: "Empirical Bayes priors"
15+
16+
sources:
17+
- authors: "Wikipedia"
18+
year: 2020
19+
title: "Empirical Bayes method"
20+
in: "Wikipedia, the free encyclopedia"
21+
pages: "retrieved on 2020-12-02"
22+
url: "https://en.wikipedia.org/wiki/Empirical_Bayes_method#Introduction"
23+
24+
def_id: "D122"
25+
shortcut: "prior-eb"
26+
username: "JoramSoch"
27+
---
28+
29+
30+
**Definition:** Let $m$ be a [generative model](/D/gm) with [likelihood function](/D/lf) $p(y \vert \theta, m)$ and [prior distribution](/D/prior) $p(\theta \vert \lambda, m)$ using [prior hyperparameters](/D/prior) $\lambda$. Let $p(y \vert \lambda, m)$ be the [marginal likelihood](/D/ml) when [integrating the parameters out of the joint likelihood](/P/ml-jl). Then, the prior distribution is called an "Empirical Bayes prior", if it maximizes the logarithmized marginal likelihood:
31+
32+
$$ \label{eq:prior-eb}
33+
\lambda_{\mathrm{EB}} = \operatorname*{arg\,max}_{\lambda} \log p(y \vert \lambda, m) \; .
34+
$$

D/prior-emp.md

Lines changed: 35 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,35 @@
1+
---
2+
layout: definition
3+
mathjax: true
4+
5+
author: "Joram Soch"
6+
affiliation: "BCCN Berlin"
7+
e_mail: "joram.soch@bccn-berlin.de"
8+
date: 2020-12-02 17:37:00
9+
10+
title: "Empirical and theoretical prior distribution"
11+
chapter: "General Theorems"
12+
section: "Bayesian statistics"
13+
topic: "Prior distributions"
14+
definition: "Empirical vs. non-empirical"
15+
16+
sources:
17+
- authors: "Soch J, Allefeld C, Haynes JD"
18+
year: 2016
19+
title: "How to avoid mismodelling in GLM-based fMRI data analysis: cross-validated Bayesian model selection"
20+
in: "NeuroImage"
21+
pages: "vol. 141, pp. 469-489, eq. 13, p. 473"
22+
url: "https://www.sciencedirect.com/science/article/pii/S1053811916303615"
23+
doi: "10.1016/j.neuroimage.2016.07.047"
24+
25+
def_id: "D119"
26+
shortcut: "prior-emp"
27+
username: "JoramSoch"
28+
---
29+
30+
31+
**Definition:** Let $p(\theta \vert m)$ a [prior distribution](/D/prior) for the parameter $\theta$ of a [generative model](/D/gm) $m$. Then,
32+
33+
* the distribution is called an "empirical prior", if it has been [derived from empirical data](/P/post-jl);
34+
35+
* the distribution is called a "theoretical prior", if it was specified without regard to empirical data.

D/prior-flat.md

Lines changed: 44 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,44 @@
1+
---
2+
layout: definition
3+
mathjax: true
4+
5+
author: "Joram Soch"
6+
affiliation: "BCCN Berlin"
7+
e_mail: "joram.soch@bccn-berlin.de"
8+
date: 2020-12-02 17:04:00
9+
10+
title: "Flat, hard and soft prior distribution"
11+
chapter: "General Theorems"
12+
section: "Bayesian statistics"
13+
topic: "Prior distributions"
14+
definition: "Flat vs. hard vs. soft"
15+
16+
sources:
17+
- authors: "Friston et al."
18+
year: 2002
19+
title: "Classical and Bayesian Inference in Neuroimaging: Theory"
20+
in: "NeuroImage"
21+
pages: "vol. 16, iss. 2, pp. 465-483, fn. 1"
22+
url: "https://www.sciencedirect.com/science/article/pii/S1053811902910906"
23+
doi: "10.1006/nimg.2002.1090"
24+
- authors: "Friston et al."
25+
year: 2002
26+
title: "Classical and Bayesian Inference in Neuroimaging: Applications"
27+
in: "NeuroImage"
28+
pages: "vol. 16, iss. 2, pp. 484-512, fn. 10"
29+
url: "https://www.sciencedirect.com/science/article/pii/S1053811902910918"
30+
doi: "10.1006/nimg.2002.1091"^
31+
32+
def_id: "D116"
33+
shortcut: "prior-flat"
34+
username: "JoramSoch"
35+
---
36+
37+
38+
**Definition:** Let $p(\theta \vert m)$ a [prior distribution](/D/prior) for the parameter $\theta$ of a [generative model](/D/gm) $m$. Then,
39+
40+
* the distribution is called a "flat prior", if its [precision](/D/prec) is zero or [variance](/D/var) is infinite;
41+
42+
* the distribution is called a "hard prior", if its [precision](/D/prec) is infinite or [variance](/D/var) is zero;
43+
44+
* the distribution is called a "soft prior", if its [precision](/D/prec) and [variance](/D/var) are non-zero and finite.

D/prior-inf.md

Lines changed: 37 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,37 @@
1+
---
2+
layout: definition
3+
mathjax: true
4+
5+
author: "Joram Soch"
6+
affiliation: "BCCN Berlin"
7+
e_mail: "joram.soch@bccn-berlin.de"
8+
date: 2020-12-02 17:28:00
9+
10+
title: "Informative and non-informative prior distribution"
11+
chapter: "General Theorems"
12+
section: "Bayesian statistics"
13+
topic: "Prior distributions"
14+
definition: "Informative vs. non-informative"
15+
16+
sources:
17+
- authors: "Soch J, Allefeld C, Haynes JD"
18+
year: 2016
19+
title: "How to avoid mismodelling in GLM-based fMRI data analysis: cross-validated Bayesian model selection"
20+
in: "NeuroImage"
21+
pages: "vol. 141, pp. 469-489, eq. 15, p. 473"
22+
url: "https://www.sciencedirect.com/science/article/pii/S1053811916303615"
23+
doi: "10.1016/j.neuroimage.2016.07.047"
24+
25+
def_id: "D118"
26+
shortcut: "prior-inf"
27+
username: "JoramSoch"
28+
---
29+
30+
31+
**Definition:** Let $p(\theta \vert m)$ a [prior distribution](/D/prior) for the parameter $\theta$ of a [generative model](/D/gm) $m$. Then,
32+
33+
* the distribution is called an "informative prior", if it biases the parameter towards particular values;
34+
35+
* the distribution is called a "weakly informative prior", if it mildly [influences the posterior distribution](/P/post-jl);
36+
37+
* the distribution is called a "non-informative prior", if it does not influence the [posterior hyperparameters](/D/post).

D/prior-maxent.md

Lines changed: 42 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,42 @@
1+
---
2+
layout: definition
3+
mathjax: true
4+
5+
author: "Joram Soch"
6+
affiliation: "BCCN Berlin"
7+
e_mail: "joram.soch@bccn-berlin.de"
8+
date: 2020-12-02 18:13:00
9+
10+
title: "Maximum entropy prior distribution"
11+
chapter: "General Theorems"
12+
section: "Bayesian statistics"
13+
topic: "Prior distributions"
14+
definition: "Maximum entropy priors"
15+
16+
sources:
17+
- authors: "Wikipedia"
18+
year: 2020
19+
title: "Prior probability"
20+
in: "Wikipedia, the free encyclopedia"
21+
pages: "retrieved on 2020-12-02"
22+
url: "https://en.wikipedia.org/wiki/Prior_probability#Uninformative_priors"
23+
24+
def_id: "D121"
25+
shortcut: "prior-maxent"
26+
username: "JoramSoch"
27+
---
28+
29+
30+
**Definition:** Let $m$ be a [generative model](/D/gm) with [likelihood function](/D/lf) $p(y \vert \theta, m)$ and [prior distribution](/D/prior) $p(\theta \vert \lambda, m)$ using [prior hyperparameters](/D/prior) $\lambda$. Then, the prior distribution is called a "maximum entropy prior", if
31+
32+
1) when $\theta$ is a [discrete random variable](/D/rvar-disc), it maximizes the [entropy](/D/ent) of the prior [probability mass function](/D/pmf):
33+
34+
$$ \label{eq:prior-maxent-disc}
35+
\lambda_{\mathrm{maxent}} = \operatorname*{arg\,max}_{\lambda} \mathrm{H}\left[ p(\theta \vert \lambda, m) \right] \; ;
36+
$$
37+
38+
2) when $\theta$ is a [continuous random variable](/D/rvar-disc), it maximizes the [differential entropy](/D/dent) of the prior [probability density function](/D/pdf):
39+
40+
$$ \label{eq:prior-maxent-cont}
41+
\lambda_{\mathrm{maxent}} = \operatorname*{arg\,max}_{\lambda} \mathrm{h}\left[ p(\theta \vert \lambda, m) \right] \; .
42+
$$

D/prior-ref.md

Lines changed: 34 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,34 @@
1+
---
2+
layout: definition
3+
mathjax: true
4+
5+
author: "Joram Soch"
6+
affiliation: "BCCN Berlin"
7+
e_mail: "joram.soch@bccn-berlin.de"
8+
date: 2020-12-02 18:26:00
9+
10+
title: "Reference prior distribution"
11+
chapter: "General Theorems"
12+
section: "Bayesian statistics"
13+
topic: "Prior distributions"
14+
definition: "Reference priors"
15+
16+
sources:
17+
- authors: "Wikipedia"
18+
year: 2020
19+
title: "Prior probability"
20+
in: "Wikipedia, the free encyclopedia"
21+
pages: "retrieved on 2020-12-02"
22+
url: "https://en.wikipedia.org/wiki/Prior_probability#Uninformative_priors"
23+
24+
def_id: "D123"
25+
shortcut: "prior-ref"
26+
username: "JoramSoch"
27+
---
28+
29+
30+
**Definition:** Let $m$ be a [generative model](/D/gm) with [likelihood function](/D/lf) $p(y \vert \theta, m)$ and [prior distribution](/D/prior) $p(\theta \vert \lambda, m)$ using [prior hyperparameters](/D/prior) $\lambda$. Let $p(\theta \vert y, \lambda, m)$ be the [posterior distribution](/D/post) that is [proportional to the the joint likelihood](/P/post-jl). Then, the prior distribution is called a "reference prior", if it maximizes the [expected](/D/mean) [Kullback-Leibler divergence](/D/kl) of the posterior distribution relative to the prior distribution:
31+
32+
$$ \label{eq:prior-ref}
33+
\lambda_{\mathrm{ref}} = \operatorname*{arg\,max}_{\lambda} \mathrm{KL} \left[ p(\theta \vert y, \lambda, m) \, || \, p(\theta \vert \lambda, m) \right] \; .
34+
$$

D/prior-uni.md

Lines changed: 34 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,34 @@
1+
---
2+
layout: definition
3+
mathjax: true
4+
5+
author: "Joram Soch"
6+
affiliation: "BCCN Berlin"
7+
e_mail: "joram.soch@bccn-berlin.de"
8+
date: 2020-12-02 17:21:00
9+
10+
title: "Uniform and non-uniform prior distribution"
11+
chapter: "General Theorems"
12+
section: "Bayesian statistics"
13+
topic: "Prior distributions"
14+
definition: "Uniform vs. non-uniform"
15+
16+
sources:
17+
- authors: "Wikipedia"
18+
year: 2020
19+
title: "Lindley's paradox"
20+
in: "Wikipedia, the free encyclopedia"
21+
pages: "retrieved on 2020-11-25"
22+
url: "https://en.wikipedia.org/wiki/Lindley%27s_paradox#Bayesian_approach"
23+
24+
def_id: "D117"
25+
shortcut: "prior-uni"
26+
username: "JoramSoch"
27+
---
28+
29+
30+
**Definition:** Let $p(\theta \vert m)$ a [prior distribution](/D/prior) for the parameter $\theta \in \Theta$ of a [generative model](/D/gm) $m$. Then,
31+
32+
* the distribution is called a "uniform prior", if its [density](/D/pdf) is constant over the entire parameter space $\Theta$;
33+
34+
* the distribution is called a "non-uniform prior", if its [density](/D/pdf) is not constant over the parameter space $\Theta$.

0 commit comments

Comments
 (0)