1 Chapter 1 Introduction and examples
1.1 What is Beseyian methods?
- Bayes’s rule provides a rational method for updating beliefs in light of new information.
1.2 What can Bayesian methods provide?
- Parameter estimates with good statistical properties
- Parsimonious descriptions of observed data
1.3 Contrast between Frequentist and Bayesian Statistics
- Frequentist statistics
- Uncertainty about the parameter estimates
- Bayesian statistics
- Unvertainty is quantified by the oberservation of data.
1.4 Bayesian learning
- Parameter — \(\theta\)
- numerical values of population characteristics
- Dataset — \(y\)
- After a dataset \(y\) is obtained, the information it contains can be used to decrease our uncertainty about the population characteristics.
- Bayesian inference
- Quantifying this change in uncertainty is the purpose of Bayesian inference
- Sample space — \(\mathcal{Y}\)
- The set of all possible datasets.
- Single dataset \(y\)
- Parameter space — \(\Theta\)
- possbile parameter values
- we hope to identify the value that best represents the true population characteristics.
- Bayesian learning begins with joint beliefs about \(y\) and \(\theta\), in terms of distribution over \(\mathcal{Y}\) and \(\Theta\)
- Prior distribution — \(p(\theta)\)
- Our belief that \(\theta\) represents that true population characteristics.
- Sampling model — \(p(\mathcal{y}|\theta)\)
- describes our belief that \(\mathcal{y}\) would be the outcome of our study if we knew \(\theta\) to be true.
- Posterior distribution — \(p(\theta|\mathcal{y})\)
- Our belief that \(\theta\) is the true value, having observed dataset \(\mathcal{y}\)
- Bayesian Update (Equation 1.1) \[p(\theta|y) = \frac{\overbrace{p(y|\theta)}^{\text{Sampling model}}\overbrace{p(\theta)}^{\text{Prior distribution}}}{\int_{\Theta}p(y|\tilde{\theta})p(\tilde{\theta})d\tilde{\theta}} \tag{1.1}\]
- Bayes’s rule tells us how to change our belief after seeing new information.
- Prior distribution — \(p(\theta)\)