Posterior Predictive Distribution
Posterior Predictive
Unlike classical learning algorithm, Bayesian algorithms do not attempt to identify “best-fit” models of the data (or similarly, make “best guess” predictions for new test inputs). Instead, they compute a posterior distribution over models (or similarly, compute posterior predictive distributions for new test inputs). These distributions provide a useful way to quantify our uncertainty in model estimates, and to exploit our knowledge of this uncertainty in order to make more robust predictions on new test points. Mathematically it is written as below using marginalization:
where is new prediction and is the datapoints sampled from distribution with unknown parameters . And we estimate the posterior using bayesian inference as:
Last updated