You have 3 free guides left 😟
Unlock your guides
You have 3 free guides left 😟
Unlock your guides

19.2 Prior and posterior distributions

2 min readjuly 19, 2024

updates our beliefs about parameters using data. It combines prior knowledge with observed evidence to form a , allowing us to make informed decisions in uncertain situations.

The process involves selecting appropriate priors, calculating posteriors using , and analyzing the impact of priors. As more data is gathered, the posterior converges towards the , regardless of the initial prior.

Bayesian Inference

Prior vs posterior distributions

Top images from around the web for Prior vs posterior distributions
Top images from around the web for Prior vs posterior distributions
  • represents initial beliefs or knowledge about a parameter before observing data
    • Denoted as P(θ)P(\theta), where θ\theta is the parameter of interest (coin bias, disease prevalence)
    • Based on domain knowledge, previous studies, or expert opinion
  • Posterior distribution represents updated beliefs or knowledge about a parameter after observing data
    • Denoted as P(θX)P(\theta|X), where XX is the observed data (coin flips, patient test results)
    • Combines prior distribution and using Bayes' theorem

Selection of prior distributions

  • Informative priors used when prior knowledge or information about the parameter is available
    • for parameters bounded between 0 and 1 (success probability)
    • for parameters with known mean and variance (average height)
  • Non-informative priors used when little or no prior knowledge about the parameter is available
    • Aim to minimize the influence of the prior on the posterior distribution
    • over the parameter space (any value equally likely)
    • , proportional to square root of Fisher information (invariant under reparameterization)

Calculation of posterior distributions

  • Bayes' theorem: P(θX)=P(Xθ)P(θ)P(X)P(\theta|X) = \frac{P(X|\theta)P(\theta)}{P(X)}
    • P(Xθ)P(X|\theta) is likelihood function, probability of observing data XX given parameter θ\theta
    • P(X)P(X) is , normalizing constant
  • Steps to calculate posterior distribution:
    1. Specify prior distribution P(θ)P(\theta)
    2. Determine likelihood function P(Xθ)P(X|\theta) based on observed data
    3. Calculate marginal likelihood P(X)P(X) by integrating or summing over all possible values of θ\theta
    4. Apply Bayes' theorem to obtain posterior distribution P(θX)P(\theta|X)

Impact of priors on posteriors

  • investigates how choice of prior distribution affects posterior distribution
    • Compare posterior distributions obtained using different priors (skeptical vs optimistic)
  • Influence of prior distribution
    • Strong prior with narrow distribution or high confidence can heavily influence posterior (expert opinion)
    • Weak prior with wide distribution or low confidence allows data to have more impact on posterior (uninformative)
    • As sample size increases, influence of prior diminishes (law of large numbers)
    • Posterior distribution converges to true parameter value, regardless of choice of prior (consistency)
© 2024 Fiveable Inc. All rights reserved.
AP® and SAT® are trademarks registered by the College Board, which is not affiliated with, and does not endorse this website.


© 2024 Fiveable Inc. All rights reserved.
AP® and SAT® are trademarks registered by the College Board, which is not affiliated with, and does not endorse this website.

© 2024 Fiveable Inc. All rights reserved.
AP® and SAT® are trademarks registered by the College Board, which is not affiliated with, and does not endorse this website.
Glossary
Glossary