Statistical inference is the process of learning about unknown quantities from observed data.
This has nothing to do with inference (ML) ("inference-time"/"test-time")
The only link is the notion of “inferring something”.
Statistical inference is about learning/uncertainty, ML inference is about making predictions with a trained model.
The Inference Problem
We observe data generated by some process with unknown parameters . What can we say about given ?
This involves two directions of reasoning:
- Forward: - If we knew the parameters, what data would we expect?
- Inverse: - Given the data we observed, what parameters are plausible?
We typically know how to compute (the likelihood) from our model, but we want (the posterior).
Approaches to Inference
bayesian inference - Uses Bayes theorem to update beliefs.
Frequentist Inference - Treats parameters as fixed but unknown. Uses procedures that have good properties when repeated.
Bayesians quantify uncertainty about parameters directly via , while frequentists quantify uncertainty about their procedures.
variational inference - Approximates complex posteriors with simpler distributions.