- We make use of Statistical Estimators in Frequentist statistics.
Frequentist Decision Theory
-
We choose any estimator or decision parameter that we want.
-
The risk of the estimator is defined as
-
The risk is not computable since it relies on
. -
The Bayes risk of an estimator is defined as
- A Bayes decision rule minimizes the expected Bayes risk.
- Murphy 6.3.1. A Bayes estimator can be obtained by minimizing the posterior expected loss for each
. - Thus, picking the optimal action on a case-by-case basis is optimal on average.
- Murphy 6.3.2. Every admissible decision rule is a Bayes decision rule with respect to some possibly improper, prior distribution.
- Thus, the best way to minimize frequentist risk is to be Bayesian
-
The maximum risk of an estimator is defined as
-
The minimax rule is one that minimizes the maximum risk
-
The minimax estimator tends to be pessimistic and difficult to compute.
-
They are equivalent to Bayes estimators under a least favorable prior.
-
-
Given two estimators
and , if , then
dominates (strictly so if we replace with ) -
An estimator is admissible if it is not strictly dominated by any other estimator.
- In practice, the sample median is often better than the sample mean because it is more robust to outliers.
- It is easy to construct an admissible estimator (see Murphy Thm. 6.3.3. Hence admissibility is not sufficient for good estimators.
Pathologies
- Confidence intervals mean we condition on unknown results, rather than on known results.
- Hypothesis testing means we favor rejecting the null rather than proving the null hypothesis.
- The
-value of frequentist statistics is dependent on when we decide to stop collecting samples. - Frequentism violates the likelihood principle, which says inference should be based on the likelihood of what is observed rather than hypothetical future data. The likelihood principle follows from two principles
- The sufficiency principle - a sufficient statistic should contain all the relevant information about an unknown parameter.
- Weak conditionality - inferences should be based on what happened rather than what might not have happened.
Topics
- Statistical Estimators
- Model Performance - specifically refer to empirical risk minimization