On Github randy3k / GFItalk
Under some mild conditions, \[ \lim_{m\to \infty} \frac{1}{m} \sum_{i=1}^{m} L(\theta_0, \delta(X_i)) = R_\delta(\theta_0). \]
\(R_\delta(\theta_0)\) measures the long run performances of \(\delta\) for \(\theta_0\).
However, in practice, it is not quite useful.
Consider an infinite sequence of problemes: \(X_i \sim P_{\theta_i}\)
(Berger, 1985) “practicing” frequentist haves quite differently from the “formal” frequentist.
Example 1
if \(x=1\), how much evidence/confidence to support \(P_0\) or \(P_1\)?
Frequentist approach
Consider the test which reject \(P_0\) when \(X=1,2\).
Upon observing \(x=1\), a frequentist will say “\(P_0\) is rejected with confidence 0.99”.
Example 2
Supposer \(X\sim N(\mu, 1)\) and it is desired to test \[ H_1: \mu \le -2 \text{ vs } H_1: \mu >2. \]
If \(x=0\), a frequentist will say “\(H_0\) is rejected with error 0.0228”.
misleading?
even worse, this frequentist statement will be valid as long as \(x\ge0\).
what is the problem here?
Fisher once said
I don’t understand yet what fiducial probability does. We shall have to live with it a long time before we know what it’s doing for us. But it should not be ignored just because we don’t yet have a clear interpretation.
Fisher called it “inverse probability”
similar to the role of probability and likelihood: \(L(\theta|x) = f(x|\theta)\)
Suppose we have - \(F(x|\theta) = P(X\le x | \theta)\)
\[ r(\theta|x) \propto - \frac{\partial F(x|\theta)}{\partial\theta} \]
Example
Supposer \(X \sim N(\theta,1)\), then \(F(x | \theta) = \Phi(x-\theta)\)
The fiducial density is \[ r(\theta|x) \propto - \frac{\partial F(x|\theta)}{\partial\theta} = \phi(x-\theta) \]
\(\theta|x \sim N(x,1)\)
fiducial inference was first proposed by Fisher in 1930
has never gained widespread acceptance
the original fiducial argument was radically different from its later versions
fiducial inference never actually developed during Fisher’s lifetime
dozens of example of non-uniqueness and non-existence
Neyman introduced confidence interval in 1934, claiming to have generalized fiducial probability
Interestingly, Fisher was one of the discussants
Fisher disputed the idea of “confidence”
“confidence” is a purely frequentist concept
“confidence” is known to omit, or suppress, part of the information supplied by the sample
In a 1935 JRSS paper, Fisher shapely criticized Neyman
“Dr. Neyman had been somewhat unwise in his choice of topics”
In 1941, Neyman published a paper “Fiducial argument and the theory of confidence interval”
During the first 2 decdeas of dispute, Fisher almost never referred directly to Neyman in print
Tsui and Weerahandi (1989) and Weerahandi (1993) proposed generalized inference and generalized \(p\)-value.
Hannig et al. (2006) noted the relationship between generalized inference and fiducial inference.
Hannig (2009) termed the new approach GFI or generalized fiducial inference
what is GFI?
switching principle
Consider the following data generating / structural / auxiliary equation \[ Y = G(U, \theta) \]
\(\theta\) is a fixed parameter
Suppose \(Y=y\)
the generalized fiducial distribution of \(\theta\) is defined as the conditional distribution of \(Q(Y, U^*)\) given \(y = G(U^*, \theta)\), i.e.,
where \(U^*\) have the same distribution of \(U\).
My favorite example: \[ Y = \theta + Z, \ Z \sim N(0,1) \]
The generalized fiducial distribution of \(\theta\) is
\[ Q(y, Z^*) | \{\exists \theta, y = \theta + Z^*\} \]
which is \(y-Z^* \sim N(y, 1)\)
\[ \begin{align*} Y_1 = \theta + Z_1\\ Y_2 = \theta + Z_2 \end{align*} \]
Given \(Y=y\) the fiducial distribution is
\[ Q({\boldsymbol{y}}, {\boldsymbol{Z}}^*) | \{y_1 - Z_1^* = y_2 - Z_2^*\} \]
which is \(N(\bar y, 1)\).
\(\{y_1 - Z_1^* = y_2 - Z_2^*\}\) is of probability 0
lead to Borel paradox - conditioning on sets of probability 0 may not be well defined
example, if \(X\) and \(Y\) are i.i.d. standard normal, what is the distribution of \(Y\) given \(Y=X\).
Under some regularity conditions,
\[ \begin{equation*} r(\theta|y) \propto J(y,\theta) f(y,\theta) , \end{equation*} \] where \(f(y,\theta)\) is the likelihood and the function \(J(y,\theta)\) is
\[ \begin{equation*} J(y,\theta)= \sum_{\substack{{\boldsymbol{i}}=(i_1,\ldots,i_p) \\ 1\leq i_1<\cdots<i_p\leq n}}\left|\det\left(\left.\frac{\partial}{\partial\theta} G(u,\theta)\right|_{u=Q(y,\theta)}\right)_{\boldsymbol{i}}\right|. \end{equation*} \]
Theorem 1 (Bernstein–von Mises theorem of GFI): let
\[r^{*}\left(s|y\right)=n^{-1/2} r(n^{-1/2}s+\hat{\theta} | y),\]
we have \[ \int_{\mathbb{R}^{p}}\left|r^{*}\left(s|y\right)-\frac{\sqrt{det\left|I\left({\theta}_{0}\right)\right|}}{\sqrt{2\pi}}e^{-s^{T}I\left({\theta}_{0}\right)s/2}\right|\, ds\stackrel{P_{\boldsymbol{\theta}_{0}}}{\rightarrow}0. \]
Theorem 2 (Consistency of confidence sets):
Under some regularity conditions,
\[ P(\theta_0 \in C_n(Y_n)) \to 1 - \alpha \]
Berger, J. (1985). “The frequentisi viewpoint and conditioning,” In Proceedings of the berkeley conference in honor ofJerzy neyman and jack kieferVol. 1, pp. 15–44.
Berger, J. O., Bernardo, J. M., and Sun, D. (2009). “The formal definition of reference priors,” The Annals of Statistics.
Dempster, A. P. (2008). “The dempster-shafer calculus for statisticians,” International Journal of Approximate Reasoning 48, 365–377.
Efron, B. (1998). “RA fisher in the 21st century,” Statistical Science.
Fisher, R. (1935). “The fiducial argument in statistical inference,” Annals of Human Genetics 6, 391–398.
Fisher, R. A. (1930). “Inverse probability,” In Mathematical proceedings of the cambridge philosophical society(Cambridge Univ Press), Vol. 26, pp. 528–535.
Fisher, R. A. (1933). “The concepts of inverse probability and fiducial probability referring to unknown parameters,” Proceedings of the Royal Society of London. Series A 139, 343–348.
Hannig, J. (2009). “On generalized fiducial inference,” Statistica Sinica 19, 491–544.
Hannig, J., Iyer, H., and Patterson, P. (2006). “Fiducial generalized confidence intervals,” Journal of the American Statistical Association 101, 254–269.
Hannig, J., Lai, R., and Lee, T. (2014). “Computational issues of generalized fiducial inference,” Computational Statistics & Data Analysis 71, Special Issue on Imprecision in Statistical Data Analysis, 849–858 (invited by special issue co–editors).
Lai, R. C. S., Hannig, J., and Lee, T. C. M. (2014). Generalized fiducial inference for ultrahigh dimensional regression.
Liu, C., and Martin, R. (2014). “Frameworks for prior-free posterior probabilistic inference,” Wiley Interdisciplinary Reviews: Computational Statistics.
Neyman, J. (1967). A selection of early statistical papers of j. neyman (University of California Press Berkeley, CA).
Tsui, K.-W., and Weerahandi, S. (1989). “Generalized p-values in significance testing of hypotheses in the presence of nuisance parameters,” Journal of the American Statistical Association 84, 602–607.
Weerahandi, S. (1993). “Generalized confidence intervals,” Journal of the American Statistical Association 88, 899–905.
Xie, M.-g., and Singh, K. (2013). “Confidence distribution, the frequentist distribution estimator of a parameter: A review,” International Statistical Review.
Zabell, S. (1992). “RA fisher and fiducial argument,” Statistical Science.