Archive
Integrated likelihood in light of de Finetti
Coletti, Giulianella; Scozzafava, Romano; Vantaggi, Barbara
Integrated likelihood in a finitely additive setting. (English summary) Symbolic and quantitative approaches to reasoning with uncertainty, 554–565, Lecture Notes in Comput. Sci., 5590, Springer, Berlin, 2009.
62A01 (62A99)
Interpreting the likelihood function under the coherence framework of de Finetti, this paper mathematically formulates the problem by defining the likelihood of a simple or composite hypothesis as a subjective probability of the observed data conditional on the truth of the hypothesis. In the probability theory of this framework, conditional probabilities given a hypothesis or event of probability zero are well defined, even for finite parameter sets. That differs from the familiar probability measures that Kolmogorov introduced for frequency-type probabilities, each of which, in the finite case, can only have zero probability mass if its event cannot occur. (The latter but not the former agrees in spirit with Cournot’s principle that an event of infinitesimally small probability is physically impossible.) Thus, in the de Finetti framework, the likelihood function assigns a conditional probability to each simple hypothesis, whether or not its probability is zero.
When the parameter set is finite, every coherent conditional probability of a sample of discrete data given a composite hypothesis is a weighted arithmetic mean of the conditional probabilities of the simple hypotheses that together constitute the composite hypothesis. In other words, the coherence constraint requires that the likelihood of a composite hypothesis be a linear combination of the likelihoods of its constituent simple hypotheses. Important special cases include the maximum and the minimum of the likelihood over the parameter set. They are made possible in the non-Kolmogorov framework by assigning zero probability to all of the simple hypotheses except those of maximum or minimum likelihood.
The main result of the paper extends this result to infinite parameter sets. In general, the likelihood of a composite hypothesis is a mixture of the likelihoods of its component simple hypotheses.
{For the entire collection see MR2907743 (2012j:68012).}
Reviewed by David R. Bickel
This review first appeared at “Integrated likelihood in a finitely additive setting” (Mathematical Reviews) and is used with permission from the American Mathematical Society.
Research topics in the Statomics Lab
Coherent fiducial distributions
D. R. Bickel and M. Padilla, “A prior-free framework of coherent inference and its derivation of simple shrinkage estimators,” Journal of Statistical Planning and Inference 145, 204–221 (2014). 2012 version

Small dimensional empirical Bayes inference
D. R. Bickel, “Simple estimators of false discovery rates given as few as one or two p-values without strong parametric assumptions,” Statistical Applications in Genetics and Molecular Biology 12, 529–543 (2013). 2011 version | erratum
To address multiple comparison problems in small-to-high-dimensional biology, this paper introduces estimators of the local false discovery rate (LFDR), reports their main properties, and illustrates their use with proteomics data. The new estimators have the following advantages:
- proven asymptotic conservatism;
- simplicity of calculation without the tuning of smoothing parameters;
- no strong parametric assumptions;
- applicability to very small numbers of hypotheses as well as to very large numbers of hypotheses.
The link to the erratum was added 31 March 2015.
Profile likelihood & MDL for measuring the strength of evidence
D. R. Bickel, “Pseudo-likelihood, explanatory power, and Bayes’s theorem [Comment on ‘A likelihood paradigm for clinical trials’],” Journal of Statistical Theory and Practice 7, 178-182 (2013).

Estimates of the local FDR
Z. Yang, Z. Li, and D. R. Bickel, “Empirical Bayes estimation of posterior probabilities of enrichment: A comparative study of five estimators of the local false discovery rate,” BMC Bioinformatics 14, art. 87 (2013). published version | 2011 version | 2010 version
This paper adapts novel empirical Bayes methods for the problem of detecting enrichment in the form of differential representation of genes associated with a biological category with respect to a list of genes identified as differentially expressed. Read more…
Optimal strength of evidence
D. R. Bickel, “Minimax-optimal strength of statistical evidence for a composite alternative hypothesis,” International Statistical Review 81, 188-206 (2013). 2011 version | Simple explanation (added 2 July 2017)
This publication generalizes the likelihood measure of evidential support for a hypothesis with the help of tools originally developed by information theorists for minimizing the number of letters in a message. The approach is illustrated with an application to proteomics data.
Confidence levels as degrees of belief
D. R. Bickel, “A frequentist framework of inductive reasoning,” Sankhya A 74, 141-169 (2013). published version | 2009 version
| relationship to a working paper | simple explanation (added 17 July 2017)
A confidence measure is a parameter distribution that encodes all confidence intervals for a given data set, model, and pivot. This article establishes some properties of the confidence measure that commend it as a viable alternative to the Bayesian posterior distribution.
Confidence (correct frequentist coverage) and coherence (compliance with Ramsey-type restrictions on rational belief) are both presented as desirable properties. The only distributions on a scalar parameter space that have both properties are confidence measures.
Local FDR estimation for low-dimensional data
M. Padilla and D. R. Bickel, “Estimators of the local false discovery rate designed for small numbers of tests,” Statistical Applications in Genetics and Molecular Biology 11 (5), art. 4 (2012). Full article | 2010 & 2012 preprints




You must be logged in to post a comment.