| Literature DB >> 35052151 |
Damián G Hernández1, Inés Samengo1.
Abstract
Inferring the value of a property of a large stochastic system is a difficult task when the number of samples is insufficient to reliably estimate the probability distribution. The Bayesian estimator of the property of interest requires the knowledge of the prior distribution, and in many situations, it is not clear which prior should be used. Several estimators have been developed so far in which the proposed prior us individually tailored for each property of interest; such is the case, for example, for the entropy, the amount of mutual information, or the correlation between pairs of variables. In this paper, we propose a general framework to select priors that is valid for arbitrary properties. We first demonstrate that only certain aspects of the prior distribution actually affect the inference process. We then expand the sought prior as a linear combination of a one-dimensional family of indexed priors, each of which is obtained through a maximum entropy approach with constrained mean values of the property under study. In many cases of interest, only one or very few components of the expansion turn out to contribute to the Bayesian estimator, so it is often valid to only keep a single component. The relevant component is selected by the data, so no handcrafted priors are required. We test the performance of this approximation with a few paradigmatic examples and show that it performs well in comparison to the ad-hoc methods previously proposed in the literature. Our method highlights the connection between Bayesian inference and equilibrium statistical mechanics, since the most relevant component of the expansion can be argued to be that with the right temperature.Entities:
Keywords: bayesian; entropy; inference; mutual information; undersampled
Year: 2022 PMID: 35052151 PMCID: PMC8775033 DOI: 10.3390/e24010125
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Figure 1Conceptual framework used to estimate the property of a stochastic system with probabilities from a limited number of observations, or samples . (a) Several probabilities can produce the same value of the property. All such -vectors belong to the same level surface of . (b) Left: Example case in which the property has circular level surfaces on the simplex embedded in three-dimensional space. Two members of the family of base functions used to expand the prior are shown, containing the different vectors shown in (a). These two members are solutions of the Maxentropy problem with two different expected values f of the property. Middle: The histogram generated by the sampled data produces a likelihood that selective favors a specific range of surface levels. Right: The most favored f value (or equivalently, value) is the one for which the prior and posterior estimates of the property coincide.
Figure 2Toy example used to illustrate the estimation of the amount of mutual information in the same system as in Section 3.1. (a) The system is governed by a bivariate probability distribution, with states characterized by two labels: x and y. There are many x states, , and all are equally probable; that is, . The variable y is binary, and its two values are depicted as red and blue, corresponding to 1 and 0, respectively. The conditional probabilities were sampled from a symmetric beta distribution with parameter . The goal is to estimate the mutual information from n samples, with . (b) The level surfaces of the mutual information, as well as some sampled -values, are displayed in a three-dimensional subspace of the full -space, for different values of the hyperparameter . (c) Prior mean mutual information as a function of the scaled hyperparameter , and its fluctuations for . (d) Prior and posterior mean mutual information as a function of the scaled hyperparameter for samples. The intersection of these curves corresponds to the MAP estimation . In gray, the posterior marginal evidence for the hyperparameter, whose width decreases as the square root of the number of states with at least two samples, . (e) Comparison of the estimation of mutual information between different methods for the considered set of samples ( is the estimator from [8]). The horizontal line corresponds to the true value of the mutual information.
Figure 3Entropy estimation in a toy example. (a) A distribution has ranked probabilities that decrease with a power law () within a finite number of states (). Distributions such as these represent a challenge for the estimation of entropy. Three possible sets of samples are displayed. (b) Prior mean entropy, plus/minus a standard deviation, as a function of the available number of states k for several values of the hyperparameter . (c) Prior and posterior mean entropy as function of for the different possible sets of multiplicities that can be obtained from samples (three examples shown in panel a, with matching colors). The intersections between prior and posterior mean entropy (MAP estimation) are marked with circles. The size of the circles is proportional to the likelihood of the corresponding multiplicity (a minimum size is imposed, for visibility). (d) Comparison of the average estimation of entropy between different methods over all the multiplicity sets (discarding the set with no coincidences, and the set with all samples in one state), weighting according to their likelihood. The horizontal line corresponds to the true value of the entropy.