Investigation on the high-order approximation of the entropy bias

The estimation of entropy from experimental data has a considerable bias when the discretization of the variable domain is comparable to the sample size. In this case, the source of the bias is the difference between the a priori distribution and the observed distribution from sampled data. In this...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Physica A 2020-07, Vol.549, p.124301, Article 124301
Hauptverfasser: Ramos, Antônio M.T., Casagrande, Helder L., Macau, Elbert E.N.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:The estimation of entropy from experimental data has a considerable bias when the discretization of the variable domain is comparable to the sample size. In this case, the source of the bias is the difference between the a priori distribution and the observed distribution from sampled data. In this paper, we estimate the entropy bias considering an infinite sum of central moments of the binomial distribution using two probability mass functions. We analyze the bias in the light of the ratio between the number of the partition of the domain and the sample size. The main motivation of this study is improving statistical hypothesis testing in which probabilities are conceived beforehand. We examine the adequacy of high-order approximation according to the ratio between the sample size and the number of domain partitions. Finally, we expand the analysis to the entropy-derived mutual information and present an application for network reconstruction. •Estimation of the entropy bias via a sum of moments of probability functions.•Adequacy of high-order expansion in information-theory functionals•Hypothesis testing using entropy-derived mutual information.•Application in network reconstruction.
ISSN:0378-4371
1873-2119
DOI:10.1016/j.physa.2020.124301