Literature DB >> 33286335

On Relations Between the Relative Entropy and χ2-Divergence, Generalizations and Applications.

Tomohiro Nishiyama1, Igal Sason2.   

Abstract

The relative entropy and the chi-squared divergence are fundamental divergence measures in information theory and statistics. This paper is focused on a study of integral relations between the two divergences, the implications of these relations, their information-theoretic applications, and some generalizations pertaining to the rich class of f-divergences. Applications that are studied in this paper refer to lossless compression, the method of types and large deviations, strong data-processing inequalities, bounds on contraction coefficients and maximal correlation, and the convergence rate to stationarity of a type of discrete-time Markov chains.

Entities:  

Keywords:  Markov chains; chi-squared divergence; f-divergences; information contraction; large deviations; maximal correlation; method of types; relative entropy; strong data–processing inequalities

Year:  2020        PMID: 33286335     DOI: 10.3390/e22050563

Source DB:  PubMed          Journal:  Entropy (Basel)        ISSN: 1099-4300            Impact factor:   2.524


  1 in total

1.  Divergence Measures: Mathematical Foundations and Applications in Information-Theoretic and Statistical Problems.

Authors:  Igal Sason
Journal:  Entropy (Basel)       Date:  2022-05-16       Impact factor: 2.738

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.