Cargando…

On Relations Between the Relative Entropy and χ(2)-Divergence, Generalizations and Applications

This paper is focused on a study of integral relations between the relative entropy and the chi-squared divergence, which are two fundamental divergence measures in information theory and statistics, a study of the implications of these relations, their information-theoretic applications, and some g...

Descripción completa

Detalles Bibliográficos
Autores principales: Nishiyama, Tomohiro, Sason, Igal
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7848888/
https://www.ncbi.nlm.nih.gov/pubmed/33286335
http://dx.doi.org/10.3390/e22050563
Descripción
Sumario:This paper is focused on a study of integral relations between the relative entropy and the chi-squared divergence, which are two fundamental divergence measures in information theory and statistics, a study of the implications of these relations, their information-theoretic applications, and some generalizations pertaining to the rich class of f-divergences. Applications that are studied in this paper refer to lossless compression, the method of types and large deviations, strong data–processing inequalities, bounds on contraction coefficients and maximal correlation, and the convergence rate to stationarity of a type of discrete-time Markov chains.