Cargando…

On the α-q-Mutual Information and the α-q-Capacities

The measures of information transfer which correspond to non-additive entropies have intensively been studied in previous decades. The majority of the work includes the ones belonging to the Sharma–Mittal entropy class, such as the Rényi, the Tsallis, the Landsberg–Vedral and the Gaussian entropies....

Descripción completa

Detalles Bibliográficos
Autores principales: Ilić, Velimir M., Djordjević, Ivan B.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8229514/
https://www.ncbi.nlm.nih.gov/pubmed/34206138
http://dx.doi.org/10.3390/e23060702
_version_ 1783712994320973824
author Ilić, Velimir M.
Djordjević, Ivan B.
author_facet Ilić, Velimir M.
Djordjević, Ivan B.
author_sort Ilić, Velimir M.
collection PubMed
description The measures of information transfer which correspond to non-additive entropies have intensively been studied in previous decades. The majority of the work includes the ones belonging to the Sharma–Mittal entropy class, such as the Rényi, the Tsallis, the Landsberg–Vedral and the Gaussian entropies. All of the considerations follow the same approach, mimicking some of the various and mutually equivalent definitions of Shannon information measures, and the information transfer is quantified by an appropriately defined measure of mutual information, while the maximal information transfer is considered as a generalized channel capacity. However, all of the previous approaches fail to satisfy at least one of the ineluctable properties which a measure of (maximal) information transfer should satisfy, leading to counterintuitive conclusions and predicting nonphysical behavior even in the case of very simple communication channels. This paper fills the gap by proposing two parameter measures named the [Formula: see text]-q-mutual information and the [Formula: see text]-q-capacity. In addition to standard Shannon approaches, special cases of these measures include the [Formula: see text]-mutual information and the [Formula: see text]-capacity, which are well established in the information theory literature as measures of additive Rényi information transfer, while the cases of the Tsallis, the Landsberg–Vedral and the Gaussian entropies can also be accessed by special choices of the parameters [Formula: see text] and q. It is shown that, unlike the previous definition, the [Formula: see text]-q-mutual information and the [Formula: see text]-q-capacity satisfy the set of properties, which are stated as axioms, by which they reduce to zero in the case of totally destructive channels and to the (maximal) input Sharma–Mittal entropy in the case of perfect transmission, which is consistent with the maximum likelihood detection error. In addition, they are non-negative and less than or equal to the input and the output Sharma–Mittal entropies, in general. Thus, unlike the previous approaches, the proposed (maximal) information transfer measures do not manifest nonphysical behaviors such as sub-capacitance or super-capacitance, which could qualify them as appropriate measures of the Sharma–Mittal information transfer.
format Online
Article
Text
id pubmed-8229514
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-82295142021-06-26 On the α-q-Mutual Information and the α-q-Capacities Ilić, Velimir M. Djordjević, Ivan B. Entropy (Basel) Article The measures of information transfer which correspond to non-additive entropies have intensively been studied in previous decades. The majority of the work includes the ones belonging to the Sharma–Mittal entropy class, such as the Rényi, the Tsallis, the Landsberg–Vedral and the Gaussian entropies. All of the considerations follow the same approach, mimicking some of the various and mutually equivalent definitions of Shannon information measures, and the information transfer is quantified by an appropriately defined measure of mutual information, while the maximal information transfer is considered as a generalized channel capacity. However, all of the previous approaches fail to satisfy at least one of the ineluctable properties which a measure of (maximal) information transfer should satisfy, leading to counterintuitive conclusions and predicting nonphysical behavior even in the case of very simple communication channels. This paper fills the gap by proposing two parameter measures named the [Formula: see text]-q-mutual information and the [Formula: see text]-q-capacity. In addition to standard Shannon approaches, special cases of these measures include the [Formula: see text]-mutual information and the [Formula: see text]-capacity, which are well established in the information theory literature as measures of additive Rényi information transfer, while the cases of the Tsallis, the Landsberg–Vedral and the Gaussian entropies can also be accessed by special choices of the parameters [Formula: see text] and q. It is shown that, unlike the previous definition, the [Formula: see text]-q-mutual information and the [Formula: see text]-q-capacity satisfy the set of properties, which are stated as axioms, by which they reduce to zero in the case of totally destructive channels and to the (maximal) input Sharma–Mittal entropy in the case of perfect transmission, which is consistent with the maximum likelihood detection error. In addition, they are non-negative and less than or equal to the input and the output Sharma–Mittal entropies, in general. Thus, unlike the previous approaches, the proposed (maximal) information transfer measures do not manifest nonphysical behaviors such as sub-capacitance or super-capacitance, which could qualify them as appropriate measures of the Sharma–Mittal information transfer. MDPI 2021-06-01 /pmc/articles/PMC8229514/ /pubmed/34206138 http://dx.doi.org/10.3390/e23060702 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Ilić, Velimir M.
Djordjević, Ivan B.
On the α-q-Mutual Information and the α-q-Capacities
title On the α-q-Mutual Information and the α-q-Capacities
title_full On the α-q-Mutual Information and the α-q-Capacities
title_fullStr On the α-q-Mutual Information and the α-q-Capacities
title_full_unstemmed On the α-q-Mutual Information and the α-q-Capacities
title_short On the α-q-Mutual Information and the α-q-Capacities
title_sort on the α-q-mutual information and the α-q-capacities
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8229514/
https://www.ncbi.nlm.nih.gov/pubmed/34206138
http://dx.doi.org/10.3390/e23060702
work_keys_str_mv AT ilicvelimirm ontheaqmutualinformationandtheaqcapacities
AT djordjevicivanb ontheaqmutualinformationandtheaqcapacities