Cargando…

Discriminant Analysis under f-Divergence Measures

In statistical inference, the information-theoretic performance limits can often be expressed in terms of a statistical divergence between the underlying statistical models (e.g., in binary hypothesis testing, the error probability is related to the total variation distance between the statistical m...

Descripción completa

Detalles Bibliográficos
Autores principales: Dwivedi, Anmol, Wang, Sihui, Tajer, Ali
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8870799/
https://www.ncbi.nlm.nih.gov/pubmed/35205483
http://dx.doi.org/10.3390/e24020188
_version_ 1784656843388223488
author Dwivedi, Anmol
Wang, Sihui
Tajer, Ali
author_facet Dwivedi, Anmol
Wang, Sihui
Tajer, Ali
author_sort Dwivedi, Anmol
collection PubMed
description In statistical inference, the information-theoretic performance limits can often be expressed in terms of a statistical divergence between the underlying statistical models (e.g., in binary hypothesis testing, the error probability is related to the total variation distance between the statistical models). As the data dimension grows, computing the statistics involved in decision-making and the attendant performance limits (divergence measures) face complexity and stability challenges. Dimensionality reduction addresses these challenges at the expense of compromising the performance (the divergence reduces by the data-processing inequality). This paper considers linear dimensionality reduction such that the divergence between the models is maximally preserved. Specifically, this paper focuses on Gaussian models where we investigate discriminant analysis under five f-divergence measures (Kullback–Leibler, symmetrized Kullback–Leibler, Hellinger, total variation, and [Formula: see text]). We characterize the optimal design of the linear transformation of the data onto a lower-dimensional subspace for zero-mean Gaussian models and employ numerical algorithms to find the design for general Gaussian models with non-zero means. There are two key observations for zero-mean Gaussian models. First, projections are not necessarily along the largest modes of the covariance matrix of the data, and, in some situations, they can even be along the smallest modes. Secondly, under specific regimes, the optimal design of subspace projection is identical under all the f-divergence measures considered, rendering a degree of universality to the design, independent of the inference problem of interest.
format Online
Article
Text
id pubmed-8870799
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-88707992022-02-25 Discriminant Analysis under f-Divergence Measures Dwivedi, Anmol Wang, Sihui Tajer, Ali Entropy (Basel) Article In statistical inference, the information-theoretic performance limits can often be expressed in terms of a statistical divergence between the underlying statistical models (e.g., in binary hypothesis testing, the error probability is related to the total variation distance between the statistical models). As the data dimension grows, computing the statistics involved in decision-making and the attendant performance limits (divergence measures) face complexity and stability challenges. Dimensionality reduction addresses these challenges at the expense of compromising the performance (the divergence reduces by the data-processing inequality). This paper considers linear dimensionality reduction such that the divergence between the models is maximally preserved. Specifically, this paper focuses on Gaussian models where we investigate discriminant analysis under five f-divergence measures (Kullback–Leibler, symmetrized Kullback–Leibler, Hellinger, total variation, and [Formula: see text]). We characterize the optimal design of the linear transformation of the data onto a lower-dimensional subspace for zero-mean Gaussian models and employ numerical algorithms to find the design for general Gaussian models with non-zero means. There are two key observations for zero-mean Gaussian models. First, projections are not necessarily along the largest modes of the covariance matrix of the data, and, in some situations, they can even be along the smallest modes. Secondly, under specific regimes, the optimal design of subspace projection is identical under all the f-divergence measures considered, rendering a degree of universality to the design, independent of the inference problem of interest. MDPI 2022-01-27 /pmc/articles/PMC8870799/ /pubmed/35205483 http://dx.doi.org/10.3390/e24020188 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Dwivedi, Anmol
Wang, Sihui
Tajer, Ali
Discriminant Analysis under f-Divergence Measures
title Discriminant Analysis under f-Divergence Measures
title_full Discriminant Analysis under f-Divergence Measures
title_fullStr Discriminant Analysis under f-Divergence Measures
title_full_unstemmed Discriminant Analysis under f-Divergence Measures
title_short Discriminant Analysis under f-Divergence Measures
title_sort discriminant analysis under f-divergence measures
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8870799/
https://www.ncbi.nlm.nih.gov/pubmed/35205483
http://dx.doi.org/10.3390/e24020188
work_keys_str_mv AT dwivedianmol discriminantanalysisunderfdivergencemeasures
AT wangsihui discriminantanalysisunderfdivergencemeasures
AT tajerali discriminantanalysisunderfdivergencemeasures