Cargando…
Discriminant Analysis under f-Divergence Measures
In statistical inference, the information-theoretic performance limits can often be expressed in terms of a statistical divergence between the underlying statistical models (e.g., in binary hypothesis testing, the error probability is related to the total variation distance between the statistical m...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8870799/ https://www.ncbi.nlm.nih.gov/pubmed/35205483 http://dx.doi.org/10.3390/e24020188 |
_version_ | 1784656843388223488 |
---|---|
author | Dwivedi, Anmol Wang, Sihui Tajer, Ali |
author_facet | Dwivedi, Anmol Wang, Sihui Tajer, Ali |
author_sort | Dwivedi, Anmol |
collection | PubMed |
description | In statistical inference, the information-theoretic performance limits can often be expressed in terms of a statistical divergence between the underlying statistical models (e.g., in binary hypothesis testing, the error probability is related to the total variation distance between the statistical models). As the data dimension grows, computing the statistics involved in decision-making and the attendant performance limits (divergence measures) face complexity and stability challenges. Dimensionality reduction addresses these challenges at the expense of compromising the performance (the divergence reduces by the data-processing inequality). This paper considers linear dimensionality reduction such that the divergence between the models is maximally preserved. Specifically, this paper focuses on Gaussian models where we investigate discriminant analysis under five f-divergence measures (Kullback–Leibler, symmetrized Kullback–Leibler, Hellinger, total variation, and [Formula: see text]). We characterize the optimal design of the linear transformation of the data onto a lower-dimensional subspace for zero-mean Gaussian models and employ numerical algorithms to find the design for general Gaussian models with non-zero means. There are two key observations for zero-mean Gaussian models. First, projections are not necessarily along the largest modes of the covariance matrix of the data, and, in some situations, they can even be along the smallest modes. Secondly, under specific regimes, the optimal design of subspace projection is identical under all the f-divergence measures considered, rendering a degree of universality to the design, independent of the inference problem of interest. |
format | Online Article Text |
id | pubmed-8870799 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-88707992022-02-25 Discriminant Analysis under f-Divergence Measures Dwivedi, Anmol Wang, Sihui Tajer, Ali Entropy (Basel) Article In statistical inference, the information-theoretic performance limits can often be expressed in terms of a statistical divergence between the underlying statistical models (e.g., in binary hypothesis testing, the error probability is related to the total variation distance between the statistical models). As the data dimension grows, computing the statistics involved in decision-making and the attendant performance limits (divergence measures) face complexity and stability challenges. Dimensionality reduction addresses these challenges at the expense of compromising the performance (the divergence reduces by the data-processing inequality). This paper considers linear dimensionality reduction such that the divergence between the models is maximally preserved. Specifically, this paper focuses on Gaussian models where we investigate discriminant analysis under five f-divergence measures (Kullback–Leibler, symmetrized Kullback–Leibler, Hellinger, total variation, and [Formula: see text]). We characterize the optimal design of the linear transformation of the data onto a lower-dimensional subspace for zero-mean Gaussian models and employ numerical algorithms to find the design for general Gaussian models with non-zero means. There are two key observations for zero-mean Gaussian models. First, projections are not necessarily along the largest modes of the covariance matrix of the data, and, in some situations, they can even be along the smallest modes. Secondly, under specific regimes, the optimal design of subspace projection is identical under all the f-divergence measures considered, rendering a degree of universality to the design, independent of the inference problem of interest. MDPI 2022-01-27 /pmc/articles/PMC8870799/ /pubmed/35205483 http://dx.doi.org/10.3390/e24020188 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Dwivedi, Anmol Wang, Sihui Tajer, Ali Discriminant Analysis under f-Divergence Measures |
title | Discriminant Analysis under f-Divergence Measures |
title_full | Discriminant Analysis under f-Divergence Measures |
title_fullStr | Discriminant Analysis under f-Divergence Measures |
title_full_unstemmed | Discriminant Analysis under f-Divergence Measures |
title_short | Discriminant Analysis under f-Divergence Measures |
title_sort | discriminant analysis under f-divergence measures |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8870799/ https://www.ncbi.nlm.nih.gov/pubmed/35205483 http://dx.doi.org/10.3390/e24020188 |
work_keys_str_mv | AT dwivedianmol discriminantanalysisunderfdivergencemeasures AT wangsihui discriminantanalysisunderfdivergencemeasures AT tajerali discriminantanalysisunderfdivergencemeasures |