Cargando…
A Feature Selection Algorithm Integrating Maximum Classification Information and Minimum Interaction Feature Dependency Information
Feature selection is the key step in the analysis of high-dimensional small sample data. The core of feature selection is to analyse and quantify the correlation between features and class labels and the redundancy between features. However, most of the existing feature selection algorithms only con...
Autor principal: | |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8727115/ https://www.ncbi.nlm.nih.gov/pubmed/34992644 http://dx.doi.org/10.1155/2021/3569632 |
_version_ | 1784626446330757120 |
---|---|
author | Zhang, Li |
author_facet | Zhang, Li |
author_sort | Zhang, Li |
collection | PubMed |
description | Feature selection is the key step in the analysis of high-dimensional small sample data. The core of feature selection is to analyse and quantify the correlation between features and class labels and the redundancy between features. However, most of the existing feature selection algorithms only consider the classification contribution of individual features and ignore the influence of interfeature redundancy and correlation. Therefore, this paper proposes a feature selection algorithm for nonlinear dynamic conditional relevance (NDCRFS) through the study and analysis of the existing feature selection algorithm ideas and method. Firstly, redundancy and relevance between features and between features and class labels are discriminated by mutual information, conditional mutual information, and interactive mutual information. Secondly, the selected features and candidate features are dynamically weighted utilizing information gain factors. Finally, to evaluate the performance of this feature selection algorithm, NDCRFS was validated against 6 other feature selection algorithms on three classifiers, using 12 different data sets, for variability and classification metrics between the different algorithms. The experimental results show that the NDCRFS method can improve the quality of the feature subsets and obtain better classification results. |
format | Online Article Text |
id | pubmed-8727115 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-87271152022-01-05 A Feature Selection Algorithm Integrating Maximum Classification Information and Minimum Interaction Feature Dependency Information Zhang, Li Comput Intell Neurosci Research Article Feature selection is the key step in the analysis of high-dimensional small sample data. The core of feature selection is to analyse and quantify the correlation between features and class labels and the redundancy between features. However, most of the existing feature selection algorithms only consider the classification contribution of individual features and ignore the influence of interfeature redundancy and correlation. Therefore, this paper proposes a feature selection algorithm for nonlinear dynamic conditional relevance (NDCRFS) through the study and analysis of the existing feature selection algorithm ideas and method. Firstly, redundancy and relevance between features and between features and class labels are discriminated by mutual information, conditional mutual information, and interactive mutual information. Secondly, the selected features and candidate features are dynamically weighted utilizing information gain factors. Finally, to evaluate the performance of this feature selection algorithm, NDCRFS was validated against 6 other feature selection algorithms on three classifiers, using 12 different data sets, for variability and classification metrics between the different algorithms. The experimental results show that the NDCRFS method can improve the quality of the feature subsets and obtain better classification results. Hindawi 2021-12-28 /pmc/articles/PMC8727115/ /pubmed/34992644 http://dx.doi.org/10.1155/2021/3569632 Text en Copyright © 2021 Li Zhang. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Zhang, Li A Feature Selection Algorithm Integrating Maximum Classification Information and Minimum Interaction Feature Dependency Information |
title | A Feature Selection Algorithm Integrating Maximum Classification Information and Minimum Interaction Feature Dependency Information |
title_full | A Feature Selection Algorithm Integrating Maximum Classification Information and Minimum Interaction Feature Dependency Information |
title_fullStr | A Feature Selection Algorithm Integrating Maximum Classification Information and Minimum Interaction Feature Dependency Information |
title_full_unstemmed | A Feature Selection Algorithm Integrating Maximum Classification Information and Minimum Interaction Feature Dependency Information |
title_short | A Feature Selection Algorithm Integrating Maximum Classification Information and Minimum Interaction Feature Dependency Information |
title_sort | feature selection algorithm integrating maximum classification information and minimum interaction feature dependency information |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8727115/ https://www.ncbi.nlm.nih.gov/pubmed/34992644 http://dx.doi.org/10.1155/2021/3569632 |
work_keys_str_mv | AT zhangli afeatureselectionalgorithmintegratingmaximumclassificationinformationandminimuminteractionfeaturedependencyinformation AT zhangli featureselectionalgorithmintegratingmaximumclassificationinformationandminimuminteractionfeaturedependencyinformation |