Cargando…

Attribute Selection Based on Constraint Gain and Depth Optimal for a Decision Tree

Uncertainty evaluation based on statistical probabilistic information entropy is a commonly used mechanism for a heuristic method construction of decision tree learning. The entropy kernel potentially links its deviation and decision tree classification performance. This paper presents a decision tr...

Descripción completa

Detalles Bibliográficos
Autores principales: Sun, Huaining, Hu, Xuegang, Zhang, Yuhong
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7514679/
https://www.ncbi.nlm.nih.gov/pubmed/33266913
http://dx.doi.org/10.3390/e21020198
_version_ 1783586643710574592
author Sun, Huaining
Hu, Xuegang
Zhang, Yuhong
author_facet Sun, Huaining
Hu, Xuegang
Zhang, Yuhong
author_sort Sun, Huaining
collection PubMed
description Uncertainty evaluation based on statistical probabilistic information entropy is a commonly used mechanism for a heuristic method construction of decision tree learning. The entropy kernel potentially links its deviation and decision tree classification performance. This paper presents a decision tree learning algorithm based on constrained gain and depth induction optimization. Firstly, the calculation and analysis of single- and multi-value event uncertainty distributions of information entropy is followed by an enhanced property of single-value event entropy kernel and multi-value event entropy peaks as well as a reciprocal relationship between peak location and the number of possible events. Secondly, this study proposed an estimated method for information entropy whose entropy kernel is replaced with a peak-shift sine function to establish a decision tree learning (CGDT) algorithm on the basis of constraint gain. Finally, by combining branch convergence and fan-out indices under an inductive depth of a decision tree, we built a constraint gained and depth inductive improved decision tree (CGDIDT) learning algorithm. Results show the benefits of the CGDT and CGDIDT algorithms.
format Online
Article
Text
id pubmed-7514679
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75146792020-11-09 Attribute Selection Based on Constraint Gain and Depth Optimal for a Decision Tree Sun, Huaining Hu, Xuegang Zhang, Yuhong Entropy (Basel) Article Uncertainty evaluation based on statistical probabilistic information entropy is a commonly used mechanism for a heuristic method construction of decision tree learning. The entropy kernel potentially links its deviation and decision tree classification performance. This paper presents a decision tree learning algorithm based on constrained gain and depth induction optimization. Firstly, the calculation and analysis of single- and multi-value event uncertainty distributions of information entropy is followed by an enhanced property of single-value event entropy kernel and multi-value event entropy peaks as well as a reciprocal relationship between peak location and the number of possible events. Secondly, this study proposed an estimated method for information entropy whose entropy kernel is replaced with a peak-shift sine function to establish a decision tree learning (CGDT) algorithm on the basis of constraint gain. Finally, by combining branch convergence and fan-out indices under an inductive depth of a decision tree, we built a constraint gained and depth inductive improved decision tree (CGDIDT) learning algorithm. Results show the benefits of the CGDT and CGDIDT algorithms. MDPI 2019-02-19 /pmc/articles/PMC7514679/ /pubmed/33266913 http://dx.doi.org/10.3390/e21020198 Text en © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Sun, Huaining
Hu, Xuegang
Zhang, Yuhong
Attribute Selection Based on Constraint Gain and Depth Optimal for a Decision Tree
title Attribute Selection Based on Constraint Gain and Depth Optimal for a Decision Tree
title_full Attribute Selection Based on Constraint Gain and Depth Optimal for a Decision Tree
title_fullStr Attribute Selection Based on Constraint Gain and Depth Optimal for a Decision Tree
title_full_unstemmed Attribute Selection Based on Constraint Gain and Depth Optimal for a Decision Tree
title_short Attribute Selection Based on Constraint Gain and Depth Optimal for a Decision Tree
title_sort attribute selection based on constraint gain and depth optimal for a decision tree
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7514679/
https://www.ncbi.nlm.nih.gov/pubmed/33266913
http://dx.doi.org/10.3390/e21020198
work_keys_str_mv AT sunhuaining attributeselectionbasedonconstraintgainanddepthoptimalforadecisiontree
AT huxuegang attributeselectionbasedonconstraintgainanddepthoptimalforadecisiontree
AT zhangyuhong attributeselectionbasedonconstraintgainanddepthoptimalforadecisiontree