Cargando…

Hierarchical Gradient Smoothing for Probability Estimation Trees

Decision trees are still seeing use in online, non-stationary and embedded contexts, as well as for interpretability. For applications like ranking and cost-sensitive classification, probability estimation trees (PETs) are used. These are built using smoothing or calibration techniques. Older smooth...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, He, Petitjean, François, Buntine, Wray
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206292/
http://dx.doi.org/10.1007/978-3-030-47426-3_18
_version_ 1783530387597688832
author Zhang, He
Petitjean, François
Buntine, Wray
author_facet Zhang, He
Petitjean, François
Buntine, Wray
author_sort Zhang, He
collection PubMed
description Decision trees are still seeing use in online, non-stationary and embedded contexts, as well as for interpretability. For applications like ranking and cost-sensitive classification, probability estimation trees (PETs) are used. These are built using smoothing or calibration techniques. Older smoothing techniques used counts local to a leaf node, but a few more recent techniques consider the broader context of a node when doing estimation. We apply a recent advanced smoothing method called Hierarchical Dirichlet Process (HDP) to PETs, and then propose a novel hierarchical smoothing approach called Hierarchical Gradient Smoothing (HGS) as an alternative. HGS smooths leaf nodes up to all the ancestors, instead of recursively smoothing to the parent used by HDP. HGS is made faster by efficiently optimizing the Leave-One-Out Cross-Validation (LOOCV) loss measure using gradient descent, instead of sampling used in HDP. An extensive set of experiments are conducted on 143 datasets showing that our HGS estimates are not only more accurate but also do so within a fraction of HDP time. Besides, HGS makes a single tree almost as good as a Random Forest with 10 trees. For applications that require more interpretability and efficiency, a single decision tree plus HGS is more preferred.
format Online
Article
Text
id pubmed-7206292
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-72062922020-05-08 Hierarchical Gradient Smoothing for Probability Estimation Trees Zhang, He Petitjean, François Buntine, Wray Advances in Knowledge Discovery and Data Mining Article Decision trees are still seeing use in online, non-stationary and embedded contexts, as well as for interpretability. For applications like ranking and cost-sensitive classification, probability estimation trees (PETs) are used. These are built using smoothing or calibration techniques. Older smoothing techniques used counts local to a leaf node, but a few more recent techniques consider the broader context of a node when doing estimation. We apply a recent advanced smoothing method called Hierarchical Dirichlet Process (HDP) to PETs, and then propose a novel hierarchical smoothing approach called Hierarchical Gradient Smoothing (HGS) as an alternative. HGS smooths leaf nodes up to all the ancestors, instead of recursively smoothing to the parent used by HDP. HGS is made faster by efficiently optimizing the Leave-One-Out Cross-Validation (LOOCV) loss measure using gradient descent, instead of sampling used in HDP. An extensive set of experiments are conducted on 143 datasets showing that our HGS estimates are not only more accurate but also do so within a fraction of HDP time. Besides, HGS makes a single tree almost as good as a Random Forest with 10 trees. For applications that require more interpretability and efficiency, a single decision tree plus HGS is more preferred. 2020-04-17 /pmc/articles/PMC7206292/ http://dx.doi.org/10.1007/978-3-030-47426-3_18 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Zhang, He
Petitjean, François
Buntine, Wray
Hierarchical Gradient Smoothing for Probability Estimation Trees
title Hierarchical Gradient Smoothing for Probability Estimation Trees
title_full Hierarchical Gradient Smoothing for Probability Estimation Trees
title_fullStr Hierarchical Gradient Smoothing for Probability Estimation Trees
title_full_unstemmed Hierarchical Gradient Smoothing for Probability Estimation Trees
title_short Hierarchical Gradient Smoothing for Probability Estimation Trees
title_sort hierarchical gradient smoothing for probability estimation trees
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206292/
http://dx.doi.org/10.1007/978-3-030-47426-3_18
work_keys_str_mv AT zhanghe hierarchicalgradientsmoothingforprobabilityestimationtrees
AT petitjeanfrancois hierarchicalgradientsmoothingforprobabilityestimationtrees
AT buntinewray hierarchicalgradientsmoothingforprobabilityestimationtrees