Cargando…
Hierarchical Gradient Smoothing for Probability Estimation Trees
Decision trees are still seeing use in online, non-stationary and embedded contexts, as well as for interpretability. For applications like ranking and cost-sensitive classification, probability estimation trees (PETs) are used. These are built using smoothing or calibration techniques. Older smooth...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206292/ http://dx.doi.org/10.1007/978-3-030-47426-3_18 |
_version_ | 1783530387597688832 |
---|---|
author | Zhang, He Petitjean, François Buntine, Wray |
author_facet | Zhang, He Petitjean, François Buntine, Wray |
author_sort | Zhang, He |
collection | PubMed |
description | Decision trees are still seeing use in online, non-stationary and embedded contexts, as well as for interpretability. For applications like ranking and cost-sensitive classification, probability estimation trees (PETs) are used. These are built using smoothing or calibration techniques. Older smoothing techniques used counts local to a leaf node, but a few more recent techniques consider the broader context of a node when doing estimation. We apply a recent advanced smoothing method called Hierarchical Dirichlet Process (HDP) to PETs, and then propose a novel hierarchical smoothing approach called Hierarchical Gradient Smoothing (HGS) as an alternative. HGS smooths leaf nodes up to all the ancestors, instead of recursively smoothing to the parent used by HDP. HGS is made faster by efficiently optimizing the Leave-One-Out Cross-Validation (LOOCV) loss measure using gradient descent, instead of sampling used in HDP. An extensive set of experiments are conducted on 143 datasets showing that our HGS estimates are not only more accurate but also do so within a fraction of HDP time. Besides, HGS makes a single tree almost as good as a Random Forest with 10 trees. For applications that require more interpretability and efficiency, a single decision tree plus HGS is more preferred. |
format | Online Article Text |
id | pubmed-7206292 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
record_format | MEDLINE/PubMed |
spelling | pubmed-72062922020-05-08 Hierarchical Gradient Smoothing for Probability Estimation Trees Zhang, He Petitjean, François Buntine, Wray Advances in Knowledge Discovery and Data Mining Article Decision trees are still seeing use in online, non-stationary and embedded contexts, as well as for interpretability. For applications like ranking and cost-sensitive classification, probability estimation trees (PETs) are used. These are built using smoothing or calibration techniques. Older smoothing techniques used counts local to a leaf node, but a few more recent techniques consider the broader context of a node when doing estimation. We apply a recent advanced smoothing method called Hierarchical Dirichlet Process (HDP) to PETs, and then propose a novel hierarchical smoothing approach called Hierarchical Gradient Smoothing (HGS) as an alternative. HGS smooths leaf nodes up to all the ancestors, instead of recursively smoothing to the parent used by HDP. HGS is made faster by efficiently optimizing the Leave-One-Out Cross-Validation (LOOCV) loss measure using gradient descent, instead of sampling used in HDP. An extensive set of experiments are conducted on 143 datasets showing that our HGS estimates are not only more accurate but also do so within a fraction of HDP time. Besides, HGS makes a single tree almost as good as a Random Forest with 10 trees. For applications that require more interpretability and efficiency, a single decision tree plus HGS is more preferred. 2020-04-17 /pmc/articles/PMC7206292/ http://dx.doi.org/10.1007/978-3-030-47426-3_18 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Zhang, He Petitjean, François Buntine, Wray Hierarchical Gradient Smoothing for Probability Estimation Trees |
title | Hierarchical Gradient Smoothing for Probability Estimation Trees |
title_full | Hierarchical Gradient Smoothing for Probability Estimation Trees |
title_fullStr | Hierarchical Gradient Smoothing for Probability Estimation Trees |
title_full_unstemmed | Hierarchical Gradient Smoothing for Probability Estimation Trees |
title_short | Hierarchical Gradient Smoothing for Probability Estimation Trees |
title_sort | hierarchical gradient smoothing for probability estimation trees |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7206292/ http://dx.doi.org/10.1007/978-3-030-47426-3_18 |
work_keys_str_mv | AT zhanghe hierarchicalgradientsmoothingforprobabilityestimationtrees AT petitjeanfrancois hierarchicalgradientsmoothingforprobabilityestimationtrees AT buntinewray hierarchicalgradientsmoothingforprobabilityestimationtrees |