Cargando…

Tree-Weighting for Multi-Study Ensemble Learners

Multi-study learning uses multiple training studies, separately trains classifiers on each, and forms an ensemble with weights rewarding members with better cross-study prediction ability. This article considers novel weighting approaches for constructing tree-based ensemble learners in this setting...

Descripción completa

Detalles Bibliográficos
Autores principales: Ramchandran, Maya, Patil, Prasad, Parmigiani, Giovanni
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6980320/
https://www.ncbi.nlm.nih.gov/pubmed/31797618
_version_ 1783490944714145792
author Ramchandran, Maya
Patil, Prasad
Parmigiani, Giovanni
author_facet Ramchandran, Maya
Patil, Prasad
Parmigiani, Giovanni
author_sort Ramchandran, Maya
collection PubMed
description Multi-study learning uses multiple training studies, separately trains classifiers on each, and forms an ensemble with weights rewarding members with better cross-study prediction ability. This article considers novel weighting approaches for constructing tree-based ensemble learners in this setting. Using Random Forests as a single-study learner, we compare weighting each forest to form the ensemble, to extracting the individual trees trained by each Random Forest and weighting them directly. We find that incorporating multiple layers of ensembling in the training process by weighting trees increases the robustness of the resulting predictor. Furthermore, we explore how ensembling weights correspond to tree structure, to shed light on the features that determine whether weighting trees directly is advantageous. Finally, we apply our approach to genomic datasets and show that weighting trees improves upon the basic multi-study learning paradigm.
format Online
Article
Text
id pubmed-6980320
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-69803202020-01-24 Tree-Weighting for Multi-Study Ensemble Learners Ramchandran, Maya Patil, Prasad Parmigiani, Giovanni Pac Symp Biocomput Article Multi-study learning uses multiple training studies, separately trains classifiers on each, and forms an ensemble with weights rewarding members with better cross-study prediction ability. This article considers novel weighting approaches for constructing tree-based ensemble learners in this setting. Using Random Forests as a single-study learner, we compare weighting each forest to form the ensemble, to extracting the individual trees trained by each Random Forest and weighting them directly. We find that incorporating multiple layers of ensembling in the training process by weighting trees increases the robustness of the resulting predictor. Furthermore, we explore how ensembling weights correspond to tree structure, to shed light on the features that determine whether weighting trees directly is advantageous. Finally, we apply our approach to genomic datasets and show that weighting trees improves upon the basic multi-study learning paradigm. 2020 /pmc/articles/PMC6980320/ /pubmed/31797618 Text en http://creativecommons.org/licenses/by/4.0/ Open Access chapter published by World Scientific Publishing Company and distributed under the terms of the Creative Commons Attribution Non-Commercial (CC BY-NC) 4.0 License.
spellingShingle Article
Ramchandran, Maya
Patil, Prasad
Parmigiani, Giovanni
Tree-Weighting for Multi-Study Ensemble Learners
title Tree-Weighting for Multi-Study Ensemble Learners
title_full Tree-Weighting for Multi-Study Ensemble Learners
title_fullStr Tree-Weighting for Multi-Study Ensemble Learners
title_full_unstemmed Tree-Weighting for Multi-Study Ensemble Learners
title_short Tree-Weighting for Multi-Study Ensemble Learners
title_sort tree-weighting for multi-study ensemble learners
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6980320/
https://www.ncbi.nlm.nih.gov/pubmed/31797618
work_keys_str_mv AT ramchandranmaya treeweightingformultistudyensemblelearners
AT patilprasad treeweightingformultistudyensemblelearners
AT parmigianigiovanni treeweightingformultistudyensemblelearners