Cargando…

Generating highly accurate prediction hypotheses through collaborative ensemble learning

Ensemble generation is a natural and convenient way of achieving better generalization performance of learning algorithms by gathering their predictive capabilities. Here, we nurture the idea of ensemble-based learning by combining bagging and boosting for the purpose of binary classification. Since...

Descripción completa

Detalles Bibliográficos
Autores principales: Arsov, Nino, Pavlovski, Martin, Basnarkov, Lasko, Kocarev, Ljupco
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5356335/
https://www.ncbi.nlm.nih.gov/pubmed/28304378
http://dx.doi.org/10.1038/srep44649
_version_ 1782515811903078400
author Arsov, Nino
Pavlovski, Martin
Basnarkov, Lasko
Kocarev, Ljupco
author_facet Arsov, Nino
Pavlovski, Martin
Basnarkov, Lasko
Kocarev, Ljupco
author_sort Arsov, Nino
collection PubMed
description Ensemble generation is a natural and convenient way of achieving better generalization performance of learning algorithms by gathering their predictive capabilities. Here, we nurture the idea of ensemble-based learning by combining bagging and boosting for the purpose of binary classification. Since the former improves stability through variance reduction, while the latter ameliorates overfitting, the outcome of a multi-model that combines both strives toward a comprehensive net-balancing of the bias-variance trade-off. To further improve this, we alter the bagged-boosting scheme by introducing collaboration between the multi-model’s constituent learners at various levels. This novel stability-guided classification scheme is delivered in two flavours: during or after the boosting process. Applied among a crowd of Gentle Boost ensembles, the ability of the two suggested algorithms to generalize is inspected by comparing them against Subbagging and Gentle Boost on various real-world datasets. In both cases, our models obtained a 40% generalization error decrease. But their true ability to capture details in data was revealed through their application for protein detection in texture analysis of gel electrophoresis images. They achieve improved performance of approximately 0.9773 AUROC when compared to the AUROC of 0.9574 obtained by an SVM based on recursive feature elimination.
format Online
Article
Text
id pubmed-5356335
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Nature Publishing Group
record_format MEDLINE/PubMed
spelling pubmed-53563352017-03-22 Generating highly accurate prediction hypotheses through collaborative ensemble learning Arsov, Nino Pavlovski, Martin Basnarkov, Lasko Kocarev, Ljupco Sci Rep Article Ensemble generation is a natural and convenient way of achieving better generalization performance of learning algorithms by gathering their predictive capabilities. Here, we nurture the idea of ensemble-based learning by combining bagging and boosting for the purpose of binary classification. Since the former improves stability through variance reduction, while the latter ameliorates overfitting, the outcome of a multi-model that combines both strives toward a comprehensive net-balancing of the bias-variance trade-off. To further improve this, we alter the bagged-boosting scheme by introducing collaboration between the multi-model’s constituent learners at various levels. This novel stability-guided classification scheme is delivered in two flavours: during or after the boosting process. Applied among a crowd of Gentle Boost ensembles, the ability of the two suggested algorithms to generalize is inspected by comparing them against Subbagging and Gentle Boost on various real-world datasets. In both cases, our models obtained a 40% generalization error decrease. But their true ability to capture details in data was revealed through their application for protein detection in texture analysis of gel electrophoresis images. They achieve improved performance of approximately 0.9773 AUROC when compared to the AUROC of 0.9574 obtained by an SVM based on recursive feature elimination. Nature Publishing Group 2017-03-17 /pmc/articles/PMC5356335/ /pubmed/28304378 http://dx.doi.org/10.1038/srep44649 Text en Copyright © 2017, The Author(s) http://creativecommons.org/licenses/by/4.0/ This work is licensed under a Creative Commons Attribution 4.0 International License. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in the credit line; if the material is not included under the Creative Commons license, users will need to obtain permission from the license holder to reproduce the material. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/
spellingShingle Article
Arsov, Nino
Pavlovski, Martin
Basnarkov, Lasko
Kocarev, Ljupco
Generating highly accurate prediction hypotheses through collaborative ensemble learning
title Generating highly accurate prediction hypotheses through collaborative ensemble learning
title_full Generating highly accurate prediction hypotheses through collaborative ensemble learning
title_fullStr Generating highly accurate prediction hypotheses through collaborative ensemble learning
title_full_unstemmed Generating highly accurate prediction hypotheses through collaborative ensemble learning
title_short Generating highly accurate prediction hypotheses through collaborative ensemble learning
title_sort generating highly accurate prediction hypotheses through collaborative ensemble learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5356335/
https://www.ncbi.nlm.nih.gov/pubmed/28304378
http://dx.doi.org/10.1038/srep44649
work_keys_str_mv AT arsovnino generatinghighlyaccuratepredictionhypothesesthroughcollaborativeensemblelearning
AT pavlovskimartin generatinghighlyaccuratepredictionhypothesesthroughcollaborativeensemblelearning
AT basnarkovlasko generatinghighlyaccuratepredictionhypothesesthroughcollaborativeensemblelearning
AT kocarevljupco generatinghighlyaccuratepredictionhypothesesthroughcollaborativeensemblelearning