Cargando…
Generating highly accurate prediction hypotheses through collaborative ensemble learning
Ensemble generation is a natural and convenient way of achieving better generalization performance of learning algorithms by gathering their predictive capabilities. Here, we nurture the idea of ensemble-based learning by combining bagging and boosting for the purpose of binary classification. Since...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5356335/ https://www.ncbi.nlm.nih.gov/pubmed/28304378 http://dx.doi.org/10.1038/srep44649 |
_version_ | 1782515811903078400 |
---|---|
author | Arsov, Nino Pavlovski, Martin Basnarkov, Lasko Kocarev, Ljupco |
author_facet | Arsov, Nino Pavlovski, Martin Basnarkov, Lasko Kocarev, Ljupco |
author_sort | Arsov, Nino |
collection | PubMed |
description | Ensemble generation is a natural and convenient way of achieving better generalization performance of learning algorithms by gathering their predictive capabilities. Here, we nurture the idea of ensemble-based learning by combining bagging and boosting for the purpose of binary classification. Since the former improves stability through variance reduction, while the latter ameliorates overfitting, the outcome of a multi-model that combines both strives toward a comprehensive net-balancing of the bias-variance trade-off. To further improve this, we alter the bagged-boosting scheme by introducing collaboration between the multi-model’s constituent learners at various levels. This novel stability-guided classification scheme is delivered in two flavours: during or after the boosting process. Applied among a crowd of Gentle Boost ensembles, the ability of the two suggested algorithms to generalize is inspected by comparing them against Subbagging and Gentle Boost on various real-world datasets. In both cases, our models obtained a 40% generalization error decrease. But their true ability to capture details in data was revealed through their application for protein detection in texture analysis of gel electrophoresis images. They achieve improved performance of approximately 0.9773 AUROC when compared to the AUROC of 0.9574 obtained by an SVM based on recursive feature elimination. |
format | Online Article Text |
id | pubmed-5356335 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2017 |
publisher | Nature Publishing Group |
record_format | MEDLINE/PubMed |
spelling | pubmed-53563352017-03-22 Generating highly accurate prediction hypotheses through collaborative ensemble learning Arsov, Nino Pavlovski, Martin Basnarkov, Lasko Kocarev, Ljupco Sci Rep Article Ensemble generation is a natural and convenient way of achieving better generalization performance of learning algorithms by gathering their predictive capabilities. Here, we nurture the idea of ensemble-based learning by combining bagging and boosting for the purpose of binary classification. Since the former improves stability through variance reduction, while the latter ameliorates overfitting, the outcome of a multi-model that combines both strives toward a comprehensive net-balancing of the bias-variance trade-off. To further improve this, we alter the bagged-boosting scheme by introducing collaboration between the multi-model’s constituent learners at various levels. This novel stability-guided classification scheme is delivered in two flavours: during or after the boosting process. Applied among a crowd of Gentle Boost ensembles, the ability of the two suggested algorithms to generalize is inspected by comparing them against Subbagging and Gentle Boost on various real-world datasets. In both cases, our models obtained a 40% generalization error decrease. But their true ability to capture details in data was revealed through their application for protein detection in texture analysis of gel electrophoresis images. They achieve improved performance of approximately 0.9773 AUROC when compared to the AUROC of 0.9574 obtained by an SVM based on recursive feature elimination. Nature Publishing Group 2017-03-17 /pmc/articles/PMC5356335/ /pubmed/28304378 http://dx.doi.org/10.1038/srep44649 Text en Copyright © 2017, The Author(s) http://creativecommons.org/licenses/by/4.0/ This work is licensed under a Creative Commons Attribution 4.0 International License. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in the credit line; if the material is not included under the Creative Commons license, users will need to obtain permission from the license holder to reproduce the material. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ |
spellingShingle | Article Arsov, Nino Pavlovski, Martin Basnarkov, Lasko Kocarev, Ljupco Generating highly accurate prediction hypotheses through collaborative ensemble learning |
title | Generating highly accurate prediction hypotheses through collaborative ensemble learning |
title_full | Generating highly accurate prediction hypotheses through collaborative ensemble learning |
title_fullStr | Generating highly accurate prediction hypotheses through collaborative ensemble learning |
title_full_unstemmed | Generating highly accurate prediction hypotheses through collaborative ensemble learning |
title_short | Generating highly accurate prediction hypotheses through collaborative ensemble learning |
title_sort | generating highly accurate prediction hypotheses through collaborative ensemble learning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5356335/ https://www.ncbi.nlm.nih.gov/pubmed/28304378 http://dx.doi.org/10.1038/srep44649 |
work_keys_str_mv | AT arsovnino generatinghighlyaccuratepredictionhypothesesthroughcollaborativeensemblelearning AT pavlovskimartin generatinghighlyaccuratepredictionhypothesesthroughcollaborativeensemblelearning AT basnarkovlasko generatinghighlyaccuratepredictionhypothesesthroughcollaborativeensemblelearning AT kocarevljupco generatinghighlyaccuratepredictionhypothesesthroughcollaborativeensemblelearning |