Cargando…
An investigation of how normalisation and local modelling techniques confound machine learning performance in a mental health study
Machine learning (ML) is increasingly deployed on biomedical studies for biomarker development (feature selection) and diagnostic/prognostic technologies (classification). While different ML techniques produce different feature sets and classification performances, less understood is how upstream da...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9156999/ https://www.ncbi.nlm.nih.gov/pubmed/35663731 http://dx.doi.org/10.1016/j.heliyon.2022.e09502 |
_version_ | 1784718553770885120 |
---|---|
author | Zhang, Xinxin Lee, Jimmy Goh, Wilson Wen Bin |
author_facet | Zhang, Xinxin Lee, Jimmy Goh, Wilson Wen Bin |
author_sort | Zhang, Xinxin |
collection | PubMed |
description | Machine learning (ML) is increasingly deployed on biomedical studies for biomarker development (feature selection) and diagnostic/prognostic technologies (classification). While different ML techniques produce different feature sets and classification performances, less understood is how upstream data processing methods (e.g., normalisation) impact downstream analyses. Using a clinical mental health dataset, we investigated the impact of different normalisation techniques on classification model performance. Gene Fuzzy Scoring (GFS), an in-house developed normalisation technique, is compared against widely used normalisation methods such as global quantile normalisation, class-specific quantile normalisation and surrogate variable analysis. We report that choice of normalisation technique has strong influence on feature selection. with GFS outperforming other techniques. Although GFS parameters are tuneable, good classification model performance (ROC-AUC > 0.90) is observed regardless of the GFS parameter settings. We also contrasted our results against local modelling, which is meant to improve the resolution and meaningfulness of classification models built on heterogeneous data. Local models, when derived from non-biologically meaningful subpopulations, perform worse than global models. A deep dive however, revealed that the factors driving cluster formation has little to do with the phenotype-of-interest. This finding is critical, as local models are often seen as a superior means of clinical data modelling. We advise against such naivete. Additionally, we have developed a combinatorial reasoning approach using both global and local paradigms: This helped reveal potential data quality issues or underlying factors causing data heterogeneity that are often overlooked. It also assists to explain the model as well as provides directions for further improvement. |
format | Online Article Text |
id | pubmed-9156999 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Elsevier |
record_format | MEDLINE/PubMed |
spelling | pubmed-91569992022-06-02 An investigation of how normalisation and local modelling techniques confound machine learning performance in a mental health study Zhang, Xinxin Lee, Jimmy Goh, Wilson Wen Bin Heliyon Research Article Machine learning (ML) is increasingly deployed on biomedical studies for biomarker development (feature selection) and diagnostic/prognostic technologies (classification). While different ML techniques produce different feature sets and classification performances, less understood is how upstream data processing methods (e.g., normalisation) impact downstream analyses. Using a clinical mental health dataset, we investigated the impact of different normalisation techniques on classification model performance. Gene Fuzzy Scoring (GFS), an in-house developed normalisation technique, is compared against widely used normalisation methods such as global quantile normalisation, class-specific quantile normalisation and surrogate variable analysis. We report that choice of normalisation technique has strong influence on feature selection. with GFS outperforming other techniques. Although GFS parameters are tuneable, good classification model performance (ROC-AUC > 0.90) is observed regardless of the GFS parameter settings. We also contrasted our results against local modelling, which is meant to improve the resolution and meaningfulness of classification models built on heterogeneous data. Local models, when derived from non-biologically meaningful subpopulations, perform worse than global models. A deep dive however, revealed that the factors driving cluster formation has little to do with the phenotype-of-interest. This finding is critical, as local models are often seen as a superior means of clinical data modelling. We advise against such naivete. Additionally, we have developed a combinatorial reasoning approach using both global and local paradigms: This helped reveal potential data quality issues or underlying factors causing data heterogeneity that are often overlooked. It also assists to explain the model as well as provides directions for further improvement. Elsevier 2022-05-21 /pmc/articles/PMC9156999/ /pubmed/35663731 http://dx.doi.org/10.1016/j.heliyon.2022.e09502 Text en © 2022 The Author(s) https://creativecommons.org/licenses/by-nc-nd/4.0/This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). |
spellingShingle | Research Article Zhang, Xinxin Lee, Jimmy Goh, Wilson Wen Bin An investigation of how normalisation and local modelling techniques confound machine learning performance in a mental health study |
title | An investigation of how normalisation and local modelling techniques confound machine learning performance in a mental health study |
title_full | An investigation of how normalisation and local modelling techniques confound machine learning performance in a mental health study |
title_fullStr | An investigation of how normalisation and local modelling techniques confound machine learning performance in a mental health study |
title_full_unstemmed | An investigation of how normalisation and local modelling techniques confound machine learning performance in a mental health study |
title_short | An investigation of how normalisation and local modelling techniques confound machine learning performance in a mental health study |
title_sort | investigation of how normalisation and local modelling techniques confound machine learning performance in a mental health study |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9156999/ https://www.ncbi.nlm.nih.gov/pubmed/35663731 http://dx.doi.org/10.1016/j.heliyon.2022.e09502 |
work_keys_str_mv | AT zhangxinxin aninvestigationofhownormalisationandlocalmodellingtechniquesconfoundmachinelearningperformanceinamentalhealthstudy AT leejimmy aninvestigationofhownormalisationandlocalmodellingtechniquesconfoundmachinelearningperformanceinamentalhealthstudy AT gohwilsonwenbin aninvestigationofhownormalisationandlocalmodellingtechniquesconfoundmachinelearningperformanceinamentalhealthstudy AT zhangxinxin investigationofhownormalisationandlocalmodellingtechniquesconfoundmachinelearningperformanceinamentalhealthstudy AT leejimmy investigationofhownormalisationandlocalmodellingtechniquesconfoundmachinelearningperformanceinamentalhealthstudy AT gohwilsonwenbin investigationofhownormalisationandlocalmodellingtechniquesconfoundmachinelearningperformanceinamentalhealthstudy |