Cargando…
Identifying bias in models that detect vocal fold paralysis from audio recordings using explainable machine learning and clinician ratings
INTRODUCTION. Detecting voice disorders from voice recordings could allow for frequent, remote, and low-cost screening before costly clinical visits and a more invasive laryngoscopy examination. Our goals were to detect unilateral vocal fold paralysis (UVFP) from voice recordings using machine learn...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Cold Spring Harbor Laboratory
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7836138/ https://www.ncbi.nlm.nih.gov/pubmed/33501466 http://dx.doi.org/10.1101/2020.11.23.20235945 |
_version_ | 1783642682933903360 |
---|---|
author | Low, Daniel M. Rao, Vishwanatha Randolph, Gregory Song, Phillip C. Ghosh, Satrajit S. |
author_facet | Low, Daniel M. Rao, Vishwanatha Randolph, Gregory Song, Phillip C. Ghosh, Satrajit S. |
author_sort | Low, Daniel M. |
collection | PubMed |
description | INTRODUCTION. Detecting voice disorders from voice recordings could allow for frequent, remote, and low-cost screening before costly clinical visits and a more invasive laryngoscopy examination. Our goals were to detect unilateral vocal fold paralysis (UVFP) from voice recordings using machine learning, to identify which acoustic variables were important for prediction to increase trust, and to determine model performance relative to clinician performance. METHODS. Patients with confirmed UVFP through endoscopic examination (N=77) and controls with normal voices matched for age and sex (N=77) were included. Voice samples were elicited by reading the Rainbow Passage and sustaining phonation of the vowel “a”. Four machine learning models of differing complexity were used. SHapley Additive exPlanations (SHAP) was used to identify important features. RESULTS. The highest median bootstrapped ROC AUC score was 0.87 and beat clinician’s performance (range: 0.74 – 0.81) based on the recordings. Recording durations were different between UVFP recordings and controls due to how that data was originally processed when storing, which we can show can classify both groups. And counterintuitively, many UVFP recordings had higher intensity than controls, when UVFP patients tend to have weaker voices, revealing a dataset-specific bias which we mitigate in an additional analysis. CONCLUSION. We demonstrate that recording biases in audio duration and intensity created dataset-specific differences between patients and controls, which models used to improve classification. Furthermore, clinician’s ratings provide further evidence that patients were over-projecting their voices and being recorded at a higher amplitude signal than controls. Interestingly, after matching audio duration and removing variables associated with intensity in order to mitigate the biases, the models were able to achieve a similar high performance. We provide a set of recommendations to avoid bias when building and evaluating machine learning models for screening in laryngology. |
format | Online Article Text |
id | pubmed-7836138 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Cold Spring Harbor Laboratory |
record_format | MEDLINE/PubMed |
spelling | pubmed-78361382021-01-27 Identifying bias in models that detect vocal fold paralysis from audio recordings using explainable machine learning and clinician ratings Low, Daniel M. Rao, Vishwanatha Randolph, Gregory Song, Phillip C. Ghosh, Satrajit S. medRxiv Article INTRODUCTION. Detecting voice disorders from voice recordings could allow for frequent, remote, and low-cost screening before costly clinical visits and a more invasive laryngoscopy examination. Our goals were to detect unilateral vocal fold paralysis (UVFP) from voice recordings using machine learning, to identify which acoustic variables were important for prediction to increase trust, and to determine model performance relative to clinician performance. METHODS. Patients with confirmed UVFP through endoscopic examination (N=77) and controls with normal voices matched for age and sex (N=77) were included. Voice samples were elicited by reading the Rainbow Passage and sustaining phonation of the vowel “a”. Four machine learning models of differing complexity were used. SHapley Additive exPlanations (SHAP) was used to identify important features. RESULTS. The highest median bootstrapped ROC AUC score was 0.87 and beat clinician’s performance (range: 0.74 – 0.81) based on the recordings. Recording durations were different between UVFP recordings and controls due to how that data was originally processed when storing, which we can show can classify both groups. And counterintuitively, many UVFP recordings had higher intensity than controls, when UVFP patients tend to have weaker voices, revealing a dataset-specific bias which we mitigate in an additional analysis. CONCLUSION. We demonstrate that recording biases in audio duration and intensity created dataset-specific differences between patients and controls, which models used to improve classification. Furthermore, clinician’s ratings provide further evidence that patients were over-projecting their voices and being recorded at a higher amplitude signal than controls. Interestingly, after matching audio duration and removing variables associated with intensity in order to mitigate the biases, the models were able to achieve a similar high performance. We provide a set of recommendations to avoid bias when building and evaluating machine learning models for screening in laryngology. Cold Spring Harbor Laboratory 2023-10-23 /pmc/articles/PMC7836138/ /pubmed/33501466 http://dx.doi.org/10.1101/2020.11.23.20235945 Text en https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (https://creativecommons.org/licenses/by-nc-nd/4.0/) , which allows reusers to copy and distribute the material in any medium or format in unadapted form only, for noncommercial purposes only, and only so long as attribution is given to the creator. |
spellingShingle | Article Low, Daniel M. Rao, Vishwanatha Randolph, Gregory Song, Phillip C. Ghosh, Satrajit S. Identifying bias in models that detect vocal fold paralysis from audio recordings using explainable machine learning and clinician ratings |
title | Identifying bias in models that detect vocal fold paralysis from audio recordings using explainable machine learning and clinician ratings |
title_full | Identifying bias in models that detect vocal fold paralysis from audio recordings using explainable machine learning and clinician ratings |
title_fullStr | Identifying bias in models that detect vocal fold paralysis from audio recordings using explainable machine learning and clinician ratings |
title_full_unstemmed | Identifying bias in models that detect vocal fold paralysis from audio recordings using explainable machine learning and clinician ratings |
title_short | Identifying bias in models that detect vocal fold paralysis from audio recordings using explainable machine learning and clinician ratings |
title_sort | identifying bias in models that detect vocal fold paralysis from audio recordings using explainable machine learning and clinician ratings |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7836138/ https://www.ncbi.nlm.nih.gov/pubmed/33501466 http://dx.doi.org/10.1101/2020.11.23.20235945 |
work_keys_str_mv | AT lowdanielm identifyingbiasinmodelsthatdetectvocalfoldparalysisfromaudiorecordingsusingexplainablemachinelearningandclinicianratings AT raovishwanatha identifyingbiasinmodelsthatdetectvocalfoldparalysisfromaudiorecordingsusingexplainablemachinelearningandclinicianratings AT randolphgregory identifyingbiasinmodelsthatdetectvocalfoldparalysisfromaudiorecordingsusingexplainablemachinelearningandclinicianratings AT songphillipc identifyingbiasinmodelsthatdetectvocalfoldparalysisfromaudiorecordingsusingexplainablemachinelearningandclinicianratings AT ghoshsatrajits identifyingbiasinmodelsthatdetectvocalfoldparalysisfromaudiorecordingsusingexplainablemachinelearningandclinicianratings |