Cargando…

Critical assessment of transformer-based AI models for German clinical notes

OBJECTIVE: Healthcare data such as clinical notes are primarily recorded in an unstructured manner. If adequately translated into structured data, they can be utilized for health economics and set the groundwork for better individualized patient care. To structure clinical notes, deep-learning metho...

Descripción completa

Detalles Bibliográficos
Autores principales: Lentzen, Manuel, Madan, Sumit, Lage-Rupprecht, Vanessa, Kühnel, Lisa, Fluck, Juliane, Jacobs, Marc, Mittermaier, Mirja, Witzenrath, Martin, Brunecker, Peter, Hofmann-Apitius, Martin, Weber, Joachim, Fröhlich, Holger
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Oxford University Press 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9663939/
https://www.ncbi.nlm.nih.gov/pubmed/36380848
http://dx.doi.org/10.1093/jamiaopen/ooac087
_version_ 1784830992624648192
author Lentzen, Manuel
Madan, Sumit
Lage-Rupprecht, Vanessa
Kühnel, Lisa
Fluck, Juliane
Jacobs, Marc
Mittermaier, Mirja
Witzenrath, Martin
Brunecker, Peter
Hofmann-Apitius, Martin
Weber, Joachim
Fröhlich, Holger
author_facet Lentzen, Manuel
Madan, Sumit
Lage-Rupprecht, Vanessa
Kühnel, Lisa
Fluck, Juliane
Jacobs, Marc
Mittermaier, Mirja
Witzenrath, Martin
Brunecker, Peter
Hofmann-Apitius, Martin
Weber, Joachim
Fröhlich, Holger
author_sort Lentzen, Manuel
collection PubMed
description OBJECTIVE: Healthcare data such as clinical notes are primarily recorded in an unstructured manner. If adequately translated into structured data, they can be utilized for health economics and set the groundwork for better individualized patient care. To structure clinical notes, deep-learning methods, particularly transformer-based models like Bidirectional Encoder Representations from Transformers (BERT), have recently received much attention. Currently, biomedical applications are primarily focused on the English language. While general-purpose German-language models such as GermanBERT and GottBERT have been published, adaptations for biomedical data are unavailable. This study evaluated the suitability of existing and novel transformer-based models for the German biomedical and clinical domain. MATERIALS AND METHODS: We used 8 transformer-based models and pre-trained 3 new models on a newly generated biomedical corpus, and systematically compared them with each other. We annotated a new dataset of clinical notes and used it with 4 other corpora (BRONCO150, CLEF eHealth 2019 Task 1, GGPONC, and JSynCC) to perform named entity recognition (NER) and document classification tasks. RESULTS: General-purpose language models can be used effectively for biomedical and clinical natural language processing (NLP) tasks, still, our newly trained BioGottBERT model outperformed GottBERT on both clinical NER tasks. However, training new biomedical models from scratch proved ineffective. DISCUSSION: The domain-adaptation strategy’s potential is currently limited due to a lack of pre-training data. Since general-purpose language models are only marginally inferior to domain-specific models, both options are suitable for developing German-language biomedical applications. CONCLUSION: General-purpose language models perform remarkably well on biomedical and clinical NLP tasks. If larger corpora become available in the future, domain-adapting these models may improve performances.
format Online
Article
Text
id pubmed-9663939
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Oxford University Press
record_format MEDLINE/PubMed
spelling pubmed-96639392022-11-14 Critical assessment of transformer-based AI models for German clinical notes Lentzen, Manuel Madan, Sumit Lage-Rupprecht, Vanessa Kühnel, Lisa Fluck, Juliane Jacobs, Marc Mittermaier, Mirja Witzenrath, Martin Brunecker, Peter Hofmann-Apitius, Martin Weber, Joachim Fröhlich, Holger JAMIA Open Research and Applications OBJECTIVE: Healthcare data such as clinical notes are primarily recorded in an unstructured manner. If adequately translated into structured data, they can be utilized for health economics and set the groundwork for better individualized patient care. To structure clinical notes, deep-learning methods, particularly transformer-based models like Bidirectional Encoder Representations from Transformers (BERT), have recently received much attention. Currently, biomedical applications are primarily focused on the English language. While general-purpose German-language models such as GermanBERT and GottBERT have been published, adaptations for biomedical data are unavailable. This study evaluated the suitability of existing and novel transformer-based models for the German biomedical and clinical domain. MATERIALS AND METHODS: We used 8 transformer-based models and pre-trained 3 new models on a newly generated biomedical corpus, and systematically compared them with each other. We annotated a new dataset of clinical notes and used it with 4 other corpora (BRONCO150, CLEF eHealth 2019 Task 1, GGPONC, and JSynCC) to perform named entity recognition (NER) and document classification tasks. RESULTS: General-purpose language models can be used effectively for biomedical and clinical natural language processing (NLP) tasks, still, our newly trained BioGottBERT model outperformed GottBERT on both clinical NER tasks. However, training new biomedical models from scratch proved ineffective. DISCUSSION: The domain-adaptation strategy’s potential is currently limited due to a lack of pre-training data. Since general-purpose language models are only marginally inferior to domain-specific models, both options are suitable for developing German-language biomedical applications. CONCLUSION: General-purpose language models perform remarkably well on biomedical and clinical NLP tasks. If larger corpora become available in the future, domain-adapting these models may improve performances. Oxford University Press 2022-11-15 /pmc/articles/PMC9663939/ /pubmed/36380848 http://dx.doi.org/10.1093/jamiaopen/ooac087 Text en © The Author(s) 2022. Published by Oxford University Press on behalf of the American Medical Informatics Association. https://creativecommons.org/licenses/by/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research and Applications
Lentzen, Manuel
Madan, Sumit
Lage-Rupprecht, Vanessa
Kühnel, Lisa
Fluck, Juliane
Jacobs, Marc
Mittermaier, Mirja
Witzenrath, Martin
Brunecker, Peter
Hofmann-Apitius, Martin
Weber, Joachim
Fröhlich, Holger
Critical assessment of transformer-based AI models for German clinical notes
title Critical assessment of transformer-based AI models for German clinical notes
title_full Critical assessment of transformer-based AI models for German clinical notes
title_fullStr Critical assessment of transformer-based AI models for German clinical notes
title_full_unstemmed Critical assessment of transformer-based AI models for German clinical notes
title_short Critical assessment of transformer-based AI models for German clinical notes
title_sort critical assessment of transformer-based ai models for german clinical notes
topic Research and Applications
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9663939/
https://www.ncbi.nlm.nih.gov/pubmed/36380848
http://dx.doi.org/10.1093/jamiaopen/ooac087
work_keys_str_mv AT lentzenmanuel criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT madansumit criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT lagerupprechtvanessa criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT kuhnellisa criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT fluckjuliane criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT jacobsmarc criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT mittermaiermirja criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT witzenrathmartin criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT bruneckerpeter criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT hofmannapitiusmartin criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT weberjoachim criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes
AT frohlichholger criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes