Cargando…
Critical assessment of transformer-based AI models for German clinical notes
OBJECTIVE: Healthcare data such as clinical notes are primarily recorded in an unstructured manner. If adequately translated into structured data, they can be utilized for health economics and set the groundwork for better individualized patient care. To structure clinical notes, deep-learning metho...
Autores principales: | , , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Oxford University Press
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9663939/ https://www.ncbi.nlm.nih.gov/pubmed/36380848 http://dx.doi.org/10.1093/jamiaopen/ooac087 |
_version_ | 1784830992624648192 |
---|---|
author | Lentzen, Manuel Madan, Sumit Lage-Rupprecht, Vanessa Kühnel, Lisa Fluck, Juliane Jacobs, Marc Mittermaier, Mirja Witzenrath, Martin Brunecker, Peter Hofmann-Apitius, Martin Weber, Joachim Fröhlich, Holger |
author_facet | Lentzen, Manuel Madan, Sumit Lage-Rupprecht, Vanessa Kühnel, Lisa Fluck, Juliane Jacobs, Marc Mittermaier, Mirja Witzenrath, Martin Brunecker, Peter Hofmann-Apitius, Martin Weber, Joachim Fröhlich, Holger |
author_sort | Lentzen, Manuel |
collection | PubMed |
description | OBJECTIVE: Healthcare data such as clinical notes are primarily recorded in an unstructured manner. If adequately translated into structured data, they can be utilized for health economics and set the groundwork for better individualized patient care. To structure clinical notes, deep-learning methods, particularly transformer-based models like Bidirectional Encoder Representations from Transformers (BERT), have recently received much attention. Currently, biomedical applications are primarily focused on the English language. While general-purpose German-language models such as GermanBERT and GottBERT have been published, adaptations for biomedical data are unavailable. This study evaluated the suitability of existing and novel transformer-based models for the German biomedical and clinical domain. MATERIALS AND METHODS: We used 8 transformer-based models and pre-trained 3 new models on a newly generated biomedical corpus, and systematically compared them with each other. We annotated a new dataset of clinical notes and used it with 4 other corpora (BRONCO150, CLEF eHealth 2019 Task 1, GGPONC, and JSynCC) to perform named entity recognition (NER) and document classification tasks. RESULTS: General-purpose language models can be used effectively for biomedical and clinical natural language processing (NLP) tasks, still, our newly trained BioGottBERT model outperformed GottBERT on both clinical NER tasks. However, training new biomedical models from scratch proved ineffective. DISCUSSION: The domain-adaptation strategy’s potential is currently limited due to a lack of pre-training data. Since general-purpose language models are only marginally inferior to domain-specific models, both options are suitable for developing German-language biomedical applications. CONCLUSION: General-purpose language models perform remarkably well on biomedical and clinical NLP tasks. If larger corpora become available in the future, domain-adapting these models may improve performances. |
format | Online Article Text |
id | pubmed-9663939 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Oxford University Press |
record_format | MEDLINE/PubMed |
spelling | pubmed-96639392022-11-14 Critical assessment of transformer-based AI models for German clinical notes Lentzen, Manuel Madan, Sumit Lage-Rupprecht, Vanessa Kühnel, Lisa Fluck, Juliane Jacobs, Marc Mittermaier, Mirja Witzenrath, Martin Brunecker, Peter Hofmann-Apitius, Martin Weber, Joachim Fröhlich, Holger JAMIA Open Research and Applications OBJECTIVE: Healthcare data such as clinical notes are primarily recorded in an unstructured manner. If adequately translated into structured data, they can be utilized for health economics and set the groundwork for better individualized patient care. To structure clinical notes, deep-learning methods, particularly transformer-based models like Bidirectional Encoder Representations from Transformers (BERT), have recently received much attention. Currently, biomedical applications are primarily focused on the English language. While general-purpose German-language models such as GermanBERT and GottBERT have been published, adaptations for biomedical data are unavailable. This study evaluated the suitability of existing and novel transformer-based models for the German biomedical and clinical domain. MATERIALS AND METHODS: We used 8 transformer-based models and pre-trained 3 new models on a newly generated biomedical corpus, and systematically compared them with each other. We annotated a new dataset of clinical notes and used it with 4 other corpora (BRONCO150, CLEF eHealth 2019 Task 1, GGPONC, and JSynCC) to perform named entity recognition (NER) and document classification tasks. RESULTS: General-purpose language models can be used effectively for biomedical and clinical natural language processing (NLP) tasks, still, our newly trained BioGottBERT model outperformed GottBERT on both clinical NER tasks. However, training new biomedical models from scratch proved ineffective. DISCUSSION: The domain-adaptation strategy’s potential is currently limited due to a lack of pre-training data. Since general-purpose language models are only marginally inferior to domain-specific models, both options are suitable for developing German-language biomedical applications. CONCLUSION: General-purpose language models perform remarkably well on biomedical and clinical NLP tasks. If larger corpora become available in the future, domain-adapting these models may improve performances. Oxford University Press 2022-11-15 /pmc/articles/PMC9663939/ /pubmed/36380848 http://dx.doi.org/10.1093/jamiaopen/ooac087 Text en © The Author(s) 2022. Published by Oxford University Press on behalf of the American Medical Informatics Association. https://creativecommons.org/licenses/by/4.0/This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research and Applications Lentzen, Manuel Madan, Sumit Lage-Rupprecht, Vanessa Kühnel, Lisa Fluck, Juliane Jacobs, Marc Mittermaier, Mirja Witzenrath, Martin Brunecker, Peter Hofmann-Apitius, Martin Weber, Joachim Fröhlich, Holger Critical assessment of transformer-based AI models for German clinical notes |
title | Critical assessment of transformer-based AI models for German clinical notes |
title_full | Critical assessment of transformer-based AI models for German clinical notes |
title_fullStr | Critical assessment of transformer-based AI models for German clinical notes |
title_full_unstemmed | Critical assessment of transformer-based AI models for German clinical notes |
title_short | Critical assessment of transformer-based AI models for German clinical notes |
title_sort | critical assessment of transformer-based ai models for german clinical notes |
topic | Research and Applications |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9663939/ https://www.ncbi.nlm.nih.gov/pubmed/36380848 http://dx.doi.org/10.1093/jamiaopen/ooac087 |
work_keys_str_mv | AT lentzenmanuel criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT madansumit criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT lagerupprechtvanessa criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT kuhnellisa criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT fluckjuliane criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT jacobsmarc criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT mittermaiermirja criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT witzenrathmartin criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT bruneckerpeter criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT hofmannapitiusmartin criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT weberjoachim criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes AT frohlichholger criticalassessmentoftransformerbasedaimodelsforgermanclinicalnotes |