Cargando…
Adapting and evaluating a deep learning language model for clinical why-question answering
OBJECTIVES: To adapt and evaluate a deep learning language model for answering why-questions based on patient-specific clinical text. MATERIALS AND METHODS: Bidirectional encoder representations from transformers (BERT) models were trained with varying data sources to perform SQuAD 2.0 style why-que...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Oxford University Press
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7309262/ https://www.ncbi.nlm.nih.gov/pubmed/32607483 http://dx.doi.org/10.1093/jamiaopen/ooz072 |
_version_ | 1783549177559515136 |
---|---|
author | Wen, Andrew Elwazir, Mohamed Y Moon, Sungrim Fan, Jungwei |
author_facet | Wen, Andrew Elwazir, Mohamed Y Moon, Sungrim Fan, Jungwei |
author_sort | Wen, Andrew |
collection | PubMed |
description | OBJECTIVES: To adapt and evaluate a deep learning language model for answering why-questions based on patient-specific clinical text. MATERIALS AND METHODS: Bidirectional encoder representations from transformers (BERT) models were trained with varying data sources to perform SQuAD 2.0 style why-question answering (why-QA) on clinical notes. The evaluation focused on: (1) comparing the merits from different training data and (2) error analysis. RESULTS: The best model achieved an accuracy of 0.707 (or 0.760 by partial match). Training toward customization for the clinical language helped increase 6% in accuracy. DISCUSSION: The error analysis suggested that the model did not really perform deep reasoning and that clinical why-QA might warrant more sophisticated solutions. CONCLUSION: The BERT model achieved moderate accuracy in clinical why-QA and should benefit from the rapidly evolving technology. Despite the identified limitations, it could serve as a competent proxy for question-driven clinical information extraction. |
format | Online Article Text |
id | pubmed-7309262 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Oxford University Press |
record_format | MEDLINE/PubMed |
spelling | pubmed-73092622020-06-29 Adapting and evaluating a deep learning language model for clinical why-question answering Wen, Andrew Elwazir, Mohamed Y Moon, Sungrim Fan, Jungwei JAMIA Open Brief Communication OBJECTIVES: To adapt and evaluate a deep learning language model for answering why-questions based on patient-specific clinical text. MATERIALS AND METHODS: Bidirectional encoder representations from transformers (BERT) models were trained with varying data sources to perform SQuAD 2.0 style why-question answering (why-QA) on clinical notes. The evaluation focused on: (1) comparing the merits from different training data and (2) error analysis. RESULTS: The best model achieved an accuracy of 0.707 (or 0.760 by partial match). Training toward customization for the clinical language helped increase 6% in accuracy. DISCUSSION: The error analysis suggested that the model did not really perform deep reasoning and that clinical why-QA might warrant more sophisticated solutions. CONCLUSION: The BERT model achieved moderate accuracy in clinical why-QA and should benefit from the rapidly evolving technology. Despite the identified limitations, it could serve as a competent proxy for question-driven clinical information extraction. Oxford University Press 2020-02-04 /pmc/articles/PMC7309262/ /pubmed/32607483 http://dx.doi.org/10.1093/jamiaopen/ooz072 Text en © The Author(s) 2020. Published by Oxford University Press on behalf of the American Medical Informatics Association. http://creativecommons.org/licenses/by-nc/4.0/ This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact journals.permissions@oup.com |
spellingShingle | Brief Communication Wen, Andrew Elwazir, Mohamed Y Moon, Sungrim Fan, Jungwei Adapting and evaluating a deep learning language model for clinical why-question answering |
title | Adapting and evaluating a deep learning language model for clinical why-question answering |
title_full | Adapting and evaluating a deep learning language model for clinical why-question answering |
title_fullStr | Adapting and evaluating a deep learning language model for clinical why-question answering |
title_full_unstemmed | Adapting and evaluating a deep learning language model for clinical why-question answering |
title_short | Adapting and evaluating a deep learning language model for clinical why-question answering |
title_sort | adapting and evaluating a deep learning language model for clinical why-question answering |
topic | Brief Communication |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7309262/ https://www.ncbi.nlm.nih.gov/pubmed/32607483 http://dx.doi.org/10.1093/jamiaopen/ooz072 |
work_keys_str_mv | AT wenandrew adaptingandevaluatingadeeplearninglanguagemodelforclinicalwhyquestionanswering AT elwazirmohamedy adaptingandevaluatingadeeplearninglanguagemodelforclinicalwhyquestionanswering AT moonsungrim adaptingandevaluatingadeeplearninglanguagemodelforclinicalwhyquestionanswering AT fanjungwei adaptingandevaluatingadeeplearninglanguagemodelforclinicalwhyquestionanswering |