Cargando…
An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification
There are a large number of symptom consultation texts in medical and healthcare Internet communities, and Chinese health segmentation is more complex, which leads to the low accuracy of the existing algorithms for medical text classification. The deep learning model has advantages in extracting abs...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7925031/ https://www.ncbi.nlm.nih.gov/pubmed/33688423 http://dx.doi.org/10.1155/2021/6664893 |
_version_ | 1783659207863566336 |
---|---|
author | Liang, Shengbin Chen, Xinan Ma, Jixin Du, Wencai Ma, Huawei |
author_facet | Liang, Shengbin Chen, Xinan Ma, Jixin Du, Wencai Ma, Huawei |
author_sort | Liang, Shengbin |
collection | PubMed |
description | There are a large number of symptom consultation texts in medical and healthcare Internet communities, and Chinese health segmentation is more complex, which leads to the low accuracy of the existing algorithms for medical text classification. The deep learning model has advantages in extracting abstract features of text effectively. However, for a large number of samples of complex text data, especially for words with ambiguous meanings in the field of Chinese medical diagnosis, the word-level neural network model is insufficient. Therefore, in order to solve the triage and precise treatment of patients, we present an improved Double Channel (DC) mechanism as a significant enhancement to Long Short-Term Memory (LSTM). In this DC mechanism, two channels are used to receive word-level and char-level embedding, respectively, at the same time. Hybrid attention is proposed to combine the current time output with the current time unit state and then using attention to calculate the weight. By calculating the probability distribution of each timestep input data weight, the weight score is obtained, and then weighted summation is performed. At last, the data input by each timestep is subjected to trade-off learning to improve the generalization ability of the model learning. Moreover, we conduct an extensive performance evaluation on two different datasets: cMedQA and Sentiment140. The experimental results show that the DC-LSTM model proposed in this paper has significantly superior accuracy and ROC compared with the basic CNN-LSTM model. |
format | Online Article Text |
id | pubmed-7925031 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-79250312021-03-08 An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification Liang, Shengbin Chen, Xinan Ma, Jixin Du, Wencai Ma, Huawei J Healthc Eng Research Article There are a large number of symptom consultation texts in medical and healthcare Internet communities, and Chinese health segmentation is more complex, which leads to the low accuracy of the existing algorithms for medical text classification. The deep learning model has advantages in extracting abstract features of text effectively. However, for a large number of samples of complex text data, especially for words with ambiguous meanings in the field of Chinese medical diagnosis, the word-level neural network model is insufficient. Therefore, in order to solve the triage and precise treatment of patients, we present an improved Double Channel (DC) mechanism as a significant enhancement to Long Short-Term Memory (LSTM). In this DC mechanism, two channels are used to receive word-level and char-level embedding, respectively, at the same time. Hybrid attention is proposed to combine the current time output with the current time unit state and then using attention to calculate the weight. By calculating the probability distribution of each timestep input data weight, the weight score is obtained, and then weighted summation is performed. At last, the data input by each timestep is subjected to trade-off learning to improve the generalization ability of the model learning. Moreover, we conduct an extensive performance evaluation on two different datasets: cMedQA and Sentiment140. The experimental results show that the DC-LSTM model proposed in this paper has significantly superior accuracy and ROC compared with the basic CNN-LSTM model. Hindawi 2021-02-23 /pmc/articles/PMC7925031/ /pubmed/33688423 http://dx.doi.org/10.1155/2021/6664893 Text en Copyright © 2021 Shengbin Liang et al. https://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Liang, Shengbin Chen, Xinan Ma, Jixin Du, Wencai Ma, Huawei An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification |
title | An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification |
title_full | An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification |
title_fullStr | An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification |
title_full_unstemmed | An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification |
title_short | An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification |
title_sort | improved double channel long short-term memory model for medical text classification |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7925031/ https://www.ncbi.nlm.nih.gov/pubmed/33688423 http://dx.doi.org/10.1155/2021/6664893 |
work_keys_str_mv | AT liangshengbin animproveddoublechannellongshorttermmemorymodelformedicaltextclassification AT chenxinan animproveddoublechannellongshorttermmemorymodelformedicaltextclassification AT majixin animproveddoublechannellongshorttermmemorymodelformedicaltextclassification AT duwencai animproveddoublechannellongshorttermmemorymodelformedicaltextclassification AT mahuawei animproveddoublechannellongshorttermmemorymodelformedicaltextclassification AT liangshengbin improveddoublechannellongshorttermmemorymodelformedicaltextclassification AT chenxinan improveddoublechannellongshorttermmemorymodelformedicaltextclassification AT majixin improveddoublechannellongshorttermmemorymodelformedicaltextclassification AT duwencai improveddoublechannellongshorttermmemorymodelformedicaltextclassification AT mahuawei improveddoublechannellongshorttermmemorymodelformedicaltextclassification |