Cargando…

An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification

There are a large number of symptom consultation texts in medical and healthcare Internet communities, and Chinese health segmentation is more complex, which leads to the low accuracy of the existing algorithms for medical text classification. The deep learning model has advantages in extracting abs...

Descripción completa

Detalles Bibliográficos
Autores principales: Liang, Shengbin, Chen, Xinan, Ma, Jixin, Du, Wencai, Ma, Huawei
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7925031/
https://www.ncbi.nlm.nih.gov/pubmed/33688423
http://dx.doi.org/10.1155/2021/6664893
_version_ 1783659207863566336
author Liang, Shengbin
Chen, Xinan
Ma, Jixin
Du, Wencai
Ma, Huawei
author_facet Liang, Shengbin
Chen, Xinan
Ma, Jixin
Du, Wencai
Ma, Huawei
author_sort Liang, Shengbin
collection PubMed
description There are a large number of symptom consultation texts in medical and healthcare Internet communities, and Chinese health segmentation is more complex, which leads to the low accuracy of the existing algorithms for medical text classification. The deep learning model has advantages in extracting abstract features of text effectively. However, for a large number of samples of complex text data, especially for words with ambiguous meanings in the field of Chinese medical diagnosis, the word-level neural network model is insufficient. Therefore, in order to solve the triage and precise treatment of patients, we present an improved Double Channel (DC) mechanism as a significant enhancement to Long Short-Term Memory (LSTM). In this DC mechanism, two channels are used to receive word-level and char-level embedding, respectively, at the same time. Hybrid attention is proposed to combine the current time output with the current time unit state and then using attention to calculate the weight. By calculating the probability distribution of each timestep input data weight, the weight score is obtained, and then weighted summation is performed. At last, the data input by each timestep is subjected to trade-off learning to improve the generalization ability of the model learning. Moreover, we conduct an extensive performance evaluation on two different datasets: cMedQA and Sentiment140. The experimental results show that the DC-LSTM model proposed in this paper has significantly superior accuracy and ROC compared with the basic CNN-LSTM model.
format Online
Article
Text
id pubmed-7925031
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-79250312021-03-08 An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification Liang, Shengbin Chen, Xinan Ma, Jixin Du, Wencai Ma, Huawei J Healthc Eng Research Article There are a large number of symptom consultation texts in medical and healthcare Internet communities, and Chinese health segmentation is more complex, which leads to the low accuracy of the existing algorithms for medical text classification. The deep learning model has advantages in extracting abstract features of text effectively. However, for a large number of samples of complex text data, especially for words with ambiguous meanings in the field of Chinese medical diagnosis, the word-level neural network model is insufficient. Therefore, in order to solve the triage and precise treatment of patients, we present an improved Double Channel (DC) mechanism as a significant enhancement to Long Short-Term Memory (LSTM). In this DC mechanism, two channels are used to receive word-level and char-level embedding, respectively, at the same time. Hybrid attention is proposed to combine the current time output with the current time unit state and then using attention to calculate the weight. By calculating the probability distribution of each timestep input data weight, the weight score is obtained, and then weighted summation is performed. At last, the data input by each timestep is subjected to trade-off learning to improve the generalization ability of the model learning. Moreover, we conduct an extensive performance evaluation on two different datasets: cMedQA and Sentiment140. The experimental results show that the DC-LSTM model proposed in this paper has significantly superior accuracy and ROC compared with the basic CNN-LSTM model. Hindawi 2021-02-23 /pmc/articles/PMC7925031/ /pubmed/33688423 http://dx.doi.org/10.1155/2021/6664893 Text en Copyright © 2021 Shengbin Liang et al. https://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Liang, Shengbin
Chen, Xinan
Ma, Jixin
Du, Wencai
Ma, Huawei
An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification
title An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification
title_full An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification
title_fullStr An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification
title_full_unstemmed An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification
title_short An Improved Double Channel Long Short-Term Memory Model for Medical Text Classification
title_sort improved double channel long short-term memory model for medical text classification
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7925031/
https://www.ncbi.nlm.nih.gov/pubmed/33688423
http://dx.doi.org/10.1155/2021/6664893
work_keys_str_mv AT liangshengbin animproveddoublechannellongshorttermmemorymodelformedicaltextclassification
AT chenxinan animproveddoublechannellongshorttermmemorymodelformedicaltextclassification
AT majixin animproveddoublechannellongshorttermmemorymodelformedicaltextclassification
AT duwencai animproveddoublechannellongshorttermmemorymodelformedicaltextclassification
AT mahuawei animproveddoublechannellongshorttermmemorymodelformedicaltextclassification
AT liangshengbin improveddoublechannellongshorttermmemorymodelformedicaltextclassification
AT chenxinan improveddoublechannellongshorttermmemorymodelformedicaltextclassification
AT majixin improveddoublechannellongshorttermmemorymodelformedicaltextclassification
AT duwencai improveddoublechannellongshorttermmemorymodelformedicaltextclassification
AT mahuawei improveddoublechannellongshorttermmemorymodelformedicaltextclassification