Cargando…

Comparison of different feature extraction methods for applicable automated ICD coding

BACKGROUND: Automated ICD coding on medical texts via machine learning has been a hot topic. Related studies from medical field heavily relies on conventional bag-of-words (BoW) as the feature extraction method, and do not commonly use more complicated methods, such as word2vec (W2V) and large pretr...

Descripción completa

Detalles Bibliográficos
Autores principales: Shuai, Zhao, Xiaolin, Diao, Jing, Yuan, Yanni, Huo, Meng, Cui, Yuxin, Wang, Wei, Zhao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: BioMed Central 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8756659/
https://www.ncbi.nlm.nih.gov/pubmed/35022039
http://dx.doi.org/10.1186/s12911-022-01753-5
_version_ 1784632606130700288
author Shuai, Zhao
Xiaolin, Diao
Jing, Yuan
Yanni, Huo
Meng, Cui
Yuxin, Wang
Wei, Zhao
author_facet Shuai, Zhao
Xiaolin, Diao
Jing, Yuan
Yanni, Huo
Meng, Cui
Yuxin, Wang
Wei, Zhao
author_sort Shuai, Zhao
collection PubMed
description BACKGROUND: Automated ICD coding on medical texts via machine learning has been a hot topic. Related studies from medical field heavily relies on conventional bag-of-words (BoW) as the feature extraction method, and do not commonly use more complicated methods, such as word2vec (W2V) and large pretrained models like BERT. This study aimed at uncovering the most effective feature extraction methods for coding models by comparing BoW, W2V and BERT variants. METHODS: We experimented with a Chinese dataset from Fuwai Hospital, which contains 6947 records and 1532 unique ICD codes, and a public Spanish dataset, which contains 1000 records and 2557 unique ICD codes. We designed coding tasks with different code frequency thresholds (denoted as [Formula: see text] ), with a lower threshold indicating a more complex task. Using traditional classifiers, we compared BoW, W2V and BERT variants on accomplishing these coding tasks. RESULTS: When [Formula: see text] was equal to or greater than 140 for Fuwai dataset, and 60 for the Spanish dataset, the BERT variants with the whole network fine-tuned was the best method, leading to a Micro-F1 of 93.9% for Fuwai data when [Formula: see text] , and a Micro-F1 of 85.41% for the Spanish dataset when [Formula: see text] . When [Formula: see text] fell below 140 for Fuwai dataset, and 60 for the Spanish dataset, BoW turned out to be the best, leading to a Micro-F1 of 83% for Fuwai dataset when [Formula: see text] , and a Micro-F1 of 39.1% for the Spanish dataset when [Formula: see text] . Our experiments also showed that both the BERT variants and BoW possessed good interpretability, which is important for medical applications of coding models. CONCLUSIONS: This study shed light on building promising machine learning models for automated ICD coding by revealing the most effective feature extraction methods. Concretely, our results indicated that fine-tuning the whole network of the BERT variants was the optimal method for tasks covering only frequent codes, especially codes that represented unspecified diseases, while BoW was the best for tasks involving both frequent and infrequent codes. The frequency threshold where the best-performing method varied differed between different datasets due to factors like language and codeset. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1186/s12911-022-01753-5.
format Online
Article
Text
id pubmed-8756659
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher BioMed Central
record_format MEDLINE/PubMed
spelling pubmed-87566592022-01-18 Comparison of different feature extraction methods for applicable automated ICD coding Shuai, Zhao Xiaolin, Diao Jing, Yuan Yanni, Huo Meng, Cui Yuxin, Wang Wei, Zhao BMC Med Inform Decis Mak Research BACKGROUND: Automated ICD coding on medical texts via machine learning has been a hot topic. Related studies from medical field heavily relies on conventional bag-of-words (BoW) as the feature extraction method, and do not commonly use more complicated methods, such as word2vec (W2V) and large pretrained models like BERT. This study aimed at uncovering the most effective feature extraction methods for coding models by comparing BoW, W2V and BERT variants. METHODS: We experimented with a Chinese dataset from Fuwai Hospital, which contains 6947 records and 1532 unique ICD codes, and a public Spanish dataset, which contains 1000 records and 2557 unique ICD codes. We designed coding tasks with different code frequency thresholds (denoted as [Formula: see text] ), with a lower threshold indicating a more complex task. Using traditional classifiers, we compared BoW, W2V and BERT variants on accomplishing these coding tasks. RESULTS: When [Formula: see text] was equal to or greater than 140 for Fuwai dataset, and 60 for the Spanish dataset, the BERT variants with the whole network fine-tuned was the best method, leading to a Micro-F1 of 93.9% for Fuwai data when [Formula: see text] , and a Micro-F1 of 85.41% for the Spanish dataset when [Formula: see text] . When [Formula: see text] fell below 140 for Fuwai dataset, and 60 for the Spanish dataset, BoW turned out to be the best, leading to a Micro-F1 of 83% for Fuwai dataset when [Formula: see text] , and a Micro-F1 of 39.1% for the Spanish dataset when [Formula: see text] . Our experiments also showed that both the BERT variants and BoW possessed good interpretability, which is important for medical applications of coding models. CONCLUSIONS: This study shed light on building promising machine learning models for automated ICD coding by revealing the most effective feature extraction methods. Concretely, our results indicated that fine-tuning the whole network of the BERT variants was the optimal method for tasks covering only frequent codes, especially codes that represented unspecified diseases, while BoW was the best for tasks involving both frequent and infrequent codes. The frequency threshold where the best-performing method varied differed between different datasets due to factors like language and codeset. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1186/s12911-022-01753-5. BioMed Central 2022-01-12 /pmc/articles/PMC8756659/ /pubmed/35022039 http://dx.doi.org/10.1186/s12911-022-01753-5 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/ (https://creativecommons.org/publicdomain/zero/1.0/) ) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
spellingShingle Research
Shuai, Zhao
Xiaolin, Diao
Jing, Yuan
Yanni, Huo
Meng, Cui
Yuxin, Wang
Wei, Zhao
Comparison of different feature extraction methods for applicable automated ICD coding
title Comparison of different feature extraction methods for applicable automated ICD coding
title_full Comparison of different feature extraction methods for applicable automated ICD coding
title_fullStr Comparison of different feature extraction methods for applicable automated ICD coding
title_full_unstemmed Comparison of different feature extraction methods for applicable automated ICD coding
title_short Comparison of different feature extraction methods for applicable automated ICD coding
title_sort comparison of different feature extraction methods for applicable automated icd coding
topic Research
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8756659/
https://www.ncbi.nlm.nih.gov/pubmed/35022039
http://dx.doi.org/10.1186/s12911-022-01753-5
work_keys_str_mv AT shuaizhao comparisonofdifferentfeatureextractionmethodsforapplicableautomatedicdcoding
AT xiaolindiao comparisonofdifferentfeatureextractionmethodsforapplicableautomatedicdcoding
AT jingyuan comparisonofdifferentfeatureextractionmethodsforapplicableautomatedicdcoding
AT yannihuo comparisonofdifferentfeatureextractionmethodsforapplicableautomatedicdcoding
AT mengcui comparisonofdifferentfeatureextractionmethodsforapplicableautomatedicdcoding
AT yuxinwang comparisonofdifferentfeatureextractionmethodsforapplicableautomatedicdcoding
AT weizhao comparisonofdifferentfeatureextractionmethodsforapplicableautomatedicdcoding