Cargando…
A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models
The International Classification of Diseases (ICD) has an important role in building applications for clinical medicine. Extremely large ICD coding label sets and imbalanced label distribution bring the problem of inconsistency between the local batch data distribution and the global training data d...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9777784/ https://www.ncbi.nlm.nih.gov/pubmed/36553921 http://dx.doi.org/10.3390/healthcare10122397 |
_version_ | 1784856191994691584 |
---|---|
author | Wang, Yaqiang Han, Xu Hao, Xuechao Zhu, Tao Shu, Hongping |
author_facet | Wang, Yaqiang Han, Xu Hao, Xuechao Zhu, Tao Shu, Hongping |
author_sort | Wang, Yaqiang |
collection | PubMed |
description | The International Classification of Diseases (ICD) has an important role in building applications for clinical medicine. Extremely large ICD coding label sets and imbalanced label distribution bring the problem of inconsistency between the local batch data distribution and the global training data distribution into the minibatch gradient descent (MBGD)-based training procedure for deep multi-label classification models for automatic ICD coding. The problem further leads to an overfitting issue. In order to improve the performance and generalization ability of the deep learning automatic ICD coding model, we proposed a simple and effective curriculum batching strategy in this paper for improving the MBGD-based training procedure. This strategy generates three batch sets offline through applying three predefined sampling algorithms. These batch sets satisfy a uniform data distribution, a shuffling data distribution and the original training data distribution, respectively, and the learning tasks corresponding to these batch sets range from simple to complex. Experiments show that, after replacing the original shuffling algorithm-based batching strategy with the proposed curriculum batching strategy, the performance of the three investigated deep multi-label classification models for automatic ICD coding all have dramatic improvements. At the same time, the models avoid the overfitting issue and all show better ability to learn the long-tailed label information. The performance is also better than a SOTA label set reconstruction model. |
format | Online Article Text |
id | pubmed-9777784 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-97777842022-12-23 A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models Wang, Yaqiang Han, Xu Hao, Xuechao Zhu, Tao Shu, Hongping Healthcare (Basel) Technical Note The International Classification of Diseases (ICD) has an important role in building applications for clinical medicine. Extremely large ICD coding label sets and imbalanced label distribution bring the problem of inconsistency between the local batch data distribution and the global training data distribution into the minibatch gradient descent (MBGD)-based training procedure for deep multi-label classification models for automatic ICD coding. The problem further leads to an overfitting issue. In order to improve the performance and generalization ability of the deep learning automatic ICD coding model, we proposed a simple and effective curriculum batching strategy in this paper for improving the MBGD-based training procedure. This strategy generates three batch sets offline through applying three predefined sampling algorithms. These batch sets satisfy a uniform data distribution, a shuffling data distribution and the original training data distribution, respectively, and the learning tasks corresponding to these batch sets range from simple to complex. Experiments show that, after replacing the original shuffling algorithm-based batching strategy with the proposed curriculum batching strategy, the performance of the three investigated deep multi-label classification models for automatic ICD coding all have dramatic improvements. At the same time, the models avoid the overfitting issue and all show better ability to learn the long-tailed label information. The performance is also better than a SOTA label set reconstruction model. MDPI 2022-11-29 /pmc/articles/PMC9777784/ /pubmed/36553921 http://dx.doi.org/10.3390/healthcare10122397 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Technical Note Wang, Yaqiang Han, Xu Hao, Xuechao Zhu, Tao Shu, Hongping A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models |
title | A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models |
title_full | A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models |
title_fullStr | A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models |
title_full_unstemmed | A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models |
title_short | A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models |
title_sort | curriculum batching strategy for automatic icd coding with deep multi-label classification models |
topic | Technical Note |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9777784/ https://www.ncbi.nlm.nih.gov/pubmed/36553921 http://dx.doi.org/10.3390/healthcare10122397 |
work_keys_str_mv | AT wangyaqiang acurriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels AT hanxu acurriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels AT haoxuechao acurriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels AT zhutao acurriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels AT shuhongping acurriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels AT wangyaqiang curriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels AT hanxu curriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels AT haoxuechao curriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels AT zhutao curriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels AT shuhongping curriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels |