Cargando…

A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models

The International Classification of Diseases (ICD) has an important role in building applications for clinical medicine. Extremely large ICD coding label sets and imbalanced label distribution bring the problem of inconsistency between the local batch data distribution and the global training data d...

Descripción completa

Detalles Bibliográficos
Autores principales: Wang, Yaqiang, Han, Xu, Hao, Xuechao, Zhu, Tao, Shu, Hongping
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9777784/
https://www.ncbi.nlm.nih.gov/pubmed/36553921
http://dx.doi.org/10.3390/healthcare10122397
_version_ 1784856191994691584
author Wang, Yaqiang
Han, Xu
Hao, Xuechao
Zhu, Tao
Shu, Hongping
author_facet Wang, Yaqiang
Han, Xu
Hao, Xuechao
Zhu, Tao
Shu, Hongping
author_sort Wang, Yaqiang
collection PubMed
description The International Classification of Diseases (ICD) has an important role in building applications for clinical medicine. Extremely large ICD coding label sets and imbalanced label distribution bring the problem of inconsistency between the local batch data distribution and the global training data distribution into the minibatch gradient descent (MBGD)-based training procedure for deep multi-label classification models for automatic ICD coding. The problem further leads to an overfitting issue. In order to improve the performance and generalization ability of the deep learning automatic ICD coding model, we proposed a simple and effective curriculum batching strategy in this paper for improving the MBGD-based training procedure. This strategy generates three batch sets offline through applying three predefined sampling algorithms. These batch sets satisfy a uniform data distribution, a shuffling data distribution and the original training data distribution, respectively, and the learning tasks corresponding to these batch sets range from simple to complex. Experiments show that, after replacing the original shuffling algorithm-based batching strategy with the proposed curriculum batching strategy, the performance of the three investigated deep multi-label classification models for automatic ICD coding all have dramatic improvements. At the same time, the models avoid the overfitting issue and all show better ability to learn the long-tailed label information. The performance is also better than a SOTA label set reconstruction model.
format Online
Article
Text
id pubmed-9777784
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-97777842022-12-23 A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models Wang, Yaqiang Han, Xu Hao, Xuechao Zhu, Tao Shu, Hongping Healthcare (Basel) Technical Note The International Classification of Diseases (ICD) has an important role in building applications for clinical medicine. Extremely large ICD coding label sets and imbalanced label distribution bring the problem of inconsistency between the local batch data distribution and the global training data distribution into the minibatch gradient descent (MBGD)-based training procedure for deep multi-label classification models for automatic ICD coding. The problem further leads to an overfitting issue. In order to improve the performance and generalization ability of the deep learning automatic ICD coding model, we proposed a simple and effective curriculum batching strategy in this paper for improving the MBGD-based training procedure. This strategy generates three batch sets offline through applying three predefined sampling algorithms. These batch sets satisfy a uniform data distribution, a shuffling data distribution and the original training data distribution, respectively, and the learning tasks corresponding to these batch sets range from simple to complex. Experiments show that, after replacing the original shuffling algorithm-based batching strategy with the proposed curriculum batching strategy, the performance of the three investigated deep multi-label classification models for automatic ICD coding all have dramatic improvements. At the same time, the models avoid the overfitting issue and all show better ability to learn the long-tailed label information. The performance is also better than a SOTA label set reconstruction model. MDPI 2022-11-29 /pmc/articles/PMC9777784/ /pubmed/36553921 http://dx.doi.org/10.3390/healthcare10122397 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Technical Note
Wang, Yaqiang
Han, Xu
Hao, Xuechao
Zhu, Tao
Shu, Hongping
A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models
title A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models
title_full A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models
title_fullStr A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models
title_full_unstemmed A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models
title_short A Curriculum Batching Strategy for Automatic ICD Coding with Deep Multi-Label Classification Models
title_sort curriculum batching strategy for automatic icd coding with deep multi-label classification models
topic Technical Note
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9777784/
https://www.ncbi.nlm.nih.gov/pubmed/36553921
http://dx.doi.org/10.3390/healthcare10122397
work_keys_str_mv AT wangyaqiang acurriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels
AT hanxu acurriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels
AT haoxuechao acurriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels
AT zhutao acurriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels
AT shuhongping acurriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels
AT wangyaqiang curriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels
AT hanxu curriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels
AT haoxuechao curriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels
AT zhutao curriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels
AT shuhongping curriculumbatchingstrategyforautomaticicdcodingwithdeepmultilabelclassificationmodels