Cargando…

Deep Multi-Objective Learning from Low-Dose CT for Automatic Lung-RADS Report Generation

Radiology report generation through chest radiography interpretation is a time-consuming task that involves the interpretation of images by expert radiologists. It is common for fatigue-induced diagnostic error to occur, and especially difficult in areas of the world where radiologists are not avail...

Descripción completa

Detalles Bibliográficos
Autores principales: Chang, Yung-Chun, Hsing, Yan-Chun, Chiu, Yu-Wen, Shih, Cho-Chiang, Lin, Jun-Hong, Hsiao, Shih-Hsin, Sakai, Koji, Ko, Kai-Hsiung, Chen, Cheng-Yu
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8951579/
https://www.ncbi.nlm.nih.gov/pubmed/35330417
http://dx.doi.org/10.3390/jpm12030417
_version_ 1784675422265409536
author Chang, Yung-Chun
Hsing, Yan-Chun
Chiu, Yu-Wen
Shih, Cho-Chiang
Lin, Jun-Hong
Hsiao, Shih-Hsin
Sakai, Koji
Ko, Kai-Hsiung
Chen, Cheng-Yu
author_facet Chang, Yung-Chun
Hsing, Yan-Chun
Chiu, Yu-Wen
Shih, Cho-Chiang
Lin, Jun-Hong
Hsiao, Shih-Hsin
Sakai, Koji
Ko, Kai-Hsiung
Chen, Cheng-Yu
author_sort Chang, Yung-Chun
collection PubMed
description Radiology report generation through chest radiography interpretation is a time-consuming task that involves the interpretation of images by expert radiologists. It is common for fatigue-induced diagnostic error to occur, and especially difficult in areas of the world where radiologists are not available or lack diagnostic expertise. In this research, we proposed a multi-objective deep learning model called CT2Rep (Computed Tomography to Report) for generating lung radiology reports by extracting semantic features from lung CT scans. A total of 458 CT scans were used in this research, from which 107 radiomics features and 6 slices of segmentation related nodule features were extracted for the input of our model. The CT2Rep can simultaneously predict position, margin, and texture, which are three important indicators of lung cancer, and achieves remarkable performance with an F(1)-score of 87.29%. We conducted a satisfaction survey for estimating the practicality of CT2Rep, and the results show that 95% of the reports received satisfactory ratings. The results demonstrate the great potential in this model for the production of robust and reliable quantitative lung diagnosis reports. Medical personnel can obtain important indicators simply by providing the lung CT scan to the system, which can bring about the widespread application of the proposed framework.
format Online
Article
Text
id pubmed-8951579
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-89515792022-03-26 Deep Multi-Objective Learning from Low-Dose CT for Automatic Lung-RADS Report Generation Chang, Yung-Chun Hsing, Yan-Chun Chiu, Yu-Wen Shih, Cho-Chiang Lin, Jun-Hong Hsiao, Shih-Hsin Sakai, Koji Ko, Kai-Hsiung Chen, Cheng-Yu J Pers Med Article Radiology report generation through chest radiography interpretation is a time-consuming task that involves the interpretation of images by expert radiologists. It is common for fatigue-induced diagnostic error to occur, and especially difficult in areas of the world where radiologists are not available or lack diagnostic expertise. In this research, we proposed a multi-objective deep learning model called CT2Rep (Computed Tomography to Report) for generating lung radiology reports by extracting semantic features from lung CT scans. A total of 458 CT scans were used in this research, from which 107 radiomics features and 6 slices of segmentation related nodule features were extracted for the input of our model. The CT2Rep can simultaneously predict position, margin, and texture, which are three important indicators of lung cancer, and achieves remarkable performance with an F(1)-score of 87.29%. We conducted a satisfaction survey for estimating the practicality of CT2Rep, and the results show that 95% of the reports received satisfactory ratings. The results demonstrate the great potential in this model for the production of robust and reliable quantitative lung diagnosis reports. Medical personnel can obtain important indicators simply by providing the lung CT scan to the system, which can bring about the widespread application of the proposed framework. MDPI 2022-03-08 /pmc/articles/PMC8951579/ /pubmed/35330417 http://dx.doi.org/10.3390/jpm12030417 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Chang, Yung-Chun
Hsing, Yan-Chun
Chiu, Yu-Wen
Shih, Cho-Chiang
Lin, Jun-Hong
Hsiao, Shih-Hsin
Sakai, Koji
Ko, Kai-Hsiung
Chen, Cheng-Yu
Deep Multi-Objective Learning from Low-Dose CT for Automatic Lung-RADS Report Generation
title Deep Multi-Objective Learning from Low-Dose CT for Automatic Lung-RADS Report Generation
title_full Deep Multi-Objective Learning from Low-Dose CT for Automatic Lung-RADS Report Generation
title_fullStr Deep Multi-Objective Learning from Low-Dose CT for Automatic Lung-RADS Report Generation
title_full_unstemmed Deep Multi-Objective Learning from Low-Dose CT for Automatic Lung-RADS Report Generation
title_short Deep Multi-Objective Learning from Low-Dose CT for Automatic Lung-RADS Report Generation
title_sort deep multi-objective learning from low-dose ct for automatic lung-rads report generation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8951579/
https://www.ncbi.nlm.nih.gov/pubmed/35330417
http://dx.doi.org/10.3390/jpm12030417
work_keys_str_mv AT changyungchun deepmultiobjectivelearningfromlowdosectforautomaticlungradsreportgeneration
AT hsingyanchun deepmultiobjectivelearningfromlowdosectforautomaticlungradsreportgeneration
AT chiuyuwen deepmultiobjectivelearningfromlowdosectforautomaticlungradsreportgeneration
AT shihchochiang deepmultiobjectivelearningfromlowdosectforautomaticlungradsreportgeneration
AT linjunhong deepmultiobjectivelearningfromlowdosectforautomaticlungradsreportgeneration
AT hsiaoshihhsin deepmultiobjectivelearningfromlowdosectforautomaticlungradsreportgeneration
AT sakaikoji deepmultiobjectivelearningfromlowdosectforautomaticlungradsreportgeneration
AT kokaihsiung deepmultiobjectivelearningfromlowdosectforautomaticlungradsreportgeneration
AT chenchengyu deepmultiobjectivelearningfromlowdosectforautomaticlungradsreportgeneration