Cargando…

Performance of deep learning technology for evaluation of positioning quality in periapical radiography of the maxillary canine

OBJECTIVES: The aim of the present study was to create and test an automatic system for assessing the technical quality of positioning in periapical radiography of the maxillary canines using deep learning classification and segmentation techniques. METHODS: We created and tested two deep learning s...

Descripción completa

Detalles Bibliográficos
Autores principales: Mori, Mizuho, Ariji, Yoshiko, Fukuda, Motoki, Kitano, Tomoya, Funakoshi, Takuma, Nishiyama, Wataru, Kohinata, Kiyomi, Iida, Yukihiro, Ariji, Eiichiro, Katsumata, Akitoshi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer Singapore 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8741711/
https://www.ncbi.nlm.nih.gov/pubmed/34041639
http://dx.doi.org/10.1007/s11282-021-00538-2
_version_ 1784629551801827328
author Mori, Mizuho
Ariji, Yoshiko
Fukuda, Motoki
Kitano, Tomoya
Funakoshi, Takuma
Nishiyama, Wataru
Kohinata, Kiyomi
Iida, Yukihiro
Ariji, Eiichiro
Katsumata, Akitoshi
author_facet Mori, Mizuho
Ariji, Yoshiko
Fukuda, Motoki
Kitano, Tomoya
Funakoshi, Takuma
Nishiyama, Wataru
Kohinata, Kiyomi
Iida, Yukihiro
Ariji, Eiichiro
Katsumata, Akitoshi
author_sort Mori, Mizuho
collection PubMed
description OBJECTIVES: The aim of the present study was to create and test an automatic system for assessing the technical quality of positioning in periapical radiography of the maxillary canines using deep learning classification and segmentation techniques. METHODS: We created and tested two deep learning systems using 500 periapical radiographs (250 each of good- and bad-quality images). We assigned 350, 70, and 80 images as the training, validation, and test datasets, respectively. The learning model of system 1 was created with only the classification process, whereas system 2 consisted of both the segmentation and classification models. In each model, 500 epochs of training were performed using AlexNet and U-net for classification and segmentation, respectively. The segmentation results were evaluated by the intersection over union method, with values of 0.6 or more considered as success. The classification results were compared between the two systems. RESULTS: The segmentation performance of system 2 was recall, precision, and F measure of 0.937, 0.961, and 0.949, respectively. System 2 showed better classification performance values than those obtained by system 1. The area under the receiver operating characteristic curve values differed significantly between system 1 (0.649) and system 2 (0.927). CONCLUSIONS: The deep learning systems we created appeared to have potential benefits in evaluation of the technical positioning quality of periapical radiographs through the use of segmentation and classification functions.
format Online
Article
Text
id pubmed-8741711
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Springer Singapore
record_format MEDLINE/PubMed
spelling pubmed-87417112022-01-20 Performance of deep learning technology for evaluation of positioning quality in periapical radiography of the maxillary canine Mori, Mizuho Ariji, Yoshiko Fukuda, Motoki Kitano, Tomoya Funakoshi, Takuma Nishiyama, Wataru Kohinata, Kiyomi Iida, Yukihiro Ariji, Eiichiro Katsumata, Akitoshi Oral Radiol Original Article OBJECTIVES: The aim of the present study was to create and test an automatic system for assessing the technical quality of positioning in periapical radiography of the maxillary canines using deep learning classification and segmentation techniques. METHODS: We created and tested two deep learning systems using 500 periapical radiographs (250 each of good- and bad-quality images). We assigned 350, 70, and 80 images as the training, validation, and test datasets, respectively. The learning model of system 1 was created with only the classification process, whereas system 2 consisted of both the segmentation and classification models. In each model, 500 epochs of training were performed using AlexNet and U-net for classification and segmentation, respectively. The segmentation results were evaluated by the intersection over union method, with values of 0.6 or more considered as success. The classification results were compared between the two systems. RESULTS: The segmentation performance of system 2 was recall, precision, and F measure of 0.937, 0.961, and 0.949, respectively. System 2 showed better classification performance values than those obtained by system 1. The area under the receiver operating characteristic curve values differed significantly between system 1 (0.649) and system 2 (0.927). CONCLUSIONS: The deep learning systems we created appeared to have potential benefits in evaluation of the technical positioning quality of periapical radiographs through the use of segmentation and classification functions. Springer Singapore 2021-05-26 2022 /pmc/articles/PMC8741711/ /pubmed/34041639 http://dx.doi.org/10.1007/s11282-021-00538-2 Text en © The Author(s) 2021 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Original Article
Mori, Mizuho
Ariji, Yoshiko
Fukuda, Motoki
Kitano, Tomoya
Funakoshi, Takuma
Nishiyama, Wataru
Kohinata, Kiyomi
Iida, Yukihiro
Ariji, Eiichiro
Katsumata, Akitoshi
Performance of deep learning technology for evaluation of positioning quality in periapical radiography of the maxillary canine
title Performance of deep learning technology for evaluation of positioning quality in periapical radiography of the maxillary canine
title_full Performance of deep learning technology for evaluation of positioning quality in periapical radiography of the maxillary canine
title_fullStr Performance of deep learning technology for evaluation of positioning quality in periapical radiography of the maxillary canine
title_full_unstemmed Performance of deep learning technology for evaluation of positioning quality in periapical radiography of the maxillary canine
title_short Performance of deep learning technology for evaluation of positioning quality in periapical radiography of the maxillary canine
title_sort performance of deep learning technology for evaluation of positioning quality in periapical radiography of the maxillary canine
topic Original Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8741711/
https://www.ncbi.nlm.nih.gov/pubmed/34041639
http://dx.doi.org/10.1007/s11282-021-00538-2
work_keys_str_mv AT morimizuho performanceofdeeplearningtechnologyforevaluationofpositioningqualityinperiapicalradiographyofthemaxillarycanine
AT arijiyoshiko performanceofdeeplearningtechnologyforevaluationofpositioningqualityinperiapicalradiographyofthemaxillarycanine
AT fukudamotoki performanceofdeeplearningtechnologyforevaluationofpositioningqualityinperiapicalradiographyofthemaxillarycanine
AT kitanotomoya performanceofdeeplearningtechnologyforevaluationofpositioningqualityinperiapicalradiographyofthemaxillarycanine
AT funakoshitakuma performanceofdeeplearningtechnologyforevaluationofpositioningqualityinperiapicalradiographyofthemaxillarycanine
AT nishiyamawataru performanceofdeeplearningtechnologyforevaluationofpositioningqualityinperiapicalradiographyofthemaxillarycanine
AT kohinatakiyomi performanceofdeeplearningtechnologyforevaluationofpositioningqualityinperiapicalradiographyofthemaxillarycanine
AT iidayukihiro performanceofdeeplearningtechnologyforevaluationofpositioningqualityinperiapicalradiographyofthemaxillarycanine
AT arijieiichiro performanceofdeeplearningtechnologyforevaluationofpositioningqualityinperiapicalradiographyofthemaxillarycanine
AT katsumataakitoshi performanceofdeeplearningtechnologyforevaluationofpositioningqualityinperiapicalradiographyofthemaxillarycanine