Cargando…

Automated hand-marked semantic text recognition from photographs

Automated text recognition techniques have made significant advancements; however, certain tasks still present challenges. This study is motivated by the need to automatically recognize hand-marked text on construction defect tags among millions of photographs. To address this challenge, we investig...

Descripción completa

Detalles Bibliográficos
Autores principales: Suh, Seungah, Lee, Ghang, Gil, Daeyoung, Kim, Yonghan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10469204/
https://www.ncbi.nlm.nih.gov/pubmed/37648714
http://dx.doi.org/10.1038/s41598-023-41489-4
_version_ 1785099390986223616
author Suh, Seungah
Lee, Ghang
Gil, Daeyoung
Kim, Yonghan
author_facet Suh, Seungah
Lee, Ghang
Gil, Daeyoung
Kim, Yonghan
author_sort Suh, Seungah
collection PubMed
description Automated text recognition techniques have made significant advancements; however, certain tasks still present challenges. This study is motivated by the need to automatically recognize hand-marked text on construction defect tags among millions of photographs. To address this challenge, we investigated three methods for automating hand-marked semantic text recognition (HMSTR)—a modified scene text recognition-based (STR) approach, a two-step HMSTR approach, and a lumped approach. The STR approach involves locating marked text using an object detection model and recognizing it using a competition-winning STR model. Similarly, the two-step HMSTR approach first localizes the marked text and then recognizes the semantic text using an image classification model. By contrast, the lumped approach performs both localization and identification of marked semantic text in a single step using object detection. Among these approaches, the two-step HMSTR approach achieved the highest F1 score (0.92) for recognizing circled text, followed by the STR approach (0.87) and the lumped approach (0.78). To validate the generalizability of the two-step HMSTR approach, subsequent experiments were conducted using check-marked text, resulting in an F1 score of 0.88. Although the proposed methods have been tested specifically with tags, they can be extended to recognize marked text in reports or books.
format Online
Article
Text
id pubmed-10469204
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-104692042023-09-01 Automated hand-marked semantic text recognition from photographs Suh, Seungah Lee, Ghang Gil, Daeyoung Kim, Yonghan Sci Rep Article Automated text recognition techniques have made significant advancements; however, certain tasks still present challenges. This study is motivated by the need to automatically recognize hand-marked text on construction defect tags among millions of photographs. To address this challenge, we investigated three methods for automating hand-marked semantic text recognition (HMSTR)—a modified scene text recognition-based (STR) approach, a two-step HMSTR approach, and a lumped approach. The STR approach involves locating marked text using an object detection model and recognizing it using a competition-winning STR model. Similarly, the two-step HMSTR approach first localizes the marked text and then recognizes the semantic text using an image classification model. By contrast, the lumped approach performs both localization and identification of marked semantic text in a single step using object detection. Among these approaches, the two-step HMSTR approach achieved the highest F1 score (0.92) for recognizing circled text, followed by the STR approach (0.87) and the lumped approach (0.78). To validate the generalizability of the two-step HMSTR approach, subsequent experiments were conducted using check-marked text, resulting in an F1 score of 0.88. Although the proposed methods have been tested specifically with tags, they can be extended to recognize marked text in reports or books. Nature Publishing Group UK 2023-08-30 /pmc/articles/PMC10469204/ /pubmed/37648714 http://dx.doi.org/10.1038/s41598-023-41489-4 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Suh, Seungah
Lee, Ghang
Gil, Daeyoung
Kim, Yonghan
Automated hand-marked semantic text recognition from photographs
title Automated hand-marked semantic text recognition from photographs
title_full Automated hand-marked semantic text recognition from photographs
title_fullStr Automated hand-marked semantic text recognition from photographs
title_full_unstemmed Automated hand-marked semantic text recognition from photographs
title_short Automated hand-marked semantic text recognition from photographs
title_sort automated hand-marked semantic text recognition from photographs
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10469204/
https://www.ncbi.nlm.nih.gov/pubmed/37648714
http://dx.doi.org/10.1038/s41598-023-41489-4
work_keys_str_mv AT suhseungah automatedhandmarkedsemantictextrecognitionfromphotographs
AT leeghang automatedhandmarkedsemantictextrecognitionfromphotographs
AT gildaeyoung automatedhandmarkedsemantictextrecognitionfromphotographs
AT kimyonghan automatedhandmarkedsemantictextrecognitionfromphotographs