Cargando…
Automated hand-marked semantic text recognition from photographs
Automated text recognition techniques have made significant advancements; however, certain tasks still present challenges. This study is motivated by the need to automatically recognize hand-marked text on construction defect tags among millions of photographs. To address this challenge, we investig...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10469204/ https://www.ncbi.nlm.nih.gov/pubmed/37648714 http://dx.doi.org/10.1038/s41598-023-41489-4 |
_version_ | 1785099390986223616 |
---|---|
author | Suh, Seungah Lee, Ghang Gil, Daeyoung Kim, Yonghan |
author_facet | Suh, Seungah Lee, Ghang Gil, Daeyoung Kim, Yonghan |
author_sort | Suh, Seungah |
collection | PubMed |
description | Automated text recognition techniques have made significant advancements; however, certain tasks still present challenges. This study is motivated by the need to automatically recognize hand-marked text on construction defect tags among millions of photographs. To address this challenge, we investigated three methods for automating hand-marked semantic text recognition (HMSTR)—a modified scene text recognition-based (STR) approach, a two-step HMSTR approach, and a lumped approach. The STR approach involves locating marked text using an object detection model and recognizing it using a competition-winning STR model. Similarly, the two-step HMSTR approach first localizes the marked text and then recognizes the semantic text using an image classification model. By contrast, the lumped approach performs both localization and identification of marked semantic text in a single step using object detection. Among these approaches, the two-step HMSTR approach achieved the highest F1 score (0.92) for recognizing circled text, followed by the STR approach (0.87) and the lumped approach (0.78). To validate the generalizability of the two-step HMSTR approach, subsequent experiments were conducted using check-marked text, resulting in an F1 score of 0.88. Although the proposed methods have been tested specifically with tags, they can be extended to recognize marked text in reports or books. |
format | Online Article Text |
id | pubmed-10469204 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-104692042023-09-01 Automated hand-marked semantic text recognition from photographs Suh, Seungah Lee, Ghang Gil, Daeyoung Kim, Yonghan Sci Rep Article Automated text recognition techniques have made significant advancements; however, certain tasks still present challenges. This study is motivated by the need to automatically recognize hand-marked text on construction defect tags among millions of photographs. To address this challenge, we investigated three methods for automating hand-marked semantic text recognition (HMSTR)—a modified scene text recognition-based (STR) approach, a two-step HMSTR approach, and a lumped approach. The STR approach involves locating marked text using an object detection model and recognizing it using a competition-winning STR model. Similarly, the two-step HMSTR approach first localizes the marked text and then recognizes the semantic text using an image classification model. By contrast, the lumped approach performs both localization and identification of marked semantic text in a single step using object detection. Among these approaches, the two-step HMSTR approach achieved the highest F1 score (0.92) for recognizing circled text, followed by the STR approach (0.87) and the lumped approach (0.78). To validate the generalizability of the two-step HMSTR approach, subsequent experiments were conducted using check-marked text, resulting in an F1 score of 0.88. Although the proposed methods have been tested specifically with tags, they can be extended to recognize marked text in reports or books. Nature Publishing Group UK 2023-08-30 /pmc/articles/PMC10469204/ /pubmed/37648714 http://dx.doi.org/10.1038/s41598-023-41489-4 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Suh, Seungah Lee, Ghang Gil, Daeyoung Kim, Yonghan Automated hand-marked semantic text recognition from photographs |
title | Automated hand-marked semantic text recognition from photographs |
title_full | Automated hand-marked semantic text recognition from photographs |
title_fullStr | Automated hand-marked semantic text recognition from photographs |
title_full_unstemmed | Automated hand-marked semantic text recognition from photographs |
title_short | Automated hand-marked semantic text recognition from photographs |
title_sort | automated hand-marked semantic text recognition from photographs |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10469204/ https://www.ncbi.nlm.nih.gov/pubmed/37648714 http://dx.doi.org/10.1038/s41598-023-41489-4 |
work_keys_str_mv | AT suhseungah automatedhandmarkedsemantictextrecognitionfromphotographs AT leeghang automatedhandmarkedsemantictextrecognitionfromphotographs AT gildaeyoung automatedhandmarkedsemantictextrecognitionfromphotographs AT kimyonghan automatedhandmarkedsemantictextrecognitionfromphotographs |