Cargando…

Hand Motion-Aware Surgical Tool Localization and Classification from an Egocentric Camera

Detecting surgical tools is an essential task for the analysis and evaluation of surgical videos. However, in open surgery such as plastic surgery, it is difficult to detect them because there are surgical tools with similar shapes, such as scissors and needle holders. Unlike endoscopic surgery, the...

Descripción completa

Detalles Bibliográficos
Autores principales: Shimizu, Tomohiro, Hachiuma, Ryo, Kajita, Hiroki, Takatsume, Yoshifumi, Saito, Hideo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8321273/
https://www.ncbi.nlm.nih.gov/pubmed/34460614
http://dx.doi.org/10.3390/jimaging7020015
_version_ 1783730811982315520
author Shimizu, Tomohiro
Hachiuma, Ryo
Kajita, Hiroki
Takatsume, Yoshifumi
Saito, Hideo
author_facet Shimizu, Tomohiro
Hachiuma, Ryo
Kajita, Hiroki
Takatsume, Yoshifumi
Saito, Hideo
author_sort Shimizu, Tomohiro
collection PubMed
description Detecting surgical tools is an essential task for the analysis and evaluation of surgical videos. However, in open surgery such as plastic surgery, it is difficult to detect them because there are surgical tools with similar shapes, such as scissors and needle holders. Unlike endoscopic surgery, the tips of the tools are often hidden in the operating field and are not captured clearly due to low camera resolution, whereas the movements of the tools and hands can be captured. As a result that the different uses of each tool require different hand movements, it is possible to use hand movement data to classify the two types of tools. We combined three modules for localization, selection, and classification, for the detection of the two tools. In the localization module, we employed the Faster R-CNN to detect surgical tools and target hands, and in the classification module, we extracted hand movement information by combining ResNet-18 and LSTM to classify two tools. We created a dataset in which seven different types of open surgery were recorded, and we provided the annotation of surgical tool detection. Our experiments show that our approach successfully detected the two different tools and outperformed the two baseline methods.
format Online
Article
Text
id pubmed-8321273
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-83212732021-08-26 Hand Motion-Aware Surgical Tool Localization and Classification from an Egocentric Camera Shimizu, Tomohiro Hachiuma, Ryo Kajita, Hiroki Takatsume, Yoshifumi Saito, Hideo J Imaging Article Detecting surgical tools is an essential task for the analysis and evaluation of surgical videos. However, in open surgery such as plastic surgery, it is difficult to detect them because there are surgical tools with similar shapes, such as scissors and needle holders. Unlike endoscopic surgery, the tips of the tools are often hidden in the operating field and are not captured clearly due to low camera resolution, whereas the movements of the tools and hands can be captured. As a result that the different uses of each tool require different hand movements, it is possible to use hand movement data to classify the two types of tools. We combined three modules for localization, selection, and classification, for the detection of the two tools. In the localization module, we employed the Faster R-CNN to detect surgical tools and target hands, and in the classification module, we extracted hand movement information by combining ResNet-18 and LSTM to classify two tools. We created a dataset in which seven different types of open surgery were recorded, and we provided the annotation of surgical tool detection. Our experiments show that our approach successfully detected the two different tools and outperformed the two baseline methods. MDPI 2021-01-25 /pmc/articles/PMC8321273/ /pubmed/34460614 http://dx.doi.org/10.3390/jimaging7020015 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) ).
spellingShingle Article
Shimizu, Tomohiro
Hachiuma, Ryo
Kajita, Hiroki
Takatsume, Yoshifumi
Saito, Hideo
Hand Motion-Aware Surgical Tool Localization and Classification from an Egocentric Camera
title Hand Motion-Aware Surgical Tool Localization and Classification from an Egocentric Camera
title_full Hand Motion-Aware Surgical Tool Localization and Classification from an Egocentric Camera
title_fullStr Hand Motion-Aware Surgical Tool Localization and Classification from an Egocentric Camera
title_full_unstemmed Hand Motion-Aware Surgical Tool Localization and Classification from an Egocentric Camera
title_short Hand Motion-Aware Surgical Tool Localization and Classification from an Egocentric Camera
title_sort hand motion-aware surgical tool localization and classification from an egocentric camera
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8321273/
https://www.ncbi.nlm.nih.gov/pubmed/34460614
http://dx.doi.org/10.3390/jimaging7020015
work_keys_str_mv AT shimizutomohiro handmotionawaresurgicaltoollocalizationandclassificationfromanegocentriccamera
AT hachiumaryo handmotionawaresurgicaltoollocalizationandclassificationfromanegocentriccamera
AT kajitahiroki handmotionawaresurgicaltoollocalizationandclassificationfromanegocentriccamera
AT takatsumeyoshifumi handmotionawaresurgicaltoollocalizationandclassificationfromanegocentriccamera
AT saitohideo handmotionawaresurgicaltoollocalizationandclassificationfromanegocentriccamera