Cargando…

Temporally guided articulated hand pose tracking in surgical videos

PURPOSE: Articulated hand pose tracking is an under-explored problem that carries the potential for use in an extensive number of applications, especially in the medical domain. With a robust and accurate tracking system on surgical videos, the motion dynamics and movement patterns of the hands can...

Descripción completa

Detalles Bibliográficos
Autores principales: Louis, Nathan, Zhou, Luowei, Yule, Steven J., Dias, Roger D., Manojlovich, Milisa, Pagani, Francis D., Likosky, Donald S., Corso, Jason J.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer International Publishing 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9883342/
https://www.ncbi.nlm.nih.gov/pubmed/36190616
http://dx.doi.org/10.1007/s11548-022-02761-6
_version_ 1784879488426835968
author Louis, Nathan
Zhou, Luowei
Yule, Steven J.
Dias, Roger D.
Manojlovich, Milisa
Pagani, Francis D.
Likosky, Donald S.
Corso, Jason J.
author_facet Louis, Nathan
Zhou, Luowei
Yule, Steven J.
Dias, Roger D.
Manojlovich, Milisa
Pagani, Francis D.
Likosky, Donald S.
Corso, Jason J.
author_sort Louis, Nathan
collection PubMed
description PURPOSE: Articulated hand pose tracking is an under-explored problem that carries the potential for use in an extensive number of applications, especially in the medical domain. With a robust and accurate tracking system on surgical videos, the motion dynamics and movement patterns of the hands can be captured and analyzed for many rich tasks. METHODS: In this work, we propose a novel hand pose estimation model, CondPose, which improves detection and tracking accuracy by incorporating a pose prior into its prediction. We show improvements over state-of-the-art methods which provide frame-wise independent predictions, by following a temporally guided approach that effectively leverages past predictions. RESULTS: We collect Surgical Hands, the first dataset that provides multi-instance articulated hand pose annotations for videos. Our dataset provides over 8.1k annotated hand poses from publicly available surgical videos and bounding boxes, pose annotations, and tracking IDs to enable multi-instance tracking. When evaluated on Surgical Hands, we show our method outperforms the state-of-the-art approach using mean Average Precision, to measure pose estimation accuracy, and Multiple Object Tracking Accuracy, to assess pose tracking performance. CONCLUSION: In comparison to a frame-wise independent strategy, we show greater performance in detecting and tracking hand poses and more substantial impact on localization accuracy. This has positive implications in generating more accurate representations of hands in the scene to be used for targeted downstream tasks.
format Online
Article
Text
id pubmed-9883342
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Springer International Publishing
record_format MEDLINE/PubMed
spelling pubmed-98833422023-01-29 Temporally guided articulated hand pose tracking in surgical videos Louis, Nathan Zhou, Luowei Yule, Steven J. Dias, Roger D. Manojlovich, Milisa Pagani, Francis D. Likosky, Donald S. Corso, Jason J. Int J Comput Assist Radiol Surg Original Article PURPOSE: Articulated hand pose tracking is an under-explored problem that carries the potential for use in an extensive number of applications, especially in the medical domain. With a robust and accurate tracking system on surgical videos, the motion dynamics and movement patterns of the hands can be captured and analyzed for many rich tasks. METHODS: In this work, we propose a novel hand pose estimation model, CondPose, which improves detection and tracking accuracy by incorporating a pose prior into its prediction. We show improvements over state-of-the-art methods which provide frame-wise independent predictions, by following a temporally guided approach that effectively leverages past predictions. RESULTS: We collect Surgical Hands, the first dataset that provides multi-instance articulated hand pose annotations for videos. Our dataset provides over 8.1k annotated hand poses from publicly available surgical videos and bounding boxes, pose annotations, and tracking IDs to enable multi-instance tracking. When evaluated on Surgical Hands, we show our method outperforms the state-of-the-art approach using mean Average Precision, to measure pose estimation accuracy, and Multiple Object Tracking Accuracy, to assess pose tracking performance. CONCLUSION: In comparison to a frame-wise independent strategy, we show greater performance in detecting and tracking hand poses and more substantial impact on localization accuracy. This has positive implications in generating more accurate representations of hands in the scene to be used for targeted downstream tasks. Springer International Publishing 2022-10-03 2023 /pmc/articles/PMC9883342/ /pubmed/36190616 http://dx.doi.org/10.1007/s11548-022-02761-6 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Original Article
Louis, Nathan
Zhou, Luowei
Yule, Steven J.
Dias, Roger D.
Manojlovich, Milisa
Pagani, Francis D.
Likosky, Donald S.
Corso, Jason J.
Temporally guided articulated hand pose tracking in surgical videos
title Temporally guided articulated hand pose tracking in surgical videos
title_full Temporally guided articulated hand pose tracking in surgical videos
title_fullStr Temporally guided articulated hand pose tracking in surgical videos
title_full_unstemmed Temporally guided articulated hand pose tracking in surgical videos
title_short Temporally guided articulated hand pose tracking in surgical videos
title_sort temporally guided articulated hand pose tracking in surgical videos
topic Original Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9883342/
https://www.ncbi.nlm.nih.gov/pubmed/36190616
http://dx.doi.org/10.1007/s11548-022-02761-6
work_keys_str_mv AT louisnathan temporallyguidedarticulatedhandposetrackinginsurgicalvideos
AT zhouluowei temporallyguidedarticulatedhandposetrackinginsurgicalvideos
AT yulestevenj temporallyguidedarticulatedhandposetrackinginsurgicalvideos
AT diasrogerd temporallyguidedarticulatedhandposetrackinginsurgicalvideos
AT manojlovichmilisa temporallyguidedarticulatedhandposetrackinginsurgicalvideos
AT paganifrancisd temporallyguidedarticulatedhandposetrackinginsurgicalvideos
AT likoskydonalds temporallyguidedarticulatedhandposetrackinginsurgicalvideos
AT corsojasonj temporallyguidedarticulatedhandposetrackinginsurgicalvideos