Cargando…
Temporally guided articulated hand pose tracking in surgical videos
PURPOSE: Articulated hand pose tracking is an under-explored problem that carries the potential for use in an extensive number of applications, especially in the medical domain. With a robust and accurate tracking system on surgical videos, the motion dynamics and movement patterns of the hands can...
Autores principales: | , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer International Publishing
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9883342/ https://www.ncbi.nlm.nih.gov/pubmed/36190616 http://dx.doi.org/10.1007/s11548-022-02761-6 |
_version_ | 1784879488426835968 |
---|---|
author | Louis, Nathan Zhou, Luowei Yule, Steven J. Dias, Roger D. Manojlovich, Milisa Pagani, Francis D. Likosky, Donald S. Corso, Jason J. |
author_facet | Louis, Nathan Zhou, Luowei Yule, Steven J. Dias, Roger D. Manojlovich, Milisa Pagani, Francis D. Likosky, Donald S. Corso, Jason J. |
author_sort | Louis, Nathan |
collection | PubMed |
description | PURPOSE: Articulated hand pose tracking is an under-explored problem that carries the potential for use in an extensive number of applications, especially in the medical domain. With a robust and accurate tracking system on surgical videos, the motion dynamics and movement patterns of the hands can be captured and analyzed for many rich tasks. METHODS: In this work, we propose a novel hand pose estimation model, CondPose, which improves detection and tracking accuracy by incorporating a pose prior into its prediction. We show improvements over state-of-the-art methods which provide frame-wise independent predictions, by following a temporally guided approach that effectively leverages past predictions. RESULTS: We collect Surgical Hands, the first dataset that provides multi-instance articulated hand pose annotations for videos. Our dataset provides over 8.1k annotated hand poses from publicly available surgical videos and bounding boxes, pose annotations, and tracking IDs to enable multi-instance tracking. When evaluated on Surgical Hands, we show our method outperforms the state-of-the-art approach using mean Average Precision, to measure pose estimation accuracy, and Multiple Object Tracking Accuracy, to assess pose tracking performance. CONCLUSION: In comparison to a frame-wise independent strategy, we show greater performance in detecting and tracking hand poses and more substantial impact on localization accuracy. This has positive implications in generating more accurate representations of hands in the scene to be used for targeted downstream tasks. |
format | Online Article Text |
id | pubmed-9883342 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Springer International Publishing |
record_format | MEDLINE/PubMed |
spelling | pubmed-98833422023-01-29 Temporally guided articulated hand pose tracking in surgical videos Louis, Nathan Zhou, Luowei Yule, Steven J. Dias, Roger D. Manojlovich, Milisa Pagani, Francis D. Likosky, Donald S. Corso, Jason J. Int J Comput Assist Radiol Surg Original Article PURPOSE: Articulated hand pose tracking is an under-explored problem that carries the potential for use in an extensive number of applications, especially in the medical domain. With a robust and accurate tracking system on surgical videos, the motion dynamics and movement patterns of the hands can be captured and analyzed for many rich tasks. METHODS: In this work, we propose a novel hand pose estimation model, CondPose, which improves detection and tracking accuracy by incorporating a pose prior into its prediction. We show improvements over state-of-the-art methods which provide frame-wise independent predictions, by following a temporally guided approach that effectively leverages past predictions. RESULTS: We collect Surgical Hands, the first dataset that provides multi-instance articulated hand pose annotations for videos. Our dataset provides over 8.1k annotated hand poses from publicly available surgical videos and bounding boxes, pose annotations, and tracking IDs to enable multi-instance tracking. When evaluated on Surgical Hands, we show our method outperforms the state-of-the-art approach using mean Average Precision, to measure pose estimation accuracy, and Multiple Object Tracking Accuracy, to assess pose tracking performance. CONCLUSION: In comparison to a frame-wise independent strategy, we show greater performance in detecting and tracking hand poses and more substantial impact on localization accuracy. This has positive implications in generating more accurate representations of hands in the scene to be used for targeted downstream tasks. Springer International Publishing 2022-10-03 2023 /pmc/articles/PMC9883342/ /pubmed/36190616 http://dx.doi.org/10.1007/s11548-022-02761-6 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Original Article Louis, Nathan Zhou, Luowei Yule, Steven J. Dias, Roger D. Manojlovich, Milisa Pagani, Francis D. Likosky, Donald S. Corso, Jason J. Temporally guided articulated hand pose tracking in surgical videos |
title | Temporally guided articulated hand pose tracking in surgical videos |
title_full | Temporally guided articulated hand pose tracking in surgical videos |
title_fullStr | Temporally guided articulated hand pose tracking in surgical videos |
title_full_unstemmed | Temporally guided articulated hand pose tracking in surgical videos |
title_short | Temporally guided articulated hand pose tracking in surgical videos |
title_sort | temporally guided articulated hand pose tracking in surgical videos |
topic | Original Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9883342/ https://www.ncbi.nlm.nih.gov/pubmed/36190616 http://dx.doi.org/10.1007/s11548-022-02761-6 |
work_keys_str_mv | AT louisnathan temporallyguidedarticulatedhandposetrackinginsurgicalvideos AT zhouluowei temporallyguidedarticulatedhandposetrackinginsurgicalvideos AT yulestevenj temporallyguidedarticulatedhandposetrackinginsurgicalvideos AT diasrogerd temporallyguidedarticulatedhandposetrackinginsurgicalvideos AT manojlovichmilisa temporallyguidedarticulatedhandposetrackinginsurgicalvideos AT paganifrancisd temporallyguidedarticulatedhandposetrackinginsurgicalvideos AT likoskydonalds temporallyguidedarticulatedhandposetrackinginsurgicalvideos AT corsojasonj temporallyguidedarticulatedhandposetrackinginsurgicalvideos |