Cargando…
A non-contact interactive system for multimodal surgical robots based on LeapMotion and visual tags
In recent years, the integration of robots in minimally invasive surgery has gained significant traction in clinical practice. However, conventional contact-based human-computer interaction poses the risk of bacterial infection, significantly limiting the role of robots in surgery. To address this l...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10616262/ https://www.ncbi.nlm.nih.gov/pubmed/37916180 http://dx.doi.org/10.3389/fnins.2023.1287053 |
_version_ | 1785129355393892352 |
---|---|
author | Zhang, Xinkang Wang, Jie Dai, Xiaokun Shen, Shu Chen, Xinrong |
author_facet | Zhang, Xinkang Wang, Jie Dai, Xiaokun Shen, Shu Chen, Xinrong |
author_sort | Zhang, Xinkang |
collection | PubMed |
description | In recent years, the integration of robots in minimally invasive surgery has gained significant traction in clinical practice. However, conventional contact-based human-computer interaction poses the risk of bacterial infection, significantly limiting the role of robots in surgery. To address this limitation, we propose an innovative interaction method rooted in gestures and visual tags, allowing surgeons to control and fine-tune surgical robots without physical contact with the environment. By encoding the six gestures collected using LeapMotion, we can effectively control the surgical robot in a non-contact manner. Moreover, utilizing Aruco technology, we have accurately identified the 3D spatial position of the visual label, and developed 12 fine-tuning operations to refine surgical instruments. To evaluate the applicability of our proposed system in surgery, we designed a relevant experimental setup. In the experiment, we achieved enough precision. These results demonstrate that our system meets the clinical standard, providing doctors with a non-contact and flexible means of interacting with robots during surgery. |
format | Online Article Text |
id | pubmed-10616262 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-106162622023-11-01 A non-contact interactive system for multimodal surgical robots based on LeapMotion and visual tags Zhang, Xinkang Wang, Jie Dai, Xiaokun Shen, Shu Chen, Xinrong Front Neurosci Neuroscience In recent years, the integration of robots in minimally invasive surgery has gained significant traction in clinical practice. However, conventional contact-based human-computer interaction poses the risk of bacterial infection, significantly limiting the role of robots in surgery. To address this limitation, we propose an innovative interaction method rooted in gestures and visual tags, allowing surgeons to control and fine-tune surgical robots without physical contact with the environment. By encoding the six gestures collected using LeapMotion, we can effectively control the surgical robot in a non-contact manner. Moreover, utilizing Aruco technology, we have accurately identified the 3D spatial position of the visual label, and developed 12 fine-tuning operations to refine surgical instruments. To evaluate the applicability of our proposed system in surgery, we designed a relevant experimental setup. In the experiment, we achieved enough precision. These results demonstrate that our system meets the clinical standard, providing doctors with a non-contact and flexible means of interacting with robots during surgery. Frontiers Media S.A. 2023-10-17 /pmc/articles/PMC10616262/ /pubmed/37916180 http://dx.doi.org/10.3389/fnins.2023.1287053 Text en Copyright © 2023 Zhang, Wang, Dai, Shen and Chen. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Neuroscience Zhang, Xinkang Wang, Jie Dai, Xiaokun Shen, Shu Chen, Xinrong A non-contact interactive system for multimodal surgical robots based on LeapMotion and visual tags |
title | A non-contact interactive system for multimodal surgical robots based on LeapMotion and visual tags |
title_full | A non-contact interactive system for multimodal surgical robots based on LeapMotion and visual tags |
title_fullStr | A non-contact interactive system for multimodal surgical robots based on LeapMotion and visual tags |
title_full_unstemmed | A non-contact interactive system for multimodal surgical robots based on LeapMotion and visual tags |
title_short | A non-contact interactive system for multimodal surgical robots based on LeapMotion and visual tags |
title_sort | non-contact interactive system for multimodal surgical robots based on leapmotion and visual tags |
topic | Neuroscience |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10616262/ https://www.ncbi.nlm.nih.gov/pubmed/37916180 http://dx.doi.org/10.3389/fnins.2023.1287053 |
work_keys_str_mv | AT zhangxinkang anoncontactinteractivesystemformultimodalsurgicalrobotsbasedonleapmotionandvisualtags AT wangjie anoncontactinteractivesystemformultimodalsurgicalrobotsbasedonleapmotionandvisualtags AT daixiaokun anoncontactinteractivesystemformultimodalsurgicalrobotsbasedonleapmotionandvisualtags AT shenshu anoncontactinteractivesystemformultimodalsurgicalrobotsbasedonleapmotionandvisualtags AT chenxinrong anoncontactinteractivesystemformultimodalsurgicalrobotsbasedonleapmotionandvisualtags AT zhangxinkang noncontactinteractivesystemformultimodalsurgicalrobotsbasedonleapmotionandvisualtags AT wangjie noncontactinteractivesystemformultimodalsurgicalrobotsbasedonleapmotionandvisualtags AT daixiaokun noncontactinteractivesystemformultimodalsurgicalrobotsbasedonleapmotionandvisualtags AT shenshu noncontactinteractivesystemformultimodalsurgicalrobotsbasedonleapmotionandvisualtags AT chenxinrong noncontactinteractivesystemformultimodalsurgicalrobotsbasedonleapmotionandvisualtags |