Cargando…

Learning intraoperative organ manipulation with context-based reinforcement learning

PURPOSE: Automation of sub-tasks during robotic surgery is challenging due to the high variability of the surgical scenes intra- and inter-patients. For example, the pick and place task can be executed different times during the same operation and for distinct purposes. Hence, designing automation s...

Descripción completa

Detalles Bibliográficos
Autores principales: D’Ettorre, Claudia, Zirino, Silvia, Dei, Neri Niccolò, Stilli, Agostino, De Momi, Elena, Stoyanov, Danail
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer International Publishing 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9307544/
https://www.ncbi.nlm.nih.gov/pubmed/35503394
http://dx.doi.org/10.1007/s11548-022-02630-2
_version_ 1784752786492096512
author D’Ettorre, Claudia
Zirino, Silvia
Dei, Neri Niccolò
Stilli, Agostino
De Momi, Elena
Stoyanov, Danail
author_facet D’Ettorre, Claudia
Zirino, Silvia
Dei, Neri Niccolò
Stilli, Agostino
De Momi, Elena
Stoyanov, Danail
author_sort D’Ettorre, Claudia
collection PubMed
description PURPOSE: Automation of sub-tasks during robotic surgery is challenging due to the high variability of the surgical scenes intra- and inter-patients. For example, the pick and place task can be executed different times during the same operation and for distinct purposes. Hence, designing automation solutions that can generalise a skill over different contexts becomes hard. All the experiments are conducted using the Pneumatic Attachable Flexible (PAF) rail, a novel surgical tool designed for robotic-assisted intraoperative organ manipulation. METHODS: We build upon previous open-source surgical Reinforcement Learning (RL) training environment to develop a new RL framework for manipulation skills, rlman. In rlman, contextual RL agents are trained to solve different aspects of the pick and place task using the PAF rail system. rlman is implemented to support both low- and high-dimensional state information to solve surgical sub-tasks in a simulation environment. RESULTS: We use rlman to train state of the art RL agents to solve four different surgical sub-tasks involving manipulation skills using the PAF rail. We compare the results with state-of-the-art benchmarks found in the literature. We evaluate the ability of the agent to be able to generalise over different aspects of the targeted surgical environment. CONCLUSION: We have shown that the rlman framework can support the training of different RL algorithms for solving surgical sub-task, analysing the importance of context information for generalisation capabilities. We are aiming to deploy the trained policy on the real da Vinci using the dVRK and show that the generalisation of the trained policy can be transferred to the real world. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s11548-022-02630-2.
format Online
Article
Text
id pubmed-9307544
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Springer International Publishing
record_format MEDLINE/PubMed
spelling pubmed-93075442022-07-24 Learning intraoperative organ manipulation with context-based reinforcement learning D’Ettorre, Claudia Zirino, Silvia Dei, Neri Niccolò Stilli, Agostino De Momi, Elena Stoyanov, Danail Int J Comput Assist Radiol Surg Original Article PURPOSE: Automation of sub-tasks during robotic surgery is challenging due to the high variability of the surgical scenes intra- and inter-patients. For example, the pick and place task can be executed different times during the same operation and for distinct purposes. Hence, designing automation solutions that can generalise a skill over different contexts becomes hard. All the experiments are conducted using the Pneumatic Attachable Flexible (PAF) rail, a novel surgical tool designed for robotic-assisted intraoperative organ manipulation. METHODS: We build upon previous open-source surgical Reinforcement Learning (RL) training environment to develop a new RL framework for manipulation skills, rlman. In rlman, contextual RL agents are trained to solve different aspects of the pick and place task using the PAF rail system. rlman is implemented to support both low- and high-dimensional state information to solve surgical sub-tasks in a simulation environment. RESULTS: We use rlman to train state of the art RL agents to solve four different surgical sub-tasks involving manipulation skills using the PAF rail. We compare the results with state-of-the-art benchmarks found in the literature. We evaluate the ability of the agent to be able to generalise over different aspects of the targeted surgical environment. CONCLUSION: We have shown that the rlman framework can support the training of different RL algorithms for solving surgical sub-task, analysing the importance of context information for generalisation capabilities. We are aiming to deploy the trained policy on the real da Vinci using the dVRK and show that the generalisation of the trained policy can be transferred to the real world. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s11548-022-02630-2. Springer International Publishing 2022-05-03 2022 /pmc/articles/PMC9307544/ /pubmed/35503394 http://dx.doi.org/10.1007/s11548-022-02630-2 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. (https://creativecommons.org/licenses/by/4.0/)
spellingShingle Original Article
D’Ettorre, Claudia
Zirino, Silvia
Dei, Neri Niccolò
Stilli, Agostino
De Momi, Elena
Stoyanov, Danail
Learning intraoperative organ manipulation with context-based reinforcement learning
title Learning intraoperative organ manipulation with context-based reinforcement learning
title_full Learning intraoperative organ manipulation with context-based reinforcement learning
title_fullStr Learning intraoperative organ manipulation with context-based reinforcement learning
title_full_unstemmed Learning intraoperative organ manipulation with context-based reinforcement learning
title_short Learning intraoperative organ manipulation with context-based reinforcement learning
title_sort learning intraoperative organ manipulation with context-based reinforcement learning
topic Original Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9307544/
https://www.ncbi.nlm.nih.gov/pubmed/35503394
http://dx.doi.org/10.1007/s11548-022-02630-2
work_keys_str_mv AT dettorreclaudia learningintraoperativeorganmanipulationwithcontextbasedreinforcementlearning
AT zirinosilvia learningintraoperativeorganmanipulationwithcontextbasedreinforcementlearning
AT deineriniccolo learningintraoperativeorganmanipulationwithcontextbasedreinforcementlearning
AT stilliagostino learningintraoperativeorganmanipulationwithcontextbasedreinforcementlearning
AT demomielena learningintraoperativeorganmanipulationwithcontextbasedreinforcementlearning
AT stoyanovdanail learningintraoperativeorganmanipulationwithcontextbasedreinforcementlearning