Cargando…

Multi-modal imaging, model-based tracking, and mixed reality visualisation for orthopaedic surgery

Orthopaedic surgeons are still following the decades old workflow of using dozens of two-dimensional fluoroscopic images to drill through complex 3D structures, e.g. pelvis. This Letter presents a mixed reality support system, which incorporates multi-modal data fusion and model-based surgical tool...

Descripción completa

Detalles Bibliográficos
Autores principales: Lee, Sing Chun, Fuerst, Bernhard, Tateno, Keisuke, Johnson, Alex, Fotouhi, Javad, Osgood, Greg, Tombari, Federico, Navab, Nassir
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Institution of Engineering and Technology 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5683202/
https://www.ncbi.nlm.nih.gov/pubmed/29184659
http://dx.doi.org/10.1049/htl.2017.0066
_version_ 1783278232980684800
author Lee, Sing Chun
Fuerst, Bernhard
Tateno, Keisuke
Johnson, Alex
Fotouhi, Javad
Osgood, Greg
Tombari, Federico
Navab, Nassir
author_facet Lee, Sing Chun
Fuerst, Bernhard
Tateno, Keisuke
Johnson, Alex
Fotouhi, Javad
Osgood, Greg
Tombari, Federico
Navab, Nassir
author_sort Lee, Sing Chun
collection PubMed
description Orthopaedic surgeons are still following the decades old workflow of using dozens of two-dimensional fluoroscopic images to drill through complex 3D structures, e.g. pelvis. This Letter presents a mixed reality support system, which incorporates multi-modal data fusion and model-based surgical tool tracking for creating a mixed reality environment supporting screw placement in orthopaedic surgery. A red–green–blue–depth camera is rigidly attached to a mobile C-arm and is calibrated to the cone-beam computed tomography (CBCT) imaging space via iterative closest point algorithm. This allows real-time automatic fusion of reconstructed surface and/or 3D point clouds and synthetic fluoroscopic images obtained through CBCT imaging. An adapted 3D model-based tracking algorithm with automatic tool segmentation allows for tracking of the surgical tools occluded by hand. This proposed interactive 3D mixed reality environment provides an intuitive understanding of the surgical site and supports surgeons in quickly localising the entry point and orienting the surgical tool during screw placement. The authors validate the augmentation by measuring target registration error and also evaluate the tracking accuracy in the presence of partial occlusion.
format Online
Article
Text
id pubmed-5683202
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher The Institution of Engineering and Technology
record_format MEDLINE/PubMed
spelling pubmed-56832022017-11-28 Multi-modal imaging, model-based tracking, and mixed reality visualisation for orthopaedic surgery Lee, Sing Chun Fuerst, Bernhard Tateno, Keisuke Johnson, Alex Fotouhi, Javad Osgood, Greg Tombari, Federico Navab, Nassir Healthc Technol Lett Special Issue on Augmented Environments for Computer-Assisted Interventions Orthopaedic surgeons are still following the decades old workflow of using dozens of two-dimensional fluoroscopic images to drill through complex 3D structures, e.g. pelvis. This Letter presents a mixed reality support system, which incorporates multi-modal data fusion and model-based surgical tool tracking for creating a mixed reality environment supporting screw placement in orthopaedic surgery. A red–green–blue–depth camera is rigidly attached to a mobile C-arm and is calibrated to the cone-beam computed tomography (CBCT) imaging space via iterative closest point algorithm. This allows real-time automatic fusion of reconstructed surface and/or 3D point clouds and synthetic fluoroscopic images obtained through CBCT imaging. An adapted 3D model-based tracking algorithm with automatic tool segmentation allows for tracking of the surgical tools occluded by hand. This proposed interactive 3D mixed reality environment provides an intuitive understanding of the surgical site and supports surgeons in quickly localising the entry point and orienting the surgical tool during screw placement. The authors validate the augmentation by measuring target registration error and also evaluate the tracking accuracy in the presence of partial occlusion. The Institution of Engineering and Technology 2017-09-14 /pmc/articles/PMC5683202/ /pubmed/29184659 http://dx.doi.org/10.1049/htl.2017.0066 Text en http://creativecommons.org/licenses/by-nc/3.0/ This is an open access article published by the IET under the Creative Commons Attribution -NonCommercial License (http://creativecommons.org/licenses/by-nc/3.0/)
spellingShingle Special Issue on Augmented Environments for Computer-Assisted Interventions
Lee, Sing Chun
Fuerst, Bernhard
Tateno, Keisuke
Johnson, Alex
Fotouhi, Javad
Osgood, Greg
Tombari, Federico
Navab, Nassir
Multi-modal imaging, model-based tracking, and mixed reality visualisation for orthopaedic surgery
title Multi-modal imaging, model-based tracking, and mixed reality visualisation for orthopaedic surgery
title_full Multi-modal imaging, model-based tracking, and mixed reality visualisation for orthopaedic surgery
title_fullStr Multi-modal imaging, model-based tracking, and mixed reality visualisation for orthopaedic surgery
title_full_unstemmed Multi-modal imaging, model-based tracking, and mixed reality visualisation for orthopaedic surgery
title_short Multi-modal imaging, model-based tracking, and mixed reality visualisation for orthopaedic surgery
title_sort multi-modal imaging, model-based tracking, and mixed reality visualisation for orthopaedic surgery
topic Special Issue on Augmented Environments for Computer-Assisted Interventions
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5683202/
https://www.ncbi.nlm.nih.gov/pubmed/29184659
http://dx.doi.org/10.1049/htl.2017.0066
work_keys_str_mv AT leesingchun multimodalimagingmodelbasedtrackingandmixedrealityvisualisationfororthopaedicsurgery
AT fuerstbernhard multimodalimagingmodelbasedtrackingandmixedrealityvisualisationfororthopaedicsurgery
AT tatenokeisuke multimodalimagingmodelbasedtrackingandmixedrealityvisualisationfororthopaedicsurgery
AT johnsonalex multimodalimagingmodelbasedtrackingandmixedrealityvisualisationfororthopaedicsurgery
AT fotouhijavad multimodalimagingmodelbasedtrackingandmixedrealityvisualisationfororthopaedicsurgery
AT osgoodgreg multimodalimagingmodelbasedtrackingandmixedrealityvisualisationfororthopaedicsurgery
AT tombarifederico multimodalimagingmodelbasedtrackingandmixedrealityvisualisationfororthopaedicsurgery
AT navabnassir multimodalimagingmodelbasedtrackingandmixedrealityvisualisationfororthopaedicsurgery