Cargando…

Image Compositing for Segmentation of Surgical Tools Without Manual Annotations

Producing manual, pixel-accurate, image segmentation labels is tedious and time-consuming. This is often a rate-limiting factor when large amounts of labeled images are required, such as for training deep convolutional networks for instrument-background segmentation in surgical scenes. No large data...

Descripción completa

Detalles Bibliográficos
Formato: Online Artículo Texto
Lenguaje:English
Publicado: IEEE 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8092331/
https://www.ncbi.nlm.nih.gov/pubmed/33556005
http://dx.doi.org/10.1109/TMI.2021.3057884
_version_ 1783687640832278528
collection PubMed
description Producing manual, pixel-accurate, image segmentation labels is tedious and time-consuming. This is often a rate-limiting factor when large amounts of labeled images are required, such as for training deep convolutional networks for instrument-background segmentation in surgical scenes. No large datasets comparable to industry standards in the computer vision community are available for this task. To circumvent this problem, we propose to automate the creation of a realistic training dataset by exploiting techniques stemming from special effects and harnessing them to target training performance rather than visual appeal. Foreground data is captured by placing sample surgical instruments over a chroma key (a.k.a. green screen) in a controlled environment, thereby making extraction of the relevant image segment straightforward. Multiple lighting conditions and viewpoints can be captured and introduced in the simulation by moving the instruments and camera and modulating the light source. Background data is captured by collecting videos that do not contain instruments. In the absence of pre-existing instrument-free background videos, minimal labeling effort is required, just to select frames that do not contain surgical instruments from videos of surgical interventions freely available online. We compare different methods to blend instruments over tissue and propose a novel data augmentation approach that takes advantage of the plurality of options. We show that by training a vanilla U-Net on semi-synthetic data only and applying a simple post-processing, we are able to match the results of the same network trained on a publicly available manually labeled real dataset.
format Online
Article
Text
id pubmed-8092331
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher IEEE
record_format MEDLINE/PubMed
spelling pubmed-80923312021-05-04 Image Compositing for Segmentation of Surgical Tools Without Manual Annotations IEEE Trans Med Imaging Article Producing manual, pixel-accurate, image segmentation labels is tedious and time-consuming. This is often a rate-limiting factor when large amounts of labeled images are required, such as for training deep convolutional networks for instrument-background segmentation in surgical scenes. No large datasets comparable to industry standards in the computer vision community are available for this task. To circumvent this problem, we propose to automate the creation of a realistic training dataset by exploiting techniques stemming from special effects and harnessing them to target training performance rather than visual appeal. Foreground data is captured by placing sample surgical instruments over a chroma key (a.k.a. green screen) in a controlled environment, thereby making extraction of the relevant image segment straightforward. Multiple lighting conditions and viewpoints can be captured and introduced in the simulation by moving the instruments and camera and modulating the light source. Background data is captured by collecting videos that do not contain instruments. In the absence of pre-existing instrument-free background videos, minimal labeling effort is required, just to select frames that do not contain surgical instruments from videos of surgical interventions freely available online. We compare different methods to blend instruments over tissue and propose a novel data augmentation approach that takes advantage of the plurality of options. We show that by training a vanilla U-Net on semi-synthetic data only and applying a simple post-processing, we are able to match the results of the same network trained on a publicly available manually labeled real dataset. IEEE 2021-02-08 /pmc/articles/PMC8092331/ /pubmed/33556005 http://dx.doi.org/10.1109/TMI.2021.3057884 Text en https://creativecommons.org/licenses/by/4.0/This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
spellingShingle Article
Image Compositing for Segmentation of Surgical Tools Without Manual Annotations
title Image Compositing for Segmentation of Surgical Tools Without Manual Annotations
title_full Image Compositing for Segmentation of Surgical Tools Without Manual Annotations
title_fullStr Image Compositing for Segmentation of Surgical Tools Without Manual Annotations
title_full_unstemmed Image Compositing for Segmentation of Surgical Tools Without Manual Annotations
title_short Image Compositing for Segmentation of Surgical Tools Without Manual Annotations
title_sort image compositing for segmentation of surgical tools without manual annotations
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8092331/
https://www.ncbi.nlm.nih.gov/pubmed/33556005
http://dx.doi.org/10.1109/TMI.2021.3057884
work_keys_str_mv AT imagecompositingforsegmentationofsurgicaltoolswithoutmanualannotations
AT imagecompositingforsegmentationofsurgicaltoolswithoutmanualannotations
AT imagecompositingforsegmentationofsurgicaltoolswithoutmanualannotations
AT imagecompositingforsegmentationofsurgicaltoolswithoutmanualannotations
AT imagecompositingforsegmentationofsurgicaltoolswithoutmanualannotations
AT imagecompositingforsegmentationofsurgicaltoolswithoutmanualannotations