Cargando…

Validation and tuning of in situ transcriptomics image processing workflows with crowdsourced annotations

Recent advancements in in situ methods, such as multiplexed in situ RNA hybridization and in situ RNA sequencing, have deepened our understanding of the way biological processes are spatially organized in tissues. Automated image processing and spot-calling algorithms for analyzing in situ transcrip...

Descripción completa

Detalles Bibliográficos
Autores principales: Vo-Phamhi, Jenny M., Yamauchi, Kevin A., Gómez-Sjöberg, Rafael
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8376178/
https://www.ncbi.nlm.nih.gov/pubmed/34370726
http://dx.doi.org/10.1371/journal.pcbi.1009274
_version_ 1783740448208060416
author Vo-Phamhi, Jenny M.
Yamauchi, Kevin A.
Gómez-Sjöberg, Rafael
author_facet Vo-Phamhi, Jenny M.
Yamauchi, Kevin A.
Gómez-Sjöberg, Rafael
author_sort Vo-Phamhi, Jenny M.
collection PubMed
description Recent advancements in in situ methods, such as multiplexed in situ RNA hybridization and in situ RNA sequencing, have deepened our understanding of the way biological processes are spatially organized in tissues. Automated image processing and spot-calling algorithms for analyzing in situ transcriptomics images have many parameters which need to be tuned for optimal detection. Having ground truth datasets (images where there is very high confidence on the accuracy of the detected spots) is essential for evaluating these algorithms and tuning their parameters. We present a first-in-kind open-source toolkit and framework for in situ transcriptomics image analysis that incorporates crowdsourced annotations, alongside expert annotations, as a source of ground truth for the analysis of in situ transcriptomics images. The kit includes tools for preparing images for crowdsourcing annotation to optimize crowdsourced workers’ ability to annotate these images reliably, performing quality control (QC) on worker annotations, extracting candidate parameters for spot-calling algorithms from sample images, tuning parameters for spot-calling algorithms, and evaluating spot-calling algorithms and worker performance. These tools are wrapped in a modular pipeline with a flexible structure that allows users to take advantage of crowdsourced annotations from any source of their choice. We tested the pipeline using real and synthetic in situ transcriptomics images and annotations from the Amazon Mechanical Turk system obtained via Quanti.us. Using real images from in situ experiments and simulated images produced by one of the tools in the kit, we studied worker sensitivity to spot characteristics and established rules for annotation QC. We explored and demonstrated the use of ground truth generated in this way for validating spot-calling algorithms and tuning their parameters, and confirmed that consensus crowdsourced annotations are a viable substitute for expert-generated ground truth for these purposes.
format Online
Article
Text
id pubmed-8376178
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-83761782021-08-20 Validation and tuning of in situ transcriptomics image processing workflows with crowdsourced annotations Vo-Phamhi, Jenny M. Yamauchi, Kevin A. Gómez-Sjöberg, Rafael PLoS Comput Biol Research Article Recent advancements in in situ methods, such as multiplexed in situ RNA hybridization and in situ RNA sequencing, have deepened our understanding of the way biological processes are spatially organized in tissues. Automated image processing and spot-calling algorithms for analyzing in situ transcriptomics images have many parameters which need to be tuned for optimal detection. Having ground truth datasets (images where there is very high confidence on the accuracy of the detected spots) is essential for evaluating these algorithms and tuning their parameters. We present a first-in-kind open-source toolkit and framework for in situ transcriptomics image analysis that incorporates crowdsourced annotations, alongside expert annotations, as a source of ground truth for the analysis of in situ transcriptomics images. The kit includes tools for preparing images for crowdsourcing annotation to optimize crowdsourced workers’ ability to annotate these images reliably, performing quality control (QC) on worker annotations, extracting candidate parameters for spot-calling algorithms from sample images, tuning parameters for spot-calling algorithms, and evaluating spot-calling algorithms and worker performance. These tools are wrapped in a modular pipeline with a flexible structure that allows users to take advantage of crowdsourced annotations from any source of their choice. We tested the pipeline using real and synthetic in situ transcriptomics images and annotations from the Amazon Mechanical Turk system obtained via Quanti.us. Using real images from in situ experiments and simulated images produced by one of the tools in the kit, we studied worker sensitivity to spot characteristics and established rules for annotation QC. We explored and demonstrated the use of ground truth generated in this way for validating spot-calling algorithms and tuning their parameters, and confirmed that consensus crowdsourced annotations are a viable substitute for expert-generated ground truth for these purposes. Public Library of Science 2021-08-09 /pmc/articles/PMC8376178/ /pubmed/34370726 http://dx.doi.org/10.1371/journal.pcbi.1009274 Text en © 2021 Vo-Phamhi et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Vo-Phamhi, Jenny M.
Yamauchi, Kevin A.
Gómez-Sjöberg, Rafael
Validation and tuning of in situ transcriptomics image processing workflows with crowdsourced annotations
title Validation and tuning of in situ transcriptomics image processing workflows with crowdsourced annotations
title_full Validation and tuning of in situ transcriptomics image processing workflows with crowdsourced annotations
title_fullStr Validation and tuning of in situ transcriptomics image processing workflows with crowdsourced annotations
title_full_unstemmed Validation and tuning of in situ transcriptomics image processing workflows with crowdsourced annotations
title_short Validation and tuning of in situ transcriptomics image processing workflows with crowdsourced annotations
title_sort validation and tuning of in situ transcriptomics image processing workflows with crowdsourced annotations
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8376178/
https://www.ncbi.nlm.nih.gov/pubmed/34370726
http://dx.doi.org/10.1371/journal.pcbi.1009274
work_keys_str_mv AT vophamhijennym validationandtuningofinsitutranscriptomicsimageprocessingworkflowswithcrowdsourcedannotations
AT yamauchikevina validationandtuningofinsitutranscriptomicsimageprocessingworkflowswithcrowdsourcedannotations
AT gomezsjobergrafael validationandtuningofinsitutranscriptomicsimageprocessingworkflowswithcrowdsourcedannotations