Cargando…

DeepFly3D, a deep learning-based approach for 3D limb and appendage tracking in tethered, adult Drosophila

Studying how neural circuits orchestrate limbed behaviors requires the precise measurement of the positions of each appendage in three-dimensional (3D) space. Deep neural networks can estimate two-dimensional (2D) pose in freely behaving and tethered animals. However, the unique challenges associate...

Descripción completa

Detalles Bibliográficos
Autores principales: Günel, Semih, Rhodin, Helge, Morales, Daniel, Campagnolo, João, Ramdya, Pavan, Fua, Pascal
Formato: Online Artículo Texto
Lenguaje:English
Publicado: eLife Sciences Publications, Ltd 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6828327/
https://www.ncbi.nlm.nih.gov/pubmed/31584428
http://dx.doi.org/10.7554/eLife.48571
_version_ 1783465373879762944
author Günel, Semih
Rhodin, Helge
Morales, Daniel
Campagnolo, João
Ramdya, Pavan
Fua, Pascal
author_facet Günel, Semih
Rhodin, Helge
Morales, Daniel
Campagnolo, João
Ramdya, Pavan
Fua, Pascal
author_sort Günel, Semih
collection PubMed
description Studying how neural circuits orchestrate limbed behaviors requires the precise measurement of the positions of each appendage in three-dimensional (3D) space. Deep neural networks can estimate two-dimensional (2D) pose in freely behaving and tethered animals. However, the unique challenges associated with transforming these 2D measurements into reliable and precise 3D poses have not been addressed for small animals including the fly, Drosophila melanogaster. Here, we present DeepFly3D, a software that infers the 3D pose of tethered, adult Drosophila using multiple camera images. DeepFly3D does not require manual calibration, uses pictorial structures to automatically detect and correct pose estimation errors, and uses active learning to iteratively improve performance. We demonstrate more accurate unsupervised behavioral embedding using 3D joint angles rather than commonly used 2D pose data. Thus, DeepFly3D enables the automated acquisition of Drosophila behavioral measurements at an unprecedented level of detail for a variety of biological applications.
format Online
Article
Text
id pubmed-6828327
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher eLife Sciences Publications, Ltd
record_format MEDLINE/PubMed
spelling pubmed-68283272019-11-06 DeepFly3D, a deep learning-based approach for 3D limb and appendage tracking in tethered, adult Drosophila Günel, Semih Rhodin, Helge Morales, Daniel Campagnolo, João Ramdya, Pavan Fua, Pascal eLife Neuroscience Studying how neural circuits orchestrate limbed behaviors requires the precise measurement of the positions of each appendage in three-dimensional (3D) space. Deep neural networks can estimate two-dimensional (2D) pose in freely behaving and tethered animals. However, the unique challenges associated with transforming these 2D measurements into reliable and precise 3D poses have not been addressed for small animals including the fly, Drosophila melanogaster. Here, we present DeepFly3D, a software that infers the 3D pose of tethered, adult Drosophila using multiple camera images. DeepFly3D does not require manual calibration, uses pictorial structures to automatically detect and correct pose estimation errors, and uses active learning to iteratively improve performance. We demonstrate more accurate unsupervised behavioral embedding using 3D joint angles rather than commonly used 2D pose data. Thus, DeepFly3D enables the automated acquisition of Drosophila behavioral measurements at an unprecedented level of detail for a variety of biological applications. eLife Sciences Publications, Ltd 2019-10-04 /pmc/articles/PMC6828327/ /pubmed/31584428 http://dx.doi.org/10.7554/eLife.48571 Text en © 2019, Günel et al http://creativecommons.org/licenses/by/4.0/ http://creativecommons.org/licenses/by/4.0/This article is distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use and redistribution provided that the original author and source are credited.
spellingShingle Neuroscience
Günel, Semih
Rhodin, Helge
Morales, Daniel
Campagnolo, João
Ramdya, Pavan
Fua, Pascal
DeepFly3D, a deep learning-based approach for 3D limb and appendage tracking in tethered, adult Drosophila
title DeepFly3D, a deep learning-based approach for 3D limb and appendage tracking in tethered, adult Drosophila
title_full DeepFly3D, a deep learning-based approach for 3D limb and appendage tracking in tethered, adult Drosophila
title_fullStr DeepFly3D, a deep learning-based approach for 3D limb and appendage tracking in tethered, adult Drosophila
title_full_unstemmed DeepFly3D, a deep learning-based approach for 3D limb and appendage tracking in tethered, adult Drosophila
title_short DeepFly3D, a deep learning-based approach for 3D limb and appendage tracking in tethered, adult Drosophila
title_sort deepfly3d, a deep learning-based approach for 3d limb and appendage tracking in tethered, adult drosophila
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6828327/
https://www.ncbi.nlm.nih.gov/pubmed/31584428
http://dx.doi.org/10.7554/eLife.48571
work_keys_str_mv AT gunelsemih deepfly3dadeeplearningbasedapproachfor3dlimbandappendagetrackingintetheredadultdrosophila
AT rhodinhelge deepfly3dadeeplearningbasedapproachfor3dlimbandappendagetrackingintetheredadultdrosophila
AT moralesdaniel deepfly3dadeeplearningbasedapproachfor3dlimbandappendagetrackingintetheredadultdrosophila
AT campagnolojoao deepfly3dadeeplearningbasedapproachfor3dlimbandappendagetrackingintetheredadultdrosophila
AT ramdyapavan deepfly3dadeeplearningbasedapproachfor3dlimbandappendagetrackingintetheredadultdrosophila
AT fuapascal deepfly3dadeeplearningbasedapproachfor3dlimbandappendagetrackingintetheredadultdrosophila