Cargando…

Automated annotation of birdsong with a neural network that segments spectrograms

Songbirds provide a powerful model system for studying sensory-motor learning. However, many analyses of birdsong require time-consuming, manual annotation of its elements, called syllables. Automated methods for annotation have been proposed, but these methods assume that audio can be cleanly segme...

Descripción completa

Detalles Bibliográficos
Autores principales: Cohen, Yarden, Nicholson, David Aaron, Sanchioni, Alexa, Mallaber, Emily K, Skidanova, Viktoriya, Gardner, Timothy J
Formato: Online Artículo Texto
Lenguaje:English
Publicado: eLife Sciences Publications, Ltd 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8860439/
https://www.ncbi.nlm.nih.gov/pubmed/35050849
http://dx.doi.org/10.7554/eLife.63853
_version_ 1784654673530060800
author Cohen, Yarden
Nicholson, David Aaron
Sanchioni, Alexa
Mallaber, Emily K
Skidanova, Viktoriya
Gardner, Timothy J
author_facet Cohen, Yarden
Nicholson, David Aaron
Sanchioni, Alexa
Mallaber, Emily K
Skidanova, Viktoriya
Gardner, Timothy J
author_sort Cohen, Yarden
collection PubMed
description Songbirds provide a powerful model system for studying sensory-motor learning. However, many analyses of birdsong require time-consuming, manual annotation of its elements, called syllables. Automated methods for annotation have been proposed, but these methods assume that audio can be cleanly segmented into syllables, or they require carefully tuning multiple statistical models. Here, we present TweetyNet: a single neural network model that learns how to segment spectrograms of birdsong into annotated syllables. We show that TweetyNet mitigates limitations of methods that rely on segmented audio. We also show that TweetyNet performs well across multiple individuals from two species of songbirds, Bengalese finches and canaries. Lastly, we demonstrate that using TweetyNet we can accurately annotate very large datasets containing multiple days of song, and that these predicted annotations replicate key findings from behavioral studies. In addition, we provide open-source software to assist other researchers, and a large dataset of annotated canary song that can serve as a benchmark. We conclude that TweetyNet makes it possible to address a wide range of new questions about birdsong.
format Online
Article
Text
id pubmed-8860439
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher eLife Sciences Publications, Ltd
record_format MEDLINE/PubMed
spelling pubmed-88604392022-02-23 Automated annotation of birdsong with a neural network that segments spectrograms Cohen, Yarden Nicholson, David Aaron Sanchioni, Alexa Mallaber, Emily K Skidanova, Viktoriya Gardner, Timothy J eLife Neuroscience Songbirds provide a powerful model system for studying sensory-motor learning. However, many analyses of birdsong require time-consuming, manual annotation of its elements, called syllables. Automated methods for annotation have been proposed, but these methods assume that audio can be cleanly segmented into syllables, or they require carefully tuning multiple statistical models. Here, we present TweetyNet: a single neural network model that learns how to segment spectrograms of birdsong into annotated syllables. We show that TweetyNet mitigates limitations of methods that rely on segmented audio. We also show that TweetyNet performs well across multiple individuals from two species of songbirds, Bengalese finches and canaries. Lastly, we demonstrate that using TweetyNet we can accurately annotate very large datasets containing multiple days of song, and that these predicted annotations replicate key findings from behavioral studies. In addition, we provide open-source software to assist other researchers, and a large dataset of annotated canary song that can serve as a benchmark. We conclude that TweetyNet makes it possible to address a wide range of new questions about birdsong. eLife Sciences Publications, Ltd 2022-01-20 /pmc/articles/PMC8860439/ /pubmed/35050849 http://dx.doi.org/10.7554/eLife.63853 Text en © 2022, Cohen et al https://creativecommons.org/licenses/by/4.0/This article is distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use and redistribution provided that the original author and source are credited.
spellingShingle Neuroscience
Cohen, Yarden
Nicholson, David Aaron
Sanchioni, Alexa
Mallaber, Emily K
Skidanova, Viktoriya
Gardner, Timothy J
Automated annotation of birdsong with a neural network that segments spectrograms
title Automated annotation of birdsong with a neural network that segments spectrograms
title_full Automated annotation of birdsong with a neural network that segments spectrograms
title_fullStr Automated annotation of birdsong with a neural network that segments spectrograms
title_full_unstemmed Automated annotation of birdsong with a neural network that segments spectrograms
title_short Automated annotation of birdsong with a neural network that segments spectrograms
title_sort automated annotation of birdsong with a neural network that segments spectrograms
topic Neuroscience
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8860439/
https://www.ncbi.nlm.nih.gov/pubmed/35050849
http://dx.doi.org/10.7554/eLife.63853
work_keys_str_mv AT cohenyarden automatedannotationofbirdsongwithaneuralnetworkthatsegmentsspectrograms
AT nicholsondavidaaron automatedannotationofbirdsongwithaneuralnetworkthatsegmentsspectrograms
AT sanchionialexa automatedannotationofbirdsongwithaneuralnetworkthatsegmentsspectrograms
AT mallaberemilyk automatedannotationofbirdsongwithaneuralnetworkthatsegmentsspectrograms
AT skidanovaviktoriya automatedannotationofbirdsongwithaneuralnetworkthatsegmentsspectrograms
AT gardnertimothyj automatedannotationofbirdsongwithaneuralnetworkthatsegmentsspectrograms