2022
DOI: 10.7554/elife.63853
|View full text |Cite
|
Sign up to set email alerts
|

Automated annotation of birdsong with a neural network that segments spectrograms

Abstract: Songbirds provide a powerful model system for studying sensory-motor learning. However, many analyses of birdsong require time-consuming, manual annotation of its elements, called syllables. Automated methods for annotation have been proposed, but these methods assume that audio can be cleanly segmented into syllables, or they require carefully tuning multiple statistical models. Here we present TweetyNet: a single neural network model that learns how to segment spectrograms of birdsong into annotated syllable… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
42
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 47 publications
(56 citation statements)
references
References 84 publications
(132 reference statements)
0
42
0
Order By: Relevance
“…We did so to account for potential circadian effects on song production. We also reassessed our results (shown in Figure 2 ) by analyzing only syllable renditions produced between 6 PM and 8 PM using new methods for automated labeling of song syllables ( Cohen et al, 2022 ). We found no statistically significant difference in learning magnitudes between the two forms of analysis ( Figure 2—figure supplement 7a , 0.167 < P boot < 0.951 on all days of training).…”
Section: Methodsmentioning
confidence: 81%
“…We did so to account for potential circadian effects on song production. We also reassessed our results (shown in Figure 2 ) by analyzing only syllable renditions produced between 6 PM and 8 PM using new methods for automated labeling of song syllables ( Cohen et al, 2022 ). We found no statistically significant difference in learning magnitudes between the two forms of analysis ( Figure 2—figure supplement 7a , 0.167 < P boot < 0.951 on all days of training).…”
Section: Methodsmentioning
confidence: 81%
“…We arbitrarily chose a subset of recordings to annotate from each species, starting with Zonotrichia leucophrys which had the largest dataset, and moving down to Calypte anna which had the smallest dataset. As few as three minutes of recording is sufficient for the TweetyNet algorithm to perform accurately [50], so we ensured that all species had at least 180 seconds worth of annotated syllables. We then added more recordings for species with more data to investigate the influence of sample size.…”
Section: Methodsmentioning
confidence: 99%
“…With respect to bioacoustics, an extractive algorithm could, for example, segment out syllables within vocalizations. A recently developed application, TweetyNet, was released to perform just this task [50,85,86] using deep learning via ANNs. Specifically, TweetyNet uses convolutional and recurrent ANNs.…”
Section: Avian Bioacousticsmentioning
confidence: 99%
See 2 more Smart Citations