2020 International Joint Conference on Neural Networks (IJCNN) 2020
DOI: 10.1109/ijcnn48605.2020.9207262
|View full text |Cite
|
Sign up to set email alerts
|

An Auto Encoder For Audio Dolphin Communication

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(10 citation statements)
references
References 12 publications
0
10
0
Order By: Relevance
“…Several recent approaches offer interesting alternatives or potential improvements. The direct use of spectrograms, either as an image or as a parameter matrix, has already been applied to mice (Premoli, et al, 2021), Atlantic spotted dolphins (Kohlsdorf, Herzing, & Starner, 2020), domestic cats (Pandeya, Kim, & Lee, 2018), common marmosets (Oikarinen, et al, 2019), etc.. However, their performance on complex and graded repertoires remain to be evaluated, and adaptation of spectrogram parameters to each species may be necessary (Knight, et al, 2020).…”
Section: Future Workmentioning
confidence: 99%
“…Several recent approaches offer interesting alternatives or potential improvements. The direct use of spectrograms, either as an image or as a parameter matrix, has already been applied to mice (Premoli, et al, 2021), Atlantic spotted dolphins (Kohlsdorf, Herzing, & Starner, 2020), domestic cats (Pandeya, Kim, & Lee, 2018), common marmosets (Oikarinen, et al, 2019), etc.. However, their performance on complex and graded repertoires remain to be evaluated, and adaptation of spectrogram parameters to each species may be necessary (Knight, et al, 2020).…”
Section: Future Workmentioning
confidence: 99%
“…The most common parametric dimensionality reduction algorithm is PCA, where a linear transform is learned between data and an embedding space. Similarly, neural networks such as autoencoders can be used to learn a set of basis features which can be complex and non-linear (Kohlsdorf et al, 2020 ; Sainburg et al, 2020c ; Goffinet et al, 2021 ; Singh Alvarado et al, 2021 ). For example, an autoencoder trained on images of faces can learn to linearize the presence of glasses or a beard (Radford et al, 2015 ; Sainburg et al, 2018b , 2021 ).…”
Section: Extracting Relational Structure and Clusteringmentioning
confidence: 99%
“…Like most areas of deep learning, substantial progress has been made on the task of audio synthesis in the past few years. Basic methods comprise autoencoders (Engel et al, 2017 ; Kohlsdorf et al, 2020 ; Sainburg et al, 2020c ), Generative Adversarial Networks (GANd) (Donahue et al, 2018 ; Engel et al, 2019 ; Sainburg et al, 2020c ; Tjandra et al, 2020 ; Pagliarini et al, 2021 ) and autoregressive approaches (Mehri et al, 2016 ; Oord et al, 2016 ; Kalchbrenner et al, 2018 ; Prenger et al, 2019 ). One advantage of GAN-based models is that their loss is not defined directly by reconstruction loss, resulting in higher-fidelity syntheses (Larsen et al, 2016 ).…”
Section: Synthesizing Vocalizationsmentioning
confidence: 99%
“…Following the success of Deep Learning, there has been a growing usage of neural network architectures for AAD. In particular, Au-toEncoders (AE) are becoming popular for unsupervised AAD [15,24]. When compared with other ML approaches (e.g., IF and OCSVM), AE present the advantage of requiring a lower computational effort [19].…”
Section: Introductionmentioning
confidence: 99%