The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019
DOI: 10.1109/icassp.2019.8682558
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Match Transient Sound Events Using Attentional Similarity for Few-shot Sound Recognition

Abstract: In this paper, we introduce a novel attentional similarity module for the problem of few-shot sound recognition. Given a few examples of an unseen sound event, a classifier must be quickly adapted to recognize the new sound event without much fine-tuning. The proposed attentional similarity module can be plugged into any metric-based learning method for few-shot learning, allowing the resulting model to especially match related short sound events. Extensive experiments on two datasets show that the proposed mo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
51
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 46 publications
(52 citation statements)
references
References 12 publications
(17 reference statements)
0
51
1
Order By: Relevance
“…In the inference step, by using the feature space, the input is classified to one of the target classes. Regarding deep metric learning in acoustic signal processing [17][18][19][20][21][22][23][24][25][26], we summarize an overview of tasks, loss functions, and sampling strategies, in Table 1. Manocha et al have worked on sound clip search task and used contrastive loss, where a feature space is learned based on a pair type that consists of the same class or different classes and a feature space distance [19].…”
Section: Related Workmentioning
confidence: 99%
“…In the inference step, by using the feature space, the input is classified to one of the target classes. Regarding deep metric learning in acoustic signal processing [17][18][19][20][21][22][23][24][25][26], we summarize an overview of tasks, loss functions, and sampling strategies, in Table 1. Manocha et al have worked on sound clip search task and used contrastive loss, where a feature space is learned based on a pair type that consists of the same class or different classes and a feature space distance [19].…”
Section: Related Workmentioning
confidence: 99%
“…The characteristics of this paradigm are highly compatible with COVID-19's disease detection tasks. Inspired by this, the experimental method we adopted when pre-training the model is consistent with [7]. Each iteration randomly selects c categories from all categories, each category contains k samples, and selects one category from each of the c categories as the test set.…”
Section: Cough Classification Algorithmmentioning
confidence: 99%
“…Based on the training strategy of few-shot learning, we introduce an attention similarity to complete the task of cough classification [7]. Unlike the previous method of calculating similarity by pooling to the same length, it can directly receive input features of different lengths and calculate the attention similarity between the input features and a certain type of features.…”
Section: Cough Classification Algorithmmentioning
confidence: 99%
“…As an alternative, few-shot learning [9][10][11][12][13][14] has been applied to audio classification [15][16][17] and sound event detection [18,19], where a classifier must learn to recognize a novel class from very few examples. Among different few-shot learning methods, metricbased prototypical networks [12] have been shown to yield excellent performance for audio [15,18,19].…”
Section: Introductionmentioning
confidence: 99%