ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9054444
|View full text |Cite
|
Sign up to set email alerts
|

Learning With Out-of-Distribution Data for Audio Classification

Abstract: In supervised machine learning, the assumption that training data is labelled correctly is not always satisfied. In this paper, we investigate an instance of labelling error for classification tasks in which the dataset is corrupted with out-of-distribution (OOD) instances: data that does not belong to any of the target classes, but is labelled as such. We show that detecting and relabelling certain OOD instances, rather than discarding them, can have a positive effect on learning. The proposed method uses an … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(8 citation statements)
references
References 11 publications
(26 reference statements)
0
8
0
Order By: Relevance
“…We train a VGG classifier on the proper training data of the iD dataset to classify the classes in the set, using data augmentation to learn iD equivariance. Fifteen datapoints from each class in the iD training set are held out as calibration data for the set (this is the same setting as Iqbal et al (2020)'s validation set in FSD).…”
Section: Results On Audio Datamentioning
confidence: 99%
“…We train a VGG classifier on the proper training data of the iD dataset to classify the classes in the set, using data augmentation to learn iD equivariance. Fifteen datapoints from each class in the iD training set are held out as calibration data for the set (this is the same setting as Iqbal et al (2020)'s validation set in FSD).…”
Section: Results On Audio Datamentioning
confidence: 99%
“…In [10], the authors worked on a dataset that was manipulated with some samples that did not belong to any class because of a labeling error. They proved that using those OOD (out of distribution) samples by separating them from the dataset can affect the training of the model in an effective way.…”
Section: Criteria and Alternatives Selectionmentioning
confidence: 99%
“…In SER, label sets in not-small datasets are inherently noisy due to reasons like sub-optimality of automatic methods used in the creation, or the difficulty of annotating audio-especially without visual cues, with large vocabularies, and because the annotation process is, sometimes, inherently subjective and ambiguous. Consequently, recent works have shown the efficacy of label noise treatment in large datasets such as AudioSet [41,85] and mid-size datasets [32,86,87].…”
Section: A Characteristicsmentioning
confidence: 99%
“…DenseNet-121. DenseNets are reported to outperform ResNets for image recognition [121], and have been recently used for SET [86,87].…”
Section: A Evaluationmentioning
confidence: 99%