2018
DOI: 10.1007/978-3-030-00764-5_2
|View full text |Cite
|
Sign up to set email alerts
|

Mixup-Based Acoustic Scene Classification Using Multi-channel Convolutional Neural Network

Abstract: Audio scene classification, the problem of predicting class labels of audio scenes, has drawn lots of attention during the last several years. However, it remains challenging and falls short of accuracy and efficiency. Recently, Convolutional Neural Network (CNN)-based methods have achieved better performance with comparison to the traditional methods. Nevertheless, conventional single channel CNN may fail to consider the fact that additional cues may be embedded in the multi-channel recordings. In this paper,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
55
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
2
1

Relationship

3
5

Authors

Journals

citations
Cited by 77 publications
(55 citation statements)
references
References 15 publications
(15 reference statements)
0
55
0
Order By: Relevance
“…It is important to note that the use of mixup augmentation leads to no addition of more external datasets, so the proposed system is still under the conditions imposed by the Albayzín 2010 evaluation. Mixup augmentation has been shown to improve model generalisation capabilities in different domains, including some audio classification tasks [64]. In our set of experiments, mixup augmentation is applied directly in the feature space.…”
Section: Mixup Data Augmentationmentioning
confidence: 99%
“…It is important to note that the use of mixup augmentation leads to no addition of more external datasets, so the proposed system is still under the conditions imposed by the Albayzín 2010 evaluation. Mixup augmentation has been shown to improve model generalisation capabilities in different domains, including some audio classification tasks [64]. In our set of experiments, mixup augmentation is applied directly in the feature space.…”
Section: Mixup Data Augmentationmentioning
confidence: 99%
“…In deep learning, data augmentation, which is to increase the data variation by altering the property of the genuine data, is an important method to improve performance of the task at hand. Techniques like adding background noise [28,29], pitch shifting [29], and sample mixing [18,30], have been proven to be useful for environmental sound recognition in general. Motivated by the work of Tokozume et al [18], we pursue a between-class (BC) data augmentation approach that mixes two samples of different classes with a random factor to generate BC examples for network training.…”
Section: Between-class Data Augmentation and Kl-divergence Lossmentioning
confidence: 99%
“…MFCC follows similar generation procedure, except for the size. To prevent over-fitting, we apply mixup-data augmentation [22] with an ratio of 0.2 [23].…”
Section: Preprocessingmentioning
confidence: 99%