2020
DOI: 10.1016/j.asoc.2020.106073
|View full text |Cite
|
Sign up to set email alerts
|

Masked Conditional Neural Networks for sound classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
13
0
2

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 39 publications
(21 citation statements)
references
References 58 publications
0
13
0
2
Order By: Relevance
“…• An Ontology-Aware Framework for Audio Event Classification [65] • Masked Conditional Neural Networks for sound classification [43] • CascadeML: An Automatic Neural Network Architecture Evolution and Training Algorithm for Multi-label Classification [46] We now discuss how our work differs from these approaches.…”
Section: Cascading Architecturesmentioning
confidence: 99%
See 1 more Smart Citation
“…• An Ontology-Aware Framework for Audio Event Classification [65] • Masked Conditional Neural Networks for sound classification [43] • CascadeML: An Automatic Neural Network Architecture Evolution and Training Algorithm for Multi-label Classification [46] We now discuss how our work differs from these approaches.…”
Section: Cascading Architecturesmentioning
confidence: 99%
“…Both [43] and our work utilize neural networks for acoustic classification. However, the conditional component of [43] does not correspond with multi-label dependence proposed in our cascading architecture but rather conditional dependence in the data itself.…”
Section: Cascading Architecturesmentioning
confidence: 99%
“…The highest accuracy is achieved by the model proposed in [44], which presents a CNN without max-pooling function and using Log-Mel audio feature extraction (CNN-Model-2 (No-maxpooling) + Log-Mel + augmentation). In [47], the application of the Masked Conditional Neural Networks for sound classification (MCLNN) to the problem of music genre classification, and ESR is presented. Both proposed models achieve high accuracy, but they are large models demanding prohibitive memory requirements for embedded platforms.…”
Section: Description Of the Modelmentioning
confidence: 99%
“…Medhat et al [14] suggest a binary-masked CNN model that adopts a controlled systematic sparseness such as embedding a filterbank-like behavior within the network to preserve the spatial locality of features in the process of e experiments on GTZAN and UrbanSound8K have achieved the accuracies of 85.1% and 74.2%, respectively. In their model, they introduce a set of hidden layers in which each neuron establishes a connection with the input by the activation function through the influence of distinct active regions in the feature vector, so the spatial information of the learned feature is saved as these active weights' locations which are fixed in position.…”
Section: Related Workmentioning
confidence: 99%