2016
DOI: 10.1109/taslp.2016.2587218
|View full text |Cite
|
Sign up to set email alerts
|

A Morphological Model for Simulating Acoustic Scenes and Its Application to Sound Event Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2017
2017
2019
2019

Publication Types

Select...
5
4
1

Relationship

4
6

Authors

Journals

citations
Cited by 22 publications
(13 citation statements)
references
References 44 publications
0
13
0
Order By: Relevance
“…Audio data for Task 2 contains instances of 11 sound classes related to office sounds: clearing throat, coughing, door knock, door slam, drawer, human laughter, keyboard, keys (placed on a table), page turning, phone ringing, and speech. Audio sequences for this task were created from isolated sound events using the sound scene synthesizer of [50]. Recordings of isolated sound events were made at LS2N,École Centrale de Nantes, using a shotgun microphone AT8035 connected to a ZOOM H4n recorder.…”
Section: A Dataset and Experimental Setupmentioning
confidence: 99%
“…Audio data for Task 2 contains instances of 11 sound classes related to office sounds: clearing throat, coughing, door knock, door slam, drawer, human laughter, keyboard, keys (placed on a table), page turning, phone ringing, and speech. Audio sequences for this task were created from isolated sound events using the sound scene synthesizer of [50]. Recordings of isolated sound events were made at LS2N,École Centrale de Nantes, using a shotgun microphone AT8035 connected to a ZOOM H4n recorder.…”
Section: A Dataset and Experimental Setupmentioning
confidence: 99%
“…One challenge in this is that these mixtures should be created such that they mimic real-life data, and this is not trivial. Until now, the synthetic data used in DCASE tasks was rather simplisticfor example the DCASE 2016 Task 2 synthetic audio dataset used a morphological model for creating the mixtures [50], but it was based on a very small number of event instances, while this year's Task 2 rare sound events dataset did not use any specific knowledge or rules for background and target event combinations.…”
Section: I C O N C L U S I O N Smentioning
confidence: 99%
“…The dataset, denoted 'OS test', contains 12 recordings of 2 minutes duration each, with different event density levels and different event-to-background ratio levels. The recordings were generated using the acoustic scene synthesizer of [31] by concatenating isolated office sounds recorded at Queen Mary University of London (using different sound sources than the ones used for the OS development dataset of subsection IV-A). This polyphonic dataset allows for direct comparison with other participating systems for the DCASE 2013 polyphonic event detection task.…”
Section: B Test Datamentioning
confidence: 99%