2019
DOI: 10.1371/journal.pcbi.1006711
|View full text |Cite
|
Sign up to set email alerts
|

A Gestalt inference model for auditory scene segregation

Abstract: Our current understanding of how the brain segregates auditory scenes into meaningful objects is in line with a Gestaltism framework. These Gestalt principles suggest a theory of how different attributes of the soundscape are extracted then bound together into separate groups that reflect different objects or streams present in the scene. These cues are thought to reflect the underlying statistical structure of natural sounds in a similar way that statistics of natural images are closely linked to the principl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
3
2

Relationship

2
8

Authors

Journals

citations
Cited by 19 publications
(10 citation statements)
references
References 127 publications
0
7
0
Order By: Relevance
“…The present model contributes to this effort by showing how adaptation, inhibition and noise may generate auditory bistability within a more comprehensive system. Past efforts that may contribute to our understanding of how brain-like computations can April 3, 2019 12/28 scale to larger systems fall short of this demonstration either because they include comprehensive dynamics of adaptation, inhibition and noise but abstract away from the specific challenges of auditory scene analysis [98][99][100][101] or because they can perform some form of auditory scene analysis but do not show any identified form of adaptation, inhibition and noise nor any bistable output [102][103][104]. Future efforts towards scaling of the present model could build on the strengths of any one of these past systems.…”
Section: Discussionmentioning
confidence: 99%
“…The present model contributes to this effort by showing how adaptation, inhibition and noise may generate auditory bistability within a more comprehensive system. Past efforts that may contribute to our understanding of how brain-like computations can April 3, 2019 12/28 scale to larger systems fall short of this demonstration either because they include comprehensive dynamics of adaptation, inhibition and noise but abstract away from the specific challenges of auditory scene analysis [98][99][100][101] or because they can perform some form of auditory scene analysis but do not show any identified form of adaptation, inhibition and noise nor any bistable output [102][103][104]. Future efforts towards scaling of the present model could build on the strengths of any one of these past systems.…”
Section: Discussionmentioning
confidence: 99%
“…Recent advances in deep neural networks (DNN) and auditory modeling have made it possible to achieve high-levels of sound recognition performance approaching human performance limits and such networks can predict various features of peripheral and central auditory processing [ 50 52 ]. Although these networks differ from the proposed HSNN in terms of the computing elements used, the computations performed and the network structure.…”
Section: Discussionmentioning
confidence: 99%
“…Some computational approaches to auditory scene analysis instead frame perceptual organization as clustering "bottom-up" features of audio (99,100). This approach has often been combined with bottom-up features inspired by neurophysiology (11,13,(101)(102)(103), but it remains difficult to identify features whose clustering is comprehensively predictive of perceptual organization. Our approach shares some abstract similarities in combining a bottom-up processing component with a top-down inference stage, but a key difference is that the bottomup component of our analysis-by-synthesis algorithm preserves uncertainty, outputting proposals for event latent variables that can be refined or rejected during inference.…”
Section: Clusteringmentioning
confidence: 99%