Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems 2021
DOI: 10.1145/3411763.3451657
|View full text |Cite
|
Sign up to set email alerts
|

A Texture Superpixel Approach to Semantic Material Classification for Acoustic Geometry Tagging

Abstract: The current state of audio rendering algorithms allows efcient sound propagation, refecting realistic acoustic properties of real environments. Among factors afecting realism of acoustic simulations is the mapping between an environment's geometry, and acoustic information of materials represented. We present a pipeline to infer material characteristics from their visual representations, providing an automated mapping. A trained image classifer estimates semantic material information from textured meshes mappi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

1
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 30 publications
1
2
0
Order By: Relevance
“…The network, pretrained on ImageNet [33], learns on 32x32 pixel resolution image patches that we extract from appearances sampled from OpenSurfaces, assembling a dataset of about 13M images, split into 9M and 4M train and evaluation sets, respectively. The model converges in 45 epochs, achieves validation accuracy of about 0.83, and, as shown in recent experiments [26], it can replace humans in the process of acoustic material tagging in enclosed scenes of real-world space. Via a one-to-many mapping, we associate each of the 34 categories from the OpenSurfaces dataset to acoustic materials.…”
Section: Acoustic Materials Classifiersupporting
confidence: 55%
See 2 more Smart Citations
“…The network, pretrained on ImageNet [33], learns on 32x32 pixel resolution image patches that we extract from appearances sampled from OpenSurfaces, assembling a dataset of about 13M images, split into 9M and 4M train and evaluation sets, respectively. The model converges in 45 epochs, achieves validation accuracy of about 0.83, and, as shown in recent experiments [26], it can replace humans in the process of acoustic material tagging in enclosed scenes of real-world space. Via a one-to-many mapping, we associate each of the 34 categories from the OpenSurfaces dataset to acoustic materials.…”
Section: Acoustic Materials Classifiersupporting
confidence: 55%
“…In-the-wild datasets, such as OpenSurfaces [24], and Mat-terport3D [25], capture material appearances in surfaces with sub-optimal lighting conditions, specular reflections and errors introduced by optic sensors, which may result in noisy image representations of materials. Classification of image patches through densely connected networks can achieve reasonable accuracy when trained on extracted features, with the condition that the data has a comprehensive representation of materials applicable to target scenes of the acoustic rendering pipeline, as shown in recent work by Colombo et al [26]. Based on captured real environments, the authors automated the mapping between the visual representation of materials, expressed as meshes whose textures are decomposed into superpixels to infer their acoustic characteristics via a learned classifier.…”
Section: Materials Recognitionmentioning
confidence: 99%
See 1 more Smart Citation