Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume 2021
DOI: 10.18653/v1/2021.eacl-main.127
|View full text |Cite
|
Sign up to set email alerts
|

A phonetic model of non-native spoken word processing

Abstract: Non-native speakers show difficulties with spoken word processing. Many studies attribute these difficulties to imprecise phonological encoding of words in the lexical memory. We test an alternative hypothesis: that some of these difficulties can arise from the non-native speakers' phonetic perception. We train a computational model of phonetic learning, which has no access to phonology, on either one or two languages. We first show that the model exhibits predictable behaviors on phone-level and word-level di… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
1
1

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 46 publications
(68 reference statements)
0
6
0
Order By: Relevance
“…Although the vast majority of previous work has been driven by the engineering applications of AWEs, there is a growing scientific interest in using deep neural networks as cognitive models of (human) speech processing [22,23,38,39]. Therefore, we argue that this cognitively motivated direction requires us to take a closer look at the embedding space and examine the degree to which we can rely on the emergent distance as an estimate of (perceptual) dissimilarity between linguistic units.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Although the vast majority of previous work has been driven by the engineering applications of AWEs, there is a growing scientific interest in using deep neural networks as cognitive models of (human) speech processing [22,23,38,39]. Therefore, we argue that this cognitively motivated direction requires us to take a closer look at the embedding space and examine the degree to which we can rely on the emergent distance as an estimate of (perceptual) dissimilarity between linguistic units.…”
Section: Discussionmentioning
confidence: 99%
“…Since AWE models have been recently adopted as cognitive models of infant phonetic learning [22] and cross-language non-native processing [23], we argue that more effort should be devoted to analyze and understand the emergent embedding space to make sure it behaves as expected. In this paper, we take a step in this direction and make the following contributions:…”
Section: Introductionmentioning
confidence: 99%
“…In this section, we train and test four neural network models on the same three datasets as before. These models have been proposed in speech technology research, in particular in low‐resource settings where transcribed data may not be available and showed high performance in word and phone discrimination tasks (Kamper et al., 2015; Kamper, 2019; Matusevych et al., 2021; Renshaw, Kamper, Jansen, & Goldwater, 2015). Fig.…”
Section: Study 2: Testing Other Modelsmentioning
confidence: 99%
“…In this section, we train and test four neural network models on the same three data sets as before. These models have been proposed in speech technology research, in particular in lowresource setting where transcribed data may not be available, and showed high performance in word and phone discrimination tasks (Kamper, 2019;Kamper et al, 2015;Matusevych, Kamper, Schatz, Feldman, & Goldwater, 2021;Renshaw, Kamper, Jansen, & Goldwater, 2015). Figure 2 schematically shows the difference between the models' architectures and input data.…”
Section: Study 2: Testing Other Modelsmentioning
confidence: 99%