2018
DOI: 10.1177/0023830918765012
|View full text |Cite
|
Sign up to set email alerts
|

Investigating Perceptual Biases, Data Reliability, and Data Discovery in a Methodology for Collecting Speech Errors From Audio Recordings

Abstract: This work describes a methodology of collecting speech errors from audio recordings and investigates how some of its assumptions affect data quality and composition. Speech errors of all types (sound, lexical, syntactic, etc.) were collected by eight data collectors from audio recordings of unscripted English speech. Analysis of these errors showed that: (i) different listeners find different errors in the same audio recordings, but (ii) the frequencies of error patterns are similar across listeners; (iii) err… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
28
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
1
1

Relationship

2
5

Authors

Journals

citations
Cited by 17 publications
(31 citation statements)
references
References 77 publications
(122 reference statements)
0
28
0
Order By: Relevance
“…The methods used to collect errors can be characterized with four methodological decisions and practices (see Alderete & Davies (2018) for a more detailed explanation of these methods). First, speech errors were collected primarily from audio recordings of natural conversations.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The methods used to collect errors can be characterized with four methodological decisions and practices (see Alderete & Davies (2018) for a more detailed explanation of these methods). First, speech errors were collected primarily from audio recordings of natural conversations.…”
Section: Methodsmentioning
confidence: 99%
“…We do this by both reviewing the empirical generalizations that characterize regularity in sublexical speech errors, and also by contributing new evidence on regularity in English speech errors. In particular, we will investigate phonological regularity in the Simon Fraser University Speech Error Database (SFUSED), a database of naturalistic speech errors collected in a way that is demonstrably less prone to methodological problems found in prior studies (Alderete & Davies, 2018). It turns out that these problems have had an effect on the characterization of phonological regularity, and so our revised characterization of regularity informs the models designed to account for it.…”
mentioning
confidence: 99%
“…Word retrieval failures and speech errors can occur at any one of these points during the spread of activation (e.g., a semantically related word may inadvertently receive more activation than the target word, resulting in a “semantic” error). Examination of such retrieval failures can provide insight into language processing and the structure of the mental lexicon, as errors do not occur randomly (Baars, 1992; Fay & Cutler, 1977); however, word retrieval failures and speech errors are infrequent in daily communication of typical adults, even when experimentally induced in laboratory settings (but see Alderete & Davies, 2019; Vitevitch et al, 2015 for methods to overcome this issue). Hence, researchers have often relied on word retrieval failures and speech errors from clinical populations, such as persons with aphasia, to provide a rich data source of word retrieval errors.…”
Section: Introductionmentioning
confidence: 99%
“…The corpus thus reflects a better sample of the true population of speech errors than other studies that do not use audio recordings and do not use pairs of trained listeners. As documented in Alderete and Davies (2018), prior studies have detection rates (an error on average every 5 to 6 minutes) that undershoot our detection rates by a wide margin. Our corpus also has a relatively low rate of uncommon but highly salient errors, like sound exchanges, and higher rates of phonotactic violations, which are further indicators of higher data reliability (Alderete & Tupper, 2018).…”
Section: The Corpus: Sfused Cantonese 10mentioning
confidence: 74%
“…This latter corpus is reported to have 13 whole syllable errors, which is 0.36% of all sound errors, far lower than Chen's estimate for Mandarin. While it could be the case that Mandarin and English simply have different rates of syllable errors, recall that speech error data collection is plagued with methodological problems (Alderete & Davies, 2018;Ferber, 1995), and so it seems more prudent to first understand why the Chen corpus (2000) has such a low rate of sound errors.…”
Section: 4mentioning
confidence: 99%