2015
DOI: 10.1515/lp-2015-0012
|View full text |Cite
|
Sign up to set email alerts
|

Models of dataset size, question design, and cross-language speech perception for speech crowdsourcing applications

Abstract: Transcribers make mistakes. Workers recruited in a crowdsourcing marketplace, because of their varying levels of commitment and education, make more mistakes than workers in a controlled laboratory setting. Methods for compensating transcriber mistakes are desirable because, with such methods available, crowdsourcing has the potential to significantly increase the scale of experiments in laboratory phonology. This paper provides a brief tutorial on statistical learning theory, introducing the relationship betw… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2016
2016
2019
2019

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 38 publications
0
4
0
Order By: Relevance
“…The development of the speech tagging tools in turn relies on the knowledge on how human listeners -the users of TTS systems -perceive and categorise prominence [1,7]. Recent advances show that crowdsourcing methods enable to directly access human prominence judgments in a relatively short time [8,9,10].…”
Section: Introductionmentioning
confidence: 99%
“…The development of the speech tagging tools in turn relies on the knowledge on how human listeners -the users of TTS systems -perceive and categorise prominence [1,7]. Recent advances show that crowdsourcing methods enable to directly access human prominence judgments in a relatively short time [8,9,10].…”
Section: Introductionmentioning
confidence: 99%
“…Despite methodological debates surrounding the trade-off between experimental control and stylistic variation (Xu 2010;Wagner et al 2015 and references therein), an increasing number of speech scientist use crowdsourcing platforms to collect linguistic data (Hasegawa-Johnson, Cole, Jyothi, & Varshney 2015). This suggests that researchers are willing to trade an increased level of variability and lack of control with quick and convenient access to large amounts of data.…”
Section: Limits and Challenges Of This Approachmentioning
confidence: 99%
“…As an illustrative example, consider the problem of mismatched crowdsourcing for speech transcription, which has garnered interest in the signal processing community [4,6,9,12,14,23]. Suppose the four possibilities for a velar stop consonant to transcribe are R = { , , , }.…”
Section: Con Dence Level Reportingmentioning
confidence: 99%