2019
DOI: 10.1162/opmi_a_00022
|View full text |Cite
|
Sign up to set email alerts
|

Segmentability Differences Between Child-Directed and Adult-Directed Speech: A Systematic Test With an Ecologically Valid Corpus

Abstract: Previous computational modeling suggests it is much easier to segment words from child-directed speech (CDS) than adult-directed speech (ADS). However, this conclusion is based on data collected in the laboratory, with CDS from play sessions and ADS between a parent and an experimenter, which may not be representative of ecologically collected CDS and ADS. Fully naturalistic ADS and CDS collected with a nonintrusive recording device as the child went about her day were analyzed with a diverse set of algorithms… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
10
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
3

Relationship

2
7

Authors

Journals

citations
Cited by 14 publications
(11 citation statements)
references
References 33 publications
1
10
0
Order By: Relevance
“…Therefore, we adopt the Natural Language Processing/Speech Technology standard and use token recall and token precision (e.g., Ludusan, Versteegh, Jansen, Gravier, Cao, Johnson & Dupoux, 2014). This is also the approach adopted by previous work that attempts to compare the overall segmentability of different registers (childversus adult-directed speech, Cristia et al, 2019;Ludusan, Mazuka, Bernard, Cristia & Dupoux, 2017), and different languages (Caines, Altmann-Richer & Buttery, 2019;Loukatou, Stoll, Blasi & Cristia, 2018;Loukatou et al, 2019), or simply evaluate proposed algorithms (e.g., Daland & Pierrehumbert, 2011;Goldwater et al, 2009;Phillips & Pearl, 2014). These scores are calculated by comparing the output string, which contains hypothesized word breaks an algorithm supplies, against the original sentence containing word breaks.…”
Section: Discussionmentioning
confidence: 99%
“…Therefore, we adopt the Natural Language Processing/Speech Technology standard and use token recall and token precision (e.g., Ludusan, Versteegh, Jansen, Gravier, Cao, Johnson & Dupoux, 2014). This is also the approach adopted by previous work that attempts to compare the overall segmentability of different registers (childversus adult-directed speech, Cristia et al, 2019;Ludusan, Mazuka, Bernard, Cristia & Dupoux, 2017), and different languages (Caines, Altmann-Richer & Buttery, 2019;Loukatou, Stoll, Blasi & Cristia, 2018;Loukatou et al, 2019), or simply evaluate proposed algorithms (e.g., Daland & Pierrehumbert, 2011;Goldwater et al, 2009;Phillips & Pearl, 2014). These scores are calculated by comparing the output string, which contains hypothesized word breaks an algorithm supplies, against the original sentence containing word breaks.…”
Section: Discussionmentioning
confidence: 99%
“…There has been some computational work comparing learning from ADS and CDS at the level of word learning and phonetic learning. Studies on segmentability use algorithms that learn to identify word units, with some studies reporting higher segmentability for CDS (Batchelder, 2002;Daland and Pierrehumbert, 2011), while Cristia et al (2019) report mixed results. Kirchhoff and Schimmel (2005) train HMM-based speech recognition systems on CDS and ADS, and test on matched and crossed test sets.…”
Section: Related Work 21 Child Directed Speech and Learnabilitymentioning
confidence: 99%
“…Interaction between children and more advanced language interlocutors (such as caregivers) plays an important role in many theories and studies on human language acquisition (e.g., Bruner, 1985;Clark, 2018). For example, although culturally dependent (Shneidman and Goldin-Meadow, 2012) and with the precise effects still up for discussion (Cristia et al, 2019), caregivers can communicate with their children in Child Directed Speech. In turn, children can for example experiment with the meaning of words, to illicit a response from their caregivers (Gillis and Schaerlaekens, 2000).…”
Section: Introductionmentioning
confidence: 99%