Proceedings of the Ninth Workshop on Innovative Use of NLP for Building Educational Applications 2014
DOI: 10.3115/v1/w14-1802
|View full text |Cite
|
Sign up to set email alerts
|

Automatic Assessment of the Speech of Young English Learners

Abstract: This paper introduces some of the research behind automatic scoring of the speaking part of the Arizona English Language Learner Assessment, a large-scale test now operational for students in Arizona. Approximately 70% of the students tested are in the range 4-11 years old. We cover the methods used to assess spoken responses automatically, considering both what the student says and the way in which the student speaks. We also provide evidence for the validity of machine scores. The assessments include 10 open… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
16
0

Year Published

2015
2015
2019
2019

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 22 publications
(22 citation statements)
references
References 21 publications
1
16
0
Order By: Relevance
“…The task types involved in this study are listed in Table 1. More details can be found in (Bernstein et al, 2010b;Cheng et al, 2014Cheng and Shen, 2010;Xu et al, 2012). Item types SentReading, SentRepeats, SentBuild, and PassReading are constrained: there is a set of pre-defined correct words and word sequences that test takers are expected to include in their response, although language models accept miscues and popular mistakes.…”
Section: Datasetsmentioning
confidence: 94%
See 1 more Smart Citation
“…The task types involved in this study are listed in Table 1. More details can be found in (Bernstein et al, 2010b;Cheng et al, 2014Cheng and Shen, 2010;Xu et al, 2012). Item types SentReading, SentRepeats, SentBuild, and PassReading are constrained: there is a set of pre-defined correct words and word sequences that test takers are expected to include in their response, although language models accept miscues and popular mistakes.…”
Section: Datasetsmentioning
confidence: 94%
“…In the last decade, the tasks used in the automatic spoken assessment research have extended from constrained ones (such as reading or repeating a sentence, or reading a passage) to open-ended ones (such as open questions or retellings of a story, passage, picture or presentation) (Bernstein et al, 2010b;Cheng et al, 2014;Evanini and Wang, 2013;Nair et al, 2005;Pearson, 2009;Zechner et al, 2009a). Providing a reasonable recognition performance in these open-ended tasks is a big challenge.…”
Section: Related Workmentioning
confidence: 98%
“…Cheng, D'Antilio, Chen, and Bernstein () presented an automated speech scoring system for an Arizona K–12 language test for English‐language learners that contained a range of items, from predictable (read or repeated prompt) to more open ended. There were 11 different item types, of which 9 were open ended.…”
Section: Overview Of Item Types Usedmentioning
confidence: 99%
“…Xie et al (2012) explored content measures based on the lexical similarity between the response and a set of reference responses. A content-scoring component based on word vectors was also part of the automated scoring engine described by Cheng et al (2014). In both these studies, content features were developed to supplement other features measuring various aspects of speaking proficiency.…”
Section: Related Workmentioning
confidence: 99%