2002
DOI: 10.1016/s0747-5632(01)00052-8
|View full text |Cite
|
Sign up to set email alerts
|

Stumping e-rater:challenging the validity of automated essay scoring

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
71
0

Year Published

2004
2004
2017
2017

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 93 publications
(78 citation statements)
references
References 9 publications
0
71
0
Order By: Relevance
“…Only one published study addressed the defensive argument directly. In Powers, et al (2001) students and teachers were asked to write "bad faith" essays deliberately in order to fool the e-rater system into assigning a higher (and lower) score than they deserved. Surely more such studies are needed in order to delineate more clearly the limitations of AES systems.…”
Section: J·t·l·amentioning
confidence: 99%
“…Only one published study addressed the defensive argument directly. In Powers, et al (2001) students and teachers were asked to write "bad faith" essays deliberately in order to fool the e-rater system into assigning a higher (and lower) score than they deserved. Surely more such studies are needed in order to delineate more clearly the limitations of AES systems.…”
Section: J·t·l·amentioning
confidence: 99%
“…Wang and Brown, 2007), and validating computerised scoring systems (e.g. Powers et al, 2001), and claims have been made that the reliability of such applications in assessing writing matches that of human raters (e.g. Dikli, 2006).…”
Section: Introductionmentioning
confidence: 99%
“…E-Rater [14] is one of the first automated systems to evaluate text difficulty based on three general classes of essay features: structure (e.g., sentence syntax, proportion of spelling, grammar, usage or mechanics errors), organization based on various discourse features, and content based on prompt-specific vocabulary. Several other tools for automated essay grading or for assessing the textual complexity of a given text have been developed and employed in various educational programs [5,15]: Lexile (MetaMetrics), ATOS (Renaissance Learning), Degrees of Reading Power: DRP Analyzer (Questar Assessment, Inc.), REAP (Carnegie Mellon University), SourceRater (Educational Testing Service), Coh-Metrix (University of Memphis), Markit (Curtin University of Technology) [16], IntelliMetric [17] or Writing Pal (Arizona State University) [18,19].…”
Section: State Of the Artmentioning
confidence: 99%