Proceedings of the Sixth Workshop on Cognitive Aspects of Computational Language Learning 2015
DOI: 10.18653/v1/w15-2402
|View full text |Cite
|
Sign up to set email alerts
|

Reading metrics for estimating task efficiency with MT output

Abstract: We show that metrics derived from recording gaze while reading, are better proxies for machine translation quality than automated metrics. With reliable eyetracking technologies becoming available for home computers and mobile devices, such metrics are readily available even in the absence of representative held-out human translations. In other words, readingderived MT metrics offer a way of getting cheap, online feedback for MT system adaptation.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
19
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 17 publications
(19 citation statements)
references
References 6 publications
0
19
0
Order By: Relevance
“…Direct assessments of adequacy and MT ranking are the official evaluation procedure for the most recent WMT translation shared task campaigns (Bojar et al, 2016(Bojar et al, , 2017. Other researchers use post-task questionnaires (Stymne et al, 2012;Doherty and O'Brien, 2014;Klerke et al, 2015;Castilho and O'Brien, 2016) to assess the perceived usefulness of MT output. Direct assessment, ranking or post-task questionnaire evaluation methods are clearly subjective and require informants to make "in vitro" judgements about the quality of MT outputs, without considering their usefulness for a specific "in vivo", real-world application.…”
Section: Evaluation Of Mt For Gistingmentioning
confidence: 99%
“…Direct assessments of adequacy and MT ranking are the official evaluation procedure for the most recent WMT translation shared task campaigns (Bojar et al, 2016(Bojar et al, , 2017. Other researchers use post-task questionnaires (Stymne et al, 2012;Doherty and O'Brien, 2014;Klerke et al, 2015;Castilho and O'Brien, 2016) to assess the perceived usefulness of MT output. Direct assessment, ranking or post-task questionnaire evaluation methods are clearly subjective and require informants to make "in vitro" judgements about the quality of MT outputs, without considering their usefulness for a specific "in vivo", real-world application.…”
Section: Evaluation Of Mt For Gistingmentioning
confidence: 99%
“…These two observations recently lead Klerke et al (2015) to suggest using eye-tracking measures as metrics in text simplification. We go beyond this by suggesting that eye-tracking recordings can be used to induce better models for sentence compression for text simplification.…”
Section: Introductionmentioning
confidence: 96%
“…Sentence compression is a basic operation in text simplification which has the potential to improve statistical machine translation and automatic summarization (Berg-Kirkpatrick et al, 2011;Klerke et al, 2015), as well as helping poor readers in need of assistive technologies (Canning et al, 2000). This work suggests using eye-tracking recordings for improving sentence compression for text simplification systems and is motivated by two observations: (i) Sentence compression is the task of automatically making sentences easier to process by shortening them.…”
Section: Introductionmentioning
confidence: 99%
“…This relationship between text and eye movements, has led to an influx of studies investigating the use of eye tracking data to improve and test computational models of language i.e. Barrett et al (2016); Demberg and Keller (2008); Klerke et al (2015). In this study we aim to incorporate eye movement data for the task of automatic readability assessment.…”
Section: Introductionmentioning
confidence: 99%
“…It can also be used to assess the performance of machine translation, text simplification and language generation systems. Eye-tracking data has previously been used to evaluate readability models (Green, 2014;Klerke et al, 2015), however, our main contribution is to explore the way that eye tracking data can help improve models for readability assessment through multi-task learning (Caruana, 1997) and parser metrics based on the surprisal theory of syntactic complexity (Hale, 2001(Hale, , 2016. Multi task learning allows the model to learn various tasks in parallel and improve performance by sharing parameters in the hidden layers.…”
Section: Introductionmentioning
confidence: 99%