Proceedings of the 24th Conference on Computational Natural Language Learning 2020
DOI: 10.18653/v1/2020.conll-1.11
|View full text |Cite
|
Sign up to set email alerts
|

Bridging Information-Seeking Human Gaze and Machine Reading Comprehension

Abstract: In this work, we analyze how human gaze during reading comprehension is conditioned on the given reading comprehension question, and whether this signal can be beneficial for machine reading comprehension. To this end, we collect a new eye-tracking dataset with a large number of participants engaging in a multiple choice reading comprehension task. Our analysis of this data reveals increased fixation times over parts of the text that are most relevant for answering the question. Motivated by this finding, we p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
19
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 24 publications
(20 citation statements)
references
References 27 publications
(24 reference statements)
1
19
0
Order By: Relevance
“…In terms of limitations, we did not investigate the breadth or depth of influence of our method of [CLS]-based aggregate attention supervision on the model attentions across layers and heads, nor the supervision of specific layers or heads as done by Strubell et al (2018). We did not explore trade-off coefficients on the multiple losses, such as the convex combination used by Malmaud et al (2020). We used a relatively small English dataset, which limited generalizability and robustness.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…In terms of limitations, we did not investigate the breadth or depth of influence of our method of [CLS]-based aggregate attention supervision on the model attentions across layers and heads, nor the supervision of specific layers or heads as done by Strubell et al (2018). We did not explore trade-off coefficients on the multiple losses, such as the convex combination used by Malmaud et al (2020). We used a relatively small English dataset, which limited generalizability and robustness.…”
Section: Discussionmentioning
confidence: 99%
“…Because BERT uses subword tokenization, to allow matching entries to be found in the ZuCo wordlevel data we split the ZuCo words into BERT tokens, evenly dividing values between each subword piece (e.g., "delicacy" → "del", "##ica", "##cy", each piece allotted a third of the ZuCo value), a technique used by Malmaud et al (2020). We preserve entity markers "<e>" and "</e>" in each sample by adding them as special tokens to the BERT tokenizer so their embeddings are learned with other tokens during fine-tuning.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Obtaining human scores for sequence tokens requires preprocessing. As ZuCo uses a word-level lexicon but BERT uses subword tokenization, in order to find matching entries we first split the ZuCo words into BERT tokens, evenly dividing values between each subword piece (e.g., when tokenizing "delicacy" → ["del", "##ica", "##cy"] in sentence j, each piece is allotted a third of the ZuCo value), a technique previously used by [26]. We pass the human ET and EEG token values z ET and z EEG through a softmax layer to obtain two distributions over sentences, vectors α ′′ ET and α ′ EEG .…”
Section: Methodsmentioning
confidence: 99%
“…Cognitive NLP tasks that have been studied in recent years include sentiment analysis [23], part-of-speech (POS) tagging [24], and named entity recognition (NER) [25]. Typically for neural network-based approaches, a recurrent architecture such as a bidirectional Long Short-Term Memory (biLSTM) network has been used; more recently a variant of BERT was used for question answering with ET prediction as the auxiliary task [26]. MTL has been used for sentiment analysis and NER with learning gaze behavior as the auxiliary task [27,28], and a combination of gaze and brain data has been applied to a suite of NLP tasks [11], including sentiment analysis, using approaches such as predicting cognitive data or using those data to augment input embeddings.…”
Section: Related Workmentioning
confidence: 99%