Proceedings of the 2020 Conference on Human Information Interaction and Retrieval 2020
DOI: 10.1145/3343413.3377960
|View full text |Cite
|
Sign up to set email alerts
|

Relevance Prediction from Eye-movements Using Semi-interpretable Convolutional Neural Networks

Abstract: We propose an image-classification method to predict the perceivedrelevance of text documents from eye-movements. An eye-tracking study was conducted where participants read short news articles, and rated them as relevant or irrelevant for answering a trigger question. We encode participants' eye-movement scanpaths as images, and then train a convolutional neural network classifier using these scanpath images. The trained classifier is used to predict participants' perceived-relevance of news articles from the… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 20 publications
(8 citation statements)
references
References 61 publications
(87 reference statements)
0
8
0
Order By: Relevance
“…A potential solution could be to use higher-level features such as the thorough reading ratio, i.e., the ratio of read and skimmed text lengths (Buscher et al, 2012), or the refixation count, i.e., the number of re-visits to a certain paragraph (Feit et al, 2020). Another solution could be found in using scanpath encodings based deep learning Castner et al (2020); Bhattacharya et al (2020b). We envision the gazebased relevance detection to be a part of future adaptive UIs that leverage multiple sensors for behavioral signal processing and analysis Oviatt et al (2018); Barz et al (2020a,b).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…A potential solution could be to use higher-level features such as the thorough reading ratio, i.e., the ratio of read and skimmed text lengths (Buscher et al, 2012), or the refixation count, i.e., the number of re-visits to a certain paragraph (Feit et al, 2020). Another solution could be found in using scanpath encodings based deep learning Castner et al (2020); Bhattacharya et al (2020b). We envision the gazebased relevance detection to be a part of future adaptive UIs that leverage multiple sensors for behavioral signal processing and analysis Oviatt et al (2018); Barz et al (2020a,b).…”
Section: Discussionmentioning
confidence: 99%
“…Jacob et al (2018) investigated whether eye movements can be used to infer the interest of a reader in a currently read article. Bhattacharya et al (2020b) encoded fixations from participants' scanpaths over documents from the g-REL corpus and trained a convolutional neural network (CNN) with the perceived relevance as prediction target. This approach is limited to small texts of similar lengths.…”
Section: Relevance Estimation From Reading Behaviormentioning
confidence: 99%
“…The VGG-19 architecture was chosen as it is a relatively shallow network with multiple small kernels which we hypothesised would be optimal for capturing any nuanced differences between the input images. Our hypothesis stems from results in an earlier paper which tests similarly a scanpath design on a wide range of out-of-the-box neural network models during a reading task 10 . To further validate our preferred VGG-19 method, we also report the results of two benchmark cases: a Support Vector Machine (SVM) configured for image classification, commonly used in scanpath classification tasks and a logistic regression model which is the most common method of traditional analysis to test the link between gaze data and choice behavior in games presented in normal-form.…”
Section: Model Selection Modelling Tasks Performance Metricsmentioning
confidence: 99%
“…Further, recent advances in Machine Learning (ML) techniques has led a significant increase in the accuracy of prediction when modelling gaze data 1,[6][7][8] . ML techniques have been applied to eye-tracking data to model human cognition in a variety of settings, including -but not limited to -detecting sarcasm 9 , identifying when a participant is in a state of confusion 7 , classifying the relevance of a passage text to a user 10 , and predicting where a participant will focus their attention during location-based games 11 . Further, humans are more frequently interacting with automated systems when engaging in strategic contexts which is a phenomenon that has been noticed by policy makers 12,13 .…”
Section: Introductionmentioning
confidence: 99%
“…This opened opportunities for adaptive user interfaces. A large body of work has focused on enhancing the query-based search of images [Faro et al 2010;Klami 2010;Klami et al 2008] or text-documents [Aula et al 2005;Bhattacharya et al 2020;Buscher et al 2008;Dumais et al 2010]. There, information about eye gaze provides feedback on the relevance of the displayed search results or text documents.…”
Section: Ui Adaptation From Gaze Behaviormentioning
confidence: 99%