2020
DOI: 10.48550/arxiv.2001.11921
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Predicting Goal-directed Attention Control Using Inverse-Reinforcement Learning

Abstract: Understanding how goal states control behavior is a question ripe for interrogation by new methods from machine learning. These methods require large and labeled datasets to train models. To annotate a large-scale image dataset with observed search fixations, we collected 16,184 fixations from people searching for either microwaves or clocks in a dataset of 4,366 images (MS-COCO). We then used this behaviorally-annotated dataset and the machine learning method of Inverse-Reinforcement Learning (IRL) to learn t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 8 publications
0
1
0
Order By: Relevance
“…For even greater specificity, we refer to priority maps in the context of a search task as “target maps” ( Zelinsky, 2008 ), a name that makes clear that the prioritization is based on a comparison of a visual input to features of a target goal. We adopted and consistently used this terminology in several recent studies ( Yang et al, 2020 ; Zelinsky et al, 2020 b), most clearly defined in Zelinsky and Bisley (2015 ), and we believe that these distinctions are particularly useful given our present goal of better understanding how different attention biases are weighted in the context of free-viewing and search tasks. In our view, when an assumption of top–down input is made, even in cases of simple text and face detection ( Liao, Shi, Bai, Wang, & Liu, 2017 ; Boyko, Basystiuk, & Shakhovska, 2018 ; Long, Ruan, Zhang, He, Wu, & Yao, 2018 ), a mixture of priority signals occurs that makes it challenging to compare models, with model performance often correlating with how much top–down input can be added to the prediction.…”
Section: Introductionmentioning
confidence: 99%
“…For even greater specificity, we refer to priority maps in the context of a search task as “target maps” ( Zelinsky, 2008 ), a name that makes clear that the prioritization is based on a comparison of a visual input to features of a target goal. We adopted and consistently used this terminology in several recent studies ( Yang et al, 2020 ; Zelinsky et al, 2020 b), most clearly defined in Zelinsky and Bisley (2015 ), and we believe that these distinctions are particularly useful given our present goal of better understanding how different attention biases are weighted in the context of free-viewing and search tasks. In our view, when an assumption of top–down input is made, even in cases of simple text and face detection ( Liao, Shi, Bai, Wang, & Liu, 2017 ; Boyko, Basystiuk, & Shakhovska, 2018 ; Long, Ruan, Zhang, He, Wu, & Yao, 2018 ), a mixture of priority signals occurs that makes it challenging to compare models, with model performance often correlating with how much top–down input can be added to the prediction.…”
Section: Introductionmentioning
confidence: 99%