2020
DOI: 10.1109/access.2020.3036681
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Modal Feature Integration Network for Human Eye-Fixation Prediction in RGB-D Images

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 56 publications
0
1
0
Order By: Relevance
“…Then, the output of Leaky ReLU layer and the output of the network block before the current block will be added together in MSFF. The addition operation here is inspired by the work on human eye-fixation prediction, where the authors also face a dualmodal information fusion problem and fuse information of different scales by addition operation [41]. Like YOLOV3, the initial fused feature maps will be fed into multiple layers for thoroughly information fusion.…”
Section: ) Multi-scale Feature Fusion Module (Msff)mentioning
confidence: 99%
“…Then, the output of Leaky ReLU layer and the output of the network block before the current block will be added together in MSFF. The addition operation here is inspired by the work on human eye-fixation prediction, where the authors also face a dualmodal information fusion problem and fuse information of different scales by addition operation [41]. Like YOLOV3, the initial fused feature maps will be fed into multiple layers for thoroughly information fusion.…”
Section: ) Multi-scale Feature Fusion Module (Msff)mentioning
confidence: 99%