2022
DOI: 10.1007/s11432-020-3181-9
|View full text |Cite
|
Sign up to set email alerts
|

Onfocus detection: identifying individual-camera eye contact from unconstrained images

Abstract: Onfocus detection aims at identifying whether the focus of the individual captured by a camera is on the camera or not. Based on the behavioral research, the focus of an individual during face-to-camera communication leads to a special type of eye contact, i.e., the individual-camera eye contact, which is a powerful signal in social communication and plays a crucial role in recognizing irregular individual status (e.g., lying or suffering mental disease) and special purposes (e.g., seeking help or attracting f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 15 publications
(5 citation statements)
references
References 38 publications
(62 reference statements)
0
4
0
Order By: Relevance
“…3) Routing layers: To study the performance of routing layers in one block, we compare our ResCaps and one modified versions, i.e., ResCaps-3L. Specifically, ResCaps and ResCaps-3L consist of two and three ResP layer(s) 4 in one block, respectively. From Table V, we find that our ResCaps consisting of two ResP layers can achieve promising performance, compared to ResCaps-3L that employs three ResP layers in one block.…”
Section: B Ablation Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…3) Routing layers: To study the performance of routing layers in one block, we compare our ResCaps and one modified versions, i.e., ResCaps-3L. Specifically, ResCaps and ResCaps-3L consist of two and three ResP layer(s) 4 in one block, respectively. From Table V, we find that our ResCaps consisting of two ResP layers can achieve promising performance, compared to ResCaps-3L that employs three ResP layers in one block.…”
Section: B Ablation Analysismentioning
confidence: 99%
“…They can recognize the image by detecting the existence of a specific entity, i.e., invariance. However, an unsophisticated perturbation on the image can fool a well-trained network to fail in recognition [1]- [4]. More worryingly, natural and non-adversarial pose changes of familiar objects in the real world are enough to trick deep networks [5], [6].…”
Section: Introductionmentioning
confidence: 99%
“…detection becomes increasingly important in understanding the intention of surrounding pedestrians in the autonomous driving environment [5], [6], [7]. Many studies on eyes contact detection use vision images taken close to a person with a clear facial appearance [8], [9], [10], [11], [12]. However, pedestrian eye detection in the wild uses distant images or videos from vehicle sensors which brings great challenges to the problem given the image quality, unconstrained surroundings and illuminations [6], [7].…”
Section: Introductionmentioning
confidence: 99%
“…However, they usually assume that the training and testing data is balanced. In practice, training or testing data appears to be long-tailed, e.g., there exist few samples for rare diseases in medical diagnosis [19,35,36,39,42,45] or endangered animals in species classification [5,31,37]. As mentioned by [32], the case becomes even worse in weakly and semi-supervised learning scenarios [10,20,27,31,33,34,38,43,44].…”
Section: Introductionmentioning
confidence: 99%