2017
DOI: 10.48550/arxiv.1702.05448
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning to Detect Human-Object Interactions

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
28
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(28 citation statements)
references
References 28 publications
0
28
0
Order By: Relevance
“…Visual Relationship Detection. Relationship detection among image constituents uses separate branches in a Con-vNet to model objects, humans, and their interactions [5,21]. A distinct approach in Santoro et al [60] treats each of the cells across channels in convolutional feature maps as an object and the relationships are modeled by a pairwise concatenation of the feature representations of individual cells.…”
Section: Related Workmentioning
confidence: 99%
“…Visual Relationship Detection. Relationship detection among image constituents uses separate branches in a Con-vNet to model objects, humans, and their interactions [5,21]. A distinct approach in Santoro et al [60] treats each of the cells across channels in convolutional feature maps as an object and the relationships are modeled by a pairwise concatenation of the feature representations of individual cells.…”
Section: Related Workmentioning
confidence: 99%
“…More recently, with the release of large datasets like HICO (Chao et al 2015), Visual Genome (Krishna et al 2017), HCVRD (Zhuang et al 2017b), V-COCO (Gupta and Malik 2015), and HICO-Det (Chao et al 2017), the problem of detecting and recognizing HOIs has attracted signification attention. This has been driven by HICO which is a benchmark dataset for recognizing human-object interactions.…”
Section: Related Workmentioning
confidence: 99%
“…These atomic recognition tasks are certainly the building blocks of a variety of approaches for HOI understanding Delaitre, Sivic, and Laptev 2011); and the progress in these atomic tasks directly translates to improvements in HOI understanding. However, the task of HOI understanding comes with its own unique set of challenges (Lu et al 2016;Chao et al 2017).…”
Section: Introductionmentioning
confidence: 99%
“…The task of recognizing objects and the relationships has been investigated by numerous studies in a various form. This includes detection of human-object interactions [7,3], localization of proposals from natural language expressions [12], or the more general tasks of visual relationship detection [17,25,38,5,19,37,34,41] and scene graph generation [33,18,35,22].…”
Section: Relationship Detectionmentioning
confidence: 99%
“…The results show that our relational embedding represents inter-dependency among all object instances, being consistent with the ground-truth relationships. To illustrate, in the first example, the ground-truth matrix refers to the relationships between the 'man'(1) and his body parts (2,3); and the 'mountain'(0) and the 'rocks' (4,5,6,7), which are also reasonably captured in our relational embedding matrix. Note that our model infers relationship correctly even there exists missing ground-truths such as cell(7,0) due to sparsity of annotations in Visual Genome dataset.…”
Section: Qualitative Evaluationmentioning
confidence: 99%