2021
DOI: 10.1109/tcsvt.2020.3032650
|View full text |Cite
|
Sign up to set email alerts
|

Improving Visual Relationship Detection With Two-Stage Correlation Exploitation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(10 citation statements)
references
References 45 publications
0
10
0
Order By: Relevance
“…Table 3 provides a detailed comparison with the existing state-of-the-art relation detection model and uses the same metrics to evaluate the recognition rate of our method. Compared with the best model TCE [14], our method is nearly 1 percentage point higher. PTAT (ours) outperforms the current state-of-the-art method, TCE, e.g., 73.65% vs. 72.20% for R@100, k=1 in predicate detection.…”
Section: ) Experiments On Visual Genomementioning
confidence: 80%
See 3 more Smart Citations
“…Table 3 provides a detailed comparison with the existing state-of-the-art relation detection model and uses the same metrics to evaluate the recognition rate of our method. Compared with the best model TCE [14], our method is nearly 1 percentage point higher. PTAT (ours) outperforms the current state-of-the-art method, TCE, e.g., 73.65% vs. 72.20% for R@100, k=1 in predicate detection.…”
Section: ) Experiments On Visual Genomementioning
confidence: 80%
“…Liu et al [1] proposed two novel modules to discover the common distribution space and the latent relationship association, which map pairs of object features into translation subspaces to induce discriminative relationship clustering. Wang et al [14] proposed a fast method for VRD based on recurrent attention and negative sampling that integrated the attention mechanism into the detection pipeline, enabling the network to focus on several specific parts of an image when scoring predicates for a given object pair. Chiou et al [8] imitated human reasoning mechanisms to propose the RVL-BERT model, which learned visual and language commonsense knowledge via self-supervised pretraining to perform relational reasoning.…”
Section: Related Work a Relationship Detectionmentioning
confidence: 99%
See 2 more Smart Citations
“…Although the existing methods have achieved superior performance in relationship detection works, there are still two key dilemmas in this field, including combination explosion and non-exclusive label problems, as follows. (1) The combination explosion problem: prior works [ 11 ] follow the naive proposing method that if it extracts N objects from an image, there are N(N-1) object-pairs in the object-pair proposal state based on N detected objects. Even worse, multiple correlated relationships usually exist between two objects, and we tend to reserve more visual relationship triplets so that the combinations grow explosively.…”
Section: Introductionmentioning
confidence: 99%