2022
DOI: 10.1016/j.patcog.2021.108217
|View full text |Cite
|
Sign up to set email alerts
|

Multi-task framework based on feature separation and reconstruction for cross-modal retrieval

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 15 publications
(2 citation statements)
references
References 9 publications
0
2
0
Order By: Relevance
“…For example, text-based retrieval methods may not be able to capture the rich visual content in images or videos. By integrating multiple modalities, cross-modal retrieval helps to represent information more comprehensively and accurately, thereby improving the overall retrieval performance [11]. Figure 1 shows a basic schematic diagram of cross-modal retrieval.…”
Section: Relate Work a Cross-modal Retrievalmentioning
confidence: 99%
“…For example, text-based retrieval methods may not be able to capture the rich visual content in images or videos. By integrating multiple modalities, cross-modal retrieval helps to represent information more comprehensively and accurately, thereby improving the overall retrieval performance [11]. Figure 1 shows a basic schematic diagram of cross-modal retrieval.…”
Section: Relate Work a Cross-modal Retrievalmentioning
confidence: 99%
“…LSH is also known as a negatives loss function, and it learns a fixed margin between the similarities of the relevant image-description embedding pairs and those of the irrelevant embedding pairs. A more recent loss function, the Max of Hinges Loss (LMH) [3], is adopted in most recent VSE networks, due to its ability to outperform LSH [20,21]. An improved version of LSH, LMH only focuses on learning the hard negatives, which are the irrelevant image-description embedding pairs that are nearer to the relevant image-description embedding pairs.…”
Section: Introductionmentioning
confidence: 99%