Proceedings of the 29th ACM International Conference on Multimedia 2021
DOI: 10.1145/3474085.3475366
|View full text |Cite
|
Sign up to set email alerts
|

Deconfounded and Explainable Interactive Vision-Language Retrieval of Complex Scenes

Abstract: In vision-language retrieval systems, users provide natural language feedback to find target images. Vision-language explanations in the systems can better guide users to provide feedback and thus improve the retrieval. However, developing explainable vision-language retrieval systems can be challenging, due to limited labeled multimodal data. In the retrieval of complex scenes, the issue of limited labeled data can be more severe. With multiple objects in the complex scenes, each user query may not exhaustive… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 35 publications
0
2
0
Order By: Relevance
“…The term "Explainable Recommendation" was first defined by Zhang et al [341]. As an important sub-field of AI and machine learning research and due to the fact that recommendation naturally involves humans in the loop, the recommender system community has been leading the research on Explainable AI ever since, which triggers a broader scope of explainability research in other AI and machine learning sub-fields [71,340], such as explainability in scientific research [181], computer vision [297], natural language processing [40,106,172,217,229], graph neural networks [265,299], database [112,291], healthcare systems [121,228,350], online education [9,20,216,264,277], psychological studies [271] and cyber-physical systems [10,12,134,135,241].…”
Section: Overview Of Explainable Recommendationmentioning
confidence: 99%
See 1 more Smart Citation
“…The term "Explainable Recommendation" was first defined by Zhang et al [341]. As an important sub-field of AI and machine learning research and due to the fact that recommendation naturally involves humans in the loop, the recommender system community has been leading the research on Explainable AI ever since, which triggers a broader scope of explainability research in other AI and machine learning sub-fields [71,340], such as explainability in scientific research [181], computer vision [297], natural language processing [40,106,172,217,229], graph neural networks [265,299], database [112,291], healthcare systems [121,228,350], online education [9,20,216,264,277], psychological studies [271] and cyber-physical systems [10,12,134,135,241].…”
Section: Overview Of Explainable Recommendationmentioning
confidence: 99%
“…Omidvar-Tehrani et al [215] conduct Explainable Points-of-Interest (POI) Recommendation as an exploratory process in which users are allowed to keep interaction with the system explanations by expressing their favorite POIs, and the interactions will impact the recommendation process. Wu et al [297] provide vision-language explanations by deconfounded learning that conducts pre-training for the vision-language model. In that way, the potential effects of confounders are removed, which will expedite accurate representation training and better explainability.…”
Section: Overview Of Explainable Recommendationmentioning
confidence: 99%