2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.01184
|View full text |Cite
|
Sign up to set email alerts
|

Learning Cross-Modal Embeddings With Adversarial Networks for Cooking Recipes and Food Images

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
211
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
3
1

Relationship

2
7

Authors

Journals

citations
Cited by 117 publications
(212 citation statements)
references
References 29 publications
1
211
0
Order By: Relevance
“…Based upon these prior works [4,7,9,29,33,36], this paper extends from cross-modal to cross-domain food retrieval. Leveraging on image-recipe pairs in a source domain, we consider the problem of food transfer as recognizing food in a target domain with new food categories and attributes.…”
Section: Introductionmentioning
confidence: 99%
“…Based upon these prior works [4,7,9,29,33,36], this paper extends from cross-modal to cross-domain food retrieval. Leveraging on image-recipe pairs in a source domain, we consider the problem of food transfer as recognizing food in a target domain with new food categories and attributes.…”
Section: Introductionmentioning
confidence: 99%
“…Jin et al analyzed images between true news and fake news in terms of, e.g., their clarity [5]. Along with the recent advances in deep learning, various RNNs and CNNs have been developed for multi-modal fake news detection and related tasks [4,7,18,21,23,24]. To learn the multi-modal (textual and visual) representation of news content, Jin et al developed VGG-19 and LSTM with an attention mechanism [4], and Khattar et al designed an encoder-decoder mechanism [7].…”
Section: Content-based Fake News Detectionmentioning
confidence: 99%
“…[Micael et al 2018] extended [Salvador et al 2017] by providing a double-triplet strategy to jointly express both the retrieval loss and the classification one for cross-modal retrieval. [Wang et al 2019;Zhu et al 2019] further introduced adversarial networks to impose the modality alignment for cross-modal retrieval. [Salvador et al 2019] proposed a new architecture for ingredient prediction that exploits co-dependencies among ingredients without imposing order for generating cooking instructions from an image and its ingredients.…”
Section: Referencementioning
confidence: 99%