2020
DOI: 10.1007/978-3-030-58604-1_18
|View full text |Cite
|
Sign up to set email alerts
|

Visual-Relation Conscious Image Generation from Structured-Text

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(9 citation statements)
references
References 17 publications
0
9
0
Order By: Relevance
“…In [122], a scene graph is used to predict initial bounding boxes for objects. Using the initial bounding boxes, relation units consisting of two bounding boxes are predicted for each individual subject-predicate-object relation.…”
Section: Scene Graphsmentioning
confidence: 99%
“…In [122], a scene graph is used to predict initial bounding boxes for objects. Using the initial bounding boxes, relation units consisting of two bounding boxes are predicted for each individual subject-predicate-object relation.…”
Section: Scene Graphsmentioning
confidence: 99%
“…This consideration is crucial for better image generation to support the story visualization task. Indeed, recent works [10,14,17,18,34] employ different techniques such as prediction networks to estimate a scene layout that gives initial and refined layout [34] or gives predictive values for the object appearances [10,14,17,18]. However, in [14], the authors used object embeddings only to represent objects in their layout without using other details from the scene graph.…”
Section: Object Layout Modulementioning
confidence: 99%
“…Johnson et al [10] first proposed to generate images from scene graphs, they implemented the sg2im method to reason related objects and relationships. Then Vo et al [36] adopted the scene structure in the conditional GAN network and put forward the stacking-GANs to infer visualrelation layouts. With the same input form, Li et al [19] proposed the PasteGAN to crop objects from the external memory tank and paste them into correct locations of the final images.…”
Section: 3mentioning
confidence: 99%
“…The graph convolutional network (GCN) can directly operate on graphs. Following [10,36,19], we also take the scene graph as input and calculate new embedding vectors for each node and edge. Additionally, we apply the same function on each graph convolutional layer, which ensures a single layer can work with arbitrarily shaped graphs.…”
Section: 3mentioning
confidence: 99%
See 1 more Smart Citation