2022
DOI: 10.48550/arxiv.2212.10015
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Benchmarking Spatial Relationships in Text-to-Image Generation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(11 citation statements)
references
References 0 publications
0
11
0
Order By: Relevance
“…In this section, we evaluate Control-GPT on a range of experimental settings to test its controllability regarding to spatial relations, object positions, and sizes based on the Visor dataset [6]. We also extend the evaluation to multiple objects and out-of-distribution prompts.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…In this section, we evaluate Control-GPT on a range of experimental settings to test its controllability regarding to spatial relations, object positions, and sizes based on the Visor dataset [6]. We also extend the evaluation to multiple objects and out-of-distribution prompts.…”
Section: Methodsmentioning
confidence: 99%
“…Human evaluation. We randomly sample 100 queries from the Visor Dataset [6], which includes 25K text prompts specifying the spatial relationships of two objects like "a carrot above a boat", "a bird below a bus". These text prompts are challenging partially because many of them are rare compositions of two unrelated objects, and associating the spatial deixis in text with regions in images is not easy.…”
Section: Querying Gpt-4 For Programmatic Sketches At Inferencementioning
confidence: 99%
See 3 more Smart Citations