2022
DOI: 10.48550/arxiv.2206.11404
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The ArtBench Dataset: Benchmarking Generative Models with Artworks

Abstract: We introduce ArtBench-10, the first class-balanced, high-quality, cleanly annotated, and standardized dataset for benchmarking artwork generation. It comprises 60,000 images of artwork from 10 distinctive artistic styles, with 5,000 training images and 1,000 testing images per style. ArtBench-10 has several advantages over previous artwork datasets. Firstly, it is class-balanced while most previous artwork datasets suffer from the long tail class distributions. Secondly, the images are of high quality with cle… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 27 publications
(57 reference statements)
0
3
0
Order By: Relevance
“…(16)(17)(18). V adv is initialized by a clean image V sampled from the ArtBench dataset (Liao et al, 2022). To ensure human imperceptibility, we consider two different attack types to constrain the optimization of V adv .…”
Section: Basic Setupsmentioning
confidence: 99%
“…(16)(17)(18). V adv is initialized by a clean image V sampled from the ArtBench dataset (Liao et al, 2022). To ensure human imperceptibility, we consider two different attack types to constrain the optimization of V adv .…”
Section: Basic Setupsmentioning
confidence: 99%
“…3 We also excluded data sets that are exclusively applicable to other research areas like aesthetic quality assessment [2], sentiment analysis [54,88], or correspondence matching [35,70]. While formal attributes at the image-level are contained in a large number of data sets, enabling the classification of artists, materials, or creation dates, among others [5,41,46,50,52,55,75,76,85,90], content-based tags are less frequent. This is due to the fact that labels referring to the image phenomena actually shown must be determined by manual annotation, driven either by crowdsourcing approaches [4] or singular institutional efforts [16,29,59].…”
Section: Related Workmentioning
confidence: 99%
“…Recent evaluation benchmarks. Apart from the automatic metrics discussed above, multiple works involve human evaluation and propose their new evaluation benchmarks [14], [16], [63], [73], [80], [86], [87]. We summarize representative benchmarks in Table 2.…”
Section: Technical Evaluation Of Text-to-image Methodsmentioning
confidence: 99%