Proceedings of the Fifth Workshop on Beyond Time and Errors: Novel Evaluation Methods for Visualization 2014
DOI: 10.1145/2669557.2669561
|View full text |Cite
|
Sign up to set email alerts
|

Repeated measures design in crowdsourcing-based experiments for visualization

Abstract: Crowdsourcing platforms, such as Amazon's Mechanical Turk (MTurk), are providing visualization researchers with a new avenue for conducting empirical studies. While such platforms offer several advantages over lab-based studies, they also feature some "unknown" or "uncontrolled" variables, which could potentially introduce serious confounding effects in the resultant measurement data. In this paper, we present our experience of using repeated measures in three empirical studies using MTurk. Each study presente… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 27 publications
(31 reference statements)
0
5
0
Order By: Relevance
“…To calculate, we first averaged performance on a participant basis per factor (e.g., Entropy bin) and then bootstrapped on user level performance. This enabled us to control for heterogeneity between participant performance, which is commonly found in crowdsourced visualization experiments [2]. Since Study 2 is a within-subjects study, we report paired Cohen's d for within-subject paired samples [23].…”
Section: Experiments Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…To calculate, we first averaged performance on a participant basis per factor (e.g., Entropy bin) and then bootstrapped on user level performance. This enabled us to control for heterogeneity between participant performance, which is commonly found in crowdsourced visualization experiments [2]. Since Study 2 is a within-subjects study, we report paired Cohen's d for within-subject paired samples [23].…”
Section: Experiments Resultsmentioning
confidence: 99%
“…• We design and develop the Du Bois wrapped bar chart 2 inspired by Du Bois' work for web-based visualizations using D3.js.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…However, very few techniques or tools are built for identical purposes in the same domains, and on similar or the exact same types of data. 54 While most traditional user-centered studies were carried out in collaboration with known ''expert'' users in the domain, recently crowdsourcing has been explored for such evaluations, 55,56 resulting in a greatly increased and diversified set of users. 57 We also chose to make use of a crowdsourcing methodology, utilizing Amazon's Mechanical Turk Service.…”
Section: Evaluation Approachmentioning
confidence: 99%
“…In case of a factorial test design, separate groups of subjects are required for each combination of the different values of the independent variables. Among others, [9] found that inconsistency occurred occasionally to most online participants, and they concluded that a within-subject design is essential for empirical crowdsourcing studies. Nevertheless, monitoring of task execution allows to immediately detect inconsistencies and filter out such participants (e.g., [10]).…”
Section: Test Designmentioning
confidence: 99%