1998
DOI: 10.1177/154193129804201902
|View full text |Cite
|
Sign up to set email alerts
|

The Evaluator Effect in Usability Studies: Problem Detection and Severity Judgments

Abstract: Usability studies are commonly used in industry and applied in research as a yardstick for other usability evaluation methods. Though usability studies have been studied extensively, one potential threat to their reliability has been left virtually untouched: the evaluator effect. In this study, four evaluators individually analyzed four videotaped usability test sessions. Only 20% of the 93 detected problems were detected by all evaluators, and 46% were detected by only a single evaluator. From the total set … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
49
0

Year Published

2001
2001
2023
2023

Publication Types

Select...
7
2

Relationship

2
7

Authors

Journals

citations
Cited by 48 publications
(51 citation statements)
references
References 17 publications
1
49
0
Order By: Relevance
“…Specifically, the evaluator effect is neither restricted to novice evaluators nor to evaluators knowledgeable of usability in general. The evaluator effect was also found for evaluators with experience in the specific UEM they have been using (Jacobsen et al, 1998;Lewis et al, 1990;Molich et al, 1998Molich et al, , 1999. Furthermore, the evaluator effect is not affected much by restricting the set of problems to only the severe problems.…”
Section: Studies Of the Evaluator Effect In Cw He And Tamentioning
confidence: 58%
See 1 more Smart Citation
“…Specifically, the evaluator effect is neither restricted to novice evaluators nor to evaluators knowledgeable of usability in general. The evaluator effect was also found for evaluators with experience in the specific UEM they have been using (Jacobsen et al, 1998;Lewis et al, 1990;Molich et al, 1998Molich et al, , 1999. Furthermore, the evaluator effect is not affected much by restricting the set of problems to only the severe problems.…”
Section: Studies Of the Evaluator Effect In Cw He And Tamentioning
confidence: 58%
“…Hence, the evaluators were requested to de-tect problems according to the nine criteria, and they were asked to report time-stamped evidence and a free-form description for each problem. Based on the evaluators' problem lists, two of the authors of Jacobsen et al (1998) independently constructed a master list of unique problem tokens. They agreed on 86% of the problem tokens, and by discussing their disagreements and the problems they did not share, a consensus was reached.…”
Section: Tamentioning
confidence: 99%
“…Previous studies (e.g., Hertzum & Jacobsen, 1999;Jacobsen, Hertzum, & John, 1998;Nielsen, 1992) have used the average detection rate of a single evaluator as their basic measure of the evaluator effect. This measure relates the evaluators' individual performances to their collective performance by dividing the average number of problems detected by a single evaluator by the number of problems detected collectively by all the evaluators.…”
Section: Measuring the Evaluator Effectmentioning
confidence: 99%
“…Evaluators may have difficulty in distinguishing between a problem that is rated as a Major Usability Problem and a Usability Catastrophe when using playability heuristics. When comparing evaluator's classification of problems to severity ratings research has shown that inter-rater reliability tends to be low [19]. This may be due to the difficulty evaluators have in distinguishing the boundaries between scales.…”
Section: Introductionmentioning
confidence: 97%