2020 Twelfth International Conference on Quality of Multimedia Experience (QoMEX) 2020
DOI: 10.1109/qomex48832.2020.9123115
|View full text |Cite
|
Sign up to set email alerts
|

Impact of the Number of Votes on the Reliability and Validity of Subjective Speech Quality Assessment in the Crowdsourcing Approach

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

1
3
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
1
1

Relationship

3
3

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 6 publications
1
3
0
Order By: Relevance
“…The larger number of simulation runs leads to a smother scatter plot and a more accurate fit. In our previous paper [40] we used 200 runs, although similar results were observed there the fitted functions only showed minor changes. Increasing the number of runs also leads to smaller confidence interval widths.…”
Section: Discussion and Future Worksupporting
confidence: 60%
See 1 more Smart Citation
“…The larger number of simulation runs leads to a smother scatter plot and a more accurate fit. In our previous paper [40] we used 200 runs, although similar results were observed there the fitted functions only showed minor changes. Increasing the number of runs also leads to smaller confidence interval widths.…”
Section: Discussion and Future Worksupporting
confidence: 60%
“…Here we use simulations and matrices mentioned in Chapter 5. Some parts of these results have been published in [40]. In this paper, we extend them by considering larger simulation runs, more QoE metrics, and a method for aggregating result of all metrics.…”
Section: Resultsmentioning
confidence: 99%
“…A previous study showed that the ITU-T Rec. P.808 provides a valid and reliable approach for speech quality assessment in crowdsourcing [5]. We provide an open-source implementation 1 of the ITU-T Rec.…”
Section: Introductionmentioning
confidence: 99%
“…It is also recommended to remove submissions, which show specific patterns in the ratings, or which are flagged by outlier detection methods. Although previous works showed that applying the best practices produce highly reliable and valid measurements in multiple studies (with some variations between them) [8], [9], there is no guaranty or method to evaluate that in absence of ground truth.…”
mentioning
confidence: 99%