Proceedings of the 27th ACM International Conference on Information and Knowledge Management 2018
DOI: 10.1145/3269206.3271747
|View full text |Cite
|
Sign up to set email alerts
|

Trustworthy Experimentation Under Telemetry Loss

Abstract: Failure to accurately measure the outcomes of an experiment can lead to bias and incorrect conclusions. Online controlled experiments (aka AB tests) are increasingly being used to make decisions to improve websites as well as mobile and desktop applications. We argue that loss of telemetry data (during upload or post-processing) can skew the results of experiments, leading to loss of statistical power and inaccurate or erroneous conclusions. By systematically investigating the causes of telemetry loss, we argu… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(1 citation statement)
references
References 19 publications
(11 reference statements)
0
1
0
Order By: Relevance
“…The challenges with experimentation motivate improved statistical techniques specialized for A/B testing. There are many techniques for fixing specific biases, sources of noise, etc: a specialized test for count data at SweetIM [36]; fixing errors with dependent data at Facebook [106]; improvements from the capabilities of A/A testing on diagnosis (which tests control vs control expecting no effect) at Yahoo [107] and Oath [108]; better calculation of overall effect for features with low coverage at Microsoft [109]; fixing errors from personalization interference at Yahoo [110]; fixing tests under telemetry loss at Microsoft [111]; correcting for selection bias at Airbnb [112]; and algorithms for improved gradual ramp-up at Google [113] and LinkedIn [114].…”
Section: Improved Statistical Methodsmentioning
confidence: 99%
“…The challenges with experimentation motivate improved statistical techniques specialized for A/B testing. There are many techniques for fixing specific biases, sources of noise, etc: a specialized test for count data at SweetIM [36]; fixing errors with dependent data at Facebook [106]; improvements from the capabilities of A/A testing on diagnosis (which tests control vs control expecting no effect) at Yahoo [107] and Oath [108]; better calculation of overall effect for features with low coverage at Microsoft [109]; fixing errors from personalization interference at Yahoo [110]; fixing tests under telemetry loss at Microsoft [111]; correcting for selection bias at Airbnb [112]; and algorithms for improved gradual ramp-up at Google [113] and LinkedIn [114].…”
Section: Improved Statistical Methodsmentioning
confidence: 99%