2022
DOI: 10.5664/jcsm.9538
|View full text |Cite
|
Sign up to set email alerts
|

Interrater reliability of sleep stage scoring: a meta-analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
62
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 74 publications
(66 citation statements)
references
References 41 publications
3
62
1
Order By: Relevance
“…When interpreting the performance of automated sleep staging algorithms it is important to keep in mind that manual scoring by humans is highly subjective [27]. Inter-rater agreement for PSG labeled by human scorers is reported as a κ of 0.76 (95% confidence interval, 0.71-0.81) [28] for 5-class sleep staging. Common mistakes between human scorers during PSG include confusion between wake and light sleep and light sleep and deep sleep [29].…”
Section: Discussionmentioning
confidence: 99%
“…When interpreting the performance of automated sleep staging algorithms it is important to keep in mind that manual scoring by humans is highly subjective [27]. Inter-rater agreement for PSG labeled by human scorers is reported as a κ of 0.76 (95% confidence interval, 0.71-0.81) [28] for 5-class sleep staging. Common mistakes between human scorers during PSG include confusion between wake and light sleep and light sleep and deep sleep [29].…”
Section: Discussionmentioning
confidence: 99%
“…If the agreement between sleep stage scorers is 85%, this is acceptable, and start to be worried if the agreement drops below 70%. As a side note, these numbers are a simplified estimate, because quantitative reliability studies use many different methods, like Cohen’s K , Fleiss K , Pearson product-moment correlation coefficient, and intraclass correlation coefficient (ICC) which reflect different statistical properties of differences [ 3 , 6 ]. Making use of algorithms and software approaches as used for cancer diagnosis (e.g.…”
Section: The Problemmentioning
confidence: 99%
“…Currently, rules for sleep staging are laid out in the AASM manual version 2.6 which provides definitions for sleep stages and the most commonly observed events related to the sleep disorders with the highest prevalence [ 1 ]. We recognize the high variability in sleep scoring results achieved by expert sleep scorers [ 2 , 3 ]. Visual sleep scoring is still, however, a very valid task because we may observe unexpected events during sleep and this teaches us much about the abnormalities observed during sleep.…”
Section: Introductionmentioning
confidence: 99%
“…Interestingly, while this is often cited as an area of concern when using wearable technology, it is rarely discussed as a concern when using statistical analysis software or other proprietary software systems commonly used in research that could influence the interpretation of results. Similarly, inter-rater reliability of PSG scoring remains an issue even in highly trained technicians [ 58 ], which begs the question of the intrinsic accuracy of any interpretation of physiological signal measurements.…”
Section: Trustmentioning
confidence: 99%