2021
DOI: 10.5664/jcsm.9174
|View full text |Cite
|
Sign up to set email alerts
|

Interrater sleep stage scoring reliability between manual scoring from two European sleep centers and automatic scoring performed by the artificial intelligence–based Stanford-STAGES algorithm

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
26
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 30 publications
(31 citation statements)
references
References 37 publications
1
26
0
Order By: Relevance
“…Therefore, it is not possible to evaluate possible differences between local and external database generalization using kappa as reference. Very recently, however, generalization of the same algorithm was evaluated on two additional external datasets, in this case reporting a combined average performance of κ = 0.61, almost in line with the reference human levels in the corresponding cohort (κ = 0.66) [ 23 ], but underperforming with respect to the original values reported in [ 21 ] (κ = 0.72–0.77).…”
Section: Analysis Of Experimental Datamentioning
confidence: 91%
See 2 more Smart Citations
“…Therefore, it is not possible to evaluate possible differences between local and external database generalization using kappa as reference. Very recently, however, generalization of the same algorithm was evaluated on two additional external datasets, in this case reporting a combined average performance of κ = 0.61, almost in line with the reference human levels in the corresponding cohort (κ = 0.66) [ 23 ], but underperforming with respect to the original values reported in [ 21 ] (κ = 0.72–0.77).…”
Section: Analysis Of Experimental Datamentioning
confidence: 91%
“…When considering data on the external dataset validation, Table 6 shows a general global decrease in the performance of the automatic methods as with respect to the corresponding indices on the local database validation scenario. Specifically, in all the works that allow comparison between local and external database generalization using the same algorithm [ 20 , 23 , 24 , 62 , 64 ] decrease in performance in noticeable when tested using external independent datasets. This trend is consistent with the results of our experimentation, as well as with data regarding human inter-rater agreement analyzed in Table 5 .…”
Section: Analysis Of Experimental Datamentioning
confidence: 99%
See 1 more Smart Citation
“…Sleep technicians must verify each epoch manually to perform the sleep scoring, and it has limitations such as labour-intensive and time-consuming and inter-rater variability ( 25 ). Kappa (κ) measures the manual sleep scoring performance to estimate interrater reliability, representing an agreement between epoch-to-epoch.…”
Section: Role Of Ai In Sleep Stage Classificationmentioning
confidence: 99%
“…Traditionally more than one technician is involved in this process to avoid biases in marking sleep stages. The accuracy of sleep scoring depends on the expertise of the technicians ( 25 ). Although PSG use in clinical sleep medicine has significant benefits, the high cost is a barrier to its accessibility to many populations.…”
Section: Introductionmentioning
confidence: 99%