2022
DOI: 10.3390/clockssleep4040039
|View full text |Cite
|
Sign up to set email alerts
|

The Impact of Missing Data and Imputation Methods on the Analysis of 24-Hour Activity Patterns

Abstract: The purpose of this study is to characterize the impact of the timing and duration of missing actigraphy data on interdaily stability (IS) and intradaily variability (IV) calculation. The performance of three missing data imputation methods (linear interpolation, mean time of day (ToD), and median ToD imputation) for estimating IV and IS was also tested. Week-long actigraphy records with no non-wear or missing timeseries data were masked with zeros or ‘Not a Number’ (NaN) across a range of timings and duration… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 13 publications
(9 citation statements)
references
References 24 publications
0
7
0
Order By: Relevance
“…In this study, there were some missing values in the raw data we used, and most of the missing values were filled in by manually tracing the raw materials. For a small amount of other missing values such as age and other quantitative data, we use mean interpolation to fill in, as the mean can represent the central trend of the data and help maintain its distribution.For qualitative data such as crime types, we use the median to fill in, which is a better choice because it can reduce the impact of extreme values while maintaining the order and level of the data 38 .…”
Section: Methodsmentioning
confidence: 99%
“…In this study, there were some missing values in the raw data we used, and most of the missing values were filled in by manually tracing the raw materials. For a small amount of other missing values such as age and other quantitative data, we use mean interpolation to fill in, as the mean can represent the central trend of the data and help maintain its distribution.For qualitative data such as crime types, we use the median to fill in, which is a better choice because it can reduce the impact of extreme values while maintaining the order and level of the data 38 .…”
Section: Methodsmentioning
confidence: 99%
“…High-frequency (100 Hz) accelerometer data were processed on Sherlock, a high-performance computing cluster provided by Stanford University, using the steps outlined in Weed et al 2022 [ 18 ]. In brief, data spanning 1 week of collection were down-sampled to 30 s epochs using the biobankAccelerometerAnalysis package in Python v3.6.1 [ 17 ].…”
Section: Methodsmentioning
confidence: 99%
“…High frequency (100 Hz) accelerometer data were processed on Sherlock, a high-performance computing cluster provided by Stanford University,using the steps outlined in Weed et al 2022 18 . In brief, data spanning 1 week of collection were down-sampled to 30 second epochs using the biobankAccelerometerAnalysis package in Python v3.6.1 17 .…”
Section: Methodsmentioning
confidence: 99%
“…Non-wear time was defined as stationary episodes lasting for at least 60 minutes in which all three axes had a standard deviation of less than 13.0 mg. If present, non-wear segments were automatically imputed using the median of similar time-of-day vector magnitude and intensity distribution data points with 30-second granularity on different days of the measurement 18 . Following these preprocessing steps, we derived the following six metrics.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation