2017 IEEE International Conference on Data Mining (ICDM) 2017
DOI: 10.1109/icdm.2017.106
|View full text |Cite
|
Sign up to set email alerts
|

Generating Synthetic Time Series to Augment Sparse Datasets

Abstract: In machine learning, data augmentation is the process of creating synthetic examples in order to augment a dataset used to learn a model. One motivation for data augmentation is to reduce the variance of a classifier, thereby reducing error. In this paper, we propose new data augmentation techniques specifically designed for time series classification, where the space in which they are embedded is induced by Dynamic Time Warping (DTW). The main idea of our approach is to average a set of time series and use th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
79
0
2

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 123 publications
(86 citation statements)
references
References 13 publications
0
79
0
2
Order By: Relevance
“…Finally, we showed how the blackbox effect of deep models which renders them uninterpretable, can be mitigated with a Class Activation Map visualization that highlights which parts of the input time series, contributed the most to a certain class identification. Although we have conducted an extensive experimental evaluation, deep learning for time series classification, unlike for computer vision and NLP tasks, still lacks a thorough study of data augmentation (Ismail Fawaz et al, 2018a;Forestier et al, 2017) and transfer learning (Ismail Fawaz et al, 2018c;Serrà et al, 2018). In addition, the time series community would benefit from an extension of this empirical study that compares in addition to accuracy, the training and testing time of these deep learning models.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Finally, we showed how the blackbox effect of deep models which renders them uninterpretable, can be mitigated with a Class Activation Map visualization that highlights which parts of the input time series, contributed the most to a certain class identification. Although we have conducted an extensive experimental evaluation, deep learning for time series classification, unlike for computer vision and NLP tasks, still lacks a thorough study of data augmentation (Ismail Fawaz et al, 2018a;Forestier et al, 2017) and transfer learning (Ismail Fawaz et al, 2018c;Serrà et al, 2018). In addition, the time series community would benefit from an extension of this empirical study that compares in addition to accuracy, the training and testing time of these deep learning models.…”
Section: Resultsmentioning
confidence: 99%
“…We implemented our framework using the open source deep learning library Keras (Chollet, 2015) with the Tensorflow (Abadi et al, 2015) back-end 1 . Following Lucas et al (2018);Forestier et al (2017); Petitjean et al (2016); Grabocka et al (2014) we used the mean accuracy measure averaged over the 10 runs on the test set. When comparing with the state-of-the-art results published in Bagnall et al (2017) we averaged the accuracy using the median test error.…”
Section: Methodsmentioning
confidence: 99%
“…We used Multi-Dimensional Scaling (MDS) [31], [32] with the objective to gain some insights on the spatial distribution of the perturbed time series compared to the original ones. MDS uses a pairwise distance matrix as input and aims at placing each object in an N-dimensional space such as the betweenobject distances are preserved as well as possible.…”
Section: Multi-dimensional Scalingmentioning
confidence: 99%
“…Other such techniques involve the addition of noise, rotation, and scaling of the values in sequences [4]. A more advanced method, DTW Barycentric Averaging (DBA) [9], generates time-series as weighted averages of multiply aligned time-series. In that method, a time-series is selected and used for aligning the remaining samples.…”
Section: Related Workmentioning
confidence: 99%
“…Therefore, time-series augmentation techniques generate new examples by randomly stretching, shrinking, removing their parts [4], or perturbing them [7]. Also, weighted aligned averages [8,9] or generative models [10] are used.…”
Section: Introductionmentioning
confidence: 99%