scispace - formally typeset
Search or ask a question

Showing papers on "Dynamic time warping published in 2018"


Journal ArticleDOI
TL;DR: Object-based time-weighted dynamic time warping (TWDTW) method achieved comparable classification results to RF in Romania and Italy, but RF achieved better results in the USA, where the classified crops present high intra-class spectral variability.

556 citations


Journal ArticleDOI
TL;DR: This paper uses dynamic time warping (DTW) algorithm to compare the various shapes of foot movements collected from the wearable IoT devices to evaluate the effectiveness of the DTW method for Alzheimer disease diagnosis.
Abstract: Alzheimer disease is a significant problem in public health. Alzheimer disease causes severe problems with thinking, memory and activities. Alzheimer disease affected more on the people who are in the age group of 80-year-90. The foot movement monitoring system is used to detect the early stage of Alzheimer disease. internets of things (IoT) devices are used in this paper to monitor the patients’ foot movement in continuous manner. This paper uses dynamic time warping (DTW) algorithm to compare the various shapes of foot movements collected from the wearable IoT devices. The foot movements of the normal individuals and people who are affected by Alzheimer disease are compared with the help of middle level cross identification (MidCross) function. The identified cross levels are used to classify the gait signal for Alzheimer disease diagnosis. Sensitivity and specificity are calculated to evaluate the DTW algorithm based classification model for Alzheimer disease. The classification results generated using the DTW is compared with the various classification algorithms such as inertial navigation algorithm, K-nearest neighbor classifier and support vector machines. The experimental results proved the effectiveness of the DTW method.

237 citations


Proceedings ArticleDOI
05 Nov 2018
TL;DR: In an effort to predict the best source dataset for a given target dataset, a new method relying on Dynamic Time Warping to measure inter-datasets similarities is proposed, leading to an improvement in accuracy on 71 out of 85 datasets.
Abstract: Transfer learning for deep neural networks is the process of first training a base network on a source dataset, and then transferring the learned features (the network’s weights) to a second network to be trained on a target dataset. This idea has been shown to improve deep neural network’s generalization capabilities in many computer vision tasks such as image recognition and object localization. Apart from these applications, deep Convolutional Neural Networks (CNNs) have also recently gained popularity in the Time Series Classification (TSC) community. However, unlike for image recognition problems, transfer learning techniques have not yet been investigated thoroughly for the TSC task. This is surprising as the accuracy of deep learning models for TSC could potentially be improved if the model is fine-tuned from a pre-trained neural network instead of training it from scratch. In this paper, we fill this gap by investigating how to transfer deep CNNs for the TSC task. To evaluate the potential of transfer learning, we performed extensive experiments using the UCR archive which is the largest publicly available TSC benchmark containing 85 datasets. For each dataset in the archive, we pre-trained a model and then fine-tuned it on the other datasets resulting in 7140 different deep neural networks. These experiments revealed that transfer learning can improve or degrade the models predictions depending on the dataset used for transfer. Therefore, in an effort to predict the best source dataset for a given target dataset, we propose a new method relying on Dynamic Time Warping to measure inter-datasets similarities. We describe how our method can guide the transfer to choose the best source dataset leading to an improvement in accuracy on 71 out of 85 datasets.

197 citations


Journal ArticleDOI
TL;DR: Several different machine learning methodologies are compared starting from well-established statistical feature-based methods to convolutional neural networks, and a novel application of dynamic time warping to bearing fault classification is proposed as a robust, parameter free method for race fault detection.

145 citations


Journal ArticleDOI
TL;DR: In this article, a shape-based approach is proposed to better classify and predict consumer energy consumption behavior at the household level, which is based on the dynamic time warping (DTW).
Abstract: Household consumer demand response (DR) is an important research and industry problem, which seeks to categorize, predict, and modify consumer’s energy consumption. Unfortunately, traditional clustering methods have resulted in many hundreds of clusters, with a given consumer often associated with several clusters, making it difficult to classify consumers into stable representative groups and to predict individual energy consumption patterns. In this paper, we present a shape-based approach that better classifies and predicts consumer energy consumption behavior at the household level. The method is based on the dynamic time warping (DTW). DTW seeks an optimal alignment between energy consumption patterns reflecting the effect of hidden patterns of regular consumer behavior. Using real consumer 24-hour load curves from Opower Corporation, our method results in a 50% reduction in the number of representative groups and an improvement in prediction accuracy measured under DTW distance. We extend the approach to estimate which electrical devices will be used and in which hours.

127 citations


Journal ArticleDOI
TL;DR: Japingz et al. as mentioned in this paper proposed an improved alignment algorithm, named shape Dynamic Time Warping (shapeDTW), which enhances DTW by taking point-wise local structural information into consideration.

112 citations


Book ChapterDOI
01 Jan 2018
TL;DR: A different type of data-driven approaches allows to perform classification directly on the raw time-series data, avoiding the features’ extraction phase, among these approaches, dynamic time warping and symbolic-based methodologies have been widely applied in many application areas.
Abstract: Chapter Overview The diffusion in power systems of distributed renewable energy resources, electric vehicles, and controllable loads has made advanced monitoring systems fundamental to cope with the consequent disturbances in power flows; advanced monitoring systems can be employed for anomaly detection, root cause analysis, and control purposes. Several machine learning-based approaches have been developed in the past recent years to detect if a power system is running under anomalous conditions and, eventually, to classify such situation with respect to known problems. One of the aspects, which makes Power Systems challenging to be tackled, is that the monitoring has to be performed on streams of data that have a time-series evolution; this issue is generally tackled by performing a features’ extraction procedure before the classification phase. The features’ extraction phase consists of translating the informative content of time-series data into scalar quantities: such procedure may be a time-consuming step that requires the involvement of process experts to avoid loss of information in the making; moreover, extracted features designed to capture certain behaviors of the system, may not be informative under unseen conditions leading to poor monitoring performances. A different type of data-driven approaches, which will be reviewed in this chapter, allows to perform classification directly on the raw time-series data, avoiding the features’ extraction phase: among these approaches, dynamic time warping and symbolic-based methodologies have been widely applied in many application areas. In the following, pros and cons of each approach will be discussed and practical implementation guidelines will be provided.

88 citations


Journal ArticleDOI
TL;DR: An opportunity to use operators' prevalent mobile devices to measure and report their equipment's cycle times in a cost-effective and continuous manner is demonstrated.

77 citations


Journal ArticleDOI
TL;DR: The longer the time needed for the search, the higher the speedup ratio achieved by the method, and it is demonstrated that the method performs similarly to UCR suite for small queries and narrow warping constraints, but performs up to five times faster for long queries and large warping windows.
Abstract: Similarity search is the core procedure for several time series mining tasks. While different distance measures can be used for this purpose, there is clear evidence that the Dynamic Time Warping (DTW) is the most suitable distance function for a wide range of application domains. Despite its quadratic complexity, research efforts have proposed a significant number of pruning methods to speed up the similarity search under DTW. However, the search may still take a considerable amount of time depending on the parameters of the search, such as the length of the query and the warping window width. The main reason is that the current techniques for speeding up the similarity search focus on avoiding the costly distance calculation between as many pairs of time series as possible. Nevertheless, the few pairs of subsequences that were not discarded by the pruning techniques can represent a significant part of the entire search time. In this work, we adapt a recently proposed algorithm to improve the internal efficiency of the DTW calculation. Our method can speed up the UCR suite, considered the current fastest tool for similarity search under DTW. More important, the longer the time needed for the search, the higher the speedup ratio achieved by our method. We demonstrate that our method performs similarly to UCR suite for small queries and narrow warping constraints. However, it performs up to five times faster for long queries and large warping windows.

72 citations


Journal ArticleDOI
TL;DR: A novel measurement called Time Alignment Measurement is proposed, which delivers similarity information on the temporal domain and demonstrates the potential of the approach in measuring performance of time series alignment methodologies and in the characterization of synthetic and real time series data acquired during human movement.

69 citations


Journal ArticleDOI
TL;DR: The importance of setting DTW’s warping window width correctly is demonstrated, and novel methods to learn this parameter in both supervised and unsupervised settings are proposed, which can produce significant improvements in classification accuracy and clustering quality.
Abstract: Dynamic Time Warping (DTW) is a highly competitive distance measure for most time series data mining problems. Obtaining the best performance from DTW requires setting its only parameter, the maximum amount of warping (w). In the supervised case with ample data, w is typically set by cross-validation in the training stage. However, this method is likely to yield suboptimal results for small training sets. For the unsupervised case, learning via cross-validation is not possible because we do not have access to labeled data. Many practitioners have thus resorted to assuming that “the larger the better”, and they use the largest value of w permitted by the computational resources. However, as we will show, in most circumstances, this is a naive approach that produces inferior clusterings. Moreover, the best warping window width is generally non-transferable between the two tasks, i.e., for a single dataset, practitioners cannot simply apply the best w learned for classification on clustering or vice versa. In addition, we will demonstrate that the appropriate amount of warping not only depends on the data structure, but also on the dataset size. Thus, even if a practitioner knows the best setting for a given dataset, they will likely be at a lost if they apply that setting on a bigger size version of that data. All these issues seem largely unknown or at least unappreciated in the community. In this work, we demonstrate the importance of setting DTW’s warping window width correctly, and we also propose novel methods to learn this parameter in both supervised and unsupervised settings. The algorithms we propose to learn w can produce significant improvements in classification accuracy and clustering quality. We demonstrate the correctness of our novel observations and the utility of our ideas by testing them with more than one hundred publicly available datasets. Our forceful results allow us to make a perhaps unexpected claim; an underappreciated “low hanging fruit” in optimizing DTW’s performance can produce improvements that make it an even stronger baseline, closing most or all the improvement gap of the more sophisticated methods proposed in recent years.

Journal ArticleDOI
TL;DR: An innovative averaging of a set of time-series based on the Dynamic Time Warping based on an existing well-established method called DBA (for DTW Barycenter Averaging) is proposed and an innovative tolerance is added to take into account the admissible variability around the average signal.

Journal ArticleDOI
TL;DR: A novel time-analysis framework is presented for large-scale comorbidity studies and the proposed methodology for the temporal assessment of common disease trajectories could serve as the preliminary basis of a disease prediction system.
Abstract: Time is a crucial parameter in the assessment of comorbidities in population-based studies, as it permits to identify more complex disease patterns apart from the pairwise disease associations. So far, it has been, either, completely ignored or only, taken into account by assessing the temporal directionality of identified comorbidity pairs. In this work, a novel time-analysis framework is presented for large-scale comorbidity studies. The disease-history vectors of patients of a regional Spanish health dataset are represented as time sequences of ordered disease diagnoses. Statistically significant pairwise disease associations are identified and their temporal directionality is assessed. Subsequently, an unsupervised clustering algorithm, based on Dynamic Time Warping, is applied on the common disease trajectories in order to group them according to the temporal patterns that they share. The proposed methodology for the temporal assessment of such trajectories could serve as the preliminary basis of a disease prediction system.

Posted Content
TL;DR: This paper proposes a data augmentation method for time series with irregular sampling, Time-Conditional Generative Adversarial Network (T-CGAN), where the generative step is implemented by a deconvolutional NN and the discriminative step by a convolutionalNN.
Abstract: In this paper we propose a data augmentation method for time series with irregular sampling, Time-Conditional Generative Adversarial Network (T-CGAN). Our approach is based on Conditional Generative Adversarial Networks (CGAN), where the generative step is implemented by a deconvolutional NN and the discriminative step by a convolutional NN. Both the generator and the discriminator are conditioned on the sampling timestamps, to learn the hidden relationship between data and timestamps, and consequently to generate new time series. We evaluate our model with synthetic and real-world datasets. For the synthetic data, we compare the performance of a classifier trained with T-CGAN-generated data, against the performance of the same classifier trained on the original data. Results show that classifiers trained on T-CGAN-generated data perform the same as classifiers trained on real data, even with very short time series and small training sets. For the real world datasets, we compare our method with other techniques of data augmentation for time series, such as time slicing and time warping, over a classification problem with unbalanced datasets. Results show that our method always outperforms the other approaches, both in case of regularly sampled and irregularly sampled time series. We achieve particularly good performance in case with a small training set and short, noisy, irregularly-sampled time series.

Journal ArticleDOI
TL;DR: This paper explores the utility of information derived from the dynamic time warping (DTW) cost matrix for the problem of online signature verification and devise a score that utilizes the information from the cost matrix and warping path alignments for authenticating the veracity of a test signature.
Abstract: This paper explores the utility of information derived from the dynamic time warping (DTW) cost matrix for the problem of online signature verification. The prior works in literature primarily utilize only the DTW scores to authenticate a test signature. To the best of our knowledge, the characteristics of the warping path (used for the alignment) in the cost matrix is hardly exploited for verification of online signatures. Accordingly, we devise a score that utilizes the information from the cost matrix and warping path alignments. We subsequently consider its fusion (using a sum rule combiner) with the DTW score for authenticating the veracity of a test signature. In addition, a minor modification is suggested with regards to the set of features employed for matching the signatures. We introduce a spacing parameter for feature extraction and demonstrate its applicability in increasing the separation between the distribution of genuine and forgery signatures for an user. Our method has been tested on two publicly available online signature databases namely the SVC-2004 Task 2 and MCYT-100. We report reduction in error rates over the traditional DTW framework.

Journal ArticleDOI
TL;DR: Experimental results show that the proposed Structured Dynamic Time Warping (SDTW) approach is robust to the diversity of same handwritten letter, and significantly outperforms state-of-the-art approaches.

Journal ArticleDOI
TL;DR: Deep Canonical Time Warping is presented, a method that automatically learns non-linear representations of multiple time-series that are maximally correlated in a shared subspace, and temporally aligned, and significantly outperform state-of-the-art methods in temporal alignment.
Abstract: Machine learning algorithms for the analysis of time-series often depend on the assumption that utilised data are temporally aligned. Any temporal discrepancies arising in the data is certain to lead to ill-generalisable models, which in turn fail to correctly capture properties of the task at hand. The temporal alignment of time-series is thus a crucial challenge manifesting in a multitude of applications. Nevertheless, the vast majority of algorithms oriented towards temporal alignment are either applied directly on the observation space or simply utilise linear projections-thus failing to capture complex, hierarchical non-linear representations that may prove beneficial, especially when dealing with multi-modal data (e.g., visual and acoustic information). To this end, we present Deep Canonical Time Warping (DCTW), a method that automatically learns non-linear representations of multiple time-series that are (i) maximally correlated in a shared subspace, and (ii) temporally aligned. Furthermore, we extend DCTW to a supervised setting, where during training, available labels can be utilised towards enhancing the alignment process. By means of experiments on four datasets, we show that the representations learnt significantly outperform state-of-the-art methods in temporal alignment, elegantly handling scenarios with heterogeneous feature sets, such as the temporal alignment of acoustic and visual information.

Journal ArticleDOI
TL;DR: The task of outlier detection is reformulate as a weighted clustering problem based on entropy and dynamic time warping for time series and the outliers are detected by an optimization problem of a new proposed cost function adapted to this kind of data.
Abstract: In the last decade, outlier detection for temporal data has received much attention from data mining and machine learning communities While other works have addressed this problem by two-way approaches (similarity and clustering), we propose in this paper an embedded technique dealing with both methods simultaneously We reformulate the task of outlier detection as a weighted clustering problem based on entropy and dynamic time warping for time series The outliers are then detected by an optimization problem of a new proposed cost function adapted to this kind of data Finally, we provide some experimental results for validating our proposal and comparing it with other methods of detection

Journal ArticleDOI
TL;DR: Four clustering models for multivariate time series are proposed, with the following characteristics: the Partitioning Around Medoids (PAM) framework is considered, and a robust metric approach is used, i.e., the exponential transformation of dissimilarity measures.

Journal ArticleDOI
TL;DR: In this paper, the authors proposed and analyzed subgradient methods for the problem of finding a sample mean in DTW spaces, which generalizes existing sample mean algorithms such as DTW Barycenter Averaging (DBA).

Journal ArticleDOI
TL;DR: This work introduces information divergence in the field of signature verification to take full advantage of the information contained in all reference samples by shifting the distance measurement between the test and reference samples to a similarity measurement between two distributions.
Abstract: The phenomenon of data clutter caused by intervariability (individual features) and intravariability (intrinsic noise of reference samples) is one of the most important reasons for performance degradation in online signature verification systems. To address this problem, we introduce information divergence in the field of signature verification to take full advantage of the information contained in all reference samples by shifting the distance measurement between the test and reference samples to a similarity measurement between two distributions (generated between reference samples as well as between the test sample and all reference samples) and simultaneously make full use of the spatial information of the reference samples. Based on that change, we propose a novel information divergence framework and provide some matcher instances that work within the proposed framework to effectively improve the performance of a signature verification system. Furthermore, to exploit the advantages of the new matching strategy, we propose a new dynamic time warping algorithm. In addition, we provide in-depth analysis of several distance normalizations and apply them to signature verification to reduce the adverse effect of “clutter” on signature data, which can effectively improve system performance. The experimental results on the MCYT-100 and SUSIG signature databases achieved equal error rates of 2.25% and 1.70% when 10 reference samples were used and 3.16% and 2.13% when 5 reference samples were used, respectively, illustrating the effectiveness of the proposed strategy in relation to other state-of-the-art strategies.

Journal ArticleDOI
TL;DR: Two methods, which are based on full factorial experiment design and optimal orthogonal experiment design, are proposed for selecting discriminative features among candidates to improve the robustness of on-line handwritten signatures.

Journal ArticleDOI
TL;DR: A linear fuzzy information granule-based dynamic time warping (LFIG_DTW) algorithm is developed for calculating the distance of two equal-length or unequal-length granular time series, and a distance-based hierarchical clustering method is designed for time-series clustering.

Journal ArticleDOI
TL;DR: The nearly 50-year-old quadratic time bound for computing Dynamic Time Warping or GED between two sequences of n points in R is broken by presenting deterministic algorithms that run in O(n2 log log log n/log log n) time.
Abstract: Dynamic Time Warping (DTW) and Geometric Edit Distance (GED) are basic similarity measures between curves or general temporal sequences (e.g., time series) that are represented as sequences of points in some metric space (X, dist). The DTW and GED measures are massively used in various fields of computer science and computational biology. Consequently, the tasks of computing these measures are among the core problems in P. Despite extensive efforts to find more efficient algorithms, the best-known algorithms for computing the DTW or GED between two sequences of points in X = Rd are long-standing dynamic programming algorithms that require quadratic runtime, even for the one-dimensional case d = 1, which is perhaps one of the most used in practice.In this article, we break the nearly 50-year-old quadratic time bound for computing DTW or GED between two sequences of n points in R by presenting deterministic algorithms that run in O(n2 log log log n/ log log n) time. Our algorithms can be extended to work also for higher-dimensional spaces Rd, for any constant d, when the underlying distance-metric dist is polyhedral (e.g., L1, Linfin).

Journal ArticleDOI
TL;DR: The study supports the use of HSMMs to assess motor performance providing a quantitative feedback to physiotherapist and patients and its correlation better with the physician's score than DTW.

Proceedings Article
01 Jan 2018
TL;DR: This paper introduces FastWWSearch: an efficient and exact method to learn WW, which shows on 86 datasets that the method is always faster than the state of the art, with at least one order of magnitude and up to 1000x speed-up.
Abstract: Time series classification maps time series to labels. The nearest neighbor algorithm (NN) using the Dynamic Time Warping (DTW) similarity measure is a leading algorithm for this task and a component of the current best ensemble classifiers for time series. However, NN-DTW is only a winning combination when its meta-parameter – its warping window – is learned from the training data. The warping window (WW) intuitively controls the amount of distortion allowed when comparing a pair of time series. With a training database of N time series of lengths L, a naive approach to learning the WW requires Θ(N·L) operations. This often results in NN-DTW requiring days for training on datasets containing a few thousand time series only. In this paper, we introduce FastWWSearch: an efficient and exact method to learn WW. We show on 86 datasets that our method is always faster than the state of the art, with at least one order of magnitude and up to 1000x speed-up.

Journal ArticleDOI
TL;DR: The dynamic time warping is more robust to the above-mentioned errors and more accurately detects damage with weak ultrasonic signatures and is compared with stretch-based methods.
Abstract: Guided wave structural health monitoring is widely researched for remotely inspecting large structural areas. To detect, locate, and characterize damage, guided wave methods often compare data to a baseline signal. Yet, environmental variations create large differences between the baseline and the collected measurements. These variations hide damage signatures and cause false detection. Temperature compensation algorithms, such as baseline signal stretch and the scale transform have been used to optimally realign data to a baseline. While these methods are effective in some conditions, their performance deteriorates in the presence of large temperature variations, long propagation distances, and high frequencies. In this paper, we use dynamic time warping to better align guided wave data and to overcome these errors. When compared with stretch-based methods, we show that the dynamic time warping is more robust to the above-mentioned errors and more accurately detects damage with weak ultrasonic signatures.

Journal ArticleDOI
TL;DR: A new time warping similarity measure (AWarp) for sparse time series that works on the run-length encoded representation of sparse timeseries and is exact for binary-valued time series and a close approximation of the original DTW distance for any-valued series.
Abstract: Dynamic time warping (DTW) distance has been effectively used in mining time series data in a multitude of domains. However, in its original formulation DTW is extremely inefficient in comparing long sparse time series, containing mostly zeros and some unevenly spaced nonzero observations. Original DTW distance does not take advantage of this sparsity, leading to redundant calculations and a prohibitively large computational cost for long time series. We derive a new time warping similarity measure (AWarp) for sparse time series that works on the run-length encoded representation of sparse time series. The complexity of AWarp is quadratic on the number of observations as opposed to the range of time of the time series. Therefore, AWarp can be several orders of magnitude faster than DTW on sparse time series. AWarp is exact for binary-valued time series and a close approximation of the original DTW distance for any-valued series. We discuss useful variants of AWarp: bounded (both upper and lower), constrained, and multidimensional. We show applications of AWarp to three data mining tasks including clustering, classification, and outlier detection, which are otherwise not feasible using classic DTW, while producing equivalent results. Potential areas of application include bot detection, human activity classification, search trend analysis, seismic analysis, and unusual review pattern mining.

Posted Content
TL;DR: This work studies a family of alignment-aware positive definite (p.d.) kernels, with its feature embedding given by a distribution of Random Warping Series (RWS), which reduces the computational complexity of existing DTW-based techniques from quadratic to linear in terms of both the number and the length of time-series.
Abstract: Time series data analytics has been a problem of substantial interests for decades, and Dynamic Time Warping (DTW) has been the most widely adopted technique to measure dissimilarity between time series. A number of global-alignment kernels have since been proposed in the spirit of DTW to extend its use to kernel-based estimation method such as support vector machine. However, those kernels suffer from diagonal dominance of the Gram matrix and a quadratic complexity w.r.t. the sample size. In this work, we study a family of alignment-aware positive definite (p.d.) kernels, with its feature embedding given by a distribution of \emph{Random Warping Series (RWS)}. The proposed kernel does not suffer from the issue of diagonal dominance while naturally enjoys a \emph{Random Features} (RF) approximation, which reduces the computational complexity of existing DTW-based techniques from quadratic to linear in terms of both the number and the length of time-series. We also study the convergence of the RF approximation for the domain of time series of unbounded length. Our extensive experiments on 16 benchmark datasets demonstrate that RWS outperforms or matches state-of-the-art classification and clustering methods in both accuracy and computational time. Our code and data is available at { \url{this https URL}}.

Journal ArticleDOI
TL;DR: This study proposes a long-term prediction approach by transforming the original numerical data into some meaningful and interpretable entities following the principle of justifiable granularity to equalize temporal sequences exhibiting sound semantics.
Abstract: In time series forecasting, a challenging and important task is to realize long-term forecasting that is both accurate and transparent In this study, we propose a long-term prediction approach by transforming the original numerical data into some meaningful and interpretable entities following the principle of justifiable granularity The obtained sequences exhibiting sound semantics may have different lengths, which bring some difficulties when carrying out predictions To equalize these temporal sequences, we propose to adjust their lengths by involving the dynamic time warping (DTW) distance Two theorems are included to ensure the correctness of the proposed equalization approach Finally, we exploit hidden Markov models (HMM) to derive the relations existing in the granular time series A series of experiments using publicly available data are conducted to assess the performance of the proposed prediction method The comparative analysis demonstrates the performance of the prediction delivered by the proposed model