scispace - formally typeset
Search or ask a question

Showing papers on "Dynamic time warping published in 2012"


Proceedings ArticleDOI
12 Aug 2012
TL;DR: This work shows that by using a combination of four novel ideas the authors can search and mine truly massive time series for the first time, and shows that in large datasets they can exactly search under DTW much more quickly than the current state-of-the-art Euclidean distance search algorithms.
Abstract: Most time series data mining algorithms use similarity search as a core subroutine, and thus the time taken for similarity search is the bottleneck for virtually all time series data mining algorithms. The difficulty of scaling search to large datasets largely explains why most academic work on time series data mining has plateaued at considering a few millions of time series objects, while much of industry and science sits on billions of time series objects waiting to be explored. In this work we show that by using a combination of four novel ideas we can search and mine truly massive time series for the first time. We demonstrate the following extremely unintuitive fact; in large datasets we can exactly search under DTW much more quickly than the current state-of-the-art Euclidean distance search algorithms. We demonstrate our work on the largest set of time series experiments ever attempted. In particular, the largest dataset we consider is larger than the combined size of all of the time series datasets considered in all data mining papers ever published. We show that our ideas allow us to solve higher-level time series data mining problem such as motif discovery and clustering at scales that would otherwise be untenable. In addition to mining massive datasets, we will show that our ideas also have implications for real-time monitoring of data streams, allowing us to handle much faster arrival rates and/or use cheaper and lower powered devices than are currently possible.

969 citations


Proceedings ArticleDOI
05 May 2012
TL;DR: In this article, an implicit authentication approach that enhances password patterns with an additional security layer, transparent to the user, is introduced, where users are not only authenticated by the shape they input but also by the way they perform the input.
Abstract: Password patterns, as used on current Android phones, and other shape-based authentication schemes are highly usable and memorable. In terms of security, they are rather weak since the shapes are easy to steal and reproduce. In this work, we introduce an implicit authentication approach that enhances password patterns with an additional security layer, transparent to the user. In short, users are not only authenticated by the shape they input but also by the way they perform the input. We conducted two consecutive studies, a lab and a long-term study, using Android applications to collect and log data from user input on a touch screen of standard commercial smartphones. Analyses using dynamic time warping (DTW) provided first proof that it is actually possible to distinguish different users and use this information to increase security of the input while keeping the convenience for the user high.

486 citations


Journal ArticleDOI
TL;DR: A novel keyword spotting method for handwritten documents is described, derived from a neural network-based system for unconstrained handwriting recognition, that performs template-free spotting, i.e., it is not necessary for a keyword to appear in the training set.
Abstract: Keyword spotting refers to the process of retrieving all instances of a given keyword from a document. In the present paper, a novel keyword spotting method for handwritten documents is described. It is derived from a neural network-based system for unconstrained handwriting recognition. As such it performs template-free spotting, i.e., it is not necessary for a keyword to appear in the training set. The keyword spotting is done using a modification of the CTC Token Passing algorithm in conjunction with a recurrent neural network. We demonstrate that the proposed systems outperform not only a classical dynamic time warping-based approach but also a modern keyword spotting system, based on hidden Markov models. Furthermore, we analyze the performance of the underlying neural networks when using them in a recognition task followed by keyword spotting on the produced transcription. We point out the advantages of keyword spotting when compared to classic text line recognition.

283 citations


Journal ArticleDOI
TL;DR: A new representation of interventions in terms of multidimensional time-series formed by synchronized signals acquired over time is proposed, which results in workflow models combining low-level signals with high-level information such as predefined phases, which can be used to detect actions and trigger an event.

255 citations


Journal ArticleDOI
TL;DR: This letter shows that its nonlinear distortion behavior is compatible with the use of a spatiotemporal segmentation of the data cube that is formed by a satellite image time series (SITS), and proves that, by taking advantage of the spatial and temporal connectivities, both the performance and the quality of the analysis can be improved.
Abstract: Satellite Image Time Series are becoming increasingly available and will continue to do so in the coming years thanks to the launch of space missions which aim at providing a coverage of the Earth every few days with high spatial resolution. In the case of optical imagery, it will be possible to produce land use and cover change maps with detailed nomenclatures. However, due to meteorological phenomena, such as clouds, these time series will become irregular in terms of temporal sampling, and one will need to compare time series with different lengths. In this paper, we present an approach to image time series analysis which is able to deal with irregularly sampled series and which also allows the comparison of pairs of time series where each element of the pair has a different number of samples. We present the dynamic time warping from a theoretical point of view and illustrate its capabilities with two applications to real-time series.

243 citations


Journal ArticleDOI
TL;DR: This work introduces dynamic time warping to stretch each beat to match a running template and combines it with several other features related to signal quality, including correlation and the percentage of the beat that appeared to be clipped to assess the clinical utility of PPG traces.
Abstract: In this work, we describe a beat-by-beat method for assessing the clinical utility of pulsatile waveforms, primarily recorded from cardiovascular blood volume or pressure changes, concentrating on the photoplethysmogram (PPG). Physiological blood flow is nonstationary, with pulses changing in height, width and morphology due to changes in heart rate, cardiac output, sensor type and hardware or software pre-processing requirements. Moreover, considerable inter-individual and sensor-location variability exists. Simple template matching methods are therefore inappropriate, and a patient-specific adaptive initialization is therefore required. We introduce dynamic time warping to stretch each beat to match a running template and combine it with several other features related to signal quality, including correlation and the percentage of the beat that appeared to be clipped. The features were then presented to a multi-layer perceptron neural network to learn the relationships between the parameters in the presence of good- and bad-quality pulses. An expert-labeled database of 1055 segments of PPG, each 6 s long, recorded from 104 separate critical care admissions during both normal and verified arrhythmic events, was used to train and test our algorithms. An accuracy of 97.5% on the training set and 95.2% on test set was found. The algorithm could be deployed as a stand-alone signal quality assessment algorithm for vetting the clinical utility of PPG traces or any similar quasi-periodic signal.

240 citations


Proceedings ArticleDOI
16 Jun 2012
TL;DR: Experimental results demonstrate that GTW can efficiently solve the multi-modal temporal alignment problem and outperforms state-of-the-art DTW methods for temporal alignment of time series within the same modality.
Abstract: Temporal alignment of human motion has been a topic of recent interest due to its applications in animation, telerehabilitation and activity recognition among others. This paper presents generalized time warping (GTW), an extension of dynamic time warping (DTW) for temporally aligning multi-modal sequences from multiple subjects performing similar activities. GTW solves three major drawbacks of existing approaches based on DTW: (1) GTW provides a feature weighting layer to adapt different modalities (e.g., video and motion capture data), (2) GTW extends DTW by allowing a more flexible time warping as combination of monotonic functions, (3) unlike DTW that typically incurs in quadratic cost, GTW has linear complexity. Experimental results demonstrate that GTW can efficiently solve the multi-modal temporal alignment problem and outperforms state-of-the-art DTW methods for temporal alignment of time series within the same modality.

185 citations


Journal ArticleDOI
01 Aug 2012
TL;DR: The principal advantage of the proposed approach is utilization of the trajectory key points from all demonstrations for generation of a generalized trajectory, resulting in a generalization procedure which accounts for the relevance of reproduction of different parts of the trajectories.
Abstract: The main objective of this paper is to develop an efficient method for learning and reproduction of complex trajectories for robot programming by demonstration. Encoding of the demonstrated trajectories is performed with hidden Markov model, and generation of a generalized trajectory is achieved by using the concept of key points. Identification of the key points is based on significant changes in position and velocity in the demonstrated trajectories. The resulting sequences of trajectory key points are temporally aligned using the multidimensional dynamic time warping algorithm, and a generalized trajectory is obtained by smoothing spline interpolation of the clustered key points. The principal advantage of our proposed approach is utilization of the trajectory key points from all demonstrations for generation of a generalized trajectory. In addition, variability of the key points' clusters across the demonstrated set is employed for assigning weighting coefficients, resulting in a generalization procedure which accounts for the relevance of reproduction of different parts of the trajectories. The approach is verified experimentally for trajectories with two different levels of complexity.

154 citations


Journal ArticleDOI
TL;DR: This work introduces a shape-motion prototype-based approach for action recognition that enables robust action matching in challenging situations and allows automatic alignment of action sequences.
Abstract: A shape-motion prototype-based approach is introduced for action recognition. The approach represents an action as a sequence of prototypes for efficient and flexible action matching in long video sequences. During training, an action prototype tree is learned in a joint shape and motion space via hierarchical K-means clustering and each training sequence is represented as a labeled prototype sequence; then a look-up table of prototype-to-prototype distances is generated. During testing, based on a joint probability model of the actor location and action prototype, the actor is tracked while a frame-to-prototype correspondence is established by maximizing the joint probability, which is efficiently performed by searching the learned prototype tree; then actions are recognized using dynamic prototype sequence matching. Distance measures used for sequence matching are rapidly obtained by look-up table indexing, which is an order of magnitude faster than brute-force computation of frame-to-frame distances. Our approach enables robust action matching in challenging situations (such as moving cameras, dynamic backgrounds) and allows automatic alignment of action sequences. Experimental results demonstrate that our approach achieves recognition rates of 92.86 percent on a large gesture data set (with dynamic backgrounds), 100 percent on the Weizmann action data set, 95.77 percent on the KTH action data set, 88 percent on the UCF sports data set, and 87.27 percent on the CMU action data set.

153 citations


Journal ArticleDOI
TL;DR: A framework to assist in the development of systems for the automatic recognition of high-level surgical tasks using microscope videos analysis to combine state-of-the-art computer vision techniques with time series analysis is proposed.
Abstract: The need for a better integration of the new generation of computer-assisted-surgical systems has been recently emphasized. One necessity to achieve this objective is to retrieve data from the operating room (OR) with different sensors, then to derive models from these data. Recently, the use of videos from cameras in the OR has demonstrated its efficiency. In this paper, we propose a framework to assist in the development of systems for the automatic recognition of high-level surgical tasks using microscope videos analysis. We validated its use on cataract procedures. The idea is to combine state-of-the-art computer vision techniques with time series analysis. The first step of the framework consisted in the definition of several visual cues for extracting semantic information, therefore, characterizing each frame of the video. Five different pieces of image-based classifiers were, therefore, implemented. A step of pupil segmentation was also applied for dedicated visual cue detection. Time series classification algorithms were then applied to model time-varying data. Dynamic time warping and hidden Markov models were tested. This association combined the advantages of all methods for better understanding of the problem. The framework was finally validated through various studies. Six binary visual cues were chosen along with 12 phases to detect, obtaining accuracies of 94%.

120 citations


Journal ArticleDOI
TL;DR: This paper presents a novel algorithm called correlation based dynamic time warping (CBDTW) which combines DTW and PCA based similarity measures and shows that CBDTW outperforms the standard SPCA and the most commonly used, Euclidean distance based multivariate DTW in case of datasets with complex correlation structure.
Abstract: In recent years, dynamic time warping (DTW) has begun to become the most widely used technique for comparison of time series data where extensive a priori knowledge is not available. However, it is often expected a multivariate comparison method to consider the correlation between the variables as this correlation carries the real information in many cases. Thus, principal component analysis (PCA) based similarity measures, such as PCA similarity factor (SPCA), are used in many industrial applications. In this paper, we present a novel algorithm called correlation based dynamic time warping (CBDTW) which combines DTW and PCA based similarity measures. To preserve correlation, multivariate time series are segmented and the local dissimilarity function of DTW originated from SPCA. The segments are obtained by bottom-up segmentation using special, PCA related costs. Our novel technique qualified on two databases, the database of signature verification competition 2004 and the commonly used AUSLAN dataset. We show that CBDTW outperforms the standard SPCA and the most commonly used, Euclidean distance based multivariate DTW in case of datasets with complex correlation structure.

Book ChapterDOI
18 Sep 2012
TL;DR: Using hourly time series, a Dynamic Time Warping algorithm is applied to measure the similarity between these time series and investigate the outlier urban areas identified through abnormal mobility patterns.
Abstract: The rapid development of information and communication technologies (ICTs) has provided rich resources for spatio-temporal data mining and knowledge discovery in modern societies. Previous research has focused on understanding aggregated urban mobility patterns based on mobile phone datasets, such as extracting activity hotspots and clusters. In this paper, we aim to go one step further from identifying aggregated mobility patterns. Using hourly time series we extract and represent the dynamic mobility patterns in different urban areas. A Dynamic Time Warping (DTW) algorithm is applied to measure the similarity between these time series, which also provides input for classifying different urban areas based on their mobility patterns. In addition, we investigate the outlier urban areas identified through abnormal mobility patterns. The results can be utilized by researchers and policy makers to understand the dynamic nature of different urban areas, as well as updating environmental and transportation policies.

Journal ArticleDOI
TL;DR: The proposed similarity outperforms the traditional DTW between the original sequences, and the model-based approach which uses ordinary continuous HMMs, and it is shown that this increase in accuracy can be traded against a significant reduction of the computational cost.
Abstract: This paper proposes a novel similarity measure between vector sequences. We work in the framework of model-based approaches, where each sequence is first mapped to a Hidden Markov Model (HMM) and then a measure of similarity is computed between the HMMs. We propose to model sequences with semicontinuous HMMs (SC-HMMs). This is a particular type of HMM whose emission probabilities in each state are mixtures of shared Gaussians. This crucial constraint provides two major benefits. First, the a priori information contained in the common set of Gaussians leads to a more accurate estimate of the HMM parameters. Second, the computation of a similarity between two SC-HMMs can be simplified to a Dynamic Time Warping (DTW) between their mixture weight vectors, which significantly reduces the computational cost. Experiments are carried out on a handwritten word retrieval task in three different datasets-an in-house dataset of real handwritten letters, the George Washington dataset, and the IFN/ENIT dataset of Arabic handwritten words. These experiments show that the proposed similarity outperforms the traditional DTW between the original sequences, and the model-based approach which uses ordinary continuous HMMs. We also show that this increase in accuracy can be traded against a significant reduction of the computational cost.

Journal ArticleDOI
TL;DR: This paper explores the automatic classification of a set of SPs based on the Dynamic Time Warping (DTW) algorithm, which is used to compute a similarity measure between two SPs that focuses on the different types of activities performed during surgery and their sequencing, by minimizing time differences.

Proceedings ArticleDOI
01 Jan 2012
TL;DR: This work transforms a beat-synchronous chroma matrix with a 2D Fourier transform and shows that the resulting representation has properties that fit the cover song recognition task, and can also apply PCA to efficiently scale comparisons.
Abstract: Large-scale cover song recognition involves calculating itemto-item similarities that can accommodate differences in timing and tempo, rendering simple Euclidean measures unsuitable. Expensive solutions such as dynamic time warping do not scale to million of instances, making them inappropriate for commercial-scale applications. In this work, we transform a beat-synchronous chroma matrix with a 2D Fourier transform and show that the resulting representation has properties that fit the cover song recognition task. We can also apply PCA to efficiently scale comparisons. We report the best results to date on the largest available dataset of around 18,000 cover songs amid one million tracks, giving a mean average precision of 3.0%.

Journal ArticleDOI
TL;DR: The notion of compact multiple alignment is introduced, which allows us to link these theories to the problem of summarizing under time warping and to use a genetic algorithm based on a specific representation of the genotype inspired by genes to consistently paint the fitness landscape.

Proceedings ArticleDOI
01 Nov 2012
TL;DR: Two approaches for identification based on biometric gait using acceleration sensor - called accelerometer and Support Vector Machine are presented.
Abstract: In this paper, we present two approaches for identification based on biometric gait using acceleration sensor - called accelerometer. In contrast to preceding works, acceleration data are acquired from built-in sensor in mobile phone placed at the trouser pocket position. Data are then analyzed in both time domain and frequency domain. In time domain, gait templates are extracted and Dynamic Time Warping (DTW) is used to evaluate the similarity score. On the other hand, extracted features in frequency domain are classified using Support Vector Machine (SVM). With the participation of total 11 volunteers over 24 years old in our experiment, we achieved the accuracy of both methods respectively 79.1% and 92.7%.

Proceedings ArticleDOI
22 Aug 2012
TL;DR: This work introduces a method for real-time gesture recognition from a noisy skeleton stream, such as the ones extracted from Kinect depth sensors, using a tailored angular representation of the skeleton joints.
Abstract: Human gesture recognition is a challenging task with many applications. The popularization of real time depth sensors even diversifies potential applications to end-user natural user interface (NUI). The quality of such NUI highly depends on the robustness and execution speed of the gesture recognition. This work introduces a method for real-time gesture recognition from a noisy skeleton stream, such as the ones extracted from Kinect depth sensors. Each pose is described using a tailored angular representation of the skeleton joints. Those descriptors serve to identify key poses through a multi-class classifier derived from Support Vector learning machines. The gesture is labeled on-the-fly from the key pose sequence through a decision forest, that naturally performs the gesture time warping and avoids the requirement for an initial or neutral pose. The proposed method runs in real time and shows robustness in several experiments.

Proceedings ArticleDOI
25 Mar 2012
TL;DR: Experimental results show that the ASMtokenizer outperforms a conventional GMM tokenizer and a language-mismatched phoneme recognizer, and the performance is significantly improved by applying unsupervised speaker normalization techniques.
Abstract: The framework of posteriorgram-based template matching has been shown to be successful for query-by-example spoken term detection (STD). This framework employs a tokenizer to convert query examples and test utterances into frame-level posteriorgrams, and applies dynamic time warping to match the query posteriorgrams with test posteriorgrams to locate possible occurrences of the query term. It is not trivial to design a reliable tokenizer due to heterogeneous test conditions and the limitation of training resources. This paper presents a study of using acoustic segment models (ASMs) as the tokenizer. ASMs can be obtained following an unsupervised iterative procedure without any training transcriptions. The STD performance of the ASM tokenizer is evaluated on Fisher Corpus with comparison to three alternative tokenizers. Experimental results show that the ASM tokenizer outperforms a conventional GMM tokenizer and a language-mismatched phoneme recognizer. In addition, the performance is significantly improved by applying unsupervised speaker normalization techniques.

Journal ArticleDOI
TL;DR: The objective of this study was the development of a remote monitoring system to monitor and detect simple motor seizures using accelerometer-based kinematic sensors from subjects undergoing medication titration at the Beth Israel Deaconess Medical Center.
Abstract: The objective of this study was the development of a remote monitoring system to monitor and detect simple motor seizures. Using accelerometer-based kinematic sensors, data were gathered from subjects undergoing medication titration at the Beth Israel Deaconess Medical Center. Over the course of the study, subjects repeatedly performed a predefined set of instrumental activities of daily living (iADLs). During the monitoring sessions, EEG and video data were also recorded and provided the gold standard for seizure detection. To distinguish seizure events from iADLs, we developed a template matching algorithm. Considering the unique signature of seizure events and the inherent temporal variability of seizure types across subjects, we incorporated a customized mass-spring template into the dynamic time warping algorithm. We then ported this algorithm onto a commercially available internet tablet and developed our body sensor network on the Mercury platform. We designed several policies on this platform to compare the tradeoffs between feature calculation, raw data transmission, and battery lifetime. From a dataset of 21 seizures, the sensitivity for our template matching algorithm was found to be 0.91 and specificity of 0.84. We achieved a battery lifetime of 10.5 h on the Mercury platform.

Proceedings ArticleDOI
25 Mar 2012
TL;DR: A fast unsupervised spoken term detection system based on lower-bound Dynamic Time Warping (DTW) search on Graphical Processing Units (GPUs) and the K nearest neighbor DTW search are presented.
Abstract: In this paper we present a fast unsupervised spoken term detection system based on lower-bound Dynamic Time Warping (DTW) search on Graphical Processing Units (GPUs). The lower-bound estimate and the K nearest neighbor DTW search are carefully designed to fit the GPU parallel computing architecture. In a spoken term detection task on the TIMIT corpus, a 55x speed-up is achieved compared to our previous implementation on a CPU without affecting detection performance. On large, artificially created corpora, measurements show that the total computation time of the entire spoken term detection system grows linearly with corpus size. On average, searching a keyword on a single desktop computer with modern GPUs requires 2.4 seconds/corpus hour.

Journal ArticleDOI
TL;DR: Novel techniques for feature parameter extraction based on MFCC and feature recognition using dynamic time warping algorithm for application in speaker-independent isolated digits recognition and the proposed Weighted MFCC is proposed.
Abstract: this paper, we propose novel techniques for feature parameter extraction based on MFCC and feature recognition using dynamic time warping algorithm for application in speaker-independent isolated digits recognition. Using the proposed Weighted MFCC (WMFCC), we achieve low computational overhead for the feature recognition stage since we use only 13 weighted MFCC coefficients instead of the conventional 39 MFCC coefficients including the delta and double delta features. In order to capture the trends or patterns that a feature sequence presents during the alignment process, we compute the local and global features using Improved Features for DTW algorithm (IFDTW), rather than using the pure feature values or their estimated derivatives. The experiments based on TI-Digits corpus demonstrate the effectiveness of proposed techniques leading to higher recognition accuracy of 98.13%.

Book ChapterDOI
29 May 2012
TL;DR: A Shape-based Clustering for Time Series (SCTS) is proposed using a novel averaging method called Ranking Shape- based Template Matching Framework (RSTMF), which can average a group of time series effectively but take as much as 400 times less computational time than that of STMF.
Abstract: One of the most famous algorithms for time series data clustering is k -means clustering with Euclidean distance as a similarity measure. However, many recent works have shown that Dynamic Time Warping (DTW) distance measure is more suitable for most time series data mining tasks due to its much improved alignment based on shape. Unfortunately, k -means clustering with DTW distance is still not practical since the current averaging functions fail to preserve characteristics of time series data within the cluster. Recently, Shape-based Template Matching Framework (STMF) has been proposed to discover a cluster representative of time series data. However, STMF is very computationally expensive. In this paper, we propose a Shape-based Clustering for Time Series (SCTS) using a novel averaging method called Ranking Shape-based Template Matching Framework (RSTMF), which can average a group of time series effectively but take as much as 400 times less computational time than that of STMF. In addition, our method outperforms other well-known clustering techniques in terms of accuracy and criterion based on known ground truth.

Proceedings ArticleDOI
01 Dec 2012
TL;DR: Experimental results show that the proposed framework improves the relative performance on a mispronounced word detection task by nearly 50% compared to an approach that only considers DTW alignment scores.
Abstract: The task of mispronunciation detection for language learning is typically accomplished via automatic speech recognition (ASR). Unfortunately, less than 2% of the world's languages have an ASR capability, and the conventional process of creating an ASR system requires large quantities of expensive, annotated data. In this paper we report on our efforts to develop a comparison-based framework for detecting word-level mispronunciations in nonnative speech. Dynamic time warping (DTW) is carried out between a student's (non-native speaker) utterance and a teacher's (native speaker) utterance, and we focus on extracting word-level and phone-level features that describe the degree of mis-alignment in the warping path and the distance matrix. Experimental results on a Chinese University of Hong Kong (CUHK) nonnative corpus show that the proposed framework improves the relative performance on a mispronounced word detection task by nearly 50% compared to an approach that only considers DTW alignment scores.

Journal ArticleDOI
TL;DR: A handwritten text biometric recognition system suitable to be applied to short sequences of text (words) with the opportunity to show that pen-up strokes possess a surprisingly high discriminative power, while the performance of the combination suggests they may carry non-redundant information with respect to pen-down strokes.

Book ChapterDOI
07 Oct 2012
TL;DR: A novel multiple shots re-identification technique is proposed which combines a standard single shot re-Identification, based on offline training to recognize humans from different views, with a Dynamic Time Warping (DTW) distance.
Abstract: This paper presents a new tracking algorithm to solve on-line the 'Tag and Track' problem in a crowded scene with a network of CCTV Pan, Tilt and Zoom (PTZ) cameras. The dataset is very challenging as the non-overlapping cameras exhibit pan tilt and zoom motions, both smoothly and abruptly. Therefore a tracking-by-detection approach is combined with a re-identification method based on appearance features to solve the re-acquisition problem between non overlapping camera views and crowds occlusions. However, conventional re-identification techniques of multi target trackers, which consist of learning an online appearance model to differentiate the target of interest from other people in the scene, are not suitable for this scenario because the tagged pedestrian moves in an environment where pedestrians walking with them are constantly changing. Therefore, a novel multiple shots re-identification technique is proposed which combines a standard single shot re-identification, based on offline training to recognize humans from different views, with a Dynamic Time Warping (DTW) distance.

Journal ArticleDOI
01 Jul 2012
TL;DR: It is argued that time series often carry structural features that can be used for identifying locally relevant constraints to eliminate redundant work and proposed salient feature based sDTW algorithms which first identify robust salient features in the given time series and then find a consistent alignment of these to establish the boundaries for the warp path search.
Abstract: Many applications generate and consume temporal data and retrieval of time series is a key processing step in many application domains. Dynamic time warping (DTW) distance between time series of size N and M is computed relying on a dynamic programming approach which creates and fills an N x M grid to search for an optimal warp path . Since this can be costly, various heuristics have been proposed to cut away the potentially unproductive portions of the DTW grid. In this paper, we argue that time series often carry structural features that can be used for identifying locally relevant constraints to eliminate redundant work. Relying on this observation, we propose salient feature based sDTW algorithms which first identify robust salient features in the given time series and then find a consistent alignment of these to establish the boundaries for the warp path search. More specifically, we propose alternative fixed core&adaptive width, adaptive core&fixed width , and adaptive core&adaptive width strategies which enforce different constraints reflecting the high level structural characteristics of the series in the data set. Experiment results show that the proposed sDTW algorithms help achieve much higher accuracy in DTW computation and time series retrieval than fixed core & fixed width algorithms that do not leverage local features of the given time series.

Journal ArticleDOI
TL;DR: This work presents a word spotting method for scanned documents in order to find the word images that are similar to a query word, without assuming a correct segmentation of the words into characters.

Proceedings ArticleDOI
03 Jul 2012
TL;DR: Insight is given about the influence of different walking speeds (slow, normal and fast) and surfaces and surfaces (flat carpeted, grass, gravel and inclined) on gait recognition.
Abstract: This paper gives an insight about the influence of different walking speeds (slow, normal and fast) and surfaces (flat carpeted, grass, gravel and inclined) on gait recognition. Gait recognition is a type of biometric authentication that operates on behavioral characteristics of human beings. This research utilizes wearable sensors, and we have used a commercially available mobile device. Gait data is collected from 48 subjects for six different walk settings in two sessions on different days to measure same-day and cross-day performance. Gait cycles are extracted and compared using dynamic time warping as distance metric. Different parameter settings are evaluated to optimize the cycle extraction process.

Proceedings Article
01 Jan 2012
TL;DR: This work shows that it can mitigate this untenable lethargy by casting DTW clustering as an anytime algorithm, and develops a novel data-adaptive approximation to DTW which can be quickly computed, and which produces approximations toDTW that are much better than the best currently known linear-time approximation.
Abstract: Given the ubiquity of time series data, the data mining community has spent significant time investigating the best time series similarity measure to use for various tasks and domains. After more than a decade of extensive efforts, there is increasing evidence that Dynamic Time Warping (DTW) is very difficult to beat. Given that, recent efforts have focused on making the intrinsically slow DTW algorithm faster. For the similarity-search task, an important subroutine in many data mining algorithms, significant progress has been made by replacing the vast majority of expensive DTW calculations with cheap-to-compute lower bound calculations. However, these lower bound based optimizations do not directly apply to clustering, and thus for some realistic problems, clustering with DTW can take days or weeks. In this work, we show that we can mitigate this untenable lethargy by casting DTW clustering as an anytime algorithm. At the heart of our algorithm is a novel data-adaptive approximation to DTW which can be quickly computed, and which produces approximations to DTW that are much better than the best currently known linear-time approximations. We demonstrate our ideas on real world problems showing that we can get virtually all the accuracy of a batch DTW clustering algorithm in a fraction of the time.