Abstract

Summarizing or averaging a sequential data set (i.e., a set of time series) can be comprehensively approached as a result of sophisticated computational tools. Averaging under Dynamic Time Warping (DTW) is one such tool that captures consensus patterns. DTW acts as a similarity measure between time series, and subsequently, an averaging method must be executed upon the behaviour of DTW. However, averaging under DTW somewhat neglects temporal aspect since it is on the search of similar appearances rather than stagnating on corresponding time-points. On the contrary, the mean series carrying point-wise averages provides only a weak consensus pattern as it may over-smooth important temporal variations. As a compromise, a pool of consensus series termed Ultimate Tamed Series (UTS) is studied here that adheres to temporal decomposition supported by the discrete Haar wavelet. We claim that UTS summarizes localized patterns, which would not be reachable via the series under DTW or the mean series. Neighbourhood of localization can be altered as a user can customize different levels of decomposition. In validation, comparisons are carried out with the series under DTW and the mean series via Euclidean distance and the distance resulted by DTW itself. Two sequential data sets are selected for this purpose from a standard repository.

1. Introduction

Classical topic of summarizing and classifying a sequential data set (i.e., a set of time series) has recently gained a new outlook because of data mining and machine learning procedures [13]. Investigating common pattern (or shape) attributes (e.g., monotonic, oscillatory, and periodic) of a set of time series is ever so attainable, but still challenging even with advanced computer technology. The basic reason is that data are ordered unlike in a usual data set. In practice, time series arise in many scenarios such as economic and financial indices, weather forecasting, disease transmission, census data, market surveys, quality control, and inventory studies [4]. Data availability is also unprecedented as digital platforms facilitate gathering, storing, and processing data in profound ways [5]. Therefore, we always need effective and efficient techniques to extract valuable information from sequential data [68].

Finding a consensus or representative time series for a set of time series (i.e., a pattern similar to many series) is often found in a handful of phenomena. Daily rainfall of a region reported by several substations, fluctuation of a financial index in different stock markets and banks, daily sales of goods in different outlets of a supermarket chain, speech and gesture recognition, and genetic sequencing of an organism are among such phenomena [911]. The domain of a time series does not necessarily stand for time; for instance, it refers to the order of nucleotides in deoxyribonucleic acid (DNA). Pattern recognition not only in series-wise but also as a whole set of series is a key concern. In this work, we are interested in the latter. A measure called Dynamic Time Warping (DTW) is frequently used to extract similarity between time series. Then, one can design an averaging method based on the behaviour of DTW that gives a consensus pattern (i.e., a pattern aligned with many series in the set) [9, 12]. DTW has been developed as a nonlinear mapping between time series to map similar features, neglecting temporal distortions [13]. Thus, a resulting series under DTW captures attributes of pattern rather than attributes of point-wise central tendency. We name the “series under DTW” as “Dynamic Time Warping Series (DTWS).” This is similar to the naming of “mean series” giving the “series under point-wise averaging.” The main drawback of DTWS is its limitations on representing local patterns (i.e., corresponding patterns upon temporal proximity). DTW approaches two time series at a time structuring a distance matrix and backtracks in the direction of minimum cumulative distances to achieve so called warping path [14, 15]. Formulation of this output (DTWS) for a set of time series via DTW is explained in detail in Section 2. Improved versions of DTW are available in the literature such as NLAAF (Non-linear Alignment and Averaging Filters), PSA (Prioritized Shape averaging), and RSA (Re-sampling Shape Averaging) which address the limitations of the algorithm in different aspects. Further, there are techniques such as Contrast Enhanced Dynamic Time Warping (CEDTW) and DTW Barycenter Averaging (DBA) which have greater accuracy and higher performance [15]. However, for this study, we stick to the basic DTW algorithm leaving improved versions into future work. This study is the first of this kind that compares UTS with consensus series under DTW.

In contrast to DTWS, the series produced by point-wise averages (the mean series) represents central tendency, but hardly absorbs pattern attributes. The mean for any set of data points yields the minimum squared deviation with the data set. Therefore, for any set of time series, the mean series assures the minimum point-wise squared deviation. Consequently, the mean series may be an over-smoothed representative, restricting to impart patterns of data series.

To address the limitations of the mean series, an alternative technique called Ultimate Tamed Series (UTS) has been proposed [1618]. On top of DTWS, it acquires local patterns with customized temporal windows. UTS is based on discrete Haar wavelet transform. In literature, Haar wavelet is recorded as the earliest wavelet approach and also the simplest in structure [1921]. In formulating UTS, a variant of Haar matrix is implemented transforming a time series into two series, termed as scaling coefficients (father wavelets) and wavelet coefficients (mother wavelets). These two wavelet types are worked out in a specific way called taming [22]. Preliminary details of wavelets and taming are recalled in Section 3. This paper extends UTS technique into higher levels of transform aligning with multiterminal binary decision diagrams available to Haar wavelet [19]. It is designed upon father wavelets that behaves as level-wise representatives of the original time series.

Similar to DTWS, UTS is also operated with two time series at a time. After operating for two time series (chosen from a given set of data series), the resultant series and the next data series are operated together and the same process is recursively carried out. Since DTWS and UTS do not accommodate commutative and associative properties, a lot of consensus series can be determined. Note that DTW is a pseudo-metric which preserves the commutative property giving the same warping distance irrespective of the order of concern for two time series. However, DTWS for two time series may vary with its order of concern due to the backtracking process in generating the consensus series. In the backtracking process, different warping paths can be generated when there is a tie between minimum neighbouring elements in the distance matrix. This scenario is explained in [23], and the DTWS formulation is described in Section 2. Thus, DTW and UTS depend on hierarchies of taking time series into the respective computational procedures [22, 23]. In this comparative study, two specific hierarchies are considered catering the highest and the lowest variations among consecutive data points in each series.

Three consensus approaches, the mean series, DTWS, and UTS, establish pattern attributes in different extents. Identifying the capability of representing patterns is worth investigating which is the main task of this paper. We test these measures for OSULeaf data set and Yoga data set available in UEA & UCR Time Series Classification Repository [24]. Section 5 contains details of data processing used to implement consensus approaches mentioned above. In the sequel, we present basic routines of DTW (Section 2) and UTS (Section 3) along with comparison tools (Section 4). Furthermore, Section 3.3 presents theoretical support for the claims tested in Section 6.

2. Dynamic Time Warping Series (DTWS)

DTW algorithm is based on a distance measure called Levenshtein distance [13, 25]. For two time series of equal length denoted in sequential terminology and , the Levenshtein string metric is recursively obtained by where is the distance between and given by . Here, and are subsequences of and , respectively (i.e., ). Using the metric , a matrix is formed where the entry is determined by the value . Next, a string called the optimal warping path , where for some , is constructed by backtracking the entries of the matrix in the direction of minimum cumulative distance starting from the entry to the entry (1,1) [26]. Note that, for each and , there is a corresponding entry from series and (say). Then, a consensus series of and named as Dynamic Time Warping Series (DTWS) is obtained by averaging the pairs (i.e., ) corresponding to each selected entry in the path . This is the process we refer by “averaging under DTW.” This consensus measure need not to have similar length to the original series, and also, the original series need not to be similar in length. brings a quantification of similarity between two series. Thus, we further use overall similarity given by to design a comparison tool later in Section 4. Here, lesser shows more similarity.

For more than two time series, DTWS is computed recursively considering the DTWS of the first two series with the third series and that resultant DTWS with the fourth series and so on [15]. In this process, pairs should be averaged in a weighted manner by considering number of series infused in each appearance of DTWS. For a pair where is the entry of the series according to the hierarchy in the set and is the entry of the averaged (according to DTW) series, the entries are paired as [15]. As mentioned, hierarchy of arranging series into the warping process comes into play here giving different DTWS for different hierarchies. Since the final output in the warping process is substantially longer than the original lengths of the series, the uniform scaling is applied to achieve a consensus series of a length similar to that of data series [14, 27]. A series can be scaled to a sequence of required length , with . Then, the new series yields a scaled DTWS. This constraint on length forms a common platform for upcoming comparisons. We execute scaling after all the series are taken into warping process allowing temporal distortions to remain flexible.

3. Ultimate Tamed Series (UTS)

In this section, we present the structure of UTS, and next two sections direct its implementation as a comparable tool to the mean series and DTWS. We start with two time series where the procedure is to be recursively used upon any number of time series. First, obtaining father and mother wavelets is described in Section 3.1 as designed via Haar wavelet. Next, the taming procedure is structured in Section 3.2 emphasizing extended work on higher levels. In the final implementation, we should decide a hierarchy of taking time series into the taming procedure. Section 3.5 sets that hierarchy, and Section 5 contains some illustrations of it in association with data. In this study, we execute computer programs designed using the MATLAB software for all the implementations of UTS and DTWS, whereas any other mathematical software can also be used.

3.1. The Discrete Haar Wavelet as a Weak Form: Father and Mother Wavelets

In mathematical analysis, weak forms are applicable when it is difficult to deal with strong forms. If we treat pattern recognition of time series as a strong form, recognizing patterns via wavelet transform resembles a weak form. Then, different transforms lead to different outlooks of statistics of time series. In general, implementation of weak forms may have their own limitations as we see in Finite Element Methods that require careful choice of weight functions and trial solutions [28, 29]. UTS can also be considered as one that arises from a weak form.

The initial step of UTS is to partition time series into two-tuples (two-component vectors). For instance, for a time series , the discrete Haar transform works on two-tuples: , when we start by . This coupling leads to partial associations since two-tuples generated by shifting the start by one data point (i.e., ) are not directly approached. In a comprehensive analysis, both these couplings can be trialed to curtail missing associations in one coupling. However, the first data point is not considered in the shifted case, and inclusion of the last data point is based on whether the series has even or odd number of points. This would not be a major drawback when we deal with lengthy time series in practice. Then, the extreme data points do not hinder much the overall context of the series.

Hereafter, we name , , as 1st coupling and , , as 2nd coupling. Furthermore, we proceed with time series containing odd number of data points: (), where we ignore the first data point in 2nd coupling and the last data-point in 1st coupling to read as

Next, father and mother wavelets are determined by right-multiplying the two-tuples by variant Haar matrix as in Definition 1. Here, the classical Haar matrix is scaled by to cater average of data points in father wavelets. It facilitates interpretations in a way that father wavelets act as representatives of the original time series in further transform levels.

Definition 1 (-transform). Right-multiplying the two-tuples of by gives
1st coupling—father wavelets: .
1st coupling—mother wavelets: .
Alterations to the 2nd coupling comes as
2nd coupling—father wavelets: .
2nd coupling—mother wavelets: .

The inverse transform of the above -transform is structured into Definition 2.

Definition 2 (inverse -transform). Data points of can be obtained back by the following inverse operations.
1st coupling: and ;
2nd coupling: and ; .
Note that the inverse of is given by , where right-multiplying the two-tuples provides back the time series .

Remark 3. The -transform in Definition 1 is continued into further levels upon father wavelet series. We address this in Section 3.2.1 with required work of UTS.

3.2. Taming Procedure

So called taming is the key procedure that articulates UTS. It is nurtured in the transformed series obtained via -transform. Here, we first present the way of taming two series and next extend it into more series and more transform levels. Only 1st coupling is demonstrated here, where the same approach is applicable to 2nd coupling.

Step 1 (father and mother wavelets). Two time series and are partitioned into two-tuples of 1st coupling as and , respectively. By -transform, we have
father wavelets of series as and mother wavelets of series as .
Here, .

Step 2 (taming criteria). We define the differences and for the father and mother wavelets of two time series as and .

The taming procedure is performed via averaging either father wavelet or mother wavelet subject to the highest of and as follows (taming criteria). Superscript stands for new wavelets produced by the taming criteria. (i)If , then and (ii)If , then and (iii)If , then and

Remark 4. Geometric motivation for the above procedure has been addressed in [1618, 22, 30]. To recall along with Figure 1, let be the line segment joining the data points and , and be the line segment joining the data points and in a time series plot. Then, and (i)do not intersect if and only if (ii)intersect at a data point or coincide if and only if (iii)intersect at a point other than to data points if and only if

In summary, we tend to tame by father wavelets if the difference of them are higher (i.e., line segments are apart) and otherwise, we tame by mother wavelets. Case (ii) compromises qualitatively moderate situation.

Step 3 (tamed series). A tamed series is obtained by the inverse -transform and . We denote showing as the end result of taming and . Thus, is considered as a consensus series.

Remark 5. Note that is the same as the mean series of and , if both father and mother wavelets are averaged in Step 2 instead of one. This coins the word taming as it sounds a lesser smoothing effect compared to usual averaging. Geometric illustration of the taming criteria is available in [17, 30].

Remark 6. Operation is noncommutative and nonassociative since we bias on first series in deciding new wavelet that is not averaged in (i) and (ii) of Step 2.

We see these as optimistic properties that allow a hierarchical procedure into a consensus measure, which is not available for the mean series. A user can customize the hierarchy according to a required descriptive analysis.

Step 4 (UTS). To find a consensus series for more than two time series, the taming procedure must be extended by taming new series with already tamed series. Figure 2 indicates the flow of this recursive procedure for a set of time series . Here, is considered as an end result giving a consensus series for all series and named it as Ultimate Tamed Series (UTS), i.e., . At the taming step, averaging must be done in a weighted manner as instead of and instead of , where now the superscript (1) stands for already tamed series and superscript (2) stands for the next series newly entering to the taming process.

The entries of the above UTS acquire local trend only with an adjacent point partitioned in 1st coupling. Indeed, sign of mother wavelet illustrates increasing/decreasing trends as it caters a difference. In order to capture pattern attributes by more data points, taming procedure should be applied to further transform levels as described in Section 3.2.1. As we suggested earlier, repeating the work for 2nd coupling yields a more comprehensive analysis.

3.2.1. Higher Level Taming

After carrying out Step 1 in Section 3.2, we continue the same for resultant father wavelets. Immediate level resulting from Step 1 is termed as level 1. Next, father and mother wavelets of level 2 are taken by coupling father wavelets of level 1 and carrying out -transform for those two-tuples. In general, father and mother wavelets of level are obtained by carrying out Step 1 for father wavelets of level . Note that still we proceed with the two time series and separately. Figure 3 depicts how a time series with eight data points is transformed into four levels as a binary diagram.

Father wavelets in level rescale original time series by averaging adjacent points (data points with temporal proximity), while mother wavelets associate with the difference of such rescaled father wavelets in the previous level. The averaging occurs due to the scaling by in -transform, giving a representative role (i.e., average) to father wavelets. We later illustrate our results with these -tuple windows.

Now, we can move into higher level taming after transforming two time series and up to required levels. Taming for higher levels is carried out via Step 2 in Section 3.2 for father and mother wavelets of respective level. Suppose we select a particular as the taming level that we implement Step 2. Next, the inverse procedure similar to Step 3 is carried out in level to obtain new father wavelets of level . Note that here, we have only the father wavelets by inverse transform in Step 3. Moreover, to make a complete taming, we must continue the inverse up to the level of original series. Then, we need corresponding mother wavelets, and for that, we propose two ways as follows. (i)Inverse with Average Mother (AM-Inverse). For all the levels from level to level 1, we use respective averages of mother wavelets corresponding to two series and .(ii)Inverse with Previous Mother (PM-Inverse). For all the levels from level to level 1, we use mother wavelets corresponding to the first series .

For a set of time series, we execute the same procedure given in Step 4 (i.e., taming already tamed series with the next series adhering to weighted averages). We denote UTS obtained by taming at level as UTS-, and to distinguish under AM-inverse and PM-inverse, we further denote it as UTS-(AM) and UTS-(PM), respectively. UTS- stands for the same UTS formulated in Section 3.2. In this manner, we can generate a pool of consensus series giving options for customizing.

Remark 7. UTS-(PM) is biased more to the pattern of earlier series than the latter series in the hierarchy (order) of taking time series into the taming procedure. It is because of taking new mother wavelets as the mother wavelets of the first series (or already tamed series). UTS-(AM) is unbiased in that sense.

We further discuss about hierarchy of taming in the Section 3.5.

3.3. Well-Posedness of -Transform and Its Inverse

This subsection presents well-posedness of -transform and its inverse enclosing existence, uniqueness, and continuous change of the concerned series. We assure these properties for a given set of data series subject to a given coupling method (1st or 2nd) and for a given mother wavelet for inverse (AM-inverse or PM-inverse). Both -transform and its inverse are encountered here. Without loss of generality, we articulate propositions between level 1 and level 2 that can be generalized between level and level .

Proposition 8 (existence). There exist father and mother wavelets of level 2 corresponding to two-tuples of father wavelets of level 1 (via -transform) and vice-versa (via inverse -transform).

Proof. This result is straightforward by Definition 1 applied to the series of father wavelets of level 1 and Definition 2 applied to the inverse operations.☐

Proposition 9 (uniqueness). (i)Father and mother wavelets of level 2 are uniquely determined by father wavelets of level 1 via -transform(ii)Conversely, father wavelets of level 1 are uniquely determined by father and mother wavelets of level 2 via inverse -transform

Proof. It is enough to show the result for one two-tuple as the same can be repeated throughout. Suppose is a two-tuple of father wavelets of level 1 corresponds to father wavelet and mother wavelet of level 2.
Then, we have the following two linear systems by adjusting right-multiplying into left-multiplying appropriately. Since the determinants of and are nonzero, both the systems have unique solutions guaranteeing the result.☐

Proposition 10 (continuous dependency). (i)Father and mother wavelets of level 2 continuously depend on the changes of father wavelets of level 1(ii)Conversely, father wavelets of level 1 continuously depend on the changes of father and mother wavelets of level 2

Proof. Here also, it is enough to show the result for one two-tuple. Suppose is a two-tuple of father wavelets of level 1 corresponds to father wavelet and mother wavelet of level 2. If we change the two-tuple into , then resulting wavelets of level 2 are changed (say into and into ). Observe that For a given , there exists such that and . Similarly, we can show that, and , establishing Proposition 10 (i).
For (ii), observe that For a given , there exists such that and .
Similarly, we can show that, and , establishing Proposition 10 (ii). ☐

Remark 11. Above propositions can be advanced by taking level 1 as original time series and level 2 as level 1. Then, data points of time series (or entries of already tamed series) replacing father wavelets of level 1 and wavelets of level 2 are replaced by those of level 1.

3.4. Qualitative Illustration

As guaranteed by the Proposition 10 (i) in the view of Remark 11, wavelets of two close time series do not vary much. Here, by the term “close,” we mean positioning of corresponding data points is close (i.e., one can say that one series is obtained by making small changes to the other series). Figure 4(a) illustrates two time series that are close to each other compared to the two time series in Figure 5(a). Corresponding wavelets are shown in Figures 4(b), 4(d), 5(b), and 5(d) that illustrate small changes in data points lead to small changes in wavelets and vice versa.

When wavelets are close, resulting new wavelets by the taming criteria are also close to corresponding wavelets. Then, the resultant UTS is also close to two time series verifying Proposition 10 (ii). This is illustrated in Figures 4(c) and 5(c).

3.5. Hierarchy for DTWS and UTS

As seen in formulation, both DTWS and UTS vary according to the hierarchy (order) of taking time series into the taming procedure. In our trials, we focus on two specific hierarchies based on variations of consecutive data points. As an aggregated measure, we consider sum of squared differences of consecutive data points of each series. For a set of time series of length , these sums are given by . Now the set of time series is hierarchically arranged according to the ascending and descending order of values. Thus, we look forward to immediate closeness/variation of data points rather than their magnitude. of a particular series does not depend on other series. In addition, it caters every consecutive variation without restricting to two-tuples in coupling technique. Therefore, hierarchy does not depend on coupling technique. These characteristics motivate deciding hierarchy via .

4. Tools for Comparison

The main task of this paper is to compare UTS-, DTWS, and the mean series in the sense of grasping consensus patterns of a set of time series. Meanwhile, one can see how far central tendency can be reflected by UTS- measures. By central tendency, we mean how each entry in a consensus series represents the center of corresponding data points. We introduce two comparison tools: Euclidean distance and DTW distance between series are utilized. They are the well-known distance measures used in similar work [9, 12, 31].

4.1. Euclidean Distance

Point-wise squared deviation is taken in the Euclidean sense to estimate central tendency acquired by consensus series. The Euclidean distance between a consensus series and each time series (say ) is given by

Next a representative of should be determined (say ) for comparisons between consensus series. Thus, represents overall deviation of time series from a consensus series in the Euclidean sense. Mean and median of ’s are two possible candidates for .

4.2. Dynamic Time Warping Distance

The string metric in Section 5 yields the optimal warping path generated for two time series. We can make use of this measure to quantify the similarity between a consensus series and each data series in the sense of pattern. Suppose stands for that similarity between and (i.e., ). This comes recursively via Equation (7), which is in accordance with Equation (1).

We recall that lesser indicates more similarity in patterns. A representative DTW distance can be obtained by taking an appropriate central measure (e.g., mean or median) of (say ). Thus, we expect to represent overall deviation of time series from a consensus series in the sense of pattern. Determining and is further discussed in Section 6.2.

Remark 12. Suppose and be two consensus series of the same set of time series. In summary, if then we would expect time series are more deviated from than when pattern attributes are concerned. If then shows more central tendency than .

5. Data Processing

We test consensus series using two data sets from UEA & UCR Time Series Classification Repository, namely, OSULeaf and Yoga [24] (http://www.timeseriesclassification.com). All the data series are stored as normalized series by -scores. It is a usual measure for many experimental settings in pattern recognition [12, 32]. Such normalization is compatible with the work of UTS too. OSULeaf data set contains one-dimensional outlines of leaves obtained by colour image segmentation and boundary extraction from digitized leaf images of six classes. Angles measured at each pixel point along the boundary of leaves form a set of series. We are motivated to choose this particular data set due to shifted fluctuating patterns in different phases. In order to preserve the meaning of summarizing a set of time series, we stick into a one class of data, class six, which includes 15 time series each having 427 data points. Note that to see a complete decomposition up to a certain level, say level , we need number of data points to be divisible by . We choose first 384 consecutive data points to attain maximal comparison up to seven levels of UTS. We start by the second data point for 2nd coupling.

The second data set, Yoga, has been generated by capturing two actors transiting between yoga poses in front of a green screen giving to two classes of data. Each image has converted to a one-dimensional series by finding the outline and measuring the distance of the outline to the center. That is, the motion of an actor captured by sensors is transformed to a time series. We proceed with class 1 containing 137 series. Out of 426 data points, we proceed with the first 384 consecutive data points, again to compatible with seven levels of UTS.

All 15 series of OSULeaf data are shown in Figure 6(a) along with the mean series and DTWS. Figure 6(b) depicts those of Yoga data. Here, only a few of data series (15) are shown for clear appearance, picking every 10th series in the ascending order of (see Section 3.5) and the series having the least and the highest .

Notably, two cases of DTWS of OSULeaf data follow opposite patterns (increasing and decreasing) in many segments of time-points. However, Yoga data do not show such a significant difference in pattern. Sensitivity of ascending and descending hierarchies to the pattern of UTS- can be scrutinized later in Section 6.

6. Results

We present consensus series (UTS-, DTWS, and the mean series) in a systematic way for comparisons. Applicability of UTS- is overseen on top of DTWS and the mean series.

6.1. Illustrations of Consensus Series

We illustrate results on four pillars: (i) coupling approach—1st coupling and 2nd coupling (from Section 3.1), (ii) taming levels and mother wavelet for inverse—average mother (AM-inverse) and previous mother (PM-inverse) (from Section 3.2.1), (iii) hierarchy of data series—ascending and descending (from Section 3.5), and (iv) comparisons—via the Euclidean distance and DTW distance (from Section 4). Pattern recognition becomes more challenging when adjacent data points are highly varied in dissimilar way. If it is required to prioritize sizable adjacent variations, the preferable combination is to execute descending hierarchy with PM-inverse. As projected by Remark 6, new wavelets in taming level are biased on that of first series. Therefore, descending hierarchy brings more sensitivity to adjacent variations. This is further supported by PM-inverse as per Remark 7. According to that, not only at the taming level but also all the earlier levels in inverse procedure are biased on the first series. In addition, a theoretical support is established by Proposition 10 (generalized version) that assures higher variations in tamed series for higher variations in data series. Finally, such variations are infused into consensus series UTS-.

In the other extreme, ascending hierarchy with AM-inverse yields much smoothed output close to the mean series. It seems OSULeaf data do not show significant adjacent variations. It is observed in Figure 7 that even descending hierarchy with PM-inverse provides UTS-, , more or less the same as the mean series. In general, if we see this similarity until we proceed into higher levels, then we have less variation in early windows of temporal proximity.

Once we consider the wavelet-based temporal aspect, four windows relevant to UTS-, , consist of 2-tuples, 4-tuples, 8-tuples, and 16-tuples, respectively. This is due to each wavelet in level is responsible for data points (Section 3.2.1). Next, UTS- has a window of 32 data points, where a wavelet in level 5 is obtained using averages of 16-tuples. Thus, taming at level 5 deals with pattern of the series of 16-tuple averages. According to Figure 8, UTS deviates from the mean series, in particular with PM-inverse, suggesting considerable variations in 16-tuple averages. In 16-tuple windows (between vertical lines), we have the relevant consensus series. It can be observed that two DTWS do not adhere to such localized patterns. Moreover, DTWS is highly sensitive to hierarchy (ascending/descending) in many windows.

A similar visualization is carried out for UTS- and UTS- in Figures 9 and 10, respectively. Note that UTS- and UTS- bring consensus patterns in 32-tuples and 64-tuples, respectively. Now, UTS- follow the pattern with much relaxed localization. It is visually verified by closeness to DTWS than the mean series. In a holistic approach, a user can adjust the windows either by size or position to see respective consensus patterns, subject to the requirements of -transform. This customizing option facilitates a setting of moving-window that can be conceptualized analogous to moving-average.

Observe that skips appear in UTS- at the vertical lines separating windows. This is a result of independent backward calculation (inverse procedure) in each window that yields loss of connection across those vertical lines. In these trials of OSULeaf data, we used first 384 data points out of total length of 427 points. This is to satisfy the requirement of level 7, where the tail or rest of the data points cannot be compiled into one complete window. If someone needs to generate UTS-, say up to level 7 including the tail, we may select last 384 data points. This is only a matter of one input command in an algorithm for UTS-.

As a further verification, we produce the same illustrations for Yoga data (Figure 11). Here also, two DTWS fails to show localized pattern attributes of the data series. However, both cases of (Asc., AM) and (Des., PM) show almost similar patterns compared to that of OSULeaf data. This interprets a lesser sensitivity of hierarchy in producing localized consensus patterns via UTS-. The same is hinted by the similarity of two DTWS in more nonlocalized aspect. In such a way, observation of UTS- along with DTWS yields more useful descriptive analysis of a set of time series.

Next, we demonstrate consensus series associated with 2nd coupling. As assured by general versions of Proposition 8 and Proposition 9, existence and uniqueness preserve for both couplings, subject to a choice for hierarchy (descending/ascending) and mother wavelet for inverse (AM-inverse/PM-inverse). Here, the aim is to see whether UTS- differ in general agreement shown by 1st coupling. We proceed with 384 data points starting from the second point of OSULeaf data. Figure 12 illustrates UTS-–UTS- analogous to that of Figure 7. It demonstrates no significant difference showing slight wobbles around the mean series similar to 1st coupling.

Figure 13 illustrates the rest of UTS- preserving similar windowing effect followed earlier, subject to a shift by one point. If UTS- (Des., PM) measures do not deviate much from 1st coupling counterparts, then it is hard to expect deviations in other options of descending/ascending, AM-inverse/PM-inverse. Therefore, we present only that category showing comparison between 1st and 2nd couplings for the sake of brevity too. According to Figure 13, UTS- behave almost the same way in both the couplings.

6.2. Comparisons of Consensus Series

Earlier section is devoted to visualize the behaviour of UTS-. In this section, we provide numerical quantification to how far they deviate from data series in the sense of the Euclidean distance and DTW distance. These two distance measures are presented in Section 4 as and , respectively, where stands for data series and is the concerned consensus series.

We have 15 data series for the trial of OSULeaf data. Table 1 contains basic statistics of corresponding . Here also, we consider UTS- of descending hierarchy with PM-inverse and ascending hierarchy with AM-inverse. For brevity, only 1st coupling is implemented. The Euclidean distances when is the mean series and DTWS are also included as baselines for comparisons.

According to Table 1 The mean series shows the least mean, while DTWS leans towards the greatest mean. Therefore, UTS- are characterized with much compromised overall deviations. Mean increases when level of taming increases in both cases of descending hierarchy with PM-inverse and ascending hierarchy with AM-inverse. Notably, this guides that higher level taming directs to lesser concern on point-wise central tendency. Median does not show the same monotonic behaviour as mean. However, median of UTS- with higher is closer to that of two DTWS than the mean series. It also shows the loss of point-wise central tendency. In contrast, we expect that UTS- with higher reflect pattern attributes rather than point-wise location. To claim this, we present statistics of DTW distances in Table 2. This pattern-oriented characteristic can be declared as a key applicability of UTS- as consensus series.

According to Table 2 also, UTS- are in a compromised context as far as the mean series and DTWS are concerned. This is implied by acquiring the least mean by DTWS and the greatest mean by the mean series. The same is observed in median as well. Both mean and median decrease when level of taming increases (except in median of UTS-L6 (Asc., AM)), giving opposite to what we observe in Table 1. This verifies that UTS- with higher align more with pattern attributes. Interestingly, UTS- still follow temporal windows, where DTWS is not abide by such localization.

The standard deviation of certain UTS is comparatively high. This may be due to outliers of distance measures resulting from outlier data series rather than unsuitability of UTS as a consensus series. Therefore, as a holistic view, we depict box-plot diagrams carrying statistics of above two tables in Figure 14. The boxes are generated for UTS measures at each level together with DTWS and mean series representative. The spread of Euclidean distances and DTW distances are produced for extreme hierarchies ascending with average mother inversion and descending with previous mother inversion to generate the box plots. Now, increasing trend in the Euclidean distance and decreasing trend in DTW distance as increases are clearly visualized.

We demonstrate the same for Yoga data in Figure 15, and the same trends can be observed as in Figure 14. In a data mining process, one may investigate all those UTS- and select a fair enough representative for further work, knowing which temporal windowing prevails.

7. Discussion

Wavelet techniques are utilized in time series analysis in order to investigate trends, recurring fluctuations, and autocorrelation effects [33, 34]. These applications refer to longitudinal aspect of time series. The approach via the discrete Haar wavelet proposed in this work adds on a utility in terms of cross-sectional characteristics of a set of time series. It is associated with the central topic of consensus patterns. Its novelty lies with the ability of customizing a consensus pattern by maintaining localized temporal concern in different extents.

The study is centralized around a pool of measures designated as Ultimate Tamed Series (UTS). A theoretical platform is also established enclosing existence, uniqueness, and continuous dependency (Proposition 8, 9, and 10). Implementation of UTS in different hierarchies (ascending and descending), taming levels and inverse procedure (AM-inverse and PM-inverse), and coupling formation (1st coupling and 2nd coupling) bring the essence of the overall performance, again highlighting customizing options. Hierarchical arrangement facilitates bringing adjacent variations of data points into proposed consensus series. A window of data points is catered at a time in taming by level . UTS- measures owe this characteristic by the routine set by the discrete Haar wavelet decomposition. Wavelet approaches gain their reputation due to such localized treatment as evident in many applications [3537]. In fact, father wavelets in different levels allow rescaling data series, while mother wavelets affix localized trends. The work relevant to this begins in Definition 1 (-transform), and it is further illustrated in Figure 3.

If we shift the implementation of UTS by one point of each data series (2nd coupling), then all outputs are different. We accept this as an optimistic feature than a limitation, since more pattern attributes can be extracted. Both classes of UTS (1st and 2nd couplings) show similar pattern if there are no contrasting variations because of the shift, which is the case for trialed data in this work (Figures 12 and 13). Otherwise, it is preferred to determine UTS- for both couplings in all illustrations since we identify discrete Haar wavelet as a weak form (Section 3.1).

As baselines, we consider two consensus series, the mean series (series of point-wise averages) and Dynamic Time Warping Series (DTWS), to show how far UTS- can illustrate patterns emphasizing local attributes. Note that we have illustrated normalized data series, and hence, the mean series acts as a guide for over-smoothed representation for pattern. Option of PM-inverse avoids over-smoothing, which is often the main advantage of UTS- over the mean series. Higher level UTS- visualize pattern attributes fairly, while lower level UTS- locate closer to the mean series. By moving back and forth with different levels of taming, one can extract important characteristics of a set of time series. Thus, UTS- measures are pooled as a collection of statistics, where a user can see more complete profile rather than seeing one consensus series. Its computational burden is redundant in modern computer technology. A DTWS (without scaling) is substantially longer compared to the original series. This expansion happens in a nonlinear manner so that uniform scaling techniques may not align the series properly to the temporal axis. Therefore, improved versions such as NLAAF and PSA [14, 26] can be tested with UTS- in a future work.

8. Conclusion

UTS- bridge the gap between the mean series and DTWS in the sense of displaying consensus patterns. The classical DTW approaches only recognize patterns somewhat neglecting the correspondence to the temporal axis while the mean series restricts into point-wise concerns. UTS- with different levels of taming allow seeing consensus patterns by expanding and shrinking the localization of time-points (Figures 14 and 15). We verify these effects via two comparison tools: Euclidean distance and DTW distance.

UTS- can also be tested for data sets that are highly varied. For that, clustering tools must be proposed to see informative consensus patterns. Several motivations can be obtained by existing clustering algorithms including well-known -means [32, 38, 39]. In the broader scope of data mining, one needs not only extracting important data but also extracting useful patterns [40]. Thus, we further propose UTS as a technique for mining consensus patterns of sequential big data.

Data Availability

UEA & UCR Time Series Classification Repository can be found at http://www.timeseriesclassification.com/index.php. OSU Leaf data is available at http://www.timeseriesclassification.com/description.php?Dataset=OSULeaf. Yoga data can be found at http://www.timeseriesclassification.com/description.php?Dataset=Yoga.

Conflicts of Interest

Authors declare that there is no conflict of interest.