k-MLE, k-Bregman, k-VARs:
Theory, Convergence, Computation
Abstract
In this paper we develop a general approach to hard clustering which we call k-MLE and provide a general convergence result for it. Unlike other hard clustering generalizations of k-means, which are based on distance or divergence, k-MLE is based on likelihood and thus has a far greater range of application. We show that ‘k-Bregman’ clustering is a special case of k-MLE and thus provide, for the first time, a complete proof of convergence for k-Bregman clustering. We give a further application: k-VARs for clustering vector autocorrelated/autoregressive time series. It does not admit a Bregman divergence interpretation. We provide simulations and real data application as well as convergence results.
Index Terms:
Time-series clustering, multivariate time series, k-MLE, k-means, k-Bregman.I Introduction
Clustering is a method for partitioning a collection of data vectors into groups (called clusters). It is a common technique for data mining and exploratory statistical analysis. It has been widely used in many fields, for instance, pattern recognition[1], image processing[2], computer vision[3], bioinformatics[4], and etc. The study of cluster analysis has a long history, going back to anthropology in the 1930s [5].
k-MLE. There are a wide range of clustering methods [1],[6],[7, 8] with new ones still emerging e.g. [9],[10]. There are two types: hard clustering methods and soft clustering methods. In hard clustering the groups are completely separate. In soft clustering the groups may overlap somewhat. So far hard clustering algorithms are all based on similarity measures, typically a distance or pseudo-distance [11]. Soft clustering methods are constructed in three ways: (i) based on likelihood via mixture models [1] fitted with the EM algorithm; (ii) based on fuzzy methods [12]; (iii) based on regularization e.g., with the entropy penalty [13]. In this paper we develop a new and general approach to hard clustering but based on the ‘classification’ likelihood [14] not similarity measures. We prove a general convergence result. And we show that a class of existing similarity/divergence methods [11] form a special case which we rename as k-Bregman. As an application we develop a new method (k-VARs) for clustering autocorrelated time series.
Convergence. For the k-means algorithm with general divergence measures, the definitive work on convergence is the seminal paper of [15]. The much later independent work of [16], while useful, is much less general. We will also discuss the work of [11] which also includes convergence analyses for k-Bregman, which we show are incomplete.
Model Selection. The problem of choosing the number of clusters is a challenging problem which garners ongoing interest. Important methods include BIC [17] and the gap method [18]. Since k-MLE is based on likelihood, it is straightforward to develop a BIC criterion for joint choice of model order and number of clusters. This is illustrated with k-VARs.
Time Series. In many applications the data to be clustered are autocorrelated time series, e.g., stock prices, locations of mobile robots, speech signals, ECG, etc. Statistically efficient clustering of such temporally autocorrelated data demands an integration of time-series structure into the clustering methodology. There is already a significant literature on extending clustering algorithms for time series, e.g., k-means integrated with DTW barycenter averaging (k-DBA) [19], clustering via ARMA mixtures[20], k-Spectral Centroid (k-SC) algorithm [21], k-Shape [22, 23], shapelet based methods [24, 25], deep learning based methods [26], etc. Some methods are only available for univariate time series. However some of these ‘time series’ clustering algorithms ignore the autocorrelation feature and so perform poorly in practice, when it is usually present.
The remainder of the paper is organized as follows. The derivation of k-MLE is presented in section II. In section III a convergence analysis is given. In section IV, k-MLE is applied to clustering autoregressive time series yielding the k-VARs algorithm. The section also covers computation, convergence and model selection. Application to data and simulations are in section V. Section VI contains conclusions. There is one appendix containing some proofs.
II k-MLE: A General Tool for Clustering
II-A Background
Given independently distributed (i.d.) data vectors (of dimension ) , each belonging to one of clusters, , we introduce the label matrix or binary cluster membership array
(1) |
The clustering problem is to estimate from the data matrix . Note that there are only finitely many ’s. We denote the column of by .
Clustering algorithms are often categorised in various ways: partition methods versus hierarchical methods; hard clustering, where each data vector is assigned to only one cluster, versus soft (a.k.a. fuzzy) clustering, where each data vector may belong to more than one cluster; similarity based versus model based. Here we introduce a new division: between deterministic label clustering (DL-C) and stochastic label clustering (SL-C). DL-C treats the membership variables as deterministic, whereas SL-C treats them as random. We note that SL-C always produces soft clustering whereas DL-C can produce either hard clustering or soft clustering.
The classic DL-C method is k-means[1], where clustering is achieved by minimizing the criterion
w.r.t. . Here is a cluster centre and is a distance or similarity measure, i.e. it has two properties: (i) positivity ; (ii) uniqueness iff . Sometimes symmetry is useful: (iii) . k-means uses Euclidean distance but other distance measures are also used, e.g. k-medoids [1].
The classic SL-C method is based on a mixture model [1] in which the are independently distributed with .
Historically almost all hard clustering methods have been based on similarity measures whereas almost all soft clustering methods are based on models. This seems to have fostered the impression that hard clustering methods could not be model based. But we will see that k-MLE is a DL-C method but is model based. This suggests that a deeper division is between DL and SL rather than between similarity based and model based methods. This paper concentrates on DL methods.
II-B k-MLE Formulation
Suppose that are i.d. with those in cluster having density where is -dimensional. Then the joint density or likelihood of the data in cluster is with corresponding log-likelihood
(2) |
where . Denoting , the joint log-likelihood of all the data is then
(3) |
We call this the deterministic label clustering likelihood (DL-CL). The corresponding clustering likelihood for a SL approach would be called a stochastic label clustering likelihood (SL-CL).
We now note two important properties of the DL-CL which now follow immediately.
-
•
Property P1: is linear in .
-
•
Property P2: is separable in .
Separability means that e.g. maximizing over can be done by separately maximizing each cluster-specific likelihood.
The DL-CL depends on binary and analog (i.e. continuous valued) parameters. This leads to a hybrid maximum likelihood estimation (MLE) problem as follows.
Definition 1 (k-MLE problem).
(4) |
where means , .
For this optimization to be well defined we need to introduce some assumptions.
Assumption 1.
-
(a)
is an open subset of with boundary . The closure is bounded (which thus means the closure is compact).
-
(b)
is continuous for .
Note that (a) and (b) ensure that is bounded. Since is a discrete, bounded set, then under Assumption 1, the k-MLE problem has at least one solution.
The k-MLE algorithm solves the hybrid k-MLE problem by cyclic ascent111Cyclic optimization is such a simple idea that it is repeatedly reinvented and renamed as if it was new. (a.k.a. coordinate ascent[27]).
Definition 2 (k-MLE algorithm).
Denote the iteration index by . We assume that initial parameter estimates are available for each cluster.
-
•
-step: given and get . Since is given, we need only maximize, for each , w.r.t. . Denote by , the index for which the log-likelihood is the largest. Then
(5) and we set . If , stop.
-
•
-step: given get .
(6) This just entails solving an MLE problem for each cluster, using the data in that cluster. If , stop.
Notice there is no requirement that the log-density is a similarity measure. So this provides a considerable generalization of DL algorithms such as k-means.
After the second author had developed this method we discovered it is not new. In [14] the DL-CL likelihood appears under the name ‘classification likelihood’. But it is only presented in the context of a multivariate Gaussian density. When the densities are multivariate spherical Gaussian then k-MLE reduces to k-means. No algorithm details are given in [14].
The DL-CL has also been developed (in a different direction) by [28] under the name ‘model-based k-means’. It is derived heuristically from a SL-CL and an algorithm is given without recognizing that it is cyclic ascent. There is no convergence analysis and no tuning parameter selection. [28] attributes the method to [29],[30]. The first reference only deals with two clusters and has no convergence analysis or tuning parameter selection. The second reference fits a mixture model with an EM algorithm and so is not a DL-C method.
II-C k-Bregman
In a seminal piece of work [11] showed ‘there exists a unique Bregman divergence corresponding to every regular exponential family’. Then, using a notion of Bregman information, they developed a hard clustering algorithm based on Bregman divergence. It is not noted in [11] but their algorithm is a coordinate descent algorithm [27]. Further [11] provided a convergence result, which we will see below leaves major questions unanswered.
We now show that the Bregman hard clustering algorithm for exponential families is a special case of k-MLE; we henceforth refer to it as k-Bregman. It is shown in [11] that if is a regular exponential family222For expontial families in general, the same result is remarked in [31]. it has a unique decomposition
(7) |
where is the Bregman divergence, is the conjugate function of , is the expectation parameter corresponding to , and the reminder term does not depend on the parameter and is a normalization factor for the probability density. Thus the log-likelihood for the whole data can be written as
(8) | ||||
(9) |
However, by the property of , the remainder term reduces to and can be dropped. Therefore for an exponential family k-MLE reduces to the k-Bregman similarity based method.
II-D Applications
The possibilities for k-MLE are extensive. Below we develop a new algorithm for clustering vector time series which is not similarity based; we also prove its convergence. We previously developed a scalar version [32] by taking a limit in an AR mixture model. That limiting argument can be extended to the vector case but yields a different algorithm which will be discussed elsewhere.
More generally, k-MLE can handle any data type e.g. multi-modal data, hybrid data with both continuous valued and discrete valued measurements; spatio-temporal data; temporal data sampled on mixed time-scales; multi-subject data and so on.
III k-MLE Convergence
This section develops a convergence analysis of the k-MLE algorithm. Our approach builds on the seminal work of [15]. Although [15] uses dissimilarity where we use log-likelihood, large parts of their development make no use of the dissimilarity properties. We also note the work of [16] (independent of [15]) which, while providing valuable additional insight on k-means, is much less general in scope and results than [15].
Note that [15] deals with minima whereas we deal with maxima. This is a trivial difference. However partly because of our likelihood framework, we need to rework some of the internal proof developments in [15]. Because there is only partial novelty here, we put most of those proofs in the appendix. However we also need non-trivial new results involving new conditions.
Let us note immediately from the construction of the k-MLE algorithm that
This ascent property, i.e. non-decrease of the log-likelihood, is necessary but far from sufficient for convergence of cyclic ascent. If the log-likelihood is bounded then the ascent property ensures, via the monotonicity, that the sequence has a limit point to which it converges. But this criterion convergence does not imply parameter convergence i.e. that have limit points or converge to any of them if they exist. Unfortunately both [28] and [11] mistakenly assert that criterion convergence implies parameter convergence. In fact the major/hard part of a proof of convergence is getting from criterion convergence to parameter convergence as we now show.
The parameter convergence analysis is in three parts. In the first part we show convergence in a finite number of steps to what we (c.f. [15]) call a partial maximum. In the second part further study is made to find conditions under which a partial maximum is a local maximum. In the third part this equivalence is established under MLE uniqueness.
III-A Convergence to Partial Maxima
It turns out to be convenient to consider a ‘relaxed’ or purely analog version of the k-MLE problem in which the membership variables are analog lying in . This is formalised by introducing an analog constraint set.
Definition 3 (analog constraint set).
(10) |
Note that is bounded and is well defined and bounded for . We can thus introduce the:
Definition 4 (concentrated333This is standard statistical and econometric terminology. log-likelihood).
(11) |
Lemma 1 ([15]).
is convex.
Proof.
See Appendix A-A. ∎
Lemma 2 (Theorem 2[15]).
The extreme points of satisfy the binary constraints of the k-MLE problem.
Proof.
Elementary. ∎
We can now introduce the concentrated k-MLE problem:
Definition 5 (k-MLE-c problem).
(12) |
Proposition 3 ([15]).
The k-MLE-c problem and the k-MLE problem have the same solution set.
Proof.
See Appendix A-B. ∎
Definition 6 (Partial Maximum [15]).
A point is called a partial maximum for the k-MLE problem if it satisfies
Theorem 4.
The iterates of the k-MLE algorithm converge to a partial maximum for the k-MLE problem in a finite number of steps.
Proof.
See Appendix A-C. ∎
III-B Convergence to Local Maxima
We introduce the
Definition 7 (maximizing solution set).
(13) |
Since there are finitely many ’s, there are finitely many ’s. Next, recalling Property P2, for fixed , is separable. Thus
(14) |
where the operator “” is the Cartesian product. We say that is a singleton set if it consists of one point. This singleton property turns out to be crucial for convergence.
We need to determine the local maxima of and so need to characterize its gradient. The one-sided directional derivative of in direction (so that is a unit vector) is defined by
Then we have the following characterization and a well-known optimality condition.
Lemma 5 (Lemma 6 [15]).
exists for any at any point , and is given by
(15) |
Proof.
See Appendix A-D. ∎
Proposition 7 ([15]).
Given where and , is a local maximum of k-MLE-c problem if and only if
(16) |
Proof.
See Appendix A-E. ∎
Now we are ready to present our main result on the local maximum of k-MLE problem.
Theorem 8.
Let be a partial maximum for k-MLE problem. Suppose is a singleton set, then is a local maximum of k-MLE problem.
Proof.
See Appendix A-F. ∎
III-C Convergence under MLE Uniqueness
To establish the singleton property of , we need conditions to ensure the uniqueness of MLE. We first introduce the following classic result.
Theorem 9 (Corollary 2.5 [34]).
Suppose that the log-likelihood function obeys the log-likelihood uniqueness conditions:
-
(i)
is twice continuously differentiable for , where is a connected open subset with boundary .
-
(ii)
where .
-
(iii)
the Hessian matrix is negative definite at every stationary point of the likelihood, i.e. .
Then
-
•
There is a unique maximum likelihood estimate .
-
•
The log-likelihood function attains: no other maxima in ; no minima or other stationary points in ; its infimum value on and nowhere else.
Remark. We state the result for the log-likelihood whereas [34] state it for the likelihood. But the statements are equivalent since the Hessians at a stationary point are equivalent.
Proposition 10.
Suppose that each cluster-specific log-likelihood satisfies the conditions of Theorem 9. Then is a singleton set for all .
Proof.
We denote ; the as , ; . In view of the representation of as a Cartesian product, we have only to show that is a singleton for each . This will follow if we show that the MLE of from the joint density is unique. This now follows from Theorem 9. ∎
Remark. In Proposition 10 rather than applying the conditions of Theorem 9 to establish MLE uniqueness, it is sometimes possible to establish MLE uniqueness directly or by other means. We will do this for k-VARs below.
We can now state the main result.
Theorem 11.
Suppose that each cluster-specific likelihood satisfies the uniqueness conditions of Theorem 9. Then, the k-MLE iterates converge to a local maximum point for the k-MLE problem.
Proof.
Remark. As indicated in the previous remark we can alternatively directly show MLE uniqueness when possible.
III-D Convergence of k-Bregman
We could now apply Proposition 10, Theorem 11 to k-Bregman. However it is simpler instead to replace Theorem 11 with yet another result from [34].
Theorem 12 (Theorem 2.6 [34]).
Suppose the log-likelihood function obeys the log-likelihood uniqueness conditions (i),(ii),(iii).
-
(i)
is twice continuously differentiable for a connected open subset with boundary .
-
(ii)
The gradient vanishes for at least one point .
-
(iii)
the Hessian matrix is negative definite at every point .
Then
-
(a)
is concave in .
-
(b)
There is a unique maximum likelihood estimate .
-
(c)
The log-likelihood has no other maxima, minima or stationary points in .
So we have to check the Theorem 12 conditions. In view of (7) we need only check Theorem 12 for exponential family densities. The density of an exponential family in (7) has the equivalent form [11, Section 4]
where is the log partition function
and is convex [11], and is an arbitrary function endowing a measure (see [11, Section 4.1]). This means that the log-density is concave and so the Hessian is negative-semi definite at each point of . But we need negative definiteness and to see what that entails we need to explicitly calculate the Hessian.
We find the score function and
hence Then we have
in which and denote the mean and variance of random variable . Now we can state the Bregman result.
Proposition 13.
Suppose the cluster-specific exponential families each have: at least one point where the gradient of the log-likelihood vanishes; positive definite variance matrices . Then, the k-MLE/k-Bregman iterates converge to a local minimum point for the k-MLE/k-Bregman problem.
Proof.
The proof is the same as that of Theorem 11 we just have to check the conditions of Theorem 12. Following the discussion above, we find that the cluster specific Hessian is
which is negative definite as required for condition (iii) in Theorem 12. Condition (i) is implicit for exponential families. Condition (ii) holds by assumption. The result now follows. ∎
IV Clustering Autocorrelated Time Series with k-VARs
We apply k-MLE to clustering of autocorrelated time series where each contains measurements of an -dimensional time series .
IV-A k-VARs Derivation
We model a time series in the cluster by a Gaussian vector autoregression (VAR), of block order . The conditional likelihood function of the VAR model is
(17) |
where for each
Also is a vector, while for , are the VAR coefficient matrices and the driving noise variance matrix. The set-up now fits in the k-MLE framework.
The classification log-likelihood is given by
(18) |
where denotes the collection of parameters and . Applying the coordinate ascent solver for k-MLE is straightforward, since we have a classic vector or multivariate regression which leads to the following algorithm:
-
•
label update:
(19) (20) -
•
parameter update:
(21) (24) (25) where denotes the number of elements of (i.e. the cardinality of set ), and .
We now consider: initialisation, stopping conditions, and fast matrix computation.
IV-B k-VARs Computation
IV-B1 Initialisation
We take a simple approach. We fit a VAR model to each time series , yielding (). Then we choose at random, one (and ) to initialise the cluster-specific parameters. We call k-VARs with this initialisation ‘k-VARs(rnd)’. In practice we can further repeat this many times and choose the initialisation that delivers the highest value of the likelihood to improve performance. If these initial parameters are all from different clusters then we call the result ‘k-VARs (oracle)’.
IV-B2 Stopping Criteria
We considered two stopping criteria: one based on parameters, the other based on log-likelihoods.
If and are the values given at the current iterate and and are the next, the stopping condition is
(26) |
for all , where is the user-specified tolerance value. One may also consider using the relative errors in (26) according to practical demands.
The log-likelihood stopping rule is
(27) |
where and denote the proposed updates. Compared to the parameter stopping rule, this one is computationally much cheaper since it avoids computing large matrix norms.
IV-B3 Fast Computation
The computation bottleneck of the k-VARs algorithm is (24) and (25). To accelerate computation, we use QR decomposition as now described.
We introduce extended vectors of the general form, and corresponding outer products of the form, . We thus define and .
We now introduce the QR decompositions
where is orthogonal and is upper triangular. We then form which yields a new expression for (24) as follows
(28) |
The QR decomposition also enables cheap computation of individual time series parameters as
(29) |
We next introduce a second set of QR decompositions
(30) |
This yields a compact update of as follows
(31) |
The initial covariance for each time series can be similarly computed by
(32) |
Finally we introduce the lower triangular Cholesky decompositions This yields reliable computation of as follows.
(33) |
where is found by rapidly solving the triangular linear system for . which is very fast since is lower triangular.
IV-B4 Summary of k-VARs and Parallelism
The complete k-VARs algorithm is summarised in Algorithm 1. In addition, it is easy to see that several sections of Algorithm 1 can be parallelised to take advantage of multi-cores, including:
-
•
(line 3) pre-computing , QR decompositions and estimations of .
-
•
(line 6-11) computing , s.
-
•
(line 12-16) updating matrix parameters .
IV-C k-VARs Convergence
Convergence analysis of k-VARs algorithm is provided by applying the general k-MLE results. As indicated in the remarks following Proposition 9 and Theorem 10 we will directly show the solution to the MLE equations is unique.
Note that the model is not from a (vector) exponential family. This is because the exponent can only be represented as an inner product of parameters with data if is known. So we cannot apply (a vector version of) k-Bregman.
Differentiating a cluster-specific likelihood gives the least squares equations shown in (21),(22). From these it is implicit that uniqueness requires that, for each : (i) the inverse of the matrix in (21) exists; (ii) in (21) has full rank. This will follow if (i),(ii) hold for every cluster of size 1. But we are now in a classic situation of requiring uniqueness of the least squares estimate of a VAR and of its noise variance from a single time series record. For cluster size 1, the matrix in (22) is very nearly a block Toeplitz matrix. So long as each time series is generated by (i) a stable VAR with (ii) full rank noise covariance matrix and , then the matrix will have full rank with probability 1 [35]. Further the estimated noise variance matrix will have full rank with probability 1 [35]. So uniqueness of the MLE is established and convergence of the k-VARs algorithm now follows.
IV-D Model Selection for k-VARs
The clustering algorithms, presented in section IV-A and IV-B, require the number of clusters and cluster-specific model orders to be specified. This section develops a Bayesian information criterion (BIC) to resolve the model selection problem. We only discuss the special case with equal ’s. The general case is computationally challenging and may not offer big improvements on large data sets.
For the case of equal ’s, we choose and as the joint minimizer of BIC on a grid as follows.
(34) |
Since the time series are assumed to be sampled independently and the models are independent then the likelihood is given by (18) with all in being set to .
A faster way to find the minimum of BIC is to use cyclic descent.
Here we alternate between the following two steps until a convergence criterion is met.
(i) Given , minimize over ;
(ii) Given , minimize over .
Of course this may get stuck in a local minimum.
But that could be managed by trying a number of random starts.
V Simulation and Application to Real Data
This section has two parts. In part I (subsections A,B,C) we we compare k-VARs with other methods on simulated data. We find that k-VARs considerably outperforms state-of-the-art methods because they ignore the autocorrelation feature. Section C studies the robustness of k-VARs with non-Gaussian driving noise and with varying signal to noise ratios. In part II (section D) we apply k-VARs to real data.
V-A Baseline Methods and Performance Indices
k-VARs is implemented in MATLAB. All simulations are executed on 2.4 GHz Intel Core i9 machines with 32GB RAM and macOS Catalina. The ‘state of the art’ comparator methods are as follows.
- •
- •
- •
- •
To avoid the drawbacks of Rand Index (RI) and the squared version of Normalised Mutual Information () discussed in [40], we used the improved measures, adjusted Rand Index (ARI) and Normalized Information Distance (NID) to evaluate clustering performance
where is the number of pairs that are in the same cluster in both and , is the number of pairs that are in different clusters in both and , is the number of pairs that are in the same cluster in but in different clusters in , is the number of pairs that are in different clusters in but in the same cluster in ; and is the average amount of information in , is the mutual information, given as
in which denotes the number of objects that are common to clusters and , is the number of objects in cluster , and is that in cluster . For both ARI and , values close to 1 indicate high clustering performance.
V-B Comparisons on Simulated Datasets
To evaluate performance in a statistical manner, we randomly generate independent datasets. Due to scalability limits on some of the competing methods, the problem size had to be restricted. Each time-series dataset is generated by simulating a class of VAR models with the following specifications:
-
•
time series dimensions: ;
-
•
model order: ;
-
•
time series length: ;
-
•
number of groups: ;
-
•
number of time series per cluster: .
The parameters of each VAR model are generated randomly and roots scaled to ensure stability. The noise covariance matrix is computed from randomly generated Cholesky factors.
The results in Figure 1, show two versions of k-VARs (defined in the caption). k-VARs (oracle), is superior to the state-of-the-art methods. And k-VARs (rnd) is competitive with them. Note also that most of the other methods require a number of tuning parameters to be chosen. We used default values suggested in the various implementations. k-VARs has no free tuning parameters and is competitive because the other methods ignore the autocorrelation feature.


Next we illustrate BIC. The following set-up applied.
-
•
cluster-specific VAR models: , ;
-
•
time series: , , .
The grid for is . log-BIC is plotted as a heatmap in Figure 2.

We see that BIC attains its minimum at the ground truth . However there is a relatively flat region in the vicinity of the minimum. We found in the simulations that good clustering performance was tolerant to some flatness in BIC near its minimum.
V-C Effect of Non-Gaussian Driving Noise and of SNR
Here we study two things. The ability of k-VARS to deal with non-Gaussian driving noise. And the effect of signal to noise ratio (SNR).
We use t-distributed driving noise with degrees of freedom (dof): . Larger dof is closer to Gaussianity. The setup is otherwise the same as for Figure 1, but with . The results in Figure 3 illustrate the robustness of k-VARs with little loss of performance for low dof.
The next simulation studies the effect of SNR for Gaussian driving noise. We modify a definition of SNR due to [41]. Denote as the zero-lag auto-covariance of a stationary VAR(p) time series whose white noise variance is . Then [41] uses SNR. This has a scaling problem. If we multiply by a diagonal scaling matrix (where is a constant) then this SNR changes. So instead we use VSNR which is unaffected by matrix scaling. The divisor of is due to a second adjustment as follows. In steady state we can write where is the one step ahead predictor and is the one step ahead prediction error and so has variance . Also it is uncorrelated with which thus has variance . In the scalar case the when i.e. ‘signal’ and noise variance are equal this would be taken to yield a SNR of 0 dB. So to reproduce that in the vector case we divide by . The model setup is the same as that for Figure 3, except we fixed . We have to modify to achieve a given SNR. The trick is to scale the roots of the VAR polynomial . The SNR values are [0,5,10,15,20] (dB), The results in Figure 4, show k-VARS to be reliable.




V-D Application to Real Dataset WAFER
We now analyse the chip frabrication ‘WAFER’ data set [42] common in time series clustering studies and available at [43]. Each data set in the database contains the measurements recorded by one sensor during the processing of one wafer by one tool.
The results in Table I, show comparisons to k-DBA, k-Shape, k-GAK, k-SC. The proposed method performs best by all measures. One may recall that the performance of k-VARs is subject to initialization and hence is restricted by local optimality. Unlike the Monte Carlo study, we can now optimise the initialisation. Recalling that k-MLE aims to maximise the mixture likelihood, we firstly perform multiple runs with random initialisations, and then apply a likelihood threshold to evaluate initialisations. The results for k-VARs in Table I show a statistical summary of 20 runs with random initialisations selected by thresholding the log-likelihood at , which value is chosen via 10 preliminary trials. The performance of k-VARs is so consistent that its 1st, 2nd, 3rd quartile are the same, hence only one value is given in Table I.
In addition, our methods provide cluster-specific parameters for further analyses, like model validation, Although the manufacturing processes of WAFER is complex and with higher-order dynamics, the results show that satisfactory clustering does not require precise modelling of each time series ( is used here). Again the reason k-VARs performs well here is that the WAFER data shows strong autocorrelation features which the other methods ignore.
k-DBA | k-Shape | k-GAK | k-SC | k-VARs | |
RI | 0.6280 | 0.5914 | 0.5008 | 0.8050 | 0.8120 |
0.0001 | 0.0001 | 0.0054 | 0.0043 | 0.0463 | |
ARI | 0.0050 | 0.0321 | 0.0026 | -0.0057 | 0.0165 |
1-NID | 0.0001 | 0.0032 | 0.0038 | 0.0011 | 0.0074 |
VI Conclusion
In this paper we have developed a general purpose deterministic label clustering method, k-MLE, that is not based on a distance or divergence measure, but on likelihood. We developed an implementation based on cyclic ascent. We also developed a model selection procedure based on BIC and illustrated it in a special case.
We provided a general theorem on parameter convergence of the cyclic ascent iterates. In so doing we drew attention to major flaws in previous clustering algorithm convergence proofs. They only obtained criterion convergence, mistakenly asserting that it implied parameter convergence whereas proving parameter convergence is much much harder.
We showed that a previous clustering method based on Bregman divergence, which we called k-Bregman, is a special case of k-MLE. Our general convergence result then provided the first valid proof of parameter convergence of k-Bregman.
We illustrated the usefulness of k-MLE by developing a new clustering algorithm, k-VARs, for autocorrelated vector time series. We developed a fast computational procedure and a BIC model selection criterion for simultaneously choosing the number of clusters and VAR order.
We only illustrated BIC for k-VARs, but as indicated in the introduction, it can be constructed for any k-MLE algorithm, since it requires only a likelihood and a parameter count.
We compared k-VARs with state of the art vector time series clustering algorithms and found it greatly outperformed them. This is simply because those algorithms mostly ignore the autocorrelation feature. We also showed robust performance in the presence of heavy-tailed driving noise and also with varying SNR.
Acknowledgments
This work was supported by a Discovery Grant DP180102417 from the Australian Research Council.
Appendix A Proofs for Convergence Analysis
A-A Proof of Lemma 1
Let then we have
as required.
A-B Proof of Proposition 3
Since the concentrated log-likelihood is convex, and since the constraint set is convex, the solution of the k-MLE-c problem is an extreme point of and so lies in . Let be this solution. Let be any that achieves . Then also solves the k-MLE problem. The converse also holds completing the proof.
A-C Proof of Theorem 4
Since are bounded sets, the iterates are a bounded (matrix) sequence and so by the Bolzano-Weierstrass theorem have at least one limit point/matrix. Next we claim that an extreme point of is visited at most once before the algorithm stops. Suppose that this is not true, i.e. for some . In view of the cyclic minimization we must have . If either of the inequalities is strict then we get a contradiction due to the stopping rule and the claim holds. If both inequalities are equalities we similarly have a contradiction since the algorithm must already have stopped. It now follows that, since there are a finite number of extreme points of , the algorithm will reach a partial maximum in a finite number of steps.
A-D Proof of Lemma 5
A-E Proof of Proposition 7
A-F Proof of Theorem 8
Since is a partial maximum, then from the definition . But when is a singleton set this is exactly the local maximum condition of Theorem 7, thereby completing the proof.
References
- [1] R. Duda, P. E. Hart, and D. G. Stork, Pattern classification. New York: Wiley, 2001.
- [2] M. Sonka, V. Hlavac, and R. Boyle, Image processing, analysis, and machine vision. Stamford, CT, USA: Cengage Learning, 2015.
- [3] R. Szeliski, Computer vision : algorithms and applications. London New York: Springer, 2011.
- [4] N. C. Jones and P. A. Pevzner, An introduction to bioinformatics algorithms. Cambridge, MA: MIT Press, 2004.
- [5] H. E. Driver and A. L. Kroeber, Quantitative expression of cultural relationships. University of California Press, 1932, vol. 31, no. 4.
- [6] T. Hastie, R. Tibshirani, and J. Friedman, “The elements of statistical learning: data mining, inference, and prediction, Springer Series in Statistics,” 2009.
- [7] B. Everitt, S. Landau, M. Leese, and D. Stahl, Cluster Analysis. Hoboken: Wiley, 2011.
- [8] C. Hennig, M. Meila, F. Murtagh, and R. Roberto, Handbook of cluster analysis. Chapman & Hall/CRC, 2016.
- [9] K. P. et al., “Convex clustering shrinkage.” in PASCAL: Workshop on Statistics and Optimization of Clustering Workshop, London, UK, 2005.
- [10] G. C. et.al., “Convex clustering: an attractive alternative to hierarchical clustering,” PLoS Comput. Biol., vol. 11, p. e1004228, 2015.
- [11] A. Banerjee, S. Merugu, I. S. Dhillon, and J. Ghosh, “Clustering with Bregman divergences,” Journal of machine learning research, vol. 6, no. Oct, pp. 1705–1749, 2005.
- [12] S. Miyamoto, H. Ichihashi, and K. Honda, Algorithms for fuzzy clustering. Berlin: Springer, 2017.
- [13] N. Karayiannis, “Meca: Maximum entropy clustering algorithm,” in Proc. 1994 IEEE 3rd Int Fuzzy Systems Conf. IEEE, 1994, pp. 630–635.
- [14] J. D. Banfield and A. E. Raftery, “Model-based Gaussian and non-Gaussian clustering,” Biometrics, pp. 803–821, 1993.
- [15] S. Z. Selim and M. A. Ismail, “K-means-type algorithms: A generalized convergence theorem and characterization of local optimality,” IEEE Trans. P.A.M.I., no. 1, pp. 81–87, 1984.
- [16] L. Bottou and Y. Bengio, “Convergence properties of the k-means algorithms,” in NIPS, 1995, pp. 585–592.
- [17] G. Schwarz, “Estimating the dimension of a model,” The annals of statistics, vol. 6, no. 2, pp. 461–464, 1978.
- [18] M. Yan and K. Ye, “Determining the number of clusters using the weighted gap statistic,” Biometrics, vol. 63, pp. 1031–1037, 2007.
- [19] F. Petitjean, A. Ketterlin, and P. Gançarski, “A global averaging method for dynamic time warping, with applications to clustering,” Pattern recognition, vol. 44, no. 3, pp. 678–693, 2011.
- [20] Y. Xiong and D.-Y. Yeung, “Time series clustering with ARMA mixtures,” Pattern Recognition, vol. 37, no. 8, pp. 1675–1689, 2004.
- [21] J. Yang and J. Leskovec, “Patterns of temporal variation in online media,” in Proceedings of the fourth ACM international conference on Web search and data mining, 2011, pp. 177–186.
- [22] J. Paparrizos and L. Gravano, “k-shape: Efficient and accurate clustering of time series,” in Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data, 2015, pp. 1855–1870.
- [23] ——, “Fast and accurate time-series clustering,” ACM Transactions on Database Systems (TODS), vol. 42, no. 2, pp. 1–49, 2017.
- [24] L. Ulanova, N. Begum, and E. Keogh, “Scalable clustering of time series with u-shapelets,” in Proceedings of the 2015 SIAM international conference on data mining. SIAM, 2015, pp. 900–908.
- [25] V. S. Siyou Fotso, E. Mephu Nguifo, and P. Vaslin, “Frobenius correlation based u-shapelets discovery for time series clustering,” Pattern Recognition, vol. 103, jul 2020.
- [26] N. S. Madiraju, S. M. Sadat, D. Fisher, and H. Karimabadi, “Deep temporal clustering: Fully unsupervised learning of time-domain features,” arXiv preprint arXiv:1802.01059, 2018.
- [27] D. Luenberger and Y. Ye, Linear and Nonlinear Programming, 3rd ed. Springer, 2008.
- [28] S. Zhong and J. Ghosh, “A unified framework for model-based clustering,” Jl. Machine Learning Research, vol. 4, pp. 1001–1037, 2003.
- [29] M. Kearns, Y. Mansour, and A. Y. Ng, “An information-theoretic analysis of hard and soft assignment methods for clustering,” in Proc. 13th Conf. Uncertainty in Artificial Intelligence,, 1997, pp. 282–293.
- [30] C. Li and G. Biswas, “Applying the hidden markov model methodology for unsupervised learning of temporal data,” Int. Jl. of Knowledge-based Intelligent Eng. Systems, vol. 6, pp. 152–160, 2002.
- [31] J. Forster and M. K. Warmuth, “Relative expected instantaneous loss bounds,” Journal of Computer and System Sciences, vol. 64, no. 1, pp. 76–102, 2002.
- [32] Z. Yue and V. Solo, “Large-scale time series clustering with k-ars,” in Proc IEEE ICASSP. IEEE, 2020, pp. 6039–6043.
- [33] J. R. Magnus and H. Neudecker, Matrix differential calculus with applications in statistics and econometrics. John Wiley & Sons, 2019.
- [34] T. Mäkeläinen, K. Schmidt, and G. P. H. Styan, “On the Existence and Uniqueness of the Maximum Likelihood Estimate of a Vector-Valued Parameter in Fixed-Size Samples,” Ann. Stat., vol. 9, no. 4, pp. 758–767, jul 1981.
- [35] H. Lütkepohl, New Introduction to Multiple Time Series Analysis. Berlin: Springer, 2005.
- [36] J. Tavenard, R. et.al. C. Holtz, M. “Tslearn, A Machine Learning Toolkit for Time Series Data,” Journal of Machine Learning Research, vol. 21, no. 118, pp. 1–6, 2020.
- [37] I. S. Dhillon, Y. Guan, and B. Kulis, “Kernel k-means: spectral clustering and normalized cuts,” in Proceedings of the tenth ACM SIGKDD Int. Conf. on Knowledge discovery & data mining, 2004, pp. 551–556.
- [38] M. Cuturi, “Fast global alignment kernels,” in Proc. 28th Int. Conf. machine learning (ICML-11), 2011, pp. 929–936.
- [39] J. Leskovec, “Codes: K-Spectral Centroid - Cluster Time Series by Shape,” 2014. [Online]. Available: http://snap.stanford.edu/data/ksc.html
- [40] N. Vinh, J. Epps, and J. Bailey, “Information theoretic measures for clusterings comparison: Variants, properties, normalization and correction for chance,” Jl. MLR, vol. 11, pp. 2837–2854, 2010.
- [41] W. Nicholson, I. W. andJ. Bien, and D. Matteson, “High dimensional forecasting by interpretable vector autoregression,” Jl .Mach. Learn. Res., vol. 21, pp. 1–52, 2020.
- [42] R. T. Olszewski, “Generalized feature extraction for structural pattern recognition in time-series data,” Tech. Rep., 2001.
- [43] M. G. Baydogan, “Multivariate Time Series Classification Datasets,” 2017. [Online]. Available: http://www.mustafabaydogan.com
- [44] L. S. Lasdon, Optimization theory for large systems. New York : London: New York : Macmillan N.Y., 1970.