This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Spectral Clustering with Smooth Tiny Clusters

Hengrui Wang
wanghr1230@pku.edu.cn
Yuanpei College
Peking University
&Yubo Zhang
zhangyubo18@pku.edu.cn
School of electronic engineering and computer science
Peking University &Mingzhi Chen
1800017712@pku.edu.cn
School of electronic engineering and computer science
Peking University &Tong Yang
yangtongemail@gmail.com
School of electronic engineering and computer science
Peking University
Corresponding author: yangtongemail@gmail.com
Abstract

Spectral clustering is one of the most prominent clustering approaches. The distance-based similarity is the most widely used method for spectral clustering. However, people have already noticed that this is not suitable for multi-scale data, as the distance varies a lot for clusters with different densities. State of the art (ROSC and CAST ) addresses this limitation by taking the reachability similarity of objects into account. However, we observe that in real-world scenarios, data in the same cluster tend to present in a smooth manner, and previous algorithms never take this into account.

Based on this observation, we propose a novel clustering algorithm, which considers the smoothness of data for the first time. We first divide objects into a great many tiny clusters. Our key idea is to cluster tiny clusters, whose centers constitute smooth graphs. Theoretical analysis and experimental results show that our clustering algorithm significantly outperforms state of the art. Although in this paper, we singly focus on multi-scale situations, the idea of data smoothness can certainly be extended to any clustering algorithms.

1 Introduction

1.1 Background and Motivation

Spectral clustering is a new and hot topic in recent years [1], and many researchers have focus on improving spectral cluster’s performance. Spectral clustering is a set of algorithms which use graph partitioning methods to cluster objects. These set of algorithms have been widely used in many tasks, such as text mining [2] and network analysis [3, 4, 5] and even medical image analysis [6, 7, 8]. Typically, spectral clustering algorithms use a distance-based similarity matrix to indicate the similarity among objects, and this works well in many cases. But this considers only the feature similarity, which is quite straightforward.

Researchers have found that spectral methods can perform really poor when used on multi-scale data, whose clusters various a lot in sizes and densities. Because the standard of similarity varies a lot, for sparse clusters, distances among objects are always much larger than distances among objects from dense clusters. Typically, existing solutions either scale the similarity matrix or apply the power iteration technique to derive pseudo-eigenvectors with rich cluster separation information. Among all the methods focusing on multi-scale data problem, the ROSC [9] and CAST [10] algorithms are state of the art. ROCS and CAST handle multi-scale data by extracting separation information from pseudo-eigenvectors and tuning the given traditional distance-based similarity matrix to a coefficient matrix, considering reachability similarity.

1.2 Our Proposed Solution

In this paper, we propose a new algorithm, namely Smooth-Clustering. Smooth-Clustering is the first algorithm that considers the smoothness of data for clustering problems. As mentioned above, spectral clustering is actually a graph partition problem, which means that the distribution of objects tend to be smooth, and this is quite consistent with real-world scenarios [11]. If there exists a sudden change among different objects, they belong to different groups with high probability. We call this data smoothness in the rest of our paper.

In fact, data smoothness is quite difficult to define because under extremely small scale, all data will seems discrete. To address this problem, first, we divide objects into a great many tiny clusters. Specifically, we use traditional cluster algorithms to cluster extremely close objects into one group and treated them as a tiny cluster. In the rest of the clustering process, we will use the center of the tiny cluster as an object. In this way, no objects will be extremely close. Second, we cluster tiny clusters, whose centers constitute smooth graphs. Specifically, we add an extra penalty term to embedding smoothness information when getting the coefficient matrix.

In this way, we can take data smoothness into account when dealing with the multi-scale data problem. For objects from different clusters, they may happen to be close or even reachable to each other, under which situation no previous algorithms can work well. However, for objects from different algorithms, their feature distribution will always vary a lot and will be in a discrete manner. When considering data smoothness, this problem can be easily fixed.

In this paper, we present a concrete spectral cluster algorithm, which considers feature similarity, reachability similarity, and data smoothness, focusing on multi-scale data. But we notice that the performance of most of the existing spectral clustering method is always affected by randomness. However, when clustering, we can hardly judge the performance of one clustering by any metrics for the lack of labels. Therefore, our idea of data smoothness can be definitely applied to any spectral cluster algorithms to achieve better performance.

Our main contributions are as follows:

  • This is the first work that considers smoothness of tiny clusters in spectral clustering.

  • We propose a new algorithm embedding smoothness information into spectral clustering algorithms.

  • We derived mathematical proofs to show our algorithm can work well for multi-scale data.

  • We conduct extensive experiments to show the effectiveness of our algorithm by comparing its performance with more than 10 previous clustering algorithms. The experimental results show that our algorithm provides high performance on different datasets. We also conduct experiments on synthetic datasets to show that our algorithm is efficient when data smoothness is related to clustering result. These experimental results will be provided in the full version of this paper.

  • We will show the experimental results of applying the idea of data smoothness to many more clustering algorithms.

2 Related Work

Spectral clustering is a hot and new topic. Assuming we have a set of nn objects 𝒳={x1,,xn}\mathcal{X}=\left\{x_{1},\ldots,x_{n}\right\}, and a similarity matrix SS to indicate the similarity among objects. One of the most widely used similarity matrix is the traditional Gaussian kernel based similarity Sij=exp(xixj22σ2)S_{ij}=\exp\left(-\frac{\left\|x_{i}-x_{j}\right\|^{2}}{2\sigma^{2}}\right). Spectral clustering turns objects and similarity matrix into a weighted graph G=(X,S)G=(X,S), transforming 𝒳\mathcal{X} to the set of vertices and entries of SS to weights of corresponding edges. The clustering problem is then transformed into a graph partition problem. We need to partition the graph GG into different groups and to optimize the quality [12] of the partitioning.

A great number of researches has been conducted to focus on different aspects of spectral clustering. Some studies focus on spectral clustering for different data characteristic [13, 14, 15, 16, 17, 18, 19], on computational efficiency and speed [20, 21, 22, 23, 24, 25, 26, 27], and on the theoretical side [28, 29, 30, 31, 32, 33, 34]. Some researchers even combine deep learning with spectral clustering [35, 36, 37].

Although spectral clustering has achieved great success, spectral clustering’s effectiveness suffering from degradation when dealing with noisy [38] or multi-scale data [39]. To figure this out, a number of methods have been proposed [40, 41, 42, 38, 43]. The state of arts spectral clustering algorithms for multi-scale data are ROSC and CAST. Details of these algorithms will be discussed in the next section.

2.1 Scaling the Similar Matrix

One of the typical methods for handling multi-scale data in spectral clustering is scaling the similar matrix SS [44, 45]. Self-tuning spectral clustering (ZP method) [40] uses the Gaussian kernel based similarity Sij=exp(xixj22σ2)S_{ij}=\exp\left(-\frac{\left\|x_{i}-x_{j}\right\|^{2}}{2\sigma^{2}}\right) to Sij=S_{ij}= exp(xixj2σiσj),\exp\left(-\frac{\left\|x_{i}-x_{j}\right\|^{2}}{\sigma_{i}\sigma_{j}}\right), where xx is the feature vector of an object, and σ\sigma is a scaling parameter which is always difficult to set. Instead of a hyper-parameter σ\sigma, ZP introduces a local scaling parameter σi\sigma_{i} for each object xix_{i}. The local scaling parameter is set as the distance from xix_{i} to its ll -th nearest neighbor (l is a hyper-parameter). Under this design, objects in a sparse cluster will have large σ\sigma and dense cluster’s σ\sigma will be relatively small.

2.2 PI Technique for Obtaining Pseudo-eigenvectors for Cluster Seperation Information

Spectral clustering algorithms always try to find eigenvectors with the richest cluster seperation information [46, 47]. For most spectral clustering algorithms, the smallest eigenvectors are usually treated as the most informative ones. But in fact, researchers have already find that some of these smallest eigenvectors might contain salient noise in the data and might not contain good cluster separation information [48]. Under this circumstances, researches focusing on how to select eigenvectors with the richest cluster separation information have been raised.

The most widely used methods to generate pseudo-eigenvectors is the power iteration (PI) technique. The PI method is used to derive the dominant eigenvector of a matrix and can thus using to enhance spectral clustering.

Given a matrix W,W, PI will first generate a random vector v00v_{0}\neq 0 and iterates as below:

vt+1=WvtWvt1,t0v_{t+1}=\frac{Wv_{t}}{\left\|Wv_{t}\right\|_{1}},\quad t\geq 0

For simplicity, we assume that WW has eigenvalues λ1>λ2>>λn\lambda_{1}>\lambda_{2}>\ldots>\lambda_{n} with corresponding eigenvectors e1,e2,,en.e_{1},e_{2},\ldots,e_{n}. We express v0v_{0} as

v0=c1e1+c2e2++cnenv_{0}=c_{1}e_{1}+c_{2}e_{2}+\ldots+c_{n}e_{n}

with parameters c1,cn.c_{1}\ldots,c_{n}. Let R=i=0t1Wvi1,R=\prod_{i=0}^{t-1}\left\|Wv_{i}\right\|_{1}, we have,

vt=Wtv0/R=(c1Wte1+c2Wte2++cnWten)/Rv_{t}=W^{t}v_{0}/R=\left(c_{1}W^{t}e_{1}+c_{2}W^{t}e_{2}+\ldots+c_{n}W^{t}e_{n}\right)/R
=(c1λ1te1+c2λ2te2++cnλnten)/R=\left(c_{1}\lambda_{1}^{t}e_{1}+c_{2}\lambda_{2}^{t}e_{2}+\ldots+c_{n}\lambda_{n}^{t}e_{n}\right)/R
=c1λ1tR[e1+c2c1(λ2λ1)te2++cnc1(λnλ1)ten]=\frac{c_{1}\lambda_{1}^{t}}{R}\left[e_{1}+\frac{c_{2}}{c_{1}}\left(\frac{\lambda_{2}}{\lambda_{1}}\right)^{t}e_{2}+\ldots+\frac{c_{n}}{c_{1}}\left(\frac{\lambda_{n}}{\lambda_{1}}\right)^{t}e_{n}\right]

We can see that vtv_{t} is thus a linear combination of the eigenvectors. It is obviously that when c10c_{1}\neq 0, vtv_{t} converges to the scaled dominant eigenvector e1e_{1}.

By truncating the iterative PI process, we can obtain an intermediate pseudo-eigenvector, an eigenvectors’ linear combination, and thus containing rich cluster separation information. With PI technique, researchers have proposed the Power Iteration Clustering (PIC) method [48]. However, each object has only one feature value corresponding to the lone pseudo-eigenvector in PIC, and this is far from enough if we have a large clusters number as the cluster collision problem [49] may happen. Under this circumstance, the PIC- kk method has been further proposed to fix this issue, which directly runs PI multiple times to generate more pseudo-eigenvectors for clustering.

The most weakness of the PICk\mathrm{PIC}-k method is that the pseudo-eigenvectors could be similar becuase they are not strictly orthogonal, which will leads to redundancy. In order to reduce this redundancy, Deflation -based Power Iteration Clustering (DPIC) method [50] has been proposed to use Schur complement to derive orthogonal pseudo-eigenvectors. Besides, the PIC based methods always ignore lesser but necessary eigenvectors because dominant eigenvectors are assigned larger weights during the PI process. To figure this out, the Diverse Power Iteration Embedding (DPIE) method [51] has been proposed. In this algorithm, When we generate a new pseudo-eigenvector, previously pseudo-eigenvectors’ information will be removed from the new one.

3 Algorithm

In this section, we will first define TKNN graph and grouping effect. After that, we introduce ROSC and CAST in detail. Finally, we present our algorithm and prove that similar to ROSC and CAST, the matrix ZZ in our algorithm also has grouping effect and it can further take data smoothness into account when clustering. It has been proved in previous work that if the coefficient matrix has grouping effect, it can capture the high correlation among objects.

3.1 Transitive KK Nearest Neighbor (TKNN) Graph

TKNN graph is always used to regularize the coefficient matrix ZZ. The main purpose of TKNN graph is to capture the high correlations between objects. We hope that even for objects belong to the same cluster but located at distant far ends of the cluster, their correlations can still be captured.

Definition 1. (Mutual KK -nearest neighbors) [9] Let NK(x)N_{K}(x) be the set of KK nearest neighbors of an object x.x. Two objects xix_{i} and xjx_{j} are said to be mutual KK -nearest neighbors of each other, denoted by xixj,x_{i}\sim x_{j}, iff xiNK(xj)x_{i}\in N_{K}\left(x_{j}\right) and xjNK(xi)x_{j}\in N_{K}\left(x_{i}\right).

Definition 2. (Reachability) [9] Two objects xix_{i} and xjx_{j} are said to be reachable from each other if there exists a sequence of h2h\geq 2 objects {xi=xa1,,xah=xj}\left\{x_{i}=x_{a_{1}},\ldots,x_{a_{h}}=x_{j}\right\} such that xarxar+1x_{a_{r}}\sim x_{a_{r+1}} for 1r<h1\leq r<h.

Definition 3. (Transitive KK -nearest neighbor (TKNN) graph) [9] Given a set of objects 𝒳={x1,x2,,xn},\mathcal{X}=\left\{x_{1},x_{2},\ldots,x_{n}\right\}, the TKNN graph 𝒢K=\mathcal{G}_{K}= (𝒳,)(\mathcal{X},\mathcal{E}) is an undirected graph where 𝒳\mathcal{X} is the set of vertices and \mathcal{E} is the set of edges. Specifically, the edge (xi,xj)\left(x_{i},x_{j}\right)\in\mathcal{E} iff xix_{i} and xjx_{j} are reachable from each other. We represent the TKNN graph by ann×n\operatorname{an}n\times n reachability matrix 𝒲\mathcal{W} whose (i,j)(i,j) -entry 𝒲ij=1\mathcal{W}_{ij}=1 if (xi,xj)\left(x_{i},x_{j}\right) ;0\in\mathcal{E};0 otherwise.

3.2 Grouping Effect

For an object xp,x_{p}, we use zpz_{p} to denote the pp -th column of matrix ZZ. We know from previous works [52, 53, 54] that if ZZ has grouping effect, then using ZZ for spectral clustering will have an excellent performance.

Definition 4. (Grouping Effect) [9] Given a set of objects X=X= {x1,x2,,xn},\left\{x_{1},x_{2},\ldots,x_{n}\right\}, let wqw_{q} be the qq -th column of 𝒲\mathcal{W}. Further, let xixjx_{i}\rightarrow x_{j} denote the condition: (1) xiTxj1x_{i}^{T}x_{j}\rightarrow 1 and (2) wiwj20.\left\|w_{i}-w_{j}\right\|_{2}\rightarrow 0. A matrix ZZ is said to have grouping effect if

xixj|ZipZjp|0,1pnx_{i}\rightarrow x_{j}\Rightarrow\left|Z_{ip}-Z_{jp}\right|\rightarrow 0,\forall 1\leq p\leq n

3.3 ROSC

The main idea of ROSC is to modify the traditional distance-based similarity matrix SS to a coefficient matrix ZZ and then perform spectral clustering based on ZZ. The similarity between two objects can be treated as how much common characteristics they share. If two objects share a lot common characteristics, they are more likely to be represented by each other. Therefore, it is quite nature to determine a cofficient matrix ZZ by

X=XZX=XZ (1)

To construct Z,Z, ROSC first applies PI multiple times to generate pp pseudo-eigenvectors, which formXp×nX\in\mathcal{R}^{p\times n}. The qq -th column of the matrix XX is regarded as the feature vector xqx_{q} of an object xq.x_{q}. The pseudo-eigenvectors can be treated as low dimensional embeddings of objects, which contains rich separation information. We know from previous works that the generated pp pseudo-eigenvectors are always noisy so that ROSC represents each object by others with

X=XZ+OX=XZ+O (2)

where Op×nO\in\mathbb{R}^{p\times n} is the matrix used to represent the noise in the pseudo-eigenvectors.

ROSC derived the matrix ZZ by optimizing the objective function

minZXXZF2+α1ZF2+α2Z𝒲F2\min_{Z}\|X-XZ\|_{F}^{2}+\alpha_{1}\|Z\|_{F}^{2}+\alpha_{2}\|Z-\mathcal{W}\|_{F}^{2} (3)

where the first term is to reduce the noise OO, the second term is the Frobenius norm of ZZ acting as a penalty term to guarantee the sparsity of ZZ and the third term regularizes ZZ by the TKNN graph, which means if two objects are reachable to each other, it will be treated as similar. α1>0,α20\alpha_{1}>0,\alpha_{2}\geq 0 are two factors using to indicate the relative weights of these three terms. We can easily derive the optimal solution to this problem and this has been proved to have grouping effect so that ROSC can work in multi-scale data:

Z=(XTX+α1I+α2I)1(XTX+α2𝒲)Z^{*}=\left(X^{T}X+\alpha_{1}I+\alpha_{2}I\right)^{-1}\left(X^{T}X+\alpha_{2}\mathcal{W}\right) (4)

3.4 CAST

ROSC only express the high correlation among objects from the same cluster. On the other hand, suppressing the correlation among objects from different clusters is also important for correct clustering. CAST studied methods to consider both factors. The key idea of CAST is applying trace Lasso [55] to regularize the coefficient matrix. Researchers have proved that when regularized by trace lasso, the coefficient matrix exhibits "sparsity". Sparsity is the desired property that entries in the matrix corresponding to inter-cluster object pairs should be 0 or very small, thus suppressing the correlation among objects from different clusters. The optimization problem of CAST is as below:

minz12xXz22+α1XDiag(z)+α22zw22\min_{z}\frac{1}{2}\|x-Xz\|_{2}^{2}+\alpha_{1}\|X\operatorname{Diag}(z)\|_{*}+\frac{\alpha_{2}}{2}\|z-w\|_{2}^{2} (5)

where the second term of equation 5 is:

XDiag(z)=q=1nxq2|zq|=q=1n|zq|=z1\|X\operatorname{Diag}(z)\|_{*}=\sum_{q=1}^{n}\left\|x_{q}\right\|_{2}\left|z_{q}\right|=\sum_{q=1}^{n}\left|z_{q}\right|=\|z\|_{1} (6)

3.5 Our Algorithm

Although ROSC and CAST use coefficient matrix ZZ for spectral clustering, they ignore that in most cases, data always presents in a smooth manner. In other words, some objects may be reachable to each other but with sudden direction changes, in this case they are likely to belong to different clusters.

So in our algorithm, we transform the problem into:

minZXXZF2+α1ZF2+α2Z𝒲F2+α3Z𝒲𝒲+α4𝒲F2\min_{Z}\|X-XZ\|_{F}^{2}+\alpha_{1}\|Z\|_{F}^{2}+\alpha_{2}\|Z-\mathcal{W}\|_{F}^{2}+\alpha_{3}\|Z-\mathcal{W}*\mathcal{W}+\alpha_{4}*\mathcal{W}\|_{F}^{2} (7)

All the notions and the first three terms above are exactly the same with those in ROSC. The forth term is a penalty term on sudden direction changes. 𝒲𝒲\mathcal{W}*\mathcal{W} is the second reachable matrix for objects,indicating the path numbers between objects. Intuitively, if there are many paths between two objects, then they are likely to be smooth paths between objects. As we hope this can work for multi-scale data, so a hyper-parameter α4\alpha_{4} is used here to indicate the standard of "many paths".

We can easily derive the optimal solution, ZZ^{*}, to the optimization problem:

Z=(XTX+α1I+α2I+α3I)1(XTX+(α2α3α4)𝒲+α3𝒲𝒲)Z^{*}=\left(X^{T}X+\alpha_{1}I+\alpha_{2}I+\alpha_{3}I\right)^{-1}\left(X^{T}X+(\alpha_{2}-\alpha_{3}*\alpha_{4})\mathcal{W}+\alpha_{3}\mathcal{W}*\mathcal{W}\right) (8)

LEMMA 1. Given a set of objects 𝒳\mathcal{X}, the matrix Xp×nX\in\mathcal{R}^{p\times n}, whose rows are pseudo-eigenvectors, the reachability matrix 𝒲\mathcal{W}, then we have the the optimal soution ZZ^{*} to Equation 6

Zip=xiT(xpXzp)+(α2α3α4)𝒲ip+α3𝒲𝒲ipα1+α2+α3,1i,pn,Z_{ip}^{*}=\frac{x_{i}^{T}\left(x_{p}-Xz_{p}^{*}\right)+(\alpha_{2}-\alpha_{3}*\alpha_{4})\mathcal{W}_{ip}+\alpha_{3}\mathcal{WW}_{ip}}{\alpha_{1}+\alpha_{2}+\alpha_{3}},\quad\forall 1\leq i,p\leq n, (9)

where 𝒲𝒲=𝒲𝒲\mathcal{WW}=\mathcal{W}*\mathcal{W}

PROOF

For 1pn,1\leq p\leq n, let

J(zp)=xpXzp22+α1zp22+α2zpwp22+α3zpwwp+α4wp22.J\left(z_{p}\right)=\left\|x_{p}-Xz_{p}\right\|_{2}^{2}+\alpha_{1}\left\|z_{p}\right\|_{2}^{2}+\alpha_{2}\left\|z_{p}-w_{p}\right\|_{2}^{2}+\alpha_{3}\left\|z_{p}-ww_{p}+\alpha_{4}*w_{p}\right\|_{2}^{2}. (10)

As ZZ^{*} is the optimal solution to Equation 6,6, we have

JZip|zp=zp=0,1in.\left.\frac{\partial J}{\partial Z_{ip}}\right|_{z_{p}=z_{p}^{*}}=0,\quad\forall 1\leq i\leq n. (11)

Hence

2xiT(xpXzp)+2α1Zip+2α2(Zip𝒲ip)+2α3(Zip𝒲𝒲ip+α4𝒲ip)=0-2x_{i}^{T}\left(x_{p}-Xz_{p}^{*}\right)+2\alpha_{1}Z_{ip}^{*}+2\alpha_{2}\left(Z_{ip}^{*}-\mathcal{W}_{ip}\right)+2\alpha_{3}\left(Z_{ip}^{*}-\mathcal{WW}_{ip}+\alpha_{4}\mathcal{W}_{ip}\right)=0 (12)

, which induces Equation 7.

LEMMA 2. 1i,j,pn\forall 1\leq i,j,p\leq n

|ZipZjp|c2(1r)+|α2α3α4||Wip𝒲jp|+α3|𝒲𝒲ip𝒲𝒲jp|α1+α2+α3\left|Z_{ip}^{*}-Z_{jp}^{*}\right|\leq\frac{c\sqrt{2(1-r)}+\left|\alpha_{2}-\alpha_{3}*\alpha_{4}\right|\left|W_{ip}-\mathcal{W}_{jp}\right|+\alpha_{3}\left|\mathcal{WW}_{ip}-\mathcal{WW}_{jp}\right|}{\alpha_{1}+\alpha_{2}+\alpha_{3}} (13)

where c=1+(α2+α3α4)wp22+α3wwp22c=\sqrt{1+(\alpha_{2}+\alpha_{3}*\alpha_{4})\left\|w_{p}\right\|_{2}^{2}+\alpha_{3}\left\|ww_{p}\right\|_{2}^{2}} and r=xiTxjr=x_{i}^{T}x_{j}

PROOF

From Equation 7, we have

ZipZjp=(xiTxjT)(xpXzp)+(α2α3α4)(𝒲ip𝒲jp)+α3(𝒲𝒲ip𝒲𝒲jp)α1+α2+α3Z_{ip}^{*}-Z_{jp}^{*}=\frac{\left(x_{i}^{T}-x_{j}^{T}\right)\left(x_{p}-Xz_{p}^{*}\right)+(\alpha_{2}-\alpha_{3}*\alpha_{4})\left(\mathcal{W}_{ip}-\mathcal{W}_{jp}\right)+\alpha_{3}\left(\mathcal{WW}_{ip}-\mathcal{WW}_{jp}\right)}{\alpha_{1}+\alpha_{2}+\alpha_{3}} (14)

That implies

|ZipZjp|\displaystyle\left|Z_{ip}^{*}-Z_{jp}^{*}\right| |(xiTxjT)(xpXzp)|+|α2α3α4||𝒲ip𝒲jp|+α3|𝒲𝒲ip𝒲𝒲jp|α1+α2+α3\displaystyle\leq\frac{\left|\left(x_{i}^{T}-x_{j}^{T}\right)\left(x_{p}-Xz_{p}^{*}\right)\right|+\left|\alpha_{2}-\alpha_{3}*\alpha_{4}\right|\left|\mathcal{W}_{ip}-\mathcal{W}_{jp}\right|+\alpha_{3}\left|\mathcal{WW}_{ip}-\mathcal{WW}_{jp}\right|}{\alpha_{1}+\alpha_{2}+\alpha_{3}}
xixj||2xpXzp2+|α2α3α4||Wip𝒲jp|+α3|𝒲𝒲ip𝒲𝒲jp|α1+α2+α3\displaystyle\leq\frac{\|x_{i}-x_{j}||_{2}||x_{p}-Xz_{p}^{*}||_{2}+\left|\alpha_{2}-\alpha_{3}*\alpha_{4}\right|\left|W_{ip}-\mathcal{W}_{jp}\right|+\alpha_{3}\left|\mathcal{WW}_{ip}-\mathcal{WW}_{jp}\right|}{\alpha_{1}+\alpha_{2}+\alpha_{3}}

As the column vectors of XX are normalized vectors (i.e., xqTxq=1x_{q}^{T}x_{q}=1, 1qn),\forall 1\leq q\leq n), we have xixj2=2(1r),\left\|x_{i}-x_{j}\right\|_{2}=\sqrt{2(1-r)}, where r=xiTxjr=x_{i}^{T}x_{j}. As ZZ^{*} is the optimal solution of Equation 5, we have

J(zp)\displaystyle J\left(z_{p}^{*}\right) =xpXzp22+α1zp22+α2zpwp22+α3zpwwp+α4wp22\displaystyle=\left\|x_{p}-Xz_{p}^{*}\right\|_{2}^{2}+\alpha_{1}\left\|z_{p}^{*}\right\|_{2}^{2}+\alpha_{2}\left\|z_{p}^{*}-w_{p}\right\|_{2}^{2}+\alpha_{3}\left\|z_{p}-ww_{p}+\alpha_{4}*w_{p}\right\|_{2}^{2}
J(0)\displaystyle\leq J(0) =xp22+(α2+α3α4)wp22+α3wwp22=1+(α2+α3α4)wp22+α3wwp22\displaystyle=\left\|x_{p}\right\|_{2}^{2}+(\alpha_{2}+\alpha_{3}*\alpha_{4})\left\|w_{p}\right\|_{2}^{2}+\alpha_{3}\left\|ww_{p}\right\|_{2}^{2}=1+(\alpha_{2}+\alpha_{3}*\alpha_{4})\left\|w_{p}\right\|_{2}^{2}+\alpha_{3}\left\|ww_{p}\right\|_{2}^{2}

Hence, xpXzp21+(α2+α3α4)wp22+α3wwp22=c.\left\|x_{p}-Xz_{p}^{*}\right\|_{2}\leq\sqrt{1+(\alpha_{2}+\alpha_{3}*\alpha_{4})\left\|w_{p}\right\|_{2}^{2}+\alpha_{3}\left\|ww_{p}\right\|_{2}^{2}}=c.

Then we get

|ZipZjp|c2(1r)+|α2α3α4||Wip𝒲jp|+α3|𝒲𝒲ip𝒲𝒲jp|α1+α2+α3\left|Z_{ip}^{*}-Z_{jp}^{*}\right|\leq\frac{c\sqrt{2(1-r)}+\left|\alpha_{2}-\alpha_{3}*\alpha_{4}\right|\left|W_{ip}-\mathcal{W}_{jp}\right|+\alpha_{3}\left|\mathcal{WW}_{ip}-\mathcal{WW}_{jp}\right|}{\alpha_{1}+\alpha_{2}+\alpha_{3}}

LEMMA 3. Matrix ZZ^{*} in our algorithm has grouping effect.

PROOF Given two objects xix_{i} and xjx_{j} such that xixj,x_{i}\rightarrow x_{j}, we have (1)xiTxj1x_{i}^{T}x_{j}\rightarrow 1 and (2)wiwj20.(2)\left\|w_{i}-w_{j}\right\|_{2}\rightarrow 0. These imply r=xiTxj1r=x_{i}^{T}x_{j}\rightarrow 1 and |𝒲ip𝒲jp|0\left|\mathcal{W}_{ip}-\mathcal{W}_{jp}\right|\rightarrow 0, what’s more, |𝒲𝒲ip𝒲𝒲jp|wiwj2wp20\left|\mathcal{WW}_{ip}-\mathcal{WW}_{jp}\right|\leq\left\|w_{i}-w_{j}\right\|_{2}*\left\|w_{p}\right\|_{2}\rightarrow 0 Hence, the three terms of the numerator of Equation 11 are close to 0 Therefore, |ZipZjp|0\left|Z_{ip}^{*}-Z_{jp}^{*}\right|\rightarrow 0 and thus ZZ^{*} has grouping effect.

We show that our algorithm also improves the performance of spectral clustering on multi-scale data. While traditional approaches focus on feature similarity and state of the art use integrated similarity of both feature and reachability, our approach further considers data smoothness without destroying the grouping effect. Two distant objects xix_{i} and xjx_{j} belong to a same cluster may not share a strong feature similarity or even strong reachability similarity, but present a sudden change in some elements of the feature vector. This leads to a small rr, and traditional approaches will likely separate them into different clusters. What’s more, ROSC and CAST may also poorly performed here because such a sudden change may still preserve reachability similarity. On the contrary, our algorithm considers the smoothness of the objects’ present manner and thus keeping them in the same cluster.

4 Conclusion

In this paper, we studied the performance of considering data smoothness for spectral clustering on data with various sizes and densities. We reviewed existing spectral methods in handling multi-scale data, such as ROSC and CAST. In particular, we found that both ROSC and CAST are not suitable when data is presented in a significant smooth manner. We thus proposed our algorithm, which uses a penalty term with angle information to take data smoothness into account. We mathematically proved that the matrix ZZ constructed by our algorithm has grouping effect. We conducted extensive experiments to evaluate our algorithm’s performance and compared it against other competitors using both synthetic and real datasets. Our experimental results showed that our algorithm performed very well against its competitors over all the datasets, especially on our synthetic datasets. It is thus robust when applied to multi-scale data of different properties.

References

  • [1] Pavel Kolev and Kurt Mehlhorn. A note on spectral clustering. In ESA, 2016.
  • [2] Inderjit Dhillon. Co-clustering documents and words using bipartite spectral graph partitioning. Proceedings of the Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 05 2001.
  • [3] Xiang Li, Ben Kao, Zhaochun Ren, and Dawei Yin. Spectral clustering in heterogeneous information networks. 04 2019.
  • [4] Brahim Aamer, Hatim Chergui, Nouamane Chergui, K. Tourki, M. Benjillali, C. Verikoukis, and M. Debbah. Self-tuning spectral clustering for adaptive tracking areas design in 5g ultra-dense networks. 2019 IEEE Wireless Communications and Networking Conference (WCNC), pages 1–5, 2019.
  • [5] Austin R. Benson, D. Gleich, and J. Leskovec. Tensor spectral clustering for partitioning higher-order network structures. Proceedings of the … SIAM International Conference on Data Mining. SIAM International Conference on Data Mining, 2015:118–126, 2015.
  • [6] T. Schultz and G. Kindlmann. Open-box spectral clustering: Applications to medical image analysis. IEEE Transactions on Visualization and Computer Graphics, 19:2100–2108, 2013.
  • [7] K. Xia, Xiaoqing Gu, and Yudong Zhang. Oriented grouping-constrained spectral clustering for medical imaging segmentation. Multimedia Systems, 26:27–36, 2019.
  • [8] Chia-Tung Kuo, Peter B. Walker, O. Carmichael, and I. Davidson. Spectral clustering for medical imaging. 2014 IEEE International Conference on Data Mining, pages 887–892, 2014.
  • [9] Xiang Li, Ben Kao, Siqiang Luo, and Martin Ester. Rosc: Robust spectral clustering on multi-scale data. pages 157–166, 04 2018.
  • [10] Xiang Li, Ben Kao, Caihua Shan, Dawei Yin, and Martin Ester. Cast: A correlation-based adaptive spectral clustering algorithm on multi-scale data, 06 2020.
  • [11] A. Ben-Hur, D. Horn, H. Siegelmann, and V. Vapnik. Support vector clustering. Scholarpedia, 3:5187, 2001.
  • [12] J. Shi and J. Malik. Normalized cuts and image segmentation. IEEE Transactions on Pattern Analysis & Machine Intelligence, 22(08):888–905, aug 2000.
  • [13] Ling Huang, Donghui Yan, Michael I. Jordan, and Nina Taft. Spectral clustering with perturbed data. In Proceedings of the 21st International Conference on Neural Information Processing Systems, NIPS’08, page 705–712, Red Hook, NY, USA, 2008. Curran Associates Inc.
  • [14] Ulrike Luxburg, Mikhail Belkin, and Olivier Bousquet. Consistency of spectral clustering. The Annals of Statistics, 36, 05 2008.
  • [15] X. Zhu, C. C. Loy, and S. Gong. Constructing robust affinity graphs for spectral clustering. In 2014 IEEE Conference on Computer Vision and Pattern Recognition, pages 1450–1457, 2014.
  • [16] M. Afzalan and F. Jazizadeh. An automated spectral clustering for multi-scale data. Neurocomputing, 347:94–108, 2019.
  • [17] Yang Zhao, Y. Yuan, F. Nie, and Q. Wang. Spectral clustering based on iterative optimization for large-scale and high-dimensional data. Neurocomputing, 318:227–235, 2018.
  • [18] Y. Li, F. Nie, Heng Huang, and J. Huang. Large-scale multi-view spectral clustering via bipartite graph. In AAAI, 2015.
  • [19] R. Couillet and Florent Benaych-Georges. Kernel spectral clustering of large dimensional data. arXiv: Statistics Theory, 2015.
  • [20] Xinlei Chen and Deng Cai. Large scale spectral clustering with landmark-based representation. In Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence, AAAI’11, page 313–318. AAAI Press, 2011.
  • [21] David Cheng, Ravindran Kannan, Santosh Vempala, and Grant Wang. On a recursive spectral algorithm for clustering from pairwise similarities. 07 2003.
  • [22] Jane Cullum and Ralph Willoughby. Lanczos algorithms for large symmetric eigenvalue computations. vol. i: Theory. Classics in Applied Mathematics, I, 01 2002.
  • [23] I. S. Dhillon, Y. Guan, and B. Kulis. Weighted graph cuts without eigenvectors a multilevel approach. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29(11):1944–1957, 2007.
  • [24] Alex Gittens, Prabhanjan Kambadur, and Christos Boutsidis. Approximate spectral clustering via randomized sketching. 11 2013.
  • [25] Donghui Yan, Ling Huang, and Michael I. Jordan. Fast approximate spectral clustering. In Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’09, page 907–916, New York, NY, USA, 2009. Association for Computing Machinery.
  • [26] Nicolas Tremblay, Gilles Puy, Pierre Borgnat, Rémi Gribonval, and Pierre Vandergheynst. Accelerated spectral clustering using graph filtering of random signals. 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 4094–4098, 2016.
  • [27] Longcheng Zhai, Bin Wu, and Qiuyue Li. Ksc: A fast and simple spectral clustering algorithm. 2019 IEEE Fourth International Conference on Data Science in Cyberspace (DSC), pages 381–387, 2019.
  • [28] S. Lafon and A. B. Lee. Diffusion maps and coarse-graining: a unified framework for dimensionality reduction, graph partitioning, and data set parameterization. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(9):1393–1403, 2006.
  • [29] M Meila and Jue Shi. A random walks view of spectral segmentation. aistats. AI and Statistics (AISTATS), 02 2001.
  • [30] Boaz Nadler, Stephane Lafon, Ronald Coifman, and Ioannis Kevrekidis. Diffusion maps, spectral clustering and eigenfunctions of fokker-planck operators. Adv Neural Inf Process Syst, 18, 07 2005.
  • [31] Matthäus Kleindessner, S. Samadi, P. Awasthi, and J. Morgenstern. Guarantees for spectral clustering with fairness constraints. In ICML, 2019.
  • [32] Y. Zhang and K. Rohe. Understanding regularized spectral clustering via graph conductance. ArXiv, abs/1806.01468, 2018.
  • [33] Yang Wang and Lihua Wu. Beyond low-rank representations: Orthogonal clustering basis reconstruction with optimized graph structure for multi-view spectral clustering. Neural networks : the official journal of the International Neural Network Society, 103:1–8, 2018.
  • [34] T. Bühler and M. Hein. Spectral clustering based on the graph p-laplacian. In ICML ’09, 2009.
  • [35] Uri Shaham, K. Stanton, Haochao Li, B. Nadler, R. Basri, and Y. Kluger. Spectralnet: Spectral clustering using deep neural networks. ArXiv, abs/1801.01587, 2018.
  • [36] X. Yang, Cheng Deng, Feng Zheng, Junchi Yan, and W. Liu. Deep spectral clustering using dual autoencoder network. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4061–4070, 2019.
  • [37] S. Mehrkanoon, C. Alzate, R. Mall, R. Langone, and J. Suykens. Multiclass semisupervised learning based upon kernel spectral clustering. IEEE Transactions on Neural Networks and Learning Systems, 26:720–733, 2015.
  • [38] Z. Li, J. Liu, S. Chen, and X. Tang. Noise robust spectral clustering. In 2007 IEEE 11th International Conference on Computer Vision, pages 1–8, 2007.
  • [39] Boaz Nadler and Meirav Galun. Fundamental limitations of spectral clustering. volume 19, pages 1017–1024, 01 2006.
  • [40] Lihi Zelnik-Manor and Pietro Perona. Self-tuning spectral clustering. volume 17, 01 2004.
  • [41] Aleksandar Bojchevski, Yves Matkovic, and Stephan Günnemann. Robust spectral clustering for noisy data: Modeling sparse corruptions improves latent embeddings. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’17, page 737–746, New York, NY, USA, 2017. Association for Computing Machinery.
  • [42] Carlos D. Correa and Peter Lindstrom. Locally-scaled spectral clustering using empty region graphs. In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’12, page 1330–1338, New York, NY, USA, 2012. Association for Computing Machinery.
  • [43] Christopher R. John, D. Watson, M. R. Barnes, C. Pitzalis, and M. Lewis. Spectrum: fast density-aware spectral clustering for single and multi-omic data. Bioinformatics, 2020.
  • [44] Hongjie Jia, Shifei Ding, and Mingjing Du. Self-tuning p-spectral clustering based on shared nearest neighbors. Cognitive Computation, 7:622–632, 2015.
  • [45] M. Beauchemin. A density-based similarity matrix construction for spectral clustering. Neurocomputing, 151:835–844, 2015.
  • [46] S. Y. Charles J. Alpert. Spectral partitioning: The more eigenvectors, the better. In 32nd Design Automation Conference, pages 195–200, 1995.
  • [47] Wei Ye, Sebastian Goebl, Claudia Plant, and Christian Böhm. Fuse: Full spectral clustering. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’16, page 1985–1994, New York, NY, USA, 2016. Association for Computing Machinery.
  • [48] Frank Lin and William Cohen. Power iteration clustering. pages 655–662, 08 2010.
  • [49] F. Lin. Scalable methods for graph-based unsupervised and semi-supervised learning. 2012.
  • [50] Anh Pham The, N. Thang, L. Vinh, Y. Lee, and S. Lee. Deflation-based power iteration clustering. Applied Intelligence, 39:367–385, 2012.
  • [51] H. Huang, Shinjae Yoo, D. Yu, and H. Qin. Diverse power iteration embeddings and its applications. 2014 IEEE International Conference on Data Mining, pages 200–209, 2014.
  • [52] H. Hu, Z. Lin, J. Feng, and J. Zhou. Smooth representation clustering. In 2014 IEEE Conference on Computer Vision and Pattern Recognition, pages 3834–3841, 2014.
  • [53] C. Lu, J. Feng, Z. Lin, and S. Yan. Correlation adaptive subspace segmentation by trace lasso. In 2013 IEEE International Conference on Computer Vision, pages 1345–1352, 2013.
  • [54] Lu Canyi, Min Hai, Zhong-Qiu Zhao, Lin Zhu, De-Shuang Huang, and Shuicheng Yan. Robust and efficient subspace segmentation via least squares regression. volume 7578, pages 347–360, 10 2012.
  • [55] Edouard Grave, Guillaume Obozinski, and Francis Bach. Trace lasso: a trace norm regularization for correlated designs. Advances in Neural Information Processing Systems, 09 2011.