Agglomerative Mean-Shift Clustering via Query Set Compression Xiao-Tong Yuan

Bao-Gang Hu

Abstract Mean-Shift (MS) is a powerful non-parametric clustering method. Although good accuracy can be achieved, its computational cost is particularly expensive even on moderate data sets. In this paper, for the purpose of algorithm speedup, we develop an agglomerative MS clustering method called Agglo-MS, along with its mode-seeking ability and convergence property analysis. Our method is built upon an iterative query set compression mechanism which is motivated by the quadratic bounding optimization nature of MS. The whole framework can be efficiently implemented in linear running time complexity. Furthermore, we show that the pairwise constraint information can be naturally integrated into our framework to derive a semi-supervised non-parametric clustering method. Extensive experiments on toy and real-world data sets validate the speedup advantage and numerical accuracy of our method, as well as the superiority of its semi-supervised version.

1

Introduction

To find the clusters of a data set sampled from a certain unknown distribution is important in many machine learning and data mining applications. Probability density estimator may represent the distribution of data in a given problem and then the modes may be taken as the representatives of clusters. As a non-parametric method, the kernel density estimation is mostly applied in practice. Given a set of N independent, identically distributed samples X = {x1 , ..., xN } drawn from a population with density function f (x), x ∈ Rd , the kernel density estimator (KDE) with kernel k(·) is defined by (1.1) N N X X 1 fˆk (x) = p(i)p(x|i) = wi k(M 2 (x, xi , Hi )) C i i=1 i=1 where p(i) = wi is the prior weight or mixing proportion PN of point xi (satisfying i=1 wi = 1), M 2 (x, xi , Hi ) = (x − xi )T H−1 i (x − xi ) is the Mahalanobis distance from ∗ Supported in part by NSFC grants #60275025 and MOST of China (No.2007DFC10740). The authors are all with the National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences.



Ran He

x to xi with covariance Hi and Ci is a normalization constant. The Mean-Shift (MS) algorithm discovered by Fukunaga and Hostetler [12] is a powerful optimization algorithm for KDE (1.1). It is expressed as the following fixed-point iteration: (1.2) l+1

x

=

N X

 1 wi g M 2 (xl , xi , Hi ) H−1 i Ci i=1

N X

!−1

 1 wi g M 2 (xl , xi , Hi ) H−1 i xi C i i=1 ′

!

where g(x) = −k (x) and k(·) is also called as the shadow of the profile g(·) [8]. By setting the query set Q = X and the reference set R = X , the Naive-MS clustering is done by grouping points in Q according to the modes they converge to via MS conducted on R. It has a wide range of applications, such as discontinuity preserving image smoothing [8], image segmentation [21] and texture classification [14]. Obviously, the Naive-MS clustering typically requires O(KN 2 ) evaluation (K is the average number of MS iterations per-query). Even for moderate data set, such an exhaustive querying mechanism will lead to severe requirements for computational time and/or storage. The recent years have witnessed a surge of interests of fast MS clustering method [4][5][14][20][21]. One well formulated method is the Gaussian blurring MS (GBMS) [7], which iteratively sharpens the query and reference set by moving each data point according to the Gaussian MS (GMS). Carreira-Perpinan [5] proves that GBMS converges cubically and further provides an accelerated version of GBMS by using an improved iteration stopping criterion. Yang et al. [21] accelerate the speed of GMS to linear running time by using improved fast gaussian transform (IFGT). Although very efficient for large size and high dimensional databases, IFGT-MS is difficult to be generalized for convex kernels other than Gaussian. For image segmentation applications, Carreira-Perpinan [4] has evaluated four kinds of acceleration strategies for GMS, which are based on the spatial structure of images and on the fact that GMS is an Expectation-Maximization (EM) algorithm [6]. The

fastest one is spatial discretisation, which can accelerate GMS by one to two orders of magnitude. However, one limitation of this strategy is that it is typically image analysis specified because it requires spatial dimensions that can be regularly discreted. Based on the dual-tree technique, Wang et al. [20] present the DT-MS for fast data clustering. Since a relative error bound is maintained at each iteration, DT-MS is proved to be stable and accurate. However, the computational cost saving by DT-MS, with comparison to IFGT-MS, is more impressive under the simple Epanechnikov kernel than under the Gaussian kernel. Recently, Yuan and Li [22] point out that convex kernel based MS is equivalent to half-quadratic (HQ) optimization for the KDE function (1.1). The HQ analysis framework of MS implies the fact that MS is a quadratic bounding (QB) optimization for KDE, which is also discovered by Fashing and Tomasi [11]. Motivated by the QB nature of MS, we develop a highly efficient query set compression mechanism to accelerate MS clustering under general kernels. The basic idea is to construct a family of d-dimensional hyperellipsoids to cover the current query set Q, promising that points inside each hyperellipsoid will be at the same basin of the d-dimensional KDE surface and hence will converge to the same mode via MS algorithm. For a given query point, the hyperellipsoid is constructed from a lower QB function of KDE defined at this point. Empirically, we have observed that the number of the covering hyperellipsoids is much smaller (in most cases a reduction of one or two orders of magnitude) than the size of Q. We then take the centers of these hyperellipsoids to form a new query set with size dramatically reduced. Such a query set covering procedure can be iteratively run until convergence is attained. Numerical experiments show that the approximation error introduced in such a query set compression mechanism is always limited and acceptable under proper kernel bandwidth. After each iteration of query set compression, the clustering can be done by grouping the points in current query set according to the hyperellipsoids they are assigned with. This naturally leads to an agglomerative clustering framework. We have analyzed the mode-seeking ability and convergence property of the proposed algorithm. A tight upper bound of the convergent query set size is derived, which promises in theory the speedup performance of our algorithm. Our second contribution is to develop a semisupervised non-parametric clustering algorithm inside the proposed query set compression framework. Until now, the MS algorithm is typically applied in an unsupervised manner. We point out that supervisory information, e.g., the pairwise constraints, can be natu-

rally integrated into the proposed query set compression framework, which leads to a novel constrained non-parametric clustering algorithm. Comparing to the constrained Kmeans (CKmeans) [18] and its variations [2][9][19], our method has the following advantages: 1) There is no potential assumption on data distribution for input data; 2) The number of output clusters is not necessarily to be known beforehand, and 3) No initialization is required to start the algorithm running. At the same time, if necessary, the CKmeans can be used as a postprocessing step to further group the output clusters by our method into desired number of clusters. The experimental evaluation on UCI and realworld data sets validates the superiority of our method. The remainder of the paper is structured as follows. In Section 2, we briefly review the QB nature of MS, which founds the base of this work, from the viewpoint of HQ analysis. In Section 3, we develop an agglomerative MS clustering method based on iterative query set compression and evaluate its numerical performance on toy and real-world clustering problems. In Section 4, by utilizing the pairwise constraint information, we further extend our unsupervised algorithm framework into a semi-supervised version. Finally, we conclude this work in Section 5. 2 Quadratic Bounding Nature of MS The fact that MS is a QB optimization is originally discovered by Fashing and Tomasi in [11], motivated by the relationship between MS and Newton-Raphson method. Actually, the QB nature of MS can be more straightforwardly derived from the HQ optimization viewpoint of MS [22]: When kernel is convex and monotonically decreasing, the MS algorithm can be explained as HQ optimization for fˆk (x). This feature can be shown by using the theory of convex conjugated functions [17] to introduce the following augmented energy function with d + N variables: (2.3) Fˆ (x, p) =

N X i=1

wi

 1 −pi M 2 (x, xi , Hi ) + ϕ(pi ) Ci

where x ∈ Rd is the KDE variable and p = (p1 , ..., pN ) is an N -dimensional vector of auxiliary variables. ϕ(·) is the dual function of k(·) [22]. Fˆ (x, p) is quadratic w.r.t. x while concave w.r.t. p. For a fixed x, the following relationship holds fˆk (x) = sup Fˆ (x, p)

(2.4)

p

and thus max fˆk (x) = max Fˆ (x, p) x

x,p

which means that maximizing fˆk (x) is equivalent to maximizing the augmented function Fˆ (x, p) on the ˆ ) of Fˆ can extended domain. A local maximizer (ˆ x, p be calculated in the following alternate maximization way from a starting point x ˆ0  ′ (2.5) pˆli = −k M 2 (ˆ xl−1 , xi , Hi ) , i = 1, ..., N

For ∀x ∈ HE(ˆ x1i ), due to the quadratic bounding property of ρˆ1 (x), we have fˆk (x) ≥ ρˆ1 (x) ≥ ρˆ1 (ˆ x0i ) = fˆk (ˆ x0i )

which means that any data point xj ∈ Q0 ∩HE(ˆ x1i ) can be alternatively chosen as xˆ1i to increase the current QB function value, rather than maximizing it. Therefore it ! ! −1 is quite hopeful that xj will converge to the same mode N N X X 1 l −1 1 l −1 l as xi does. This is close to the essence behind gener(2.6) xˆ = wi pˆi Hi wi pˆi Hi xi C C i i alized EM, in which the M step is allowed to merely i=1 i=1 increase the expectation obtained in E step rather than which is actually the MS algorithm (1.2). maximizing it. Now we may reasonably claim that At a fixed point x ˆl−1 ∈ Rd , a lower QB function for points inside Q0 ∩ HE(ˆ x1i ) will converge to the same fˆk (x) is given by local maximum and can be safely clustered together by ˆl) (2.7) ρˆl (x) , Fˆ (x, p MS. Sequentially, we can run such a hyperellipsoid construction procedure on the query set Q0 for each point which according to (2.4) satisfies fˆk (ˆ xl−1 ) = ρˆl (ˆ xl−1 ) not yet being covered by any existing hyperellipsoids, and fˆk (x) ≥ ρˆl (x) for ∀x. Obviously the step (2.6) is until the whole set is completely scanned. As a result, equivalent to solving we obtain a family S0 of hyperellipsoids S which covers the l entire query data set, i.e., Q ⊆ x1i ). (2.8) xˆ = arg max ρˆl (x). 0 HE(ˆ x1i )∈S0 HE(ˆ x Actually, S0 can be viewed as a compressor of Q0 . The which indicates that MS is a QB optimization for size of S0 generally depends on data set distribution, KDE. The QB viewpoint of MS motivates the following kernel bandwidth and data scanning order for hyperelacceleration strategy for MS clustering. lipsoid construction. When S0 is relatively dense, we could alternatively apply the well known set covering 3 Agglomerative MS Clustering greedy algorithm [13] to find a subset of S0 , to cover Q0 . The key point of our agglomerative MS algorithm is to A formal description of our query set covering mechaconstruct a family of d-dimensional hyperellipsoids to nism is given in a form of algorithm function in Algocover the current query set Q, promising that points rithm 1. inside each hyperellipsoid will converge to a common 1: Function: Query Set Covering(Query data set Q, Reflocal maximum of KDE via MS. We then use the centers erence data set R) of these hyperellipsoids to form a new query set as the 2: Initialization: S = ∅. compressor of the original one. We may iteratively run 3: for each xi ∈ Q do such a set covering mechanism several times until it 4: if ∃HE(ˆ x1j ) ∈ S such that xi ∈ HE(ˆ x1j ) then converges. After each iteration level, the clustering is 5: Associate xi with HE(ˆ x1j ) . done by grouping the current query points according to 6: else the hyperellipsoids they are associated with, which leads 7: Run one iteration of MS from xi with reference to hierarchical clustering. In the following derivation, set R and construct the HE(ˆ x1i ), as stated in we assume that the covariance is homogenous, i.e., sectionS3.1. 8: S = S {HE(ˆ x1i )} Hi = H. 3.1 Query Set Covering Let’s start with Q0 = R = X . Given a data point xi ∈ Q0 , we initialize x ˆ0i ← xi . It is known from eq. (2.8) that the output x ˆ1i from the first MS iteration is the maximizer of a QB function ρˆ1 (x). Thus we may write down: ′

ρˆ1 (x) = sM 2 (x, xˆ1i , H) + C PN w pˆ1 ′ where s = − i=1 Ci i i < 0 and C is the constant term. Taking x ˆ1i as center, we define the following ddimensional hyperellipsoid: HE(ˆ x1i ) = {x|M 2 (x, xˆ1i , H) ≤ M 2 (ˆ x0i , x ˆ1i , H)}

9: end if 10: end for 11: Return S

Algorithm 1: The Query Set Covering Function

3.2 Iterative Query Set Compression Given currently constructed hyperellipsoid set S0 , we may take the centers of the hyperellipsoids in it to form a compressed query set, i.e., Q1 = {ˆ x1i |HE(ˆ x1i ) ∈ S0 }. The above presented set covering operation can be directly applied on Q1 . After sufficient times of iteration until convergence, we will obtain a sparse enough query set

Q∞ . At each iteration level l, we could group points in Ql according to the hyperellipsoids in Sl they are associated with. Such a query set compression framework naturally leads to an agglomerative clustering of Q0 . A formal description of the proposed method, namely AggloMS, is given in Algorithm 2. We canP see that the computational cost for Agglo-MS is O( L l=1 |Ql |N ), here L is a sufficient large iteration number that promises the convergence. Since |QL | ≤ ... ≤ |Q1 | ≪ N and L ≈ K typically hold (see the algorithm analysis later on), the Agglo-MS can speedup the Naive-MS by a facP tor of KN/( L l=1 |Ql |). Initial Query Set Covering : Let query set Q0 = X and reference set R = X . S0 =Query Set Covering(Q0 , R) Iterative Query Set Compression Phase: Set l = 1 while |Ql | > |Ql−1 | do Let query set Ql = {ˆ x1i |HE(ˆ x1i ) ∈ Sl−1 }. Sl =Query Set Covering(Ql , R) Group the points in query set Ql according to the hyperellipsoids in Sl they are associated with. 10: l←l+1 11: end while 1: 2: 3: 4: 5: 6: 7: 8: 9:

until it converges, and the Q∞ will surly contain at least these m modes returned by the Naive-MS. Based on the above two points, we claim that the Agglo-MS possesses the same mode-seeking ability as that of the Naive-MS if the query set covering is done with some proper data scanning orders. Our numerical observation shows that such a property of the Agglo-MS always holds under most data scanning orders. 3.3.2 Convergence Property of |Ql | We discuss here the convergence property of |Ql |, which is important in algorithm speedup performance analysis. First, due to the fact that |Q1 | ≥ |Q2 | ≥ ..., and |Ql | ≥ 1, we have the following convergence proposition: Proposition 3.1. The sequence {|Ql |, l = 1, 2, ...} generated in the Agglo-MS algorithm converges.

We now focus on the value of |Q∞ |, which is more interested in by end-users. It is known that the NaiveMS clustering is done by grouping the queries according to the modes they separately converge to. In practice, the convergent modes correspond to one common KDE maximum may not be exactly the same due to computational error. We take those modes fall inside a sphere with small enough radius ρ as identical, and nonidentical Algorithm 2: The Agglo-MS Clustering algorithm otherwise. Here ρ serves as the resolution of Naive-MS clustering, i.e., the smaller ρ is, the more clusters will 3.3 Algorithm Study We give in this section some be output. For the clarity of description, we introduce algorithm analysis for the Agglo-MS. The discussion the following concept of ρ-Mode Number for Naive-MS: includes: 1) An insight view of the mode-seeking ability of Agglo-MS, and 2) The convergence property of query set size |Ql |, which is the main concern of algorithm Definition 1. ( ρ-Mode Number) Given a resolution parameter ρ, the number of nonidentical modes lospeedup performance. cated by Naive-MS algorithm on data set X under bandwidth H is referred as ρ-Mode Number, which is denoted 3.3.1 On Mode-Seeking Ability First, we tell that by M (X , H, ρ). the Agglo-MS generally will find no more modes than Naive-MS does. Actually, each point survives in the currently compressed query set Ql is the l-th MS iteration output from some initial point in Q0 . Therefore the final obtained Q∞ is a subset of the convergent modes returned by Naive-MS on Q0 . This also implies that L ≈ K. Second, we show that the Agglo-MS will find no less modes than the Naive-MS does if the query set covering is done with some proper data scanning orders. Suppose that m different modes are located by the Naive-MS on Q0 and let C1 , ..., Cm be the corresponding m clusters. For each cluster Cj , let x∗j = arg maxx∈Cj fˆk (x). It is easy to know that a hyperellipsoid constructed from x∗j will cover no other points in Q0 than x∗j . We may scan from x∗j , j = 1, ..., m, to do set covering on Q0 and compress it into Q1 . Obviously the Naive-MS will find at least these same m modes on Q1 . Such a way of scanning order generation can be repeated on Ql , l ≥ 1,

We aim to establish the relationship between |Ql | and M (X , H, ρ). To do this, we slightly modify the set covering mechanism in algorithm 1 as follows: The if condition in line #4 is revised as: ∃HE(ˆ x1j ) ∈ S W 1 1 such that xi ∈ HE(ˆ xj ) HS(ˆ xj , 4ρ). Here HS(ˆ x1j , 4ρ) 1 is a hypersphere centered at x ˆj with radius 4ρ. We refer algorithm 1 after this modification as the modified query set covering, with which we have the following proposition on an upper bound of |Q∞ |. Proposition 3.2. Given a resolution parameter ρ, the sequence {|Ql |, l = 1, 2, ...} generated in the Agglo-MS algorithm with the modified query set covering mechanism satisfies lim |Ql | 6 M (X , H, ρ)

l→∞

Proof. For each xli ∈ Ql , denote x∞ i be its corresponding MS convergent point. It is known that ∃L > 0 such that if l > L then kxli − x∞ i k < ρ, ∀i. Also, we have that the convergent set {x∞ , i i = 1, ..., |Ql |} is covered by M (X , H, ρ) hyperspheres with radius ρ. Based on these preliminary knowledge, we now prove the proposition with reduction to absurdity. Assume that liml→∞ |Ql | = T > M (X , H, ρ). Due ′ to the discreteness of cardinality we get that ∃L > 0 ′ such that if l > L then |Ql | = T . Now we fix the ′ iteration level l > max(L, L ). Since T > M (X , H, ρ), from the drawer principle we know that there exist at least two points xli and xlj whose convergent points x∞ i and x∞ j will fall inside the same covering hypersphere. Now, we consider the (l + 1)th iteration of query set compression from xli (note that hyperellipsoids will be constructed for each point in query set when it converges). According to the triangle inequality, we ∞ ∞ ∞ have kxl+1 − xlj k < kxl+1 − x∞ i k + kxi − xj k + kxj − i i l xj k < ρ + 2ρ + ρ = 4ρ. Therefore, according to the modified set covering mechanism xlj shall be associated with the hyperellipsoid constructed from xl+1 in the i (l + 1)th level of set compression. This indicates that |Ql+1 | < |Ql | = T , which leads to contradiction. Proposition 3.2 tells us that when ρ is properly chosen and l is large enough, Ql will be rather sparse. Therefore significant speedup of Agglo-MS over Naive-MS can be achieved.

and the query points are clustered in an agglomerative way during this procedure. In the following experiments, we choose the NaiveMS as baseline algorithm and make performance comparison among the Agglo-MS, IFGT-MS and ms1-MS (for image segmentation). In our implementation of the Agglo-MS, we use the modified query set covering function. 3.5 A Walkthrough Example We apply in this section the Agglo-MS on a 2D toy data set (shaped as shown in Figure 1(b), size 1, 821) to illustrate its working mechanism. In this case, isotropic covariance H = σ 2 I is used, hence the hyperellipsoids are degenerated into disks. Let’s start from the point A(−1.5, 0.6) (the green dot in Figure 1(b)) with bandwidth σ = 0.4 to perform the query set covering. After one iteration of MS, point A shifts to point B(−1.29, 0.62) (the red dot in Figure 1(b)). Now, we construct the first disk, − − → taking B as its center and kABk as the radius. In Figure 1(a), we plot the mesh of KDE (in black) and the QB function ρˆ1 (x) (in blue) at A. Obviously, any point lies inside this disk will increase ρˆ1 (x), hence is hopeful to converge to the same mode as that of A via MS. After one pass of scanning, a total of 223 disks are obtained to cover the initial query data set, as illustrated in Figure 1(c). One worrying aspect of our set covering mechanism is the possible intersections between disks from different clusters. To more clearly illustrate such a phenomenon, we have selected a local window (in green) located around one cluster boundary area in Figure 1(c) and magnified it in Figure 1(d). Those disks in blue straddle clusters. This is also where the numerical error of Agglo-MS generates from. The detailed quantitative evaluation on this aspect will be addressed later on. We now enter the iterative query set compression phase. After l = 5 times of iteration a query set with 104 points is obtained, and the corresponding compressed query set and clustering result are shown in Figure 1(e)&1(f). The convergent (l = 60) compressed query set and the corresponding clustering result are shown in Figure 1(g)&1(h). The query set size vs. iteration number curves for different bandwidths are given in Figure 2(a). We evaluate the numerical performance of accelerated MS algorithms using the CPU running time and the following defined ε-error rate (ε-ER): (3.9) ! N ∞,Naive-MS 1 X kˆ x∞,X-MS − x ˆ k i i ε-ER = δ >ε N i=1 kˆ x∞,Naive-MS k i

3.4 Summary of Method To summarize the algorithm development so far, an accelerated MS clustering method based on iterative query set compression is derived. The computational complexity, mode-seeking ability and convergence property of the proposed algorithm are analyzed. In methodology, our Agglo-MS algorithm differs itself from the existing acceleration strategies, including IFGT-MS [21], LSH-MS [14], ms1MS [4] and DT-MS [20]. The IFGT-MS approximates MS calculation at each query point by taking its nearby points as reference set and adopting the fast multipole trick. It is only applicable to Gaussian kernel. The LSH-MS also makes a fast search of the neighborhood around a query point to approximately compute MS iteration. For image segmentation tasks, the ms1-MS smartly stops the current MS iteration if the ascent-path intersects with a previously visited pixel. This is close to the trick used in [8]. The DT-MS achieves speedup by recursively considering a region of query points and a region of reference points, which are represented by nodes of a query tree and a reference tree respectively. The proposed Agglo-MS, on the other hand, iteratively compress the query set until convergence is attained, where δ(x) is the delta function that equals one if

3

3

2

2

1.6

1

1.4

0

1.2

B 1 0

A

−1

−1

1

−2

−2

0.8

−3 −4 −2

−3 −1

0

(a)

1

−4 −2

2

0.6 −1

(b)

0

1

0.4 0.5

2

(c)

3

3

3

2

2

2

2

1

1

1

1

0

0

0

0

−1

−1

−1

−1

−2

−2

−2

−2

−3

−3

−3

−3

−4 −2

−4 −2

−4 −2

0

1

2

−1

0

(e)

1

2

−1

(f)

0

1.5

(d)

3

−1

1

1

−4 −2

2

−1

0

(g)

1

2

(h)

Figure 1: Illustration of iterative query set compression working mechanism on a 2D toy dataset. See text for the detailed description. 1000

35 σ = 0.1 σ = 0.3 σ = 0.8

600 400

IFGT-MS

25

15 10

20

30

40

Iteration Number

50

60

(a) Query Set Size vs. Iteration Number

0.3

IFGT-MS

0.2

0.1

5 10

Agglo-MS

0.15

200 0

0.35

0.25

20

ε-ER

Speedup Ratio

Query Set Size

800

0.4 Agglo-MS

30

0.05

0 0.1

0.2

0.3

0.4

σ

0.5

0.6

0.7

(b) Speedup vs. bandwidth

0.8

0 0.1

0.2

0.3

0.4

σ

0.5

0.6

0.7

0.8

(c) ε-ER vs. bandwidth

Figure 2: Quantitative evaluation curves for the 2D toy problem. boolean variable x is true and equals zero otherwise, while x ˆ∞,X-MS and x ˆ∞,Naive-MS are convergent modes i i returned by X-MS (X stands for ’Agglo’, ’IFGT’ or ’ms1’ in this work) and Naive-MS respectively from an initial query point xi in Q0 . We set the parameters ρ = ε = 10−3 , and choose the Gaussian kernel throughout the experiments in this paper. The algorithm codes 1 are basically written in C++, and run on a P4 Core2 2.4G Hz CPU with 2G RAM. For this example, the quantitative results under bandwidth σ = 0.4 are listed in table 1. The optimal clustering results (|Q∞ | = 5)

by Agglo-MS are achieved under σ ∈ (0.25, 0.54). The speedup vs. bandwidth curves of the Agglo-MS and IFGT-MS over the Naive-MS are given in Figure 2(b), from which we can see that Agglo-MS outperforms IFGT-MS in speedup performance. We also show the ε-ER vs. bandwidth curves of the Agglo-MS and IFGTMS in Figure 2(c), from which we can see that the approximation error are comparable between these two algorithms. Note that when bandwidth is small (σ<0.3) or large (σ>0.6), relatively high ε-ER is introduced in Agglo-MS. This is because small bandwidth tends to lead to over clustering by MS method, which makes 1 The IFGT-MS code is implemented using the Agglo-MS sensitive to the intersection between the IFGT library provided by the authors at disks from different clusters, while too large bandwidth http://www.umiacs.umd.edu/~ vikas/Software/IFGT/IFGT_code.htm. will lead to large covering disks, which also increase The important turnable parameters, e.g., p max -polynomial or- the chance of involving points belonging to different der, K-number of cells and r-ratio of the cut off radius, can be clusters. automatically chosen by the program. We further carefully tune the parameters p Naive-MS.

max

and r to achieve comparable result to

Table 1: CPU running time (in milliseconds) and ε-ER of Agglo-MS, IFGT-MS (p max = 5, r = 2) and Naive-MS on the 2D synthetic dataset. σ = 0.4. Methods |Ql | CPU Time ε-ER l=1 223 40 — Agglo-MS l=5 104 53 — l = 60 5 117 0.0016 IFGT-MS — 975 0 Naive-MS — 1,759 —

3.6 Real-World Experiments We now assess the performance of Agglo-MS on some real-world clustering tasks, including image segmentation and high dimensional data set clustering. 3.6.1 Image Segmentation The most important application of MS clustering algorithm is for unsupervised image segmentation. We test in this part the performance of Agglo-MS in image segmentation tasks. We follow the approach in [21], where each datum is represented by spatial and range features (i, j, L∗ , u∗ , v ∗ ), where (i, j) is the pixel location in the image and (L∗ , u∗ , v ∗ ) is the normalized LUV color feature. Figure 4(b)&4(f) show the results by Agglo-MS on the color image hand. The speedup vs. bandwidth curves of Agglo-MS, IFGT-MS and ms1-MS over Naive-MS on this image are given in Figure 3(a), from which we can see that the Agglo-MS is always much faster than the other two. The ε-ER vs. bandwidth curves of the AggloMS, IFGT-MS and ms1-MS are plotted in Figure 3(b), from which we can see that the approximation error are comparable among these three algorithms. The reasonable segmentations are achieved under the bandwidth interval σ ∈ (0.1, 0.22). We give in Figure 4 the selected segmentation results under bandwidths σ = 0.1, 0.2 by different MS clustering methods. 400

0.1 Agglo-MS

Agglo-MS

IFGT-MS

IFGT-MS

0.08

ms1-MS

300

ms1-MS

250

ε-ER

Speedup Ratio

350

200

0.06 0.04

150 100

0.02

50 0 0.2

0.4

0.6

σ

0.8

1

(a) Speedup vs. bandwidth

0 0.2

0.4

0.6

σ

0.8

1

(b) ε-ER vs. bandwidth

Figure 3: Quantitative evaluation curves for the hand image.

Some images from [21] and the Berkeley segmentation dataset2 are also used for evaluation. Four se-

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)

Figure 4: Segmentation results under σ = 0.1 (top row) and σ = 0.2 (bottom row) on the hand image. For each row, from left to right: Naive-MS, Agglo-MS, IFGT-MS and ms1-MS.

lected groups of segmentation results are given in Figure 5. The quantitative comparison between Agglo-MS, IFGT-MS and ms1-MS on these images are listed in table 2. As expected, the Agglo-MS significantly outperforms the other two in speedup performance on these four test images. The ε-ER achieved by Agglo-MS is overall less than 3% and comparable to the other two. 3.6.2 High Dimensional Cases To evaluate the speedup and numerical performance of the Agglo-MS in high dimensional data clustering tasks, we have applied it on several real-world databases, including the CMU PIE face database3, the MNIST handwritten digit database4 and the TDT2 document database5 . We first briefly describe these data sets and then give the quantitative results on them. CMU PIE Data set The CMU PIE face database contains 68 subjects with 41,368 face images as a whole. We follow [15] to use 170 face images for each individual in our experiment. The total count of points in the data set is 11,554. The size of each cropped gray scale image is 32 × 32 pixels. We do clustering in a subspace embedded by spectral regression (SR) [3], with dimension reduced from 1024 into 67. MNIST Document Data set The MNIST database of handwritten digits has a training set of 60,000 examples,and a test set of 10,000 examples. The digits have been size-normalized and centered in a fixed-size (28 × 28) bilevel image. our clustering is done on the training set embedded by SR, with dimension reduced from 784 to 9. TDT2 Document Data set The TDT2 corpus con3 http://www.ri.cmu.edu/projects/project_418.html 4

http://yann.lecun.com/exdb/mnist/

2 http://www.eecs.berkeley.edu/Research/Projects/CS/vision/bsds/ 5 http://www.nist.gov/speech/tests/tdt/tdt98/index.htm

(a)

(b)

(c)

(d)

(e)

(f)

(g)

(h)

(i)

(j)

(k)

(l)

(m)

(n)

(o)

(p)

Figure 5: Selected image segmentation results. For each image group, from left to right: Naive-MS, Agglo-MS, IFGT-MS and ms1-MS.

Table 2: Quantitative results by Agglo-MS, IFGT-MS (p max = 3, r = 2), ms1-MS and Naive-MS on four test images. Images Sizes σ Agglo-MS CPU IFGT-MS Time (s) ms1-MS Naive-MS Agglo-MS Speedup IFGT-MS ms1-MS Agglo-MS ε-ER IFGT-MS ms1-MS |Q∞ |/M (X , H, ρ)

House 255 × 192 0.1 1.780 5.628 8.015 116.587 65.50 2.64 14.55 0.029 0.021 0.027 5/5

Base Dive 432 × 294 0.1 3.927 23.863 14.607 169.290 43.11 5.15 11.59 0.025 0.054 0.024 7/7

Hawk 481 × 321 0.06 9.313 28.753 12.372 155.829 16.73 2.58 12.60 0.001 0.046 0.001 3/3

Cowboy 481 × 321 0.1 5.888 27.82 18.124 717.521 121.86 3.66 39.59 0.005 0.015 0.005 5/5

sists of 11,201 on-topic documents which are classified into 96 semantic categories. In this experiment, we use Table 3: Performance of Agglo-MS vs. Naive-MS on high dimensional datasets. the top 30 categories, thus leaving us with 9,394 docuDatasets PIE MNIST TDT2 ments. The clustering is also performed in a subspace Dimensions 67 9 29 embedded by SR, with dimension reduced from 36,771 Classes 68 10 30 to 29. Sizes 11,554 60,000 9,394 The provided IFGT-MS code fails on these data σ 15 10 0.8 sets. The quantitative results by Agglo-MS and NaiveCPU Agglo-MS 52.748 13.273 4.789 MS are listed in table 3, from which we can see that Time (s) Naive-MS 505.111 381.646 125.729 the Agglo-MS also significantly improves the clustering Speedup 9.58 28.75 26.25 speed with acceptable approximation error in high ε-ER 0.092 0.065 0.002 dimensional space. |Q∞ |/M (X , H, ρ) 76/76 113/113 22/22 4

Extension for Constrained Clustering

Our Agglo-MS can be naturally extended into a constrained non-parametric clustering framework, with cluster purity much improved. Typically, the constrained clustering focused on the use of background information in the form of instance level must-link and cannot-link constraints. A must-link constraint enforces that two instances must be placed in the same clus-

ter while a cannot-link constraint enforces that two instances must not be placed in the same cluster. The previous work on constrained clustering can be divided into two categories: 1) where the constrains help the algorithm construct a distortion/distance function for optimization [1][16][19] and 2) where the constraints are used as hints to guide the algorithm to search a feasi-

ble solution [9][10][18]. The first type of work makes the assumption that points surrounding a pair of mustlink/cannot-link points should be close to/far from each other, while the second type just requires that the two points be in the same/different clusters. Our constrained Agglo-MS (CAgglo-MS) method falls into the second category. The must-link and cannot-link constraints are used to guide query set covering at each iteration level to promise the feasibility of the output clusters. At the same time, the constraints themselves are automatically updated as query set evolves. To further get the desired number of clusters, the CKmeanslike methods [2][9][18][19]) can be alternatively applied on the final compressed query set with updated constraints. 4.1 Definition of Constraints We consider the problem of clustering the set X under the following types of constraints:

is not allowed to be associated with the same hyperellipsoid during the set covering. This will make sure that, at each iteration level l, the obtained clusters are CLfeasible under current cannot-link constraints Cl . Here, the cannot-link constraint Cl is updated according the following rule: Given a pair of cannot-link points in Cl−1 , the centers of the two hyperellipsoids they separately associated with in Sl are labeled as cannot-link in Cl . Therefore, through simple induction on l we get that the output clusters at each iteration level are CLfeasible under the initial cannot-link constraint C0 = C. The following proposition summarizes the feasibility issues discussed above. Proposition 4.1. The output clusters at each level l by CAgglo-MS algorithm are ML-feasible and CL-feasible under the constraints M and C.

Typically, CAgglo-MS will output more clusters than Agglo-MS does on the same data set. This can also 1. Must-Link Constraints: Each must-link conbe verified by the following knowledge: Since CAgglostraint involves a pair of points xi and xj (i 6= j). MS is CL-feasible, we know from [10] that |Q∞ | is In any feasible clustering, points xi and xj must bounded below by a number kmin , which is upper be in the same cluster. Denote M be the set of bounded by one plus the maximum degree of a node must-link constraint pairs. in an undirected graph Gc = {Vc , Ec } whose vertices set Vc = Q0 and edges set Ec = C. If necessary, as what 2. Cannot-Link Constraints: Each cannot-link is done in the following experiments, we may further constraint also involves a pair of distinct points xi group the output clusters of CAgglo-MS so that more and xj . In any feasible clustering, points xi and xj compact results can be obtained, . must not be in the same cluster. Denote C be the set of cannot-link constraint pairs. 4.3 Performance Evaluation A constrained clustering algorithm is referred to as MLfeasible and CL-feasible if the constraints M and C are 4.3.1 Data Preparation The data sets used in our constrained clustering experiments include 4 data sets satisfied respectively. from the UCI repository6 : Iris, Wine, Sonar and 4.2 The Algorithm and Feasibility The CAgglo- Ionosphere; and two subsets from the MNIST and MS clustering method is formally given in Algorithm TDT2. For MNIST we choose the digits {3, 6, 8, 9} 4. We show through the following detailed construc- to form the MNIST.Multi4 while for TDT2 we chose tion procedure that CAgglo-MS is both ML-feasible and the top ten documents to form the TDT2.Multi10. CL-feasible. As is well known, must-link constraints are We then randomly sample 5% from the MNIST.Multi4 transitive. Therefore, a given collection M of must-link and TDT2.Multi10. Table 4 summarizes the properties constraints can be transformed into an equivalent col- of these data sets. The constraints are generated as lection M = {M1 , ..., Mr } of constraints, by computing follows: for each constraint, we pick out one pair of the transitive closure of W . In the CAgglo-MS, we take data points randomly from the input data sets (the M1 , ..., Mr as r singletons, and together with the com- labels of which are available for evaluation purpose but Sr plemental set X − i=1 Mi to form the initial query unavailable for clustering). If the labels of this pair of set. Such a way of initialization promises that the out- points are the same, then we generate a must-link. If put clusters at each iteration level l is ML-feasible due the labels are different, a cannot-link is generated. The to the fact that Agglo-MS is an agglomerative cluster- amounts of constraints are determined by the size and ing algorithm. To promise the CL-feasibility, we revise difficulty of data set. the algorithm 1 into a cannot-link constrained query set covering mechanism, as is formally described in Al6 http://archive.ics.uci.edu/ml/ gorithm 3. The key point is that each cannot-link pair

4.3.2 Experimental Design In the first group of experiments, we aim to compare the clustering performance of CAgglo-MS and the original Agglo-MS. We 1: Function: CL Query Set Covering(Query data set Q, Reference use the Precision [2] to evaluate the purity of output clusters by CAgglo-MS and Agglo-MS. The reason to data set R, Cannot-link constraints C) 2: Initialization: S = ∅. adopt Precision as measurement is that the number of 3: for each xi ∈ Q do output clusters by Agglo-MS and CAgglo-MS may differ W 4: if ∃HE(ˆ x1j ) ∈ S such that xi ∈ HE(ˆ x1j ) HS(ˆ x1j , 4ρ) from the underlying class numbers. x1j ), C) is false and Violate Cannot Link (xi , HE(ˆ Our second group of experiments is designed to then compare CAgglo-MS with two popular constrained K5: Associate xi with HE(ˆ x1j ). means clustering methods MPCKmeans [2] and CK6: else 7: Run one iteration of MS from xi with reference means [18], as well as the traditional K-means. To outset R and construct the HE(ˆ x1i ), as stated in put the desired number of clusters in our algorithm, we alternatively apply MPCKmeans to further group the sectionS3.1. output clusters from the CAgglo-MS with the updated 8: S = S {HE(ˆ x1i )} CL-Constraints. We refer to such a combined method 9: end if 10: end for as CAgglo-MS-Kmeans. The astute reader might notice 11: Return S that during the query set compression in CAgglo-MS, —————————————————————— the CL-Constraints Cl will be stronger than Cl−1 since 12: Function: the cannot-link property is propagated from point pairs Violate Cannot Link (Data point xi , Hyperellipsoid HE, to cluster pairs. The final obtained CL-Constraint C∞ Cannot-link constraints C) will then always be over strengthened and go against 13: if ∃xj such that (xi , xj ) ∈ C and xj is already associated the compact clustering of Q∞ . To alleviate such an inwith HE then herent drawback, in our implementation, we just carry 14: Return true out the MPCKmeans on the first five Ql (l ≤ 5) with 15: else the corresponding CL-Constrained Cl and pick the one 16: Return false that performs the best as final clustering output. The 17: end if Algorithm 3: Cannot-Link Constrained Query Set F-Measure score [2] is adopt as quantitative measurement for clustering performance evaluation. Covering 4.3.3 Results The Precision curves of the CAggloMS and Agglo-MS on the test data sets are plot in Figure 6, from which we can see that integrating semisupervision information into query set compression does help to significantly improve the purity of the output 1: Constrained Initial Query Set Covering : Sr ′ ′ S 2: Let X = X − i=1 Mi . Set query set Q0 = X {xi ∈ clusters. We illustrate in Figure 7 the evolution of Mi , i = 1, ..., r} and reference set R = X . Set the cannot-link constraints during the running of CAggloMS on the Iris data set with 200 constraints. From constraints C0 = C 3: S0 =CL Query Set Covering(Q0 , R, C0 ) Figure 7(c) we can obviously see that the CL-Constraint 4: Constrained Iterative Set Compression Phase: is much strengthened and at least six clusters are 5: Set l = 1 required on Q150 to fulfil the corresponding C150 . 6: while Convergence is not attained do The F-Measure curves of CAgglo-MS-Kmeans, 7: Let query set Ql = {ˆ x1i |HE(ˆ x1i ) ∈ Sl−1 }. MPCKmeans, CKmeans and the traditional K-means 8: Construct cannot-link constraints Cl so that: on these data sets are given in Figure 8. Our CAgglo∀(xm , xn ) ∈ Cl−1 and xm ∈ HE(ˆ x1i ), xn ∈ HE(ˆ x1j ), MS-Kmeans achieves the best performance on the four 1 1 we have (ˆ xi , x ˆ j ) ∈ Cl . UCI data sets and the TDT2.Multi10 subset. On 9: Sl =CL Query Set Covering(Ql , R, Cl ) 10: Group each point in the query set Ql according to the the MINIST.Multi4 subset, CAgglo-MS-Kmeans and MPCKmeans perform comparably, and both are supehyperellipsoid it is associated with in Sl . rior to CKmeans when constraints are less than 650 11: l ←l+1 12: end while while inferior to it as constraints increase above 650. Algorithm 4: Constrained Agglo-MS Clustering

0.98

1 0.9 0.8

Precision

Precision

0.94

1 0.9

Precision

0.96

0.8

0.92

CAgglo-MS

Agglo-MS

Agglo-MS

0.6

0.6

0.88

CAgglo-MS

0.7

0.7

Agglo-MS

0.9

CAgglo-MS

0.5 0.86 10

50

90

130

170

Number of Constraints

0.5

200

200

400

600

800

Number of Constraints

(a) Iris

1000

200

(b) Wine

600

800

1000

(c) Sonar

1

1

1

0.95

0.98

0.99 0.98

0.9

0.96

0.85

Precision

0.97

Precision

Precision

400

Number of Constraints

0.94

CAgglo-MS 0.8 Agglo-MS

0.96

CAgglo-MS

0.92

0.93

0.88

0.65 200

400

600

800

Number of Constraints

Agglo-MS

0.94

0.9

0.7

CAgglo-MS

0.95

Agglo-MS

0.75

1000

500

1000

1500

Number of Constraints

(d) Ionosphere

2000

0.92

200

(e) MNIST.Multi4

400

600

800

Number of Constraints

1000

(f) TDT2.Multi10

Figure 6: The Precision vs. number of constraints curves of CAgglo-MS and Agglo-MS.

3.5

4.5 4

Dimension 2

4

Dimension 2

4.5

4

Dimension 2

4.5

3.5

3

2.5

3.5

3

2.5

2 4

5

6

7

Dimension 1

2 4

8

(a) Iris: Original CL-constraints

3

2.5

5

6

Dimension 1

7

2 4

8

(b) Iris: CL-constraints for Q20

5

6

Dimension 1

7

8

(c) Iris: CL-constraints for Q150

Figure 7: Evolution of cannot-link constraints during query set compression in CAgglo-MS, on the UCI Iris data set. The points from the three classes are differently colored as dots. The cannot-link constraints are represented by red lines. The black stars represent the current query points. Here, the first two dimensions of input feature are plotted for the sake of visualization. 0.96

1

1

0.9

0.9

0.9

F-Measure

F-Measure

0.92

MPCKmeans CKmeans Kmeans

0.8

0.88

0.8

CAgglo-MS-Kmeans MPCKmeans CKmeans Kmeans

0.7

0.86 0.84

F-Measure

CAgglo-MS-Kmeans 0.94

0.6

CAgglo-MS-Kmeans MPCKmeans

0.7

CKmeans Kmeans

0.6

0.82 0.8

50

100

150

0.5 0

200

Number of Constraints

200

(a) Iris

600

800

1000

0.5 0

CKmeans

CAgglo-MS-Kmeans

0.8

MPCKmeans CKmeas

0.7

Kmeans

0.6

400

600

Number of Constraints

(d) Ionosphere

800

1000

0.5 0

0.8

1000

CAgglo-MS-Kmeans MPCKmeans

0.7

Kmeans

CKmeans Kmeans

0.6

200

F-Measure

MPCKmeans

0.7

800

0.9

F-Measure

CAgglo-MS-Kmeans

600

1

0.9

F-Measure

0.9

400

Number of Constraints

(c) Sonar

1

0.8

200

(b) Wine

1

0.5 0

400

Number of Constraints

0.6

200

400

600

Number of Constraints

800

(e) MNIST.Multi4 (5%)

1000

0.5 0

200

400

600

Number of Constraints

800

1000

(f) TDT2.Multi10 (5%)

Figure 8: The F-Measure vs. number of constraints curves of CAgglo-MS-Kmeans, MPCKmeans, CKmeans and traditional K-means.

Table 4: Datasets used for evaluation of CAgglo-MS Datasets Iris Wine Sonar Ionosphere MNIST.Multi4 (5%) TDT2.Multi10 (5%)

Sizes 150 178 208 351 1190 372

Dimensions 4 13 60 34 9 29

Classes 3 3 2 2 4 10

5 Conclusion In this paper, we report our research progress on algorithm improvement for the popularly used MS nonparametric clustering method. As the first contribution, we develop the so called Agglo-MS algorithm based on a highly efficient hyperellipsoid query set covering mechanism. The mode-seeking ability and convergence property of the Agglo-MS are analyzed. The Agglo-MS is applicable to general convex kernels. Another advantage of Agglo-MS over some existing ones, e.g., IFGTMS and LSH-MS, lies in that it is free of parameter tuning, hence is more flexible when applied in practice. Extensive evaluation on several toy and real-world clustering tasks validates the time efficiency and numerical accuracy of Agglo-MS in both low and high dimensional spaces. The second contribution of this work is to elegantly integrate the pairwise constraint information into the Agglo-MS to develop a semi-supervised non-parametric clustering algorithm, namely CAggloMS. The experimental evaluation on UCI and real-world data sets validates the superiority of CAgglo-MS. We expect that, through combining with dimensionality reduction techniques, one can achieve competitive solutions via Agglo-MS and CAgglo-MS in many clustering tasks.

[6] [7] [8]

[9]

[10]

[11] [12]

[13]

[14]

[15]

[16]

[17] [18]

References [19] [1] S. Basu, M. Bilenko, and R. Mooney. A probabilistic framework for semi-supervised clustering. In Knowledge Discovery and Data Mining. ACM, 2004. [2] M. Bilenko, S. Basu, and R. Mooney. Integrating constraints and metric learning in semi-supervised clustering. In International Conference on Machine Learning, volume 1, pages 81–88, 2004. [3] D. Cai, X. He, and J. Han. Spectral regression for efficient regularized subspace learning. In International Conference on Computer Vision. IEEE, 2007. [4] M. Carreira-Perpinan. Acceleration strategies for gaussian mean-shift image segmentation. In Computer Vision and Pattern Recognition, volume 1, pages 1160– 1167. IEEE, 2006. [5] M. Carreira-Perpinan. Fast nonparametric clustering

[20]

[21]

[22]

with gaussian blurring mean-shift. In International Conference on Machine Learning, pages 153–160, 2006. M. Carreira-Perpinan. Gaussian mean-shift is an em algorithm. IEEE TPAMI, 29(5):767–776, 2007. Y. Cheng. Mean shift, mode seeking, and clustering. IEEE TPAMI, 17(7):790–799, 1995. D. Comaniciu and P. Meer. Mean shift: A robust approach toward feature space analysis. IEEE TPAMI, 24(5):603–619, May 2002. I. Davidson and S. Ravi. Clustering with constraints: Feasibility issues and the k-means algorithm. In International Conference on Data Mining. SIAM, 2005. I. Davidson and S. Ravi. Hierarchical clustering with constraints: theory and practice. In Principles and Practice of Knowledge Discovery in Databases (PKDD), 2005. M. Fashing and C. Tomasi. Mean shift is a bound optimization. IEEE TPAMI, 27:471–474, Mar. 2005. K. Fukunaga and L. Hostetler. The estimation of the gradient of a density function, with application in pattern recognition. IEEE Transactions on Information Theory, 21:32–40, 1975. M. R. Garey and D. S. Johnson. Computers and Intractability, A Guide to the Theory of NPCompleteness. Freeman, New York, 1979. B. Georgescu, I. Shimshoni, and P. Meer. Mean shift based clustering in high dimensions: a texture classification example. In International Conference on Computer Vision, volume 1, pages 456–463. IEEE, 2003. X. He, S. Yan, Y. Hu, P. Niyogi, and H. Zhang. Face recognition using laplacianfaces. IEEE TPAMI, 27(3):1–13, Mar. 2005. D. Klein, S. Kamvar, and C. Manning. From instancelevel constrains to space-level constraints: Making the most of prior knowledge in data clustering. In International Conference on Machine Learning, pages 307–314, 2002. R. Rockfellar. Convex Analysis. Princeton Press, 1970. K. Wagstaff, C. Cardie, S. Rogers, and S. Schroedl. Constrained k-means clustering with background knowledge. In International Conference on Machine Learning, pages 577–584, 2001. F. Wang, T. Li, and C. Zhang. Semi-supervised clustering via matrix factorization. In International Conference on Data Mining. SIAM, 2008. P. Wang, D. Lee, A. Gray, and J. Rehg. Fast mean shift with accurate and stable convergence. In International Conference on Artificial Intelligence and Statistics, volume 2, pages 604–611, 2007. C. Yang, R. Duraiswami, N. A. Gumerov, and L. Davis. Improved fast gauss transform and efficient kernel density estimation. In International Conference on Computer Vision, volume 1, pages 664–671. IEEE, 2003. X. T. Yuan and S. Z. Li. Half quadratic analysis for mean shift: with extension to a sequential data mode-seeking method. In International Conference on Computer Vision. IEEE, 2007.

Agglomerative Mean-Shift Clustering via Query Set ... - CiteSeerX

To find the clusters of a data set sampled from a certain unknown distribution is important in many machine learning and data mining applications. Probability.

2MB Sizes 3 Downloads 299 Views

Recommend Documents

Agglomerative Mean-Shift Clustering via Query Set ... - CiteSeerX
learning and data mining applications. Probability ..... Figure 1: Illustration of iterative query set compression working mechanism on a 2D toy dataset. See text for the ..... MS and LSH-MS, lies in that it is free of parameter tuning, hence is more

Agglomerative Mean-Shift Clustering via Query Set Compression ∗
To find the clusters of a data set sampled from a certain unknown distribution is important in many machine learning and data mining applications. Probability.

Agglomerative Hierarchical Speaker Clustering using ...
news and telephone conversations,” Proc. Fall 2004 Rich Tran- ... [3] Reynolds, D. A. and Rose, R. C., “Robust text-independent speaker identification using ...

Mining Search Engine Query Logs via Suggestion Sampling - CiteSeerX
and notice is given that copying is by permission of the Very Large Data .... The performance of suggestion sampling and mining is measured ...... Estimating the efficiency of backtrack programs. Mathematics of Computation, 29(129):121–136,.

web usage mining using rough agglomerative clustering
is analysis of web log files with web pages sequences. ... structure of web sites based on co-occurrence ... building block of rough set theory is an assumption.

Web page clustering using Query Directed Clustering ...
IJRIT International Journal of Research in Information Technology, Volume 2, ... Ms. Priya S.Yadav1, Ms. Pranali G. Wadighare2,Ms.Sneha L. Pise3 , Ms. ... cluster quality guide, and a new method of improving clusters by ranking the pages by.

Achieving Anonymity via Clustering - Stanford CS Theory
2Department of Computer Science, Stanford University,. Stanford, CA .... year with a maximum of 100 years. In this ... clustering with minimum cluster size r = 2, applied to the table in .... the clause nodes uj have degree at most 3 and cannot be.

Achieving anonymity via clustering - Research at Google
[email protected]; S. Khuller, Computer Science Department, Unversity of Maryland, .... have at least r points.1 Publishing the cluster centers instead of the individual ... with a maximum of 1000 miles, while the attribute age may differ by a

Web Query Recommendation via Sequential ... - Semantic Scholar
wise approaches on large-scale search logs extracted from a commercial search engine. Results show that the sequence-wise approaches significantly outperform the conventional pair-wise ones in terms of prediction accuracy. In particular, our MVMM app

Web Query Recommendation via Sequential ... - Semantic Scholar
Abstract—Web query recommendation has long been con- sidered a key feature of search engines. Building a good Web query recommendation system, however, is very difficult due to the fundamental challenge of predicting users' search intent, especiall

CLUSTERING of TEXTURE FEATURES for CONTENT ... - CiteSeerX
storage devices, scanning, networking, image compression, and desktop ... The typical application areas of such systems are medical image databases, photo ...

Hierarchical Planar Correlation Clustering for Cell ... - CiteSeerX
3 Department of Computer Science. University of ..... (ECCV-12), 2012. Bjoern Andres, Julian Yarkony, B. S. Manjunath, Stephen Kirchhoff, Engin Turetken,.

Query By Committee Made Real - CiteSeerX
algorithms such as the k-nearest neighbors can be optimal. However ... labels. QBC works in an online fashion where each instance is considered only once to de- ... ing algorithm [11] and the perceptron based active learning algorithm [12].

Hierarchical Planar Correlation Clustering for Cell ... - CiteSeerX
3 Department of Computer Science. University of California, Irvine .... technique tries to find the best segmented cells from multiple hierarchical lay- ers. However ...

Using Ephemeral Clustering and Query Logs to ...
Nov 29, 2011 - with Internet access has changed the way to access the In- ternet content ... the whole list to look for the relevant image using network- ing transfer. ..... chose the top fastest rising and top falling queries from 2010 provided by .

Contextual Query Based On Segmentation & Clustering For ... - IJRIT
In a web based learning environment, existing documents and exchanged messages could provide contextual ... Contextual search is provided through query expansion using medical documents .The proposed ..... Acquiring Web. Documents for Supporting Know

Contextual Query Based On Segmentation & Clustering For ... - IJRIT
Abstract. Nowadays internet plays an important role in information retrieval but user does not get the desired results from the search engines. Web search engines have a key role in the discovery of relevant information, but this kind of search is us

Efficient Maximum Margin Clustering via Cutting Plane ...
where the data samples X are mapped into a high ... kernel trick, this mapping could be done implicitly. ...... document vectors in a 8014-dimensional space.

Mining Search Engine Query Logs via Suggestion ... - EE, Technion
Many search engines and other web applications suggest ... In online advertising, advertisers bid for search key- ... Rank of a page x is the (normalized) amount of impressions ... The second challenge was to create an efficient sampler.

Multi-view clustering via spectral partitioning and local ...
(2004), for example, show that exploiting both the textual content of web pages and the anchor text of ..... 1http://www.umiacs.umd.edu/~abhishek/papers.html.

Efficient Maximum Margin Clustering via Cutting Plane ...
Apr 25, 2008 - Automation, Tsinghua Univ. SIAM International Conference on Data Mining ..... Handle large real-world datasets efficiently. Bin Zhao, Fei Wang, ...

Multi-Objective Multi-View Spectral Clustering via Pareto Optimization
of 3D brain images over time of a person at resting state. We can ... (a) An illustration of ideal- ized DMN. 10. 20 .... A tutorial on spectral clustering. Statistics and ...

Mining Search Engine Query Logs via Suggestion ... - EE, Technion
Many search engines and other web applications suggest ... Example applications include com- ..... ple, if the base data set is a query log, the popularity of.

A Framework for Minimal Clustering Modification via ...
1https://sites.google.com/site/ .... We experiment with two real world data sets (social network ... ten found clustering simply puts most instances in one clus-.