Spectral Clustering for Medical Imaging Chia-Tung Kuo∗ , Peter B. Walker† , Owen Carmichael‡ , Ian Davidson∗ ∗ Department of Computer Science, University of California, Davis [email protected], [email protected] † United States Navy [email protected] ‡ Department of Neurology, University of California, Davis [email protected]

Abstract—Spectral clustering is often reported in the literature as successfully being applied to applications from image segmentation to community detection. However, what is not reported is that great time and effort are required to construct a graph Laplacian to achieve these successes. This problem which we call Laplacian construction is critical for the success of spectral clustering but is not well studied by the community. Instead the best Laplacian is typically learnt for each domain from trial and error. This is problematic for areas such as medical imaging since: i) the same images can be segmented in multiple ways depending on the application focus and ii) we don’t wish to construct one Laplacian; rather we wish to create a method to construct a Laplacian for each patient’s scan. In this paper we attempt to automate the process of Laplacian creation with the help of guidance towards the application focus. In most domains creating a basic Laplacian is plausible, so we propose adjusting this given Laplacian by discovering important nodes. We formulate this problem as an integer linear program with a precise geometric interpretation which is globally minimized using large scale solvers such as Gurobi. We show the usefulness on a real world problem in the area of fMRI scan segmentation where methods using standard Laplacians perform poorly.

I. I NTRODUCTION AND MOTIVATION Spectral clustering is extensively used in many areas. Classic success stories include image segmentation [1], community detection [2] and database clustering [3]. However, underlying all these successes is the time consuming trial and error task of constructing an appropriate Laplacian for the spectral clustering algorithm. In particular decisions need to be made such as if the graph is completely connected or should it be a k-nearest neighbor graph or even an -ball graph. Similarly the method of creating the edge weights needs to be determined and it includes decisions such as if the edge weights are thresholded (rounded) to 0 or 1. The possible number of combinations to try is immense and trying them all is prohibitive. The importance of creating the Laplacian cannot be understated. Consider the problem of image segmentation. Though the reported results [1] are impressive, they are much due to the affinity measure which multiplies the square of the pixel location distance by the square of the pixel intensity distance. However, even having the right affinity measure is not sufficient. Creating the

(a) k-Nearest Neighbor Graph

(b)  Ball Graph

Figure 1. Why Laplacian creation is important. The spectral clustering of a classic image using the standard k-nearest neighbor Laplacian (a) and the same clustering using an -ball graph (b). For -ball Laplacians, a foreground and a background can be separated with a 2-way cut but multi-way cuts result in totally uninterpretable mess.

appropriate edge set is also critical; creating the wrong edge set with the right affinity measure can still yield poor results as shown in Figure 1. Creating Laplacians for medical imaging is even more challenging for several reasons. First, how the image is to be segmented depends heavily on the application. If we wish to segment the images to better understand what differentiates Alzheimer effected and healthy individuals, this will require a different Laplacian compared to if we wish to better understand patients whose cognitive abilities on spatial reasoning differ greatly. Secondly, we are not interested in adjusting a single Laplacian; rather since we have many scans and wish to create a Laplacian for each scan, we require a method of creating Laplacians from the underlying data. However, methods such as kernel learning and kernel adjustment are not appropriate. Whereas kernel learning attempts to learn a single kernel over all data points, our problem is more challenging in that we wish to learn a method of creating a Laplacian for each image. Our problem cannot not be addressed using metric learning which reweighs/combines features, either. This is so as the resultant combined features may not yield a legal graph Laplacian that is both positive semi-definite (PSD) and obeys requirements such as transitivity and symmetry. In this paper we explore the simplified problem setting with two useful and realistic assumptions: i) The domain expert already has

We propose an integer linear programming formulation that addresses the problem of adjusting a collection of given graph Laplacians, each of which is known to belong to one of a collection of distinct populations. The output of the optimization is a binary vector indicating the set of important nodes in the graphs that solves a variant of Problem 1. • We show how to use the above measure of node importance in constrained spectral clustering (CSC) work [4] (section III) and our proposed weighted spectral clustering (WSC) scheme. Our methods significantly outperform the regular spectral clustering in differentiating normal and demented individuals (Figure 3). • Importantly, we also show the usefulness of our approach as a predictive device (section V) that is particularly useful in medical imaging settings where only small data sets are available and standard classification methods fail poorly (see Table I). The rest of the paper is organized as follows. We discuss the related work in section II. In section III, we present our formulation and solution for learning a set of important nodes. Section IV describes our novel node weighted spectral clustering to better incorporate the knowledge learnt from our formulation. In section V we demonstrate the usefulness of our proposed approach on real world data set. Finally section VI concludes our paper. •

x   x   x  

o    o        o  

***  

x          Healthy  pa-ents   o          MCI  pa-ents   *          Demented  pa-ents  

Figure 2. A diagrammatic representation of the output of the MaxMin Laplacian Adjustment problem. The diameter of any population is guaranteed to be less than the separation between any two populations. Here the distance function is given by the user and in all our experiments is the absolute-valued entry-wise difference between the Fiedler vectors of the Laplacians.

an initial Laplacian for each image and ii) The domain expert can provide guidance/hints. Problem Statement. In this paper we explore a problem setting where we assume we are given a collection of preliminary graph Laplacians {L1 , L2 , . . . , Ln }, derived from the medical scans, each of which belongs to one of r distinct populations: A1 . . . Ar . Our goal then is to find an adjustment of these Laplacians {L˜1 , L˜2 , . . . , L˜n } that respects the side information that entities in the same population should be similar to each other but different to those in other populations. Formally: Problem 1: Max-Min Laplacian Adjustment

II. R ELATED WORK Input: A collection of preliminary Laplacians {L1 . . . Ln } for n subjects, each belonging to one of r populations A1 . . . Ar . Output: Adjusted Laplacians {L˜1 , L˜2 , . . . , L˜n } such that: d(L˜i , L˜j ) ≤ d(L˜i , L˜k ) ∀i, j ∈ IAx , ∀k ∈ IAy ∀x, y, x 6= y where d(·, ·) is some distance measure between pairs of Laplacians. The above adjustment problem is called Max-Min due to its geometric interpretation. The constraints mean that according to the distance function d the maximum diameter of any population (say A1 ) must be less than or equal to the minimal separation between any two populations (say A1 , A2 ) as shown in Figure 2. To address Problem 1 we focus on learning which nodes are important and which are unimportant in distinguishing populations in the form of an importance weighting vector over the nodes. This weighting vector can be used in a number of ways with spectral clustering, such as encoding must-link and cannotlink constraints in constrained spectral clustering (CSC) [4] or directly encoding the node importance with our novel weighted spectral clustering (WSC) procedure (Algorithm 1). We summarize our contributions below.

Recent work [5] studied, under a statistical framework, the effects of different graph construction methods on the resulting (normalized) min cuts of the graphs and showed that the min cuts can be drastically different for graphs constructed from the same raw data but with different methods (even with different hyperparameters in the same family of methods, such as k-nearest neighbor). The results call for a more automated and domain guidance oriented approach in manipulating/constructing affinities/kernels for spectral clustering. To our knowledge no previous work has studied the problem of graph construction in our proposed setting which learns them from user-defined guidance. Instead there are three bodies of work that explore graph construction: i) Learning low ranked graphs, ii) Learning sparse graphs and iii) Learning balanced graphs. All three areas use guidance that is intrinsic to the data rather than from user-defined guidance. We now summarize related work in each. Learning Low Rank Graphs: The objectives of most of these work [6], [7] is to embody the belief that a low rank graphs prevents overfitting (and thus achieves better generalization) and could better capture more interpretable underlying structure, e.g. block diagonal kernels. [8] formulates the low rank matrix approximation that maintains non-negativity and positive semi-definiteness into a convex optimization and proposes a novel variant of Augmented Lagragian Multiplier method to solve the optimization.

Learning Sparse and Balanced Graphs: [9] learns a sparse and low rank matrix approximation by solving another convex optimization involving `1 norm and trace norm using a proximal descent algorithm. Another method of inducing sparsity is to weigh each node in the graph and then threshold (remove) those edges below a certain weight to achieve a balanced graph. Some other works [10], [11] address semi-supervised learning based on graph learning methods where sparsity is encouraged on the graph. III. F ORMULATION AND NOTATION Suppose we are given a collection of graph Laplacians {L1 , . . . , Ln } of the same number of nodes and each Laplacian Li belongs to one of the two populations l(Li ) ∈ L = {A, B}. The case where there are more than two populations is straightforward, but we present the case with two populations for notational simplicity. Let IA = {i|l(Li ) = A} be the set of indices for which the corresponding entity (i.e. Laplacian) belongs to population A and define IB similarly. Let d(·, ·) be a distance measure between Laplacians for two entities. In our formulation we optimize over a 0/1 indicator vector x whose length is the number of nodes in the graph; xi = 1 means the i-th node is important in making adjusted Laplacians consistent with our guidance and xi = 0 means the node is likely of less importance. We further define 1 = (1, 1, . . . , 1)T to be the vector containing 1 in all entries and tol ≥ 0 is a user-defined constant for tolerance. Now we present the following integer linear program (ILP) for guided node selection. maximize x

1T x

˜i, L ˜ j ) ≤ d(L ˜i, L ˜ k ) + tol ∀i, j ∈ IA , k ∈ IB subject to d(L ˜i, L ˜ j ) ≤ d(L ˜i, L ˜ k ) + tol ∀i, j ∈ IB , k ∈ IA d(L xi ∈ {0, 1} ∀i (1) Notice that in the formulation above the adjusted Lapla˜ i are not defined beforehand; rather we define the cians L distance in the constraints in terms of how we would want the adjusted Laplacians to be measured (see equation (2) for details). The objective (i.e. maximize 1T x) suggests that we leave out fewest nodes (i.e. not important in distinguishing the populations) and thus this formulation seeks the least modification in the number of nodes such that the resulting adjusted Laplacians satisfy the max-min property. The adjusted Laplacians (reweighed in our current work) are constructed using the learnt output x afterwards. In our case, since we want the similarity between two Laplacians to be defined based on their eigenvectors and there is no simple analytic formula to define the eigenvector of an adjusted Laplacian, we solve a variant of the Max-Min problem by defining the constraints using a simpler measure

of distance as follows. Once the distance measure is defined (thus constraints in (1) are defined), the ILP is solved using modern large scale solvers such as Gurobi1 and CPLEX2 . A Distance Function for Spectral Clustering. It is well known that the second principal eigenvector (the Fiedler vector) solves the relaxed min-cut problem. The cut can be viewed as partitioning the image with one side containing the foreground activity (the cognitive network) and the other side background activity. In our application domain, our goal is to have similar min-cuts for scans/images in the same population while scans across different populations shall have different min-cuts. Therefore we compute the mincut vi for each Laplacian Li , define an adjusted cut to be v˜i = vi ◦ x, where ◦ is the Hadamard (entry-wise) product, and define the distance measure as X X ˜i, L ˜j ) = d(L |˜ vi − v˜j | = |vi ◦ x − vj ◦ x| all entries

all entries

= |vi ◦ x − vj ◦ x|T 1 = |vi − vj |T x (2) where | · | denotes entry-wise absolute value. Since x is the indicator variable with each entry being 0 or 1, this measure amounts to selecting a subset of entries in the original cut and measures the absolute difference between a pair of cuts only on those selected entries. Alignment Step. An important step before applying this measure is the sign “alignment” of eigenvectors. Since any nonzero scalar multiple of an eigenvector is also an eigenvector of the same eigenvalue and represents the same cut, one should always first normalize the eigenvectors and “align” the signs. This alignment could be done manually via visual inspection. Alternatively domain knowledge can be used to address the problem. IV. T WO M ETHODS FOR U SING N ODE I MPORTANCE We propose two approaches to utilize the binary vector x learnt from optimization in equation (1). The first approach is to encode the selected node set (xi = 1) into constraints and apply earlier work on constrained spectral clustering (CSC) [4]. Here we interpret an important node as belonging to the network and they should be placed on the same side of the cut. In the second approach we propose and use a novel node weighted spectral clustering where the nodes are re-weighed using the binary indicator vector. A. Utilizing Selected Nodes with CSC CSC incorporates the knowledge of pairwise constraints on the nodes into spectral clustering. The objective is the same as the regular spectral clustering but has an additional constraint v T Qv ≥ α where Q is an n × n matrix with Qi,j = 1 indicating node i and node j should be on the 1 http://www.gurobi.com/ 2 http://www-01.ibm.com/software/commerce/optimization/ cplex-optimizer/

same side of the cut (MUST-LINK) and Qi,j = −1 for having the two nodes on different sides (CANNOT-LINK). The parameter α encodes how well the user requires the resulting cut to satisfy these constraints. Thus a natural way to incorporate the knowledge from x is then to define the constraint matrix Q = xxT + (¬x)(¬x)T and perform CSC on the graphs, where (¬x) is the negation of the binary vector. This effectively adds the constraints that important nodes should be together and unimportant nodes should be together. It is worth noting that, however, CSC [4] may possibly ignore the constraints if they are too restrictive.

theorem states the spectrum of the Laplacian is preserved under our reweighting scheme. Due to the page limit, we refer the interested readers to the appendix for the proof3 . Theorem 1: Graph Spectra Conservation Theorem. Given an edge weighted Laplacian Le and a diagonal node weighting matrix N . Then our weighted Laplacian 1 1 Le,n = N 2 Le N 2 has the same graph spectra properties as Le in that both have the same positive index of inertia and negative index of inertia.

B. Incorporating Node Importance Guidance via Weighted Spectral Clustering Here we present a scheme to adjust spectral clustering for edge and node weighted graphs. In our work we focus on weighting the nodes only to exploit the output from our ILP formulation (1). Note that simply removing the rows and columns for the unimportant nodes will disrupt the spectra of the graphs resulting in invalid Laplacians (e.g. not positive semi-definite). How to incorporate this information into the spectral clustering algorithm without disrupting the spectra of the graph is the challenge here. The key insight is that the weight matrix can be written as the outer-product of the graph written in “edge” format. Let B be an “edge” format matrix for a graph G. Here B is n × m (where m = |E|) and each column in B is an indicator vector stating which nodes are part of the edge. If the i-th column in B represents an edge between nodes j and k then bj,i = 1 and bk,i = −1. We only consider undirected edges and hence it is not important which entry is +1 or -1. Then L = BB T defines the unnormalized Laplacian of the graph. If R is a m × m diagonal matrix and Ri,i > 0 indicates the weight on edge i then the edge weighted Laplacian is given by Le = BRB T . Clearly both L and Le are symmetric positive semi-definite. We now add node weights to the graph derived from the binary solution x to our ILP (1). Given the binary variable x, we define(an n × n diagonal matrix N whose entries 1 if xi = 0 are Ni,i = where w > 1 is a constant w if xi = 1 specifying how much more weight a selected node should carry. Then the node and edge weighted Laplacian matrix is given by 1 1 (3) Le,n = N 2 Le N 2 p This definition scales the rows and columns of Le by Ni,i , respectively. Note that the off-diagonal entries in a Laplacian are non-positive and spectral P P clustering aims to minimize the cut cost, v T Lv = i j vi Li,j vj . Since having two large weights on Ni,i and Nj,j makes Li,j negative and large in magnitude, this strongly encourages the resulting vi and vj to have the same sign (i.e. on the same side of the cut). Overall our proposed node weighted spectral clustering is summarized in Algorithm 1. The following

1 2 3

Input: collection of Laplacians {L1 , . . . , Lk } and weights encoded in the diagonal matrix N Output: cuts {v1 , . . . , vk } for i = 1, . . . , k do 1 1 (i) Le,n ← N 2 Li N 2 ; (i) vi ← Fiedler vector of Le,n (i.e. regular spectral (i) clustering on Le,n ); Algorithm 1: Weighted Spectral Clustering

V. E MPIRICAL S TUDY In this section we aim to empirically evaluate the effectiveness of our proposed approach3 on a medical imaging data set. Such data sets are freely available from the ADNI website4 after appropriate disclosures are signed. We intend to address the followings questions: • How does Laplacian construction fare when separating different populations on the same set of images in comparison with other existing methods? • Can our work be used for prediction? This is another important question since the data sets in medical imaging are typically very small and traditional learning algorithms perform poorly. A. fMRI scan data Data Setup. In this set of experiments we work with a data set of functional magnetic resonance imaging (fMRI) scans taken from 40 subjects at rest, of whom 21 were diagnosed as demented and the other 19 were normal. A further 21 patients were considered mildly cognitive impaired (MCI). These subjects are age controlled to be all elderly and we refer to the normal population simply as elderly or normal hereafter. Each scan consists of snapshots of 3D images (61 × 73 × 61) over 200+ time points. For the purpose of computational cost and ease of presentation we only work with one particular middle slice (of size 61 × 73) of the images though extension to using 3D images shall be straightforward; further as these images are rectangular but had been aligned well, we adopt a common practice of masking out the background with a brain-shaped mask 3

Appendix and codes can be found at http://kuo.idav.ucdavis.edu.

4 http://adni.loni.usc.edu

and our resulting scans now each contain 1730 voxels as shown in Figure 4(a). As well as the clinical diagnosis of dementia, for each person a series of cognitive scores on executive, spatial and semantic reasoning were obtained.

0.91

1.15

0.9 0.9 0.85

1.1

0.89 0.88

0.8

1.05 0.87

B. Analysis Separating Normal from Demented Individuals

0.75

0.86 1 0.85

0.7

Core Task. One common task of interest is to identify the differences in voxel activations and/or regions between the normal subjects and the demented subjects. We start off by constructing an absolute correlation graph for each scan resulting in a 1730 × 1730 affinity matrix where the i, j-th entry is the absolute-valued Pearson’s correlation between the activation time series of the i-th voxel and the j-th voxel. Such correlation analysis on fMRI connectivity is commonly used in the field of neuroimaging [12]. We then construct normalized Laplacians [3] from these absolute correlation graphs as the preliminary Laplacians. We compute the second principal eigenvectors of these preliminary Laplacians to be used in the constraints as shown in equation (2). We align the normalized eigenvectors such that the side with more entries (i.e. foreground) are positive. It is worth noting that in this context it only makes sense to enforce the maximum diameter constraints on the elderly population. Since two patients who were both diagnosed as demented may vary in their deformity, enforcing similarity on all pairs of demented subjects is not appropriate in this context. Additionally we set a non-zero tolerance for the inequality constraints in (1). We test out a range of tolerance values and set the tolerance to 20 in our experiments such that the number of nodes selected (i.e. number of 1’s in the solution x) is roughly the size of a known default mode network (DMN) mask (Figure 4(c)) provided by neurology professionals. We then solved our proposed integer program (1) using the modern large scale optimization software Gurobi producing a set of nodes that differentiated normal people from demented people. Results of Using Important Nodes. We use the important nodes in CSC and WSC as described earlier in section IV, where CSC encodes that the important nodes should be clustered together and WSC encodes that some nodes are more important than others. For WSC we again test a range of values for the weight w and set it to 5. We observe that further increasing w only scales the cut costs for both normal and demented populations without additional distinguishablity. For each method we create box plots as follows. We compute v T Lv (the cut cost) for all normal and demented individuals and plot their range separately. Figure 3 shows the box plots and T-test results for our methods and the regular spectral clustering (SC). We can clearly see from the box plots that CSC and WSC are much better in terms of distinguishing the two populations in cut costs and such differences are reflected in the corresponding small p-values in their T-tests (at 95% confidence level). The regular SC using the unadjusted Laplacian, on the other

0.84

0.95

0.65 0.83 0.9 Elderly

Elderly

Demented

(a) Regular SC

Demented

Elderly

(b) Constrained SC

Demented t statistic p value Elderly -1.6016 0.1175 (d) T-test for Regular SC

Demented

(c) Weighted SC

Demented t statistic p value Elderly -3.1697 0.0030 (e) T-test for Constrained SC

Demented t statistic p value Elderly -3.483 0.0013 (f) T-test for Weighted SC Figure 3. Box plots and T-test results of the min-cut costs grouped by populations using regular spectral clustering (SC), constrained spectral clustering (CSC) and weighted spectral clustering (WSC). Note the y-axes are not comparable across different plots since the scales of the weights are different. 0

0

10

10

20

20

30

30

40

40

50

50

60

60 0

20

40 nz = 1730

60

(a) Preprocessing mask

0

20

40 nz = 423

60

(b) Selected node set

(c) DMN

Figure 4. (a) is the voxels in the brain used as preprocessing mask, (b) the region of the selected nodes and (c) is the idealized default mode network (DMN) typically found in young healthy people at rest.

hand, is unable to draw significant conclusion between the two populations of cut costs. It is important to note that all three techniques were given the exact same preliminary Laplacians. C. Prediction The results we have shown in the previous section are useful for showing the benefits of our method in an exploratory setting. Here we show how our weighted spectral clustering method can be used to make predictions. This is a very important but challenging task. Due to the great cost of collecting data, most medical imaging data sets consist of no more than 100 scans that may have taken several years to collect. As per Figure 3 (right) the mean cut cost of the normal people is approximately 0.975 and the demented people 1.025. This leads to the simple classification rule as follows: if the weighted cut cost is greater than 1 then the person is classified as demented. For the data the algorithm was built on we obtain an accuracy of 15/19 for the normal patients and 17/21 for

Learning Algorithm SVM - Gaussian Kernel SVM - Polynomial Kernel Decision Tree Boosted Decision Trees Bagged Decision Trees

Prediction Accuracy 0.56 0.45 0.34 0.43 0.54

Table I T HE ACCURACY OF VARIETY OF METHODS AT PREDICTING WHETHER MCI PEOPLE WILL EVENTUALLY BECOME DEMENTED . A LL MCI PEOPLE IN OUR STUDY WILL EVENTUALLY BECOME DEMENTED AND OUR METHOD PREDICTS APPROXIMATELY 90% ACCURACY. O UR TRAINING DATA HAS 19 INSTANCES LABELED AS NORMAL AND 21 INSTANCES LABELED AS DEMENTED .

the demented individuals. A far more challenging prediction problem is to predict mildly cognitive impaired (MCI) people all of which will eventually become demented. For these people we expect them to be predicted as demented. For the 21 MCI people in our test set 19 were correctly predicted as demented. We of course do not expect such a high classification accuracy for larger scale studies. As a comparison we tried a variety of discriminative classifiers since generative classifiers are expected to perform poorly on such problems given the small training set size and lack of appropriate priors. We found that the Weka implementations [13] of SVM, decision trees, and neural networks perform no better than guessing the majority label of demented and often worse. Results are shown in Table I. The challenge here is that there are only 40 training instances with 1700+ features. Preprocessing the data using dimension reduction techniques such as PCA produces even worse results since estimating the variance from such a small sample is misleading. It is important to note, that the transductive spectral methods [14] cannot be used in this setting since the spectral clustering is on a single image not on a collection of images together. VI. C ONCLUSION In this paper we explore the under-studied problem of adjusting Laplacians under the guidance of knowing that they belong to several underlying populations. We present an integer linear program formulation for selecting a set of nodes in the graphs that enforce the property that, according to a user defined distance function, any pair of graphs from the same population are closer together than any pair of graphs from different populations. We utilize the selected set of nodes in two ways; we explore encoding these nodes into constraints to be used in constrained spectral clustering algorithm [4]. We also propose a weighted spectral clustering algorithm that can incorporate this knowledge by reweighing the nodes. Empirical evaluation on a real world fMRI imaging data set demonstrates that our methods are able to distinguish normal and demented subjects (at the 95% confidence level) much better than the regular spectral clustering.We further show that our setting is useful

for making predictions for medical imaging settings where training data sets are tiny. VII. ACKNOWLEDGMENT The authors gratefully acknowledge support of this research via ONR grant N00014-11-1-0108 and NSF Grant NSF IIS-1422218. R EFERENCES [1] J. Shi and J. Malik, “Normalized cuts and image segmentation,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 22, no. 8, pp. 888–905, Aug. 2000. [2] S. Fortunato, “Community detection in graphs,” Physics Reports, vol. 486, no. 3-5, pp. 75 – 174, 2010. [3] U. Luxburg, “A tutorial on spectral clustering,” Statistics and Computing, vol. 17, no. 4, pp. 395–416, 2007. [4] X. Wang and I. Davidson, “Flexible constrained spectral clustering,” in KDD. New York, NY, USA: ACM, 2010, pp. 563–572. [5] M. Maier, U. von Luxburg, and M. Hein, “How the result of graph clustering methods depends on the construction of the graph,” ESAIM: Probability and Statistics, vol. 17, pp. 370–418, 1 2013. [6] F. Shang, J. L., Y. Liu, and F. Wang, “Learning spectral embedding via iterative eigenvalue thresholding,” in CIKM, 2012, pp. 1507–1511. [7] F. Shang, J. L., and W. F., “Semi-supervised learning with mixed knowledge information,” in KDD, 2012, pp. 732–740. [8] D. Luo, C. H. Q. Ding, H. Huang, and F. Nie, “Forging the graphs: A low rank and positive semidefinite graph learning approach.” in NIPS,, 2012, pp. 2969–2977. [9] P.-A. Savalle, E. Richard, and N. Vayatis, “Estimation of simultaneously sparse and low rank matrices.” in ICML. Omnipress, 2012. [10] R. He, W.-S. Zheng, B.-G. Hu, and X.-W. Kong, “Nonnegative sparse coding for discriminative semi-supervised learning,” in CVPR. IEEE, 2011, pp. 2849–2856. [11] S. Yan and H. Wang, “Semi-supervised learning by sparse representation.” in SDM. SIAM, 2009, pp. 792–801. [12] K. J. Friston, “Functional and effective connectivity: a review,” Brain connectivity, vol. 1, no. 1, pp. 13–36, 2011. [13] M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann, and I. H. Witten, “The weka data mining software: an update,” ACM SIGKDD explorations newsletter, vol. 11, no. 1, pp. 10– 18, 2009. [14] D. Zhou and C. J. Burges, “Spectral clustering and transductive learning with multiple views,” in ICML. ACM, 2007, pp. 1159–1166.

Spectral Clustering for Medical Imaging

integer linear program with a precise geometric interpretation which is globally .... simple analytic formula to define the eigenvector of an adjusted Laplacian, we ... 2http://www-01.ibm.com/software/commerce/optimization/ cplex-optimizer/ ...

406KB Sizes 0 Downloads 444 Views

Recommend Documents

Spectral Clustering - Semantic Scholar
Jan 23, 2009 - 5. 3 Strengths and weaknesses. 6. 3.1 Spherical, well separated clusters . ..... Step into the extracted folder “xvdm spectral” by typing.

Spectral Clustering for Time Series
the jth data in cluster i, and si is the number of data in the i-th cluster. Now let's ... Define. J = trace(Sw) = ∑K k=1 sktrace(Sk w) and the block-diagonal matrix. Q =.... ..... and it may have potential usage in many data mining problems.

Spectral Clustering for Complex Settings
2.7.5 Transfer of Knowledge: Resting-State fMRI Analysis . . . . . . . 43 ..... web page to another [11, 37]; the social network is a graph where each node is a person.

Parallel Spectral Clustering
Key words: Parallel spectral clustering, distributed computing. 1 Introduction. Clustering is one of the most important subroutine in tasks of machine learning.

Spectral Embedded Clustering
2School of Computer Engineering, Nanyang Technological University, Singapore ... rank(Sw) + rank(Sb), then the true cluster assignment ma- trix can be ...

Spectral Embedded Clustering - Semantic Scholar
A well-known solution to this prob- lem is to relax the matrix F from the discrete values to the continuous ones. Then the problem becomes: max. FT F=I tr(FT KF),.

Parallel Spectral Clustering - Research at Google
a large document dataset of 193, 844 data instances and a large photo ... data instances (denoted as n) is large, spectral clustering encounters a quadratic.

Active Spectral Clustering - Computer Science, UC Davis
tion, social network analysis and data clustering can be abstracted into a graph ... Previous research [5] showed that in batch constrained clustering, not all given ...

Spectral Clustering with Limited Independence
Oct 2, 2006 - data in which each object is represented as a vector over the set of features, ... and perhaps simpler “clean-up” phase than known algo- rithms.

Flexible Constrained Spectral Clustering
Jul 28, 2010 - H.2.8 [Database Applications]: Data Mining. General Terms .... rected, weighted graph G(V, E, A), where each data instance corresponds to a ...

Parallel Spectral Clustering Algorithm for Large-Scale ...
Apr 21, 2008 - Spectral Clustering, Parallel Computing, Social Network. 1. INTRODUCTION .... j=1 Sij is the degree of vertex xi [5]. Consider the ..... p ) for each computer and the computation .... enough machines to do the job. On datasets ...

Parallel Spectral Clustering Algorithm for Large-Scale ...
1 Department of ECE, UCSB. 2 Department of ... Apr. 22, 2008. Gengxin Miao Et al. (). Apr. 22, 2008. 1 / 20 .... Orkut is an Internet social network service run by.

Consensus Spectral Clustering in Near-Linear Time
chine learning and data mining applications [11]. The spectral clustering approaches are prohibited in such very large-scale datasets due to its high ...

Multi-view clustering via spectral partitioning and local ...
(2004), for example, show that exploiting both the textual content of web pages and the anchor text of ..... 1http://www.umiacs.umd.edu/~abhishek/papers.html.

Diffusion Maps, Spectral Clustering and Eigenfunctions ...
spectral clustering and dimensionality reduction algorithms that use the ... data by the first few eigenvectors, denoted as the diffusion map, is optimal under a ...

Self-Taught Spectral Clustering via Constraint ...
Oracle is available, self-teaching can reduce the number ... scarce and polling an Oracle is infeasible. ... recover an almost perfect constraint matrix via self-.

Multi-way Constrained Spectral Clustering by ...
for data analysis. Typically, it works ... tor solutions are with mixed signs which makes incor- porating the ... Based on the above analysis, we propose the follow-.

Kernel k-means, Spectral Clustering and Normalized Cuts
[email protected]. Yuqiang Guan. Dept. of Computer ... republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee.

Multi-Objective Multi-View Spectral Clustering via Pareto Optimization
of 3D brain images over time of a person at resting state. We can ... (a) An illustration of ideal- ized DMN. 10. 20 .... A tutorial on spectral clustering. Statistics and ...

Nanometer Scale Spectral Imaging of Quantum Emitters in Nanowires ...
Published: December 23, 2010 r 2010 American Chemical Society. 568 dx.doi.org/10.1021/nl103549t |Nano Lett. 2011, 11, 568-573. LETTER pubs.acs.org/NanoLett. Nanometer Scale Spectral Imaging of Quantum Emitters in Nanowires and Its Correlation to Thei

Nanometer Scale Spectral Imaging of Quantum ...
Dec 23, 2010 - electronic and electromagnetic environment at the nanometer scale. In the special ... tron energy loss spectroscopy (EELS) in a scanning transmission electron .... data set at position indicated by the thin yellow cylinders in (d).

Nanometer Scale Spectral Imaging of Quantum Emitters in Nanowires ...
Dec 23, 2010 - low energy gap semiconductor embedded in a larger gap one, which mimics the ... photon sources,5 where quantum confinement is exploited for tailoring the ... It seems thus interesting to switch to the alternative given.

On Constrained Spectral Clustering and Its Applications
Our method offers several practical advantages: it can encode the degree of be- ... Department of Computer Science, University of California, Davis. Davis, CA 95616 ...... online at http://bayou.cs.ucdavis.edu/ or by contacting the authors. ...... Fl

Consensus Spectral Clustering in Near-Linear Time
quality large-scale data analysis. I. Introduction. Clustering is one of the most widely used techniques for data analysis, with applications ranging from statistics, ...