Maximum Margin Embedding Bin Zhao, Fei Wang, Changshui Zhang State Key Laboratory of Intelligent Technology and Systems, Tsinghua National Laboratory for Information Science and Technology (TNList), Department of Automation,Tsinghua University, Beijing, China Abstract We propose a new dimensionality reduction method called Maximum Margin Embedding (MME), which targets to projecting data samples into the most discriminative subspace, where clusters are most well-separated. Specifically, MME projects input patterns onto the normal of the maximum margin separating hyperplanes. As a result, MME only depends on the geometry of the optimal decision boundary and not on the distribution of those data points lying further away from this boundary. Technically, MME is formulated as an integer programming problem and we propose a cutting plane algorithm to solve it. Moreover, we prove theoretically that the computational time of MME scales linearly with the dataset size. Experimental results on both toy and real world datasets demonstrate the effectiveness of MME.

1

Introduction

In machine learning and data mining, we are often faced with data of very high dimensionality. Directly dealing with these data is usually intractable due to the computational cost and the curse of dimensionality. In recent years, many embedding (or dimensionality reduction) methods have been proposed, among which the most well-known ones include Principal Component Analysis (PCA) [5] and manifold based embedding algorithms such as ISOMAP [10], locally linear embedding (LLE) [7], Hessian LLE (HLLE) [3], Laplacian eigenmap [1], local tangent space alignment (LTSA) [13], charting [2], conformal eigenmap [8], semi-definite embedding [11], etc. However, there are still some limitations for directly applying these afore mentioned embedding methods as they are not closely related to the following classification or clustering task. Specifically, although PCA is a popular unsupervised method which aims at extracting a subspace in which the variance of the projected data is maximized (or, equivalently, the reconstruction error is minimized), by its

very nature PCA is not defined for classification tasks. That is to say, after the projection computed by PCA, the Bayes error rate may become arbitrarily bad even though the original input patterns are perfectly classifiable. Moreover, as the majority of manifold based embedding algorithms adopt a similar framework as PCA where certain form of reconstruction loss is minimized, they also suffer from the same problem as PCA. Different from previously proposed embedding methods which find the optimal subspace by minimizing some form of average loss or cost, our goal in this paper is to directly find the most discriminative subspace, where clusters are most well-separated. We propose maximum margin embedding (MME) to achieve this goal by finding the maximum margin separating hyperplanes which separate data points from different classes with the maximum margin and projecting input patterns onto the normal of these hyperplanes. These projections will later result in the maximal margin separation of data points from different classes. Moreover, MME is insensitive to the actual probability distribution of patterns lying further away from the separating hyperplanes. Furthermore, MME is robust to noise as it is insensitive to small perturbations of patterns lying further away from the decision boundary. Finally, the large margin criterion implemented in MME enables the good generalization on future data. We formulate maximum margin embedding (MME) as an integer programming problem and propose a cutting plane algorithm to solve it. Specifically, we construct a nested sequence of successively tighter relaxations of the original MME problem, and each optimization problem in this sequence could be efficiently solved using the constrained concave-convex procedure (CCCP). Moreover, we prove theoretically that the cutting plane algorithm takes time O(snd) to converge with guaranteed accuracy, where n is the total number of samples in the dataset, s is the average number of non-zero features, i.e. the sparsity, and d is the targeted subspace dimension. Our experimental evaluations on both toy and real world datasets demonstrate that MME extracts a subspace more suitable for classification or clus-

tering. Moreover, as there are no time-consuming eigendecomposition problems involved, MME converges much faster than several state-of-the-art embedding methods. The rest of this paper is organized as follows. In section 2, we present the principles of maximum margin embedding and formulate it as an integer programming problem. The cutting plane algorithm for efficiently solving this optimization problem is presented in detail in section 3. In section 4, we provide a theoretical analysis on the time complexity of our MME algorithm. Experimental results on both toy and real-world datasets are provided in section 5, followed by the conclusions in section 6.

2

Maximum Margin Embedding

Conventionally, embedding methods target to projecting input patterns into certain subspace, where data points from different classes are scattered far away from each other. However, previous methods calculate the optimal projecting direction by minimizing some form of average loss or cost, such as the reconstruction loss. These attempts could be understood as indirectly maximizing the scattering between data points from different classes. Different from these traditional embedding algorithms, maximum margin embedding targets to directly finding the optimal separating hyperplanes for the input points and projecting the data points onto the normal vectors of these hyperplanes. This projection will later result in the maximal margin separation of data points from different classes. Motivated by the success of large margin methods in the supervised scenario, we define the optimal separating hyperplane as the one that separates data points from different classes with the maximum margin. The benefits of projecting data points onto the maximum margin separating hyperplane are threefold: (1) as the maximum margin separating hyperplanes are determined by those data points lying close to the decision boundary, MME is insensitive to the actual probability distribution of patterns lying further away from the classifying hyperplane. Therefore, when a large mass of the input patterns lies far away from the ideal classifying hyperplane, we can expect MME to win against those methods that minimize some form of average loss since they unnecessarily take into account the full distribution of the input patterns. (2) MME is robust to noise as it is insensitive to small perturbations of patterns lying further away from the separating hyperplane. (3) By projecting the input data points onto the maximum margin separating hyperplane, the obtained projection vectors are likely to provide good generalization on future data. Suppose we want to embed the input patterns into a ddimensional subspace, in order to extract “new” information unrelated to the previously obtained projecting vectors, all extracted separating hyperplanes should be orthogonal

to each other. To accomplish this, we calculate the maximum margin separating hyperplanes serially and incorporate a suitable orthogonality constraint such that the r-th (r = 2, . . . , d) separating hyperplane should be orthogonal to all the previously extracted r − 1 hyperplanes. The problem now boils down to how to find the maximum margin separating hyperplanes. Different from the supervised scenario where the data labels are known, we consider the data embedding problem in the unsupervised case, where the optimal separating hyperplane is the one with the maximum margin over all possible labelings for the data points. Hence, we first need to find the labels for the input patterns such that if we subsequently run an SVM with this labeling, the margin would be maximal among all possible labelings. By projecting the input data points onto the normal of the hyperplane associated with the above calculated data labels, the projected data points from different classes could be separated maximally. Therefore, the major computation involved in maximum margin embedding is finding the maximum margin separating hyperplanes, which could be formulated as the following optimization problem: min

y,w,b,ξi ≥0

n CX 1 T w w+ ξi 2 n i=1

(1)

s.t. yi (wT φ(xi ) + b) ≥ 1 − ξi , i = 1, . . . , n AT w = 0 y ∈ {−1, +1}n Pn where i=1 ξi is divided by n to better capture how C scales with the dataset size and AT w = 0 with A = [w1 , . . . , wr−1 ] constrains that w should be orthogonal to all previously calculated projecting vectors. Moreover, φ(·) maps the data samples in X into a high (possibly infinite) dimensional feature space. Therefore, maximum margin embedding first solve problem (1) without orthogonality constraints and obtain (w1 , b1 ). Then, suppose the hyperplanes (w1 , b1 ), . . . , (wr−1 , br−1 ) have already been calculated, then the r-th hyperplane (wr , br ) is obtained as the solution to problem (1) with A = [w1 , . . . , wr−1 ]. However, the above optimization problem has a trivially “optimal” solution, which is to assign all patterns to the same class, and the resultant margin will be infinite. Moreover, another unwanted solution is to separate a single outlier or a very small group of samples from the rest of the data. To alleviate these trivial solutions, we propose the following class balance constraint −l ≤

n X ¡ i=1

¢ wT φ(xi ) + b ≤ l

(2)

where l ≥ 0 is a constant controlling the class imbalance. For the multi-class scenario, MME is still formulated as in Eq.(1). This could be understood as implicitly grouping data samples into two “super-clusters” and the extracted

projecting direction is the most discriminative one for these two “super-clusters”.

3

The Proposed Method

The difficulty with problem (1) lies in the fact that we have to minimize the objective function w.r.t. the labeling vector y, in addition to w, b and ξi . In this section, we will propose a cutting plane algorithm to solve this problem

3.1

Cutting Plane Algorithm

Theorem 1 Maximum margin embedding could be equivalently formulated as min

w,b,ξi ≥0

s.t.

n CX 1 T ξi w w+ 2 n i=1

(3)

|wT φ(xi ) + b| ≥ 1 − ξi , i = 1, . . . , n n X ¡ T ¢ −l ≤ w φ(xi ) + b ≤ l i=1

increased by 2n −n. The algorithm we propose in this paper targets to find a small subset of constraints from the whole set of constraints in problem (4) that ensures a sufficiently accurate solution. Specifically, we employ an adaptation of the cutting plane algorithm[6], where we construct a nested sequence of successively tighter relaxations of problem (4). Moreover, we will prove theoretically in section 4 that we can always find a polynomially sized subset of constraints, with which the solution of the relaxed problem fulfills all constraints from problem (4) up to a precision of ² [4]. That is to say, the remaining exponential number of constraints are guaranteed to be violated by no more than ², without the need for explicitly adding them to the optimization problem. Specifically, the cutting plane algorithm keeps a subset Ω of working constraints and computes the optimal solution to problem (4) subject to the constraints in Ω. The algorithm then adds the most violated constraint in problem (4) into Ω. In this way, a successively strengthening approximation of the original MME problem is constructed by a cutting plane that cuts off the current optimal solution from the feasible set [6]. The algorithm stops when no constraint in (4) is violated by more than ², i.e.,

T

A w=0

n

where the labeling vector y is calculated as yi sign(wT φ(xi ) + b).

∀c ∈ {0, 1}n : =

By reformulating MME as problem (3), the number of variables involved is reduced by n, but there are still n slack variables ξi in problem (3). To further reduce the number of variables, we have the following theorem Theorem 2 Problem (3) could be equivalently formulated as follows: min w,b,ξ≥0

1 T w w+Cξ 2

s.t. ∀c ∈ {0, 1}n : −l ≤

n X ¡ i=1

(4) n n 1X 1X ci |wTφ(xi )+b| ≥ ci −ξ n i=1 n i=1

¢ wTφ(xi ) + b ≤ l

Here, the feasibility of a constraint is measured by the corresponding value of ξ, therefore, the most violated constraint is the one that would result in the largest ξ. With each constraint in problem (4) represented by a vector c, we have Theorem 3 The most violated constraint could be computed as follows ½ 1 if |wT φ(xi )+b| < 1 ci = (6) 0 otherwise The cutting plane algorithm iteratively selects the most violated constraint under the current hyperplane parameter and adds it into the working constraint set Ω until no violation of constraints is detected. Assume the current working constraint set is Ω, MME could be formulated as follows min w,b,ξ≥0

AT w = 0 Any solution (w∗ , b∗ ) to problem (4) is also Pna solution to problem (3) (and vice versa), with ξ ∗ = n1 i=1 ξi∗ .

Although problem (4) has 2 constraints, one for each possible vector c = (c1 , . . . , cn ) ∈ {0, 1}n , it has only one slack variable ξ that is shared across all constraints, thus, the number of variables is further reduced by n − 1. Each constraint in this formulation corresponds to the sum of a subset of constraints from problem (3), and the vector c selects the subset. However, the number of constrains in problem (4) is n

n

1X 1X ci |wTφ(xi )+b| ≥ ci−(ξ +²) (5) n i=1 n i=1

s.t.

1 T w w+Cξ (7) 2 n n 1X 1X ∀c ∈ Ω : ci |wTφ(xi )+b| ≥ ci −ξ n i=1 n i=1 −l ≤

n X ¡ i=1

¢ wTφ(xi ) + b ≤ l

T

A w=0 The cutting plane algorithm for MME is presented in Algorithm 1. As there are two “repeat” steps involved in Algorithm 1, for the rest of this paper we denote the outer loop as LOOP-1 and the inner loop as LOOP-2.

Algorithm 1 Cutting plane algorithm for maximum margin embedding Initialize the values for C, l and ², set A = φ, r = 0 repeat Set Ω = φ repeat Solve problem (7) for (w, b) under the current working constraint set Ω, select the most violated constraint c with Eq.(6), and set Ω = Ω ∪ {c} until The newly selected constraint c is violated by no more than ² A = A ∪ {w} and r = r + 1 until r equals to the expected subspace dimension

Optimization via the CCCP

3.2

In each iteration of the LOOP-2, we need to solve problem (7) to obtain the optimal classifying hyperplane under the current working constraint set Ω. Although the objective function in (7) is convex, the constraints are not, and this makes problem (7) difficult to solve. Fortunately, the constrained concave-convex procedure (CCCP) is designed to solve those optimization problems with a concave-convex objective function under concave-convex constraints [9]. Specifically, the objective function in (7) is quadratic and the last two constraints are linear. Moreover, the first constraint as shown in Eq.(8) is, though non-convex, a difference between two convex functions. n

∀c ∈ Ω :

n

1X 1X ci −ξ − ci |wT φ(xi )+b| ≤ 0 (8) n i=1 n i=1

Hence, we can solve problem (7) with the CCCP. Given an initial point (w0 , b0 ), the CCCP Pn computes (wt+1 , bt+1 ) from (wt , bt ) by replacing n1 i=1 ci |wT φ(xi ) + b| in the constraint with its first-order Taylor expansion at (wt , bt ), 1 T w w+Cξ (9) 2 n n 1X 1X s.t. ∀c ∈ Ω : ci −ξ − ci sign(wtTφ(xi )+bt ) n i=1 n i=1 £ ¤ · wTφ(xi )+b ≤ 0 n X ¡ T ¢ w φ(xi ) + b ≤ l −l ≤

min

w,b,ξ≥0

i=1

AT w = 0 and the above QP problem could be solved in polynomial time. Following the CCCP, the obtained solution (w, b) from this QP problem is then used as (wt+1 , bt+1 ) and the iteration continues until convergence. Putting everything together, according to the formulation of the CCCP [9], we solve problem (7) with Algorithm 2.

Algorithm 2 Solve problem (7) via the constrained concave-convex procedure Initialize (w0 , b0 ) repeat Find (wt+1 , bt+1 ) as the solution to the quadratic programming problem (9). Set w = wt+1 , b = bt+1 and t = t + 1 until stopping criterion satisfied.

3.3

Justification of the Cutting Plane Algorithm

The following theorem characterizes the accuracy of the solution computed by the cutting plane algorithm. Theorem 4 For any dataset X = (x1 , . . . , xn ) and any ² > 0, the cutting plane algorithm for maximum margin embedding returns a point (w, b, ξ) for which (w, b, ξ + ²) is feasible in problem (4). Based on the above theorem, ² indicates how close one wants to be to the error rate of the best projecting hyperplane and can thus be used as the stopping criterion.

4

Time Complexity Analysis

In this section, we will provide theoretical analysis on the time complexity of the cutting plane algorithm. We will mainly focus on the high-dimensional sparse data, where s ¿ N and N is the dimension of the original feature space. For non-sparse data, by simply setting s = N , all our theorems still hold. Assuming the targeted subspace dimension is d, we will first analyze the computational time involved in LOOP-2. By multiplying this time with d, we get the time complexity of our cutting plane algorithm. Specifically, we will show that while calculating wr , each iteration of the cutting plane algorithm takes polynomial time. Since the algorithm is iterative, we will next prove that the total number of constraints added into the working set Ω, i.e. the total iterations involved in LOOP-2, is upper bounded. Theorem 5 Each iteration of LOOP-2 takes time O(sn) for a constant working set size |Ω|. Theorem 6 For any ² > 0, C > 0, l > 0 and any dataset X = {x1 , . . . , xn }, LOOP-2 terminates after adding at most CR ²2 constraints, where R is a constant number independent of n and s. Theorem 7 For any dataset X = {x1 , . . . , xn } with n samples and sparsity of s, and any fixed value of C > 0, l > 0 and ² > 0, assuming the targeted subspace dimension is d, the cutting plane algorithm takes time O(snd).

Experiments

In this section, we will validate the accuracy and efficiency of MME on both toy and real world datasets. Moreover, we will also analyze the scaling behavior of MME with the dataset size. All the experiments are performed on a 1.66GHZ Intel CoreTM 2 Duo PC running Windows XP with 1.5GB main memory.

We use 11 datasets in our experiments, which are selected to cover a wide range of properties: Ionosphere, Sonar, Wine, Breast, Iris and Digits from the UCI repository, 20 newsgroup, WebKB, Cora, USPS and ORL. For 20 newsgroup, we choose the topic rec which contains autos, motorcycles, baseball and hockey from the version 20-news-18828. For WebKB, we select a subset consists of about 6000 web pages from computer science departments of four schools (Cornell, Texas, Washington, and Wisconsin). For Cora, we select a subset containing the research paper of subfield data structure (DS), hardware and architecture (HA), machine learning (ML), operating system (OS) and programming language (PL). For USPS, we use images of digits 1, 2, 3 and 4 from the handwritten 16 × 16 digits dataset.

5.2

Visualization Capability of MME

We will first demonstrate the visualization capability of MME, where we compare with PCA, ISOMAP and LLE. Two dimensional projections on four datasets from UCI repository are shown in Figure 1, from which we observe that MME separates data samples from different classes with a larger margin, which will clearly improve the performance of the following classification or clustering algorithms.

5.3

ISOMAP

LLE

MME

PCA

ISOMAP

LLE

MME

PCA

ISOMAP

LLE

MME

Datasets

Comparisons and Accuracy

To investigate whether MME can extract useful subspaces that preserve information necessary for the classification task, we ran MME on a number of classification problems and calculate the accuracy of Nearest Neighbor classifier on the embedded data samples. We changed the number of dimensions of the estimated subspace and measured the leave-one-out classification accuracy that could be achieved by projecting data onto the extracted subspace. Besides our MME algorithm, we also implements some other competitive algorithms and present their results for comparison, i.e., PCA[5], ISOMAP[10], LLE[7] and unsupervised LDA (uLDA)[12]. For MME, a linear kernel is used, and the values of C and l are tuned using grid search. The value of ² is

PCA

ISOMAP

LLE

MME

Figure 1. Dataset visualization results of (from left to right) PCA, ISOMAP, LLE and MME applied to (from top to bottom) the “digits”, “breast cancer”, “wine” and “iris” datasets. fixed at 0.01. For ISOMAP and LLE, a k-NN graph is used, where we select k from 1, . . . , 50 and report the highest accuracy. Results in Figure 2 clearly show advantage of MME over other embedding methods.1

5.4

Dataset Size n vs. Speed

In the theoretical analysis section, we state that the computational time of MME scales linearly with the number of samples. We present numerical demonstration for this statement in figure 3, where a log-log plot of how computational time increases with the size of the dataset is shown. Specifically, lines in the log-log plot correspond to polynomial growth O(nd ), where d is the slope of the line. Figure 3 shows that the CPU-time of MME scales roughly O(n), which is consistent with theorem 7. WebKB & 20−News

2

10

CPU−Time (seconds)

5.1

PCA

WK−CL WK−TX WK−WT WK−WC 20−News O(n)

1

10

0

10

Cora−DS Cora−HA Cora−ML Cora−OS Cora−PL USPS O(n)

1

10

0

10

−1

−1

10

Cora & USPS

2

10

CPU−Time (seconds)

5

2

10

3

10 Number of Samples

10

2

10

3

10 Number of Samples

Figure 3. CPU-time (seconds) of MME as a function of the dataset size n. 1 uLDA failed to work on USPS and text datasets due to an out-ofmemory problem.

ISOMAP

0.6

PCA ISOMAP LLE uLDA MME

LLE

0.7

uLDA

0.5

MME

0.65 2

4

6 Dimension

8

0.4

10

WebKB−Washington

1

2

4

6 Dimension

8

2

4

6 Dimension

8

0.7 0.65

PCA

0.6

LLE

0.5 0

10

20

40 60 Dimension

80

0.7 0.65

PCA ISOMAP LLE MME

0.6

MME

0.55 0.5 0

100

Cora−HA

0.7

20

0.6

0.5

0.5

0.4

40 60 Dimension

80

100

Cora−ML

0.6

0.65 20

40 60 Dimension

80

PCA ISOMAP LLE MME

0.6 0.55 0

100

Cora−OS

0.7

20

40 60 Dimension

80

0.25

PCA ISOMAP LLE MME

0.2 100

0

Cora−PL

0.5

0.3

20

40 60 Dimension

80

Accuracy

0.65

0.35

0.4

PCA ISOMAP LLE MME

0.3 0.2 0

100

20−Newsgroup

1

0.4

0.8

0.9

0.35

0.7

0.85

0.4 0.35

0.3 0.25 PCA ISOMAP LLE MME

0.2 0.15

20

40 60 Dimension

80

100

0.1 0

20

40 60 Dimension

80

0.6 0.5

PCA ISOMAP LLE MME

0.4 0.3 100

0.2 0

80

0.1 0

100

20

40 60 Dimension

80

0.8 0.75

PCA ISOMAP LLE MME

0.65 0

20

40 60 Dimension

80

100

ORL

0.8

0.7

100

PCA ISOMAP LLE MME

1

0.95

Accuracy

PCA ISOMAP LLE MME

Accuracy

0.5 0.45

Accuracy

0.45

0.6 0.55

40 60 Dimension

USPS

1

0.9

0.65

20

0.3 0.2

Accuracy

PCA ISOMAP LLE MME

0.7

Accuracy

0.8 0.75

0

ISOMAP

0.55

Cora−DS

0.5

Accuracy

Accuracy

Accuracy

PCA ISOMAP LLE uLDA MME

0.4

0.7

Accuracy

0.7

0.45

0.75

0.85

0

0.75

0.55

0.95 0.9

0.8

0.75

0.6

WebKB−wisconsin

0.8

0.8

0.8

0.65

10

0.85

0.85

0.75

WebKB−Texas

0.9

0.85

Accuracy

PCA

0.75

0.7

WebKB−Cornell

0.9

0.9

Accuracy

0.8

Accuracy

0.85

Wine

0.95

0.8

0.9 Accuracy

Sonar

0.9

Accuracy

Ionosphere

1 0.95

20

40 60 Dimension

80

0.6 0.4

PCA ISOMAP LLE uLDA MME

0.2

100

0 0

20

40 60 Dimension

80

100

Figure 2. Leave-one-out classification accuracy of Nearest Neighbor classifier on the embedded data samples.

6

Conclusions

We propose maximum margin embedding (MME), to project data samples into the most discriminative subspace, where clusters are most well-separated. Detailed theoretical analysis of the algorithm is provided, where we prove that the computational time of MME scales linearly with the sample size n with guaranteed accuracy. Moreover, experimental evaluations on several real world datasets show that MME performs better than several state-of-the-art embedding methods, both in efficiency and accuracy.

Acknowledgments This work is supported by the project (60675009) of the National Natural Science Foundation of China.

References [1] M. Belkin and P. Niyogi. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Computation, 15(6):1373–1396, 2003. [2] M. Brand. Charting a manifold. In NIPS, 2003. [3] D. L. Donoho and C. E. Grimes. Hessian eigenmaps: locally linear embedding techniques for highdimensional data. In Proceedings of the National Academy of Arts and Sciences, volume 100, 2003.

[4] T. Joachims. Training linear svms in linear time. In SIGKDD, 2006. [5] I. T. Jolliffe. Principal Component Analysis. SpringerVerlag, New York, 1986. [6] J. E. Kelley. The cutting-plane method for solving convex programs. Journal of the Society for Industrial Applied Mathematics, 8:703–712, 1960. [7] S. T. Roweis and L. K. Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290:2323–2326, 2000. [8] F. Sha and L. K. Saul. Analysis and extension of spectral methods for nonlinear dimensionality reduction. In ICML, 2005. [9] A. J. Smola, S. Vishwanathan, and T. Hofmann. Kernel methods for missing variables. In AISTATS, 2005. [10] J. B. Tenenbaum, V. de Silva, and J. C. Langford. A global geometric framework for nonlinear dimensionality reduction. Science, 290:2319–2323, 2000. [11] K. Q. Weinberger, F. Sha, and L. K. Saul. Learning a kernel matrix for nonlinear dimensionality reduction. In ICML, 2004. [12] J. Ye, Z. Zhao, and M. Wu. Discriminative k-means for clustering. In NIPS. 2007. [13] Z. Y. Zhang and H. Y. Zha. Principal manifolds and nonlinear dimensionality reduction via tangent space alignment. SIAM Journal on Scientific Computing, 26(1):313–338, 2004.

Maximum Margin Embedding

is formulated as an integer programming problem and we .... rate a suitable orthogonality constraint such that the r-th ..... 5.2 Visualization Capability of MME.

488KB Sizes 2 Downloads 512 Views

Recommend Documents

Solution: maximum margin structured learning
Structured Learning for Cell Tracking. Xinghua Lou ... Machine learning for tracking: • Local learning: fail .... Comparison: a simple model with only distance and.

Unsupervised Maximum Margin Feature Selection ... - Semantic Scholar
Department of Automation, Tsinghua University, Beijing, China. ‡Department of .... programming problem and we propose a cutting plane al- gorithm to ...

Maximum Margin Supervised Topic Models - Semantic Scholar
efficient than existing supervised topic models, especially for classification. Keywords: supervised topic models, max-margin learning, maximum entropy discrimi- nation, latent Dirichlet allocation, support vector machines. 1. Introduction. Probabili

MedLDA: Maximum Margin Supervised Topic ... - Research at Google
Tsinghua National Lab for Information Science and Technology. Department of Computer Science and Technology. Tsinghua University. Beijing, 100084, China.

Efficient Maximum Margin Clustering via Cutting Plane ...
where the data samples X are mapped into a high ... kernel trick, this mapping could be done implicitly. ...... document vectors in a 8014-dimensional space.

Efficient Maximum Margin Clustering via Cutting Plane ...
Apr 25, 2008 - Automation, Tsinghua Univ. SIAM International Conference on Data Mining ..... Handle large real-world datasets efficiently. Bin Zhao, Fei Wang, ...

Efficient Maximum Margin Clustering via Cutting Plane ...
retrieval and text mining, web analysis, marketing, com- putational biology, and ... nology (TNList), Department of Automation, Tsinghua Univer- sity, Beijing, China. ..... and (w, b, ξi) are optimal solutions to problem (3.7) and. (3.9) respectivel

Efficient Maximum Margin Clustering via Cutting Plane ...
Consequently, the algorithms proposed in [19] and [18] can only handle very small datasets containing several hundreds of samples. In real world applications such as scientific information retrieval and text mining, web analysis, and compu- tational

M3IC: Maximum Margin Multiple Instance Clustering
a practical perspective, clustering plays an outstanding role in data mining applications such as information retrieval, text mining, Web analysis, marketing, computational biology, and many others [Han and Kamber, 2001]. However, so far, almost all

Unsupervised Maximum Margin Feature Selection with ...
p XLXT vp. (14). s.t. ∀i ∈ {1,...,n},r ∈ {1,...,M} : d. ∑ k=1. (vyik−vrk)xik ..... American Mathematical. Society, 1997. 3. [4] C. Constantinopoulos, M. Titsias, and A.

Page 1 1mm top margin mm top margin (5 yeaC mm top margin ...
4mm left margin Printer margin test for Letter (portrait) 4mm right margi. 5mm left margin. 6mm left margin. 7mm left margin is. 8mm left margin Expand small ...

Affinity Weighted Embedding
Jan 17, 2013 - contain no nonlinearities (other than in the feature representation in x and y) they can be limited in their ability to fit large complex datasets, and ...

Applicable Exposure Margin
Feb 27, 2017 - futures and options contracts on individual securities, the applicable ... Telephone No. Fax No. Email id. 18002660057. +91-22-26598242.

Embedding Denial
University of Melbourne [email protected]. April 10, 2011. 1 Introduction ...... denial fit to express disagreement? We've got half of what we want: if I assert.

Intensive margin, extensive margin, and allowance for ...
allowance for spouses: A discrete choice analysis. Shun-ichiro .... Section 4 explain our data, and the results ... Under the tax code in 1997, when our data.

Cauchy Graph Embedding
ding results preserve the local topology of the ... local topology preserving property: a pair of graph nodes ..... f(x)=1/(x2 + σ2) is the usual Cauchy distribution.

Factor-based Compositional Embedding Models
Human Language Technology Center of Excellence. Center for .... [The company]M1 fabricates [plastic chairs]M2 ... gf ⊗ hf . We call efi the substructure em-.

Circulant Binary Embedding - Sanjiv Kumar
to get the binary code: h(x) = sign(RT. 1 ZR2). (2). When the shapes of Z, R1, R2 are chosen appropriately, the method has time and space complexity of O(d1.5) ...

Tissue Embedding Center.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item.

Tangent-Corrected Embedding
lying instead on local Euclidean distances that can be mis- leading in image space. .... 0 means the pair is parallel (aligned) and 1 means the pair is orthogonal ...