Improving Support Vector Machine Generalisation via Input Space Normalisation Shawkat Ali and Kate A. Smith Faculty of Information Technology Monash University, Victoria 3800, Australia. [email protected], [email protected] Abstract Data pre-processing always plays a key role in learning algorithm performance. In this research we consider data pre-processing by normalization for Support Vector Machines (SVMs). We examine the normalization effect across 112 classification problems with SVM using the rbf kernel. We observe a significant classification improvement due to normalization. Finally we suggest a rule based method to find when normalization is necessary for a specific classification problem. The best normalization method is also selected by SVM itself. Keywords: Normalization, Classification, Support Vector Machines I. INTRODUCTION Support Vector Machines (SVMs) [1-3] are a kind of elegant machine learning tool, which can solve classification, regression and novelty detection problems with better generalisation compared with other traditional learning algorithms. Let us consider, a linear SVM over a r training set {( xi , y i ), i = 1,L, n}, x i ∈ ℜ m , y i ∈ {− 1,1},

r

where the task is to estimate a vector ω and a scalar b to r r construct an optimal hyperplane defined by ω ⋅ x + b ; as follows: 1 r r ω ⋅ω min r ω ,b 2 (1) r r subject to y i ( ω ⋅ x i + b ) ≥ 1 which leads to the dual optimisation problems such as: r r 1 max W (α ) = ∑ α i − ∑ α i α j y i y j xi ⋅ x j α 2 i i, j

subject to ∑ y iα i = 0 i αi ≥ 0

(2)

Now the goal is to estimate the optimisation parameter properly. The generalisation performance of SVMs and many other function estimation algorithms depend on appropriate parameter estimation. Data normalization has already showed significant improvement of generalisation performance with SVM and other methods [4-6, 22-25]. Herbrich and Graepel [7] have shown in terms of image classification that normalization is a data pre-processing method which plays an important role in SVM classification with real world problems. In this research we examine the issue of data normalization across a wide range of classification problems and propose a simple rule based methodology to find where normalization is beneficial. Finally we select an appropriate normalization

method with the help of popular learning algorithms with priority given to SVM itself for a particular classification problem. Our methodology seeks to understand the characteristics of the data (for particular classification problems), and explain why a normalized input space might offer better generalisation. First we classify a wide range of natural classification problems with the most popular learning algorithms including SVM with radial basis function (rbf) kernel. Automatic rbf kernel parameter selection is described in [8]. After that we use all these problems with SVM rbf with modified input space with different normalized methods. Then we identify the dataset characteristics matrix by statistical measures to generate rules where normalization is necessary or not. Finally we use a set of popular learning algorithms to predict the appropriate normalization method selection for a particular classification problem with SVM. All 112 classification problems (see Appendix I) are considered from the UCI Repository [9] and Knowledge Discovery Central [10] database. Over all the experiments we consider 10 Fold Cross Validation (10FCV) performances. Our paper is organised as follows: In Section II, we provide some theoretical frameworks regarding SVM input space normalization. Section III analyses the experimental results. All statistical measures to identify the dataset characteristics matrix are summarized in Section IV. A priori normalization/non normalization method selection will be explained in V. The most suitable normalization method selection is described in section V. Finally we conclude our research towards the end in Section VI. II. SVMS INPUT SPACE One speciality of SVM is that it transforms the data by adopting kernel. During the transformation some kernels essentially normalize the data points automatically. For example, rbf kernel [3] as follows:  r r 2  xi − x j  r r  where σ > 0 (3) K xi , x j = exp − 2   2 σ     We propose some additional normalization for SVM before we transform the data points in the kernel space. The normalized data is used in the SVM kernel space rather than natural data input. We provide the preference on global attributes values normalization

(

)

rather than single attribute normalization. Two types of normalization [11] are examined in this research as described in the following sections. Min-Max Normalization The min-max normalisation basically needs the minimum and maximum values from the data matrix. The data matris also depend on the upper and lower bound specification. The formulation of the Min-Max normalization is: D (i ) − min( D ) (4) D ′(i ) = * (U − L ) + L max( D ) − min( D ) where D ′ is the normalized data matrix, D is the natural data matrix and U and L are the upper and lower normalization bound. This type of normalization method is used to normalize the data matrix into a desired bound. The most popular bound is between 0 and 1. We also change the bound values to between 0 and –1, as well as between 1 and –1. Zero-Mean Normalization This type of normalisation generally require the global mean value and the standard deviation of the data matrix. The formulation of the Zero-Mean normalization is as follows: D−D D′ =  (5) 

 σ  where D is the mean of the natural data matrix D and

σ

is the standard deviation of the same data matrix. In this normalization method, the mean of the normalized data points is reduced to zero. Due to this, the mean and standard deviation of the natural data matrix is required. The normalization performance of UCI dataset ‘xab’ is shown in Figure 1. The natural data distribution is highly positive skewed. But the normalized ‘xab’ dataset is shown balanced skewed. The hyperplane construction procedure of SVM with rbf kernel for UCI data set ‘wpbc’ is shown in Figure 2.

(a)

Figure 1. Graphically represents the natural and different normalized distribution of ‘xab’ UCI dataset. The data distribution scale is completely different after getting normalization. During the classification of SVM using natural data points, several optimal decision boundaries are constructed, but only a single boundary is constructed for normalized data points. The misclassification error is higher for natural data points than normalized data points.

SVM with rbf kernel: natural data points learning.

(b) SVM with rbf kernel: Min-Max normalized (1&-1) data points learning.

(c) SVM with rbf kernel: Min-Max normalized (0&-1) data points learning.

(d) SVM with rbf kernel: Zero-Mean normalized data points learning.

Figure 2. The hyperplane construction procedure for natural and normalized data points of UCI ‘wpbc’ data set. SVM considers the rbf kernel with width 1.

80 70 60

A v e ra g e % a c c u ra c y

50 40 30 20 10

me ro ze

m

m

in -

in -

ma

ma

x

x( 1&

(0 &

-1 )

-1 )

da al na tu r

an

lo g

0

ta

We now compare the scaling method with normalization performance. This method also transforms the data points within a certain range. Log Scaling The formulation of the log scaling [12] method of the data matrix is as follows: D′(i ) = log( D (i )) (6) This type of scaling transforms the data points within a logarithmic scale.

Name of methods

Figure 4. Average accuracy performance of different normalization and scaling methods compared to natural data points classification performance. Figure 3. The hyperplane construction procedure for log scaled data points of UCI ‘wpbc’ data set. SVM considers the rbf kernel with width 1. 10

C PU tim e in Sec.

8 7 6 5 4 3 2 1

ze

ro -

me

an

log

-1 ) (1 & in-

m

in-

ma

ma

x

x

(0 &

al

da

-1 )

ta

0

m

III. EXPERIMENTAL RESULTS Classification and Computational Performance We tested the performance of the different normalization methods with SVM. All the methods average performance based on accuracy and computational time is shown in Figure 4 and 5. We observed normalization performance is generally much better than natural data sets classification performance for SVM. The log scaling performance is not carried out the similar significance with normalization. In terms of computational cost, SVM needed more time to classify the normalized data points than natural data. We observed among the different methods of normalization 84% of the 112 data sets performed better with some kind of normalization. The rest of the data sets showed better classification performance without normalizing or scaling. We found some visual reason why normalization is not always effective. First, those data sets with a combination of negative continuous and discrete attributes values, second a combination of categorical and continuous and finally those holding categorical attributes values that all are not aspect data normalization. We observed after getting normalisation these data points become very close in the feature space, where optimal hyperplane construction procedure is complex. In the following section we describe the methodology we use to assist in the appropriate selection of the best method for a given dataset. First each dataset is described by a set of measurable meta characteristics; we then combine this information with the

9

na tu r

It is observed that log scaled data points are very difficult to classify by SVM as also shown in Figure 3. The classification error is higher than normalized data classification performance.

Name of methods

Figure 5. Average computational performance of different normalization and scaling method with comparing natural data points classification performance. performance results; and finally use a rule-based induction method to provide rules describing when each method for SVM is likely to perform well. IV. DATASETS CHARACTERISTICS MEASUREMENT Each dataset can be described by simple, distance and distribution based statistical measures [13,14]. These three sets of measures characterise the datasets in different ways. First, the simple classical statistical measures identify the data characteristics based on variable to variable comparisons. Then, the distance based measures identify the data characteristics based on sample to sample comparisons. Finally, the density based measures consider the single data point from a matrix to identify the datasets characteristics. We average most of statistical measures over all the variables and take these as global measures of the dataset characteristics.

Simple Statistical Measures Descriptive statistics can be used to summarise any large dataset into a few numbers that contain most of the relevant characteristics of that dataset. The following table lists the statistical measures used in this work as provided by the Matlab Statistics Toolbox [12] and some other different sources [15, 16]. Geometric mean Max. and Min. eigenvalue Harmonic mean Skewness Trim mean Kurtosis Standard deviation Correlation Coefficient Interquartile Range Prctile Distance Based Measures Distance based measures calculate the dissimilarity between samples. We measure the different distance between each pair of observations for each dataset is listed in the below table. Euclidean distance Mahalanobis distance City Block distance Distribution Based Measures The probability distribution of a random variable describes how the probabilities are distributed over the various values that the random variable can take on. We measure the probability density function (pdf) and cumulative distribution function (cdf) for all datasets by considering different types of distributions are listed in the following table. Chi-square pdf Chi-square cdf Normal pdf Binomial pdf Exponential pdf

Normal cdf Discrete uniform cdf F pdf

Gamma pdf Lognormal pdf

Poisson pdf Student's t pdf

Rayleigh pdf

Noncentral T cdf (nctcdf)

These measures are all calculated for each of the datasets to produce a dataset characteristics matrix. Finally by combining this matrix with the performance results we can derive rules to suggest when certain methods are appropriate. V. RULE GENERATION The trial-and-error approach could be a very common procedure to select the normalization or non normalization method for SVM classification. It is a computationally complex task to find an appropriate method by following this procedure. If we are interested in applying a specific method to a particular problem we have to consider which method is more suitable for which problem. The suitability test can be done from rules developed with the help of the data characteristics properties. Rule based learning algorithms, especially decision trees (also called classification trees or hierarchical classifiers), are a divide-and-conquer approach or a top-down induction method, that have been studied

with interest in the machine learning community. Quinlan [17] introduced C4.5 and then C5.0 algorithms to solve classification problems. C5.0 works in three main steps. First, the root node at the top node of the tree considers all samples and passes them through to the second node called ‘branch node’. The branch node generates rules for a group of samples based on an entropy measure. In this stage C5.0 constructs a very big tree by considering all attribute values and finalises the decision rule by pruning. It uses a heuristic approach for pruning based on statistical significance of splits. After fixing the best rule, the branch nodes send the final class value in the last node called the ‘leaf node’ [17,18]. C5.0 has two parameters: the first one is called the pruning confidence factor (c) and the second one represents the minimum number of branches at each split (m). The pruning factor has an effect on error estimation and hence the severity of pruning the decision tree. The smaller value of c produces more pruning of the generated tree and a higher value results in less pruning. The minimum branches m indicates the degree to which the initial tree can fit the data. Every branch point in the tree should contain at least two branches (so a minimum number of m = 2 ). For detail formulations see [17]. Now that the characteristics of each dataset can be quantitatively measured, we can combine this information with the empirical evaluation of normalization/natural classification performance and construct the dataset characteristics matrix. Thus, the result of the jth method on the ith dataset is calculated as:

eij − max(ei ) Rij = 1 − min (ei ) − max(ei )

(7)

where eij is the percentage of correct classification for the jth method on dataset i, and ei is a vector of accuracy for dataset i. The class values in the matrix are assigned based on the performance. If the normalization method is showed better performance than natural data set classification, then class value is 1, otherwise 2. Based on the 112 classification problems we can then train a rule-based classifier (C5.0) to learn the relationship between dataset characteristics and normalization/natural method performance. We split the matrix 90% to construct the model tree. The process is then repeated using a 10 fold cross validation approach so that 10 trees are constructed. From these 10 trees, the best rules are found for normalization/non normalization method selection based on the best test set results. The generalisation of these rules is then tested by applying each of the randomly extracted test sets and calculating the average accuracy of the rules as discussed below in Table 1.

Rules for Normalization/Non Normalization Selection The rules for normalization method selection are generated with c = 85% and m = 4 as follows: Rule 1: IF (nctcdf > 0) OR (median > 68.9) THEN dataset should need to do normalization. Rule 2: IF (nctcdf < 0) OR (median < 68.9) THEN dataset should not need to do normalization. Include principle rules are data driven was to determine if data will benefit from normalization, or is it already normalized enough. For example, we can explain the nature of the rules by following noncentral T cumulative distribution function (nctcdf) value p as shown in Figure 6. Noncentral T cumulative distribution [20-21] is highly positively skewed with respect to normal distribution as shown in Figure 6(a). We observed from the experiment, when p value of the noncentral T cumulative distribution function is less than 0, then the distribution nature of the data is closely normal. In that case, data normalization is not required for SVM classification. On the other hand, when p value is greater than 0 that means data is closely distributed as like noncentral T cumulative distribution. These types of data are needed to normalization for SVM classification. The 10 fold cross validation performance of the rule generation is summarized in Table 1. We summarised the best rules from 10FCV performance. Average 10FCV performance is more than 89%. However, which method is best for individual datasets has been shown to be quite data dependent. These rules might be useful to determine where normalization is most appropriate for which problem. Now, we can decide either normalization is necessary or not. Then we need to find which normalization method is suitable for a particular problem. We analysed the performance of each normalization and scaling method with comparing natural datasets classification performance. We found the log scaling method is not a good way to transformed SVM input space. Due to this we consider Min-Max both method and zero-mean method to normalize the SVM input space before start mining a problem. We used a set of learning algorithms including neural network (NN), decision tree (C4.5), Naive Bayes (NB) (for details see [19]) and SVM to predict the appropriate normalization method selection for a particular classification problem. We repeat the meta learning technique with using same above data characteristics as described in section 4 and individual performance results of different normalization methods for this prediction. The class membership attribute has designed by {+1 & -1}. If any learning algorithm predict +1 that means the current method is appropriate for the present dataset. The 10FCV learning algorithms performance is shown in Table 2.

Table 1: Confusion matrix based on 10FCV results for the normalization/non normalization method selection rule. Data Condition Y N

Normalization Method Best Y N 88 6 6 12

Accuracy = 89.40%

(a) Noncentral T cumulative distribution, when p = 0.15.

(b) Noncentral T cumulative distribution, when p < 0.

(c) Noncentral T cumulative distribution, when p > 0. Figure 6. Noncentral T cumulative distribution with different p values. Table 2: Normalization method selection performance with different learning algorithms. Test set Classification Performance (% accuracy) MLP NB C4.5 SVM Max-Min method [0&-1] 75 66.67 66.67 83.33 Max-Min method [1&-1] 75 50 91.67 100 Zero-Mean method 91.67 50 83.33 100 SVM was shown best performance for a specific normalization method selection. So, SVM can help itself when this algorithm needs proper normalization method.

computational time to classify the normalized data sets rather than natural data sets. We proposed a priori rule based method when normalization is necessary for SVM with a specific problem. SVM is also used by itself to find out the most suitable normalization methods for a specific problem. This research could be re-examining with other popular SVM kernels.

VI. DISCUSSIONS We investigate the normalization affect across a large scale classification problem with one of the popular classification algorithm SVM. The normalization of SVM input space can significantly influence the higher accuracy performance of the classification procedure. We find out the reason why some problems are not required normalization method. Some normalization method required less

VII. APPENDIX: DATASETS DESCRIPTION. # Datasets 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19

Dataset names

20

abalone adp adult+stretch adult-stretch allbp ann1 ann2 aph art australian balance-scale bcw bcw_noise bld bld_noise bos bos_noise breast-cancer breast-cancerwisconsin bupa

21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56

c cleveland-heart cmc cmc_noise crx dar dhp dna dna_noise DNA-n dph echocardiogram flare german glass hayes-roth h-d hea hea_noise heart hepatitis horse-23 horse-colic house-votes-84 ionosphere iris khan labor-neg lenses letter-a lung-cancer lymphography mha monk1 monk2 monk3

# samples 1253 1351 20 20 840 1131 1028 909 1051 690 625 699 683 345 345 910 910 286

# attributes 8 11 4 4 6 6 6 18 12 14 4 9 18 6 15 13 25 6

# classes 3 3 2 2 3 3 3 2 2 2 3 2 2 2 2 3 3 2

699 345

9 6

2 2

1500 303 1473 1473 490 1378 1500 2000 2000 1275 590 131 1389 1000 214 160 303 270 270 270 155 368 368 435 351 150 1063 40 24 1334 32 148 1269 556 601 554

15 13 9 15 15 9 7 60 80 60 10 7 10 24 10 5 13 13 20 13 19 22 27 16 33 4 5 16 5 16 56 18 8 6 6 6

2 5 3 3 2 5 2 3 3 3 2 2 2 2 6 3 2 2 2 2 2 2 2 2 2 3 2 2 3 2 2 8 4 2 2 2

# Datasets 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74

Dataset names

75 76 77

pvro rph shuttle-landingcontrol sick-euthyroid sma smo smo_noise sonar splice switzerland-heart t_series tae tae_noise thy_noise tic-tac-toe titanic tmris tqr trains-transformed ttt va-heart veh veh_noise vot_noise wdbc wine wpbc xaa xab xac xad xae xaf xag xah xai yha zoo

78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112

mushroom musk1 musk2 nettalk_stress new-thyroid page-blocks pendigits-8 pha phm phn pid pid_noise pima poh post-operative primary-tumor pro promoter

# samples 1137 476 1154 1141 215 1149 1399 1070 1351 1500 532 532 768 527 90 339 1257 106

# attributes 11 166 15 7 5 10 16 9 11 9 7 15 8 11 8 17 12 57

# classes 2 2 2 5 3 5 2 5 3 2 2 2 2 2 3 2 2 2

590 1093

18 8

2 2

15 1582 409 1429 1299 208 1589 123 62 151 151 3772 958 2201 100 1107 10 958 200 846 761 391 569 178 199 94 94 94 94 94 94 94 94 94 1601 101

6 15 7 8 15 60 60 8 2 5 10 35 9 3 3 11 16 9 8 18 30 30 30 13 33 18 18 18 18 18 18 18 18 18 9 16

2 2 4 3 3 2 3 5 2 3 3 3 2 2 2 2 2 2 4 4 4 2 2 3 2 4 4 4 4 4 4 4 4 4 2 7

VIII. ACKNOWLEDGEMENT This research was supported by Monash University Post Graduation Publication Award 2004. REFERENCES [1] V. Vapnik, The Nature of The Statistical Learning Theory, Springer-Verlag, New York, 1995. [2] V. N. Vapnik, Statistical Learning Theory, John Wiley & Sons, Inc, 1998. [3] V. N. Vapnik, “An Overview of Statistical Learning Theory,” IEEE Transaction on Neural Networks, vol. 10(5), 1999, pp. 988-999. [4] A. Graf, and S. Borer, “Normalization in Support Vector Machines,” in Proc. DAGM Pattern Recognition. Berlin, Germany: Springer-Verlag, 2001. [5] M. Pontil and A. Verri, “Support Vector Machines for 3-D Object Recognition,” IEEE Trans. Pattern Anal. Machine Intell., vol. 20, 1998, pp. 637–646. [6] A. B. A. Graf, A. J. Smola and S. Borer, “Classification in a Normalized Feature Space Using Support Vector Machines,” IEEE Transactions on Neural Networks, vol. 14(3), 2003, pp. 597-605. [7] R. Herbrich and T. Graepel, “A PAC-bayesian margin bound for linear classifiers: Why SVM’s work,” Advances in Neural Information Processing Systems, vol. 13, 2001. [8] S. Ali, and K. A. Smith, “Kernel Width Selection for SVM Classification-A Meta-Learning Approach,” Accepted by International Journal of Data Warehousing and Mining, Idea Publishers, USA, 2005. [9] C. Blake, and C. J. Merz, “UCI Repository of Machine Learning Databases,” http://www.ics.uci.edu/~mlearn /MLRepository.html,Irvine, CA: University of California, 2002. [10] T.-S. Lim, (2002), “Knowledge Discovery Central, Datasets,” http://www.KDCentral.com/. [11] R. L. Kennedy, Y. Lee, B. V. Roy, C. D. Reed, and R. P. Lippman, Solving Data Mining Problems Through Pattern Recognition, Prentice-Hall, NJ, 1997. [12] Statistics toolbox user’s guide, Version 3, The MathWorks, Inc. USA, 2001. [13] K. A. Smith, F. Woo, V. Ciesielski, and R. Ibrahim, “Modelling The Relationship Between Problem Characteristics and Data Mining Algorithm Performance Using Neural Networks”, C. Dagli et al. (eds.), Smart Engineering System Design: Neural Networks, Fuzzy Logic, Evolutionary Programming, Data Mining, and Complex Systems, ASME Press, vol. 11, 2001, pp. 357-362. [14] K. A. Smith, F. Woo, V. Ciesielski, and R. Ibrahim, “Matching Data Mining Algorithm Suitability to Data Characteristics Using a Self-Organising Map”, in A. Abraham and M. Koppen (eds.), Hybrid Information Systems, Physica-Verlag, Heidelberg, 2002, pp. 169180.

[15] W. Mandenhall and T. Sincich, Statistics for Engineering and The Sciences, 4th eds. Prentice Hall, 1995. [16] A. C. Tamhane and D. D. Dunlop, Statistics and Data Analysis, Prentice Hall, 2002. [17] R. Quinlan C4.5: Programs for Machine Learning, Morgan Kaufman Publishers, San Mateo, CA, 1993. [18] R. P. W. Duin, A note on comparing classifier, Pattern Recognition Letters, vol. 1, 1996. pp. 529-536. [19] I. H. Witten, and E. Frank, Data Mining: practical machine learning tool and technique with Java implementation, Morgan Kaufmann, San Francisco, 2000. [20] M. Evans, N. Hastings, and B. Peacock, Statistical Distributions, Second Edition, John Wiley and Sons, 1993. [21] N. Johnson, and S. Kotz, Distributions in Statistics: Continuous Univariate Distributions2, John Wiley and Sons, 1970. [22] R. M. Lewis and M. W. Trosset, “Sensitivity analysis of the strain criterion for multidimensional scaling,” (Accepted for publication), Computational Statistics & Data Analysis, v.50(1),January 2006, pp. 135-153. [23] T. Wigren, “Recursive prediction error identification and scaling of non-linear state space models using a restricted black box parameterization”, Automatica, (Accepted for publication), 2005. [24] H-F. Köhn, “Combinatorial individual differences scaling within the city-block metric”, Computational Statistics & Data Analysis, (In Press), 2005. [25] H. Jing, R.T. Pivik and R. A. Dykman, “A new scaling method for topographical comparisons of event-related potentials”, Journal of Neuroscience Methods, (In Press), 2005.

Improving Support Vector Machine Generalisation via ...

Abstract. Data pre-processing always plays a key role in learning algorithm performance. In this research we consider data pre-processing by normalization for Support Vector. Machines (SVMs). We examine the normalization effect across 112 classification problems with SVM using the rbf kernel. We observe a significant ...

294KB Sizes 1 Downloads 227 Views

Recommend Documents

Improving Support Vector Machine Generalisation via Input ... - IJEECS
[email protected]. Abstract. Data pre-processing always plays a key role in learning algorithm performance. In this research we consider data.

Improving Support Vector Machine Generalisation via Input ... - IJEECS
for a specific classification problem. The best normalization method is also selected by SVM itself. Keywords: Normalization, Classification, Support Vector.

Support Vector Echo-State Machine for Chaotic ... - Semantic Scholar
Dalian University of Technology, Dalian ... SVESMs are especially efficient in dealing with real life nonlinear time series, and ... advantages of the SVMs and echo state mechanisms. ...... [15] H. Jaeger, and H. Haas, Harnessing nonlinearity: Predic

Support Vector Echo-State Machine for Chaotic Time ...
Keywords: Support Vector Machines, Echo State Networks, Recurrent neural ... Jordan networks, RPNN (Recurrent Predictor Neural networks) [14], ESN ..... So the following job will be ...... performance of SVESM does not deteriorate, and sometime it ca

Exploiting Geometry for Support Vector Machine Indexing
describing its key operations: index creation, top-k ..... 4.1.3 Constructing intra-ring index For each ring, KDX ..... index = Bin search( Arr[R ][ inverted index[x]], τ ).

Support vector machine based multi-view face detection and recognition
theless, a new problem is normally introduced in these view- ...... Face Recognition, World Scientific Publishing and Imperial College. Press, 2000. [9] S. Gong ...

Fuzzy Logic and Support Vector Machine Approaches to ... - IEEE Xplore
IEEE TRANSACTIONS ON PLASMA SCIENCE, VOL. 34, NO. 3, JUNE 2006. 1013. Fuzzy Logic and Support Vector Machine Approaches to Regime ...

Support Vector Echo-State Machine for Chaotic ... - Semantic Scholar
1. Support Vector Echo-State Machine for Chaotic Time. Series Prediction ...... The 1-year-ahead prediction and ... of SVESM does not deteriorate, and sometime it can improve to some degree. ... Lecture Notes in Computer Science, vol.

Support vector machine based multi-view face ... - Brunel University
determine the bounding boxes on which face detection is performed. .... words, misalignment in views may lead to a significant drop in performance.

"Support vector machine active learning with ... -
"Support vector machine active learning with applications ... The Journal of Machine Learning Research 2 ... Increasing the pool size improves results (Figure 7) ...

Support Vector Machine Fusion of Multisensor Imagery ...
support vector machine (SVM) fusion for the classification of multisensors images ... These tools are also critical for biodiversity science and conservation [2].

Video Concept Detection Using Support Vector Machine with ...
Video Concept Detection Using Support Vector Machine with Augmented. Features. Xinxing Xu. Dong Xu. Ivor W. Tsang ... port Vector Machine with Augmented Features (AFSVM) for video concept detection. For each visual ..... International Journal of Comp

Support Vector Machines
Porting some non-trivial application to SVM tool and analyze. OR а. Comparison of Neural Network and SVM using tools like SNNS and. SVMLight. : 30 ...

Improving Predictive State Representations via ... - Alex Kulesza
Computer Science & Engineering. University of Michigan. Abstract. Predictive state representations (PSRs) model dynam- ical systems using appropriately ...

Improving Word Representations via Global Visual Context
Department of Electrical Engineering and Computer Science. University of Michagan [email protected]. Abstract. Visually grounded semantics is a very ...

Improving Word Representations via Global Visual Context
Department of Electrical Engineering and Computer Science ... In this work, we propose to use global visual context to help learn better word ... In this way, we are able to measure how global visual information contributes (or affects) .... best and

Programming Exercise 6: Support Vector Machines
You can then submit this file through the web .... Anyone knows how much it costs to host a web portal ? > .... Figure 12: Top predictors for spam email. 14 ...

Improving quantum microscopy and lithography via ...
Jul 27, 2004 - 2 Department of Aerospace and Mechanical Engineering, Princeton University, Princeton, ... 3 Institute for Quantum Studies and Department of Physics, Texas A&M University, College ..... If the pump field is tuned close to.

Privacy Preserving Support Vector Machines in ... - GEOCITIES.ws
public key and a signature can be used. .... authentication code (MAC) which is derived from the ... encryption-decryption and authentication to block the.

Model Selection for Support Vector Machines
New functionals for parameter (model) selection of Support Vector Ma- chines are introduced ... tionals, one can both predict the best choice of parameters of the model and the relative quality of ..... Computer Science, Vol. 1327. [6] V. Vapnik.