Abstract
Twin support vector machine (TWSVM) and projection twin support vector machine (PTSVM), are two extensions of traditional support vector machine (SVM). However, TWSVM and PTSVM did not consider the local geometrical structure information of training samples. Therefore, a locality preserving projection twin support vector machine (LPPTSVM) is presented by introducing the basic idea of the locality preserving projection (LPP) into PTSVM. This method not only inherits the ability of TWSVM and PTSVM for dealing with the XOR problem, but also fully considers the local geometrical structure between samples and shows the local underlying discriminatory information. For linear LPPTSVM method, regularization technique is used to overcome the singularity problem, and then the nonlinear LPPTSVM method is constructed by the empirical kernel mapping. Experimental results conducted on the artificial datasets and UCI datasets illustrate the effectiveness of the LPPTSVM method.
Keywords
Introduction
Support vector machine (SVM) is one of the excellent machine learning tools for binary classification and regression1,2 and has been widely applied to a variety of real-world problems ranging from image classification, 3 text categorization, 4 bioinformatics, 5 etc. However, one of the main challenges for SVM is the high computational complexity and SVM solution does not fully take into account the class distribution. 6 Recently, multi-hyperplane support vector machine such as twin support vector machine (TWSVM) 7 and projection twin support vector machine (PTSVM), 8 as an extension direction of SVM, has been one of the hot research topics in the field of pattern recognition. In 2006, Mangasarian and Wild 9 proposed a nonparallel plane classifier for binary data classification, which is termed as the generalized eigenvalue proximal support vector machine (GEPSVM). The essence of GEPSVM is to seek two nonparallel planes, so that data points of each class are proximal to one of them. GEPSVM discarded the parallelism condition of proximal support vector machine (PSVM) 10 and required that each hyperplane be as close as possible to one of the data sets and as far as possible from the other data sets. Hence, multiple surface support vector machines have been widely investigated, e.g. TWSVM, 7 PTSVM, 8 MVSVM, 11 TBSVM, 12 and so on.13–17 However, in the process of learning, the above mentioned classification methods are almost not fully consider the local geometrical structure information of the sample. In order to effectively reveal the local geometrical structure of the sample, many scholars conducted a lot of research and achieved rich results, e.g. isometric mapping (IM), 18 locally linear embedding (LLE), 19 Laplacian eigenmap (LE) 20 and locality preserving projections (LPP). 21 Especially, the LPP method can keep the local geometrical structure between the samples. Furthermore, LPP can be easily extended to nonlinear embedded and can find the low dimensional nonlinear manifold structure. 22 In order to utilize the advantages of LPP and combine with SVM, Laplacian support vector machine (LSVM) is proposed in Benkin et al. 23 and minimum class locality preserving variance support vector machine (MCLPVSVM) is proposed in Wang et al. 24 However, these methods are the results of direct combination of LPP and traditional SVM. There are not many researches on the combination of LPP and multi-hyperplane support vector machines.
Based on the above analysis, in this paper, we propose a novel locality preserving projection twin support vector machine, termed LPPTSVM. LPPTSVM method introduces the basic idea of LPP into PTSVM and a regularization term is used to overcome the singularity problem. This method has the following advantages: first, LPPTSVM inherits the advantages of TWSVM and PTSVM, which can well deal with crossplane (XOR) problem; second, LPPTSVM fully considers the local geometrical structure information of samples, which can improve the generalization ability of the algorithm to a certain extent; third, regularization technique is used to overcome the singularity problems, which can ensure the stability of the algorithm.
The rest of this paper is organized as follows. The related works will be briefly reviewed in the Related works section. The Locality preserving projection twin support vector machine section proposes our linear LPPTSVM and its nonlinear version, and in addition, the successive overrelaxation (SOR) algorithm is also proposed in this section. Experimental results are described in the Experimental results section, and conclusion and future works are presented in the last section.
Related works
Consider a binary classification problem in the
Twin support vector machine
For the linear case, the TWSVM
7
determines two nonparallel hyperplanes
Projection twin support vector machine
The central idea of linear projection twin support vector machine
8
is to find a projection axis for each class, such that within-class variance of the projected data points of its own class is minimized meanwhile the projected data points of the other class scatter away as far as possible. Thus, the primal problems of linear PTSVM are a pair of QPPs
Obviously, from equations (2) to (5), we can see that the objective function of TWSVM and PTSVM does not consider local geometrical structure between the samples.
Locality preserving projection twin support vector machine
Linear LPPTSVM
In order to overcome the singularity problem of scatter matrix, a regularization term is added to objective function similar to literatures,12,15 and we introduce the basic idea of LPP into PTSVM and obtain the primal optimal problem as follows.
In order to get the solution to primal problem equation (8), we need to derive its dual problem. Therefore, the Lagrangian function is introduced as follows
Obviously, equation (11) implies that
Since
Finally, putting equations (12) and (15) into the Lagrangian function equation (10), we obtain the dual problem as follows
In the same way, we can obtain the Wolfe dual problem of equation (9) as follows
By using Lagrangian method and KKT condition, we obtain
Once
Nonlinear LPPTSVM
For nonlinear classification problem, the most commonly used approach is to first introduce nonlinear mapping and map the sample from the input space into high-dimensional feature space, and then use kernel techniques to perform linear algorithm in feature space. However, local preserving within-class scatter matrix
First, for given kernel function
Second, we use empirical kernel mapping to map the training samples from input space into empirical feature space, that is
Therefore, the new training samples in empirical feature space can be expressed as
At last, we take
Implementation
In this section, we discuss the implementation of our proposed LPPTSVM. In our LPPTSVM, the dual problem can be rewritten as the following unified form
As we can see, if
In our proposed LPPTSVM, most of the computational cost is incurred in solving the dual QPP equation (24). In order to solve the above QPP quickly, we use a very efficient optimization technique called successive overrelaxation (SOR) algorithm, which can be seen in the literatures.12,15,27 SOR is an excellent QPP solver because it is able to deal with very large datasets that need not reside in memory. 27
Select the parameter Select the parameter Compute And define Stop if Compute And define Stop if Select the parameter Compute And define Stop if Step 1.
Step1.
Step2.
Step3.
Step 2.
Step3.
Step1.
Step2.
Step3.
Experimental results
In order to evaluate the proposed LPPTSVM, we investigate its classification accuracies and computational efficiencies on three artificial datasets and a number of real-world UCI benchmark datasets. In the experiments, we focus on the comparison between our proposed LPPTSVM and some state-of-the-art classifiers, e.g. TWSVM, 7 PTSVM 8 and MVSVM. 11 All methods are implemented in MATLAB R2013a on a personal computer (PC) with an Intel (R) Core (TM) processor (3.40 GHz) and 4 GB random-access memory (RAM). PTSVM and our proposed LPPTSVM are solved by SOR algorithm. The eigenvalue problem in MVSVM is solved by MATLAB function ‘eig.m’ and the QPPs problems in TWSVM are solved by the optimization toolbox QP in MATLAB. The “Accuracy” used to evaluate methods is defined as Accuracy = (TP + TN)/(TP + FP + TN + FN), where TP, TN, FP, and FN are the number of true positives, true negatives, false positives, and false negatives, respectively. The parameters are selected by employing the standard 10-fold cross-validation methodology.
Toy examples
In this sub, three artificial datasets, including crossplane (XOR), complex XOR dataset and Two-moons manifold dataset
23
have been used to show that our proposed LPPTSVM can deal with linearly inseparable problems. In the experiment, XOR dataset contains 200 samples (100 positive samples and 100 negative samples) and Complex XOR dataset contains 260 samples (100 positive samples and 160 negative samples). Figure 1 shows the XOR and complex XOR dataset. Two-moons manifold dataset contains 100 samples (50 positive samples and 50 negative samples), and Figure 2 shows two kinds of Two-moons datasets with different complexity.
Crossplane (XOR) (a) and Complex XOR (b) datasets. Two kinds of Two-moons datasets with different complexity. (a) Two-moons-1 dataset and (b) Two-moons-2 dataset.

Classification accuracy of TWSVM, PTSVM, and LPPTSVM on XOR datasets.
Classification accuracy of TWSVM, PTSVM, and LPPTSVM on Two-moons datasets.
UCI datasets
Test results of linear MVSVM, TWSVM, PTSVM, and LPPTSVM.
Test results of nonlinear MVSVM, TWSVM, PTSVM, and LPPTSVM.
Conclusions
In this paper, a novel locally preserving projection twin support vector machine is proposed by combined with LPP and PTSVM. This method not only inherits the advantages of TWSVM and PTSVM, but also fully considers the local geometrical structure information between samples. Experimental results obtained on artificial datasets and real-world UCI datasets illustrate the effectiveness of the proposed LPPTSVM. It should be pointed out that there are many parameters in our LPPTSVM, so the parameter selection is a practical problem and should be investigated in the future.
Footnotes
Acknowledgment
The authors would like to thank Dr Yuan-Hai Shao from Zhejiang University of Technology for his valuable discussion and help.
Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This work was supported in part by the National Natural Science Foundation of China under grant nos 61373055 and 61103128, and the Youth Fund Project of Anqing normal university under grant no. KJ201308.
