The Scientific World Journal

Volume 2015, Article ID 497617, 12 pages

http://dx.doi.org/10.1155/2015/497617

## A Learning Framework of Nonparallel Hyperplanes Classifier

^{1}College of Mathematics and Systems Science, Xinjiang University, Urumqi 830046, China^{2}State Key Lab of Biochemical Engineering, Institute of Process Engineering, Chinese Academy of Sciences, Beijing 100190, China^{3}Zhijiang College, Zhejiang University of Technology, Hangzhou 310024, China

Received 21 June 2014; Revised 19 September 2014; Accepted 19 September 2014

Academic Editor: Qiankun Song

Copyright © 2015 Zhi-Xia Yang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

A novel learning framework of nonparallel hyperplanes support vector machines (NPSVMs) is proposed for binary classification and multiclass classification. This framework not only includes twin SVM (TWSVM) and its many deformation versions but also extends them into multiclass classification problem when different parameters or loss functions are chosen. Concretely, we discuss the linear and nonlinear cases of the framework, in which we select the hinge loss function as example. Moreover, we also give the primal problems of several extension versions of TWSVM’s deformation versions. It is worth mentioning that, in the decision function, the Euclidean distance is replaced by the absolute value , which keeps the consistency between the decision function and the optimization problem and reduces the computational cost particularly when the kernel function is introduced. The numerical experiments on several artificial and benchmark datasets indicate that our framework is not only fast but also shows good generalization.

#### 1. Introduction

Classification problem is an important issue in machine learning and data mining, which is mainly comprised of binary and multiclass classification. Support vector machine (SVM), proposed by Burges [1] and Cortes and Vapnik [2], is an excellent tool for classification. In contrast with conventional artificial neural networks (ANNS) which aim at reducing empirical risk, SVM is principled and implements the structural risk minimization (SRM) that minimizes the upper bound of the generalization error [3–5]. Within a few years after its introduction, SVM has been successfully applied to pattern classification and regression estimation like face detection [6, 7], text categorization [8], time series prediction [9], bioinformatics [10], and so forth.

Recently, for binary classification, Mangasarian and Wild [11] proposed the generalized eigenvalue proximal support vector machine (GEPSVM) via two nonparallel hyperplanes. In their approach, the data points of each class are proximal to one of two nonparallel hyperplanes. The nonparallel hyperplanes are determined by eigenvectors corresponding to the smallest eigenvalues of two related generalized eigenvalue problems. Inspired by GEPSVM [11], Jayadeva et al. [12] developed twin SVM (TWSVM) with two nonparallel hyperplanes. However, the two hyperplanes are got by solving two quadratic programming (QP) problems, similar to the standard SVM. Furthermore, TWSVM differs from the standard SVM in fundamental way. In TWSVM, one solves a pair of smaller size QP problems rather than a single QP problem in the standard SVM. Therefore, TWSVM works faster than the standard SVM. Subsequently, there are many extensions for TWSVM including the improvements on TWSVM (TBSVM) [13], the least square TWSVM (LS-TWSVM) [14–17], nonparallel plane proximal classifier (NPPC) [18], smooth TWSVM [19], geometric algorithm [20], and twin support vector regression (TWSVR) [21]. TWSVM was also extended to deal with multiclassification TWSVM [22–24]. More precisely, in [22], TWSVM was extended straight from binary classification to multiclass classification, in which each primal problem covers all patterns except the patterns of the th class in the constraints for the th () hyperplane. In [23], the authors extended TWSVM based on the idea of “one-versus-rest” (1-v-r) from binary classification to multiclass classification, in which there are two quadratic programming (QP) problems for each reconstructing binary classification. However, they both have not kept the advantage of TWSVM which has lower computational complexity than that of the standard SVM. In [24], Yang et al. proposed multiple birth SVM (MBSVM) with much lower computational complexity than that of both [22, 23] by solving smaller size of QP problems for -class classification; only the empirical risk is considered like TWSVM. However, in TBSVM [13], the structural risk minimization principle is implemented by introducing the regularization term.

In this paper, we propose a novel learning framework of nonparallel hyperplanes support vector machines based on TWSVM and its extension versions, called NPSVMs, which not only provide a unified view for TWSVM and its many extension versions but also can deal with binary and multiclass classification problems. For binary classification, if the loss function is the hinge loss function, then the framework can become TWSVM [12] or TBSVM [13] with different parameters; if the loss function is the square loss function, then the framework is LS-TWSVM [14]; if the loss function is the convex combination of the linear and square loss functions, then the framework is NPPC [18]. Actually, we can also get smooth TWSVM [19] by replacing 2-norm with 1-norm in the framework. However, for multiclass classification, the framework does not directly extend, in which we switch the roles of the patterns of the -th class and the rest class and replace “min” with “max” in the decision function. Moreover, we only use the absolute value rather than the Euclidean distance in the decision function due to the twofold reasons: reducing the computational cost particularly when the kernel function is introduced and making the consistency since it is the corresponding absolute value that appears in the primal problems. Concretely, we discuss the linear and nonlinear cases of the framework, in which we select the hinge loss function as example. Moreover, we also give the primal problems of extensions of LS-TWSVM, 1-norm LS-TWSVM, NPPC, and smooth TWSVM. Finally, the numerical experiments on several artificial and benchmark datasets indicate that our frameworks are not only fast but also show good generalization.

The paper is organized as follows. Section 2 introduces the brief reviews of SVMs. Section 3 proposes our frameworks, in which Section 3.1 discusses the linear framework, Section 3.2 extend into the nonlinear framework, Section 3.3 gives SOR algorithm for solving the hinge NPSVMs, and Section 3.4 discusses several other extension approaches. Finally, Section 4 deals with experimental results and Section 5 contains concluding remarks.

#### 2. Brief Reviews of SVMs

##### 2.1. Twin Support Vector Machine

Given the following training set for the binary classification:where is the th data point, the input is a pattern, the output is a class label, , and is the number of data points. In addition, let and be the number of data points in positive class and negative class, respectively, and . Furthermore, the matrices and consist of the inputs of Class and the inputs of Class 2, respectively.

The gaol of TWSVM [12] is to find two nonparallel hyperplanes in -dimensional input space:such that one hyperplane is close to the patterns of one class and far away from the patterns of the other class to some extent. TWSVM is in spirit of GEPSVM [11]. But both of GEPSVM and TWSVM are different from the standard SVM. For TWSVM, each hyperplane is generated by solving a QP problem looking like the primal problem of the standard SVM. The primal problems of TWSVM can be presented as follows:where and are nonnegative parameters and and are vectors of ones of appropriate dimensions. In the QP problem (4), the objective function tends to keep hyperplane (2) close to the patterns of Class 1 and the constraints require the hyperplane (2) to be at a distance of at least 1 from the patterns of Class 2. The QP problem (5) has similar property. Moreover, we note that the constraints do not contain all patterns in the training set (1) but are determined by only the patterns of one class in both classes. Therefore, in [12], the authors claimed that TWSVM is approximately four times faster than the standard SVM.

Define and . It has been shown that when both and are positive definites, the Wolfe duals of (4) and (5) are written as follows:respectively, where and are Lagrangian multipliers.

In order to avoid the possible ill-conditioning of and , TWSVM introduces a term (), where is an identity matrix of appropriate dimensions. Thus, the nonparallel hyperplanes (2) and (3) can be obtained from the solutions and of the QP problems (6) and (7). Considerwhere , . Moreover, a new pattern is assigned to Class (), depending on which of the two nonparallel hyperplanes given by (2) and (3) lies closer to; that is,

##### 2.2. Multiple Birth Support Vector Machine

Given the training setwhere the input , , is the pattern and the output is the class label. The task is to seek hyperplanes,and assign the class label according to which hyperplane a new pattern is farthest from.

For convenience, denote the number of data points of the th class in the training set (10) as and define the following matrixes: the patterns belonging to the th class are represented by the matrix , . In addition, define the matrixthat is, consists of the patterns belonging to all classes except the th class, . The primal problems of MPSVM [24] are comprised of the following QP problem:where and are the vectors of ones, is the slack variable, and is the penalty parameter, . The dual problem of QP problem (13) is formulated as follows:where the penalty parameter , , and , . Similarly, in order to avoid the possibility of the ill-conditioning of the matrix in some situations, one introduces a regularization term , where is a fixed small scalar and is the identity matrix with appropriate size.

After getting the solution to the above QP problem (13) with , a new pattern is assigned to class (), depending on which of the hyperplanes given by (11) lies farthest from; that is, the decision function is represented aswhere is the absolute value.

#### 3. The Framework of Nonparallel Hyperplanes Classifiers

In this section, we propose a learning framework of nonparallel hyperplanes classifier, which gives a unified form for TWSVM and its many extension versions and extend them into multiclass classification problem. We first develop the linear framework and then extend it to nonlinear framework.

##### 3.1. Linear Framework

Given the training set (10), the task is to find nonparallel hyperplanes:one for each class. For obtaining the unknown hyperplanes, we construct the following standard framework for each unknown hyperplane:where the matrix is comprised of the patterns in the th class, the matrix is defined (12), and are the parameters, and are vectors of ones of appropriate dimensions, , and is the loss function (e.g., square loss, hinge loss, etc.). In the optimization problem (17), the first term approximatively minimizes the sum of the squared Euclidean distances from the patterns except for the th class to hyperplanes; the second term is the Tikhonov regularization term [25] and can implement the structural risk minimization principle like TBSVM [13]; the third term constitutes the loss function which is defined different loss functions corresponding to different models.

For a new pattern , we assign to class () according to the following decision function:where is the absolute value. Note that we only use the absolute value in the decision function. There are two main reasons: one is that the first term of the optimization problem (17) just minimizes the sum of the square rather than the sum of square Euclidean distance from the patterns to hyperplanes, so it should keep consistency between the optimization problem and the decision function; another is that it reduces the computational cost particularly when the kernel function is introduced afterwards.

In fact, if , the parameter is equal to 0, and the loss function is hinge loss function, that is, , then the optimization problem (17) becomes TWSVM [12]. Moreover, if the parameter is alterable, then it is TBSVM [13]. And if the loss function is the square loss function, that is, , it is LS-TWSVM [14]. And if the loss function is a convex combination of linear and square loss, that is, , where , then it is NPPC [18]. Other extension versions of TWSVM also can be contained in the optimization problem (17), for instance, smooth TWSVM, 1-norm LS-TWSVM [17], and so forth, in which we just need to select proper norm or loss function.

More importantly, our framework can solve multiclass classification problem, which is extension of TWSVM, TBSVM, LS-TWSVM, NPPC, and so forth. It should be pointed out that our framework is not straight extension of TWSVM and its deformation versions. Concretely, from the optimization problem (17), we can see that the first term contains the patterns except for those of the th class and the third term just involves the patterns of the th class. This strategy cannot lead to significant increase of the complexity of the optimization when the number of classes increases. We will dwell on in specific algorithm afterwards.

Now, we give the detailed algorithm to the hinge loss function as an example, called hinge NPSVM (HNPSVM). And then the optimization problem (17) is the following formulation with the hinge loss function:where the matrix is comprised of the patterns in the th class, the matrix is defined (12), and are the parameters, and are vectors of ones of appropriate dimensions, and . Actually, the problem is equivalent to the following quadratic programming:where the matrix is comprised of the patterns in the th class, the matrix is defined (12), and are the parameters, and are vectors of ones of appropriate dimensions, and .

In fact, for , we have QP problems like (20). In particular, when is equal to 2, that is, , the QP problems (4) and (5) can be obtained as a special case of (20) with . For simplicity, assume that the number of each class points is almost balanced; namely, the number of the th class is . Then, note that the constraints just involve the patterns of the th class, so the complexity of the the problem (20) is no more than . However, if TWSVM is directly extended to multiclass classification case like [22], we will get a different optimization problem, in which the roles of patterns of the th class and the rest class are switched. Thus, the complexity of the optimization problem will increase significantly and is determined by the patterns except for the patterns of the th class in the training set (10), which is no more than . Obviously, our approach is approximately times faster than the model in [22]. On the other hand, when the number of each class points is unbalanced, our apprach still is faster than the model in [22] because the complexity of our optimization problem just is decided by the number of the patterns of the th class rather than the patterns of the rest classes. Therefore, our HNPSVM keeps the computation complexity low.

It is well known that the solution of primal problem (20) is obtained from the solutions of their dual problems. So we now derive their dual problems. The Lagrangian function of the problem (20) is given bywhere , are nonnegative Lagrange multiplier vectors. The Karush-Kuhn-Tucker (KKT) necessary and sufficient optimality conditions [26] for the QP problem (20) are given bySince , according to (24), we haveNext, from (22) and (23), we can obtainwhere is an identity matrix of appropriate dimensions. Let ; (29) can be written aswhere and . And then putting (30) into the Lagrangian function (21) and using (22)–(28), we can get the dual problem of the primal problem (20):where and are parameters and . Obviously, if we have the solution of the QP problem (31), then we obtain the nonparallel hyperplanes (16) by (30).

It is worth mentioning that the parameter replaces as in (8), so is no longer a fixed small scalar but a weighting factor which determines the trade-off between the regularization term and the empirical risk in the problem (20). Therefore, the high and low of the value of reflects the structure of minimization principle and our HNPSVM includes MBSVM.

##### 3.2. Nonlinear Framework

Similarly, we also extend the linear framework of NPSVMs to nonlinear case. For a -class classification (10), our goal is to find kernel-generated hyperplanes:where and is an appropriately chosen kernel function.

In order to obtain the hyperlanes (32), we construct the following framework formulation:where the matrix is comprised of the patterns in the th class, the matrix is defined (12), and are the parameters, and are vectors of ones of appropriate dimensions, , and is the loss function (e.g., square loss or hinge loss, etc.). Similarly, as discussed in the last subsection, the problem (33) can be reduced to the nonlinear formulations of the difference approaches (e.g., TWSVM, TBSVM, LS-TWSVM, NPPC, etc.) when the difference loss functions or parameters are selected for .

A new pattern is assigned to the th class by the following decision functions:where is the absolute value. Note that, in this decision function (34), we just compute the absolute value rather than Euclidean distance from the pattern to the hyperplanes. This strategy reduces the complexity of computation because Euclidean distance should be from the pattern to the th hyperplanes. Thus, the decision function (34) not only saves the computation quantity but also keeps the consistency with the first term of the problem (33).

Now, we still select the hinge loss function as example. Then, the problem (33) can be formulated as follows:where the matrix is comprised of the patterns in the th class, the matrix is defined by (12), and are parameters, and are vectors of ones of appropriate dimensions, and . Similarly, derived process with the linear case, its dual problem is formulated as:where and are parameters, , , and . And the augmented vector is given by .

##### 3.3. SOR Algorithm

In our HNPSVMs, the QP problems (31) and (36) can be rewritten as the following unified forms:where is positive definite. For example, the above problem becomes the problem (36), when , .

The above problem (37) can be solved efficiently by the following successive overrelaxation (SOR) algorithm; see [27].

*Algorithm 1. *SOR for the QP problem (36) is as follows.(1)Select the parameter and the initial value .(2)Suppose that is obtained by the times iterate; compute according to the following iterate formula:where . And define , where and are the strictly lower triangular matrix and the diagonal matrix, respectively.(3)Stop if is less than some desired tolerance. Else, replace by and by and go to 2.

SOR is an excellent TWSVM solver, because it can process efficiently very large datasets that need not reside in memory. Furthermore, it has been proved that this algorithm converges linearly to a solution in [27, 28]. It should be pointed out that we employ the Sherman-Morrison-Woodbury formula [29] for the inversion of matrix and, hence, need only to invert matrix with a lower order , instead of the order . Further, in practise, if the number of patterns in the th classe is large, then the rectangular kernel technique [30, 31] can be applied to reduce the dimensionality of our nonlinear classifiers.

##### 3.4. Several Others Approaches

In this section, we briefly give several extension versions based on our framework by selecting different loss function or replacing 2-norm.

First, if the square loss function is chosen, that is, , then we can get the following formulation from the framework (17):where the matrix is comprised of the patterns in the th class and the matrix is defined by (12), and are parameters, and are vectors of ones of appropriate dimensions, and . This is extension version of LS-TWSVM [14].

Second, if we replace 2-norm with 1-norm in the problem (39), then we can get the extension of 1-norm LS-TWSVM [17] as follows:where the matrix is comprised of the patterns in the th class, the matrix is defined by (12), and are parameters, and are vectors of ones of appropriate dimensions, and .

Third, if the loss function is a convex combination of linear and square loss, that is, , where , then we can obtain extension version of NPPC [18] as follows:where the matrix is comprised of the patterns in the th class, the matrix is defined by (12), and are parameters, and are vectors of ones of appropriate dimensions, and .

Forth, if the square hinge loss function is selected, that is, , then we can get the extension version of smooth TWSVM as follows:where the matrix is comprised of the patterns in the th class, the matrix is defined by (12), and are parameters, and are vectors of ones of appropriate dimensions, and .

These approaches have the same decision function (18) and can be extended into nonlinear case. And their solving methods can construct based on their binary algorithms.

#### 4. Numerical Experiments

In this section, we present experimental results of our binary HNPSVM (BHNPSVM) and multiclass HNPSVM (MHNPSVM) on both artificial and benchmark datasets. In experiments, we focus on the comparison between our methods and some state-of-the-art classification methods, including SVM, GEPSVM, TWSVM, “1-v-1,” “1-v-r,” and MBSVM. All the classification methods are implemented in MATLAB 7.0 [32] environment on a PC with Intel P4 processor (2.9 GHz) with 1 GB RAM. In order to give the fastest training speed, we employ Libsvm [33] to implement the SVM, “1-v-1,” and “1-v-r”. Our BHNPSVM and MHNPSVM and TWSVM and MBSVM are implemented using SOR technique; GEPSVM is implemented by simple MATLAB functions like “eig,” respectively. As for the problem of selecting parameters, we employ standard 10-fold cross-validation technique [34]. Furthermore, the parameters for all methods are selected from the set .

##### 4.1. Toy Examples

Firstly, we consider a simple two-dimensional “Cross Planes” dataset as Example 1, which was tested in [11, 13] to indicate that nonparallel hyperplanes classifiers can handle the cross planes dataset much better compared with parallel ones. Now, we show that our BHNPSVM also can handle cross-planes type data well due to use of our decision function. The “Cross Planes” dataset is generated by perturbing points lying on two intersecting lines. Figures 1(a)–1(d) show the dataset and the linear classifiers obtained by SVM, GEPSVM, TWSVM, and our BNPSVM. It is easy to see that the result of our BNPSVM is more reasonable than that of SVM, and better than that of GEPSVM and TWSVM. In addition, we list the accuracy and CPU time for these four classifiers in Table 1. From Table 1, we can see that our BNPSVM obtains the best accuracy while not the slowest computing time.