Scientific Programming

Scientific Programming / 2021 / Article

Research Article | Open Access

Volume 2021 |Article ID 5560465 | https://doi.org/10.1155/2021/5560465

Wenhao Xie, Yanhong She, Qiao Guo, "Research on Multiple Classification Based on Improved SVM Algorithm for Balanced Binary Decision Tree", Scientific Programming, vol. 2021, Article ID 5560465, 11 pages, 2021. https://doi.org/10.1155/2021/5560465

Research on Multiple Classification Based on Improved SVM Algorithm for Balanced Binary Decision Tree

Academic Editor: Michele Risi
Received08 Jan 2021
Accepted22 Mar 2021
Published05 Apr 2021

Abstract

Support vector machines (SVMs) are designed to solve the binary classification problems at the beginning, but in the real world, there are a lot of multiclassification cases. The multiclassification methods based on SVM are mainly divided into the direct methods and the indirect methods, in which the indirect methods, which consist of multiple binary classifiers integrated in accordance with certain rules to form the multiclassification model, are the most commonly used multiclassification methods at present. In this paper, an improved multiclassification algorithm based on the balanced binary decision tree is proposed, which is called the IBDT-SVM algorithm. In this algorithm, it considers not only the influence of “between-classes distance” and “class variance” in traditional measures of between-classes separability but also takes “between-classes variance” into consideration and proposes a new improved “between-classes separability measure.” Based on the new “between-classes separability measure,” it finds out the two classes with the largest between-classes separability measure and uses them as the positive and negative samples to train and learn the classifier. After that, according to the principle of the class-grouping-by-majority, the remaining classes are close to these two classes and merged into the positive samples and the negative samples to train SVM classifier again. For the samples with uneven distribution or sparse distribution, this method can avoid the error caused by the shortest canter distance classification method and overcome the “error accumulation” problem existing in traditional binary decision tree to the greatest extent so as to obtain a better classifier. According to the above algorithm, each layer node of the decision tree is traversed until the output classification result is a single-class label. The experimental results show that the IBDT-SVM algorithm proposed in this paper can achieve better classification accuracy and effectiveness for multiple classification problems.

1. The Introduction of Support Vector Classifier

Setting a set, which includes m samples: , in which . If the hyperplane can separate the samples of this set exactly and the distance from the nearest sample to the hyperplane is the maximum, this hyperplane is called the optimal hyperplane, which is also called the maximum margin hyperplane (see Figure 1).

Different from other classification algorithms, SVM can well solve the problems of data classification in high-dimensional space by the kernel functions. It maps the samples from the input space to the higher dimensional space by this mapping: , and it can construct the optimal classification hyperplane to separate the samples in the high-dimensional feature space. In this way, the problem above is transformed into the following optimization problem:in which C is the penalty parameter, which controls the degree of penalty for the misclassification samples. In addition, the greater the value of C, the greater the penalty for error.

It converts the above problems into the following dual problem:in which is the Lagrange multiplier and is the kernel function, which represents the inner product of and . If is positive definite or semidefinite, formula (2) is a convex quadratic programming problem. And the final decision function is as follows:

2. The Analysis of the Multiclassification SVM Algorithms

2.1. An Overview of the Multiclassification SVM Methods

Currently, there are two main multiclassification methods for SVM including the direct method and the indirect methods [2, 3]. The principles of these methods are briefly introduced as follows:(1)The direct method directly modifies the quadratic programming form of SVM, designs the multiple objective functions, solves a large quadratic programming problem based on the SVM model, and completes the multiple classification problems at one time.Let be the training set, which meets the conditions: , . It will look for a decision function: and implements the mapping from to when .By modifying the quadratic programming equation, it can obtain the objective function of the optimization problem based on SVM multiclassification as follows [4]:in which is the multiclassification label corresponding to the sample .This algorithm makes a direct improvement on the original SVM classification model, and the theory is simple. However, due to the large number of variables, the solution process of the objective function is difficult, especially in the case of a large number of samples, which will greatly increase the difficulty of calculation and solution and increase the training time. At the same time, compared with other algorithms, this method does not have a good classification accuracy, so it is not widely used in practice.(2)By contrast, the indirect methods design multiple binary classifiers first and then sort and combine the samples according to some rules so as to realize the multilabel identification for them [5]. The advantage of the indirect methods is obvious. The principle of the methods is to integrate multiple binary classifiers, each classification is equivalent to training one binary classifier separately, and processing number of the support vectors and training time are greatly reduced in every time, so it has strong operability and practicability. In recent years, the indirect methods which use the integrated binary classifiers have become the research hotspots of the SVM multiclassification research.

2.2. The Introduction for the Binary Decision Tree Method Based on SVM

There are many indirect classification methods based on SVM: one-versus-one method, that is, OVO-SVM method [6]; one-versus-all method, that is, OVA-SVM [7, 8]; directed acyclic graph method, that is, DAG-SVM method [9, 10], error-correcting output codes method [11, 12], binary decision tree method [13, 14], and hierarchical multiclass SVM algorithm [15, 16], in which the binary decision tree method is divided into the balanced binary decision tree and the imbalanced binary decision tree.

The principle of the SVM multiclass classification method based on Decision Tree (DT) is to construct a decision tree so that each layer can separate one or more classes from the rest of the classes. The process goes on and on until it determines the final category of the samples by the leaf node. For classes of samples, we can achieve the classification of samples by training only decision planes.

Different construction methods can be chosen to construct the different decision trees. If the number of two classes of nodes is equal in each classification, such a decision tree structure is called a balanced decision tree. A decision tree with an imbalanced number of nodes on both sides is called an imbalanced decision tree (see Figure 2). Obviously, in the process of classification, the imbalanced binary tree selects one class and the rest classes to be divided every time, while the balanced binary tree selects multiple classes and multiple classes to be divided each time.

For M multiclassification problems, the binary decision tree only needs to train M 1 classifiers, but the OVO and directed acyclic graph need to train M (M 1)/2 classifiers. In contrast, the former can save the repeated training time. At the same time, when using the binary decision tree method, the judgment of class be a single class, and there is no case that an object outputs multiple class labels or no judgment result. So, it avoids the indivisibility of the problems of the OVA and OVO algorithm. However, the binary decision tree algorithm also has its own drawbacks, and its classification effect is affected by the binary tree structure. In the binary tree classification, the “error accumulation” phenomenon often occurs; that is, the error occurring at a certain node will spread to the next layer of nodes, making the classification error of the next level further expand or even lose the meaning of classification. Therefore, the higher the node where the error occurs, that is, the earlier it occurs, the larger the scope of the error’s influence and the worse the classifier’s effect.

In general, the first step of the method of a binary tree based on SVM is to construct between-classes separability measure, then find out the two classes or two class combinations with the largest measure value, identify one or a group as a positive class and another one or a group as a negative class, and train the classification decision hyperplane through SVM; in the same way, the lower nodes are grouped and trained through the between-classes separability measure until the leaf nodes are output.

2.3. Between-Classes Separability Measure

As mentioned above, the “error accumulation” phenomenon is easy to occur in the binary tree multiclassification algorithm. In order to solve this problem better, many scholars have done a lot of work: literature [17, 18] attempt to construct a more reasonable decision tree by improving the measure of separation between classes. In reference [19], the equivalent distance is used as the separability measure. During the training, it grows from the leaves to the root. Each time the classifier is trained, two local class clusters with the minimum equivalent distance are found for separation. In order to further improve the classification performance of decision tree nodes, literature [20] proposes an improved dual-support vector machine algorithm. Literature [21] proposes a SVM multiclassification decision tree based on the cumulative fitness genetic algorithm. This algorithm constructs a binary tree layer by layer from the root node, selects a new population, and searches for the global optimal solution to classify training samples by redefining the genetic fitness function.

For the multiclassification decision tree, the earlier the classification error occurs and the closer to the upper nodes, the worse the classification effect. Therefore, to construct a reasonable algorithm and tree structure, it is necessary to first separate the classes that are not easily misclassified. In other words, the closer the nodes are to the upper layers, the greater the difference between the two classes should be. In this way, the occurrence of the classification errors can be controlled as far as possible away from the root nodes. Therefore, separating the classes with the greatest differences as early as possible is the key to implementing the above steps. In order to achieve this goal, it is necessary to construct a reasonable between-classes separability measure.

2.3.1. The Introduction of between-Classes Separability Measure

The between-classes separability measure refers to the degree of separation between different classes. In the feature space, each type of data object corresponds to a class field, which is the minimum convex set containing such data objects. The best separation of two class fields means that there is no intersection between them. In addition, if the two classes have overlapping regions, the smaller the proportion of data elements in the overlapping region to the total data elements, the easier the two classes of data are to be separated and the better the separability is [22]. As the sample space shown in Figure 3, class 2 has the best separability because it does not intersect with the other classes. However, both class 3 and class 4 intersect with class 1 and the sample proportion of intersection area between class 4 and class 1 is larger than that between class 3 and class 1. Therefore, class 3 has a better separability than class 4. Through the above analysis of the separability between-classes, we can first separate class 2 with the best separability from the rest classes to train a SVM binary classifier; then choose class 3 with better separability and separate it from class 1 and 4; and finally, separate class 1 and 4. The imbalanced binary tree constructed in accordance with the above ideas is shown in Figure 4.

It can be seen that for the SVM multiclassification methods based on decision tree, constructing a reasonable separability measure has an important impact on the classification performance of the decision tree, and it is also the focus of current research. Therefore, it is the key to reduce the “error accumulation,” improve the classification accuracy, and improve the classification efficiency by separating the easily distinguishable classes first and reducing the error rate of the upper nodes as much as possible.

2.3.2. The Description of the Existing Algorithms

The binary tree classification algorithm based on SVM proposed in literature [23] uses the minimum distance method as the between-classes separability measure. This algorithm first finds the two classes of samples with the longest distance between classes to form two different datasets, while the rest of the classes classify the two sets according to the nearest principle and finally forms two categories of samples, positive and negative, for classifier training. Then, the samples of each class are recursively recurred according to the above algorithm until each output is a single category. This algorithm is called BDT-SVM algorithm. Literature [17] defines the separation measure by “the distance between the center of two types of samples and the ratio of the variance sum of the two types of samples themselves.” The higher the value is, the higher the degree of separation is. In this paper, this algorithm is called VBDT-SVM algorithm. In literature [24], the separability measure proposed in literature [17] is used as fitness function to optimize the decision tree with the genetic algorithm. Different classification strategies are adopted according to the degree of node separability measure in classification, and better results are obtained.

The above algorithms construct the between-classes separability measure from different angles, and both of them achieve good classification effect. However, there are still some shortcomings, which are shown as follows:(1)The BDT-SVM algorithm only considers the distance between classes as the index of class separation degree. Although it is simple, it ignores the distribution characteristics within samples. In addition to the influence of the distance between classes, the distribution characteristics of samples also affect the degree of separation between classes. Literature [17] points out that when the distance between the class are same, the degree of separation between the classes is closely related to their own distribution, and the degree of separation between the classes is inversely proportional to the dispersion degree of the sample’s own distribution. Therefore, the VBDT-SVM algorithm proposed in literature [17] defines the measure of between-classes separability as the ratio of the distance between the center point of two types of samples and the sum of the variance of the two types of samples themselves (see formula (5)). The higher the ratio is, the easier the separation of the two classes is.where represents the distance between the samples of class i and class j, and represent the class variances of class i and class j, respectively (see formula (9)), and represent the distribution tightness of the two classes of samples themselves.(2)Formula (5) not only considers the distance of class centers also considers the class variance, which is the discrete degree of the distribution of the samples in the same class. However, the algorithm still exists deficiencies; for example, it does not consider that when the between-classes distance and class variance are same, the separability of classes also depends on the distribution tightness of one class objects relative to the other classes of objects, that is, the between-classes variance (see Definition 1).

3. The Design of IBDT-SVM Algorithm

In summary, in order to construct a more reasonable binary decision tree, it is necessary to solve the following problems:(1)It should construct a new between-classes separability measure and find the two classes with the greatest difference according to the between-classes separability measure at the upper node as far as possible so as to avoid bringing the errors to the nodes of the next layer.(2)It should construct a new building-tree scheme and a reasonable classifier as close to the root node as possible. In this way, it is necessary to carry out a second revision for each classifier in order to improve the classification accuracy and avoid the “error accumulation” to the greatest extent.

Based on the above considerations, this improved algorithm divides the samples with the greatest difference step by step according to the new between-classes separability measure to train the classifier and then classifies the rest classes according to the class-grouping-by-majority principle to form the new training samples and trains the classifier again. Thus, an improved SVM multiple classification algorithm based balanced binary decision tree, known as IBDT-SVM algorithm, is obtained.

3.1. The Improved Between-Classes Separability Measure

In order to solve the problems existing in the traditional between-classes separability measure, inspired by three decision-making of clustering thought [25], this paper proposes between-classes separability measure about q neighbors, and the new between-classes separability measure mainly considers the following three factors as follows. (1) The Between-Classes Variance. Starting from the number and distance between the samples, it reflects the closeness of the relationship between a certain class of objects and its neighboring classes. It is inversely proportional to the between-classes separability measure; that is, the smaller the value, the greater the separability measure of the two classes of samples. (2) Class Variance. It reflects the compactness of distribution of samples of the class itself and is inversely proportional to the between-classes separability measure; that is, the smaller the value, the greater the separation of the class from other classes. (3) Between-Classes Distance. It is the distance between two class centers and is proportional to the between-classes separability measure; the greater the value is, the greater the separability of the two classes of samples is.

Definition 1. (between-classes variance). Considering one sample’s q neighbors in the other neighbor class, its value indicates the degree of separation of one class objects from another class.(1)The degree of separation between object in class and Class is expressed as follows:where represents the nearest neighbors of the object ; represents the sum of the distances between and the objects that are in class but also belong to .(2)Separability between all objects in class and class is as follows:Similarly, separability between all objects in class and class is as follows:where

Definition 2. (class variance). Suppose class has samples and the center is , then ; suppose class has samples and the center is , then ; and are defined as class variances of class and class , respectively; that is:On the basis of considering the between-classes variance, class variance, and between-classes distance, it improves the separability measure function again, as defined in the following.

Definition 3. (between-classes separability measure). The degree of separation between class and class is defined as follows:where represents the distance between the centers of class and class , that is, the between-classes distance.

3.2. Class Grouping Algorithm Based on the Principle of Class-Grouping-by-Majority

The IBDT-SVM multiclassification algorithm proposed in this paper improves the between-classes separability measure on the basis of considering the between-classes distance, class variance, and between-classes variance. According to the improved between-classes separability measure, the two classes with the highest separability measure are first found and the classification model is trained. Then, it uses the class-grouping-by-majority principle [26] to group the other classes into these two groups and use them as the training samples to retrain the classifier. This algorithm loops on each decision surface until each output is a leaf node, that is, a single sample point. This method can ensure that all classes can be separated as far as possible at each classification node and the classification of the rest classes as reasonable as possible. For the data with uneven or sparse distribution, the error caused by the classification method of minimum distance between class centers can be avoided.

3.2.1. The Introduction of the Class-Grouping-by-Majority Principle

Now take the classification of two-dimensional samples of class “1,” “2,” and “3” (see Figure 5) as an example and introduce the principle of class-grouping-by-majority. If class “1” and class “2” are used as training sets to train the SVM classifier, the samples after classification are scattered on both sides of the initial classifier (see Figure 5(a)). For samples of class “3,” they are distributed to both sides of the initial classifier, but the class “2” side contains most of the samples of class “3.” According to the class-grouping-by-majority algorithm, all examples of class “3” and class “2” were reclassified into one class, and the examples of class “1” are classified as another class samples. In this way, the SVM classifier is retrained and the final classifier is obtained (see Figure 5(b)). The structure of the finial binary decision tree is shown in Figure 5(c).

3.2.2. Class Grouping Algorithm Based on the Principle of Class-Grouping-by-Majority

The idea of the IBDT-SVM algorithm is that on each layer of the decision tree first finding the two classes based on the improved maximum between-classes separability measure mentioned above which have the greatest difference, namely, the two classes that are easiest to be separated. It denotes the two classes as class and class . It takes the two classes’ samples as the positive and negative samples to train the classifier, denoted as old-classifier. Then, based on the principle of class-grouping-by-majority, it groups the other classes closer to the two classes to form the two new classes: and . Then, it takes the two classes: class and class as the positive and negative samples to train the SVM classifier again, denoted as new-classifier. The cycle continues until all classification results are single class.

Take a 4-classes classification problem as an example (see Figure 6). Suppose that among the four classes, class 1 and class 2 have the maximum separability measure, we find out class 1 and class 2 first and take them as the positive and negative samples, respectively, to train the classifier, that is, old-classifier. Then, it puts the training samples of class 3 and class 4 into the classifier for testing according to the class-grouping-by-majority principle. There are four possibilities:(1)If most of the samples of class 3 are on the positive side of the old-classifier, while most of the samples of class 4 are on the negative side of the old-classifier, then the classified result is {1, 3}, {2, 4} (see Figure 6(a))(2)If most of the samples of class 4 are on the positive side of the old-classifier, while most of the samples of class 3 are on the negative side of old-classifier, the classification result is {1, 4}, {2, 3} (see Figure 6(b))(3)If most of the samples of the two classes such as class 3 and class 4 are all biased to the positive side of the old-classifier, then the classification result is {1, 3, 4}, {2} (see Figure 6(c))(4)If most of the samples of the two classes such as class 3 and class 4 are all biased to the negative side of the old-classifier, then the classification result is {1}, {2, 3, 4} (see Figure 6(d))

For the above situations, the reclassified samples need to be treated as the positive and negative objects to retrain the classifier, and the new-classifier is obtained (see Figure 7).

3.3. IBDT Algorithm

The idea of the improved binary decision tree algorithm called as IBDT algorithm in this paper is as follows.

Assume that there are M classes of input data, and class i contains samples, then . The IBDT algorithm process is as follows.

Step 1. Set the initial q value and calculate the separability measure values of every two classes of samples in M class which is calculated according to formulas (6)–(10), namely, the value of .

Step 2. Set the two classes which have the maximum between-classes separability measure be class and class and then find them out;

Step 3. The classifier is trained by using class and class as the training samples, denoted as old-classifier.

Step 4. According to the principle of the class-grouping-by-majority, the remaining class samples are classified into class and class , respectively, and form two major classes and , and then the two major class samples are marked as the positive samples and negative samples, training the SVM classifier to form the classifier, denoted as new-classifier.

Step 5. Repeat the above operations until the each outputting sample is labeled as a single category.

4. The Numerical Experiments and Results

We used five multiclass datasets in the UCI database such as Segmentation, Statlog, Iris, Breast tissue, and Page Blocks to verify the classification accuracy of the improved algorithm proposed in this paper–IBDT-SVM. We use the tenfold cross-validation algorithm to select the training set and the test set. Each time a subset is selected as the test set, and the average cross-validation recognition accuracy of 10 times is taken as the last result.

The above five datasets are classified using the following algorithms including OVO, OVA, BDT-SVM, VBDT-SVM algorithm, and the IBDT-SVM algorithm proposed in this paper, and the classification accuracy is compared.

In this study, we used five multiclass datasets from the UCI database, with different sample sizes, number of attribute values, and number of categories. The full name and corresponding abbreviation of the dataset used in this experiment are shown in Table 1, and the characteristics of the above multiclass datasets are shown in Table 2, including data set name, sample number, number of attributes of samples (data dimension), and number of categories.


Dataset IDDataset namesDataset abbreviations

1Image segmentationSegmentation
2Statlog (Landsat satellite)Statlog
3IrisIris
4Breast tissueBreast tissue
5Page blocks classificationPage Blocks


Dataset IDDataset abbreviationsSample sizesNumber of attribute valuesNumber of categories

1Segmentation2310197
2Statlog6435367
3Iris15043
4Breast tissue106106
5Page Blocks5473105

The classification accuracy of 10 experiments using the OVO algorithm, OVA algorithm, BDT-SVM algorithm, VBDT-SVM algorithm, and IBDT-SVM algorithm for the above 5 datasets is shown in Tables 37). The average classification accuracy of the above results is shown in Table 8). It can be seen that the IBDT-SVM algorithm has a good classification effect.


12345678910

OVO0.9230.9610.9610.9570.9570.9650.9610.9570.9570.960
OVA0.8800.9400.9440.9350.9350.9480.9390.9350.9390.943
BDT-SVM0.9440.9870.9720.9870.9740.9850.9810.9870.9780.991
VBDT-SVM0.9740.9760.9890.9810.9830.9810.9870.9870.9810.985
IBDT-SVM0.9870.9890.9910.9910.9910.9720.9780.9780.9850.985


12345678910

OVO0.9670.9480.9060.8330.9450.9430.9690.9620.9450.896
OVA0.9220.8910.8570.7570.8780.8880.9230.9160.8870.839
BDT-SVM0.9610.9610.9550.8480.9530.9560.9470.9460.9410.935
VBDT-SVM0.9610.9390.9560.9500.9570.9520.9520.9500.9440.945
IBDT-SVM0.9400.9460.9570.9570.9520.9610.9530.9450.9550.952


12345678910

OVO1.000.9330.9701.000.9330.9330.9331.001.001.00
OVA0.9330.8671.0000.9330.9330.9330.8670.9331.01.0
BDT-SVM0.9330.9330.9670.900.9670.9330.9670.9000.9000.967
VBDT-SVM0.9670.9671.0000.9331.0000.9330.9000.9330.9000.900
IBDT-SVM0.9721.0000.9670.9800.9670.9670.9670.9330.9670.980


12345678910

OVO0.2860.3850.4170.3330.3640.3000.5560.5560.6250.625
OVA0.2860.3080.3330.3330.2730.3000.2220.3330.3750.375
BDT-SVM0.8100.7620.7620.6670.5710.6190.7140.6670.8100.714
VBDT-SVM0.5710.7620.5710.7140.5710.5710.6190.4290.5710.762
IBDT-SVM0.7620.7140.7140.7140.8100.8100.6190.6670.8570.571


12345678910

OVO0.9440.9470.9360.9320.9420.9490.9290.8850.9360.938
OVA0.9420.9540.9400.9360.9420.9480.9340.8790.9340.945
BDT-SVM0.9150.9000.9110.9150.9060.9260.9230.9150.9000.909
VBDT-SVM0.9100.9050.9140.8880.8980.9170.9160.9190.9210.899
IBDT-SVM0.9280.9180.9280.9500.9300.9280.9500.9300.9280.960


Data setsOVOOVABDT-SVMVBDT-SVMIBDT-SVM

Segmentation0.9560.9340.9790.9820.985
Statlog0.9310.8760.9400.9500.952
Iris0.9700.9400.9370.9430.970
Breast tissue0.4450.3140.7100.6140.724
Page Blocks0.9340.9350.9120.9090.935

The comparison results of the classification accuracy of OVO algorithm, OVA algorithm, BDT-SVM algorithm, VBDT-SVM algorithm, and IBDT-SVM algorithm for the above five different datasets in 10 experiments are shown in Figures 8(a)8(e). It can be seen from Figure 8, compared with the other four multiclassification algorithms, the IBDT-SVM algorithm proposed in this paper has better stability and higher classification accuracy. It shows the average classification accuracy of the above five algorithms by a bar chart (see Figure 9).

5. Conclusion

In this paper, an improved multiclassification algorithm based on the balanced binary decision tree, called IBDT-SVM algorithm, is proposed. This algorithm improves the original BDT-SVM algorithm’s between-classes separability measure so that the calculation of separability between-classes not only considers the distance between classes but also considers the class variance and the variance between-classes. According to the improved between-classes separability measure, the two classes with the greatest difference will be used to train the old classifier. In addition, considering the classification error of nonuniform data or sparse data using the minimum center distance classification method, the class-grouping-by-majority principle is used to classify the rest samples, and the new-classifier is trained again so as to ensure the better classification effect.

The algorithm proposed by this paper is simple and easy, its need to construct the classifier of the number is less, and the experiments show that it can improve the classification accuracy than the original binary tree classification algorithm. In the future work, we also can consider to construct a better algorithm or decision tree structure, such as considering through the improved SVM algorithm to further optimize the decision tree classification mechanism; constructing the SVM decision tree classification based on the distance between each center and root node; or combining the advantages of other algorithms to solve the classification accuracy of multiclassification, such as forming a multiclassification model through rough set theory with complementary advantages.

Data Availability

The imbalanced dataset data used to support the findings of this study are deposited in UCI dataset and KEEL dataset, and they are available openly (the URL of the UCI dataset is http://archive.ics.uci.edu/ml/datasets.php and the URL of the KEEL dataset is https://sci2s.ugr.es/keel/dataset.php).

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (project no. 61976244) and the Special project of Shaanxi Provincial Department of Education (project no. JK190646).

References

  1. W. H. Xie, G. Q. Liang, Z. H. Dong et al., “An improved oversampling algorithm based on the samples’ selection strategy for classifying imbalanced data,” Mathematical Problems In Engineering, vol. 2019, Article ID 3526539, 13 pages, 2019. View at: Publisher Site | Google Scholar
  2. P. X. Yang and P. C. Zhang, “Distributed generation system interconnection protection based on multi-class SVM,” Automation of Electric Power Systems, vol. 39, no. 22, pp. 110–151, 2015. View at: Google Scholar
  3. D. Li, W. Yang, and S. Wang, “Classification of foreign fibers in cotton lint using machine vision and multi-class support vector machine,” Computers and Electronics in Agriculture, vol. 74, no. 2, pp. 274–279, 2010. View at: Publisher Site | Google Scholar
  4. E. J. Bredensteiner and K. P. Bennett, “Multicategory classification by support vector machines,” Computational Optimization, vol. 12, no. 3, pp. 53–79, 1999. View at: Publisher Site | Google Scholar
  5. E. Allwein, R. E. Schapire, and Y. Singer, “Reducing multiclass to binary: a unifying approach for margin classifiers,” Journal of Machine Learning Research, vol. 1, no. 2, pp. 113-114, 2000. View at: Google Scholar
  6. R. Debnath, N. Takahide, and H. Takahashi, “A decision based one-against-one method for multi-class support vector machine,” Pattern Analysis and Application, vol. 7, no. 2, pp. 164–175, 2004. View at: Publisher Site | Google Scholar
  7. R. Rifkin and A. Clautau, “In defense of one-vs-all classification,” Journal of Machine Learning Research, vol. 2004, no. 5, pp. 101–141, 2004. View at: Google Scholar
  8. L. Xu, G. Z. Zhao, and H. Gu, “Novel one-vs-rest classifier based on SVM and multi-spheres,” Journal of Zhejiang University, vol. 43, no. 2, pp. 303–308, 2009. View at: Google Scholar
  9. J. C. Platt, N. Cristianini, and J. Shawe-Taylor, “Large margin DAGs for multi classification,” Advances in Neural Information Processing Systems, vol. 12, pp. 547–553, 2000. View at: Google Scholar
  10. C. Zhang, Z. F. Ma, X. X. Niu et al., “A method based on dynamic sequence analysis of an API and DAG-SVM multi-class support vector machines of unknown virus detection,” Journal of Chinese Computer Systems, vol. 33, no. 12, pp. 2724–2728, 2012. View at: Google Scholar
  11. O. Pujol, P. Radeva, and J. Vitrià, “Discriminant ECOC: a heuristic method for application dependent design of error correcting output codes,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, no. 6, pp. 1007–1012, 2006. View at: Publisher Site | Google Scholar
  12. T. G. Dietterich and G. Bakiri, “Solving multiclass learning problems via error-correcting output codes,” Journal of Artificial Intelligence Research, vol. 2, pp. 263–286, 1995. View at: Publisher Site | Google Scholar
  13. D. Gjorgjevikj, “A multi-class SVM classifier utilizing binary decision tree,” Informatica, vol. 33, no. 2, pp. 225–233, 2009. View at: Google Scholar
  14. Z. H. Diao, C. J. Zhao, X. Y. Guo et al., “A new SVM muli-class classification algorithm based on balance decision tree,” Control and Decision, vol. 26, no. 1, pp. 149–156, 2011. View at: Google Scholar
  15. R. Y. Chen and J. F. Jiang, “Pose-varied face recognition based on neural network and hierarchical support vector machines,” Computer Engineering, vol. 32, no. 24, pp. 210–215, 2006. View at: Google Scholar
  16. B. Lu and J. B. Yang, “A strategy of multi-level classification based on SVM,” Computer Engineering, vol. 31, no. 1, p. 113, 2005. View at: Google Scholar
  17. S. Y. Xia, H. Pan, and L. Z. Jin, “Multi-class SVM method based on non-balanced binary tree,” Computer Engineering and Applications, vol. 45, no. 17, pp. 167–169, 2009. View at: Google Scholar
  18. X. R. Li, G. Y. Zhao, and L. Y. Zhao, “Design of decision-tree-based support vector machines multi-class classifier based on vector project,” Journal of Control and Decision, vol. 23, no. 7, pp. 745–750, 2008. View at: Google Scholar
  19. R. Wang and J. Chen, “Application of support vector machine-based decision tree classifier in rotor fault diagnosis,” Journal of Vibration and Impact, vol. 29, no. s, pp. 258–260, 2010. View at: Google Scholar
  20. J. Y. Xie, B. Q. Zhang, and W. Z. Wang, “A partial binary tree algorithm for multiclass classification based on twin support vector machines,” Journal of the Nanjing University (Natural Science), vol. 47, no. 4, pp. 354–363, 2011. View at: Google Scholar
  21. Q. S. Zhu and K. Cheng, “SVM decision-tree multi-classification strategy based on genetic algorithm with cumulative fitness,” Computer Application Research, vol. 33, no. 291, pp. 64–67, 2016. View at: Google Scholar
  22. F. Y. Sun, Study on Multi-Class Classification Method of Support Machine Based on Decision Tree, Northeast Normal University, Changchun, China, 2015.
  23. E. Y. Wu and J. Lv, “Research on multiclass classification algorithm based on binary tree SVM,” Journal of Chongqing Normal University (Natural Science), vol. 33, no. 3, pp. 102–106, 2016. View at: Google Scholar
  24. S. F. Ding, J. G. Sun, and D. L. Chen, “Improved SVM decision-tree and its application in remote sensing classification,” Application Research of Computers, vol. 29, no. 3, pp. 1146–1148, 2012. View at: Google Scholar
  25. H. Yu and C. K. Mao, “Automatic three-way decision clustering algorithm based on k-means,” Computer Applications, vol. 36, no. 8, pp. 2061–2065, 2016. View at: Google Scholar
  26. P. Kantavat, B. Kijsirikul, P. Songsiri, K.-I. Fukui, and M. Numao, “Efficient decision trees for multi-class support vector machines using entropy and generalization error estimation,” International Journal of Applied Mathematics and Computer Science, vol. 28, no. 4, pp. 705–717, 2018. View at: Publisher Site | Google Scholar

Copyright © 2021 Wenhao Xie et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views357
Downloads283
Citations

Related articles

Article of the Year Award: Outstanding research contributions of 2020, as selected by our Chief Editors. Read the winning articles.