Authors MA (category) Other Methods fitness function Remarks Standard MA Desai et al. (1997) [60 ] GA (EA) - no. of misclassification (i) multiclass problem (ii) compare with NN, LOGIT, DA Ong et al. (2005) [61 ] GP (EA) - mean absolute error (i) rule extraction (ii) compare with LOGIT, NN, TREE, rough sets Finlay (2006) [62 ] GA (EA) - GINI coefficient (i) large credit data application (ii) compare with LOGIT, LR, NN Cai et al. (2009) [63 ] GA (EA) - error rate (i) include misclassification cost into fitness function Huang et al. (2006) [64 ] 2stage GP (EA) - mean absolute error (i) rules extraction (ii) compare with LOGIT, TREE, kNN, GP Abdou et al. (2009) [65 ] GP (EA) - (sum of square error) +(i) study effect of misclassification costs (classification error)(ii) compare with profit analysis and weight-of-evidence measure Dong et al. (2009) [66 ] SA (IB) - accuracy similarity function (i) rules extraction ( is a coefficient)(ii) compare with DA, kNN, TREE Martens et al. (2010) [67 ] ACO (SI) - coverage + confidence (i) rules extraction (ii) compare with TREE, SVM, majority vote Yang et al. (2012) [68 ] CS (SI) - error rates (i) compare with Altman’s Z-score and SVM Uthayakumar et al. (2017) [69 ] ACO (SI) - (i) rules extraction (ii) compare with LOGIT, NN, RF, RBF network Hybrid MA-MA Jiang et al. (2011) [70 ] SA+GA NN (i) SA optimizes GA (IB+EA) (ii) parameter tuning (iii) compare with standalone NN and GA-optimized NN Aliehyaei and Khan (2014) [71 ] ACO (SI), GP (EA), - mean absolute error (i) rules extraction by ACO to input to GP ACO+GP (ii) compare with ACO and GP Hybrid MA-DM Fogarty and Ireson (1993) [72 ] GA (EA) BN accuracy (i) features extraction (ii) large credit data application (iii) compare with default rule, BN, kNN, TREE Hand and Adams (2000) [73 ] SA (IB) WOE, weighted likelihood (i) features discretization WOE (ii) compare with LOGIT, DA, two other discretization methods Drezner et al. (2001) [74 ] TS (IB) LR (i) features selection (wrapper approach) (ii) compare with Altman’s Z-score Lacerda et al. (2005) [75 ] GA (EA) NN average of individuals with (i) hyperparameters tuning rank of individual (ii) compare with NN, consecutive learning algorithm, SVM from a population Huang et al. (2007) [35 ] GA (EA) SVM accuracy (i) hyperparameters tuning, feature selection (wrapper approach) (ii) compare with NN, GP, TREE Zhang et al. (2008) [38 ] GP (EA) SVM accuracy+expected (i) rules extraction misclassification cost (ii) compare with SVM, GP, TREE, LOGIT, NN Liu et al. (2008) [76 ] GP (EA) DA Information value (i) features extraction (select derived characteristics) (ii) large dataset from finance enterprise (iii) compare with DA Wang et al. (2008) [77 ] GA (EA) NN 1/MSE (i) parameters tuning (ii) compare with NN Zhang et.al (2008) [78 ] GA (EA) TREE (1-info entropy)+(i) features extraction (1-info entropy)(ii) compare with TREE, NN, GP, GA-optimized SVM, rough set Jiang (2009) [79 ] SA (IB) TREE (i) (i) rules extraction (ii) (ii) compare with TREE (iii) Marinakis et al. (2009) [80 ] ACO (SI), kNN, 1-NN, not mentioned (i) multiclass problem, features selection (wrapper approach) PSO (SI) weighted kNN (ii) compare with wrapper models of GA and TS with kNN (and variants) Sustersic et al. (2009) [81 ] GA (EA) NN accuracy and RMSE , (i) feature selection (wrapper approach) , preset threshold(ii) compare with NN, LOGIT (features from PCA) Zhou et al. (2009) [42 ] GA (EA) SVM error rate (i) hyperparameters tuning (ii) compare with LOGIT, kNN, DA, TREE Zhou et al. (2009) [41 ] GA (EA) WSVM AUC (i) hyperparameters tuning, feature selection (wrapper approach) (ii) compare with LR, LOGIT, NN, TREE, kNN, Adaboost Marinaki et al. (2010) [82 ] HBO (SI) kNN, 1-NN, accuracy (i) multiclass problem, features selection (wrapper approach) weighted kNN (ii) compare with wrapper models of GA, PSO, TS, ACO with kNN (and variants) Huang and Wu (2011) [83 ] GA (EA) kNN, BN, TREE, LR, accuracy (i) feature selection (wrapper approach) SVM, NN, Adaboost, (ii) compare with kNN, BN, TREE, LOGIT, SVM (features from Logitboost, Multiboost filter selection approach) Yu et al. (2011) [44 ] GA (EA) LS-SVM accuracy (i) hyperparameters tuning (ii) study class imbalance problem (iii) compare with results from [8 , 45 ] Correa and Gonzalez (2011) [84 ] BPSO (SI), NN AUC (i) hyperparameters tuning GA (EA) (ii) large credit dataset (iii) cost study on application, behavioural and collection scoring (iv) compare with LOGIT, NN, Global Optimum Oreski et al. (2012) [85 ] GA (EA) NN accuracy (i) hyperparameters tuning, feature selection (wrapper approach) (ii) study effect of misclassification costs (iii) compare their proposed model with features from filter selection approach Danenas and Garsva (2012) [47 ] GA (EA),PSO (SI) SVM TPR (i) model selection (ii) hyperparameters tuning Chen et al. (2013) [49 ] ABC (SI) SVM not mentioned (i) multiclass problem (ii) hyperparameters tuning (iii) compare with SVM (tuned with GA and PSO) Garsva and Danenas (2014) [51 ] PSO (SI) SVM (i) TPR, (ii) accuracy (i) model selection (ii) hyperparameters tuning (iii) study class imbalance problem (iv) compare among all SVM, LS-SVM tuned with PSO, DS, SA Oreski and Oreski (2014) [86 ] GA (EA) NN accuracy (i) feature selection (wrapper approach) (ii) compare with standard wrapper-based NN with GA Danenas and Garsva (2015) [52 ] PSO (SI) SVM TPR (i) model selection (ii) hyperparameters tuning (iii) compare with LOGIT, RBF network Wang et al. (2010) [87 ] TS (IB) NN, SVM, LOGIT entropy (i) features selection (filter approach) (ii) compare with NN, SVM, LOGIT with full features Wang et al. (2012) [88 ] SS (SI) NN, TREE, LOGIT entropy (i) features selection (filter approach) (ii) compare with NN, TREE, LOGIT with full features Waad et al. (2014) [89 ] GA (EA) LOGIT, SVM, TREE (i) feature selection (filter approach) (ii) compare with LOGIT, SVM,TREE (features from other filter selection and rank aggregation methods) Hsu et al. (2018) [53 ] ABC (SI) SVM , if (i) multiclass problem , if (ii) hyperparameters tuning (iii) compare with LOGIT and SVM (tuned with GS, GA, PSO) Jadhav et al. (2018) [54 ] GA (EA) SVM, kNN, NB accuracy (i) features selection (wrapper approach) (ii) compare with standalone SVM, kNN, NB and their wrappers with GA Wang et al. (2018) [55 ] multiple population SVM accuracy (i) features selection (wrapper approach) GA (EA) (ii) compare with MPGA-SVM, GA-SVM, SVM Krishnaveni et al. (2018) [90 ] HS (SI) 1-NN accuracy (i) features selection (wrapper approach) (ii) computational time reduction (iii) compare with standalone SVM, TREE, kNN, NB, NN and their wrappers with GA and PSO