- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

ISRN Artificial Intelligence

VolumeΒ 2012Β (2012), Article IDΒ 628496, 10 pages

http://dx.doi.org/10.5402/2012/628496

## Prediction of Ultimate Bearing Capacity of Cohesionless Soils Using Soft Computing Techniques

^{1}Department of Civil Engineering, TKM College of Engineering, Kerala, Kollam 691005, India^{2}Department of Civil Engineering, National Institute of Technology, Kerala, Calicut 673601, India^{3}Department of Civil Engineering, National Institute of Technology, Karnataka, Surathkal, Mangalore 575025, India^{4}Department of Civil Engineering, College of Engineering, Thiruvananthapuram 695016, India

Received 31 July 2011; Accepted 7 September 2011

Academic Editor: M.Β Abbod

Copyright Β© 2012 S. Adarsh et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

This study examines the potential of two soft computing techniques, namely, support vector machines (SVMs) and genetic programming (GP), to predict ultimate bearing capacity of cohesionless soils beneath shallow foundations. The width of footing (), depth of footing (), the length-to-width ratio () of footings, density of soil ( or ), angle of internal friction (), and so forth were used as model input parameters to predict ultimate bearing capacity (). The results of present models were compared with those obtained by three theoretical approaches, artificial neural networks (ANNs), and fuzzy inference system (FIS) reported in the literature. The statistical evaluation of results shows that the presently applied paradigms are better than the theoretical approaches and are competing well with the other soft computing techniques. The performance evaluation of GP model results based on multiple error criteria confirms that GP is very efficient in accurate prediction of ultimate bearing capacity cohesionless soils when compared with other models considered in this study.

#### 1. Introduction

Design of foundations is performed based on two criteria: ultimate bearing capacity and limiting settlement. The ultimate bearing capacity is governed by shear strength of the soil and is estimated by theories proposed by Terzaghi [1], Meyerhof [2], Hansen [3], Vesic [4], and others. However, the different bearing capacity formulae shows wide degree of variability while estimating bearing capacity of dense sand on cohesionless soils. Also the bearing capacities are validated through laboratory studies performed on small-scale models. Due to the βscale effectβ for the large-scale foundations on dense sand, shearing strain show that considerable variation along the slip line and the average mobilized angle of shearing resistance along the slip line are smaller than the maximum value () obtained by plane shear tests [5]. Thus, the use of may lead to an overestimated bearing capacity value for the calculations based on different formulae [1β4].

In the recent past, the use of soft computing techniques has attracted many researchers and applied quite successfully for solving many complex geotechnical engineering problems. Artificial neural networks (ANNs) may probably be the most popular among these tools, applied for prediction of bearing capacity of cohesionless soils [5], bearing capacity of piles, settlement predictions, liquefaction, and slope stability problems [6]. Support vector machines (SVMs) are recent addition to the soft computing family that uses statistical learning theory as the working principle. SVM and its variants are applied for geotechnical problems such as prediction of pile load capacity [7], settlement of foundations [8], slope stability [9], and liquefaction potential [10].

The evolutionary computational techniques may be a better alternative for solving regression problems as they follow an optimization strategy with progressive improvement towards the global optima. They start with possible trial solutions within a decision space, and the search is guided by genetic operators and the principle of βsurvival of the fittestβ [11]. Genetic Algorithm (GA) is one of the most popular and powerful evolutionary optimization technique [11] explored by [12], but it cannot be used to evolve complex models such as equations. This limitation is overcome by Genetic Programming (GP) introduced by Koza [13], which works on the principle of GA. GP writes expressions or computer programs instead of strings in GA. In this paper, SVM and GP are used as alternate paradigms to predict bearing capacity of cohesionless soils under shallow foundations.

#### 2. Support Vector Machine

Support vector machine (SVM) is a relatively recent addition to the family of soft computing techniques evolved from the concept of statistical learning theory explored by Boser et al. [14]. SVM performs the regression by using a set of nonlinear functions that are defined in a high-dimensional space. SVM has been used to solve nonlinear regression problems by the principle of structural risk minimization (SRM), where the risk is measured using Vapnikβs accuracy intensive loss function () [15]. SVM uses a risk function consisting of the empirical error and a regularization term. More details on SRM can be found in Cortes and Vapnik [16]. Considering a set of input-output pairs as training dataset, , where is the input, is the output, is the -dimensional vector space, and is the one-dimensional vector space. In this problem, the width of footing (), depth of footing (), the length-to-width ratio () of footings, density of soil ( or ) angle of internal friction (), and so forth were used as model input parameters to predict ultimate bearing capacity (). Hence, for this problem, and .

The intension of SVM is to fit a function that can approximately predict the value of output on supplying a new set of predictors (input variables).

The -intensive loss function can be described as follows: otherwise, This defines an -tube so that if the predicted value is within the tube, the loss is zero; otherwise the loss is equal to the absolute value of the deviation minus . This concept is depicted in Figure 1.

SVM attempts to find that a function that gives the deviation of ββ from the actual output is as flat as possible.

Consider a linear function of the form, where is an adjustable weight vector and is the scalar threshold. Fitness means the search for a small value of ββ. It can be represented as a minimization problem with an objective function comprising the Euclidian norm as follows: Some allowance for errors () may also be introduced. Two slack parameters and have been introduced to penalize the samples with error more than ββ. Thus the infeasible constraints of the optimization problem are eliminated. The modified formulation takes the following form: The constant determines the tradeoff between the flatness of and the amount up to which the deviations larger than ββ are tolerated [17]. The above optimization problem is solved by Vapnik [15] using Lagrange multiplier method. The solution is given by where and are known as support vectors and is the number of support vectors.

Some Lagrange multipliers () will be zero, which implies that these training solutions are irrelevant to the final solution (known as sparseness of the solution). The training objects with nonzero Lagrange multipliers are called support vectors. When linear regression is not appropriate, input data have to be mapped into a high-dimensional feature space through nonlinear mapping and the linear regression needs to be performed in the high-dimensional feature space [14]. Kernel function is used to transform nonlinear data from the input to the feature space in linear form. Then linear fitting in new space will be equal to nonlinear fitting in original space: where is the kernel function, and are inputs, and is the dot product in the high-dimensional space.

Thus, (6) can be replaced by The concept of nonlinear mapping is depicted in Figure 2.

The functions which satisfy Mercerβs theorem can be used for fitting the data [14]. Polynomial functions, radial basis function (RBF), and splines are the most commonly used kernel functions for data fitting using SVM. The mathematical forms of some popular kernel functions can be found in [18].

#### 3. Genetic Programming

Genetic Programming (GP) is an automatic programming technique for evolving computer programs to solve, or approximately solve, problems introduced by Koza [13]. GP is basically an optimization paradigm that can also be effectively applied to the genetic symbolic regression (GSR). GSR involves finding a mathematical expression in symbolic form relating finite values of set of independent variables () and a set of dependent variables () [19]. GP works on Darwinβs natural selection theory in evolution. Here, a population is progressively improved by selectively discarding the not-so-fit population and breeding new children to form better populations. Like other evolutionary algorithms, the solution is started with a random population of individuals (equations or computer programs). Each possible solution set can be visualized as a βparse treeβ comprising the terminal set (input variables) and functions (general operators such as +, β, *, /, logarithmic or trigonometric). The βfitnessβ is a measure of how closely a trial solution solves the problem. The objective functionβthe minimization of error between estimated and observed valuesβis the fitness function. The solution set in a population associated with the βbest fitβ individuals will be reproduced more often than the less fit solution sets. It iteratively transforms a population of computer programs into a new generation of programs by applying analogs to naturally occurring genetic operators like reproduction, mutation, and crossover. The different genetic operations can be found in detail in [13]. The basic procedure of GP is presented as a flow chart in Figure 3.

In the recent past, GP is effectively applied to solve a wide range of geotechnical engineering problems [20β22]. GP can evolve an explicit equation or equivalent computer program relating the input and output variables which is a more understandable depiction of the cause-effect process. Some literature suggests that the program-based GP approach (i.e., the GP algorithms which give a computer program which helps for estimating the predictant value for a given set of predictors) can perform equally well with an equation-based approach and other soft computing tools like ANN [23β26]. A program-based GP approach is adopted for the present study.

#### 4. Model Development and Results

The primary step in model development for the estimation of bearing capacity of cohesionless soils underneath shallow foundations is identification of parameters that affect the bearing capacity. The basic form of equation for bearing capacity of cohesionless soil is [5] where is the width of foundation, is the depth of foundation, is the unit weight of sand, , is the bearing capacity factors, , is the shape factors, and , the depth factors. These factors primarily depend on the angle of shearing resistance, unit weight of the sand, and the geometry of the foundation.

The main factors affecting the bearing capacity are its width (least lateral dimension, ), length of footing (), shape (square, rectangular, and circular), and depth of embedment (). The depth of foundation has the greatest effect on the bearing capacity of all the physical properties of the foundation. There are some other factors such as compressibility and thickness of the soil layer beneath the foundation that contribute to a lesser degree [5]. The effect of compressibility is small, except for loose densities, and is generally less important in bearing capacity computation [5]. Moreover, there are insufficient data to consider compressibility as well as thickness of soil stratum. Therefore, they are not considered in this study.

##### 4.1. Database

The data used in the present study has been adopted from Padmini et al. [5]. The five input parameters used for the model development in this study are width of footing (), depth of footing (), footing geometry (), unit weight of sand (), and angle of shearing resistance (). Ultimate bearing capacity () is the single output. The data thus compiled comprises a total of 97 data sets, which consists of results of load test data of square, rectangular, and strip footings of different sizes tested in sand beds of various densities. Out of the total 97 sets of data, 78 are used for training and 19 are used for validation in all the experiments considered in this study. The data division is done in such a way that the same 19 sets of data used by Padmini et al. [5] are kept as the validation dataset to enable a comparison of results of the present study with those obtained by ANN and FIS by Padmini et al. [5].

##### 4.2. Development of SVM Model

The data mining software WEKA 3.6.1 proposed by Witten and Frank [27] is used for developing SVM model. In this study an -variant of SVM (-SVM) is used for support vector regression, and the loss function () is fixed as 0.001. Initially, a polynomial kernel of degree () 2 is used to fit a nonlinear model. The selection of regularization parameter and kernel-specific parameters ( and for polynomial and RBF kernel, resp.) may influence the results. A large value of indicates that the objective function is only to minimize the empirical risk, which makes the learning machine more complex. On the other hand, a smaller may cause learning errors with poor approximation [28].

A trial and error approach is followed to find the optimal value of for model with polynomial kernel. The parameter of 100 is found to be quite successful in giving satisfactory performance. Then a radial basis function (RBF) kernel is used to fit a nonlinear model in the present study to build an SVM model. The combination of control parameters such as and gives very good training performance. The plot between observed and predicted values of training dataset is shown in Figure 4 (polykernel) and Figure 5 (RBF kernel). These plots indicate that the model is well trained.

##### 4.3. Development of GP Model

The genetic programming software DISCIPULUS [29] is used for developing GP model. The models are created in the form of βevolvedβ computer programs as GP uses Darwinian natural selection to create them. Using this model, the output of statistically similar input data can be predicted with very much accuracy. The initial control parameters used for the problem are population size: 500, crossover probability: 0.95, and mutation probability: 0.5. The basic arithmetical functions (such as addition, multiplication, subtraction, and division (+, *, β, /)) constitute the function set. The fitness function is selected as the root-mean-square error between the measure and predicted values of ultimate bearing capacity. The best program generated by GP software for predicting the UBC of cohesionless soils is given in the appendix. The plot between observed and predicted values of training dataset is shown in Figure 6. This plot indicates that the model is well trained.

#### 5. Results and Discussions

The efficiency of the developed models is analyzed by different statistical performance evaluation criteria such as correlation coefficient (), coefficient of efficiency (), root-mean-square error (RMSE), mean bias error (MBE), and mean absolute relative error (MARE). The equations of different performance evaluation measures were presented in Table 1, in which stands for the observed output value, represents the computed output value, is the mean of observed values, represents the mean of computed values, and represents the number of data points. The different performance evaluation criteria estimated for training dataset are presented in Table 2. The predictions for testing dataset using different models are presented in Table 3, and the performance evaluation for these predictions is presented in Table 4. However, it is to be noted that the ANN and FIS results presented in Table 4 are deduced based on the relative error (RE) values reported by Padmini et al. [5]. From Table 4 it can be inferred that the correlation coefficient and coefficient of efficiency are the highest (0.997 and 0.996) and the error criteria such as RMSE, MBE, and MARE are the least (44.967, 4.01, and 7.69) for the GP-based modeling.

Further the scatter plots between observed and predicted values of UBC for SVM models are presented in Figure 7 (polykernel) and Figure 8 (RBF kernel). The 5% error bar lines are plotted along with these scatter plots. Such a plot can be used to indicate the range of standard deviation and to determine whether the differences are statistically significant [8]. A perusal of plots shows that, for GP-based predictions, all points lie within the specified confidence interval of 95%. Thus, it can be inferred that all the soft computing methods outperform the theoretical approaches in the prediction of bearing capacity. Similar plot for predictions with GP model is presented in Figure 9. Also from Table 4, it is seen that the value and value are closer to unity and different error criteria are much lesser for any of the applied soft computing tools when compared with theoretical models.

A statistical evaluation of the predictions by the different soft computing models for the testing dataset is performed and presented in Table 5. The standard deviation, average deviation, and coefficient of variation values of GP model results (607.91, 454.59, and 1.059) show close agreement with that of observed values (600.02, 459.48, and 1.052) followed by that of SVM (RBF) model. This confirms the robustness of the newly applied paradigms.

The different performance evaluation measures of SVM-based modelling (in Tables 4 and 5) show that the performance of RBF-based SVM is competent with ANN and FIS results. Also SVM involves only lesser number of control parameters (such as and ), and ANN involves large number of such parameters and their optimal combination is a tedious process. Thus, the SVM approach is quite simple to implement. Further, the performance evaluation of GP-based results (Tables 4 and 5) shows that the , and different error criteria are better for the GP model when compared with the theoretical methods, the SVM, and interpreted results of ANN and FIS. Thus, GP is proven to be a reliable alternative soft computing technique for prediction of ultimate bearing capacity of shallow foundation on cohesionless soil.

#### 6. Conclusions

In this paper the application of two relatively recent soft computing techniquesβSVM and GPβis investigated for the prediction of ultimate bearing capacity of cohesionless soils beneath shallow foundations. SVM results are competent and demand the optimal selection of only a few number of control parameters when compared with ANN. Performance evaluation based on multiple error criteria shows that error is the least and correlation coefficient () and coefficient of efficiency () are the highest for the GP-based modeling than SVMs, ANN, FIS, and the different theoretical models considered in this study. The GP-based modeling is found to be superior in terms of quality, and it gives the output in the form of computer programs which enables the user to apply for a new set of input data to predict the ultimate bearing capacity. Thus, GP can be recommended as a robust soft computing paradigm to predict the ultimate bearing capacity of soil.

#### Appendices

#### A. Note

The C++ Program to predict the ultimate bearing capacity of cohesionless soils is given here. V[0] to V[4] represent the input parameters width of footing (), the depth of footing (), the length-to-width ratio (), the field density (), the angle of shearing resistance (). f[0], f[1], and so forth, are the temporary computation variables that the programs GP software creates. The output of these programs is the value remaining in f[0] after the program executes. This program needs to be run in the DISCIPULUS software environment to get the predictant value for a new set of predictors.

#### B. Best Program

#define TRUNC(x)(((x)>=0) ? floor(x): ceil(x)) | |

#define C_FPREM (_finite(f[0]/f[1]) ? f[0]-(TRUNC(f[0]/f[1]) | |

β*f[1]): f[0]/f[1]) | |

#define C_F2XM1 (((fabs(f[0])<=1) && | |

β(!_isnan(f[0]))) ? (pow(2,f[0])-1): | |

β((!_finite(f[0]) && !_isnan(f[0]) && | |

β(f[0]<0)) ? -1: f[0])) | |

β | |

float DiscipulusCFunction(float v[]) | |

{ | |

βββlong double f[8]; | |

β | |

βββlong double tmp = 0; | |

β | |

βββint cflag = 0; | |

βββf[0]=f[1]=f[2]=f[3]=f[4]=f[5]=f[6]=f[7]=0; | |

βββL0: f[0]/=-1.364008665084839f; | |

βββL1: f[0]+=f[1]; | |

βββL2: f[0]=βf[0]; | |

βββL3: f[0]β=v[0]; | |

βββL4: f[0]+=v[4]; | |

βββL5: f[0]+=v[4]; | |

βββL6: f[0]=cos(f[0]); | |

βββL7: f[0]+=f[0]; | |

βββL8: f[0]+=f[0]; | |

βββL9: f[0]+=f[0]; | |

βββL10: f[0]*=v[1]; | |

βββL11: f[0]+=v[4]; | |

βββL12: f[0]β=1.252994060516357f; | |

βββL13: f[0]*=pow(2,TRUNC(f[1])); | |

βββL14: cflag=(f[0] < f[1]); | |

βββL15: f[0]=sqrt(f[0]); | |

βββL16: f[0]*=0.2877938747406006f; | |

βββL17: tmp=f[1]; f[1]=f[0]; f[0]=tmp; | |

βββL18: f[0]β=v[3]; | |

βββL19: f[0]*=β0.494312047958374f; | |

βββL20: f[0]*=0.7790718078613281f; | |

βββL21: f[0]*=f[1]; | |

βββL22: f[0]=fabs(f[0]); | |

βββL23: f[0]=cos(f[0]); | |

βββL24: f[0]=βf[0]; | |

βββL25: f[0]=sqrt(f[0]); | |

βββL26: if (cflag) f[0] = f[1]; | |

βββL27: f[0]+=v[4]; | |

βββL28: f[0]*=0.9955191612243652f; | |

βββL29: f[0]*=0.4281637668609619f; | |

βββL30: f[0]β=f[0]; | |

βββL31: f[0]β=v[3]; | |

βββL32: f[0]/=v[0]; | |

βββL33: f[0]β=0.9955191612243652f; | |

βββL34: f[0]+=v[4]; | |

βββL35: cflag=(f[0] < f[1]); | |

βββL36: f[0]β=f[1]; | |

βββL37: f[0]/=f[0]; | |

βββL38: f[0]*=pow(2,TRUNC(f[1])); | |

βββL39: f[0]β=f[1]; | |

βββL40: f[0]=βf[0]; | |

βββL41: f[0]=fabs(f[0]); | |

βββL42: f[0]=βf[0]; | |

βββL43: f[0]*=0.4281637668609619f; | |

βββL44: f[0]*=f[0]; | |

βββL45: f[0]*=f[0]; | |

βββL46: f[0]/=f[0]; | |

βββL47: f[0]*=β0.7297487258911133f; | |

βββL48: f[0]/=1.084159851074219f; | |

βββL49: f[0]+=v[4]; | |

βββL50: tmp=f[0]; f[0]=f[0]; f[0]=tmp; | |

βββL51: f[0]/=f[1]; | |

βββL52: f[0]+=0.7790718078613281f; | |

βββL53: f[0]+=v[4]; | |

βββL54: f[0]β=f[1]; | |

βββL55: f[0]*=f[1]; | |

βββL56: f[0]*=0.4281637668609619f; | |

βββL57: f[0]β=v[3]; | |

βββL58: f[0]β=β0.9486191272735596f; | |

βββL59: if (cflag) f[0] = f[1]; | |

βββL60: f[0]/=v[2]; | |

βββL61: f[0]+=v[4]; | |

βββL62: f[0]β=v[2]; | |

βββL63: f[0]β=v[2]; | |

βββL64: f[0]β=v[2]; | |

βββL65: f[0]*=v[1]; | |

βββL66: f[0]+=v[4]; | |

βββL67: f[0]*=f[1]; | |

βββL68: f[0]*=0.4281637668609619f; | |

βββL69: f[0]β=v[3]; | |

βββL70: f[1]*=f[0]; | |

βββL71: f[0]*=f[0]; | |

βββL72: f[0]β=f[1]; | |

βββL73: f[1]+=f[0]; | |

βββL74: if (!cflag) f[0] = f[1]; | |

βββL75: f[0]β=1.987620830535889f; | |

βββL76: f[0]β=v[4]; | |

βββL77: f[0]β=1.987620830535889f; | |

βββL78: f[0]β=v[4]; | |

βββL79: f[0]β=v[4]; | |

βββL80: f[0]β=0.7361507415771484f; | |

βββL81: f[0]β=1.987620830535889f; | |

βββL82: f[0]β=v[4]; | |

ββL83: f[0]β=1.987620830535889f; | |

βββL84: f[0]β=v[4]; | |

βββL85: f[0]β=1.501374244689941f; | |

βββL86: f[0]+=v[2]; | |

βββL87: f[0]β=v[4]; | |

ββL88: f[0]β=1.530829906463623f; | |

βββL89: if (!cflag) f[0] = f[1]; | |

βββL90: f[0]+=v[3]; | |

βββL91: f[0]β=v[4]; | |

βββL92: f[0]+=β1.907608032226563f; | |

βββL93: if (!cflag) f[0] = f[1]; | |

βββL94: f[0]+=v[3]; | |

βββL95: f[0]+=v[3]; | |

βββL96: f[0]+=v[3]; | |

βββL97: f[0]+=v[3]; | |

βββL98: f[0]+=v[3]; | |

βββL99: f[0]+=v[3]; | |

βββL100: f[0]+=v[3]; | |

βββL101: f[0]+=v[3]; | |

βββL102: f[0]+=v[3]; | |

βββL103: f[0]+=v[3]; | |

βββL104: f[0]+=v[3]; | |

βββL105: f[0]+=v[3]; | |

βββL106: f[0]+=v[3]; | |

βββL107: f[0]+=v[3]; | |

βββL108: f[0]+=v[2]; | |

βββL109: f[0]+=v[3]; | |

βββL110: | |

βββif (!_finite(f[0])) f[0]=0; | |

βββreturn f[0]; | |

}. |

#### Acknowledgments

This paper is a part of a research work carried out at the Department of Civil Engineering, TKM College of Engineering Kollam, Kerala, India, in 2010. The authors thank the Department of Civil Engineering, TKM College of Engineering Kollam for providing all necessary help. They also thank the anonymous reviewer/s who helped to improve the quality of the paper.

#### References

- K. Terzaghi,
*Theoretical Soil Mechanics*, John Wiley & Sons, New York, NY, USA, 1943. - G. G. Meyerhof, βSome recent research on the bearing capacity of foundations,β
*Canadian Geotechnical Journal*, vol. 1, no. 1, pp. 16β26, 1963. View at Google Scholar - J. B. Hansen, βA general formula for bearing capacity,β
*Danish Geotechnical Institute Bulletin*, vol. 11, 1961. View at Google Scholar - A. S. Vesic, βAnalysis of ultimate loads of shallow foundations,β
*Journal of Soil Mechanics and Foundation Division*, vol. 99, no. 1, pp. 45β73, 1973. View at Google Scholar Β· View at Scopus - D. Padmini, K. Ilamparuthi, and K. P. Sudheer, βUltimate bearing capacity prediction of shallow foundations on cohesionless soils using neurofuzzy models,β
*Computers and Geotechnics*, vol. 35, no. 1, pp. 33β46, 2008. View at Publisher Β· View at Google Scholar Β· View at Scopus - M. A. Shahin, H. R. Maier, and M. B. Jaksa, βArtificial neural network applications in Geotechnical Eng,β
*Australian Geomechanics*, vol. 36, no. 1, pp. 49β62, 2001. View at Google Scholar - P. Samui, βPrediction of friction capacity of driven piles in clay using the support vector machine,β
*Canadian Geotechnical Journal*, vol. 45, no. 2, pp. 288β295, 2008. View at Publisher Β· View at Google Scholar Β· View at Scopus - P. Samui and T. G. Sitharam, βLeast-square support vector machine applied to settlement of shallow foundations on cohesionless soils,β
*International Journal for Numerical and Analytical Methods in Geomechanics*, vol. 32, no. 17, pp. 2033β2043, 2008. View at Publisher Β· View at Google Scholar Β· View at Scopus - P. Samui, βSlope stability analysis: a support vector machine approach,β
*Environmental Geology*, vol. 56, no. 2, pp. 255β267, 2008. View at Publisher Β· View at Google Scholar Β· View at Scopus - M. Pal, βSupport vector machines-based modelling of seismic liquefaction potential,β
*International Journal for Numerical and Analytical Methods in Geomechanics*, vol. 30, no. 10, pp. 983β996, 2006. View at Publisher Β· View at Google Scholar Β· View at Scopus - J. H. Holland,
*Adaptation in Natural and Artificial System*, Ann Arbour Science Press, Ann Arbor, Mich, USA, 1975. - D. E. Goldberg,
*Genetic Algorithms in Search, Optimization and Machine Learning*, Addison-Wesley, Reading, Mass, USA, 1989. - J. R. Koza,
*Genetic Programming: On the Programming of Computers by Means of Natural Selection*, MIT Press, Cambridge, Mass, USA, 1992. - B. E. Boser, I. M. Guyon, and V. N. Vapnik, βA training algorithm for optimal margin classifiers,β in
*5th Annual ACM Workshop on COLT*, D. Haussler, Ed., pp. 144β152, ACM Press, Pittsburgh, Pa, USA, 1992. View at Google Scholar - V. N. Vapnik,
*Statistical Learning Theory*, John Wiley & Sons, New York, NY, USA, 1998. - C. Cortes and V. Vapnik, βSupport-vector networks,β
*Machine Learning*, vol. 20, no. 3, pp. 273β297, 1995. View at Publisher Β· View at Google Scholar Β· View at Scopus - A. J. Smola and B. Schölkopf, βA tutorial on support vector regression,β
*Statistics and Computing*, vol. 14, no. 3, pp. 199β222, 2004. View at Publisher Β· View at Google Scholar Β· View at MathSciNet Β· View at Scopus - S. Rajasekaran, S. Gayathri, and T. L. Lee, βSupport vector regression methodology for storm surge predictions,β
*Ocean Engineering*, vol. 35, no. 16, pp. 1578β1587, 2008. View at Publisher Β· View at Google Scholar Β· View at Scopus - S. T. Khu, S. Y. Liong, V. Babovic, H. Madsen, and N. Muttil, βGenetic programming and its application in real-time runoff forecasting,β
*Journal of the American Water Resources Association*, vol. 37, no. 2, pp. 439β451, 2001. View at Google Scholar Β· View at Scopus - A. A. Javadi, M. Rezania, and M. M. Nezhad, βEvaluation of liquefaction induced lateral displacements using genetic programming,β
*Computers and Geotechnics*, vol. 33, no. 4-5, pp. 222β233, 2006. View at Publisher Β· View at Google Scholar Β· View at Scopus - A. Johari, G. Habibagahi, and A. Ghahramani, βPrediction of soil-water characteristic curve using genetic programming,β
*Journal of Geotechnical and Geoenvironmental Engineering*, vol. 132, no. 5, pp. 661β665, 2006. View at Publisher Β· View at Google Scholar Β· View at Scopus - B. S. Narendra, P. V. Sivapullaiah, S. Suresh, and S. N. Omkar, βPrediction of unconfined compressive strength of soft grounds using computational intelligence techniques: a comparative study,β
*Computers and Geotechnics*, vol. 33, no. 3, pp. 196β208, 2006. View at Publisher Β· View at Google Scholar Β· View at Scopus - S. B. Charhate, M. C. Deo, and S. N. Londhe, βInverse modeling to derive wind parameters from wave measurements,β
*Applied Ocean Research*, vol. 30, no. 2, pp. 120β129, 2008. View at Publisher Β· View at Google Scholar Β· View at Scopus - S. Gaur and M. C. Deo, βReal-time wave forecasting using genetic programming,β
*Ocean Engineering*, vol. 35, no. 11-12, pp. 1166β1172, 2008. View at Publisher Β· View at Google Scholar Β· View at Scopus - K. Ustoorikar and M. C. Deo, βFilling up gaps in wave data with genetic programming,β
*Marine Structures*, vol. 21, no. 2-3, pp. 177β195, 2008. View at Publisher Β· View at Google Scholar Β· View at Scopus - S. S. Kashid, S. Ghosh, and R. Maity, βStreamflow prediction using multi-site rainfall obtained from hydroclimatic teleconnection,β
*Journal of Hydrology*, vol. 395, no. 1-2, pp. 23β38, 2010. View at Publisher Β· View at Google Scholar Β· View at Scopus - I. H. Witten and E. Frank,
*Data Mining*, Morgan Kaufmann, San Francisco, Calif, USA, 2000. - P. S. Yu, S. T. Chen, and I. F. Chang, βSupport vector regression for real-time flood stage forecasting,β
*Journal of Hydrology*, vol. 328, no. 3-4, pp. 704β716, 2006. View at Publisher Β· View at Google Scholar Β· View at Scopus - F. D. Francone,
*Discipulus Owner’s Manual Version 3.0 DRAFT*, Machine Learning Technologies, Littleton, Colo, USA, 1998.