- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

Journal of Renewable Energy

Volume 2013 (2013), Article ID 952613, 9 pages

http://dx.doi.org/10.1155/2013/952613

## The Effectiveness of Feature Selection Method in Solar Power Prediction

Power Engineering Research Group (PERG), Central Queensland University, Rockhampton, QLD 4701, Australia

Received 13 December 2012; Accepted 11 July 2013

Academic Editor: Czarena Crofcheck

Copyright © 2013 Md Rahat Hossain et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

This paper empirically shows that the effect of applying selected feature subsets on machine learning techniques significantly improves the accuracy for solar power prediction. Experiments are performed using five well-known wrapper feature selection methods to obtain the solar power prediction accuracy of machine learning techniques with selected feature subsets. For all the experiments, the machine learning techniques, namely, least median square (LMS), multilayer perceptron (MLP), and support vector machine (SVM), are used. Afterwards, these results are compared with the solar power prediction accuracy of those same machine leaning techniques (i.e., LMS, MLP, and SVM) but without applying feature selection methods (WAFS). Experiments are carried out using reliable and real life historical meteorological data. The comparison between the results clearly shows that LMS, MLP, and SVM provide better prediction accuracy (i.e., reduced MAE and MASE) with selected feature subsets than without selected feature subsets. Experimental results of this paper facilitate to make a concrete verdict that providing more attention and effort towards the feature subset selection aspect (e.g., selected feature subsets on prediction accuracy which is investigated in this paper) can significantly contribute to improve the accuracy of solar power prediction.

#### 1. Introduction

Feature selection can be considered one of the main preprocessing steps of machine learning [1]. Feature selection is different from feature extraction (or feature transformation), which creates new features by combining the original features [2]. The advantages of feature selection are manyfold. First, feature selection significantly saves the operating time of a learning procedure by eliminating irrelevant and redundant features. Second, without the intervention of irrelevant, redundant, and noisy features, learning algorithms can centrally point on most essential features of data and build simpler but more precise data models. Third, feature selection can help build a simpler and more common model and get a better insight into the fundamental perception of the task [3–5]. The feature selection aspect is fairly significant for the reason that with the same training data, it may happen that an individual regression algorithm can perform better with different feature subsets. The success of machine learning on a particular task is affected by many factors. Among those factors first and foremost is the representation and quality of the instance data [6]. The training stage becomes critical with the existence of noisy, irrelevant, and redundant data. Sometimes, the real life data contain too much information; among those, very little is useful for desired purpose. Therefore, it is not important to include every piece of information from the raw data source for modelling.

Usually, features are differentiated [7] as(1)relevant: this class of features has strong impact on the output;(2)irrelevant: opposite to relevant features, irrelevant features do not have any bias on the output;(3)redundant: a redundancy occurs when a feature captures the functionality of other;

All the algorithms to perform feature selection consist of two common aspects. One is the search method which is actually a selection algorithm to generate designed feature subsets and attempts to reach the most advantageous ones. Another aspect is called evaluator which is basically an evaluation algorithm to make a decision about the goodness of the planned feature subset and finally returns the assessment about righteousness of the search method [8]. On the other hand, lacking an appropriate stopping condition, the feature selection procedure could run exhaustively or everlastingly all the way throughout the raw dataset. It may be discontinued whenever any attribute is inserted or deleted but ultimately not producing a better subset or whenever a subset is produced which provides the maximum benefits according to some assessing functions. A feature selector may stop manipulating features when the merit of a current feature subset stops improving or conversely does not degrade. A usual feature selection procedure revealed in Figure 1 consists of four fundamental steps: (A) subset generation; (B) subset evaluation; (C) stopping criterion; and (D) result validation [9]. The procedure begins with subset generation that utilizes a definite search strategy to generate candidate feature subsets. Then, each candidate subset is evaluated according to a definite evaluation condition and compared with the previous best one. If it is better, it replaces the previous best. The process of subset generation and evaluation is repetitive until a given stopping condition is satisfied. Finally, the selected best feature subset is validated by some test data. Figure 1 graphically demonstrates the previous mentioned steps and procedures of feature selection process.

Based on some evaluation functions and calculations, feature selection methods find out the best feature from different candidate subsets. Usually, feature selection methods are classified into two general groups (i.e., filter and wrapper) [10]. Inductive algorithms are used by wrapper methods as the evaluation function whereas filter methods are independent of the inductive algorithm. Wrapper methods work along wrapping the feature selection in conjunction with the induction algorithm to be used, and to accomplish this wrapper methods use cross validation. With the same training data, it may happen that individual regression algorithm can perform better with different feature subsets [11]. Widely used wrapper selection methods are briefly discussed in the following section. Section 3 deals with real life data collection and analysis of the dataset. Sections 4 and 5 show the experimental results using machine learning techniques and selected feature subsets. Comparison of the results and graphical presentation of those results are presented in those two sections. Section 6 demonstrates the performance of those obtained prediction results by paired -tests. The results from the experiments demonstrate that LMS, MLP, and SVM supplied with selected feature subsets provide better prediction accuracy (i.e., reduced MAE and MASE) than when they are without the selected feature subsets. Concluding remarks are provided in final section of this paper.

#### 2. Wrapper Methods of Feature Selection

In the field of machine learning, the feature subset selection that is also named as attribute subset selection, variable selection, or variable subset selection is an important method that helps to select an appropriate subset of significant features for model development. In machine learning based advanced and sophisticated applications (e.g., solar power prediction), feature selection methods have become an obvious need. Generally, feature selection methods are categorised into three different classes [3, 12]: *filter selection methods* select feature subsets as a preprocessing act, autonomously of the selected predictor; *wrapper selection methods* exploit the predictive power of the machine learning technique to select suitable feature subsets; and to conclude, *embedded selection methods* usually select feature subsets in the course of training. Wrapper methods were used for the experiments in this paper. Therefore, in the subsequent section, widely used and accepted wrapper methods are placed.

The wrapper methods use the performance (e.g., regression, classification, or prediction accuracy) of an induction algorithm for feature subset evaluation. Figure 2 shows the ideas behind wrapper approaches [3]. For each generated feature subset , wrappers evaluate its goodness by applying the induction algorithm to the dataset using features in subset . Wrappers can find feature subsets with high accuracy because the features match well with the learning algorithms.

The easiest method among all the wrapper selection algorithms is the forward selection (FS). This method starts the procedure without having any feature in the feature subset and follows a greedy approach so that it can sequentially add features until no possible single feature addition results in a higher valuation of the induction function. Backward elimination (BE) begins with the complete feature set and gradually removes features as long as the valuation does not degrade. Description about forward selection (FS) and backward selection (BS) can be found in [13] where the authors proved that wrapper selection methods are better than methods having no selection.

Starting with an empty set of features, the best first search (BFS) produces every possible individual feature extension [3]. BFS exploits the greedy hill climbing approach in conjunction with backtracking to search the space of feature subsets. BFS has all the flexibility to start with an empty subset and search in forward direction. Alternatively, it can start having full set of attributes and search in backward direction or it can start randomly from any point and move towards any direction. Extension of the BFS is the linear forward selection (LFS). A limited number of attributes are taken into consideration by LFS. This method either selects the top attributes by initial ordering or it can carry put a ranking [14, 15].

Subset size forward selection (SSFS) is the extension of LFS. SSFS carries out an internal cross validation. An LFS is executed on every fold to find out the best possible subset size [15, 16]. Through the individual evaluations, attributes are ranked by the ranker search. It uses this search in combination with attribute evaluators [16].

GA performs a search using the simple genetic algorithm described in Goldberg’s study [17]. Genetic algorithms are random search techniques based on the principles of natural selection [17]. They utilize a population of competing solutions evolved to an optimal solution. For feature selection, a solution is a fixed length binary sequence representing a feature subset. The value of each position—typically 1 or 0—in the sequence represents the presence or absence of a particular feature, respectively. The algorithm proceeds in an iterative manner where each successive generation is produced by applying genetic operators to the members of the current generation. Nonetheless, GAs naturally involves a huge quantity of evaluations or iterations to achieve optimal solution. Other than all these conventional methods, we have experimentally verified an unconventional approach. In this method, we calculated the correlation coefficient for each (except the target attribute) of the competing attributes with respect to the target attribute of the used dataset. For this purpose, we used Pearson’s correlation coefficient formula which is described in the next section. After the attribute wise calculation, we selected those attributes whose correlation coefficient values are positive only as feature subset. The attributes having negative correlation coefficient are ignored for this case. We named this method positive correlation coefficient selection (PCCS) [18].

#### 3. Real Life Data Collection and Analysis

One of the key conditions to successfully perform the experiments of this paper is to collect recent, reliable, accurate, and long-term historical weather data of the particular location. However, finding accurate multiyear data near the experiment site has always proved to be challenging because these data are not readily available due to the cost and difficulty in measurement [19]. There are only few sites in Australia providing research-quality data of solar radiation; so, these data are generally not available. Rockhampton, a subtropical town in North Australia, was chosen for the experiments of this research. The selected station is “Rockhampton Aero,” having latitude of −23.38 and longitude of 150.48. According to the Renewable Energy Certificates (RECs) zones within Australia, Rockhampton is identified within the most important zone [20]. The recent data were collected from Commonwealth Scientific and Industrial Research Organization (CSIRO), Australia.

Data were also collected from the Australian Bureau of Meteorology (BOM), the National Aeronautics and Space Administration (NASA), and the National Oceanic and Atmospheric Administration (NOAA). All of this mission-critical information on solar radiation was being estimated from cloud cover and humidity at airports. Free data are available from National Renewable Energy Laboratory (NREL) and NASA. These are excellent for multiyear averages but perform poorly for hourly and daily measurements. After analyzing the raw data collected from different sources, the data provided by CSIRO were finally selected. The data used in this paper are based on hourly global solar irradiance ground measurements which are a significant aspect of the dataset. These data were gathered for a period of five years from 2006 to 2010.

The attributes in the used dataset are average air temperature, average wind speed, current wind direction, average relative humidity, total rainfall, wind speed, wind direction, maximum peak wind gust, current evaporation, average absolute barometer, and average solar radiation. Table 1 represents the statistical properties of the raw data.

#### 4. Applying Feature Selection Techniques on the Dataset

All the research works related to solar radiation prediction select the input features or attributes randomly. Unlike the conventional way, this research experimented with the maximum number of features and found out the best possible combination of features for the individual learning models of the hybrid model. To perform the experiments for selecting significant feature subsets for individual machine learning technique, the traditional BFS, LFS, SSFS, ranker search, GS, and our very own PCCS selection methods are used. To carry out experiments, three algorithms for machine learning technique, namely, least median square [21], multilayer perceptrons [22], and support vector machine [23], are used.

Evaluating the degree of fitness, that is, how well a regression model fits to a dataset, is usually obtained by correlation coefficient. Assuming the actual values as and the predicted values as , the correlation coefficiency is known by the equation: where, To find out the correlation coefficient of the model, the full training set is partitioned into ten mutually exclusive and same-sized subsets. The performance of the subset depends on the accuracy of predicting test values. For every individual algorithm, this cross validation method was run over ten times, and finally, the average value for 10 cross validations was calculated. In -cv, a dataset is uniformly partitioned into folds of similar size . For the sake of clarity and without loss of generality; it is supposed that is multiple of . Let be the complement dataset of . Then, the algorithm induces a classifier from , and estimates its prediction error with . The -cv prediction error estimator of is defined as follows [24]: where if and only if and equal to zero otherwise. So, the -cv error estimator is the average of the errors made by the classifiers in their respective divisions .

According to Zheng and Kusiak in [24], the mean absolute error (MAE) and mean absolute percent error (MAPE) are used to measure the prediction performance; we have also used these evaluation metrics for our experiments. The definitions are expressed as where , , , , and .

Error of the experimental results was also analyzed according to mean absolute scaled error (MASE) [25]. MASE is scale free, less sensitive to outlier; its interpretation is very easy in comparison to other methods and less variable to small samples. MASE is suitable for uncertain demand series as it never produces infinite or undefined results. It indicates that the prediction with the smallest MASE would be counted the most accurate among all other alternatives [25]. Equation (5) states the formula to calculate MASE as where,

##### 4.1. Prediction of the Machine Learning Techniques Using the Selected Feature Subsets

Various feature subsets were generated or selected using different wrapper feature selection methods. Afterwards, six-hours-ahead solar radiation prediction by the selected machine learning techniques, namely, LMS, MLP, and SVM, was performed. For this instance, the selected feature subsets were supplied to the individual machine learning techniques. The intention of this experiment was to observe whether this initiative produces any improvement in the error reduction of those selected machine learning techniques or not. For these experiments, any tuning of the particular algorithms to a definite dataset was avoided. For all the experiments, default values of learning parameters were used. In general, in the following tables, one can see the CC, MAE, MAPE, and MASE of six-hours-ahead prediction for each machine learning technique supplied with different feature subsets. For all the experiments, “W” is used to indicate that a particular machine learning technique supplied with the selected feature subsets statistically outplays the one without applying feature selection (WAFS) methods.

Tables 2 and 3 represent the obtained CC and MAE for applying LMS, MLP, and SVM machine learning technique for six hours in advance prediction on the used dataset before and after feature selection process.

In Tables 4 and 5, the MAPE and MASE are shown before and after feature selection processes are applied to LMS, MLP, and SVM machine learning technique for the same purpose.

The results from the experimental results show that the PCCS is somewhat a superior feature selection method for LMS algorithm considering all the instances. It is noticeable that all the feature selection methods contributed to improve the CC of LMS algorithm. However, in the case of MAE, all the selection algorithms except the GS improve the results for LMS. In both the case of MAPE and MASE, BFS is the only selection method which does not improve the results for LMS. It is found from those results that the ranker search is to some extents superior feature selection method for MLP algorithm. It is noticeable that all the feature selection methods present a nearly close CC for MLP algorithm but in the case of MAE, MAPE, and MASE, ranker search is the only selection method which improves the results. Finally, the obtained results illustrate that again the ranker search is to some extent a superior feature selection method for SVM. It is also noticeable that all the feature selection methods present either nearly close or equal CC for SVM. However, in the case of MAE, MAPE, and MASE, LFS is the only one which is unable to improve the results for SVM.

#### 5. Prediction Results: Before versus after Applying the Feature Selection Techniques

In Table 6, the prediction errors (MAE and MASE) of the individual machine learning techniques are compared on the basis of before supplying selected feature subsets and after supplying selected feature subsets on them. The comparative results show that errors are reduced for all the instances after supplying selected feature subsets. The terms MAE_BEFORE and MASE_BEFORE represent the results for having no selected feature subsets for MAE and MASE, respectively, whereas the terms MAE_AFTER and MASE_AFTER represent the results having selected feature subsets for MAE and MASE, respectively.

Figure 3(a) graphically demonstrates the prediction accuracy comparison of LMS, MLP, and SVM before and after applying selected feature subsets in terms of MAE. Figure 3(b) graphically demonstrates the prediction accuracy comparison of LMS, MLP, and SVM before and after applying selected feature subsets in terms of MASE.

Figure 4(a) graphically illustrates the significance of applying and without applying (WAFS) various feature selection methods on LMS in terms of CC. Figure 4(b) graphically illustrates the significance of applying and without applying (WAFS) feature selection methods on MLP in terms of MAE. In both the cases, results are improved after applying selected feature subsets.

Figure 5 graphically illustrates the significance of applying and without applying (WAFS) feature selection methods on SVM in terms of MAPE. The graphical representation clearly shows that the result is improved in terms of MAPE after applying selected feature subsets.

#### 6. Performance Analysis through Statistical Tests

A statistical test provides a mechanism for making quantitative decisions about a process or processes. The intention is to determine whether there is enough evidence to “reject” an inference or hypothesis about the process. For a single group, the means of two variables are usually compared by the paired-samples -test. It considers the differences between the values of two variables in a case-by-case basis and examines whether the average varies from 0 or not. It carries out various types of outputs such as statistical description for the test variables, the correlation among the test variables, meaningful statistics for the paired differences, the -test itself, and a 95% confidence interval. The paired-samples -tests are carried out in order to justify whether any significant difference exists between the actual and predicted results achieved by the selected three machine learning techniques for ensemble method. The -test is executed with the SPSS package—PASW Statistics 20 [26]. In this paper, the paired -test is employed to verify the performance of the machine learning techniques used in the experiments. Here, the null hypothesis and the alternative hypothesis are termed as and , respectively, where means there is no significant difference between the actual and predicted mean values, means there is significant difference between the actual and predicted mean values.

In Tables 7, 8, and 9, the paired samples statistics, the paired sample correlations, and paired samples tests for the actual and predicted values of LMS, MLP, and SVM are represented, respectively. Observing Table 9, we find that for pair 1, , ; for pair 2, , ; for pair 3, , ; and for pair 4, , . Due to the means of the actual and predicted values of each pair and the direction of the values, it can be concluded that there was no statistically significant difference between actual and predicted values for all the cases. Therefore, the test failed to reject the null hypothesis .

#### 7. Conclusions

Feature selection is a fundamental issue in both the regression and classification problems especially for the dataset having a very high volume of data. Applying feature selection methods on machine learning techniques may significantly contribute to increase performance in terms of accuracy. In this paper, various methods of feature selection methods have been briefly described. In particular, the wrapper feature selection methods are found better, which is also justified by the results obtained from the experiments performed in this paper. From the experiments performed in this paper, it is found that for LMS, the MAE before and after applying selected feature subsets is 77.19 and 73.37, respectively, and MASE 0.63 and 0.59, respectively. In the case of MLP, the MAE before and after applying selected feature subsets is 91.02 and 84.31, respectively, and MASE 0.74 and 0.68, respectively. For SVM, the MAE before and after applying selected feature subsets is 126.88 and 122.11, respectively, and MASE 1.03 and 0.99, respectively. The comparison between the results clearly shows that LMS, MLP, and SVM provide better prediction accuracy (i.e., reduced MAE and MASE) with selected feature subsets than without selected feature subsets. Experimental results of this paper facilitate to make a concrete verdict that providing more attention and effort towards the feature subset selection aspect (e.g., selected feature subsets on prediction accuracy which is investigated in this paper) can significantly contribute to improve the accuracy of solar power prediction.

It is mentionable that for these experiments, the machine learning techniques were applied with the default learning parameter settings. In the near future, the new experiments will be performed with the intention to achieve better prediction accuracy of the selected machine learning techniques by applying both the optimized or tuned learning parameter settings and selected feature subsets on them.

#### Nomenclature

: | Actual values |

: | Alternative hypothesis |

: | Class attribute; |

: | Complement dataset of |

cv: | Cross validation |

: | Divisions |

-cv: | Error estimator |

: | Induction algorithm |

M: | Mean |

: | Null hypothesis |

: | Number of folds |

: | Number of instances |

: | Set of points |

: | Standard deviation |

: | Training set |

: | Two-tailed significance value |

: | Unknown errors or residuals |

: | Unlabeled instance. |

*Greek Symbols*

: | Classifiers |

: | Randomized error estimator |

: | Significance level = 0.05. |

*Abbreviations*

BOM: | Bureau of meteorology |

BE: | Backward elimination |

BS: | Backward selection |

BFS: | Best first search |

CSIRO: | Commonwealth Scientific and Industrial Research Organization |

CC: | Correlation coefficient |

Deg: | Degree |

DegC: | Degree Celsius |

FCBF: | Fast correlation-based feature selection |

FS: | Forward selection |

GA: | Genetic algorithm |

GS: | Genetic search |

hPa: | Hectopascals |

Km/h: | Kilometer per hour |

LFS: | Linear forward selection |

LMS: | Least median square |

MAE: | Mean absolute error |

MASE: | Mean absolute scaled error |

MAPE: | Mean absolute percent error |

MLP: | Multilayer perceptron |

PASW: | Predictive analytic software |

RECs: | Renewable energy certificates |

mm: | Millimeter |

NASA: | National Aeronautics & Space Admin |

NOAA: | National Oceanic & Atmospheric Admin |

NREL: | National Renewable Energy Laboratory |

PCCS: | Positive correlation coefficient selection |

SPSS: | Statistical Package for Social Sciences |

SSFS: | Subset size forward selection |

SVM: | Support vector machine |

WAFS: | Without applying feature selection |

W/m^{2}: | Watt per meter square. |

#### References

- L. Yu and H. Liu, “Feature selection for high-dimensional data: a fast correlation based filter solution,” in
*Proceedings of the 20th International Conference on Machine Learning*, pp. 856–863, August 2003. View at Scopus - F. Tan,
*Improving feature selection techniques for machine learning [Ph.D. thesis]*, 2007, Computer Science Dissertations: Paper 27, http://digitalarchive.gsu.edu/cs_diss/27. - R. Kohavi and G. H. John, “Wrappers for feature subset selection,”
*Artificial Intelligence*, vol. 97, no. 1-2, pp. 273–324, 1997. View at Google Scholar · View at Scopus - D. Koller and M. Sahami, “Toward optimal feature selection,” in
*Proceedings of the 13th International Conference on Machine Learning*, pp. 284–292, 1996. - M. Dash and H. Liu, “Feature selection for classification,”
*Intelligent Data Analysis*, vol. 1, no. 3, pp. 131–156, 1997. View at Google Scholar - T. Mitchell,
*Machine Learning*, McGraw Hill, 1997. - D. A. Bell and H. Wang, “A formalism for relevance and its application in feature subset selection,”
*Machine Learning*, vol. 41, no. 2, pp. 175–195, 2000. View at Publisher · View at Google Scholar · View at Scopus - M. Karagiannopoulos, D. Anyfantis, S. B. Kotsiantis, and P. E. Pintelas, “Feature selection for regression problems,” in
*Proceedings of the 8th Hellenic European Research on Computer Mathematics & Its Applications (HERCMA '07)*, pp. 20–22, 2007. - H. Liu and L. Yu, “Toward integrating feature selection algorithms for classification and clustering,”
*IEEE Transactions on Knowledge and Data Engineering*, vol. 17, no. 4, pp. 491–502, 2005. View at Publisher · View at Google Scholar · View at Scopus - P. Langley, “Selection of relevant features in machine learning,” in
*Proceedings of the AAAI Fall Symposium on Relevance*, pp. 1–5, 1994. - A. L. Blum and P. Langley, “Selection of relevant features and examples in machine learning,”
*Artificial Intelligence*, vol. 97, no. 1-2, pp. 245–271, 1997. View at Google Scholar · View at Scopus - I. Guyon and A. Elisseeff, “An introduction to variable and feature selection,”
*Journal of Machine Learning Research*, vol. 3, pp. 1157–1182, 2003. View at Google Scholar - R. Caruana and D. Freitag, “Greedy attribute selection,” in
*Proceedings of the 11th International Conference on Machine Learning*, San Francisco, Calif, USA, 1994. - M. Gütlein, E. Frank, M. Hall, and A. Karwath, “Large-scale attribute selection using wrappers,” in
*Proceedings of the IEEE Symposium on Computational Intelligence and Data Mining (CIDM '09)*, pp. 332–339, April 2009. View at Publisher · View at Google Scholar · View at Scopus - M. Gutlein,
*Large scale attribute selection using wrappers [Diploma of Computer Science thesis]*, Albert-Ludwigs-Universität-Freiburg, 2006. - M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann, and H. Witten, “The WEKA data mining software: an update,”
*ACM SIGKDD Explorations Newsletter*, vol. 11, no. 1, pp. 10–18, 2009. View at Google Scholar - E. Goldberg,
*Genetic Algorithms in Search, Optimization and Machine Learning*, Addison-Wesley, 1989. - M. R. Hossain, A. M. T. Oo, and A. B. M. S. Ali, “The combined effect of applying feature selection and parameter optimization on machine learning techniques for solar Power prediction,”
*American Journal of Energy Research*, vol. 1, no. 1, pp. 7–16, 2013. View at Publisher · View at Google Scholar - S. Coppolino, “A new correlation between clearness index and relative sunshine,”
*Renewable Energy*, vol. 4, no. 4, pp. 417–423, 1994. View at Google Scholar · View at Scopus - http://www.solarguys.com.au/.
- P. J. Rousseeuw, “Least median of squares regression,”
*Journal of the American Statistical Association*, vol. 79, no. 388, pp. 871–880, 1984. View at Google Scholar - S. Haykin,
*Neural Networks: A Comprehensive Foundation*, Prentice Hall, 1999. - S. K. Shevade, S. S. Keerthi, C. Bhattacharyya, and K. R. K. Murthy, “Improvements to the SMO algorithm for SVM regression,”
*IEEE Transactions on Neural Networks*, vol. 11, no. 5, pp. 1188–1193, 2000. View at Publisher · View at Google Scholar · View at Scopus - H. Zheng and A. Kusiak, “Prediction of wind farm power ramp rates: a data-mining approach,”
*Journal of Solar Energy Engineering*, vol. 131, no. 3, Article ID 031011, 8 pages, 2009. View at Publisher · View at Google Scholar · View at Scopus - R. J. Hyndman and A. B. Koehler, “Another look at measures of forecast accuracy,”
*International Journal of Forecasting*, vol. 22, no. 4, pp. 679–688, 2006. View at Google Scholar *IBM SPSS Statistics for Windows-Version 20.0*, IBM Corporation, New York, NY, USA, 2011.