Research Article  Open Access
Lida Barba, Nibaldo Rodríguez, "Hybrid Models Based on Singular Values and Autoregressive Methods for Multistep Ahead Forecasting of Traffic Accidents", Mathematical Problems in Engineering, vol. 2016, Article ID 2030647, 14 pages, 2016. https://doi.org/10.1155/2016/2030647
Hybrid Models Based on Singular Values and Autoregressive Methods for Multistep Ahead Forecasting of Traffic Accidents
Abstract
The traffic accidents occurrence urges the intervention of researchers and society; the human losses and material damage could be abated with scientific studies focused on supporting prevention plans. In this paper prediction strategies based on singular values and autoregressive models are evaluated for multistep ahead traffic accidents forecasting. Three time series of injured people in traffic accidents collected in Santiago de Chile from 2000:1 to 2014:12 were used, which were previously classified by causes related to the behavior of drivers, passengers, or pedestrians and causes not related to the behavior as road deficiencies, mechanical failures, and undetermined causes. A simplified form of Singular Spectrum Analysis (SSA), combined with the autoregressive linear (AR) method, and a conventional Artificial Neural Network (ANN) are proposed. Additionally, equivalent models that combine Hankel Singular Value Decomposition (HSVD), AR, and ANN are evaluated. The comparative analysis shows that the hybrid models SSAAR and SSAANN reach the highest accuracy with an average of 1.5% and 1.9%, respectively, from 1 to 14step ahead prediction. However, it was discovered that HSVDAR shows a higher accuracy in the farthest horizons, from 12 to 14step ahead prediction, which reaches an average of 2.2%.
1. Introduction
Traffic accidents with fatalities and severely injured people are a socioeconomic problem of focus. According to the WHO [1], the traffic accidents cost between 1% and 3% of the GDP of a country, regardless of invaluable emotional damage for the victims and their families. Several studies have been developed to explain the nature of the problem, most through classification; diverse methods have been used to detect the key factors that influence the incidents severity. Abellán et al. (2013) used decision trees to extract some key patterns of severe accidents in Granada; the rules were defined in function of the variables related to atmospheric factors, driver characteristics, road conditions, or a factor combination [2]. Chang and Chien applied a method based on nonparametric classification and regression tree to establish the empirical relationship between injury severity and driver/vehicle characteristics in truckinvolved accidents [3]. De Oña et al. (2013) use Latent Class Clustering and Bayesian networks to identify the variables involved in traffic accidents; the accident type and sight distance were detected in all the traffic accidents on rural highways in Granada [4]. Recently, Shiau et al. (2015) presented Fuzzy Robust Principal Component Analysis (FRPCA) combined with Backpropagation Neural Network (BPNN) to identify the relationships between the variables of road designs, ruleviolation items, and accident types; the results showed an 85.89% classification accuracy [5]. Lin et al. (2015) used real time traffic accidents in a highway in Virginia; they found that the best model was based on Frequent Pattern and a Bayesian network; the model predicted 61.1% of accidents while having a false alarm rate of 38.16% [6]. The risk indicating variables of traffic accidents have been identified taking into consideration diverse factors; some factors are related to road design parameters [7, 8], environment conditions [9], traffic signs, or interactions among some factors [10].
The data provided by the Chilean Commission of Traffic Safety (CONASET) [11] shows a high and increasing rate of fatalities and injuries in traffic accidents from 2000 to 2014. Santiago is the most populated Chilean region, whose time series of injured people is used in this analysis. The abundance of information and research about traffic accidents, severities, risks, and occurrence factors, among others, might appear somewhat unapproachable in terms of the prevention plans.
In this case study, with categorical causes defined by CONASET and the ranking method, the potential causes of injuries in traffic accidents, which can be counteracted through campaigns directed to the change of attitude in drivers, passengers, and pedestrians toward road safety, have been identified. Two groups have been created based on the primary and secondary causes which are present in around of injured people, and a third group was created with the remaining 25%. At first, nonstationary and nonlinear characteristics were found in the time series, turning the forecasting into a difficult task. Some researches exploit the potentialities of the SSA technique to extract components in a time series; SSA commonly has been used to extract trend, seasonality, and/or noise [12]; the extracted components are used to explain the complex behavior of some time series in diverse ambit, from nature [13, 14] to industrial process [15], or economic indicators [16].
On the other hand, the combination of SSA and autoregressive linear and nonlinear methods is a recent alternative which has demonstrated robustness and universal capacities in terms of shortterm forecasting [17–19]. SSA combined with artificial intelligence techniques was used by Xiao et al. (2014) for monthly air transport demand forecasting [20]; a similar combination was made by Abdollahzade et al. (2015) with a nonlinear and chaotic time series [21].
To our knowledge [22], onestep ahead forecasting based on HSVD combined with ARIMA and neural networks reaches high accuracy in traffic accidents prediction. HSVD has similarities with respect to SSA in the steps of embedding, decomposition, and grouping. The main difference is in the last step previous to the component extraction; HSVD avoids diagonal averaging; instead, a particular extraction process is proposed. Although this difference exists, in terms of computational complexity both SSA and HSVD have equal floating point operations, which is due to the Rbidiagonalization (RSVD) [23] implemented in both cases. For that reason, comparisons in terms of implementation and forecasting results are presented.
This forecasting approach is described in two stages, preprocessing and prediction. In the first stage Singular Spectrum Analysis and Singular Value Decomposition (SVD) of Hankel are implemented to obtain the components of low and high frequency from the observed time series; an additive component of low frequency is extracted, whereas the component of high frequency is computed by subtraction. The result is a pair of smoothed time series which can be predicted robustly by computing loworder autoregressive methods in the second stage. The direct method is applied in the second stage to develop multistep ahead prediction, with multiinput and singleoutput. The inputs are the components lagged values, whose optimal number is identified through the Autocorrelation Function.
Conventional SSA is put into practice in four steps, embedding, decomposition, grouping, and diagonal averaging over all the elementary matrices [24]. In this work, the SSA implementation is simplified in three steps: embedding, decomposition, and diagonal averaging; only one elementary matrix is needed and it is computed with the first SVD eigentriple. The time series of length is embedded in a trajectory matrix of dimension , where is the window length. The general rule used to delimit the windows length is ; a large decomposition is given with a high value of , while a short decomposition is the opposite. During the literature review, some strategies have been used to select the window length; some instances are [25], weighted correlation [26], and extreme of autocorrelation [27]. The method used in this work to select the optimal window length is based on the Shannon entropy of the singular values. The second step of SSA is Singular Value Decomposition of the matrix obtained in the embedding; with the first eigentriple an elementary matrix is computed. Finally, by diagonal averaging over the elementary matrix, the elements of low frequency component are extracted.
The contribution is an accurate multistep ahead forecasting methodology based on singular values decomposition and autoregressive models through the comparison of four hybrid models. The prediction is focused on causes of the traffic accidents with injured people, which is oriented to support prevention plans of the government and police. The paper is organized as follows. Section 2 describes the Methodology. Section 3 shows efficiency criteria to evaluate the prediction accuracy. Section 4 characterizes the Case Study. Section 5 presents the Empirical Research Result. Finally Section 6 concludes the paper.
2. Methodology
Initially, the ranking technique is applied to find the potential causes of at least 75% of the events registered in the historical time series of injured people in traffic accidents in Santiago de Chile. The causes related to drivers, pedestrians, and passengers behavior were prioritized.
The forecasting methodology applied in the analyzed hybrid models is described in two stages, preprocessing and prediction, as Figure 1 illustrates. In the preprocessing stage Singular Spectrum Analysis and Singular Value Decomposition of Hankel are used to extract an additive component of low frequency from the observed time series, and by simple subtraction between the observed time series and the component of low frequency, the component of high frequency is obtained. In the prediction stage, linear and nonlinear models are implemented.
2.1. Singular Spectrum Analysis
Singular Spectrum Analysis extracts the component of low frequency from the observed time series. The component of high frequency is obtained by simple subtraction between the observed time series and the component . Consider Conventional SSA is implemented in four steps, embedding, decomposition, grouping, and diagonal averaging [24]. In this work, SSA is simplified in three steps: embedding, decomposition, and diagonal averaging.
The embedding step maps the time series of length to a sequence of multidimensional lagged vectors; the Hankel matrix structure is used in the embedding: where the elements . The window length has an important role in the forecasting model; it has an initial value of , and is computed as follows:The Singular Value Decomposition of the real matrix has the form where each is the th eigenvalue of the matrix arranged in decreasing order of magnitudes. is the corresponding orthonormal eigenvectors system of the matrix .
Standard SVD terminology calls the th singular value of the matrix ; is the th left singular vector and is the th right singular vector of . The collection is called th eigentriple of the SVD.
The computation of the optimal window length is based on the eigenvalues differential entropy. The first eigenvalues obtained with contain a high spread of energy; therefore, the window length is evaluated in the range by means of the differential entropy as follows: where is the th differential entropy, is the th Shannon entropy, and is the th normalized eigenvalue also known as eigenvalue energy. The embedding step is executed again with this decomposition that reaches a high energy spread and lower differential entropy. With the optimal the embedding and decomposition are computed again.
The first eigentriple is used to obtain the elementary matrix , which will be used in the extraction of the low frequency component: The step of diagonal averaging is applied over to extract the elements of the component ; the process is shown below: Once is obtained, the component is computed with (1).
2.2. Hankel Singular Value Decomposition
The preprocessing based on Singular Value Decomposition of Hankel is implemented in three steps: embedding, decomposition, and extraction. HSVD implements the steps of embedding and decomposition as SSA (presented in Section 2.1). The elementary matrix is also computed with the first eigentriple obtained in the decomposition step (as (6)).
In the extraction step, the elements of the low frequency component are obtained from the first row and the last column of the matrix , which has the same structure as matrix (trajectory matrix); therefore, the elements of are where is a matrix.
2.3. Prediction with the Autoregressive Method
The prediction is the second stage of the traffic accidents forecasting methodology (illustrated in Figure 1). In order to obtain the traffic accidents prediction , during the preprocessing stage the low frequency component and the high frequency component were obtained. The components are estimated through the autoregressive method and the addition of the components is computed to obtain the prediction as follows: where represents the time instant and represents the horizon, with values . The component is used as exogenous variable in the computation of the , due to a high influence of over .
The predicted components via AR model are defined with where is the number of lagged values and and are the coefficients of and , respectively.
The coefficients estimation is based on linear Least Square Method (LSM); the components and are defined with the linear relationship expressed in matrix form: where and are the regressor matrices of and , respectively; and are the coefficients vectors of and , respectively. The coefficients are computed with the MoorePenrose pseudoinverse matrices, and , as follows:
2.4. Prediction with the Autoregressive Neural Network
In this case study, a single hidden layer Autoregressive Neural Network is used to approach each component; the ANN has a standard multilayer perceptron (MLP) structure of three layers [28]. The training subset is iteratively used to adjust the connections weights via learning algorithm; the ANN with the lowest error is selected to implement the solution with the testing subset.
The nonlinear inputs are the lagged terms, which are contained in the regressor matrix ; at hidden layer is applied the sigmoid transfer function, and at output layer is obtained the prediction. The ANN output is where is the predicted value, is the time instant, and are the linear and nonlinear weights of the ANN connections, respectively; the sigmoid transfer function is computed with The ANN structure for prediction is denoted with , with inputs, hidden nodes, and 1 output while the ANN structure for is denoted with , with inputs, hidden nodes, and 1 output . LevenbergMarquardt is the learning algorithm applied for weight updating in both neural networks [29].
3. Efficiency Criteria
The forecasting accuracy is evaluated with conventional metrics and an improved evaluation metric. The conventional metrics are Mean Absolute Percentage Error (MAPE), Root Mean Square Error (RMSE), Determination Coefficient , and Relative Error . The Modified NashSutcliffe Efficiency (MNSE) metric is computed in order to improve the evaluation criteria, which is sensitive to significant overfitting or underfitting [30]. Consider where is the observed signal, is the predicted signal, is the testing sample size, and is the variance.
Furthermore, two statistical tests are computed to evaluate the differences and superiorities of either model, the Wilcoxon test and Pitman’s correlation test, respectively.
The Wilcoxon () signed rank test evaluates the pairwise differences in the squares of each multistep ahead residuals; the differences are ranked in ascending order, with no regard to the sign, and the ranks are assigned from one to the number of the forecast errors available for comparison. The sum of the ranks of positive differences is then computed to obtain [31]. The probability of finding a test statistic as or more extreme than the observed value under the null hypothesis is found using the statistics given by Pitman’s correlations test is applied to identify the superiority of a model in pairwise comparisons [32]; the test is based on the computation of the correlation between and as follows: where is the covariance and . is the residual vector obtained with model 1, and is the residual vector obtained with model 2. Model 1 would show superiority in front of model 2 at significance level if .
4. Case Study
The Chilean police (Carabineros de Chile) collects the features of the traffic accidents, and CONASET records the data. The regional population is estimated in 6,061,185 inhabitants, equivalent to 40.1% of the national population. Santiago shows a high rate of the events with severe injuries from 2000 to 2014, with 260,074 injured people.
The entire series of 783 registers is shown in Figure 2, which have been collected in the mentioned period from January to December, with weekly sampling. The highest number of injured people was observed in weeks 84, 184, 254, and 280, while the lowest number of injured people was observed in weeks 344, 365, 426, 500, and 756.
One hundred causes of traffic accidents have been defined by CONASET and grouped into categories. In this case study, three analysis groups have been created. In groups 1 and 2, those causes directly related to behavior of drivers, passengers, or pedestrians have been prioritized. Group 3 involves the rest of the causes. Figures 3(a), 4(a), and 5(a) show the observed time series of the groups, InjuredG1, InjuredG2, and InjuredG3, respectively; the values have been normalized via division by the maximum value in each time series.
(a)
(b)
(c)
(a)
(b)
(c)
(a)
(b)
(c)
The categories of groups 1 and 2 are as follows: reckless driving, recklessness in passenger, recklessness in pedestrian, alcohol in driver, alcohol in pedestrian, and disobedience to signal. In Table 1 are shown 20 causes of groups 1 and 2, which cover of the events with injured people. The causes are listed sequentially; the cause with the highest importance has value 1 (with the highest number of injured people), and the cause with minor importance has value 20.

From Table 1, the first three causes of injuries in traffic accidents are as follows: unwise distance, inattention to traffic conditions, and disrespect to red light. The cause with the lowest importance for injuries in traffic accidents is drunk pedestrian. The categories with the highest number of causes are as follows: imprudent driving and disobedience to signal.
Two groups of analysis were formed from the information presented in Table 1; the first group labeled with InjuredG1 corresponds to the first ten primary causes, and the second group labeled with InjuredG2 corresponds to the ten secondary causes. The first group overspreads around 60% of injured people in traffic accidents, whereas the second group overspreads around 15%. The third group, labeled with InjuredG3, corresponds to the categories road deficiencies, mechanical failures, and undetermined/noncategorized causes; this group overspreads around 25% of injured people.
Complementary information was observed about traffic accidents conditions with high rate of injured people. With regard to vehicles, automobiles are involved in 54% of events, followed by vans and trucks with 19%, bus and trolley with 16%, motorcycles and bicycles with 8%, and others with 3%. With regard to environmental conditions, 85% of events was observed with cloudless conditions. Additionally, 97% of the traffic accidents with injured people have taken place in urban areas, whereas 3% correspond to rural area. With regard to relative position, 46% of the events have been produced in intersections controlled by traffic signals or police officers with 46%, followed by accidents that happened in straight sections with 37%, and other relative positions with 17%.
5. Empirical Research Result
The results of the methodology implementation with linear and nonlinear models are described by stages: components extraction and prediction.
5.1. Components Extraction
The methodology presented in Section 2 describes the preprocessing stage and the prediction stage. The preprocessing stage is based on two types of methods, Singular Spectrum Analysis and Singular Value Decomposition of Hankel.
Both preprocessing techniques SSA and HSVD embed the time series in a structure of two dimensions; the initial window length used is . Once matrix is obtained, the decomposition is computed. The differential energy of the eigenvalues is obtained with (5a). A high energy content was observed in the first eigenvalues; the lowest differential energy was observed in decompositions based on values of 15, 17, and 16 of window length, for InjuredG1, InjuredG2, and InjuredG3, respectively, and these values were set as window length.
The embedding process is implemented again with the optimal window , and the decomposition is recomputed. The first elementary matrix is used by SSA and HSVD to obtain the low frequency component. In SSA, to extract the elements of , diagonal averaging is applied, while HSVD uses direct extraction from the first row and last column of . Finally with (1) the component of high frequency was computed.
Each data set of low and high frequency has been divided into two subsets, training and testing; the training subset involves of the samples, and consequently the testing subset involves the remaining .
The decomposition results by means of SSA and HSVD. Figures 3(b), 4(b), and 5(b) show the low frequency components, whereas Figures 3(c), 4(c), and 5(c) show the high frequency components of InjuredG1, InjuredG2, and InjuredG3, respectively.
The nonstationary trend in the signals was verified for InjuredG1, InjuredG2, and InjuredG3 through Kwiatkowski, Phillips, Schmidt, and Shin test (KPSS) [33]. The test assesses the null hypothesis that the signals are trend stationary; it was rejected at 5% significance level; consequently, nonstationary unitroot processes are present in the signals.
The time series InjuredG1 is observed in Figure 3(a); similar dynamic is observed with respect to full series, taking into consideration that this group contains the predominant causes. The components extracted by SSA and HSVD are shown in Figure 3(b); longmemory periodicity features were observed. The resultant components are presented in Figure 3(c); shortterm periodic fluctuations were identified. The components obtained with SSA and HSVD are similar; however, a slight difference is observed in the components of low frequency; SSA extracts smoother components than HSVD.
Both techniques SSA and HSVD show that the principal ten causes (in Table 1 with importance 1 to 10) present the highest incidence between years 2002 and 2005 (weeks 106 to 312); it descends from 2006 until half 2012 (around week 710); an increment is observed between weeks 711 and 732 (second semester of 2013 and first semester of 2014).
Figure 4 shows the time series InjuredG2, the low frequency component, and the high frequency component. As previous series, the components of low frequency show similar dynamic of slow fluctuations with decreasing trend (), while the high frequency shows fast fluctuations. In this group is also observed via SSA smoother than via HSVD. Both techniques SSA and HSVD show that the secondary ten causes (described in Table 1 with importance 11 to 20) present the highest incidence of injured people in years 2000, 2003, and 2004; it descends from 2005; therefore, forward downtrend is observed in the number of injured people in traffic accidents due to the 10 secondary causes.
Figure 5 shows InjuredG3 and its components and ; as in previous analysis the components of low frequency show longmemory periodicity features, whereas the components of high frequency show shortterm periodic fluctuations, and via SSA is smoother than via HSVD.
Both techniques SSA and HSVD show that the causes are related to road deficiencies, mechanical failures, and undetermined/noncategorized causes (with 25% of incidence). The highest incidence is observed in year 2001, from year 2002 it presents strong decay until 2006, and forward uptrend is observed with a temporal decrease in 2009.
Prevention plans and punitive laws have been implemented in Chile during the analyzed period, via education, drivers licensing reforms, zero tolerance law, Emilia’s law, and transit law reforms, among others. The effect of a particular preventive or punitive action is not analyzed in this work; however, the proposed shortterm prediction methodology based on observed causes and intrinsic components is a contribution to government and society in preventive plans formulation, its implementation, and the consequent evaluation.
5.2. Prediction
The prediction is implemented by means of the autoregressive models, linear (AR) or nonlinear (ANN). The models use the lagged terms of the components and ; the optimal number of the lagged terms was fixed in weeks, which was found through the computation of the Autocorrelation Function over the observed time series of injured people due to all causes.
The models based on Singular Spectrum Analysis (SSAAR and SSAANN) receive the components of SSA preprocessing stage, whereas the models based on Singular Value Decomposition of Hankel (HSVDAR and HSVDANN) receive the components of HSVD preprocessing stage.
The ANN has single hidden layer structure , with 32 inputs, 1 hidden node, and 1 output. The LM algorithm was used iteratively to adjust the linear and nonlinear weights.
The direct method was used to develop multistep ahead forecasting; in Tables 2 and 3 are presented the average prediction results with hybrid models SSAAR, SSAANN, HSVDAR, and HSVDANN with the three time series. The arithmetic mean of the resultant metrics is presented in Tables 2 and 3; the results shows that the accuracy decreases as the time horizon increases; therefore, the best accuracy was obtained for the nearest weeks, and the lowest accuracy was obtained for the farthest weeks. The best mean accuracy was reached by using SSAAR model, with of 92.6%, of 99.3%, of , and of . The lowest mean accuracy was obtained with HSVDANN model, with of 85.6%, of 98.2%, of , and of . The second best average accuracy was reached by SSAANN, and the third best accuracy was reached by HSVDAR.


The highest gain in average from 1 to 14step ahead prediction is 7.6%, while average is 93.3%. However, it was observed that HSVDAR shows a higher accuracy in farthest horizons, from 12 to 14step ahead prediction, which reaches these average results: of 89.4%, of 98.9%, of 2.2%, and of 1.0%; the gain in average from 12 to 14step ahead prediction is 12.1%, and on average is 83.3%.
Prediction horizons higher than 14 weeks provide inaccurate results.
From previous tables, similar accuracy was identified in the prediction through SSAAR, HSVDAR, and SSAANN, for 11step ahead prediction. The predicted signals are shown in Figures 6, 7, and 8, whereas the metrics residuals are presented in Tables 4, 5, and 6.



(a)
(b)
(c)
(d)
(e)
(f)
(a)
(b)
(c)
(d)
(e)
(f)
(a)
(b)
(c)
(d)
(e)
(f)
The results for 11step ahead prediction of InjuredG1 are shown in Figures 6(a), 6(c), 6(e), and Table 4. From these figures and metrics, a good fit is observed; the highest accuracy was reached via SSAAR with of 87.6%, of 98.6%, of 2.0%, and of 1.1%.
In Figures 6(b), 6(d), and 6(f), the Relative Error of InjuredG2 prediction is shown. The model SSAAR shows 94.7% of the predicted points with Relative Error lower than , HSVDAR 86.7%, and SSAANN 83.1%. The gain was computed by means of residual metrics and the two best models (SSAAR and HSVDAR); the highest gain was observed in with .
The results for 11step ahead prediction of InjuredG2 are shown in Figures 7(a), 7(c), and 7(e) and Table 5; all models achieve a good fit; SSAAR and HSVDAR reach the highest and similar accuracy. In Figures 7(b), 7(d), and 7(f), the Relative Error is shown; SSAAR presents 85.8% of the predicted points with a Relative Error lower than , HSVDAR 85.3%, and SSAANN 80%. The gain was computed based on residual metrics and the two best models; the highest gain was observed in with .
The results for 11step ahead prediction of InjuredG3 are presented in Figures 8(a), 8(c), and 8(e) and Table 6; all models achieve also a good fit and similar accuracy. In Figures 8(b), 8(d), and 8(f), the Relative Error is illustrated; SSAAR presents 92.4% of predicted points with a Relative Error lower than , HSVDAR 94.2%, and SSAANN 91.6%. The gain was computed based on residual metrics and the two best models; the highest gain was observed in with .
In the next section the differences and/or superiorities of either linear model SSAAR or HSVDAR are identified through the application of the statistical tests.
5.3. Models Statistical Tests
The performance of the linear hybrid models SSAAR and HSVDAR is evaluated with the Wilcoxon hypothesis test and with Pitman’s correlations test ((16)–(17a), (17b), and (17c)). The Wilcoxon hypothesis test results are shown in Table 7, and Pitman’s correlation test results are shown in Table 8.


From Table 7, in 37 comparisons between residuals of SSAAR and HSVDAR, the test rejects the null hypothesis that there is no difference in the prediction at 5% significance level. In the remaining 5 comparisons the null hypothesis that there is no difference in the prediction is accepted. In this case, there is no difference in 12step ahead prediction for time series InjuredG1; the same situation was found in 10 and 11step ahead prediction for time series InjuredG2 and InjuredG3.
Pitman’s correlations test is applied with the residual values to identify the superiority of SSAAR over HSVDAR or the opposite. The correlations between and are shown in Table 8.
The null hypothesis of Pitman’s correlation is true at 5% significance level if , where testing samples. The results of the correlations are shown in Table 8. From Table 8, the results present similarities with respect to Wilcoxon test. In 5 predictions there is no superiority of either model (SSAAR and HSVDAR). SSAAR shows superiority with respect to HSVDAR in 30 predictions (when for nearest horizons), whereas the opposite is observed in 7 predictions; HSVDAR shows superiority with respect to SSAAR (when for farthest horizons).
6. Conclusions
In this paper has been developed multistep ahead traffic accidents forecasting approach based on singular values and autoregressive models. The nonstationary and nonlinear time series of injured people in traffic accidents of Santiago de Chile was used.
Before the models methodology stages, ranking was applied to detect the relevant causes of injuries in traffic accidents; causes related to behavior of drivers, pedestrians, or passengers are predominant. Unwise distance, inattention to traffic conditions, and disrespect to red light are the first important causes of injuries in traffic accidents in concordance with previous studies that determine disrespect towards the road signs as a principal cause of traffic accidents. Complementary information was observed about traffic accidents conditions with high rate of injured people, automobiles type, environmental conditions, and relative position, among others.
This approach was described in two stages, preprocessing and prediction; in the first stage two methods for components extraction were developed, Singular Spectrum Analysis and Singular Value Decomposition of Hankel, whereas in the second stage the linear autoregressive model and an Autoregressive Neural Network with LevenbergMarquardt algorithm were used.
Four hybrid models were implemented: SSAAR, HSVDAR, SSAANN, and HSVDANN. The models were evaluated for 14week ahead forecasting; comparative analysis shows that the proposed models SSAAR and SSAANN achieved the highest accuracy with an average of 92.6% and 90.3%, respectively; the highest gain in average achieved by SSAAR is 7.6%. However, it was observed that HSVDAR shows a higher accuracy in farthest horizons from 12 to 14 steps, which reaches an average of 89.4%; in this case the highest gain achieved by HSVDAR in is 12%.
The statistical tests application through Wilcoxon and Pitman has shown that SSAAR is superior to HSVDAR in 30 of 42 comparisons of resultant efficiency criteria (at nearest horizon) at 5% significance level, 5 comparisons show equivalence, and 7 comparisons show the superiority of HSVDAR over SSA (at farthest horizon).
In further works, more strategies of components extraction will be explored; spectral analysis could help to explain the nature of traffic accidents in other geographic zones. Detailed work focused on the causes of traffic accidents will be done to support prevention plans aimed at promoting good habits on roads and highways.
Competing Interests
The authors declare that they have no competing interests regarding the publication of this paper.
References
 World Health Organization, 2015, http://www.who.int.
 J. Abellán, G. López, and J. de Oña, “Analysis of traffic accident severity using decision rules via decision trees,” Expert Systems with Applications, vol. 40, no. 15, pp. 6047–6054, 2013. View at: Publisher Site  Google Scholar
 L.Y. Chang and J.T. Chien, “Analysis of driver injury severity in truckinvolved accidents using a nonparametric classification tree model,” Safety Science, vol. 51, no. 1, pp. 17–22, 2013. View at: Publisher Site  Google Scholar
 J. De Oña, G. López, R. Mujalli, and F. J. Calvo, “Analysis of traffic accidents on rural highways using Latent Class Clustering and Bayesian Networks,” Accident Analysis and Prevention, vol. 51, pp. 1–10, 2013. View at: Publisher Site  Google Scholar
 Y.R. Shiau, C.H. Tsai, Y.H. Hung, and Y.T. Kuo, “The application of data mining technology to build a forecasting model for classification of road traffic accidents,” Mathematical Problems in Engineering, vol. 2015, Article ID 170635, 8 pages, 2015. View at: Publisher Site  Google Scholar
 L. Lin, Q. Wang, and A. W. Sadek, “A novel variable selection method based on frequent pattern tree for realtime traffic accident risk prediction,” Transportation Research Part C: Emerging Technologies, vol. 55, pp. 444–459, 2015. View at: Publisher Site  Google Scholar
 M. Deublein, M. Schubert, B. T. Adey, J. Köhler, and M. H. Faber, “Prediction of road accidents: a Bayesian hierarchical approach,” Accident Analysis and Prevention, vol. 51, pp. 274–291, 2013. View at: Publisher Site  Google Scholar
 S. Ognjenovic, R. Donceva, and N. Vatin, “Dynamic homogeneity and functional dependence on the number of traffic accidents, the role in urban planning,” in Proceedings of the International Scientific Conference Urban Civil Engineering and Municipal Facilities (SPbUCEMF '15), vol. 117, p. 551, 2015. View at: Google Scholar
 K. Keay and I. Simmonds, “Road accidents and rainfall in a large Australian city,” Accident Analysis and Prevention, vol. 38, no. 3, pp. 445–454, 2006. View at: Publisher Site  Google Scholar
 M. Aron, R. Billot, N. E. Faouzi, and R. Seidowsky, “Traffic indicators, accidents and rain: some relationships calibrated on a French urban motorway network,” Transportation Research Procedia, vol. 10, pp. 31–40, 2015. View at: Publisher Site  Google Scholar
 Comisión Nacional de Seguridad de Tránsito, 2015, http://www.conaset.cl.
 H. Viljoen and D. G. Nel, “Common singular spectrum analysis of several time series,” Journal of Statistical Planning and Inference, vol. 140, no. 1, pp. 260–267, 2010. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 C. A. F. Marques, J. A. Ferreira, A. Rocha et al., “Singular spectrum analysis and forecasting of hydrological time series,” Physics and Chemistry of the Earth, Parts A/B/C, vol. 31, no. 18, pp. 1172–1179, 2006. View at: Publisher Site  Google Scholar
 L. Telesca, M. Lovallo, A. Shaban, T. Darwich, and N. Amacha, “Singular spectrum analysis and FisherShannon analysis of spring flow time series: An application to Anjar Spring, Lebanon,” Physica A, vol. 392, no. 17, pp. 3789–3797, 2013. View at: Publisher Site  Google Scholar
 H. Hassani, S. Heravi, and A. Zhigljavsky, “Forecasting European industrial production with singular spectrum analysis,” International Journal of Forecasting, vol. 25, no. 1, pp. 103–118, 2009. View at: Publisher Site  Google Scholar
 H. Hassani, A. Webster, E. S. Silva, and S. Heravi, “Forecasting U.S. tourist arrivals using optimal singular spectrum analysis,” Tourism Management, vol. 46, pp. 322–335, 2015. View at: Publisher Site  Google Scholar
 J. Wang, S. Jin, S. Qin, and H. Jiang, “Swarm intelligencebased hybrid models for shortterm power load prediction,” Mathematical Problems in Engineering, vol. 2014, Article ID 712417, 17 pages, 2014. View at: Publisher Site  Google Scholar
 H. Li, L. Cui, and S. Guo, “A hybrid shortterm power load forecasting model based on the singular spectrum analysis and autoregressive model,” Advances in Electrical Engineering, vol. 2014, Article ID 424781, 7 pages, 2014. View at: Publisher Site  Google Scholar
 W. Zhang, Z. Su, H. Zhang, Y. Zhao, and Z. Zhao, “Hybrid wind speed forecasting model study based on SSA and intelligent optimized algorithm,” Abstract and Applied Analysis, vol. 2014, Article ID 693205, 14 pages, 2014. View at: Publisher Site  Google Scholar
 Y. Xiao, J. J. Liu, Y. Hu, Y. Wang, K. K. Lai, and S. Wang, “A neurofuzzy combination model based on singular spectrum analysis for air transport demand forecasting,” Journal of Air Transport Management, vol. 39, pp. 1–11, 2014. View at: Publisher Site  Google Scholar
 M. Abdollahzade, A. Miranian, H. Hassani, and H. Iranmanesh, “A new hybrid enhanced local linear neurofuzzy model based on the optimized singular spectrum analysis and its application for nonlinear and chaotic time series forecasting,” Information Sciences, vol. 295, pp. 107–125, 2015. View at: Publisher Site  Google Scholar
 L. Barba, N. Rodríguez, and C. Montt, “Smoothing strategies combined with ARIMA and neural networks to improve the forecasting of traffic accidents,” The Scientific World Journal, vol. 2014, Article ID 152375, 12 pages, 2014. View at: Publisher Site  Google Scholar
 G. H. Golub and C. F. V. Loan, Matrix Computations, The Johns Hopkins University Press, 1996.
 N. Golyandina, V. Nekrutkin, and A. A. Zhigljavsky, Analysis of Time Series Structure, Chapman & Hall/CRC, 2001.
 J. Elsner and A. Tsonis, Singular Spectrum Analysis: A New Tool in Time Series Analysis, Plenum, 1996.
 H. Hassani, R. Mahmoudvand, and M. Zokaei, “Separability and window length in singular spectrum analysis,” Comptes Rendus Mathematique, vol. 349, no. 1718, pp. 987–990, 2011. View at: Publisher Site  Google Scholar  MathSciNet
 R. Wang, H.G. Ma, G.Q. Liu, and D.G. Zuo, “Selection of window length for singular spectrum analysis,” Journal of the Franklin Institute, vol. 352, no. 4, pp. 1541–1560, 2015. View at: Publisher Site  Google Scholar  MathSciNet
 J. A. Freeman and D. M. Skapura, Neural Networks, Algorithms, Applications, and Programming Techniques, AddisonWesley, 1991.
 M. Hagan, H. Demuth, and M. Bealetitle, Neural Network Design, Hagan Publishing, 2002.
 P. Krause, D. P. Boyle, and F. Bäse, “Comparison of different efficiency criteria for hydrological model assessment,” Advances in Geosciences, vol. 5, pp. 89–97, 2005. View at: Publisher Site  Google Scholar
 R. L. Ott and M. Longnecker, An Introduction to Statistical Methods and Data Analysis, Cengage Learning, Boston, Mass, USA, 6th edition, 2001.
 K. Hipel and A. McLeod, Time Series Modelling of Water Resources and Environmental Systems, Elsevier, 1994.
 D. Kwiatkowski, P. C. B. Phillips, P. Schmidt, and Y. Shin, “Testing the null hypothesis of stationarity against the alternative of a unit root: how sure are we that economic time series have a unit root?” Journal of Econometrics, vol. 54, no. 1–3, pp. 159–178, 1992. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2016 Lida Barba and Nibaldo Rodríguez. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.