Abstract

Banks, financial, and credit institutions encountering the weakening financial system and increased risk factors cause high inflation and great losses for an economy. Detecting financial risks in advance could help financial institutions avoid losses, and the financial system could be eventually affected less. Early warning systems for banks could be helpful to identify financial risks and take measures to deal with hazardous situations. Various approaches have already been put forward. However, inaccuracy issues in risk detection are one of the main issues. Combining semantic hierarchy with the GMDH neural network to predict financial risks is proposed. A semantic hierarchy approach based on converting risk-related values and picking influential variables could be practical in risk detection. Besides, the GMDH algorithm utilizing neural networks based on available data has the capability of predicting possible risks that could occur in the future. The outcomes of the proposed method when compared to non-data mining methods suggest that it improves accuracy by almost 20%.

1. Introduction

With the globalization of economies and intensifying competition among banks, profit margins decreased and the risk factors increased [1]. Weaknesses in the banking system could occur due to possible reasons of boycotts imposed by the international community towards banking system, reduced government support, crisis risk, losses from transactions, exchange rate fluctuations and debt, increased market uncertainty, excessive asset holdings, cash balances on balance sheets, rising bankruptcy trends of large credit companies, and innovations of monetary and financial instruments [2]. Both risk management and forecasting play significant roles. Therefore, the risk detection of financial institutions has received much attention [3]. Financial alerting systems to assess risks for financial institutions have been a research area [4].

The current financial crisis has fueled the initiatives of policymakers to construct alert systems to predict financial crises in advance, assuming that these systems could function and provide signals based on models and forecasts utilizing some indicators before a crisis could occur. The past crises proved to be very costly for both advanced and emerging economies [5]. Besides, the world economic system could be vulnerable and fragile. Moreover, each crisis having a unique character caused by such situations ranging from political unrest to shock of the trade market could lead to two consequences. While the first one is characterized by not being predictable, the second one is characterized by the complexity of its nature. On the other hand, convincing policymakers to take precautionary measures against any possible crisis is so difficult. Economically and politically costly measures do not take effect until a crisis develops. Precautionary measures have to be taken to prevent economic damage, and policies should be implemented to address it. Any warning system is expected to alert policymakers before a crisis occurs, so measures that are implemented could prevent it from spreading through the economy. Various methods to reduce damage have been proposed, but a series of measures could hardly be welcomed. For example, although homeowners prefer their home values to rise rapidly, financial institutions are reluctant to put more capital aside. How crises could spread through economies is reported in [6]. Consequently, any warning system could inform financial institutions about imminent dangers and help strengthen them. Hence, it allows policymakers to take precautions before a crisis.

Various methods have been implemented to forecast crises. However, these approaches have drawbacks such as inadequate accuracy and high-risk prediction. Due to the issues that existed in the previous methods, this manuscript proposes a novel method combining the GMDH deep neural network with semantic technique. Therefore, the proposed method is utilized to generate a financial alert system. The rest of the paper is organized as follows: Section 2 summarizes various models dealing with financial warning systems and risk-based management. The GMDH neural network is presented in Section 3. The proposed method is introduced in Section 4. The results are presented in Section 5, and Section 6 presents the conclusion.

2. Financial Warning Systems

An early warning system (EWS), which is a monitoring, reporting, and alerting system, forecasts quantitatively success levels, probable anomalies, issues, risks, potential bankruptcies, and transactions affairs [4]. Hence, avoiding or mitigating potential issues could be possible by EWSs. Financial statements such as balance sheets and income tables providing financial information are utilized in EWSs that are, in essence, analysis techniques identifying current states of financial intuitions.

In the construction of an EWS after a crisis is defined, which types of events are related to a crisis should be identified. Then, constructing an analytical model could systematically determine possible economic damage. Afterward, it is expected that the paths from which a crisis begins and penetrates other sectors are determined. Finally, favorable conditions in the financial sector should be determined. An effective and efficient alert system includes several features such as the ability to chart the outlook for global finance and macroeconomics, assess the economic impact of countries and sectors, and analyze paths of impacts in sectors. Various financials are noted in countries [7]. Policymakers must comprehend how these EWSs function to prevent crises. However, constructing such systems is overlooked since policymakers doubt their efficiency. In addition to operating these systems, many appropriate and effective policies should be considered to reduce risks. Global communication, for instance, is very important in this regard. Admittedly, no such systems perfectly signal before any crisis occurs. False alarms could be possible [8, 9]. Besides, factors such as efficient leading-edge analysis, precrisis scrutiny, and widespread global coordination are essentials [10]. Recently, research has been conducted on risk pacing by EWSs.

A novel random forest-based alert system investigating bank-level financial statements to determine patterns that put banks at risk was proposed, and it performs better than the conventional methods [11].

To determine the monetary misery of individual banks, AI was proposed in [12]. A decision tree involving subjective financial area and macroeconomic provisions was constructed by Quinlan’s C5.0 computation to assess a model, whose final structure contains 12 elements and 19 hubs when compared to the benchmark model of a logit.

The hidden Markov model, Dynamic Linear Switching System model, and Simple Linear Dynamics System model were employed under the title of Dynamic Bayesian Networks utilizing the Markovian framework. The likelihood of an imminent crisis based on given dynamics could be computed. When compared with other methods of signal extraction and logit, it could be better utilized as EWS [13].

Monetary emergencies pertinent to small-scale and open economies employ large-scale powerful factor models that extract useful data out of bulk datasets and provide a mixed recurrence index merging worldwide and homegrown monetary policies with monetary markers. Thailand, South Korea, Singapore, Malaysia, Philippines, and Indonesia were examined by the model. The extracted elements and other driving markers utilized by the logit replace the model in foreseeing foundational occasions that have a great impact. The usefulness of the extracted components on provident control over a model employing sufficiently long history pointers is shown by in-test and out-of-test results [14].

Predictive methods as EWSs utilized in a large undergraduate STEM course were examined. Research has currently shed light on harnessing them as tools for identifying students at risk.

Student interaction with Learning Management Systems (LMSs) was utilized. Eight prediction methods and the optimal time of implementing an EWS were investigated. Findings pertinent to a statistics course at the university level were reported since LMSs had all resources related to the course and enabled weekly assessments that helped students find their study pattern [15].

A combination of models used as an EWS to exhibit scholarly studies of undergraduates was proposed. Moodle collaboration information, attributes, and grades of 802 undergraduates from a CO college were utilized. Four components, Access, Questionnaire, Task, and Age, were found to be significant to expect scholarly studies from undergraduates. Moodle helped identify factors composing Access such as visits to discussions and glossaries. Factors identified by visits and endeavors in polls constitute the Questionnaire. Task is composed of factors identified with counseled and submitted undertakings. Age was strikingly distinguished as a negative indicator. Furthermore, cluster examination suggested some connections with Moodle that are firmly identified [16].

Two methods, linear regression and family-based data analysis [17,18], generate better risk predictions for banks when compared to other methods.

Adjustments of financial indicators under emergency might affect the data extraction of EWSs in the state of a strategic relapse when genuine situations of foundational banking emergencies occur. Considering possible limits of the regular EWSs and properties of the AI calculation, the EWSs utilizing experts’ voting structure could likely fit into foundational banking emergencies. The irregular wood classifier taking into account experts’ voting process is the most effective among AI classifiers. The experts’ voting EWS orchestrating multivariate data could be more appropriate to provide alarms when shifted settings exist [19].

The credit-based variables as EWSs in the emerging economies were examined. The bank-related data such as collected total credit to the private sector was assessed as an indicator utilizing the Accuracy Criterion (AUC). The nominal credit growth and the change in the credit to GDP ratios are found to be the best beckoning attributes. The nominal credit growth significantly outperformed the credit-to-GDP gap in almost all specifications. These findings presented converse outcomes compared with the results of the East-West Institute since the credit-to-GDP gap is the best performance in the developed economies [20].

An incorporated EWS predicting the disturbance of the securities exchange was proposed in [21]. The proposed EWS initially characterizes securities exchange emergencies concerning a marker utilizing switching ARCH that separates probabilities of the great unpredictability system. A crossover calculation is employed to make everyday forecasts. In the observational assessment based on ten-year Chinese stock information, the proposed EWS produces better outcomes. The linear regression producing higher accuracy than other methods was reported by [22].

ANNs utilizing various sampling schemes were introduced. Three models were employed based on past stream scenes containing high convergences of airborne particulate matter. Model 3 was better in forecasting estuarial dust occasions when precisely taking into account the qualities of media and contamination origins. Hence, it helped avoid preemptive fiasco mitigated measures to stay away from the antagonistic wellbeing results of residue storms inside estuarial districts since the proposed model efficiently provided better estimates [23].

Precrisis periods in the Eurozone utilizing the EWS were researched, which led to a novel proposed model to deal with uncertainty. A set of macroeconomic risk indices was utilized to pick better ones. Then, four indices, as explanatory variables, were employed in all types of logit models to assess precrisis probabilities [24].

The underlying indicators triggering the Islamic banking crisis were examined by [25]. The method called Markov Switching Linear Regression (MS-VAR) utilizes monthly data gathered from the official database covering January 2004 to March 2017. It was concluded that Islamic banking is more resistant to domestic and foreign risks than conventional banking. When Islamic banks are compared to conventional banks, the former’s z-score is greater (11.933 > 11679) [25].

Measuring the level of financial fixing, monetary market circumstances, and fundamental monetary danger based on the adequacy and exactness of monetary status pointers led the relapse strategy to be utilized so the determination of significant pointers could be obtained. A unique weighted monetary condition based on the time-differing boundary vector autoregressive model consists of five factors, namely, loan fee, land value, cash supply, swapping scale, and stock value, which adequately mirror the real monetary circumstance. China’s monetary circumstance from 2013 to 2017 was analyzed by this strategy. Changes in monetary markers help estimate the level of money fixing and monetary economic situations in advance [26].

The available research specific to China examining mechanisms of the industry environment found a profound impact on corporate financial risk and investigated seven dimensions of industry environmental risk. An index system was constructed employing Structural Equation Model. The development of the financial risk was found to be highly related to the industry environment. EWSs classify seven subsystems and are capable of identifying potential financial risks, which is conducive to developing effective financial policies concerning external changes [27].

Banking and money emergencies were an impressive expense for homegrown and worldwide monetary frameworks. EWSs were particularly investigated in the turn of events and a combination of conditions in monetary fields. Foreseeing and recognizing antagonistic and risky occasions in the fields of business, money, and finance were contemplated [28].

Prospective CoVaR as a market-based measurement of systemic risk was proposed in emerging markets. The presented ΔCoVaR values represented a network of asset exposures. Two novel network-based indicators were proposed based on exposure networks. Besides, they were utilized as EWS to predict recession and crisis at the enterprise and entire market levels utilizing the relevance of new indicators. The outcomes of the EWS for market returns suggest that models predict a market crisis, which provides an alarm signal with up to 7 periods before the crisis. The implementation of those helped policymakers develops appropriate macro and micro policies to prevent systemic risks [29].

A logistic model predicting bank failures based on most of the downward trends of NIC banks was investigated. G8 banks performed well in growth phases, although the performances of the NIC banks were poor in downturns [30].

3. GMDH Neural Network

Deep learning neural network implemented in financial risk forecasting has four types of layers, as shown in Figure 1.

3.1. Input Layer

The parameters are normalized before the semantic aspect is applied. The input layer represents the input data obtained from feature extraction.

3.2. Latent Layers

Training of neural networks in a supervised manner needs initial weights assigned by a Gaussian distribution, and backpropagation algorithm updates them, which is a slow process and is trapped in a local minimum. Pretraining a layer by a nonsupervised method could resolve this issue. Therefore, deep learning neural network supplemented with this feature will be constructed.

3.3. Output Layer

Two types of output layers are utilized, which are called Output Layer 1 and 2. The output layer called Output Layer 1 is employed to rebuild a network, called input and utilized for training; extract information from the last layer of latent layers; and employ backward error propagation. On the other hand, Output Layer 2, called a softmax layer, is embedded in a softmax approach to conduct a regression. Probabilities of belonging to each class are obtained.

3.4. Softmax Regression Function

The logistic regression has a nonlinear structure that combines the attributes in a linear form called the probit function. The response variable takes binary values represented by

The response variable is a Bernoulli-type random variable Y whose probability is denoted by η. The probability of success depends on the predictor; i.e., η = η (x). If X is assumed to be a regularizing predictor random variable, η (x) is a conditional expectation of Y defined by

However, a unitary binding function in (3) transforms the expectation into the linear combination of the predictions called generalized linear models.

The inverse of the logarithmic function is the logistic function defined by

θ is optimized by the cost function defined by

4. The Proposed Method

The proposed method is described and presented concerning the available trend. Figure 2 depicts the proposed financial smoothing system that utilizes the data collected from past samples. A filter is utilized to eliminate corrupt data, which is called preprocessing. Better data processing time and model robustness are reached. In other words, data should be prepared for both simulation and analysis. Semantic hierarchy helps determine salient features utilizing semantic logic. The selection of semantic features enables the model to be more accurate in risk prediction with the lowest error rate. The dataset is split into two categories, called training and test datasets. While the training data is employed to construct a model, the test data is utilized to assess the model.

The balanced sampling scheme splitting data into training and test samples is implemented. The GMDH deep neural network is constructed with salient features employing 70 to 80% of the data. The rest, 20 to 30%, is allocated to forecasting. The efficiency and performance of the proposed method are finally assessed.

4.1. Data Preprocessing

Preprocessing could be conducted utilizing various methods: data cleansing, data collection, data transmission, and data reduction. The proposed strategy is to examine the data and to predict if the rows or columns have missing values that are replaced by calculating the mean. The steps are depicted in Figure 3.

4.2. Data Preparation

Data preparation is related to converting the data into a usable form for cloud computing servers.

4.3. Normalizing Metadata

Normalizing values of each attribute converts values to the range between 0 and 1, then the row of the data matrix is rotated to arrange. All datasets are then transformed into a matrix. Normalization is defined by

Maximum and minimum values are represented by Xmax and Xmin, respectively. Hence, all data is contained in [0, 1]. Existing samples are preprocessed, and the normalization process is separately conducted on each block. Then, the semantic hierarchy utilizes those outputs.

4.4. Deletion of Outlier Samples Utilizing DBSCAN

DBSCAN clustering algorithm groups data; identifies categories; and labels data points as prime, boundary, and anomalous points. The advantage of the DBSCAN over other clustering methods is that it specifies and predicts samples well, which helps deep learning models be refined and improves the processing speed. The DBSCAN, a density-based spatial model, requires two user-defined parameters called the epsilon (eps) neighborhood and the minimum number of points (minPts). Points within the eps distance are called neighbors for a given point. If the number of neighboring points is greater than minPts, this is called the cluster of points.

Boundary points are defined as nonprime points but are neighborhoods of principal points. The DBSCAN differs from the conventional clustering approaches since it could define anomalous points not suitable for clusters. Figure 4 presents the results of clustering utilizing the eps distance threshold as a clustering criterion. The DBSCAN also requires a minimum number of points belonging to a cluster. When minPts is 3, groups 1, 2, 4, and 6 are designated as clusters. In contrast, DBSCAN discriminates groups 3 and 5 as outliers due to not having enough points. When minPts is assigned as 5, groups 3, 5, and 6 are found as outliers.

The inputs are the user-defined datasets and parameter values of eps and minPts.

Input: data collection, eps neighborhood distance, minPts (minimum number of points).

Output: distant spots and clusters determined.

Variables: M, n: the values of row and column of D, distance: distance-vector, indicators: indicating a distance less than eps, class no: displaying clusters by default.

4.5. Applying the Semantic Hierarchy Algorithm

Defining semantic values of attributes and picking salient attributes are performed in two steps. Semantic hierarchy is employed to separate data values by type. Defining different scenarios and generating different outcomes are possible based on semantic hierarchy. For instance, a dataset having property of the shelf life of the system whose values range from 1 to 100 minutes is assumed. The hierarchies changing from 1, 2, 3, up to n could be defined. Transforming the values from 1 to 100 into three words or representative numbers of low (0), medium (1), and long (2) size could be one representation when assuming the level is 3. Therefore, semantic hierarchies could be defined based on the need. The proposed semantic congruence coefficient (SCC) criteria with properties could be defined as follows:0 ≤ SCC (G, G′) ≤ 1If graphs G and G 'are completely uniform, then SCC (G, G′) = 1If graphs G and G′ do not have any semantically identical vertex pairs, then SCC (G, G′) = 0

The larger the homomorphic subgraph between G and G′ is, the greater the number of vertices, the number of edges, or the degree of semantic similarity of vertices will be. Hence, the SCC will be larger.

To compute SCC between graphs, we need to find the simplest congruent graph with the fewest vertices possible, which is denoted by  ⟶ x ⟶ b. Both input and output degrees of the node x1 are utilized. The node x is removed and replaced by a ⟶ b, which is only for offline graphs.

An algorithm similar to the largest shared subgraph algorithm needs to be conducted except for the subgraphs of two graphs overlapping. The output will provide the SCC values between the two graphs. Comparing the SCC values of the two graphs with a threshold value helps determine a sample of whether a financial risk could appear.If SSHC (G, G′) > threshold, then Label = “RISK”ElseLabel = “NONE RISK”

Besides, the amount of uncertainty in the nasal passage could be determined based on the distance between SCC and threshold.

4.6. Separation of Training and Test Samples

The GMDH deep neural network utilizes training and test datasets. Some sampling schemes could be utilized to extract training dataset.(i)Random sampling(ii)Classified sampling(iii)Balanced sampling

Random sampling is one of the simplest schemes that functions randomly and splits samples into training and test datasets. On the other hand, the main disadvantage is that it may not be sampled from a particular batch and ultimately results in reduced accuracy of data classification and validation. Classified sampling is another method that performs based on a probability-based scheme and selects the samples with some drawbacks. Balanced sampling is one of the methods that balance the required samples from the available classes, which finally picks balanced data.

4.7. GMDH Deep Neural Network

The proposed algorithm is based on deep neural network learning. Training dataset is utilized to construct the GMDH model. Then, outputs for each sample are produced with two outcomes, yes or no. Group Method of Data Handling (GMDH), a family of induction algorithms employing computer-based mathematical models dealing with multiparameter datasets, optimizes parameters of the models in areas such as data mining, knowledge discovery, prediction of complex systems, optimization, and pattern recognition. The GMDH neural network performs better than Single Exponential Smoothing, Double Exponential Smoothing, ARIMA, and Open Neural Network. The inductive method, which sorts gradually complex polynomial models and selects the best solution with external criterion, is employed in the GMDH algorithms [19].

Figure 5 presents a set of neurons for the GMDH algorithms. Different pairs of nerve cells are represented in each layer and are connected by quadratic polynomials. A connection of new neurons appears in the next layer, which maps inputs into outputs. The rhythm problem is formally defined as a method of finding a function f that could be estimated by the input-output dataset. Therefore, the single-input data pairs define the following equations.

Equation (9) represents the predicted output. To minimize the squared difference between the predicted and actual outputs, a minimization problem is defined by

A generic equation could be represented by complex polynomials known as Ivanchenko polynomial defined by

However, two-variable representation to predict y is defined by

The coefficients ai in (16) are computed to minimize the difference between the observed and predicted response variable based on each pair (xi, xj) by utilizing the least squares. Then, the polynomial tree is constructed. Therefore, the coefficients of each quadratic function Gi are determined [19].

The polynomial structure of the regression can depend on the samples . Since different samples could exist the least squares approximation could be utiilized to find unknown parameters, it means least squares .

The following matrix relation could be obtained by the quadratic subunit in (16) for each row of M data.where the vector of unknown coefficients of the fourth-degree polynomial is represented by

The vector is the output values of the samples.

Utilizing the least-squares method leads to the solution of the normal equations defined by

The vector determines the best quadratic coefficients. However, a solution obtained from normal equations is almost able to bypass errors.

5. Results

A database with 17232 samples and 869 features was utilized implementing Mat Lab 2015a.

5.1. Evaluation Criteria

The criterion employed to assess the proposed method are as follows:

TP (true positive) and FP (false positive) denote the number of samples projected correctly and mistakenly to financial risk, respectively.

FN (False negative) designates the number of samples that cause a false-positive normal error.

The error and average error rates are computed by (23) and (24), respectively.

5.2. Evaluating the Results of the Proposed Model without Semantic Techniques

Figure 6 presents the results with no semantic technique. The accuracy rate of the risk forecast, accuracy, call rate, error rate, MAE, RMSE, MAPE, and standard deviation (SD) values are 99.822%. 79.43%, 20.8%, 0.177%, 0.177, 0.421, 12.135, and 0.299, respectively.

5.3. Evaluation of Results with a Semantic Technique
5.3.1. Semantic Level 2

Figure 7 depicts the results with semantic technique and semantic level 2. The accuracy rate of the risk forecast, accuracy, call rate, error rate, MAE, RMSE, MAPE, and SD are 99.899%, 84.902%, 83.901%, 0.100%, 0.100, 0.317, 7.065, and 0.179, respectively.

5.3.2. Semantic Level 3

Figure 8 summarizes the results with both semantic technique and level 3. The accuracy rate of the risk forecast, accuracy, call rate, error rate, MAE, RMSE, MAPE, and SD are 99.899%, 87.901%, 87.900%, 0.100%, 0.100%, 0.317, 7.065, and 0.176, respectively.

5.3.3. Semantic Level 5

Figure 9 reports the results with semantic technique and level 5. The accuracy rate of the risk forecast, accuracy, call rate, error rate, MAE, RMSE, MAPE, and SD are 99.911%, 88.912%, 88.912%, 0.088, 0.088, 0.297, 6.214, and 0.171, respectively.

6. Conclusion

This manuscript introduces a novel approach to the prediction of financial risks by constructing a model. SHT and GMDH deep learning neural networks are employed. While the SHT is employed to convert data to a specific range of values and to select salient features, the GMDH deep learning neural network is utilized to construct a model and to predict risk. By simulating the proposed model on financial data and presenting a semantic approach to risk prediction, a 99% accuracy level is reached. Comparing the results of the proposed model with those of other non-data mining methods reveals that the proposed method improved accuracy by almost 20%.

Data Availability

All the data can be obtained from the author upon reasonable request.

Conflicts of Interest

The author declares no conflicts of interest.

Acknowledgments

The research was supported by the Research on Multiple Conjectural Causes and Countermeasures for the Distortion of Investor Protection Policy in China’s Stock Market (No. 17BGL059).