Abstract

Bank customer validation is carried out with the aim of providing a series of services to users of a bank and financial institutions. It is necessary to perform various analytical methods for user’s accounts due to the high volume of banking data. This research works in the field of money laundering detection from real bank data. Banking data analysis is a complex process that involves information gathered from various sources, mainly in terms of personality, such as bills or bank account transactions which have qualitative characteristics such as the testimony of eyewitnesses. Operational or research activities can be greatly improved if supported by proprietary techniques and tools, due to the vast nature of this information. The application of data mining operations with the aim of discovering new knowledge of banking data with an intelligent approach is considered in this research. The approach of this research is to use the spiking neural network (SNN) with a group of sparks to detect money laundering, but due to the weakness in accurately identifying the characteristics of money laundering, the maximum entropy principle (MEP) method is also used. This approach will have a mapping from clustering and feature extraction to classification for accurate detection. Based on the analysis and simulation, it is observed that the proposed approach SNN-MFP has 87% accuracy and is 84.71% more functional than the classical method of using only the SNN. In this analysis, it is observed that in real banking data from Mellat Bank, Iran, in its third and fourth data, with a comprehensive analysis and reaching different outputs, there have been two money laundering cases.

1. Introduction

Significant developments in information and communication technology over the last several decades have brought about changes in a number of domains, including international commerce. As a result, trade, economic, banking, customs, and other operations have transformed. The development of diverse approaches for distinct IT operations for varied activities has also been facilitated by advancements in the realm of computers and intelligent ways. E-commerce is one of the most significant economic successes of information technology, and quick and simple Internet connection has also made a perfect environment for commercial and economic transactions by giving users access to the virtual world [1]. Online payment methods, e-banking, and e-commerce infrastructure are some of the crucial concerns. Any banking service that uses electronic tools and does not need the consumer to physically be at a certain location is considered electronic banking [2].

Banks are impacted by developments as a result of financial liberalisation and globalisation, and they are responding by increasing the range of services they provide to clients. Even though there is more competition for banking services, banks are working to boost their clients’ adoption of online banking [3]. Because of this rivalry between banks, more and more are seeking new locations in which to extend their services, and the number of banks providing Internet banking services is rising daily. More clients can be reached, and a wider range of services may be offered thanks to Internet banking. Customers may get financial services more quickly and efficiently and save time by managing their accounts with the help of online banking. Organisations have quickly adopted and employed information systems due to the evident advantages of information technology, which include improving business accuracy and speed, worldwide quality, cost reduction, and customer happiness [4].

In actuality, e-banking helps businesses save expenses and maintain their competitive edge over conventional banks [3]. Given the advantages of Internet banking and a large number of Iranians using the Internet, statistics show that less than 40% of cardholders at 5% of each bank make online purchases, and only 5% of customers use the bank’s Internet services, indicating that the full potential of cyberspace is not being utilised to provide these services [4]. The researcher’s conclusions indicate that if customers reject or underutilize the new banking technology and services, they will not get any return on their investments. They want to get cutting-edge facilities and technologies [5]. Banks need to come up with new competitive tactics in this environment, and one thing that affects such strategies is consumer behaviour. Because the growth of e-banking relies on consumer approval, it is crucial to investigate customer behaviour and variables influencing e-banking adoption.

Receiving and paying money is one of the most crucial aspects of any bank. This may be carried out online or offline, using the bank’s Internet services or by visiting one of the bank’s branches. To find every potential activity, these transactions must be managed and watched over. However, it is thought to be very difficult for the workforce to monitor them owing to the massive amount of transactional data. As a result, these transactional data are intelligently monitored using machine learning-based techniques. Data mining techniques are one aspect of machine learning that can recognise and extract new knowledge and information from data. The identification of questionable activities in bank accounts and the identification of money laundering, which calls for specialised oversight, are two of the most significant components of transactional data. In order to achieve this goal, this study presents a learning-based approach for the analysis and detection of money laundering. To achieve this goal, data collection pertaining to bank financial transactions is required. The data used in this study come from a Mellat Bank branch in Iran in February 2016. There are 3500 user-performed transactions in this dataset in a given month, including both legitimate and questionable ones. Data have no output and just transaction input values and time. Thus, unsupervised learning is the process of learning new information.

The main approach of this research is to use the principles of data mining which reversibly tries to learn the data in general. Therefore, a new and optimal method will be used in combination. So, the spiking neural network (SNN) is used as part of training and testing data. Although this approach has been suggested before and used for a variety of reasons, it is not a good idea since the spark structure may also extract characteristics. However, via this study, the method has been refined. Therefore, a method called the maximum entropy principle (MEP) is used to improve it. The MEP method can achieve dimension reduction, feature selection, and feature extraction by considering entropy criteria and iteration loops, which increases the execution speed of the Spark method and can increase the improvement in the ROC diagram and the AUC rate. These evaluation criteria have capability to examine financial transaction accuracy in detection in datasets. The type of training method will be unsupervised which will increase the speed of implementation in the training and testing phases. The proposed method is expected to have a functional advantage over the previous methods. So, the main contribution of this article is using the structural method for money laundering analysis and detection with SNN-MEP as unsupervised learning. The rest of this paper is as follows: In Section 2, a literature review is discussed based on recent articles for similar goals. Then, in Section 3, a new method of SNN-MEP is proposed. After that in Section 4, a simulation is performed and results are analyzed correctly. In Section 5, a conclusion as the main discussion of this research is reviewed. In general, the principles of some literature are reviewed as data mining in [6, 7], SNN in [8], and transactions in databases and datasets in [913].

We address the area under the ROC curve (AUC) and the receiver operating characteristic (ROC) curve for clinical domain binary classification issues. A boosting strategy for maximising the AUC serves as the foundation for our statistical approach, which combines many feature variables. The feature variables from several basic classifiers are dynamically integrated into one strong classifier in this iterative process. Through the use of a penalty term for nonsmoothness, we examine regularisation to stop overfitting to the data in the algorithm. We can better comprehend the relationship between each feature variable and the binary outcome variable with the use of this regularisation technique, which also enhances classification performance. We provide evidence of the value of score plots created component wise using the boosting approach. We demonstrate the usefulness of our strategy by describing two simulated studies and an actual data analysis.

2. Literature Review

2.1. Internet Banking Systems

In a study, they examined the factors that determine the acceptance of the Internet banking system by users [14]. There are not many studies in the field of money laundering detection, but a series of other processes have been carried out in it which can lead to further work. Therefore, this section tries to explain a series of similar methods that can help to detect money laundering and any banking disorder. To examine the factors affecting the acceptance of Internet banking, they used the variables of trust, experience of using the Internet, use of other banking services, perceived ease, and usefulness. In their study, they found that despite investments in the use of information technology in the field of banking, some users do not use it despite access to technology. They also found that trust is an important and effective element on consumer behaviour and determines the success of technologies such as e-commerce and transaction analysis.

In a study entitled Internet banking implementation, they found that full capacity of cyberspace is not used to provide these services in Turkey [15]. Trust is not perfect as a person believes that using online banking is safe and that no privacy factor threatens his or her privacy. They found that security and privacy concerns were the biggest barriers to online banking in the country. Trust in Internet banking is also much more important and complex than traditional banking due to its virtual environment. Therefore, to make online purchases, customers must trust the online business and payment through the bank. Without trust, the consumer will refuse any online transaction.

In a study entitled mobile acceptance survey in China, they found that users, especially in developing countries, are accustomed to face-to-face monetary transactions and are more cautious in using online banking [16]. They also found that users use technology when they find it useful, so banks should try to inform their customers about the benefits of using Internet banking over traditional banking. Informing customers that using the benefits of Internet banking services can help increase their productivity, make it easier to communicate with the bank, and improve people’s business performance.

In a study entitled definitive Internet banking in Greece [17], the following results were obtained: Perceived usefulness has a high impact on the attitude towards Internet banking services, the perception of usefulness has a significant effect on the intention to use Internet banking between attitudes and there is a strong significant relationship between the intentions to use Internet banking. There is also a significant relationship between trust in Internet banking and the intention to use Internet banking. In addition, customers’ perceptions of usefulness, credibility, and ease of use play a major role in accepting Internet banking. They eventually concluded that when the attitude of users could be changed, the acceptance of Internet banking would also increase.

In [18] entitled outlier data recognition algorithm using big data processing and Internet of Things (IoT) architecture, the authors examined the presentation of the outlier data recognition procedure using the K-means algorithm and big data processing using the Hadoop database and Mahout. In another article entitled detection and prediction of distance-based outlier data [19], the authors examined the method of distance-based outlier data detection that finds higher outlier data in an unlabeled dataset and provides a subset of it. A set of calls for solving outlier dates and new hidden objects can be used to predict outlier data. Considering only one subset of all two-by-two distances from the dataset, the solution set contains a sufficient number of points to detect high outliers. The properties of the solution set are investigated, and its calculation algorithms over time under the required quadratic equation are presented.

2.2. Data Management Systems

In a two-step clustering process for outlier data detection [20], a new model of a two-step clustering algorithm is proposed for outlier point detection. Using the heuristic technique that if a novel input pattern is sufficiently far from all center clusters, it will be assigned as cluster center, the classic K-means algorithm modification in phase 1 was used for the first time. The results show that the data point in the same cluster may most likely resemble all outliers or nonoutliers, and then, a minimum spanning tree is used in phase 2 by removing the longest edge. At the end, small clusters of trees with fewer nodes are selected as outliers.

In the continuation of this section, a series of methods that have been presented to date are also reviewed. The main concern in lending organisations is credit management and customer needs’ assessment which deals with data breaches, theft detection, and other lending scams. Data breach is a critical issue in the field of global security in lending. In the United Kingdom, 93% of large organisations and 87% of small organisations suffer from data breaches [21]. The average cost in the UK against data breaches reported in [22] is about 1.4 $ million, and the recovery period is normally about 9 months and 3 days. Although various technical solutions for the accreditation and needs assessment of current funds have been offered in various banks in recent years and some are being upgraded, this issue is still considered an important approach [23]. Many approaches in this field have been proposed with the neural network methods which are one of the newest methods proposed in [2429] which have advantages and disadvantages of their kinds, including computational complexity, execution time. The algorithm's lack of generalizability and inadequate accuracy is shown above.

Additionally, using the Portugal-related data set, a neural network-based approach is provided to forecast banks' marketing performance [30]. In another approach presented in [31], the classification of bank customers in order to assess their needs and credit in the need for a loan is considered which is based on a Random Forest algorithm. In [32], the identification of prioritization of criteria affecting the benchmarking of bank customers using the hierarchical analysis process is presented. The results of this research can show the needs of customers in lending. This research is using banking data of Mazandaran province, Iran, but it has been suggested that this topic can be implemented with different data. Another case study conducted for a bank in Tehran province, Iran, is presented in [33] which proposed a method as an application to various datasets. This method classifies the bank customers’ data with a data mining and data clustering approach in order to identify the people in need of a loan and perform accreditation in the bank. In [34], the accreditation and the needs assessment of bank’s customers with a data mining approach have been considered in order to estimate the growth of the fund as well as the number of people in need of loans. The use of logic functions is proposed in this research. The results show high accuracy of the proposed method, but computational complexity is also high. In [35], the authors examined the factors affecting the intention to use the bank’s mobile payment system. These findings show that attitude, mental norm, and behavioural control have a positive effect on the intention to use the mobile payment system.

In [36], the authors examined the effect of mobile banking on electronic customer satisfaction. The sample of 360 customers out of 400 who used mobile banking services in Jordan Ahli Bank, Union Bank, HSBC Bank, and Sarmayeh Bank and the hypothesis were tested through simple regression, and the results represented that the use of mobile banking services to achieve electronic customer satisfaction is effective. The results indicated that there is a statistically significant effect of the overall dimensions of mobile banking services on customer satisfaction. After a simple regression, it was indicated that privacy and accessibility are more influential than other dimensions of mobile banking.

2.3. Electronic Banking Systems

In [37], some influencing factors of acceptance in electronic banking were surveyed with an analysis from the perspective of bank customers. Preliminary data were collected from 387 valid questionnaires that were randomly distributed among customers of 26 licensed Jordanian banks. Multiple regression analysis was used to test the hypotheses. The main findings of this study are avoiding uncertainty that has an important and positive effect on perceived ease of use and perceived usefulness. Perceived risk has a greater impact on customer attitudes, which in turn affects customers’ intentions to use e-banking. Based on the idea expressed in this research, four reference articles can be cited, which in the adjacent areas are the proposed approach, which includes references [3841]. Since the issue of money laundering detection and any unauthorized transactions in the field of anomaly detection is based on a dataset, data mining methods can be extracted and used from other research in other fields. For example, the authors of [38] objectively improved the efficiency of a bank branch. The proposed method is robust and multivariate clustering which has been proposed as a method called data envelopment analysis (DEA). The main structure is clustering and complete data analysis covering each field of data for the previous section. In this research, the structure of data coverage in the forthcoming project is used. In [39], a clustering of big data with the definition of membership functions and fuzzy variables is performed as fuzzy type II, which can be used for modeling and simulation to obtain suspicious as well as uncertain data. This research is not in the field of banking systems, but its method can be used. In [40], a three-cluster big data ensemble approach has been performed that has trained and tested a total of 19 real big datasets in the world and has achieved interesting results in all of them. The algorithm is used for K-means optimization in clustering in an ensemble manner that reduces processing costs and improves data mining operations. The three-cluster structure can be used in this research. Also, in [41], anomalies are detected from the dataset of the wireless sensor network (WSN), which can detect any intrusion and fraud. The use of the sampled window structure is modeled to improve data clustering and identify any category of anomalous operations as an important issue. It is clear that the scope of this research is in network systems, but the clustering structure of this research can be used with the method mentioned in this project, integration and windowing operations to improve the combined model of SNN and MEP.

There are some other studies that have been conducted in the field of bank customer’s validation. Of these, the most recent one in [42] provided a multilevel tutorial on label ratios for classifying bank customers. In this research, an attempt has been made to use a multilevel extreme learning machine (ELM) for learning from label proportions (LLP). With a neural network-like structure, ELM has a higher computational speed and better generalizability. As a result, it is convenient to deal with large-scale, multitier problems. In order to maintain the stability of the stable model when increasing the size of the bag, a small number of label samples in its model, called LLP-ELM, have been performed in a semisupervised learning framework. Experiments prove that there are benefits to improving this research for large class sizes. In practice, it is worthwhile to consider the proposed algorithm for learning multiple classes of label ratios for bank customer classification and other multiclass scenarios. A descriptive study on customer satisfaction with mBoB (Mobile Bhutan Bank) services has examined the level of satisfaction and challenges facing mBoB users in Bhutan [43]. For this purpose, preliminary data were collected using a structural questionnaire consisting of 28 questions. A convenient sampling method was used to select 150 respondents around the site. The collected data were analyzed using descriptive statistics such as simple frequency, percentage, and mean. This study indicated that factors for ease of use, security, and 24-hour access are statistically significant and have positive effects on customer satisfaction with the use of mBoB.

Customer sharing based on Internet banking data can be used which has been studied in [44]. Clustering is an unsupervised data mining technique that can be used to divide the customer. This research creates clustering models on customer profile data using XYZ Internet banking. Clustering methods were performed using the K-means method and the K-medoids method based on the RFM score of the customer’s online banking transactions. In this research, a knowledge discovery methodology has been used. The performance of both methods was measured and compared. The results represented that the K-means method based on the cluster K distance (AWC) performed better than the K-medoids method. According to the Davies–Bouldin index, K-means perform slightly better than K-medoids. Also, in [45], the authors proposed a novel machine learning method for money laundering based on the Python programming language. Detecting money laundering using neural networks in order to stop terrorism is suggested in [46]. Also, in [47, 48], the authors used another method of machine learning for fight with money laundering in banks. In [49], a new method was proposed for detecting suspicious activities in bank data for money laundering detection as a new application and prototype. Also, two different antiterrorism models for money laundering detection were proposed in [50] for the Bitcoin (BTC) cryptocurrency index and [5153] for BankX in the United Kingdom.

3. Proposed Method

In the proposed model, first, the SNN with sparks performs clustering operations and then classification at the data level and then, the MEP method can reduce dimensions and select and extract features by considering entropy criteria and iterative loops based on classification operations were obtained, which increases the execution speed of the SNN and can also help to improve the ROC diagram and AUC rate and other evaluation criteria and can accurately examine financial transactions in datasets. Therefore, the proposed approach has two parts: the first part comprises clustering and classification operations with SNN and the second part comprises dimensional reduction, selection, and feature extraction with MEP.

3.1. Part 1: SNN Model Clustering

To determine the initial framework of the Spark structure, a global model is defined as and the clustering schematic is defined as which is a member of the clusters in the global model. In the most existing clustering results, a cluster is usually represented by a single set, which divides the global model into two regions. From a decision perspective to view the representative of a single set, it means that the objects in the set belong to this cluster and the objects in the set definitely do not belong to this cluster. This is a typical result of bilateral decisions. In this research, this issue is called cluster representation based on bilateral decisions. However, a two-way representation of a cluster cannot indicate that objects may belong to that cluster and cannot intuitively indicate the amount of data infiltration during the cluster formation process. Therefore, it makes more sense to use three regions to represent a cluster than to use two regions. Therefore, a cluster representation based on tripartite decisions is proposed [40]. In contrast to the cluster resolution overview, this study presents a three-dimensional cluster as a pair of sets that are in the following equation [40]:

In this regard, and . It is assumed that there is a structure in the form . Then, , , and are naturally three regions as the core region, the margin region, and the trivial region, respectively. If , the object belongs to the cluster ; if , the object may belong to the cluster ; and if , the object belongs to no cluster of . These subsets have a series of features [40]:

In these equations, if , the representation of in equation (7) will be , which is a single set and . It represents bilateral decisions. In other words, the representation of a set is a special case of the representation of three-dimensional clusters. In addition, according to formula (2), it is sufficient to artificially represent an area by the core area and the margin area. Alternatively, for , a cluster schematic can be provided by the specifications in the following equation [40]:

In the equation, the attribute implies that a cluster cannot be empty. This makes a cluster physically meaningful. The attribute states that each object must definitely belong to a cluster or may belong to a cluster which ensures that each object is properly clustered. According to the family of clusters , there is a family of clusters such as equation (4) formed by tripartite representations [40]:

It is obvious that the family is formed under bilateral clusters which are as follows (5) [40]:

Soft clustering and hard clustering can be formed as described in equation (6) under representations which are available for clustering if [40]:

If equation (6) is established, it is called soft clustering, and otherwise, it is called hard clustering. Until a condition of equation (6) is met, there must be at least one object belonging to more than one cluster. Obviously, a tripartite representation has advantages: displaying a specific set is a special case of representing tripartite clusters. It intuitively shows which objects are the core of the cluster and which are the edges of the cluster. This type diversifies the overlap and reduces the search space when focusing on objects with the overlap/border. Here, an evaluation-based tripartite cluster model is introduced that generates three regions using an evaluation function and a pair of thresholds. This model partially addresses the problem of data mining of a global set of three areas of banking data simultaneously. It is assumed that there is a threshold pair and . Although estimates are restrictive based on a general order, they have a computational advantage. By comparing valuation values with a pair of thresholds, three areas in the bank data can be obtained for initial processing for clustering. Based on the evaluation function , tripartite decisions are made in the following equation [40]:

Based on the evaluator function , a new algorithmic structure can be proposed in Spark. In fact, to formulate the evaluation function, we can refer to similarity measures or remote measures, probability, possibility functions, fuzzy membership functions, Bayesian validation measures, submeasures, and so on. This study, using the evaluation function , attempts to optimize Spark’s orthogonality in a spike neural network with CRF. The spiking neural network presented in this research has high flexibility, in addition to using the linear function to activate cells or neurons or sparks in the hidden layer, uses nonlinear activating functions such as sigmoid or sinusoidal, or uses nonderivatives as well as intermittent activation functions. By default, Spark has an equation, as follows:

According to this equation, represents the weights between the input layer and the hidden layer, represents the weights between the output layer and the input layer, is the threshold value of the neurons in the secretory layer or bias, and is an activator or stimulus function. Input layer weights and bias are randomly assigned. The activation function is assigned at the beginning of the number of input layer neurons or and the number of hidden layer neurons or . According to this information, if the known parameters in the overall equilibrium are combined and rearranged, the output layer will be as expressed in the following equation:

In all models of training-based algorithms, the main goal is to minimise errors as much as possible. The output error function is obtained by the actual output in Spark, which can be divided into the training part and the test part . For both functions, the output obtained by the actual output must be equal to . When this equation is performed and the results are satisfied, an unknown parameter is specified. Although sparks have been used to consider the local dependencies of labels, they are not sufficient for banking data. This is mainly because anatomical structures have complex shapes that are different for modeling. In addition, the temporal or spatial relevance of banking data also plays an important role in clustering that should be considered in the method. Therefore, it is better to correct the probability mapping obtained by Spark. The matrix can be a very unlikely matrix, meaning that the amount of data in the training set phase may not be equal to the total number of data properties. Therefore, the operation of reversing and finding weights or will be a major issue and challenge. To overcome this challenge in Spark, a fully connected CRF matrix is used, which can develop the approximate inverse calculation of nonreversible matrices, which can perform clustering operations with high accuracy and incredible speed compared to the method.

In recent years, CRFs have been widely used in many data mining and processing applications because of their good performance for modeling complex dependencies in spatial data. To cluster banking data, CRFs can be used not only to model the relationship between data features but also to model the dependency between their local data. As mentioned earlier, CRFs have been formulated as neural networks for various data mining and processing operations. However, the process of teaching their method is cumbersome and computationally complex. In contrast, in this paper, CRFs will be used as a suitable processing method. Using a fully connected CRF matrix and layer, is the output matrix and is a fully connected inverse CRF inverted matrix of . Therefore, according to Spark optimization, which is CRF-Spark in this section, the problem of output weights in Spark was solved and converted to . In general, CRF-Spark becomes a chain of repetitive modules over time in the training phase. CRF-Spark will be able to work like a conveyor, that is, add or subtract information to neurons. Unlike deep learning structures and other models, such as the support vector machine or new Bayesian, no weight updating operations are performed during training. CRF-Spark can specify features during data mining. By minimising the CRF energy performance, an appropriate model is taught that can be modeled as

In this regard, are processing data and are parts of the original data or . is a negative logarithmic probability in which is the probability obtained by Spark for each data . While measuring the capabilities of two matrix pairs of CRF in the fully connected layer, it deals with the relationship between each data pair which is defined aswhere is the number of Gaussian kernels and refers to a weight for the Gaussian kernel is the function label which is compatible. The variable represents the appearance of the kernel trying to assign the same class tags to neighboring and adjacent data with the same intensity. Also, the variable represents the smoothness of the core, which aims at removing unnecessary areas. These two steps are represented as the following equations:where and are the application intensities and , and are the corresponding spatial coordinates, and represent the characteristic vectors of each data pair, namely, performance intensity and spatial information, and , , and represent Gaussian kernel parameters, respectively. However, some parts of data clusters may not be processed and mined properly in this way, so an optimization of this algorithm will be performed in layers. In general, the layers of the Spark method are the use of an input layer with a number of neurons (spike). Next, the structure of the training and test layers has convolutional, pooling, and fully connected layers with CRF. Then, a softmax function is embedded for it and an output layer to display the work. The windowing of the training layers is in the form of a matrix, which is considered as 9 × 9 in the convolve layer, 7 × 7 in the pooling layer with two random pooling, and 5 × 5 as the maximum pooling section. The structure of the fully connected layer is CRF as described in the previous section, and its window structure is 9 × 9. The softmax function is also 7 × 7. In fact, the initial Spark training and processing operation takes place in the training layer, including the convolutional layer and the pooling layer which are interconnected, that is, first a convolutional layer, followed by a random pooling layer, another convolutional layer, and a maximum pooling layer. At the end of this training layer, there is a fully connected CRF layer. Then, outside the training layer, there is a softmax function that performs the optimization with MEP. It should be noted that the number of neurons in each segment is decisive. The main framework of the clustering operation with the Spark model is specified in Figure 1.

3.2. Part 2: MEP for SNN Optimization Based on Classification Structure

A general distribution of the Spark structure is created that the entropy of a distribution is expressed as follows [41]:

If, in this equation, the distribution is discrete, there is an equation as follows [41]:

According to these equations, is a probabilistic value and is a maximum entropy-based distribution. The principle of maximum entropy states that the most appropriate distribution for modeling a given dataset is the highest entropy among all data that meet the limits of prior knowledge. These constraints are always given as a relation based on the intended distribution. The mode of entropy maximisation for operations is to reduce the dimensions, select, and extract features to improve clustering operations and change the classification structure in the following equation [41]:

In this regard, is the mapping in terms of distribution, is the expected value for the data test, and is the result of the Spark operation which represents the entropy for the next step in maximisation. is the distance between the clusters created in the Spark operation which originates from the global set and is supposed to interpret the operation into a classification structure and create classes to specify the type of banking operation of the data. is also a real set of entropy to be maximised, and or the global set of clusters based on the Spark model is a member of .

4. Simulation and Results

The data obtained are the result of observation and testing that was obtained from Mellat Bank, Iran, in February 2016. When data enter the program as input, they are normalized. The data need to enter the SNN as a whole. The data are placed in the input layer completely, each column is known as the data attribute, each row is known as a spike (neuron), and the sum of these spikes is called the sparks. The total number of features is 9 items that are placed in the input layer. Data will be entered as input from the input layer to the hidden (middle) layer of the SNN. In these middle layers, training operations are performed. These layers include the convolutional layer with a 3 × 3 windowing filter and then the random pooling layer with a 3 × 3 windowing filter. The total number of features in this middle layer in the training phase is 18. In fact, in the data training phase of the SNN, the initial clustering operation is performed along with training, but it is observed that 18 features are a bit high and the features should be reduced. Therefore, there is a function called softmax in which the MEP algorithm is executed which can perform feature extraction operations (dimension reduction, feature selection, and extracting the best features). In this section, the windowing filter in the softmax function is 9 × 9. The size of windowing filters should be individual, and their size depends on the design of the neural network and is determined experimentally. During the feature extraction operation, which has three main steps (dimension reduction, feature selection, and extracting the best features), the windowing dimensions should be slightly larger than those of the training layer. Now that the MEP has been implemented in the softmax function of the SNN and feature extraction has been performed, it can be seen that there are 3 top features that have been selected. Then, there is an output layer that converts clustering mapping in the training phase along with feature extraction with the MEP into a classification mode. Thus, the two key participants in this study are the data that identify money laundering suspects and the healthy data that do not include money laundering. The general configuration of the SNN with the explanations provided can be seen in Figure 2.

There are nine primary data properties in the input layer, eighteen general data properties in the middle layer during the training phase, and the SoftMax function runs the MEP algorithm. During the training and clustering phases, the properties of the eighteen properties are reduced to three properties, and the same three features are used for classification in the output layer. There are two weights for the middle layer: the first weight is 0 and the second weight interfering with the softmax function is 1, and the bias size is always constant 1. The type of activation function in the training phase based on clustering is defined as sigmoid. In the softmax function with the MEP executable algorithm, both weight and bias are considered equal to 1, and its activation function is considered linear in the feature extraction operation. In fact, in the two middle layers (with convolutional and random pooling layers) and softmax function, training operations are performed and clustering and feature extraction are performed, respectively. In the output layer, the final classification is performed. Levenberg Marquardt was used as an adaptive marker in the neural network training process, along with descending gradients with momentum (traingdx) and mean square error (MSE) was used as the performance assessment criteria over 100 iteration rounds. The efficiency rate of the SNN-MEP based on entropy can be seen in the performance output of this network which is shown in Figure 3.

It is observed that in 1000 iterations, the minimum is close to the target or the best network conditions. The training states can also be seen in Figure 4.

As shown in the top portion of Figure 4, the gradient rate of the network was 0.0016979 across 1000 repetitions, with ascents in each component of the gradient. The two middle layers are for training to recognise based on clustering and SoftMax function based on feature extraction by the MEP method, up to three main features and approximately five features could be present in the training rate, which is shown in the lower figure in 1000 repetition rounds. Additionally, the error evaluation, which is equal to 0 per 1000 repetition rounds, is observed in the middle figure. Based on clustering and feature extraction, the prognosis for the first three days of training data is as follows (Figure (5)). Since the SNN can obtain data during training and other operations such as clustering and feature extraction that are taken into consideration in this study, it was historically regarded for prediction operations, which is why this prediction is shown to provide an improved understanding of the suggested strategy.

Figure 5 indicates the amount of money loaded in 3 days from the main data, which is related to 30 days to 85 days, during a span of three days (which increases the rate of money increase). Prediction in the data can help show embezzlement or any other action intended, such as money laundering. The red graph shows the loading of money into the data box. The blue graph shows the average load of money into the data box. The black graph shows the daily amount of money loaded into the data box, and the turquoise chart shows the minimum amount of money loaded into the data box. The following is a graph of the output of the clustering phase, which can be seen if there is noise in this phase, the output of which is in the form of Figure 6.

Based on Figure 6, it can be seen that there are two main noises in it. These noises can indicate money laundering or a high volume of features. Therefore, feature extraction operations are performed using the MEP method in the softmax function. Now, it can measure the accuracy of the operation, the total output of which is shown in Figure 7.

According to Figure 7, accuracy measurements in percentage (out of 100%) over a period of 80 days of the total data can be observed. There are two graphs in this output: the first blue diagram is the diagram of the proposed approach of this research which shows the SNN-MEP method and the red diagram shows the single use of the SNN. The proposed approach with the blue diagram is somewhat better than the red diagram of the SNN alone which can be seen from a high value to a low detection percentage within 10 to 20 days of observation. It can be said that in this period of time, money laundering has been carried out which has drastically reduced the accuracy of the approach, or it can be clustered and classified due to high noise in the data in the combined operation phase. However, from the 20th to the 60th day, it is observed that the accuracy is increasing, and on the 61st to the 70th day, an increase and a decrease are observed; then, from the 70th to the 80th day, the accuracy is again increased. Based on a general analysis of Figure 6 and its comparison with Figure 7, it can be concluded that there is a problem in the data that have reduced the accuracy in some parts and increased in some parts. The overall accuracy of the proposed hybrid approach on the 80th day is 87% and has an error of 13%, and the method of using only the SNN on the 80th day is about 84.71% accurate which shows the improvement in the proposed approach compared to the classic SNN. In the following, the fitness function is considered for this proposed approach, the output of which is shown in Figure 8.

Based on Figure 8, it can be seen that the output of the red graph is for the lower levels of the classification, the blue graph is for the upper levels of the classification, and the green graph is for the middle level. Based on this, it can be seen that there is a real noise in the total amount of data in the sparks along with their properties which is higher and lower with the blue and red graphs, above the normal level, i.e., the green graph and an error. It is also observed in Figure 7 that the proposed approach has a 13% error. Next, an evaluation is performed from the fitting function according to Figure 8, the output of which is shown in Figure 9.

Based on Figure 9, the fitness function ratio is performed on the evaluation of 50 datasets suspected of money laundering which is the data pattern. It can be seen that the proposed approach with a red graph to a higher proportion than the red graph which is the classical approach has been able to identify a higher level of the original data pattern and receive a better analytical output for evaluation. These 50 datasets show a high volume with noise, and it is possible to detect money laundering by making it smaller, which is performed with 25 datasets, and the output is shown in Figure 10.

Based on this output, in 25 classes of 5 data from the total data, there is a general amount of money laundering and nonmoney laundering. It can be seen that the 4th dataset in the two sections has a high recognition of the proposed approach with the blue diagram, and the 3rd dataset also has this problem in different parts of the data classes. Therefore, money laundering is seen in datasets 4 and 3 of the total data, and the rest of the data are healthy and has no money laundering.

5. Conclusion

Bank customer accreditation is considered a critical process in the statistics of any bank or financial organisation. It is significant because user transactions must be identified in order to assess and differentiate between healthy and unhealthy currencies. This research will provide a perceptive approach consistent with its emphasis on the use of authentic Iranian bank data to identify money laundering. Annual money laundering efforts pose a danger to global economic stability. These businesses may profit from criminal activities, endangering the stability of international financial systems. Due to these reasons, money laundering is seen as a severe problem in many countries. Both academics and analysts have a specific interest in the use of software technology to enhance the detection of money laundering activities. While ensuring that they provide more sincere and efficient customer service, banks and other financial institutions are striving to ensure that unlawful conduct is discovered and stopped. In the current economic climate, bank financial authorities are making an effort to steer clear of the latest frauds. Effective technology solutions are necessary in the fight against money laundering. Better data and analysis are necessary to support researchers in focusing on dubious behaviour. Legislation that is always changing and recent examples of large financial institutions breaking money laundering laws demonstrate the need for a better technology base in order to handle antimoney laundering activities efficiently. The method this research suggests for identifying money laundering is using an intelligent technology. Two different clustering and classification mappings for the spiking neural network (SNN) are used in this approach. This strategy uses the maximum entropy principle (MEP) technique since the SNN is unable to minimise the quantity of data during feature extraction and optimum feature selection. A training layer for feature extraction operations based on the MEP technique and a training layer for clustering utilising convolutional and random/maximum pooling layers are often included in the softmax function in an SNN. Clustering and feature extraction mappings are identified and listed below as methods for data classification and money laundering detection. The present study’s overall approach improves the diagnosis in two datasets 3 and 4 of all the data in a time frame of 30 to 80 days, with an accuracy as high as 87% compared to the previous technique, which is 84.71% accurate and employs just the SNN.

Abbreviations

ROC:Receiver operating characteristic
AUC:Area under the ROC curve
CRF:Conditional random fields.

Data Availability

The data used to support the findings of this study are included within the article.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this article.

Authors’ Contributions

Xiaorong Qiu, Ye Xu, Yingzhong Shi, and S. Kannadhasan Deepa were responsible for conceptualization, methodology, writing, editing, and results. S. Balakumar was responsible for methodology, writing, editing, and results.