BioMed Research International

BioMed Research International / 2021 / Article
Special Issue

Artificial Intelligence for Medical Image Analysis

View this Special Issue

Research Article | Open Access

Volume 2021 |Article ID 6653879 | https://doi.org/10.1155/2021/6653879

Mohsen Ahmadi, Abbas Sharifi, Shayan Hassantabar, Saman Enayati, "QAIS-DSNN: Tumor Area Segmentation of MRI Image with Optimized Quantum Matched-Filter Technique and Deep Spiking Neural Network", BioMed Research International, vol. 2021, Article ID 6653879, 16 pages, 2021. https://doi.org/10.1155/2021/6653879

QAIS-DSNN: Tumor Area Segmentation of MRI Image with Optimized Quantum Matched-Filter Technique and Deep Spiking Neural Network

Academic Editor: Huiyu Zhou
Received27 Nov 2020
Revised29 Dec 2020
Accepted06 Jan 2021
Published21 Jan 2021

Abstract

Tumor segmentation in brain MRI images is a noted process that can make the tumor easier to diagnose and lead to effective radiotherapy planning. Providing and building intelligent medical systems can be considered as an aid for physicians. In many cases, the presented methods’ reliability is at a high level, and such systems are used directly. In recent decades, several methods of segmentation of various images, such as MRI, CT, and PET, have been proposed for brain tumors. Advanced brain tumor segmentation has been a challenging issue in the scientific community. The reason for this is the existence of various tumor dimensions with disproportionate boundaries in medical imaging. This research provides an optimized MRI segmentation method to diagnose tumors. It first offers a preprocessing approach to reduce noise with a new method called Quantum Matched-Filter Technique (QMFT). Then, the deep spiking neural network (DSNN) is implemented for segmentation using the conditional random field structure. However, a new algorithm called the Quantum Artificial Immune System (QAIS) is used in its SoftMax layer due to its slowness and nonsegmentation and the identification of suitable features for selection and extraction. The proposed approach, called QAIS-DSNN, has a high ability to segment and distinguish brain tumors from MRI images. The simulation results using the BraTS2018 dataset show that the accuracy of the proposed approach is 98.21%, average error-squared rate is 0.006, signal-to-noise ratio is 97.79 dB, and lesion structure criteria including the tumor nucleus are 80.15%. The improved tumor is 74.50%, and the entire tumor is 91.92%, which shows a functional advantage over similar previous methods. Also, the execution time of this method is 2.58 seconds.

1. Introduction

Brain tumors, which are well known to be one of the most common diseases of the nervous system, can cause many damages to human health and can also result in death. In this matter, the most common type of brain tumor among adults is glioma [1]. These tumors can be classified based on their grades as follows: Low-Grade Gliomas (LGG) exhibit benign trends and provide better patient awareness, whereas High-Grade Gliomas (HGG) are malignant, which may lead to receiving worse patient awareness [2]. The medical image of brain tumors helps assess disease development before and after treatment. Several imaging techniques, such as MRI, CT, PET, and SPECT imaging, have been used to examine brain tumors. However, MRI imaging is now the main imaging technique that can be used for glioma’s diagnosis and treatment, because it has advantages such as good soft-tissue disparity, multiplied parameters, shooting in the desired direction, noninvasive photography, and so on. It also has various sequences, such as T1 weight images, T1 or T1ce-enhanced contrast, T2 weight, and Fluid Attenuation Inversion Retrieval (FLAIR). These sequences offer additional details about different parts of brain tumors [3]. For instance, the tumor area via peritumoral edema may be diagnosed in FLAIR and T2 images. Conversely, the tumor nucleus area without peritumoral edema is more prominent in images of T1 and T1ce. In this way, the different main MRI methods focus on detailed information of images, which describe the features of brain tumors under several sides.

For a medical diagnosis, accurate segmentation of these tumors is critical and needs therapeutic planning. Segmentation of brain tumors in an automatic way and existing infrastructures from medical imaging allow for accurate diagnosis of tumors. It can help plan surgery and the treatment of brain tumors by providing a more efficient and better diagnosis [4]. In particular, it is critical to divide these tumor tissues, such as enhancing core, necrosis, edema, and nonenhancing core in terms of the natural brain tissue, containing white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF). Nevertheless, the precise automatic segmentation of these tumors is a challenging issue due to several reasons. In image segmentation operations, the outlines between the normal tissues and brain tumor are blurred because of the partial size effects, the gradient filtering intensity, and the magnetic field artifacts. Moreover, brain tumors are very varied in terms of size, shape, and location in patients. It is recommended to utilize a novel, robust, and fast method with the utmost care in the field of image segmentation. The segmentation of different images is a separate issue, and the right method should be designed according to each structure that should be segmented with a specific purpose. Deep convolutional neural networks have done very well in recent years in brain tumor segmentation [5]. In this regard, the convolutional neural network (so-called CNN) is a popular deep learning model that can elicit some favorite features for the original data classification [6].

This article proposes a new optimal framework of the brain tumor segmentation of MRI images that uses the structure of an optimized deep spiking network with the Quantum Artificial Immune System (QAIS). This framework is fully integrated with the QAIS-DSNN and conditional random field (CRF) combination. In the first step, a multiplied level architecture network is proposed to consider interdependence segmentation among neighboring pixels and supplementary information in various layers and measures. The background textual information of the three-dimensional MRI images is essential for brain tumor segmentation that is not taken into account by the CNNs. The study also introduces connected CRFs to correct the mapping probability attained by QAIS-DSNN.

2. Literature Review

To date, several methods have been proposed for MRI imaging. This section examines an overview of several classified methods.

2.1. Research on Deep Learning-Based Methods

The importance of MRI imaging methods for brain tumors in recent years with deep learning principles and methods due to high applications and relevant results has been highly regarded. In [7], the design of different types of convolutional neural network architecture is proposed in the form of windowing with a deep layer in different grades of gliomas specimens using small nuclei. A two-way convolutional neural network model has been proposed in [8], and one channel provides detailed features of local and the other provides universal feature extraction. In [9], a convolutional neural network architecture has been created as a cascaded CNN to obtain the local dependencies of tags, achieving better performance in segmentation. Besides, they selected a two-step training strategy to address label imbalance distribution. Recently, there are advantages of multiscale features of the convolutional neural network in segmentation work [1016]. In general, there are two methods to elicit features of multiplied scales: the first method is to use feature mapping of different network levels to show multiscale features [10].

In this respect, a multiscale convolutional neural network has been suggested to divide the retinal vein in [17]. Scale images are identified at different stages of the convolutional neural network to obtain the retinal arteries’ probability mappings. Also, in [18], the structure of the Fully Convolutional Neural Network (FCNN) was developed for training with CRF; however, the process of training them was extremely time-consuming and expensive in terms of memory consumption. The second case is the transfer of versions on a different scale from the input image using the same network [10]. Also, multiscale features have been obtained by the convolutional neural network in [18]. This paper adopted three-dimensional CRFs to process segmentation results, but configuring three-dimensional CRFs is a complex process. In [18], different sizes from a convolutional neural network architecture have been used as cascaded CNN to record multiscale features. Due to this research and, of course, many other types of research that are beyond the scope of this research, the convolutional neural network has achieved significant achievements. The ability to learn neural networks with architecture and fixed parameters is limited, and the useful information, for three-dimensional MRI data, may be overlooked.

Some researchers use two-dimensional [19] or three-dimensional convolutional neural network models [18, 20, 21] to deal with three-dimensional images. For brain tumor segmentation, a three-dimensional semantic segmentation network based on the encoder-decoder architecture was developed in this way [22]. A hierarchical segmentation system that has varied the segmentation into three binary tasks has been proposed [19, 23, 24]. They also taught models of segmentation from sagittal, coronal, and axial perspectives. In the practical step, to achieve the final results, they averaged the SoftMax outputs obtained in the mentioned perspectives. Even though these methods do work very well, they raise both memory consumption and fiscal complexity. Thus, fiscal models, such as conditional random fields (CRFs) and Markov Random Fields (MRFs), are mainly employed to investigate spatial text information. In [25], a neonatal structure of a deep neural network called the Growing Deep Convolutional Neural Network (GCNN) is presented to segment MRI images to diagnose brain tumors.

There is also another method combined with GCNN that is a Stationary Wavelet Transform (SWT). The hybrid deep learning method is simulated with the use of BraTS2018 dataset and evaluated using the peak signal-to-noise ratio (PSNR), the average square error, and so on. In [26], a complete convolutional neural network with pyramidal features is presented as an Atrous convolution for brain tumor segmentation by MRI images. This research uses data sets from BraTS2013, BraTS2015, and BraTS2018, the results of which have a functional advantage over lesion structure, including tumor nucleus, improved tumor, and the whole tumor, compared to previous methods, especially the convolutional neural network. These results are based on the Dice criterion, 76.88% for the tumor nucleus, 74.43% for the optimized nucleus, and 86.58% for the entire tumor. Also, in [27], the convolutional neural network is used in three dimensions based on a method called Test-Time Augmentation. This research uses BraTS2018 data and shows the results of its evaluation with a lesion structure, including tumor nucleus, improved tumor, and whole tumor, with functional superiority over many convolutional methods and deep networks. These results were in two ways using the Dice criterion, which was 90.21% for the tumor nucleus, 79.72% for the optimized nucleus, and 85.83% for the entire tumor. In a similar study, in [28], the convolutional neural network is considered to be multicascaded (CNN) and conditional random field proposed as MCCNN. The results of this study, based on the lesion structure criteria, were 71.78% for the improved nucleus, 88.24% for the total tumors, and 74.81% for the tumor nucleus. For breast imaging monitoring and data system ranking, Kang et al. [29] indicated a dominant fuzzy full-connected layer. The aim of the model was to establish complementary scoring properties for semantic segmentation with fuzzy rules.

2.2. A Review of Deep Spiking Neural Network

First of all, it should be noted that sparks are the neurons of the neural network that use spikes instead of neurons in the spiking neural network, and a set of neurons in an input layer with spikes is called a spark. spiking neural networks (SNNs) are driven by the processing of biological knowledge, which communicates in parallel scattered and nonsynchronous binary signals. In neuromorphic hardware, SNNs indicate some appropriate features such as fast inference, low energy use, and event-dependent processing of information. It creates interesting applicants to apply deep learning (DL) networks effectively and a selection process for several learning tasks at a computer. Here, SNNs consider a wide range of training methods, including the conversion of convolutional deep networks to SNNs, limited preconversion training, and a variety of biological motivations [30].

Neural networks are usually read if they have at least two hidden layers of nonlinear input conversion. In this study, only feedback networks are considered to calculate mapping from input to output. Spiking neural networks were initially studied as biological information processing models in which neurons exchange information through spikes. Here, all spikes are expected to be stereotypical events; in this way, data processing is minimized to two main factors: First of all, the timing of spikes, for example, firing frequency, the relative timing of pre-/postsynapse spikes, and special patterns of movement. Secondly, the identification of the synapses used means it is possible to connect nerve cells, whether the synapse is stimulating or inhibitory. With regard to the degree of detail of the simulation neurons, the two neurons are the point at which the input spikes alter their (somatic) membrane potential immediately or are built together with complex (dendritic) spatial structures as multichamber models. Hence, the dendritic currents will communicate before that. There were also changes to physical capacity. Here, several models of spike neurons, such as Hodgkin’s Huxley model, integrate-and-fire, and spike response, explain the evolution of membrane potential and the spike of different rates of detail in development. Essentially, the membrane potential of the stream merges with the entry of the spikes and generates a new spike since the threshold is crossed. After the spike is obtained, the small axon is sent to all the linked nerve cells by a delay via the axon, based on which the membrane potential is adjusted to a certain base. Figure 1 shows this.

Direct communication between spiking and analog neural networks is formed by assuming a stable state, by considering the activation of an analog neuron is equal to the firing rate of a spiking neuron. Many geometric models used those rate codes to describe brain computational processes. Nevertheless, more complex processes can also form the neural spike models, which depend on some reference signals or relative timing between spikes, such as network fluctuations. Temporary codes are very important in biology; even a spike or small time-consuming changes in neuron firing may cause different reactions, as most decisions must be calculated before a reliable estimation of the spike [30].

In addition to the biological definition of SNNs, they contain a pragmatic functional representation in the field of neural engineering; SNNs are commonly referred to as spikes and are event-based. An event here is a collection of digital information defined by a time marker’s origin and destination address. Unlike biologically motivated SNNs, it may have several bits of load information. The source for this protocol is the address index or AER (Address Event Representation) protocol, which is used after processing to link to event-based sensors through digital connection to neural chips or digital hardware. Event-based visual sensors use the loading bit to differentiate between silent and visual events; however, the loading bit can also be used to send other types of information to postsynapse targets, potentially to calculate more advanced functions than the fire integration method or integrate and fire used. The reason for researching SNNs is that, in real-world activities, the brains display considerable cognitive function. With continuing efforts to enhance our perception of brain-like calculations, models closer to biology are closer to achieving human intelligence than more abstract, or at least more computationally effective, models [30].

In this way, SNN methods are ideally appropriate to process the space-time information based on neuro sensors, which are themselves energy-efficient. Sensors collect precise environmental information, and SNNs can use some useful time codes for their calculations. This information processing is also the focus of the event, which is denoted whenever a small amount of information is not recorded in the SNN; it does not do much calculation, but the SSN creates more spikes when an activity explosion is recorded. It leads to a very efficient way of calculation, assuming that information from the outside world is usually scattered. Also, time-domain input is another precious piece of information compared to framework-based approaches, where an artificial timeline is introduced entered by the sensor. It can result in an effective calculation features such as optical current or stereo inequality combined with spike-sensitive learning rules. In deep SNNs, asynchronous axis-based computing mode results in the rapid dissemination of prominent information through multiple network layers. In practical terms, SNNs must be run on neuromorphic hardware to take advantage of this effect. This process is a quasisimultaneous data processing combined with an event-based sensor, which implies that after the first input spikes are registered, the first estimated output of the final layer is immediately available. Also, for multilayered networks, it is right as the spikes extend immediately to the higher layers as soon as enough activity is generated by the bottom layer. You do not have to wait to complete of the complete input series, which is unlike traditional deep neural networks, where it is important to completely charge all layers until the final output is calculable. The primary performance spikes are inevitably based on incomplete data. It was thus concluded that deep SNNs would increase their efficiency in classification and decrease the processing time of the spike more than their input. To decrease the expected delays in inference, SNNs can also be specifically fitted. SNNs are the computational model chosen to run highly energy-efficient neuromorphic hardware devices supporting a data-driven processing mode and maintaining local calculations, thus prohibiting access to expensive memory [30].

Here, despite recent advances, one of the major deep SNN disadvantages is their accuracy in standard metrics such as MNIST, CIFAR, or ImageNet which is not as good as that of their machine learning counterparts. The existence of the benchmarks present in traditional frame-based images can perhaps be attributed to this, to some extent. A sort of conversion of the picture to the Spark sequence that is typically inefficient is therefore required. The lack of training algorithms that take advantage of Spark neurons’ features, such as efficient timescales, is another limiting factor. In contrast, several approaches employ many approximations according to the rate of use of convolutional deep learning neural networks, denoting that no progress can be expected. Deep SNNs may be practical in these cases and maybe faster, in which they get more efficient than convolutional systems, where SNN runs on diagonal neural hardware. For SNNs, the training algorithms are difficult to analyze due to their noncomputational and discontinuous computational methods, which generated direct use of successful techniques behind the scenes, especially for deep neural networks be difficult [30]. In traditional AI standards, the performance of SNNs should only be considered as concept proof, but not as the ultimate research goal. If biology is the model of spike networks, it can be concluded that they are designed for behavioral tasks such as making decisions based on continuous current input when moving in the real world. Whereas brains may solve these things, they are certainly not optimal for it. Recently, the Internet environment lacks good metrics and evaluation metrics that can measure effective performance in the real world [30].

3. Proposed Approach

The preprocessing phase of the proposed approach is aimed at reducing the initial noise. In the following, the operation of segmentation and extraction of features is aimed at distinguishing tumor masses from the data set. The preprocessing section applies a method called Quantum Matched-Filter Technique, followed by a CRF-based QAIS-DSNN combination approach.

3.1. Preprocessing Phase

Initially, there will be a preprocessing phase involving noise reduction. Every single image is displayed in a combination of local threshold and active contouring using a two-dimensional array of pixels; their values are integers in the range of [0,255]. Local thresholds initialize images in two steps. First, the input noise image is considered the primary image to which image noise removal will be applied. This operation is mainly utilized as a local search operator to enhance the initial images, using the Quantum Matched-Filter Technique (QMFT). The use of local thresholds and active contours has been used in this paper because they are computationally faster than other methods in the literature. Thus, at the end of the first step, there will be a decomposed image. In the second step, thresholding is done on the detail coefficients, and one of these decomposed sections is randomly selected and sent to a reconstruction operation. The reconstruction section can be defined: (i)Gaussian blur: uses a Gaussian filter to filter the image. Between pixels and pixels, the filter size is accidentally selected(ii)Mean filter (averaging filter): filters the image using an average filter(iii)Intensity change: all image pixels are multiplied by a similar criterion randomly selected in the range [0.7, 1.3](iv)Implement light-intensive sections in quantum and reverse processing that performs the QMFT

Then, the following operations are performed: (i)One-point row: a pixel row is chosen randomly(ii)One-point column: it is identical to the preceding form, but instead of a row, the column is considered(iii)Point-to-point random: accidentally, every pixel is selected from the decomposition until a new image is created(iv)Identify all points in a row and column in the image to reduce the majority of noise as QMFT

After analysis, when the selected range value [0.1] is less than the local search rate in the QMFT, a new image of the local search operator may pass. As the decomposition is complete, the entire image is sorted by its pixel value. Then the best aspect ratio in the image is considered as a quantum value in the sequel. A signal in MRI images may be broken down into multiple displaced or resized displays of features known at the feature extraction stage. Local thresholds and active contours can be used to analyze an image into its components. It is possible to perform image segmentation operations after applying QMFT along with local and active contouring thresholds. In this case, the local threshold coefficients and the active contour based on QMFT can be destroyed to eliminate some details. Local thresholds and QMFT-based active contours have a tremendous advantage in separating fine detail in an image. Active contour can be used to isolate very fine details of an image. At the same time, local thresholds can detect large details, combining fine and large details, and reading all rows and columns linearly and diagonally. Quantum satisfies QMFT to minimize the noise in the MRI image. QMFT based on local thresholds and active contours can create a sparse display. A local and active contouring threshold function with QMFT has two main features, the first of which is a function of oscillation or wave appearance, such as

In this case, most of the energy in Ψ (t) is limited to a limited time, which is in the form of

The proposed method is generally calculated to reduce the noise in

In Equation (3), the term ensures a certain degree of validity and accuracy between the rated and original image, in which denotes the rated image while means the noisy image. The parameter is defined as the sum of the variable adjustment periods, β and λ are the balancing parameters, and is the sum of the image’s points. The purpose of minimizing Equation (3) is to decrease total image diversity while maintaining accuracy and validity. The balancing values are changed from 1 to the size of the image for both β and λ to minimize Equation (3).

3.2. Segmentation with QAIS-DSNN Combined Approach

The deep spiking neural network presented in this study, due to its high flexibility, can use a linear and nonlinear functions such as sigmoid or sinusoidal in hidden layers. Use nonderivative as well as intermittent activation. By default, DSNN has

According to Equation (4), displays the weights between the input and the hidden layers, and displays the weights between the output and the input layers (). The value of the neuron threshold is in the hidden layer or the bias. is an activator or stimulus function. The weights of the input layer, , and bias, , are randomly assigned. The beginning of the neuron number on the input layer and the neuron number on the hidden layer is assigned to activation function . If the known parameters in the general equilibrium are combined and controlled on the basis of this information, the output layer will be similar to

The main goal is to minimize errors as much as possible in all models of training-based algorithms. The the output error function is obtained by the actual output in DSNN, which can be done with two training sections, and the test section, . The output generated by the real output,, must be identical with the same for both functions. An unknown parameter is specified when this equation is performed and the results are satisfying. While spikes have been used to understand local label dependencies, for medical images such as MRI, they are not appropriate. Typically, that is because anatomical forms have complex shapes for models that are distinct.

Moreover, either the temporal or the spatial relationship of MRI data also plays a critical role in classification, which should be paid attention to with regard to the method. Therefore, it is better to modify the mapping of the probability achieved by DSNN. A rather low-probability matrix may be the -matrix, which means that the amount of data in the training process will not be identical to the total number of data characteristics. But it would be a big challenge to reverse and find weights or β. A fully connected CRF matrix is used to overcome this challenge in DSNN, which can develop an approximate reversal of the matrix that cannot be reversed. It can reduce the size, selection, and extraction of features at the segmentation with high precision and incredible speed compared to other methods. Currently, CRFs have been implemented in many medical imaging applications because they perform well when modeling some complex spatial data dependencies. In this way, to segment brain tumors, CRFs can be used not only to model the relationship between an image pixel and poster properties but also to make local pixel properties and their labels dependent. As discussed earlier, in [11] and [26], CRFs were employed to visualize images through image formulation as neural networks. Nevertheless, the process of training their method is cumbersome and mathematically complex. In contrast, CRFs will be utilized as a suitable hash method. Using the fully connected matrix and CRF layer, the output matrix and the matrix are all inverted and generalized by . Therefore, due to the improvement of DSNN as CRF-DSNN in this section, the problem of the output weights in DSNN has been resolved and converted to . In general, CRF-DSNN becomes a series of repeating units over time in the training phase. CRF-DSNN will be able to act as a belt conveyor and add or subtract information to neurons. Unlike deep learning structures and other classification models, such as backup conveyor machines or nanoscale works, no weight update is performed during training. CRF-DSNN can define features at the segmentation. By reducing CRF energy performance, a suitable model is taught that can be modeled as where are the designations of the segmentation and properties are specific pixels of the original image or . is the negative logarithmic probability where is a probability obtained by DSNN per pixel . While measuring the capabilities of a matrix pair of CRFs in a fully connected layer, it deals with the relationship between each pixel that is defined as where , the number of Gaussian nuclei and indicate a weight for the Gaussian nucleus mth, and is the label of consistent function. displays the core appearance, which tries to assign the same class labels to neighboring and adjacent pixels with the same intensity. displays the kernel smoothness, which is associated with eliminating unnecessary areas. These two steps are shown as

and are the light intensities of the pixel and and and are the corresponding spatial coordinates. and mean the characteristics of each pixel pair, i.e., the brightness intensity and spatial information. , , and show the parameters of the Gaussian nucleus, respectively. However, some points in the mass may not be segmented in this way, so this algorithm optimization will be done in layers. In general, the DSNN method’s layers are the use of the input layer with the number of neurons (spikes). Then, the structure of the training and testing layer used convolution, pooling, and fully connected layers along with CRF. Then, a SoftMax layer is embedded for it and then an output layer to display the work. The training layer window is in the form of a matrix, in the convolution layer, in the pooling layer, and in the maximum section (Maxpool). The structure of the fully connected layer is . The SoftMax layer is also .

The Quantum Artificial Immune System (QAIS) is used to optimize the segmentation process during high-altitude neural network training in the SoftMax layer section. The QAIS uses a factor called an antigen. In an MRI image, all antigens are detected through a memory-based adult detection system, which has a fault tolerance experiment with a choice of the colon and immune mutations. Colonial choices and immune mutations are the other two factors of the QAIS algorithm. The more MRI data, the more copies are duplicated. In this algorithm, reproduction is plural, especially like a crossover in the genetic algorithm [12, 13]. Antibodies focus on modern quantum memory detection systems in mass segmentation in real time and examine detection and cross-sectional states against the MRI image structure.

The display of MRI image data is performed by a set of antigens , in which the antigens determine the ad. They display one bit of binary string bits’ properties that are represented by MRI image data antigens. These bits contain Trait codes. Also, is the spatial state in the QAIS that is presented by and displays all the activities of the primary population in the image in segmentation. is the natural state number in the QAIS algorithm, which is considered as a constant value. There are two states of self-adjusting and non-self-adjusting in the artificial immune system algorithm. The self-adjusting state () displays all MRI image data and the non-self-adjusting state () displays all the segmented data. Therefore, there is a relationship between self-adjusting and non-self-adjusting states, represented by and equations.

The safety diagnostic set is also , where is the antibody, is the concentration of the antibody, is the tolerance of error, indicates the age of memory and the maturity of genes, is a set of real numbers, and is the case number natural genes. Memory detection set and the gene recognition maturity group are shown as . There is also immaturity, which is defined as the immaturity of genes that are expressed as . In these relationships, , where α represents the threshold for error in detecting immature status, λ represents the gene life cycle, and β represents the threshold value for detecting gene maturity.

In order to establish and evaluate the structure of diagnostic development, an immature gene detector becomes the mature state detector, which will be successful in the fault tolerance phase. When the adaptive time between the adult gene detector in the gene’s life cycle and the antigens activated exceeds the β threshold, the adult detector clones or collects itself and then evolves into a memory detector. It means that genes and antigens will have a memory. Once the antigens are recognized by a specialist, he/she assembles the mature diagnostic compound. To ensure that antigens are effectively detected and that a variety of antibodies are detected in the reagent (mature or immature), they will detect known or unknown attacks. A total of three operators are used for the QAIS algorithm to improve the transverse distribution of MRI image data, which includes dependency assessment, reproduction selection, or safety and mutation combination, which are described separately.

Hamming distance is used to compute the correlation for antigen detection. For example, the error tolerance mode is considered to create a model of correlation assessment. An unsuccessful identifier can succeed, if the immature identifier has never been compared to all elements of the self-organizing group in the α variable. On the contrary, it can lead to the death of genes and antigens. The is assumed, and Equation (10) shows how is determined by the .

According to Equation (10), indicate whether the is compatible with and is the size of the detector, so is used to calculate the correlation between s and . Likewise, as represents the correlation threshold. Equation (11) is used to implement the mature error detector of the immature , and Equation (12) is used to add the self-enforced identifier time when the results of the Equation (11) return to 1, and if , the immature identity must develop into a mature

The colonial or combination choice operator performs cellular operators in mature and memory diagnosis. Equation (13) is used to detect cloning state and a mixture of genes and antigens.

According to Equation (13), is a colonial or combination constant. shows all the combinations. The colonial determinant or combination factor is used to analyze the performance of cellular operators in mature and memory diagnosis. Equation (13) is used to detect the cloning state and a mixture of genes and antigens. In Equation (14), and display the colonic selection group or group of memory and mature detectors. After making a colonial selection or group of genes and antigens in a generation, the cloned or combined section is added to the adult diagnostic group, and the same detector , in the colonial selection group, or the and combination will be removed.

The goal of the immune mutation operator is to enhance the detector’s diversity with the mutation of the antibody generation in the corresponding detector, which is used to improve the ability to detect antigens. Considering the bit, the detector set is matched by the antigen; these bits are used by 0.1 instead of randomly. displays the size of . The mutant detector is used as an immature detector by the self-regulating set. To detect the adult mode, if the adaptive time is greater than the activated threshold β, the stimulus operation is performed using Equation (13) and then combined with the memory detector according to

Equation (15) is assumed to represent the arranged numbers of the reagent that can be matched with antigens. Therefore, the memory diagnosis segment is combined with Equation (16), but this occurs when the memory diagnosis segment can be successfully matched with antigens.

Equation (17) also illustrates a different type of antigen removed in the MRI image data for display.

Intensity and variety are two important features of swarm intelligence algorithms. The intensity is in the search of the best-obtained solutions and choosing the best candidate points. It is worthwhile to mention that the diversification procedure can allow the optimizer to explore the search space more efficiently. Inertial weight parameters indicate changes in optimal global attractiveness that affect the convergence rate and update each mass’s position in the combination algorithm QAIS-DSNN. In the proposed QAIS-DSNN hybrid algorithm, the inertial weights are set to a large value to emphasize exploration, i.e., 0.9, which are set in the initial search mode, finally reduced to 0.1 linearly for the importance of linear optimization. Inspired by the classic artificial immune system, it is guaranteed that quantitatively, global characteristics for optimal segmentation can be determined when using the spiking neural network. As the number of repetitions increases, the initial population is encouraged to local search. Finally, the population should only carefully search for a local area without discovery to find out if there are any other masses. As a result, the first quantum combination strategy is to provide a linear weight reduction of the new frequency. The model is created as where is the inertial weight range and is the maximum number of repetitions, in which denotes the number of repetitions. As such, bothand are linearly decreased from 0.9 to 0.1 in the repeat cycle. The developed mixed QAIS-DSNN algorithm may be trapped in local improvement due to the presence of different iterative cycles in the tumor fractionation improvement process, in addition to the high research capacity. Therefore, to solve this problem, it is possible to provide a comparative update strategy for the parameter that is best to assist neurons and primary residents of the proposed algorithm out of the optimal local areas. In this strategy, is considered the best at a great value in the initial phase of finding the optimum value of the QAIS-DSNN algorithm with strong exploration ability (global search) and gradually decreasing with increasing frequency for accurate searching.

Equation (19) displays the optimal value for the better adaptive update scheme used here. where means the number of repetitions, whereas denotes the maximum number of repetitions. The next step is to introduce a novel method for updating the neurons and the initial population to accelerate global convergence. Initially, the status vector is updated by where is the total number of variables, is the upper limit, and is the lower limit of the variable in the variable. is the search environment, the same as the main input image. Randomization is then performed that prevents trapping in the optimal local solution, so randomization is introduced in Equation (21) with the value of , which is a randomization parameter. where is the position of each neuron, and the initial population of the combination approach and is a random number generated, represented as a uniform distribution in the range of . In general, the flowchart of the proposed approach is shown in Figure 2.

3.3. Investigating the Computational Complexity of the QAIS-DSNN Method

Here, the computational complexity of the developed QAIS-DSNN algorithm is investigated. Computational complexity includes temporal and spatial complexity. The time complexity of the QAIS-DSNN algorithm depends on two steps including calculation of the motion and updating of the positions of the neurons and the initial population. Therefore, the complexity of time can be defined in where is the maximum repetition cycle, denotes the number of neurons or initial population, and is the dimensions of the problem.

4. Simulation and Results

BraTS data is a collection of brain tumor MRI images, including 145 folders for patients under different conditions. The dataset consists of 4 versions from 2012 to 2018. Database versions are getting better every year. The primary data is in DICOM format which have been converted to JPEG format for easier use through DICOM Viewer software. The input images are three-dimensional. Due to the large size of the images in the BraTS, we used 1000 video input samples to study the proposed approach. The simulation will be done in MATLAB 2015b environment and a system with 7-core processor specifications with 6 MB of cache and 3.6 MHz and 6 GB of memory in Windows 10. When the simulation is performed, all BraTS2018 data are trained and tested by the proposed method. For visualization, an example of images is shown to examine the proposed approach’s results, step by step. Initially, the input image is given to the system, as shown in Figure 3.

In all BraTS2018 data images, an initial noise reduction is required, using the QMFT algorithm, in which the image is read linearly, columnar, and diagonally without any repetition to reduce noise. The schematic of this output is in the form of Figure 4, and the result of the image that the noise reduction operation and its initial highlighting is in Figure 5.

The value of peak signal-to-noise ratio (PSNR) is illustrated in Figure 6. The analysis is done for 1120 MRI images. The mean value of PSNR is 87.35. It can be seen that the noise reduction provides an interesting picture in which a good segmentation can be applied. For this purpose, a deep neural network spiking or DSNN method is applied to the noise reduction operation output, and the BraTS2018 video data set is trained, which will be 75% training and 25% test. But the combined DSNN approach with the QAIS algorithm is made in this section so that the overall result is visible. According to the DSNN structure, it is observed that five input layers are considered, in which all BraTS2018 video data are placed. Then, there are three rows of training layers, the first of which is the training deep layer. In this row, from the deep layer, one by one, the convolution layer with windowing, and then the random polarizing layer with windowing, again the convolution layer with windowing, and then the maximum polarization layer as is located. The stimulus function of this layer is a zygomatic logarithm in that the number of general layers is 20. Then, the fully connected layer is associated with CRF, which is considered as 10 layers.

Then, there is a SoftMax layer with the QAIS algorithm designed to optimize DSNN segmentation during training and testing, more accurate mass detection, and feature selection operations. Its drive function is linear. There are separate settings for the QAIS algorithm. The initial population of this algorithm is considered to be 200. The colonial rate is 0.04, and its repetition rate is 10 cycles for optimizing the DSNN algorithm segmentation and selecting features in the SoftMax layer. In the end, there is an output layer that is a layer to display the output. The number of raw data training and testing rounds in QAIS-DSNN is equal to 7000 rounds. The QAIS-DSNN core is resilient back-propagation, and its performance is measurable with average error squares. The training process is illustrated in Figure 7. The termination criterion for the training process is mean square error as . Regrading Figure 7, 1342 epochs lead to converge the training process. When the proposed approach is applied to multiple images, the overall result will be Figure 8.

The ROC chart and the AUC rate are the proposed approach in Figure 9. This curve is known as one of the most important evaluation criteria, which measures the efficiency of classification operations in a system. In general, in a binary classification system in which the differentiation threshold differs, the ROC curve is a graphical representation of the degree of sensitivity or correct prediction versus false prediction. The ROC curve is also shown by plotting the correct positives against the predicted false positives. A number which measures and evaluates an aspect of performance is the area below the ROC curve. This area below the curve is called the AUC. A value above 0.7 to 1 indicates an excellent level of prediction and classification performance. According to Figure 9, it is observed that the value of AUC is a number below one, which shows the optimization of the proposed approach as much as possible. The presence of some similar sections with cancerous masses in the available data and presented method led to the creation of a series of minor errors that have not been adapted to the fitting line. The blue circles are the criterion values, and the red line is the ROC diagram on which the data is fitted. In some areas where the data is a bit far away, an error occurs and leads to a decrease in inaccuracy. Also in the middle line is regression called the ROC peak relative to regression, and the area below it is AUC. After applying the proposed approach, it is necessary to compare the proposed approach with other proposed methods, which are examined in terms of different evaluation criteria to determine the guarantee of the proposed approach. For this purpose, Table 1 shows a comparison in terms of average error squares, signal-to-noise ratio, and accuracy. Also, a comparison has been made in terms of Dice evaluation criteria for tumor nucleus, total tumor, and tumor areas, the results of which can be seen in Table 2.


The peak signal-to-noise ratio (decibels)Mean square errorMethodReference

72.10.23KNNMittal et al. [25]
88.40.045Genetic algorithmMittal et al. [25]
74.20.021SVMMittal et al. [25]
73.50.256SOMMittal et al. [25]
94.20.012CNNMittal et al. [25]
96.640.001GCNNMittal et al. [25]
77.70.001BAT-IT2FCMAlagarsamy et al. [31]
97.790.006QIAS-DSNNProposed approach


Tumor improvement or ET areasTumor nucleus or TCTotal tumor or WTMethodReference

81.84%88.34%91.2%ADNN-PSOIrfan Sharif et al. [32]
85.83%79.72%90.21%3D cascaded CNN-TTAWang et al. [27]
79.19%85.40%90.31%Cascaded CNNWang et al. [27]
77.07%73.04%89.56%Multiclass WNet+TTAWang et al. [27]
71.78%74.81%88.24%MCCNNHu et al. [28]
72.29%76.75%86.23%Two-stageZhou et al. [26]
70.9%75.1%85.1%Ordinary fusionZhou et al. [26]
73.44%76.58%86.38%3D UNetZhou et al. [26]
72.55%75%84.94%APFNetZhou et al. [26]
74.43%76.88%86.56%APF+3D-CRFZhou et al. [26]
74.50%80.15%91.92%QAIS-DSNNProposed approach

The next comparison is the percentage-based accuracy for segmentation to distinguish the mass region from the images, which are averaged from the BraTS data set. The results are reported in Table 3.


Accuracy (%)MethodReference

98.20%GCNNMittal et al. [25]
95%3D cascaded CNN-TTAWang et al. [27]
88.50%MCCNNHu et al. [28]
96.12%BAT-IT2FCMAlagarsamy et al. [31]
92%ADNN-PSOIrfan Sharif et al. [32]
98%PSO-LDA-GA-ANNSharif et al. [33]
98.21%QAIS-DSNNProposed approach

Finally, a comparison is made in terms of computational complexity in terms of time between the method presented in this study and the methods in reference [25], the results of which are shown in Table 4. It is noteworthy that this study has listed the system used during the processing of the proposed method, and this comparison is made on a case-by-case basis with reference [25].


Computational complexity in terms of time (seconds)MethodReference

0.84 secondsGCNNMittal et al. [25]
1.6 secondsCNNMittal et al. [25]
3.2 secondsGenetic algorithmMittal et al. [25]
2.7 secondsSVMMittal et al. [25]
4.2 secondsSOMMittal et al. [25]
3.8 secondsKNNMittal et al. [25]
2.58 secondsQAIS-DSNNProposed approach

Based on the results of the comparisons in terms of evaluation, it is observed that the proposed approach is optimal in terms of the mean error squares of most methods, but the GCNN [25] and BAT-IT2 FCM [38] algorithms have better results than the research approach. In terms of signal-to-noise ratio, the proposed method of compared algorithms has had better results. In terms of Dice evaluation criteria, most research is on the same level. There are differences in the parts of the whole tumor, the tumor nucleus, or the improved part of the tumor, depending on the different methods available. In terms of accuracy, the prediction approach has better results, but with the GCNN [25] algorithm, it is 0.01% more efficient. Also, the results were obtained at the level of convolutional methods, and the computational complexity of the proposed approach has been implemented in the system; however, the computational complexity can be seen by combining the existing algorithms.

5. Conclusion

This article is innovative in the field of noise reduction and segmentation of MRI images to detect the area of tumor masses. Also, we used the QMFT method to find noise and reconstruct it with adjacent pixels to process them horizontally, vertically, and diagonally. It is formed in the fastest time and has been able to move the noise by identifying and reviewing neighbors and matching the pixel data with neighbors based on the edge of the image. Then, the segmentation operation was performed with a QAIS-DSNN combination approach. In this approach, the deep neural network of spiking with CRF is considered, so that after the input layer—including neurons (spikes)—the training layer has convolution and polarizing layers. All of them are connected to the CRF format. Then, there is a SoftMax layer outside the training layer, which is optimized for segmentation and detection to accurately identify tumor features, in this method with the QAIS. The simulation results show that the proposed QAIS-DSNN approach has a functional advantage over the previous methods evaluation criteria. Among these evaluation results, we can point out the accuracy in segmentation and detection of the exact mass area in MRI images with an accuracy of 98.21%. Also, the average rate of error squares is 0.006, and the peak rate of the signal-to-noise ratio is 97.79 decibels. The use of lesion structural criteria includes a tumor nucleus of 80.15%, improved tumor of 74.50%, and a total tumor of 91.92%, which is a functional advantage over similar previous methods. Reducing computational complexity compared to previous methods and improving execution time by 2.58 seconds also confirms it. In the future, the plan is to use a huge dataset or transfer this system to the breast, lung, and some other tumor detection tasks. Moreover, we are going to add this automated segmentation method for CNN-based segmentation ground truth images. Furthermore, the presented filtering system can be added as a layer in the CNN method and change the resolution of the matrix in each iteration.

Data Availability

The data that support the findings of this study are openly available in BraTS2013, BraTS2015, and BraTS2018 at https://www.med.upenn.edu/sbia/brats2018/data.html

Disclosure

The funding sources had no involvement in the study design, collection, analysis or interpretation of data, and writing of the manuscript or in the decision to submit the manuscript for publication.

Conflicts of Interest

We declare no conflict of interest.

References

  1. S. Bauer, R. Wiest, L. P. Nolte, and M. Reyes, “A survey of MRI-based medical image analysis for brain tumor studies,” Physics in Medicine and Biology, vol. 58, no. 13, pp. R97–R129, 2013. View at: Publisher Site | Google Scholar
  2. D. N. Louis, A. Perry, G. Reifenberger et al., “The 2016 world health organization classification of tumors of the central nervous system: a summary,” Acta Neuropathologica, vol. 131, no. 6, pp. 803–820, 2016. View at: Publisher Site | Google Scholar
  3. B. H. Menze, A. Jakab, S. Bauer et al., “The multimodal brain tumor image segmentation benchmark (BRATS),” IEEE Transactions on Medical Imaging, vol. 34, no. 10, pp. 1993–2024, 2015. View at: Publisher Site | Google Scholar
  4. S. Bakas, “Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features,” Scientific Data, vol. 4, article 170117, 2017. View at: Google Scholar
  5. G. Litjens, T. Kooi, B. E. Bejnordi et al., “A survey on deep learning in medical image analysis,” Medical Image Analysis, vol. 42, pp. 60–88, 2017. View at: Google Scholar
  6. S. Hassantabar, M. Ahmadi, and A. Sharifi, “Diagnosis and detection of infected tissue of COVID-19 patients based on lung X-ray image using convolutional neural network approaches,” Chaos, Solitons & Fractals, vol. 140, article 110170, 2020. View at: Publisher Site | Google Scholar
  7. S. Pereira, A. Pinto, V. Alves, and C. A. Silva, “Brain tumor segmentation using convolutional neural networks in MRI images,” IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1240–1251, 2016. View at: Publisher Site | Google Scholar
  8. W. Mengqiao, Y. Jie, C. Yilei, and W. Hao, “The multimodal brain tumor image segmentation based on convolutional neural networks,” in 2017 2nd IEEE International Conference on Computational Intelligence and Applications (ICCIA), pp. 336–339, Beijing, China, 2017. View at: Publisher Site | Google Scholar
  9. M. Havaei, A. Davy, D. Warde-Farley et al., “Brain tumor segmentation with deep neural networks,” Medical Image Analysis, vol. 35, pp. 18–31, 2017. View at: Publisher Site | Google Scholar
  10. G. Wang, M. A. Zuluaga, W. Li et al., “DeepIGeoS: a deep interactive geodesic framework for medical image segmentation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 41, no. 7, pp. 1559–1572, 2019. View at: Publisher Site | Google Scholar
  11. X. Zhao, Y. Wu, G. Song, Z. Li, Y. Zhang, and Y. Fan, “A deep learning model integrating FCNNs and CRFs for brain tumor segmentation,” Medical Image Analysis, vol. 43, pp. 98–111, 2018. View at: Publisher Site | Google Scholar
  12. M. Ahmadi, S. Jafarzadeh-Ghoushchi, R. Taghizadeh, and A. Sharifi, “Presentation of a new hybrid approach for forecasting economic growth using artificial intelligence approaches,” Neural Computing and Applications, vol. 31, no. 12, pp. 8661–8680, 2019. View at: Publisher Site | Google Scholar
  13. S. Dorosti, S. J. Ghoushchi, E. Sobhrakhshankhah, M. Ahmadi, and A. Sharifi, “Application of gene expression programming and sensitivity analyses in analyzing effective parameters in gastric cancer tumor size and location,” Soft Computing, vol. 8, pp. 1–22, 2019. View at: Google Scholar
  14. F. Isensee, P. Kickingereder, D. Bonekamp et al., “Brain tumor segmentation using large receptive field deep convolutional neural networks,” in Bildverar- beitung für die Medizin 2017, pp. 86–91, Springer-Vieweg, Berlin, Germany, 2017. View at: Google Scholar
  15. S. Chen, C. Ding, and M. Liu, “Dual-force convolutional neural networks for accurate brain tumor segmentation,” Pattern Recognition, vol. 88, pp. 90–100, 2019. View at: Publisher Site | Google Scholar
  16. Z. Liu, L. Chen, L. Tong et al., Deep Learning Based Brain Tumor Segmentation: A Survey, 2020, http://arxiv.org/abs/2007.09479.
  17. K. Hu, Z. Zhang, X. Niu et al., “Retinal vessel segmentation of color fundus images using multiscale convolutional neural network with an improved cross-entropy loss function,” Neurocom- puting, vol. 309, pp. 179–191, 2018. View at: Publisher Site | Google Scholar
  18. K. Kamnitsas, C. Ledig, V. F. J. Newcombe et al., “Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation,” Medical Image Analysis, vol. 36, pp. 61–78, 2017. View at: Publisher Site | Google Scholar
  19. G. Wang, W. Li, S. Ourselin, and T. Vercauteren, “Automatic brain tumor segmentation using cascaded anisotropic convolutional neural networks,” in Proceedings International MICCAI Brainlesion Workshop, pp. 178–190, Springer, Quebec City, QC, Canada, 2017. View at: Google Scholar
  20. S. Zheng, S. Jayasumana, B. Romera-Paredes et al., “Conditional random fields as recurrent neural networks,” in Proceedings of the IEEE international conference on computer vision, pp. 1529–1537, 2015. View at: Google Scholar
  21. M. Kistler, S. Bonaretti, M. Pfahrer, R. Niklaus, and P. Büchler, “The virtual skeleton database: an open access repository for biomedical research and collaboration,” Journal of Medical Internet Research, vol. 15, no. 11, article e245, 2013. View at: Publisher Site | Google Scholar
  22. A. Myronenko, “3D MRI brain tumor segmentation using autoencoder regularization,” pp. 311–320, Springer. View at: Google Scholar
  23. L. C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, “Semantic image segmentation with deep convolutional nets and fully connected CRFs,” 2014, http://arxiv.org/abs/1412.7062. View at: Google Scholar
  24. S. Zheng, S. Jayasumana, B. Romera-Paredes et al., “Conditional random fields as recurrent neural networks,” in 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1529–1537, Santiago, Chile, 2015. View at: Publisher Site | Google Scholar
  25. M. Mittal, L. M. Goyal, S. Kaur, I. Kaur, and D. Jude Hemanth, “Deep learning based enhanced tumor segmentation approach for MR brain images,” Applied Soft Computing, vol. 78, pp. 346–354, 2019. View at: Google Scholar
  26. Z. Zhou, Z. He, and Y. Jia, “AFPNet: a 3D fully convolutional neural network with atrous-convolution feature pyramid for brain tumor segmentation via MRI images,” Neurocomputing, vol. 402, pp. 235–244, 2020. View at: Publisher Site | Google Scholar
  27. G. Wang, W. Li, S. Ourselin, and T. Vercauteren, “Automatic brain tumor segmentation using convolutional neural networks with test-time augmentation,” International MICCAI Brainlesion Workshop, BrainLes 2018: Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries, pp. 61–72, 2018. View at: Google Scholar
  28. H. Kai, G. Qinghai, Z. Yuan et al., “Brain tumor segmentation using multi-cascaded convolutional neural networks and conditional random field,” IEEE Access, vol. 7, no. 8, pp. 92615–92629, 2019. View at: Google Scholar
  29. C. Kang, X. Yu, S. H. Wang et al., “A heuristic neural network structure relying on fuzzy logic for images scoring,” IEEE Transactions on Fuzzy Systems, vol. 29, no. 1, pp. 34–45, 2020. View at: Google Scholar
  30. M. Pfeiffer and T. Pfeil, “Deep learning with spiking neurons: opportunities and challenges,” Frontiers in neuroscience, vol. 12, p. 774, 2018, https://www.frontiersin.org/articles/10.3389/fnins.2018.00774/full. View at: Google Scholar
  31. S. Alagarsamy, K. Kamatchi, V. Govindaraj, Y.-D. Zhang, and A. Thiyagarajan, “Multi-channeled MR brain image segmentation: a new automated approach combining BAT and clustering technique for better identification of heterogeneous tumors,” Biocybernetics and Biomedical Engineering, vol. 39, no. 4, pp. 1005–1035, 2019. View at: Publisher Site | Google Scholar
  32. M. I. Sharif, J. P. Li, M. A. Khan, and M. A. Saleem, “Active deep neural network features selection for segmentation and recognition of brain tumors using MRI images,” Pattern Recognition Letters, vol. 129, pp. 181–189, 2020. View at: Publisher Site | Google Scholar
  33. M. Sharif, J. Amin, M. Raza, M. Yasmin, and S. C. Satapathy, “An integrated design of particle swarm optimization (PSO) with fusion of features for detection of brain tumor,” Pattern Recognition Letters, vol. 129, pp. 150–157, 2020. View at: Publisher Site | Google Scholar

Copyright © 2021 Mohsen Ahmadi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views113
Downloads114
Citations

Related articles