BioMed Research International

BioMed Research International / 2020 / Article

Research Article | Open Access

Volume 2020 |Article ID 8125392 | https://doi.org/10.1155/2020/8125392

Shakhawan H. Wady, Raghad Z. Yousif, Harith R. Hasan, "A Novel Intelligent System for Brain Tumor Diagnosis Based on a Composite Neutrosophic-Slantlet Transform Domain for Statistical Texture Feature Extraction", BioMed Research International, vol. 2020, Article ID 8125392, 21 pages, 2020. https://doi.org/10.1155/2020/8125392

A Novel Intelligent System for Brain Tumor Diagnosis Based on a Composite Neutrosophic-Slantlet Transform Domain for Statistical Texture Feature Extraction

Academic Editor: Jinyuan Zhou
Received19 Nov 2019
Revised10 Apr 2020
Accepted08 Jun 2020
Published11 Jul 2020

Abstract

Discrete wavelet transform (DWT) is often implemented by an iterative filter bank; hence, a lake of optimization of a discrete time basis is observed with respect to time localization for a constant number of zero moments. This paper discusses and presents an improved form of DWT for feature extraction, called Slantlet transform (SLT) along with neutrosophy, a generalization of fuzzy logic, which is a relatively new logic. Thus, a novel composite NS-SLT model has been suggested as a source to derive statistical texture features that used to identify the malignancy of brain tumor. The MR images in the neutrosophic domain are defined using three membership sets, true (), false (), and indeterminate (); then, SLT was applied to each membership set. Three statistical measurement-based methods are used to extract texture features from images of brain MRI. One-way ANOVA has been applied as a method of reducing the number of extracted features for the classifiers; then, the extracted features are subsequently provided to the four neural network classification techniques, Support Vector Machine Neural Network (SVM-NN), Decision Tree Neural Network (DT-NN), -Nearest Neighbor Neural Network (KNN-NN), and Naive Bayes Neural Networks (NB-NN), to predict the type of the brain tumor. Meanwhile, the performance of the proposed model is assessed by calculating average accuracy, precision, sensitivity, specificity, and Area Under the Curve (AUC) of the Receiver Operating Characteristic (ROC) curve. The experimental results demonstrate that the proposed approach is quite accurate and efficient for diagnosing brain tumors when the Gray Level Run Length Matrix (GLRLM) features derived from the composite NS-SLT technique is used.

1. Introduction

Most contemporary vision algorithms cannot accurately perform based on image intensity values which are directly derived from the initial gray level representation. Image intensity values are highly redundant, while the amount of important information within the image might be small. The Slantlet-based transformation of the initial MR image representation into a feature representation explicitly emphasizes the useful image features without losing essential image information, reduces the redundancy of the image data, and eliminates any irrelevant information [1]. Medical images perform a crucial role in disease analysis, education, investigation, etc. In the medical domain, due to the enormous development of digital medical images, an automated classification system of brain tumors is required to help radiologists accurately identify brain tumors or perform investigation based on brain Magnetic Resonance Imaging (MRI) [2, 3]. Since 2006, numerous systems were developed in the area of medical image, which relies mainly on the extraction of low-level features such as texture, intensity, shape, and color in order to understand, characterize, and classify medical images efficiently [2]. Medical image classification is a key issue in the field of image recognition, and it is intended to classify medical images into different categories. Basically, the classification of medical images can be divided into two phases of development. Effective image features are extracted from the first stage, and the second step is to use the features to construct an image dataset model [4]. Moreover, texture analysis, the mathematical method for quantitative analysis of image pattern variation, had shown promising diagnostic potential in different brain tumors that relate to an object’s surface properties and its association with the adjacent region [57].

A brain tumor is one of the worst diseases that has risen due to an abnormal brain cell growth affecting the function of nervous systems. Various types of tumors in the brain may be benign or malignant. Cells of a benign brain tumor (low-grade glioma (LGG)) rarely invade healthy adjacent cells and have different boundaries and slow development of progression. Malignant brain tumor (HGG, BM, or recurrent glioma) cells readily invade brain or spinal cord neighboring cells and have fluid boundaries and rapid growth levels [810]. The early stage of tumor diagnosis relies on the doctor’s knowledge and experience to help patients to recover and survive. An automated brain tumor classification system is an efficient tool to help physicians to successfully follow their treatment options [11, 12]. During the past years, several automatic methods for brain image analysis have been developed to detect and classify brain tumors using MR images.

The research paper [13] addresses a fully automated system for the identification of tumor slices and the delineation of the tumor region on the basis of two-dimensional anatomic MR images. Features were extracted using Gabor wavelet and statistical feature extraction techniques, and they achieved the highest classification result with statistical features in comparison to Gabor wavelet features. Subashini and Gandhi [14] and his coworkers published an article on automatic detection and classification of MRI brain tumors using LabVIEW. A dataset of 80 images was utilized to test this approach, and they achieved 92.5% of classification accuracy. In another work [15], the authors proposed a 2-level DWT method to extract features from MR images. In the method, feature selection using PCA and DNN models was used for brain MRI classification into normal and three categories of malignant brain tumors. Gupta et al. [16] proposed a noninvasive system for brain glioma detection on brain MRIs using texture and morphological features with ensemble learning. Simulations were scored 97.37% and 98.38 on JMCD and BraTS, respectively. In [17], the authors developed a clinical support system to enhance brain tumor detection and classification using images from the BraTS dataset. The tumor region’s features were collected by the GLCM extraction technique and classified using LOBSVM with 97.69% accuracy. An approach of a deep learning (DL) model based on a CNN for the classification of brain tumor MR images was suggested by Sultan et al. [18]. The proposed system attained a substantial performance with the best overall accuracy of 98.7%. In Reference [18], the authors have addressed the new liver and brain tumor classification approach using CNN, DWT, and LSTM for feature extraction, signal processing, and signal classification, respectively. Experimental results showed that hybrid CNN-DWT-LSTM algorithms were substantially better performing, and they achieved overall performance of 98.6%. In 2019, Ullah et al. [19] developed a modified scheme to differentiate between normal and abnormal brain MR images based on a median filter, DWT, color moments, and ANN. In [20], the author proposed a machine learning approach based on delta-radiomic features of DSC-MR images. The developed algorithm was used for classifying HG and LG GBMs with an average of 90% accuracy.

Over the past few decades, many methods have been proposed in the literature for feature extraction. These techniques were based on features extracted from spatial and frequency domains, and it was observed that very few studies have been conducted on brain tumor diagnosis based on the neutrosophic domain. Amin and his colleagues [21] developed a new system of neutrosophic ranking for classifying tumors in BUS images. In the system, original BUS images were transformed into a neutrosophic set domain and various features were extracted from statistical and morphological features. Sert and Avci [22] proposed a neutrosophic set EMFSE system using maximum fuzzy entropy and fuzzy c-partition methods to identify the enhancing part of the tumor in a brain MR image. The authors in [23] proposed an effective automatic brain tumor segmentation scheme based on the NS-EMFSE method for classifying brain tumors as benign and malignant with the SVM and KNN classifier. A dataset of 500 samples was taken from various cancer categories for the TCGA-GBM dataset to test this approach, and they achieved the highest performance by the SVM classifier with 95.62%.

1.1. Neutrosophy

Neutrosophy is a branch of philosophy, introduced by F. Smarandache in 1980, which generalized dialectics and studied the origin, nature, and scope of neutralities, in addition to their interactions with numerous ideational spectra [24]. In neutrosophy theory, every event has a definite degree of truth (), falsity (), and indeterminacy () that have to be considered independently from each other [23, 2528]. Therefore, is an idea, theory, event, concept, or entity; is the opposite of ; and the neutrality {} means neither nor , that is, the neutrality between the two extremes [29, 30].

1.2. Concept of Neutrosophic Set

A neutrosophic set is a generalization of the theory of fuzzy set, intuitionistic fuzzy set, paraconsistent set, dialetheist set, paradox set, and tautological set where each element of the universe has a degree of truth, falsity, and indeterminacy, respectively. Unlike in fuzzy sets, the neutrosophic set presents the additional domain () which provides a more effective way to handle higher degrees of uncertainty. Let be a universe of discourse set and a neutrosophic set in is characterized by three neutrosophic components: , , and are defined to estimate the membership degree (truth membership degree), nonmembership degree (falsity membership degree), and the indeterminacy membership degree of an element independently. The neutrosophic schema in the general case is shown in Figure 1.

The novelty of the proposed approach is to apply Slantlet transform in each of the neutrosophic sets to extract statistical texture features, which has not been explored and performed on MICCAI BraTS dataset. Furthermore, different individual and combined feature extraction methods using composite NS-SLT were compared through their classification accuracies to select the effective approach with four types of neural network classification techniques. To evaluate the performance, extensive experiments were carried out which show that the proposed composite system achieves excellent results and classifies images accurately.

2. Materials and Methods

The overall design of the proposed framework is shown in Figure 2. First, MR images of patients are acquired, cropped, and resized in the preprocessing step; then, statistical texture features are extracted from SLT in the neutrosophic domain. Afterwards, feature selection is performed to choose the most salient features, followed by applying four neural network classifiers to identify the tumor as benign or malignant derived from the extracted features. Finally, the performance is evaluated by using certain parameters. The detail of these given methods has been presented in the subsequent subsections.

2.1. Dataset

Images in the MICCAI Brain Tumor Segmentation 2017 Challenge (BraTS 2017) were used to analyze and evaluate our proposed approach, which is one of the standard and benchmarked datasets [9, 3133]. It is comprised of 210 preoperative MR images of patients from high-grade glioma (HGG) volumes and 75 MRIs from low-grade glioma (LGG) volumes collected from multiple centers. For each patient, there are four MRI modalities, including the native T1-weighted (T1), contrast-enhanced T1-weighted (T1ce), T2-weighted (T2), and T2 fluid-attenuated inversion recovery (FLAIR) (Figure 3). After their preprocessing, the data provided are distributed, i.e., skull-stripped, coregistered to the same anatomical template, and with the same resolution interpolated into and with a sequence size of . In order to homogenize data, each modality scan is rigidly coregistered with T1Ce modality, because in most cases, T1Ce has the highest spatial resolution. Therefore, for our experiments, 285 brain MRI tumor (T1Ce) images are used, out of which 210 were cancerous (malignant) tumors from HGG and 75 were benign tumors from LGG.

2.2. Preprocessing

In the preprocessing stage, the input images (axial images) were initialized. The middle slice in an MRI volume is considered to have all the tissue regions. The pixels (nonobject) in the background are usually very prominent in MR images, and the processing time of brain extraction can be reduced considerably by separating target pixels from background pixels. Therefore, in this step, the bounding box cropping approach is computed in order to extract the brain portion alone as the AOI by removing the unwanted background from the input image. Before importing the input MR images into the system, the cropped MR images are resized into pixels.

2.3. The Image in Neutrosophic Domain

Let be a universe of discourse and be a set included in , which is composed of bright pixels. The image in the neutrosophic domain () is represented using three distinctive membership components (, , and ), where defines the truth scale, defines the scale of false, and characterizes the scale of intermediate. All considered components are autonomous from each other. A pixel () of an image in the neutrosophic domain is characterized as [2628, 30, 34] and belongs to set in the following way: it is true membership function in the bright pixel set, indeterminacy membership function in the set, and a falsity-membership function in the set, where varies in , varies in , and varies in . There is a valuation for each component in [0, 1]. In the image domain, pixel is transformed into a neutrosophic domain by calculating in equations (1), (2), (3), (4), (5) and (6), where considered as a probability that pixel belongs to white set (object), indeterminate set, and nonwhite set (background), respectively (see Figure 4). This is the primary benefit of neutrosophy in image processing, and it can be taken at the same time when the decision is made for each pixel in the image. In [22, 23, 3538], the following basic equations were proposed for transforming images from a pixel domain to the neutrosophic domain: where represents the intensity value of an image in the pixel domain; , , and are true, indeterminacy, and false sets, respectively, in the neutrosophic domain; can be defined as the local mean value of ; and is the homogeneity value of at (), which is described by the absolute value of the difference between intensity value of an image and its local mean value .

2.4. Slantlet Transform (SLT)

The Slantlet transform is an improved orthogonal DWT variant with two zero moments and better time localization which was first utilized by Selesnick to evaluate nonstationary signals [39]. DWT is usually carried out by filter bank iteration, where a tree structure is utilized. Slantlet transform is inspired by an equivalent DWT implementation, in which a filter bank in a parallel structure is implemented [40]. DWT utilizes a product form of basic filters in some of these parallel branches, and the filter bank “Slantlet” uses a similar structure in parallel. However, there is no product type of implementation for the component filter branches, which means that SLT has extra independence. SLT will produce a filter bank, where each filter has its length in the power of 2; this results in a periodic output for the analysis filter bank and reduces the samples which support approaches one-thirds, as increases [41].

For a mathematical perspective of the transformation of Slantlet, let us take a generalized representation of Figure 5, for () scales. The filters in scale must be , , and to analyze the signal where each filter has an appropriate support. For , the SLT filter bank uses () number of pairs of channels, i.e., channels in total. The low pass filter is then combined with its adjacent filter, where a downsampling of is followed by any filter. The channel pairs of each constitute a , followed by a downsampling by and the downsample by a reversed time version The following expressions are represented by the following, as the filters , , and implement linear forms in pieces:

Two issues must be taken into account when computing SLT on MR images. Firstly, input signal length should be power of two, or higher than, the analysis filter bank length of the SLT, since all filter lengths are power of two in SLT filter bank. Secondly, the matrix of transformation has to be constructed. In a 2D SLT decomposition, there is usually an image that is divided into two parts: approximation and detailed parts. The approximation part includes one low-frequency LL subband, and detailed parts include three high-frequency subbands: LH, HL, and HH, as Figure 6 illustrates, where H and L represent the high- and low-frequency bands, respectively. The low-frequency subband component (LL) includes the inventive information of the original image. On the contrary, the LH, HL, and HH subbands retain the information associated with the contour, edge, and the image’s other details. In the image, high coefficients characterize the important information; the low (insignificant) coefficients meanwhile are deliberated as trivial information or noise. Therefore, such small coefficients should be avoided for the best results. In this work, the SLT was utilized on MR images in spatial and neutrosophic domains to extract the statistical features of the images.

2.5. Feature Extraction

Feature extraction is the process of transforming the raw pixel values from an image into a set of features, normally distinctive properties of input patterns that can be used in the selection and classification tasks. Feature extraction techniques are usually divided into the geometrical, statistical, model-based, and signal processing [14, 16, 18, 42]. This stage involves obtaining important features extracted from MR images. The main features can be used to indicate the texture property, and the information is stored in the knowledge base for the system training. Three sets of statistical texture features (GLDS, GLRLM, and GLCM) are included for feature extraction in the proposed system. The obtained texture features by different methods are used individually and fused with each other for the classification process. Table 1 shows all 22 statistical textural features extracted from each technique.


TechniqueTextural featuresNo. of extracted features

GLCMHomogeneity, energy, entropy, symmetry, contrast, correlation, moment 1, moment 2, moment 3, moment 410
CLRLMShort run emphasis, long run emphasis, gray level nonuniformity, run percentage, run length nonuniformity, low gray level run emphasis, high gray level run emphasis8
GLDSAngular second moment, contrast, mean, entropy4

2.5.1. Gray Level Cooccurrence Matrix (GLCM)

GLCM is one of the most widespread techniques of texture analysis that quantitatively measured the frequency of different combinations of pixel brightness values (gray levels) which occur in an image, and it has been used in a number of applications, e.g., [4248]. In this step, texture features that contain information about the image are computed by GLCM to extract second-order statistic texture features (Table 1).

(1) Neutrosophic Image Homogeneity. Homogeneity also called inverse difference moment is a value that measures the similarity of the distribution of elements in the gray level cooccurrence matrix which is defined in [48]. The values vary between 0 and 1, and a higher value reveals a smoother texture feature.

Mathematically, homogeneity of an image in the spatial domain is defined as where denotes element of GLCM; is the number of gray levels in the image; and demonstrates the number of rows and columns in the image.

The neutrosophic image homogeneity is defined as the summation of the homogeneities of three sets , , and . The basic equations to transform images from the pixel domain to the neutrosophic domain are calculated as follows:

(2) Neutrosophic Image Energy.

(3) Neutrosophic Image Entropy.

(4) Neutrosophic Image Contrast.

(5) Neutrosophic Image Symmetry.

(6) Neutrosophic Image Correlation.

(7) Neutrosophic Image Moment 1.

(8) Neutrosophic Image Moment 2.

(9) Neutrosophic Image Moment 3.

(10) Neutrosophic Image Moment 4.

2.5.2. Gray Level Run Length Matrix (GLRLM)

The concept, GRLM, is based on the reality that many neighboring pixels with the same gray level are characterized by coarse texture features [42, 44, 45, 47]. For a given image, GLRLM is calculated by representing the total runs of pixels having gray level and run length in a particular direction. Textural features are calculated from a set of components used to explore the essence of the textures of the image. Many numerical texture measurements can be calculated from the original run-length matrix . At the end, eight original features of run length statistics for the neutrosophic domain are derived (Table 1).

(1) Neutrosophic Image Short Run Emphasis (SRE). where denotes the number of runs of pixels that have gray level and length group ; is the total number of runs in the image; is the number of gray levels (bins); and is the number of run lengths (bins):

(2) Neutrosophic Image Long Run Emphasis (LRE).

(3) Neutrosophic Image Gray Level Nonuniformity (GLN).

(4) Neutrosophic Image Run Percentage (RP). where is the total number of pixels in the image:

(5) Neutrosophic Image Run Length Nonuniformity (RLN).

(6) Neutrosophic Image Low Gray Level Run Emphasis (LGRE).

(7) Neutrosophic Image High Gray Level Run Emphasis (HGRE).

2.5.3. Gray Level Difference Statistics (GLDS)

The GLDS emphasizes the histogram of the absolute differences in the gray level between the two pixels that are separated by a displacement vector to calculate the tumor region’s texture coarseness [49]. Let be the displacement vector between two image pixels and the gray level difference at distance :

is the histogram of the gray level differences at the specific distance . One distinct histogram exists for each distance . The following four statistical features were derived from the histogram of gray level differences in the neutrosophic domain (Table 1).

(1) Neutrosophic Image Angular Second Moment.

(2) Neutrosophic Image Contrast.

(3) Neutrosophic Image Mean.

(4) Neutrosophic Image Entropy.

2.6. Feature Selection

The large number of texture features causes difficulty in ranking, prolongs computational time, and involves more memory space. Thus, the selection of features was regarded as part of the design of the proposed system. In our paper, the analysis of variance (ANOVA) technique was used to reduce the dimension of data based on its significance and variance and avoid losing too much information (Table 2). ANOVA is a powerful tool for determining if two or more sets of data have a statistically significant difference [50]. A normalization process on the input feature set was performed as part of data preparation prior to applying the ANOVA method.


TechniquesNo. featuresFeature selection method (ANOVA)
Scenario 1 (NS)Scenario 2 (SLT)Scenario 3 (NS-SLT)
No. features valueNo. features valueNo. features value

GLDS4222
GLRLM8333
GLCM10442
Fusion of GLRLM and GLDS12655
Fusion of GLCM and GLDS14756
Fusion of GLCM and GLRLM18985
Fusion of GLCM, GLRLM and GLDS2210107

2.7. Classification of Brain Tumors

Classification is a machine learning technique in which training data are used for building models and the model is used to predict new data [9, 16, 21, 51, 52]. In order to evaluate algorithm performance, the developed model is evaluated using testing data. Classification includes a wide range of decision-making approaches that are used in the CAD system [4]. Pixel-based image classification techniques analyze the numerical properties of selected image feature vectors and organize data into categories. In this study, four different classification techniques have been used, namely, DT-NN, SVM-NN, KNN-NN, and NB-NN, as classifiers to classify brain tumors.

3. Experimental Results and Discussions

All experiments were conducted in MATLAB using brain tumor images described in Section 2.1. Four pattern recognition neural network classifiers have been used. In addition, several statistical features such as GLDS, GLRLM, and GLCM (Table 1) were derived from different proposed scenarios (NS, SLT, and composite NS-SLT). The entire dataset was divided into training and testing sets with the ratio of 80 : 20 percent with the 10-fold cross-validation procedure. Performances of the three various scenarios were analyzed through a number of different measures [53, 54]. Further, performance evaluation accuracy of the statistical prediction system can also be done by calculating and analyzing the ROC curve. The ROC curve is a plot of the true-positive rate (sensitivity) versus the false-positive rate (1-specificity) for different thresholds over the entire range of each classifier output values. In contrast with the classification accuracies obtained from truth tables, ROC analysis is independent of class distribution or error costs.

All results were first analyzed using boxplot diagrams that provided an overview of statistical values and distributions of benign and malignant brain tumors, as shown in Figure 7. Comparing sample medians regarding GLRLM-SRE (Figures 7(j)7(l)), GLCM energy (Figures 7(p)7(r)), and GLCM symmetry features (Figures 7(s)7(u)), it is clearly visible that composite NS-SLT followed by texture feature extraction methods was significantly better compared to NS and SLT methods individually. Also, GLRLM-GLNU (Figures 7(g)7(i)) and GLRLM-RP (Figures 7(m)7(o)) features using both composite NS-SLT and SLT methods showed better performance than the NS-based texture method; however, GLDS-ASM and GLDS mean features (Figures 7(a)7(f)) yield poor results, because an overlap of statistical features was observed between benign and malignant brain tumor categories in all scenarios. As a result, the composite NS-SLT method has an effective ability for brain tumor classification in comparison to other implemented techniques.

For each scenario, a different composition of each group of statistical and textural features was made. Table 2 presents the performance of each scenario followed by various pattern recognition classifiers (after applying ANOVA), starting by deriving each group (GLDS, GLRLM, and GLCM) features individually to see which group performs better in the classification stage with the minimum number of features. The performance metrics of NS, SLT, and composite NS-SLT scenarios for each of the proposed individual category of textural feature extraction corresponding to each scenario are shown in Table 3 and Figure 8. The GLRLM features derived by composite NS-SLT recorded the highest average classification accuracy rate with SVM-NN classifier 98.94% and an AUC of 0.99. As with all classifiers, GLRLM and GLCM features derived from composite NS-SLT achieved excellent average classification accuracy except for the GLDS features which achieved the lowest average classification results with KNN-NN and DT-NN classifiers, respectively.


FeaturesClassifier methodsTechniquesPerformance metrics
Accuracy (%)PrecisionSensitivitySpecificityAUC

GLDSDT-NNNS
SLT
NS-SLT
SVM-NNNS
SLT
NS-SLT
KNN-NNNS
SLT
NS-SLT
NB-NNNS
SLT
NS-SLT

GLRLMDT-NNNS
SLT
NS-SLT
SVM-NNNS
SLT
NS-SLT
KNN-NNNS
SLT
NS-SLT
NB-NNNS
SLT
NS-SLT

GLCMDT-NNNS
SLT
NS-SLT
SVM-NNNS
SLT
NS-SLT
KNN-NNNS
SLT
NS-SLT
NB-NNNS
SLT