Medical diagnostics, a technique used for visualizing the internal structures and functions of human body, serves as a scientific tool to assist physicians and involves direct use of digital imaging system analysis. In this scenario, identification of brain tumors is complex in the diagnostic process. Magnetic resonance imaging (MRI) technique is noted to best assist tissue contrast for anatomical details and also carries out mechanisms for investigating the brain by functional imaging in tumor predictions. Considering 3D MRI model, analyzing the anatomy features and tissue characteristics of brain tumor is complex in nature. Henceforth, in this work, feature extraction is carried out by computing 3D gray-level cooccurence matrix (3D GLCM) and run-length matrix (RLM) and feature subselection for dimensionality reduction is performed with basic differential evolution (DE) algorithm. Classification is performed using proposed extreme learning machine (ELM), with refined group search optimizer (RGSO) technique, to select the best parameters for better simplification and training of the classifier for brain tissue and tumor characterization as white matter (WM), gray matter (GM), cerebrospinal fluid (CSF), and tumor. Extreme learning machine outperforms the standard binary linear SVM and BPN for medical image classifier and proves better in classifying healthy and tumor tissues. The comparison between the algorithms proves that the mean and standard deviation produced by volumetric feature extraction analysis are higher than the other approaches. The proposed work is designed for pathological brain tumor classification and for 3D MRI tumor image segmentation. The proposed approaches are applied for real time datasets and benchmark datasets taken from dataset repositories.

1. Introduction

Brain tumor is a common malignancy found in humans. Basic knowledge of tumor activities and interactions in malignancies enables efficient diagnosis and clinical understanding for the physician. As a result, brain tumor optimization methodologies possess higher potential for carrying out clinical diagnostics. In the growing scenario, medical imaging techniques (X-Ray, CT, MRI), functional imaging techniques (single photon emission computed tomography—SPECT, positron emission tomography—PET, and functional magnetic resonance imaging—fMRI), and signal diagnostic techniques such as ECG and EEG are used widely to achieve a prior analysis of the malignancies and for appropriate therapy planning. Locating brain tumor segmentation and classification of brain and tissues within magnetic resonance (MR) images is integral to the treatment of brain cancer. Segmenting brain tumors in MR images involves classifying each voxel as normal brain tissues and tumor tissues, based on the description of that voxel. This task, a prerequisite for treating brain cancer using radiation therapy, is typically done manually by expert medical doctors, who find this process laborious and time-consuming. Replacing this manual effort with a good automated pattern recognition model would save expert time; the resulting labels may also be more accurate and more consistent.

Texture analysis [1] is an efficient measure to estimate the structural orientation, roughness, smoothness, or regularity differences of diverse regions in an image scene, showing promising results as an image analysis method for detecting nonvisible and visible lesions, with a number of applications in magnetic resonance imaging MRI. Most classification techniques offer gray-level pixel based statistical features. Also, many statistical and machine learning perceptions have been identified for medical image classification. The major disadvantages in the existing approaches [26] are based on the independency in the considered process from any functional form, even when no prior assumptions are possible and when only data is available. As a result the process of feature selection and classification remains a challenge due to the facts of lack of large diversity in shape and appearance of clear edge between adjacent tissues [7], intensities of same tissues varying in different locations, noisy images, accuracy in classifier design and interpretability of the volumetric feature analysis.

The choice of highly discriminating texture features is the most important factor for a success in texture segmentation, but this has been neglected in earlier approaches. 2D cooccurrence or run-length features may lack the sensitivity to identify larger scale or more coarse changes in spatial frequency. Secondly, there are a great many pixels (such as 256 × 256 × 124) for 3D MR images. Consequently, segmentation will have high computational complexity and require large memory storage. This problem can be solved by applying the 2D methods sequentially, and even experts segment the images in this way. However, this will lose some of the geometric information.

To overcome the above difficulty, in recent years, some of the research work shows various 3D texture features [810]. All these approaches applied a 3D GLCM and 3D RLM method to computed tomography (CT) images to separate various organs of human body. Few other approaches [2, 11] computed statistical, gradient, and Gabor filter features at multiple scales and orientations in 3D to capture the entire range of shape, size, and orientation of the tumor on automated segmentation of prostatic adenocarcinoma from high-resolution MR images of brain. Certain approaches analyzed mean and standard deviation of FLAIR signal intensities of texture analysis on the hippocampal body [12]. A volumetric congruent local binary pattern (LBP) algorithm for 3D neurological image retrieval was also carried out. This proposed work performs volumetric three dimensional texture analysis to extract the better features of the considered images and these extracted features are further utilized for classification the images.

Therefore in this paper, basic differential evolution [13] is employed to select appropriate features for dimensionality reduction and thus DE is used as a feature subselection process after the features are extracted employing 3D GLCM and RLM. After the selection of optimal features using DE algorithm, then the refined group search optimizer based extreme learning machine is used to carry out the classification process. In the proposed RGSO based ELM, the weights and bias to ELM are optimized using RGSO for better simplification and classification of pathological brain tissues and tumors. The schematic flow for 3D image classification framework is shown in Figure 1.

2. Proposed 3D Volumetric Feature Extraction and Subselection Model

In general, feature extraction is a process that transforms the input data into a set of features. A feature is said to be a rich one when it is characterized by appropriate texture, shape, intensity, and location. All these said features are selected with an expert opinion from the radiologists. 3D texture measures describe the distribution using joint second order statistics of volume element (voxel) values in an image. Since the 2D GLCM algorithm is not able to describe the pixel pair displacements in 3D images, the approach extends to the GLCM and RLM for 3D volumetric data. Basically the image is represented by a function containing three space variables , , and , where , , and . The function can take any value , where is total number of intensity levels in the image. The optimal feature set is categorized as follows.Image intensity (e.g.: mean, standard deviation)Texture features (e.g.: gray-level cooccurrence and run-length matrices).

2.1. 3D Gray-Level Cooccurrence Matrices

The gray-level cooccurrence matrix was originally proposed in the year 1973 [14]. The 3D GLCM developed in this work is extended from original 2D GLCM, a three-dimensional cooccurrence matrix, , which is also an matrix, where is the number of gray levels of an image. In order to improve the computational efficiency, the number of gray levels (i.e., data-bit) is usually reduced. The 3D GLCM is basically defined as where represents the th column and represents the th row in 3D GLCM; represents the relation of a voxel pair (voxel pairs are defined by a distance and direction which can be represented by a displacement vector , where and represent the displacement (in pixels) along -axis and -axis in spatial domain and is the number of bands in the spectral domain). and are the gray values in the position and of a moving window , and , , are denoted as the position of a moving window in the volumetric data. The three-component vector with each component equal to −1, 0, or 1 is used to define different directions [15]. The vector describes neighbour’s direction in Cartesian coordinate system [6].

Four variables that are required to use GLCM include the quantization level of the image, the size of the moving window, the direction and distance of pixel pairs, and the statistics used as texture measurements. According to these four parameters, texture images can be extracted using 3D GLCM and can be used as features for analysis or classification.

2.2. Run-Length Matrices

Run-length statistics capture the coarseness of a texture in specified directions. A run is defined as a string of consecutive pixels which have the same gray-level intensity along a specific linear orientation to describe the frequency of appearance. Fine textures tend to contain more short runs with similar gray-level intensities, while coarse textures have more long runs with significantly different gray-level intensities [1, 16].

A run-length matrix is defined as follows: each element represents the number of runs with pixels of gray-level intensity equal to and length of run equal to along a specific orientation. For a given 3D image, presented as a series of slices in a preferred slice orientation, a run-length matrix is defined as follows: each element represents the number of runs with pixels of gray-level intensity equal to and length of run equal to along the direction. An orientation is defined using a displacement vector , where , , and are the displacements for the -axis, -axis, and -axis, respectively. Unlike the 2D texture characterization, the volumetric texture requires 13 different displacements from a total of 26 possible displacements in a three dimensional space [17].

The length of the run is the total number of pixel points in the run and the gray-level run-length feature is estimated using where is the maximum gray level and is the maximum run length which is equal to max . The element specifies the estimated number of times that a given picture contains a run length for a gray level in the direction of the angle and , , and denote the length of the ROI in , , and directions. The textural features are measured from .

Once the run-length matrices are calculated along each direction, several texture descriptors are calculated to capture the texture properties and are differentiated among various features. These descriptors can be used either with respect to each direction or by combining them if a global view of the texture information is required. Eleven descriptors are typically extracted from the run-length matrices: short run emphasis (SRE), long run emphasis (LRE), high gray-level run emphasis (HGRE), low gray-level run emphasis (LGRE), pair-wise combinations of the length and gray-level emphasis (SRLGE, SRHGE, LRLGE, and LRHGE), run-length nonuniformity (RLNU), gray-level nonuniformity (GLNU), and run percentage (RPC). Some of these descriptors reflect specific characteristics in the image. For example, SRE measures the distribution of short runs in an image, while run percentage measures both the homogeneity and the distribution of runs of an image in a specific direction.

2.3. 3D GLCM and RLM Based Feature Extraction

The extraction of optimal features is based on the characteristics of texture, shape, intensity of image, and spectral density. The key approach employed to extract volumetric features is based on 3D GLCM, gradient vector composition, and RLM. 3D volume analysis process occurs in 13 directions, which results in 13 3D GLCM and RLM. Figure 2 shows the 3D volumetric directions.

In a 3D volumetric space, the directions are selected by linking a voxel to each of its nearest 26 () neighbours, respectively, leading to 13 different directions from a total of 26 possible directions. Each of these slices is processed at once producing only one run-length encoding matrix for all consecutive slices forming the 3D image and, thus, the run-length computation for the volumetric texture is faster. In this case, it should be observed that 11 texture features are calculated for characterizing the texture for each subregion as in Table 1.

The spatial dependence of gray-level values across multiple slices is captured by 3D GLCM. The GLCM matrix is calculated for 0, 45, 90, and 135 degrees for and a distance scale of 1. Here 4 spatial distances and 13 directions are chosen, computing 52 (13 × 4) directional vectors and cooccurrence matrices. Henceforth, statistical measures considered are variance, entropy, energy, contrast, and homogeneity, which are calculated for each matrix. Employing these statistical measures, the feature vector of 260 (5 × 52) components is formed. The remaining features include 11 RLM and 2 gradient vector and gradient orientation features. The optimal features set with 18 texture descriptors were calculated.

2.4. Feature Subselection Model

The process of subselection in image processing selects optimal feature subset from the existing feature sets to achieve output conceptions in image classification. The aim of the subselection process is to extract best minimal subset from the original element set, instead of transforming the data to new dimensions. Various measures will be taken to analyze features from different perspectives, but few features extracted tend to behave similar without any specific variations. Also, the improvement in the dimensionality of texture features decreases the memory and the computational time. The criterion here is to identify the features that are correlated or predict the class label. The objective of this search in subset includes maximization of this criterion. Feature subselection model typically incorporates a search strategy for exploring the space of feature subsets [18]. More sophisticated search strategies such as genetic algorithm (GA) [19, 20] or simulated annealing (SA) can be employed to better explore the search space. This work invokes differential evolution (DE) technique for feature subselection.

In recent years, there has been a growing interest in evolutionary algorithms for diverse fields of science and engineering. The differential evolution algorithm (DE) [13] is a relatively novel optimization technique for solving numerical-optimization problems. The algorithm has successfully been applied to several sorts of problems as it has claimed a wider acceptance and popularity, following its simplicity, robustness, and good convergence properties.

2.4.1. Need for DE Algorithm

The DE algorithm is population based including a simple and direct searching algorithm for globally optimizing multimodal functions. Just like the genetic algorithms (GA), it employs crossover and mutation operators as selection mechanisms. As previously mentioned, an important difference among other evolutionary computational techniques, such as genetic algorithms (GA), is that the GA relies on the crossover operator which provides the exchange of information required to build better solutions. DE algorithm fully relies on the mutation operation as its central procedure. The applicability of DE algorithm, in comparison with that of the other approaches like GA, simulated annealing, or Tabu search, includes its exploration and exploitation capability because of its nonuniform crossover and mutation operations. The advantage of DE algorithm in exploring is utilized for its convergence for feature selection process. This enables the search to be focused on the most promising area of the solution space. The mutation operation is based on the differences of randomly sampled pairs of solutions within the population. Besides being simple and capable of globally optimizing multimodal search spaces, the DE algorithm shows other benefits: it is fast, easy to use, and can very easily adapt in the case of integer or discrete optimizations. It is quite effective for nonlinear constraint optimization, including penalty functions.

2.4.2. DE Algorithm Flow Process [13]

Classic DE algorithm begins by initializing a population of and -dimensional vectors with parameter values which are randomly and uniformly distributed between the prespecified lower initial parameter bound and the upper initial parameter bound :

The subscript is the generation index, while and are the parameter and population indexes, respectively. Hence, is the th parameter of the th population vector in generation .

To generate a trial solution, DE algorithm first mutates a best solution vector from the current population by adding it to the scaled difference of two other vectors from the current population, with being the mutant vector. Vector indexes and are randomly selected considering that both are distinct and different from the population index (i.e., ). The mutation scale factor is a positive real number typically less than 1.

The next step considers one or more parameter values of the mutant vector to be exponentially crossed with those belonging to the th population vector . The result is the trial vector The crossover constant controls the section of parameters belonging to the mutant vector which contributes to the trial vector. In addition, the trial vector always inherits the mutant vector parameter with a random index to ensure that the trial vector differs by at least one parameter from the vector to be compared .

Finally, a selection operation is used to improve the solutions. If the cost function of the trial vector is less than or equal to the target vector, then the trial vector replaces the target vector on the next generation. Otherwise, the target vector remains in the population for at least one new generation: Here, represents the cost function. These steps are repeated until a termination criterion is attained or a predetermined generation number is reached.

The differential evolution process reduces the dimensionality into 7 most decisive features out of the 18 features extracted. The optimal features selected after this process include gradient vector parameter, gray-level nonuniformity, energy, variance, entropy, sum variance, and short run emphasis. Therefore, from the existing 18 features, the optimal 7 feature subsets are extracted and applied for the next step.

3. ELM-RGSO for Brain Tumor and Tissue Classification

Over the past few decades it has been noted that artificial neural networks [21] play a major role in pattern recognition and image classification applications. It is because of their generalization and conditioning capabilities, requirement of minimal training points, and faster convergence time. ANNs are found to perform better and result in faster output in comparison with that of the conventional classifiers. Various neural network architectures [21] like radial basis function network, probabilistic neural network, back propagation neural network, and support vector machines and its variants were used for pattern and image classification applications; the underlying limitations in all these cases include the selection time incurred due to the preprocessing speed delay. For better classification accuracy, more training data is to be utilized in comparison to that of the testing data. Also, the selection of appropriate training algorithms is important, which enables the considered application to not experience the local or global minima problem. The above addressed criteria for improving the training performance, so as to result in better classification accuracy, have been noted in a neural network architecture, extreme learning machine (ELM) classifier proposed in [22, 23], which handles the training for single hidden layer feedforward neural networks.

3.1. Basic Extreme Learning Machine (ELM) Classifier

Extreme learning machine classifier [23] is a single-hidden layer feedforward neural network, wherein the weights for the input and hidden layer, as well as the respective biases, are randomly assigned without any training process. ELM is considered as a network architecture, which reduces the training time since its network output weights are computed using Moore-Penrose inverse and using the norm least square solution. ELM [24, 25] is best suited for larger training samples and also the effect of number of hidden neurons using different ratios of the number of features of testing and training data was examined. This classifier is compared with that of the conventional neural network classifiers using the classification rate for brain tissue and tumor classification. Figure 3 shows the basic ELM architecture.

The basic ELM classifier algorithm is given as follows.

Give a training set , kernel function , and hidden neuron .

Step 1. Select suitable activation function and number of hidden neurons for the given problem.

Step 2. Assign arbitrary input weight and bias .

Step 3. Calculate the output matrix at the hidden layer

Step 4. Calculate the output weight β as where is the Moore-Penrose generalized pseudoinverse of hidden layer output matrix.

3.2. Group Search Optimizer (GSO) Algorithm

The group search optimizer algorithm [26] used in this paper is based on the biological producer-scrounger (PS) model, which assumes group members search either for “finding” (producer) or for “joining” (scrounger) opportunities. Animal scanning mechanisms (e.g., vision) are incorporated to develop the GSO algorithm. GSO also employs “rangers” which perform random walks to avoid entrapment in local minima. The population of the GSO algorithm is called a group and each individual in the population is called a member. The GSO algorithm is implemented for this work because of its nature of random walk in various directions. The movements of the members to find the solution are carried out in a fast manner by eliminating the less efficient members in the group. This results in the accurate and faster convergence of the algorithmic process and, henceforth, this GSO algorithm has been chosen to be implemented for the proposed work.

In GSO [27], there are three types of members in a group: producers, scroungers, and dispersed members. There is only one producer and remaining members are either scroungers or dispersed members. Dispersed members are less efficient members who perform random walks. At each iteration, a group member, located in the most promising area, conferring the best fitness value, is chosen as the producer. The other group members are selected as scroungers or rangers by random. Then, each scrounger takes a random walk towards the producer, and each ranger takes a random walk in the arbitrary direction.

In -dimensional search space, the th member at th iteration has a current position and a head angle . Search direction of th member is a unit vector that can be calculated from via a Polar to Cartesian coordinate transformation.

At th iteration the producer behaves as follows.(1)The producer will scan at zero degree and then scan laterally by randomly sampling three points in the scanning field: one point at zero degree, one point in the right hand side hypercube and one point in the left hand side hypercube is maximum pursuit angle and is maximum pursuit distance. is a normally distributed random number with mean 0 and standard deviation 1 and is a uniformly distributed random sequence in the range .(2)The producer will then find the best point with the best resource (fitness value). If the best point has a better resource than its current position, then it will fly to this point or it will stay in its current position and turn its head to a new randomly generated angle. Consider where is the maximum turning angle.(3)If the producer cannot find a better area after iterations, it will turn its head back to zero degree, where is a constant.

At th iteration, th scrounger walks randomly towards the producer. Consider where is a uniform random sequence in the range . Operator is the Hadamard product or the Schur product, which calculates the entry wise product of the two vectors. If a scrounger finds a better location than the current producer and other scroungers, then it will switch as producer in the next iteration.

The group members, who are less efficient foragers than the dominant, will be dispersed from the group. If the th group member is dispersed, it will perform ranging. At the th iteration, it generates a random head angle using (4); and then it chooses a random distance and moves to the new point

To maximize their chances of finding resources, the GSO algorithm employs the fly-back mechanism [9] to handle the problem specified constraints. When the optimization process starts, the members of the group search the solution in a feasible manner. If any member moves into the infeasible region, it will be forced to move back to the previous position to guarantee a feasible solution.

3.3. Developed Refined GSO (RGSO)

In the original GSO, 75% rest members will perform scrounging and the remaining 25% rest members will perform ranging. In this paper, the ranging operation for the remaining 25% rest members will not be done as it is done in original GSO. Instead, these members for ranging operation will learn from the “worst” member in its group. This refinement of learning from “worst” member leads to discovering better solution regions in complex optimization search spaces.

Compared to the original GSO [26], RGSO algorithm searches more promising regions to find the global optimum. The difference between GSO and RGSO is that the differential operator is applied to only accept the basic GSO generating new better solution for each krill instead of accepting all the krill updating adopted in krill herd (KH). This is rather greedy. The original GSO is very efficient and powerful but highly prone to premature convergence. Therefore, to evade premature convergence and further improve the exploration ability of the original GSO, a differential guidance is used to tap useful information in all the krill individuals to update the position of a particular krill individual. Equation (18) expresses the differential mechanism. Consider where is the first element in the dimension vector . is the th element in the dimension vector . is the first element in the dimension vector . is the random integer generated separately for each , between 1 and , but .

Accordingly, the refined GSO is presented in Algorithm 1.

Set := 0;
Randomly initialize positions and head angles of all members;
WHILE (the termination conditions are not met)
For (each members in the group)
  Calculate fitness:    Calculate the fitness value of current member:
  Choose producer:  Find the producer of the group;
  Perform producing: (1) The producer will scan at zero degree and then scan laterally by randomly sampling three points
                 in the scanning field using (10) to (12).
             (2) Find the best point with the best resource (fitness value). If the best point has a better resource
               than its current position, then it will fly to this point. Otherwise it will stay in its current
               position and turn its head to new angle using (13).
             (3) If the producer cannot find a better area after iterations, it will turn its head back to zero
               degree using (14);
  Perform scrounging: Randomly select 75% from the rest members to perform scrounging using (15);
  Perform ranging:    For the rest 25% members, they will perform ranging:
               Find the worst point with the worst resource (fitness value). If the worst point has a better
               resource than its current position, then it will fly to this point. These members will update
               their position based on this worst resource.
 Set := ;

Since the problem of interest in this research is complex in nature, the above refined GSO will be able to discover better regions to train the ELM for brain tumor classification.

3.4. RGSO Based ELM for Brain Tumor Classification

This proposed methodology combines the concept of RGSO for optimizing the weights in ELM neural network. This refined GSO with ELM enables the selection of input weights to increase the generalization performance and the conditioning of the single layer feedforward neural network. The steps of the proposed approach are as follows.

Step 1. Initialize positions and head angles with a set of input weights and hidden biases: . These will be randomly initialized within the range of on dimensions in the search space.

Step 2. For each member in the group, the respective output final weights are computed at ELM as given in (9).

Step 3. Now invoke refined GSO as in Table 1.

Step 4. Then the fitness, which is the mean square error (MSE) of each member, is evaluated as given below: where is the number of training samples and the terms and are the error of the actual output and target output of the th output neuron of th sample. Thus, fitness function is defined by the MSE. In order to avoid overfitting of the single layer feedforward neural network, the fitness of each member is adopted as the mean squared error (MSE) on the validation set only instead of the whole training set as in [28].

Step 5. Find the producer of the group based on the fitness.

Step 6. Update the position of each member as given in (13) to (17).

Step 7. Stopping criteria: the algorithm repeats Steps 26 until certain criteria are met, along with hard threshold value as maximum number of iterations. On reaching the stopping criteria, the algorithm returns the optimal weights with minimal MSE as its solution.

Thus refined GSO (RGSO) with ELM finds the best optimal weights and bias so that the fitness reaches the minimum to achieve better generalization performance, with minimum number of hidden neurons, considering both the advantages of both ELM and RGSO. In the process of selecting the input weights, the refined GSO considers not only the MSE on validation set but also the norm of the output weights [29]. The proposed RGSO based ELM will combine the feature of RGSO into ELM to compute the optimal weights and bias to make the MSE minimal.

4. Experimental Results and Discussion

4.1. Clinical Datasets

The details of tumor benchmark dataset, normal healthy benchmark dataset, and real time clinical data from hospitals used in the simulation are given in Table 2. The clinical specimen utilized in the present study consists of brain MR-images of 70 clinical routine cases with verified and untreated intracranial tumors. Each image sequence with a 3 mm brain slice-interval (i.e., the voxel size was 0.4492 mm × 0.4492 mm × 33 mm) in axial plane was measured. Figure 4 shows the 10 SPL dataset.

The simulation tool employed for implementing the proposed work is MATLAB (Version 7.11) and was executed in computer with Intel core i5 processor with 2.53 GHz speed and 3 GB of RAM.

4.2. Volumetric Feature Analysis of Datasets

In medical applications, the small homogeneous tissue region is selected manually and is referred to as region of interest (ROI). In the 3D texture analysis, the same homogeneous region was maintained over ten additional neighboring slices from the reference slice (five up and five below) to form a cubic volume of interest (VOI) of about 1330 voxels in each class and in each subject. Four volume(s) of interest (VOI) of different brain tissues and tumor parts were identified for each patient by an experienced neuroradiologist: (i) solid (active tumor), (ii) white matter (WM), (iii) gray matter (GM), and (iv) cerebrospinal fluid (CSF) [30].

Homogeneous VOIs (400 to 1600 voxels) were carefully selected avoiding signals from adjacent tissues. A normalization process was carried out based on histogram and ROI/VOI. It delineates training areas for all classes over each VOI reducing the initial image’s gray levels number to 128 in order to shorten calculation time and to avoid sparse matrices. 5 parameters of cooccurrence matrix, 11 parameters of run-length matrix, and 2 parameters of Sobel and Laplace gradient along with its orientations were calculated for 3D methods, with five distance = 1, 2, 3, 4, and 5. The 3D feature extraction method from GLCM, RLM, and gradient model considers the 26 neighbors of a voxel. Tables 3 and 4 show the 3D VOI GLCM texture measures for real time clinical datasets and SPL datasets, respectively.

The relevant features are identified using area under the curve (AUC) value from receiver operating characteristics (ROC) curve and values from 2-tailed Student’s -test. Statistical analysis is evaluated based on and AUC values, where the best features are ranked. Table 5 shows the relevant features as indicated by high value of AUC and low value for 3D VOI features.

4.3. Feature Subselection Using DE

The rich aggregate bank of features of the tissue image was calculated by using the gray-level run-length matrix method, GLCM, gradients, and intensity. In the proposed system, twenty-seven parametric features were extracted, with a total of 1755 features for each VOI. The possible outcome of the employing DE is to provide the optimal set of features which can be used as input to the classifier. These parameters were selected using differential evolution which provides excellent selection capabilities for any dataset. So out of the 18 features, an optimal 7 decisive features are selected and are taken for the next step to classification process. Table 6 shows the optimal subselected features using differential evolution approach.

From Table 6, the maximum and minimum values of subselected feature parameters for healthy and tumorous brains can be noted. Figures 5 and 6 show the MRI brain case 2 SPL dataset and its volumetric feature extraction, respectively.

4.4. Brain Tissue and Tumor Classification Using ELM with RGSO

Hybrid ELM-RGSO is now invoked after the subselection process for efficient and accurate tumor classification. The details of the simulation and the performance of the classifier for the dataset are presented in this section. The weights and bias to the input layer and hidden layer of the ELM architecture are optimized using the proposed RGSO. The inputs () are 7 and the hidden neurons () are 5, for which the best optimal weight parameters are selected employing RGSO. The number of outputs specifies each tissue, that is, cerebrospinal fluid (CSF), white matter (WM), gray matter (GM), and tumor. Consequently, the structure of the single SLFN ELM neural network is 7-5-4.

Table 7 compares the ELM-RGSO simulation results with existing tested classifiers. The ELM-RGSO classification performance is appreciably good for 7 dimensional spaces. Thus, extreme learning machine outperforms the standard binary linear SVM and BPN for medical image classifier and proves better classification of healthy and tumor tissues. The comparison proves that the mean and standard deviation produced by volumetric feature extraction analysis is higher than the other approaches: connected threshold (CT), neighborhood connected (NC), and confidence connected (CC). The results are almost similar to the approach in [31] (original SPL data analysis). But it can be viewed that a significant difference exists in the computation time amongst the approaches.

The validation on the above tumor classification is evaluated quantitatively by calculation of the similarity index (SI) or dice coefficient between the automatic and the manual classifications. Consider where with is a segmented volume and is the overlap of and . The SI is also used as a measure for the interobserver variability. The true positive fraction (TPF), or sensitivity, and the specificity or false positive fractions (FPF) are also used for evaluation. Table 8 depicts the average value of the TPF, FPF, and SI of various classifiers. The proposed method showed high accuracy for all tissue classes and the SIs were close to the interobserver SI of the manual segmentations when compared to the state-of-the-art model using kNN classifier as given in [5]. However, for tissue types with less overlap, the SI measure shows a better distinction between the segmentation methods.

The ELM-RGSO classifier showed the highest overlap with the manual segmentation for all tissues. Table 9 depicts the average value based on the tumor segmentation methods based on feature extraction model and compared to the state-of-the-art techniques [31]. The proposed model using 3D feature extraction proves better and improves values on sensitivity and specificity for the 10 cases of SPL dataset. The algorithm segmented tumor and expert segmented tumor are shown in Figures 7 and 8.

4.5. Classification Analysis

The aim of this work is to classify normal and tumor images correctly. In the analyses of the images, each image is classified into one of two prognostic classes (normal and tumor image). Based on the features that have been extracted, the classifier separates these features. Combining both the publically available datasets (SPL and McConnell Brain Centre images), the dataset of 30 patients with texture measurements are well separated for the two classes; that is, the patients were carefully selected with respect to the average texture value of the measures.

Table 10 depicts the classification accuracy between various classifiers with proposed classification model. It is evident that images with low texture values are patients with good prognosis (normal brain) and vice versa. The images were then classified by ELM-RGSO classifier with leave-one-out cross-validation; that is, the classifier was trained with 19 patients and then the one patient not used in training was classified. This is rotated in such a way that all patients are used as a test set. In the above, the samples chosen are 70 cases of real time data and 30 cases of SPL and MBC set. The samples are converted into their respective pixel values and are inputted to the ELM model. All the samples were loaded sequentially; hence the delay process was within the permissible extent.

The proposed ELM-RGSO classifier, which classifies the patients to the class which is most probable, will result in one patient data being wrongly classified, leading to a correct classification rate for the 30 patients of 98.29%. Table 11 tabulates comparison of ELM, IELM-PSO classifier, and ELM-RGSO classifier.

5. Conclusion

The major issue in the development of pattern recognition towards brain tumor tissue classification is the formation of feature extraction analogy and the classifier. In this work, a refined GSO based extreme learning machine classifier is proposed to develop the brain tumor tissue classification along with the DE as feature subselector. Simulation result shows that the proposed volumetric feature extraction method works well and produces a minimum number of features with higher classification accuracy when compared with conventional 2D and 3D feature extraction methods.

The developed hybrid GSO and ELM [23] along with DE as feature subselector shows highest improvement in comparison with other literature studies in neural networks. On the whole, the significant finding of this work employing differential evolution, ELM, refined GSO, and volumetric analysis validates the correlation of magnetic resonance image measures as well as anatomical and histopathological parameters of research. Thus, a hybrid GSO and ELM classifier is developed in this research work for brain tumor tissue categorization in 3D MR images to reduce the computational cost and time prevailed in all earlier classifiers.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.


The authors thank PSG IMSR and Hospitals for providing image data for carrying out this research work.