BioMed Research International

BioMed Research International / 2019 / Article

Research Article | Open Access

Volume 2019 |Article ID 5173589 | https://doi.org/10.1155/2019/5173589

Roozbeh Zarei, Jing He, Siuly Siuly, Guangyan Huang, Yanchun Zhang, "Exploring Douglas-Peucker Algorithm in the Detection of Epileptic Seizure from Multicategory EEG Signals", BioMed Research International, vol. 2019, Article ID 5173589, 19 pages, 2019. https://doi.org/10.1155/2019/5173589

Exploring Douglas-Peucker Algorithm in the Detection of Epileptic Seizure from Multicategory EEG Signals

Academic Editor: Andrei Surguchov
Received24 Mar 2019
Accepted16 Jun 2019
Published07 Jul 2019

Abstract

Discovering the concealed patterns of Electroencephalogram (EEG) signals is a crucial part in efficient detection of epileptic seizures. This study develops a new scheme based on Douglas-Peucker algorithm (DP) and principal component analysis (PCA) for extraction of representative and discriminatory information from epileptic EEG data. As the multichannel EEG signals are highly correlated and are in large volumes, the DP algorithm is applied to extract the most representative samples from EEG data. The PCA is utilised to produce uncorrelated variables and to reduce the dimensionality of the DP samples for better recognition. To verify the robustness of the proposed method, four machine learning techniques, random forest classifier (RF), k-nearest neighbour algorithm (k-NN), support vector machine (SVM), and decision tree classifier (DT), are employed on the obtained features. Furthermore, we assess the performance of the proposed methods by comparing it with some recently reported algorithms. The experimental results show that the DP technique effectively extracts the representative samples from EEG signals compressing up to over 47% sample points of EEG signals. The results also indicate that the proposed feature method with the RF classifier achieves the best performance and yields 99.85% of the overall classification accuracy (OCA). The proposed method outperforms the most recently reported methods in terms of OCA in the same epileptic EEG database.

1. Introduction

Epilepsy is one of the most common neurological disorders of the human brain that affects approximately 65 million people of the world [1]. It is characterised by unprovoked recurring seizures which are induced by abnormal and synchronous discharges of a group of neurons in the brain [2]. Although numerous molecular mechanisms underlying different forms of epilepsy have been identified, the etiology of majority of them cannot be explained by a simple defect altering ionic homeostasis [3]. Transient and unexpected electrical disturbances of the brain are recognised as the possible causatives for epileptic seizures. In the majority of cases, seizures occur unexpectedly, without a sign of warning to alert and prepare the person for an onset of a seizure. Such abrupt and uncontrollable nature of the disease can cause physical injury due to loss of motor control, loss of consciousness, or delayed reactivity during seizures. Impairment of consciousness can be life-threatening, especially if they occur while the person is driving, swimming, climbing heights, or alone. Electroencephalogram (EEG) is most commonly used technique for diagnosis of epileptic seizure in the medical community [4]. EEG record electrical activity along the scalp, via the placement on the scalp of multiple electrodes; it measures voltage fluctuations resulting from ionic current flows within the brain [5, 6]. Epileptic activity can create clear abnormalities on a standard EEG and leaves its signature on it [7]. Epileptic seizure activities in the brain commonly manifest spikes or spike wave complexes in EEG signals which are usually analysed visually by expert or neurologists [8, 9]. However, the visual scanning of EEG signal is very time-consuming and costly; it may be inaccurate, very complex, subject to judgement, and human error [10] as EEG signals contain a huge amount of data (in size and dimension). Therefore, there is an increasing need for developing automated epileptic seizure detection algorithms not only to alleviate the neurologist’s burden of analysing long-term EEG signals but also to ensure a proper diagnosis and evaluation of neurological diseases.

In past two decades, several EEG signal processing techniques have been developed for automated epileptic seizure detection based on various feature extraction and classification techniques. The key challenge of any detection method is the extraction of the distinguishing features from EEG signals as it significantly affects the performance of the classifier. Representative characteristics or features extracted from EEG data can describe the key properties or morphologies of the signals for perfect detection of epileptic seizure [11]. As feature extraction is the most important part of detection process which plays key role in the performance of a classifier, this study aims to develop a new efficient feature extraction technique for the classification of epileptic seizure from EEG signals.

Several feature extraction methods have been applied in epileptic seizure detection, such as correlation [12], linear prediction error energy [13], fast Fourier transform (FFT) [14], wavelet transform [1517], empirical mode decomposition (EMD) [18, 19], Lyapunov exponent [20], Correlation dimension [21], approximate entropy (ApEn) [22, 23], clustering technique [24], Sampling technique [10, 25], Complex network [6, 26, 27], and Optimum allocation [7, 28]. These feature extraction techniques can be grouped into four categories [29], namely, time-domain [12, 13], frequency domain [14], time-frequency domain [15, 16, 18, 19], and nonlinear methods [2022]. Once features are extracted from EEG signals, a classifier is employed to differentiate between normal and epileptic EEG. Many classification methods have been proposed for seizure detection such as various types of artificial neural networks (ANNs) [3032], support vector machine (SVM) [5, 12, 33, 34], Decision tree (DT) [35], k-nearest neighbour [36], and Random Forests (RF) [37].

Due to complex characteristics of EEG signals (e.g., nonstationary, aperiodic, and poor signal-to-noise ratio), sometimes it is very hard to achieve reasonable performance in the detection of epileptic seizure. Some of the existing feature extraction methods are not a good choice for obtaining characteristic features from nonstationary epileptic EEG data (e.g., Fourier transformation) [14, 38], and therefore most of their performances are limited regarding success rate and effectiveness [39, 40]. Moreover, the majority of the existing methods cannot appropriately handle large EEG data. Although most of the EEG recordings are multicategories in a real clinical application, most of the current methods are applied for binary EEG classification problems (Normal signal vs. ictal signal) [32, 4145] and only a few methods focus on multiclass EEG classification [37, 39, 40, 4648]. Considering these issues, this paper proposes a new feature extraction technique based on Douglas-Peucker algorithm (DP) and principal component analysis (PCA) for classification of multiclass EEG signals.

The DP [49] is the most well-known line simplification algorithm which is widely used in cartographic and computer graphic applications to reduce the complexity and storage requirements of curves by removing curve’s no-characteristic points and extracting characteristic points [5053]. It is also applied in biomedical applications such as Electroencephalogram (ECG) signals compression [5456]. The main theme of this algorithm is to shorten a line by detecting and preserving the most significant points of a line while neglecting less important points. Although the DP technique has a high capability to represent the original patterns of time series data and reduce the size of data, it has not been considered before for epileptic detection in the EEG signal analysis to the best of author’s knowledge. Thus this study introduces for the first time the idea of using the DP methods for extracting representative sampling points from huge amount of raw EEG data.

The main aim of this research is to develop a novel feature extraction technique for detection of epileptic seizure from multicategory EEG signal for properly handling big size EEG data. Moreover, this paper investigates the effectiveness of DP algorithm in the detection of epileptic seizure from EEG data and also discovers an effective classifier for the proposed features. In the proposed methodology, first the nonstationary epileptic EEG signals are partitioned into some nonoverlapping segments (called Segm) to make them stationary (discussed in detail in Section 3.1.1). Then the DP algorithm is effectively employed to extract representative sampling points from each Segm and also to reduce the size of each Segm by removing redundant points. At the next stage, the PCA is used to reduce the dimensionality of DP data and also to produce uncorrelated variables which are considered as features, denoted as DP_PCA feature set. In order to select an efficient classifier for DP_PCA feature set, this study employs four popular machine learning techniques namely, RF, k-nearest neighbour algorithm (k-NN), SVM, and DT on the extracted features. To evaluate the consistency and performance of the proposed methods, tenfold cross-validation is applied to create training and testing set. The performance of each method is evaluated by sensitivity (Se), specificity (Sp), overall classification accuracy (OCA), false positive rate (FPR), kappa statistic, and receiver operating characteristic (ROC) curve area. In order to further evaluate the performances, the proposed method is compared with other six existing algorithms. The experiment results show that the RF classifier is the best classifier for DP_PCA feature set compared to other three classifiers. The results also indicate that the proposed method outperforms the existing methods [37, 39, 40, 4648] regarding Se, Sp, and OCA.

The rest of the paper is organized as follows: in Section 2, we describe the prior studies in multiclass EEG signals classification. Section 3 presents the methodology of the proposed method. Section 3 also describes the experimental data and implementation. Section 4 discusses the experimental results and discussions. Finally, Section 5 draws the conclusion for this paper.

2. Previous Work

In the last decade, various methods have been proposed for the classification of EEG signals [1, 2, 8, 9, 1216, 22, 29, 36, 5760]. However, only a few approaches have dealt with multiclass EEG classification problems [37, 39, 40, 4648]. For comparative reasons, the most recent and relevant studies dealing with multiclass EEG classification problems on a benchmark epileptic EEG dataset [61, 62] are reviewed.

Most recently, Emigdio et al. [37] developed a method based on Holderian regularity and the Matching Pursuit (MP) algorithm for feature extraction in the epileptic EEG signal classification. The feature sets were constructed by combining features extracted from EEG signals through regularity analysis, the MP algorithm and simple time-domain statistical analysis. These feature sets were then fed to a Random Forests classifier for classification of epileptic states. The performance of the method was tested on the Bonn data set [61, 62] considering different classification problems (binary classification problems and multiclass classification problems). The results showed that the overall classification accuracy was 97.6% for the five-class classification problem.

Murugavel and Ramakrishnan [48] introduced an approach based on a hierarchical multiclass SVM (H-MSVM) with extreme learning machine (ELM) as the kernel for the classification of epileptic EEG signals. The wavelet transform was used to decomposed the EEG data into six subbands and then six features such as largest Lyapunov exponent, statistical values, and approximate entropy were extracted from each subband. The extracted features were employed as the input to the classifier. The artificial neural network (ANN) and multiclass SVM were also utilised to identify the five-category EEG signals. The experimental results showed that the H-MSVM classifier with ELM kernel yielded a better performance regarding classification accuracy and computation complexity compared to the ANN and SVM classifiers. The H-MSVM achieved an overall classification accuracy of 94%.

Ubeyli [47] reported a method based on Lyapunov exponents and a probabilistic neural network (PNN) classifier for classification of EEG signals. The Lyapunov exponents were obtained from each EEG signal using Jacobi-based algorithms and considered as feature vectors. The statistic over the Lyapunov exponents was used to reduce the dimensionality of the extracted feature vectors. The selected features were fed to the PNN and multilayer perceptron neural network (MLPNN) classifiers. The classification results show that the PNN with Lyapunov exponents features achieved an overall classification accuracy of 98.05% while the MLPNN produced a 92.20% accuracy rate.

Ubeyli [46] presented automated diagnostic systems combined with spectral analysis techniques for classification of EEG signals. Eigenvector methods were used to calculate the wavelet coefficients and power spectral density (PSD) values which considered as features. The selected features then were fed to seven classification algorithms: SVM, PNN, mixture of experts (ME), modified mixture of experts (MME), recurrent neural networks (RNN), MLPNN, and combined neural networks (CNN). The experimental results showed that the SVM and MME classifiers achieved better performance compared to other five classifiers. The classification accuracy for the SVM, MME, PNN, ME, RNN, CNN and MLPNN classifiers with the obtained features were 99.20%, 98.68%, 95.30%, 95%, 94.85%, 93.48%, and 90.48%, respectively.

Ubeyli [39] developed a method based on multiclass SVMs with the error correcting output codes (ECOC) and eigenvector methods for the classification of EEG signals. The PSD values of the EEG signals were obtained using three different eigenvector methods such as the MUSIC [63], Pisarenko [64], and minimum-norm [65]. The statistics over the set of the power levels of the PSDs were considered as features and fed to the multiclass SVMs. The MLPNN classifier was also applied to the same feature set. The total classification accuracy obtained by SVM with the ECOC and the MLPNN was 99.30% and 92.90%, respectively.

Guler and Ubeyli [40] proposed the multiclass SVM with the ECOC for the classification of multiclass EEG signals. They also tested the probabilistic neural network (PNN) and multilayer perceptron neural network (MLPNN) classifiers on the same epileptic EEG data. The wavelet coefficients and Lyapunov exponents were used to extract features from the EEG data. The extracted features were employed as the input of the three classifiers. The results showed that the multiclass SVM classifier achieved better performance than the other two classifiers. The total classification accuracy for the SVM, PNN, and MLPNN was 99.28%, 98.05%, and 93.63%, respectively.

3. Methods and Materials

3.1. Proposed Approach

The paper introduces a novel method based on DP in the multiclass EEG signal classification. In this study, the DP approach is developed to select representative samples from the original EEG signals that reflect an entire database. Next, The PCA is used to reduce the dimension of the obtained DP sample set which is considered as a feature set. Finally, the extracted features are tested by four machine learning methods, including RF, k-NN, SVM, and DT. As shown in Figure 1, the entire process of proposed method is divided into five major parts: data segmentation, Douglas-Peucker algorithm, dimension reduction by PCA, DP_PCA feature set, and the classification part by the RF, k-NN, DT, and SVM. A detailed description of these five parts is provided in the following sections.

3.1.1. Data Segmentation

Most of the EEG signal processing methods require stationarity of the signals. Although EEG signal may not be stationary, usually smaller windows or parts of those signals will exhibit stationarity [7]. An EEG signal is stationary for a small amount of time. That is the reason the recorded EEG signals of every class are split into several nonoverlapping segments based on a particular time period to properly account for possible stationarities. Hence the EEG signals of each class are segmented into some fixed-size nonoverlapping time windows (called ‘Segm’) to obtain representative values of a specific time period. Each Segm consists of EEG channel data within a time window. Figure 2 illustrates an example of determining the segments Segms in an EEG signal of a class. It is worthwhile to mention that the number of Segms (k) is determined empirically over time for any experiment design.

3.1.2. Douglas-Peucker Algorithm

The DP [49] is one of the most popular methods for line (trajectory) simplification. The algorithm simplifies a line by detecting and preserving the most significant points of a line while neglecting less important points. In this study, the DP technique is used to extract the representative samples from different ‘Segms’. Let the data series (trajectory) S be described by the set of N points . The main idea of DP algorithm is to determine a new data series with fewer and most significant points without deviating from the original data series by at most a simplification tolerance . As an initial step of DP, the algorithm approximates the data series S with a line segment constructed from the first to the last data point. Then it calculates the perpendicular Euclidean distance between each intermediate data point and the line segment and retains the point which has the maximum distance . The algorithm compares with the given simplification tolerance . If the maximum distance is less than the simplification tolerance , the algorithm removes all intermediate points in data series. Otherwise, it uses data point to split the data series to two subseries and and recursively repeats the procedure for each subseries. The DP algorithm terminates when the in a subseries is lower than the simplification tolerance or the subseries contains only two data points. Figure 3 illustrates an example of DP sample point extraction. The original data series contain eight points (). The distances from the points to the line segment are first computed. Since the maximum distance at point exceeds the given simplification tolerance , the data series are divided at this point into two subseries (step 2 in Figure 3). In the left subseries, the distance from to the line segment is lower than value, so the point is ignored. In the right subseries, the distance from the point to the line segment also exceeds the simplification tolerance , hence a new split is performed at the point , and the process is repeated for each part, respectively. Figure 3 shows that the original data series having 8 points finally becomes a 4-points data series after this process.

The value of simplification tolerance determines the degree of simplification. Therefore, it is an important task in DP algorithm to determine the most significant value. Choosing a small value of will produce a minimally simplified data series (i.e., only a few redundant data points will be removed from data series) while selecting a large one will provide a highly simplified data series that might lead to losing some of the significant points from the data series. The following formula (1) is used to calculate the most significant value for each data series.

where N is the number of points in data series; is the Euclidean distance between two points and ; is the overall distance of the data series; and T is a real number which is determined empirically. The value of is changed as different percentages of the overall distance of the data series by setting different values for T.

As shown in Figure 1, the DP process consists of the following steps to extract the representative samples from various Segms.

Step 1. Consider all the channels of the EEG data of a class.

Step 2. The EEG data of that class is split into k Segms considering a particular time period. Suppose the sizes of the Segms are , respectively.

Step 3. The overall distance of each Segm is calculated. Then by setting the T value, the value of simplification tolerance for each Segm is calculated using (1).

Step 4. The representative samples from each Segm are extracted using DP algorithm. Let be the sizes of samples obtained from the Segms whose sizes are , respectively. The representative samples selected from each Segm in a class make a vector set denoted as DP_Sample as shown in Figure 1.

Step 5. The vector sets of all classes construct a matrix (denoted as DP_samples set) that is used as input to the PCA, as discussed in the next section.

3.1.3. Dimension Reduction by PCA

The PCA is a well-known statistical method for feature extraction and dimensionality reduction [6668]. It uses an orthogonal transformation to convert a set of observations of possibly correlated variables into a smaller set of uncorrelated variables called principal components (PC). These components represent the most important linear characteristics of the data. The multichannel EEG signals recorded from different scalp sites are highly correlated. They contain a large amount of redundant information. Therefore, it would be useful to remove this redundant information by converting the EEG signals into a set of new linearly uncorrelated variables (i.e., the PC space) and utilise these new variables as features for better classification of EEG signals. In this paper, the PCA is used to reduce the dimensionality of the DP_Samples set and also to obtain EEG features for classification of epileptic EEG signals.

Let where each row represents a data point in a p-dimensional space (considering a p-channel EEG signals as a p-dimensional space) and n is the number of the points selected by DP. PCA can be formulated as the following optimisation problem:

where U is a matrix consisting of q dominant eigenvectors. This problem can be solved by deriving an eigenvalue decomposition problem of the covariance matrix.

where

is the covariance matrix, is the global mean defined as , are the eigenvalues and they are sorted in descending order, and are the corresponding eigenvectors. In order to reduce the dimensionality of the DP_Samples set, only the first q eigenvector () which corresponds to the q largest eigenvalues is selected by the following equation to represent the DP_Samples set.

For a given precision parameter (considering =90% in this study), the matrix  U consisting of q dominant eigenvectors is constructed and the q-dimensional feature set denoted as DP_PCA feature set is computed as follows:

3.1.4. DP_PCA Feature Set

The new feature set donated as DP_PCA feature set is generated by reducing the dimensionality of the DP_Samples set using PCA method as discussed in Section 3.1.3. This feature vector set is divided into a training set and a testing set using a tenfold cross-validation method, which is discussed in Section 3.3. As shown in Figure 1, this feature set is fed to each of the four classifiers discussed in the next section.

3.1.5. Classification by the RF, KNN, SVM, and DT

This study considers four classifiers: RF, k-NN, DT, and SVM for testing the performance of the proposed feature extraction method. A brief explanation of these classification methods is provided in the following sections.

Random Forest. The RF is an ensemble learning technique developed by Breiman [69]. It consists of many individual classification trees, where each tree is constructed using a tree classification by selecting a random subset of input features and a different bootstrap sample from the training data. The RF aggregates the results of all classification trees to classify new samples. Each tree casts a unit vote at the input data and then the forest selects the class with the most votes for the input data. Figure 4 illustrates the structure of random forest classifier.

The RF algorithm proceeds as follows:(1)From the training data set, m training subsets are generated using the bootstrapping technique (randomly sampling with replacement). Each training subset has the same size as the training data set and contains approximately one-third of the samples of the training data set.(2)For each training subsets, a decision tree is built with the following criteria: at each node in building a decision tree, a random number of f features are selected from the F input features () and the best split (e.g., the largest Gini measure) among these f features is used to divide the node. The tree is grown to the maximum size with no pruning. The tree growing algorithm used in RF is random trees.(3)The m trees are combined into an RF ensemble and use a majority voting scheme to predict the class of new data by evaluating votes from each tree.

K-Nearest Neighbour Algorithm. The k-NN is a supervised learning algorithm for classifying objects based on closest training observations in the feature space [66, 70]. Although the k-NN is the simplest algorithm among all machine learning algorithms, it can still yield high performance, without a priori assumptions about the distributions from which the training samples are drawn [66]. Given a query vector and a set of N labelled instances , the aim of the classifier is to identify the class label of on the predefined P classes. The k-NN classification algorithm tries to find the k-nearest neighbours of , and uses a majority vote to determine the class label of . Without prior knowledge, the k-NN classifier usually applies Euclidean distances as the distance metric [71]. A detailed discussion of this method can be found in [66, 70, 72].

Support Vector Machine. The SVM is a machine learning algorithm based on statistical learning theory and structural risk minimisation principle presented by Vapnik [73]. The main idea of SVM is to map the input data into a higher dimensional space and then determines an optimal separating hyperplane between the two classes of data in the transformed space [74, 75]. For nonlinear classifier models, when the data are not linearly separable, SVMs map inseparable input data into a high-dimensional space by constructing a linear kernel function to make the input data linearly separable in new space and allows better fitting of the hyperplane to the input dataset. Although the SVM is originally designed as a two-class classifier, some methods have been proposed to extend the application of SVM to multiclass classifications. One common used procedure in practice is to employ a set of pairwise classifiers, based on one-against-one decomposition [75]. The decision function of binary SVM classifier can be expressed as follows:

where sgn is the signum function, is kernel function, and b is the bias of the training samples. There are several kernel functions such as linear kernel, polynomial kernel, RBF kernel, and sigmoid kernel. In this paper, the polynomial kernel is considered as the best kernel function for identifying multicategories EEG signals as it was found to give the best classification performance.

The regularisation parameter C is used to control the trade-off between training error and model complexity and can be calculated as follows:where N is the size of the training set.

In the multiclass classification problem, the SVMs work by using a collection of decision functions . The class decision can be obtained by the following formula [75]:

where kl indicates each pair of classes selected from separated target classes and n is the number of separated target classes. The algorithm proceeds as follows: it assigns a label to the class: arg max , (k = 1, 2,..., n). The pairwise classification then converts the n-class classification problem into n(n1)/2 two-class problems which cover all pairs of classes. An overview of SVM classifier can be found in [7375].

J48 Decision Tree. J48 decision tree is an implementation of the C4.5 algorithm [76] in the WEKA (The Waikato Environment for Knowledge Analysis) [77, 78]. C4.5 is an extension of the ID3 algorithm. It uses the top-down construction technique to recursively split the data set into smaller subsets based on the value of an attribute [76, 79]. This classifier builds a decision tree for the given dataset using the concept of information entropy. In a decision tree, each attribute can be used to make a decision by splitting the data into smaller subsets. At each node of the tree, the algorithm evaluates each attribute of the data for dividing the data into smaller subsets and chooses the attribute that gives the highest information gain. Once an attribute is selected, the data set is split into subsets, and the splitting process is repeated for each subset until further splitting is not gainful. In the resulting tree structure; each inner node in the tree corresponds to one of the input attributes, each branch represents a value or range of values of that attribute, and each leaf accounts for a classification.

3.2. The Epileptic EEG Data

The epileptic EEG data used in this work is obtained from publicly available EEG database of Department of Epileptology, University of Bonn, Germany [61, 62]. The whole database contains five subsets denoted as Sets A, B, C, D, and E. Each subset is containing 100 single-channel EEG signals with a duration of 23.6 s. The subsets A and B are recorded extracranially, whereas subsets C, D, and E are recorded intracranially. Set A and Set B were collected from five healthy volunteers with eyes open and eyes closed, respectively. Sets C and D were collected from five epileptic patients during interictal periods. Set C was recorded from the hippocampal formation on the opposite side of the epileptogenic zone while Set D was recorded from the epileptogenic zone. Set E was collected from all of the recording zones in Sets C and D during seizure activity (ictal periods). All EEG recordings were recorded using a 128-channel amplifier system with a sampling rate of 173.61 Hz and 12-bit A/D resolution. Signals were filtered using a 0.53–40 Hz (12 dB/octave) band pass filter, and artifacts such as muscle and eye movements were removed by visual inspection. A summary description of the five set EEG data is provided in Table 1. Exemplary EEG time series from each of the five classes (Set A-Set E) are shown in Figure 5


Set ASet BSet CSet DSet E

SubjectsFive healthy subjectsFive healthy subjectsFive epileptic subjectsFive epileptic subjectsFive epileptic subjects
Patient’s stateAwake and eyes open (normal)Awake and eyes closed (normal)Seizure-free (interictal)Seizure-free (interictal)Seizure activity (ictal)
Electrode typeSurfaceSurfaceIntracranialIntracranialIntracranial
Electrode placementInerenational 10-20 systemInerenational 10-20 systemEpileptogenic zoneEpileptogenic zoneEpileptogenic zone
Number of channels100100100100100
Time duration (s)23.623.623.623.623.6

3.3. Implementation

This section presents the implementation of the proposed method on the epileptic EEG data [61, 62]. As discussed in Section 3.2, the complete dataset contains five sets (denoted as A, B, C, D, and E), each containing 100 channels data of 23.6 s. Each channel consists of 4096 data samples. The implementation of the proposed method comprises five steps as follow:(1)Each class data is segmented into four Segms (K=4), each containing 100 channels data of 5.9 s. As each channel consists of 4096 data samples, the sizes of the four Segms are =1024, =1024, =1024, and =1025, respectively.(2)To determine the value of simplification tolerance , the overall distance of each of the four Segms in each class are calculated, and then the value of T is changed from 0.01 to 0.1 with step size 0.01 in (1) to identify the most significant value for each Segm. From the experiment, it is considered that T=0.06, and then the value of for each Segm is calculated using (1). Table 2 presents the obtained value of for each Segm in each of the five classes. From Table 2, it is observed that the values of are not equal due to the differences in the overall distance of Segms.(3)Using the obtained value of shown in Table 2, the representative samples from each Segm are extracted using the DP algorithm. Figures 6 and 7 show typical results of DP for the healthy subject (class A) and the epileptic patient (class E), respectively. In Figures 6 and 7, the first Segm of class A and class E is considered, respectively. It can be seen from Figures 6 and 7 that the DP samples can effectively represent the original signals with fewer points which indicate the ability of DP to select most significant points from each signal.  Table 3 provides the number of the representative samples chosen by DP for each Segm in each of the five classes. It can be seen from Table 3 that the number of the representative samples for each Segm is not equal; e.g., in Set A (Class 1), the total number of 581, 612, 599, and 562 samples is selected by DP from Segm 1, Segm 2, Segm 3, and Segm 4, respectively. The total number of the representative samples is 2354 for Set A.The representative samples selected from all Segms in a class create a vector set denoted as DP_Sample as shown in Figure 1. For example, the selected representative samples from each of the four Segms of class 1 create a DP_Sample 1 vector. The DP_Sample 1 is constructed as 581, 612, 599, and 562 which contains all 2354 selected samples in class 1. The vector sets are created similarly: DP_Sample 2, DP_Sample 3, DP_Sample 4, DP_Sample 5 from class 2, class 3, class 4, and class 5, respectively. All DP_Samples from the five-class EEG data construct a matrix denoted as DP_sample set that is used as an input to the PCA. The DP_sample set contains all 10755 selected samples from five classes (2354 for class 1, 2008 for class 2, 2237 for class 3, 2398 for class 4, 1758 for class 5). It can be seen from Table 3 that the DP reduces the data samples size of the five class from 20480 sample points to 10755 sample point (47.49% sample reduction). Here note that each sample has 100 dimensions as each class contains 100 channels of EEG data. Therefore, the DP_sample set consists of 10755 samples of 100 dimensions.(4)The PCA is applied to reduce the dimensionality of the DP_ Sample set. Only the first q eigenvectors are selected to represent the DP_sample set based on the accumulation of their respective eigenvalues exceed 90% of total sum of eigenvalues (see (5)). Figure 8 illustrates the cumulative eigenvalues for all 100 eigenvectors. It is observed that the accumulation of the first 53 eigenvalues exceeds 90% of total sum of eigenvalues. Therefore, only the first 53 eigenvectors are considered for obtaining the DP_PCA feature set. The obtained DP_PCA feature set contains 10755 samples of 53 dimensions.(5)The DP_PCA feature set is divided into a training set and a testing set using a 10-fold cross-validation method to evaluate the performances of the proposed methods. The DP_PCA feature set is split into ten mutually exclusive subsets (10-folds) of approximately equal sizes. Training and testing are performed ten times. Each time, one of the folds is used as a testing set and the remaining nine folds are combined into a set for training.


Segm 1Segm 2Segm 3Segm 4

Set A96.8084.4588.3197.38
Set B137139.87179.87167.87
Set C66.3799.7762.7568.71
Set D52.1475.9663.5354.37
Set E908.46821.18747.53727.49


Sample size Segm 1Segm 2Segm 3Segm 4Total

Set A581 (1024)612 (1024)599 (1024)562 (1025)2354 (4096)
Set B564 (1024)583 (1024)429 (1024)432 (1025)2008 (4096)
Set C589 (1024)408 (1024)622 (1024)618 (1025)2237 (4096)
Set D669 (1024)489 (1024)532 (1024)708 (1025)2398 (4096)
Set E421 (1024)438 (1024)450 (1024)449 (1025)1758 (4096)

Note: in each cell, the number inside the parentheses is the Segm size (e.g., (1024) in the 1st cell) and the number outside of the parentheses is the sample size (e.g., 581 in the 1st cell) obtained by the DP.

In this research, the performances of the proposed methods are evaluated based on different statistical measures, such as Se, OCA, FPR, kappa statistic, and ROC curve area. Their formulas are given below:

where denotes the overall observed agreement between the classifier and the true classes, and represents the expected proportion of agreement. Besides, the Area Under the ROC Curve (AUC) is measured to compare the overall performance of the classifiers. The ROC curve is obtained by plotting the sensitivity versus false positive rates [47].

4. Results and Discussions

This section presents the experimental results of the proposed methods on the epileptic EEG datasets. First, the effectiveness of each of the four mentioned classifiers is evaluated on the DP_PCA feature set to select the most appropriate classifier as discussed in Section 4.1. Then a comparison between the proposed method and six existing methods is provided in Section 4.2. All mathematical calculations are carried out in MATLAB R signal processing tool (version 7.11, R2010b). The classification executions for all four classifiers: RF, k-NN, SVM, and J48 classifiers are executed in WEKA machine learning toolkit [77, 78]. The LIBSVM tools (version 3.2) [80] is used in WEKA for the SVM classification. It is worth mentioning that the default parameter values for each classifier in WEKA are used as there are no specific guidelines for selecting these parameters.

4.1. Classification Results for Each Classifier

As mentioned before, four machine learning methods, RF, k-NN, SVM, and J48 classifiers, are tested for detection of the multiclass EEG signals. Table 4 presents the classification results of all classifiers on the DP_PCA feature set. The performance results are given by averaging over the results of the 10-fold cross-validation test and expressed as the mean standard deviation. As shown in Table 4, the RF classifier achieves the highest classification performance in terms of average classification accuracy which is 99.85%. The k-NN classifier stands at the second position and the SVM classifier achieves the third position with the average classification accuracy of 98.31% and 96.11%, respectively. The J48 classifier yields the lowest average classification accuracy among all tested methods. It can be seen from Table 4 that the RF classifier produces the best performance in terms of sensitivity among all classifiers and obtains the sensitivity rate of 99.79% for Set A, 99.85% for Set B, 99.96% for Set C, 99.71% for Set D, and 100% for Set E. It is observed that both k-NN and SVM classifiers yield high sensitivity rates in Sets A, B, C, and D, but they failed to correctly classify the epileptic patient during seizure activity class (Set E). Table 4 also shows that the standard deviation for every classifier is very low which indicates the consistency of the mentioned classifiers for the DP_PCA features set.


Classifiers Se (mean standard deviation)10-fold cross validation OCA
Set ASet BSet CSet DSet E(mean standard deviation)

RF99.79 0.3699.850.2499.96 0.1499.71 0.34100 0.0099.85 0.13
k-NN100 0.0099.75 0.35100 0.00100 0.0089.932.0298.31 0.32
SVM100 0.0099.85 0.34100 0.0099.87 0.2076.563.5396.11 0.61
J4895.80 1.5491.68 1.3193.52 1.7594.83 1.3395.39 1.4394.27 0.78

To provide more detailed information about how the 10-fold cross-validation system produces the classification performance regarding sensitivity and accuracy in each of the ten folds for each of the four classifiers, the classification results in each of the ten folds are provided in Figures 9 and 10. Figure 9 illustrates the classification results in terms of the Se in each of the ten folds in each class (Set) for the reported classifiers and Figure 10 shows the classification results for all classifiers in terms of the OCA in each of the ten folds. The error bars in these figures represent the standard errors.

Figure 9 presents the patterns of the Se for each class. From Figure 9(a), little fluctuation is noted in the Se patterns among the ten folds in each of the five classes for the RF classifier. These results indicate the stability and robustness of the RF classifier. From Figures 9(b) and 9(c), it is seen that the Se patterns for Set A, Set B, Set C, Set D are almost similar but the patterns for Set E is different and dramatically lower than the other patterns for both k-NN and SVM classifiers. This indicates the weakness of the k-NN and SVM classifiers for detecting the epileptic signals during seizure activity class (Set E). Figure 9(d) shows that the J48 classifier produces similar Se patterns for all Sets. It can be seen from Figure 9 that the fluctuations in the Se patterns among the different folds are negligible in each class for all classifiers. These results demonstrate the consistency of the classification methods.

Figure 10 shows the overall classification accuracies against each of the 10-folds for all classifiers. As can be seen from Figure 10, the RF classifier yields the best performance for each of the 10-folds compared to the k-NN, SVM and the J48 classifiers. It is observed that the fluctuations of the performance of the RF classifier are smaller among the different folds compared to other classifiers, indicating the stability of the RF classifier for the DP_PCA features Set. This figure also shows that the k-NN classifier produces a better performance than both SVM and J48 classifiers in each of the 10-folds. The lowest performance is obtained by the J48 classifier in each of the 10-folds.

Table 5 provides the FPR for the four classifiers in each of the ten folds for Set A, Set B, Set C, Set D, and Set E. In Table 5, the overall results of each class (Set) are also reported in terms of mean standard deviation of the FPR over a 10-fold cross-validation. It is observed from Table 5 that the RF classifier produces an overall FPR of 0.04% for Set A, 0.03% for Set B, 0.11% for Set C, 0.01% for Set D, and 0.00% for Set E, while these values are 0.18%, 1.09%, 0.36%, 0.50%, and 0.00%, respectively, for the k-NN classifier; and 1.06%, 1.63%, 1.14%, 1.07%, and 0.00%, respectively, for the SVM classifier; and 1.18%, 1.81%, 2.20%, 1.75%, and 0.29%, respectively, for the J48 classifier. The results show that in most cases the FPRs are zero in each of the folds in all classes for the RF classifier. It is also observed that the overall FPRs of the RF classifier are lower than those of the k-NN, the SVM, and the J48 classifiers in all classes.


Classifiers DatasetFPR(%)
Set ASet BSet CSet DSet E

RF1-fold0.000.110.230.000.00
2-fold0.000.000.000.000.00
3-fold0.000.230.120.120.00
4-fold0.000.000.230.000.00
5-fold0.000.000.120.000.00
6-fold0.000.000.000.000.00
7-fold0.120.000.120.000.00
8-fold0.240.000.000.000.00
9-fold0.000.000.230.000.00
10-fold0.000.000.000.000.00
Overall0.040.080.030.080.110.100.010.040.000.00

k-NN1-fold0.001.490.351.200.00
2-fold0.240.800.230.840.00
3-fold0.121.260.590.600.00
4-fold0.121.260.230.000.00
5-fold0.121.490.350.000.00
6-fold0.240.690.470.480.00
7-fold0.480.570.350.600.00
8-fold0.120.800.700.480.00
9-fold0.121.370.000.480.00
10-fold0.241.140.350.240.00
Overall0.180.131.090.340.360.200.500.360.000.00

SVM1-fold1.071.940.940.480.00
2-fold0.711.140.821.200.00
3-fold1.071.141.290.840.00
4-fold0.601.940.940.360.00
5-fold1.191.710.821.190.00
6-fold1.311.711.531.440.00
7-fold1.310.911.411.200.00
8-fold1.551.720.821.440.00
9-fold0.712.061.411.200.00
10-fold1.072.061.411.320.00
Overall1.060.311.630.421.140.291.070.380.000.00

J481-fold0.592.172.002.030.44
2-fold1.431.601.061.560.00
3-fold1.661.602.002.630.22
4-fold1.190.692.822.030.44
5-fold1.192.402.461.910.22
6-fold1.191.142.470.960.78
7-fold1.072.512.121.680.11
8-fold0.831.602.001.200.11
9-fold1.791.832.351.200.44
10-fold0.832.522.702.280.11
Overall1.18 0.381.81 0.612.20 0.501.75 0.530.29 0.38

In order to explore the best classifier for the DP_PCA features set, the performances of all four classifiers are compared in terms of kappa statistics and AUC. Figure 11 presents the AUC for the RF, k-NN, SVM, and J48 classifiers for the DP_PCA features set, separately for each of five classes and their overall AUC as well. The AUC is used as a measure for assessing the classifier performance (e.g., a higher value of the area indicates better performance of the classifier). As can be seen in Figure 11, each of the four classifiers achieve high AUC close to 1 for each class (Set), and the RF classifier produces slightly higher AUC in each class comparing to the other three classifiers. Figure 12 shows the performance of all reported classifiers in terms of kappa statistic. In this research, kappa statistics test is used to evaluate the consistency of the four classifiers: the RF, k-NN, SVM, and J48 on the DP_PCA features set. The kappa value (k) indicates the consistency of the classifier. The consistency is considered as mild if k0.2, fair if 0.21 k 0.40, moderate if 0.41 k 0.60, good if 0.61 k 0.80, and excellent if k 0.81. The maximum value of kappa is one which defines total consistency. As can be seen in Figure 12, the kappa values are very high (close to 1) for all four classifiers, and the RF classifier achieves the highest kappa value (K=0.998). From Figures 11 and 12, it is clear that the RF classifier yields better performance with the DP_PCA features set in the EEG signals classification than the other three classifiers. Therefore, The RF classifier is selected as the best classifier for the DP_PCA features set in epileptic EEG signal classification.

4.2. Comparison

Although there are many studies in the literature for epileptic EEG classification, most are restricted to the two-class classification problems dealing with the benchmark epileptic EEG data [32, 4145]. Few studies have focused on the multiclass EEG signal classification [37, 39, 40, 4648] (discussed in Section 2). To further evaluate the efficiency of the proposed method, a comparison of the proposed method with other six reported methods is provided. Table 6 provides a comparative study between the proposed method and the three reference algorithms for the same benchmark epileptic EEG dataset. This table reports the overall classification performance of the five categories of EEG signals in terms of sensitivity, specificity, and the classification accuracy. The specificity is the complement of false positive rate (100 - FPR). The highest overall classification performances among all reported methods are highlighted in italic. From Table 6, it is observed that the proposed method achieves the highest performance in each statistical parameter of each class compared to the six reference methods. The OCA of the proposed method is 99.85% while they are 99.30%, 99.28%, 99.20%, 98.05%, 97.60%, and 93.63% for methods reported in [37, 39, 40, 4648], respectively. These results indicate the proposed method outperforms all six referenced methods and improves the OCA by at least 0.55%.


MethodsDatasetsStatistical parameters (%)
Sensitivity (Se)Specificity (Sp)Overall classification accuracy (OCA)

Proposed MethodSet A99.7999.9699.85
Set B99.8599.97
Set C99.9699.90
Set D99.7199.99
Set E100.00100.00

Emigdio et al. (2016) [37] Set A95.3099.2697.60
Set B96.8398.69
Set C97.2399.53
Set D98.4099.42
Set E99.2899.80

Murugavel and Ramakrishnan (2016) [48] Set A93.2598.4293.63
Set B93.6398.36
Set C94.0098.16
Set D94.1397.17
Set E93.1399.54

Ubeyli (2010) [47] Set A98.0099.6298.05
Set B98.2599.53
Set C98.1399.37
Set D98.0099.15
Set E97.8899.87

Ubeyli (2009) [46] Set A99.2599.8499.20
Set B99.1399.81
Set C99.2599.72
Set D99.3899.62
Set E99.00100.00

Ubeyli (2008) [39] Set A99.3899.8199.30
Set B99.2599.87
Set C99.1399.78
Set D99.5099.65
Set E99.25100.00

Guler and Ubeyli (2007) [40] Set A99.2599.8499.28
Set B99.3899.84
Set C99.2599.75
Set D99.3899.65
Set E99.13100.00

5. Conclusion

This research introduces a new concept based on DP algorithm for extracting representative information from multicategory EEG signals data in the epileptic seizures identification. This study also investigated which machine leaning model (e.g., RF, k-NN, SVM and DT) is suitable for the proposed feature exaction method. The experimental results demonstrate that the proposed method is very effective and efficient for extracting distinguishable features from the epileptic EEG data. The high classification performances achieved by all reported classifiers confirm the consistency of the extracted features to detect epileptic EEG signals. The results show that the proposed RF classifier with the DP_PCA features yields the best overall performance as compared to the other classifiers. The results also indicate that our proposed method outperforms the existing methods for the same epileptic EEG database. To conclude, the DP algorithm is reliable for extracting the representative samples from the original EEG data and the RF with the proposed feature set is an effective classifier for the classification of multiclass EEG signals. A limitation of the current study is that the computational complexity of the proposed method grows as the EEG data size increases. Therefore, the proposed method may not work effectively and can take more time to process and classify very large EEG data. For future work, we plan to reduce the complexity of the proposed method by replacing the PCA and DP algorithms with existing low computational complexity techniques of PCA and DP.

Data Availability

The epileptic EEG data used to support the findings of this study are available from publicly available EEG database of Department of Epileptology, University of Bonn, Germany.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

Acknowledgments

The present paper was extracted from part of Ph.D. thesis of first author [82] and financially supported by the Australian Research Council Linkage Project (LP100200682).

References

  1. D. J. Thurman, E. Beghi, C. E. Begley et al., “Standards for epidemiologic studies and surveillance of epilepsy,” Epilepsia, vol. 52, supplement 7, pp. 2–26, 2011. View at: Publisher Site | Google Scholar
  2. R. S. Fisher, C. Acevedo, A. Arzimanoglou et al., “ILAE official report: a practical clinical definition of epilepsy,” Epilepsia, vol. 55, no. 4, pp. 475–482, 2014. View at: Publisher Site | Google Scholar
  3. A. Surguchov, I. Surgucheva, M. Sharma, and V. Singh, “New putative epigenetic mechanism of epilepsy,” Frontiers in Neurology, vol. 8, no. 3, 2017. View at: Google Scholar
  4. S. Siuly and Y. Zhang, “Medical big data: Neurological diseases diagnosis through medical data analysis,” Data Science and Engineering, vol. 1, no. 2, pp. 54–64, 2016. View at: Publisher Site | Google Scholar
  5. N. Moghim and D. W. Corne, “Predicting epileptic seizures in advance,” PLoS ONE, vol. 9, no. 6, Article ID e99334, 2014. View at: Publisher Site | Google Scholar
  6. H. Wang, G. Zhuo, Y. Zhang et al., “Analyzing EEG signal data for detection of epileptic seizure: Introducing weight on visibility graph with complex network feature,” in Proceedings of the Australasian Database Conference, pp. 56–66, 2016. View at: Google Scholar
  7. S. Siuly and Y. Li, “Designing a robust feature extraction method based on optimum allocation and principal component analysis for epileptic EEG signal classification,” Computer Methods and Programs in Biomedicine, vol. 119, no. 1, pp. 29–42, 2015. View at: Publisher Site | Google Scholar
  8. G. C. Ray, “An algorithm to separate nonstationary part of a signal using mid-prediction filter,” IEEE Transactions on Signal Processing, vol. 42, no. 9, pp. 2276–2279, 1994. View at: Publisher Site | Google Scholar
  9. S. Mukhopadhyay and G. C. Ray, “A new interpretation of nonlinear energy operator and its efficacy in spike detection,” IEEE Transactions on Biomedical Engineering, vol. 45, no. 2, pp. 180–187, 1998. View at: Publisher Site | Google Scholar
  10. S. Siuly, E. Kabir, H. Wang, and Y. Zhang, “Exploring sampling in the detection of multicategory EEG signals,” Computational and Mathematical Methods in Medicine, vol. 2015, Article ID 576437, 12 pages, 2015. View at: Publisher Site | Google Scholar
  11. F. Alçin, S. Siuly, V. Bajaj, Y. Guo, A. Sengu, Y. Zhang et al., “Multi-category EEG signal classification developing time-frequency texture features based Fisher Vector encoding method,” Neurocomputing, vol. 218, pp. 251–258, 2016. View at: Google Scholar
  12. S. Chandaka, A. Chatterjee, and S. Munshi, “Cross-correlation aided support vector machine classifier for classification of EEG signals,” Expert Systems with Applications, vol. 36, no. 2, pp. 1329–1336, 2009. View at: Publisher Site | Google Scholar
  13. S. Altunay, Z. Telatar, and O. Erogul, “Epileptic EEG detection using the linear prediction error energy,” Expert Systems with Applications, vol. 37, no. 8, pp. 5661–5665, 2010. View at: Publisher Site | Google Scholar
  14. K. Polat and S. Güneş, “Classification of epileptiform EEG using a hybrid system based on decision tree classifier and fast Fourier transform,” Applied Mathematics and Computation, vol. 187, no. 2, pp. 1017–1026, 2007. View at: Publisher Site | Google Scholar | MathSciNet
  15. H. Ocak, “Automatic detection of epileptic seizures in EEG using discrete wavelet transform and approximate entropy,” Expert Systems with Applications, vol. 36, no. 2, pp. 2027–2036, 2009. View at: Publisher Site | Google Scholar
  16. H. Adeli, S. Ghosh-Dastidar, and N. Dadmehr, “A wavelet-chaos methodology for analysis of EEGs and EEG subbands to detect seizure and epilepsy,” IEEE Transactions on Biomedical Engineering, vol. 54, no. 2, pp. 205–211, 2007. View at: Publisher Site | Google Scholar
  17. A. R. Hassan, S. Siuly, and Y. Zhang, “Epileptic seizure detection in EEG signals using tunable-Q factor wavelet transform and bootstrap aggregating,” Computer Methods and Programs in Biomedicine, vol. 137, pp. 247–259, 2016. View at: Publisher Site | Google Scholar
  18. W. Wu, X. Gao, B. Hong, and S. Gao, “Classifying single trial EEG during motor imagery by iterative spatio-spectral pattern learning (ISSPL),” IEEE Transactions on Biomedical Engineering, vol. 55, no. 6, pp. 1733–1743, 2008. View at: Publisher Site | Google Scholar
  19. S. Li, W. Zhou, Q. Yuan, S. Geng, and D. Cai, “Feature extraction and recognition of ictal EEG using EMD and SVM,” Computers in Biology and Medicine, vol. 43, no. 7, pp. 807–816, 2013. View at: Publisher Site | Google Scholar
  20. N. F. Güler, E. D. Übeyli, and I. Güler, “Recurrent neural networks employing Lyapunov exponents for EEG signals classification,” Expert Systems with Applications, vol. 29, no. 3, pp. 506–514, 2005. View at: Publisher Site | Google Scholar
  21. S. Ghosh-Dastidar, H. Adeli, and N. Dadmehr, “Mixed-band wavelet-chaos-neural network methodology for epilepsy and epileptic seizure detection,” IEEE Transactions on Biomedical Engineering, vol. 54, no. 9, pp. 1545–1551, 2007. View at: Publisher Site | Google Scholar
  22. V. Srinivasan, C. Eswaran, and N. Sriraam, “Approximate entropy-based epileptic EEG detection using artificial neural networks,” IEEE Transactions on Information Technology in Biomedicine, vol. 11, no. 3, pp. 288–295, 2007. View at: Publisher Site | Google Scholar
  23. M.-J. Chuang, S.-T. Wu, S.-H. Tang et al., “The HDAC inhibitor LBH589 induces ERK-dependent prometaphase arrest in prostate cancer via HDAC6 inactivation and down-regulation,” PLoS ONE, vol. 8, no. 9, Article ID e73401, 2013. View at: Publisher Site | Google Scholar
  24. S. Siuly, Y. Li, and P. P. Wen, “Clustering technique-based least square support vector machine for EEG signal classification,” Computer Methods and Programs in Biomedicine, vol. 104, no. 3, pp. 358–372, 2011. View at: Publisher Site | Google Scholar
  25. Siuly, Y. Li, and P. Wen, “EEG signal classification based on simple random sampling technique with least square support vector machine,” International Journal of Biomedical Engineering and Technology, vol. 7, no. 4, pp. 390–409, 2011. View at: Publisher Site | Google Scholar
  26. S. Supriya, S. Siuly, H. Wang, J. Cao, and Y. Zhang, “Weighted Visibility Graph with Complex Network Features in the Detection of Epilepsy,” IEEE Access, vol. 4, pp. 6554–6566, 2016. View at: Publisher Site | Google Scholar
  27. S. Supriya, S. Siuly, and Y. Zhang, “Automatic epilepsy detection from EEG introducing a new edge weight method in the complex network,” IEEE Electronics Letters, vol. 52, no. 17, pp. 1430–1432, 2016. View at: Publisher Site | Google Scholar
  28. E. Kabir, Siuly, and Y. Zhang, “Epileptic seizure detection from EEG signals using logistic model trees,” Brain Informatics, vol. 3, no. 2, pp. 93–100, 2016. View at: Publisher Site | Google Scholar
  29. U. R. Acharya, S. V. Sree, G. Swapna, R. J. Martis, and J. S. Suri, “Automated EEG analysis of epilepsy: a review,” Knowledge-Based Systems, vol. 45, pp. 147–165, 2013. View at: Publisher Site | Google Scholar
  30. Y. Song, J. Crowcroft, and J. Zhang, “Automatic epileptic seizure detection in EEGs based on optimized sample entropy and extreme learning machine,” Journal of Neuroscience Methods, vol. 210, no. 2, pp. 132–146, 2012. View at: Publisher Site | Google Scholar
  31. A. Subasi and E. Erçelebi, “Classification of EEG signals using neural network and logistic regression,” Computer Methods and Programs in Biomedicine, vol. 78, no. 2, pp. 87–99, 2005. View at: Publisher Site | Google Scholar
  32. U. Orhan, M. Hekim, and M. Ozer, “EEG signals classification using the K-means clustering and a multilayer perceptron neural network model,” Expert Systems with Applications, vol. 38, no. 10, pp. 13475–13481, 2011. View at: Publisher Site | Google Scholar
  33. Y. Kumar, M. L. Dewal, and R. S. Anand, “Epileptic seizure detection using DWT based fuzzy approximate entropy and support vector machine,” Neurocomputing, vol. 133, no. 8, pp. 271–279, 2014. View at: Publisher Site | Google Scholar
  34. N. Nicolaou and J. Georgiou, “Detection of epileptic electroencephalogram based on permutation entropy and support vector machines,” Expert Systems with Applications, vol. 39, no. 1, pp. 202–209, 2012. View at: Publisher Site | Google Scholar
  35. Z. Iscan, Z. Dokur, and T. Demiralp, “Classification of electroencephalogram signals with combined time and frequency features,” Expert Systems with Applications, vol. 38, no. 8, pp. 10499–10505, 2011. View at: Publisher Site | Google Scholar
  36. S. Xie and S. Krishnan, “Dynamic principal component analysis with nonoverlapping moving window and its applications to epileptic EEG classification,” The Scientific World Journal, vol. 2014, Article ID 419308, 10 pages, 2014. View at: Publisher Site | Google Scholar
  37. E. Z-Flores, L. Trujillo, A. Sotelo, P. Legrand, and L. N. Coria, “Regularity and Matching Pursuit feature extraction for the detection of epileptic seizures,” Journal of Neuroscience Methods, vol. 266, pp. 107–125, 2016. View at: Publisher Site | Google Scholar
  38. S. Siuly and Y. Li, “Discriminating the brain activities for brain–computer interface applications through the optimal allocation-based approach,” Neural Computing and Applications, vol. 26, no. 4, pp. 799–811, 2015. View at: Publisher Site | Google Scholar
  39. E. D. Übeyli, “Analysis of EEG signals by combining eigenvector methods and multiclass support vector machines,” Computers in Biology and Medicine, vol. 38, no. 1, pp. 14–22, 2008. View at: Publisher Site | Google Scholar
  40. I. Guler and E. D. Ubeyli, “Multiclass support vector machines for EEG-signals classification,” IEEE Transactions on Information Technology in Biomedicine, vol. 11, no. 2, pp. 117–126, 2007. View at: Publisher Site | Google Scholar
  41. K. Natarajan, U. R. Acharya, F. Alias, T. Tiboleng, and S. K. Puthusserypady, “Nonlinear analysis of EEG signals at different mental states,” Biomedical Engineering Online, vol. 3, no. 1, article 7, p. 1, 2004. View at: Publisher Site | Google Scholar
  42. U. R. Acharya, F. Molinari, S. V. Sree, S. Chattopadhyay, K. H. Ng, and J. S. Suri, “Automated diagnosis of epileptic EEG using entropies,” Biomedical Signal Processing and Control, vol. 7, no. 4, pp. 401–408, 2012. View at: Publisher Site | Google Scholar
  43. U. R. Acharya, S. V. Sree, P. C. A. Ang, R. Yanti, and J. S. Suri, “Application of non-linear and wavelet based features for the automated identification of epileptic EEG signals,” International Journal of Neural Systems, vol. 22, no. 2, Article ID 1250002, pp. 565–579, 2012. View at: Publisher Site | Google Scholar
  44. M. Niknazar, S. R. Mousavi, B. Vosoughi Vahdat, and M. Sayyah, “A new framework based on recurrence quantification analysis for epileptic seizure detection,” IEEE Journal of Biomedical and Health Informatics, vol. 17, no. 3, pp. 572–578, 2013. View at: Publisher Site | Google Scholar
  45. G. Chen, “Automatic EEG seizure detection using dual-tree complex wavelet-Fourier features,” Expert Systems with Applications, vol. 41, no. 5, pp. 2391–2394, 2014. View at: Publisher Site | Google Scholar
  46. E. D. Übeyli, “Decision support systems for time-varying biomedical signals: EEG signals classification,” Expert Systems with Applications, vol. 36, no. 2, pp. 2275–2284, 2009. View at: Publisher Site | Google Scholar
  47. E. D. Übeyli, “Lyapunov exponents/probabilistic neural networks for analysis of EEG signals,” Expert Systems with Applications, vol. 37, no. 2, pp. 985–992, 2010. View at: Publisher Site | Google Scholar
  48. A. S. M. Murugavel and S. Ramakrishnan, “Hierarchical multi-class SVM with ELM kernel for epileptic EEG signal classification,” Medical & Biological Engineering & Computing, vol. 54, no. 1, pp. 149–161, 2016. View at: Publisher Site | Google Scholar
  49. D. H. Douglas and T. K. Peucker, “Algorithms for the reduction of the number of points required to represent a digitized line or its caricature,” Cartographica: The International Journal for Geographic Information and Geovisualization, vol. 10, no. 2, pp. 112–122, 1973. View at: Google Scholar
  50. U. C. Benz, P. Hofmann, G. Willhauck, I. Lingenfelder, and M. Heynen, “Multi-resolution, object-oriented fuzzy analysis of remote sensing data for GIS-ready information,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 58, no. 3-4, pp. 239–258, 2004. View at: Publisher Site | Google Scholar
  51. A. Orzan, A. Bousseau, P. Barla, H. Winnemöller, J. Thollot, and D. Salesin, “Diffusion curves: A vector representation for smooth-shaded images,” Communications of the ACM, vol. 56, no. 7, pp. 101–109, 2013. View at: Google Scholar
  52. K. Zhang, J. Yan, and S.-C. Chen, “Automatic construction of building footprints from airborne LIDAR data,” IEEE Transactions on Geoscience and Remote Sensing, vol. 44, no. 9, pp. 2523–2533, 2006. View at: Publisher Site | Google Scholar
  53. J. A. García and J. Fdez-Valdivia, “Boundary simplification in cartography preserving the characteristics of the shape features,” Computers & Geosciences, vol. 20, no. 3, pp. 349–368, 1994. View at: Publisher Site | Google Scholar
  54. B. Boucheham, Y. Ferdi, and M. C. Batouche, “Recursive versus sequential multiple error measures reduction: A curve simplification approach to ECG data compression,” Computer Methods and Programs in Biomedicine, vol. 81, no. 2, pp. 162–173, 2006. View at: Publisher Site | Google Scholar
  55. B. Boucheham, Y. Ferdi, and M. C. Batouche, “Piecewise linear correction of ECG baseline wander: A curve simplification approach,” Computer Methods and Programs in Biomedicine, vol. 78, no. 1, pp. 1–10, 2005. View at: Publisher Site | Google Scholar
  56. B. Boucheham, “ShaLTeRR: A contribution to short and long-term redundancy reduction in digital signals,” Signal Processing, vol. 87, no. 10, pp. 2336–2347, 2007. View at: Publisher Site | Google Scholar
  57. N. Hazarika, J. Z. Chen, A. C. Tsoi, and A. Sergejew, “Classification of EEG signals using the wavelet transform,” Signal Processing, vol. 59, no. 1, pp. 61–72, 1997. View at: Publisher Site | Google Scholar
  58. H. Adelia, Z. Zhoub, and N. Dadmehrc, “Analysis of EEG records in an epileptic patient using wavelet transform,” Journal of Neuroscience Methods, vol. 123, no. 1, pp. 69–87, 2003. View at: Publisher Site | Google Scholar
  59. R. Sharma and R. B. Pachori, “Classification of epileptic seizures in EEG signals based on phase space representation of intrinsic mode functions,” Expert Systems with Applications, vol. 42, no. 3, pp. 1106–1117, 2015. View at: Publisher Site | Google Scholar
  60. R. Zarei, J. He, S. Siuly, and Y. Zhang, “A PCA aided cross-covariance scheme for discriminative feature extraction from EEG signals,” Computer Methods and Programs in Biomedicine, vol. 146, pp. 47–57, 2017. View at: Publisher Site | Google Scholar
  61. R. G. Andrzejak, K. Lehnertz, F. Mormann, C. Rieke, P. David, and C. E. Elger, “Indications of nonlinear deterministic and finite-dimensional structures in time series of brain electrical activity: dependence on recording region and brain state,” Physical Review E: Statistical, Nonlinear, and Soft Matter Physics, vol. 64, no. 6, Article ID 061907, 2001. View at: Google Scholar
  62. EEG time series (epileptic EEG data), http://www.meb.uni-bonn.de/epileptologie/science/physik/eegdata.html, 2017.
  63. R. O. Schmidt, “Multiple emitter location and signal parameter estimation,” IEEE Transactions on Antennas and Propagation, vol. 34, no. 3, pp. 276–280, 1986. View at: Publisher Site | Google Scholar
  64. V. F. Pisarenko, “The retrieval of harmonics from a covariance function,” Geophysical Journal International, vol. 33, pp. 347–366, 1973. View at: Publisher Site | Google Scholar
  65. R. Kumaresan and D. W. Tufts, “Estimating the angles of arrival of multiple plane waves,” IEEE Transactions on Aerospace and Electronic Systems, vol. 19, no. 1, pp. 134–139, 1983. View at: Google Scholar
  66. R. O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification, John Wiley Sons, 2001. View at: MathSciNet
  67. L. I. Smith, “A tutorial on principal components analysis,” Cornell University, USA, vol. 51, p. 52, 2002. View at: Google Scholar
  68. R. Zarei, J. He, G. Huang, and Y. Zhang, “Effective and efficient detection of premature ventricular contractions based on variation of principal directions,” Digital Signal Processing, vol. 50, pp. 93–102, 2016. View at: Publisher Site | Google Scholar
  69. L. Breiman, “Random forest,” Machine Learning, vol. 45, no. 1, pp. 5–32, 2001. View at: Publisher Site | Google Scholar | MathSciNet
  70. T. M. Cover and P. E. Hart, “Nearest neighbor pattern classification,” IEEE Transactions on Information Theory, vol. 13, no. 1, pp. 21–27, 1967. View at: Publisher Site | Google Scholar
  71. Y. Song, J. Huang, D. Zhou, H. Zha, and C. L. Giles, “Iknn: Informative k-nearest neighbor pattern classification,” in Proceedings of the European Conference on Principles of Data Mining and Knowledge Discovery, pp. 248–264, Springer, 2007. View at: Google Scholar
  72. J. Han, J. Pei, and M. Kamber, Data Mining: Concepts and Techniques, Elsevier, 2011.
  73. V. N. Vapnik and V. Vapnik, Statistical Learning Theory, vol. 1, Wiley, New York, NY, USA, 1998.
  74. R. K. Begg, M. Palaniswami, and B. Owen, “Support vector machines for automated gait classification,” IEEE Transactions on Biomedical Engineering, vol. 52, no. 5, pp. 828–838, 2005. View at: Publisher Site | Google Scholar
  75. X. Yin, B. W.-H. Ng, B. M. Fischer, B. Ferguson, and D. Abbott, “Support vector machine applications in terahertz pulsed signals feature sets,” IEEE Sensors Journal, vol. 7, no. 12, pp. 1597–1607, 2007. View at: Publisher Site | Google Scholar
  76. J. R. Quinlan, C4. 5: Programs for Machine Learning, Elsevier, 2014.
  77. I. H. Witten and E. Frank, Data Mining: Practical Machine Learning Tools and Techniques, Morgan Kaufmann, 2005.
  78. E. Frank, M. Hall, G. Holmes et al., “Weka-a machine learning workbench for data mining,” in Data Mining and Knowledge Discovery Handbook, vol. 23, pp. 1269–1277, Springer, 2009. View at: Google Scholar | MathSciNet
  79. G. Ilczuk, R. Mlynarski, A. Wakulicz-Deja, A. Drzewiecka, and W. Kargul, “Rough set techniques for medical diagnosis systems,” in Proceedings of the Computers in Cardiology, pp. 837–840, IEEE, September 2005. View at: Google Scholar
  80. C.-C. Chang and C.-J. Lin, “LIBSVM: a library for support vector machines,” ACM Transactions on Intelligent Systems and Technology (TIST), vol. 2, no. 3, p. 27, 2011. View at: Google Scholar
  81. K. Revett, M. Szczuka, P. Jahankhani, and V. Kodogiannis, “Attribute selection for EEG signal classification using rough sets and neural networks,” in Proceedings of the International Conference on Rough Sets and Current Trends in Computing, pp. 408–417, Springer, 2006. View at: Publisher Site | Google Scholar | MathSciNet
  82. R. Zarei, Developing enhanced classification methods for ecg and eeg signals [PhD Thesis], Victoria University, 2017.

Copyright © 2019 Roozbeh Zarei et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views933
Downloads431
Citations

Related articles